JFIFxxC      C  " }!1AQa"q2#BR$3br %&'()*456789:CDEFGHIJSTUVWXYZcdefghijstuvwxyz w!1AQaq"2B #3RbrJFIFxxC      C  " }!1AQa"q2#BR$3br %&'()*456789:CDEFGHIJSTUVWXYZcdefghijstuvwxyz w!1AQaq"2B #3Rbrsite-packages/pip-19.3.1.dist-info/LICENSE.txt000064400000002102151464452710014325 0ustar00Copyright (c) 2008-2019 The pip developers (see AUTHORS.txt file) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. site-packages/pip-19.3.1.dist-info/WHEEL000064400000000156151464452710013300 0ustar00Wheel-Version: 1.0 Generator: bdist_wheel (0.33.6) Root-Is-Purelib: true Tag: py2-none-any Tag: py3-none-any site-packages/pip-19.3.1.dist-info/METADATA000064400000006255151464452710013622 0ustar00Metadata-Version: 2.1 Name: pip Version: 19.3.1 Summary: The PyPA recommended tool for installing Python packages. Home-page: https://pip.pypa.io/ Author: The pip developers Author-email: pypa-dev@groups.google.com License: MIT Keywords: distutils easy_install egg setuptools wheel virtualenv Platform: UNKNOWN Classifier: Development Status :: 5 - Production/Stable Classifier: Intended Audience :: Developers Classifier: License :: OSI Approved :: MIT License Classifier: Topic :: Software Development :: Build Tools Classifier: Programming Language :: Python Classifier: Programming Language :: Python :: 2 Classifier: Programming Language :: Python :: 2.7 Classifier: Programming Language :: Python :: 3 Classifier: Programming Language :: Python :: 3.5 Classifier: Programming Language :: Python :: 3.6 Classifier: Programming Language :: Python :: 3.7 Classifier: Programming Language :: Python :: 3.8 Classifier: Programming Language :: Python :: Implementation :: CPython Classifier: Programming Language :: Python :: Implementation :: PyPy Requires-Python: >=2.7,!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.* pip - The Python Package Installer ================================== .. image:: https://img.shields.io/pypi/v/pip.svg :target: https://pypi.org/project/pip/ .. image:: https://readthedocs.org/projects/pip/badge/?version=latest :target: https://pip.pypa.io/en/latest pip is the `package installer`_ for Python. You can use pip to install packages from the `Python Package Index`_ and other indexes. Please take a look at our documentation for how to install and use pip: * `Installation`_ * `Usage`_ Updates are released regularly, with a new version every 3 months. More details can be found in our documentation: * `Release notes`_ * `Release process`_ If you find bugs, need help, or want to talk to the developers please use our mailing lists or chat rooms: * `Issue tracking`_ * `Discourse channel`_ * `User IRC`_ If you want to get involved head over to GitHub to get the source code, look at our development documentation and feel free to jump on the developer mailing lists and chat rooms: * `GitHub page`_ * `Dev documentation`_ * `Dev mailing list`_ * `Dev IRC`_ Code of Conduct --------------- Everyone interacting in the pip project's codebases, issue trackers, chat rooms, and mailing lists is expected to follow the `PyPA Code of Conduct`_. .. _package installer: https://packaging.python.org/en/latest/current/ .. _Python Package Index: https://pypi.org .. _Installation: https://pip.pypa.io/en/stable/installing.html .. _Usage: https://pip.pypa.io/en/stable/ .. _Release notes: https://pip.pypa.io/en/stable/news.html .. _Release process: https://pip.pypa.io/en/latest/development/release-process/ .. _GitHub page: https://github.com/pypa/pip .. _Dev documentation: https://pip.pypa.io/en/latest/development .. _Issue tracking: https://github.com/pypa/pip/issues .. _Discourse channel: https://discuss.python.org/c/packaging .. _Dev mailing list: https://groups.google.com/forum/#!forum/pypa-dev .. _User IRC: https://webchat.freenode.net/?channels=%23pypa .. _Dev IRC: https://webchat.freenode.net/?channels=%23pypa-dev .. _PyPA Code of Conduct: https://www.pypa.io/en/latest/code-of-conduct/ site-packages/pip-19.3.1.dist-info/RECORD000064400000142732151464452710013421 0ustar00../../../bin/pip,sha256=isMYezCDrq79FYcQQqNmVGeOh26HvQz3wntdy5zsiUc,219 ../../../bin/pip3,sha256=isMYezCDrq79FYcQQqNmVGeOh26HvQz3wntdy5zsiUc,219 ../../../bin/pip3.8,sha256=isMYezCDrq79FYcQQqNmVGeOh26HvQz3wntdy5zsiUc,219 pip-19.3.1.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4 pip-19.3.1.dist-info/LICENSE.txt,sha256=W6Ifuwlk-TatfRU2LR7W1JMcyMj5_y1NkRkOEJvnRDE,1090 pip-19.3.1.dist-info/METADATA,sha256=3VTJN81wKoCumM54Ay7Je-xOiIkScmdYQu9WD43uK5o,3245 pip-19.3.1.dist-info/RECORD,, pip-19.3.1.dist-info/WHEEL,sha256=8zNYZbwQSXoB9IfXOjPfeNwvAsALAjffgk27FqvCWbo,110 pip-19.3.1.dist-info/entry_points.txt,sha256=tbYijp8dTAsoGzseo-3RkvhZNzpr8DW4lFJdKFuARUk,113 pip-19.3.1.dist-info/top_level.txt,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4 pip/__init__.py,sha256=3XVcbBDOOau8B2kbTs7mairuEe_dyGELVtflVt0LMVs,23 pip/__main__.py,sha256=JhJZiiWUmlPTwFOZ7_L68hWG4nUVQTDTtwc1o_yWdWw,628 pip/__pycache__/__init__.cpython-38.pyc,, pip/__pycache__/__main__.cpython-38.pyc,, pip/_internal/__init__.py,sha256=4lqJ4waZxiP4QFg34znQ3j5zdyDsrWZkERPeLAs7lT8,80 pip/_internal/__pycache__/__init__.cpython-38.pyc,, pip/_internal/__pycache__/build_env.cpython-38.pyc,, pip/_internal/__pycache__/cache.cpython-38.pyc,, pip/_internal/__pycache__/collector.cpython-38.pyc,, pip/_internal/__pycache__/configuration.cpython-38.pyc,, pip/_internal/__pycache__/download.cpython-38.pyc,, pip/_internal/__pycache__/exceptions.cpython-38.pyc,, pip/_internal/__pycache__/index.cpython-38.pyc,, pip/_internal/__pycache__/legacy_resolve.cpython-38.pyc,, pip/_internal/__pycache__/locations.cpython-38.pyc,, pip/_internal/__pycache__/main.cpython-38.pyc,, pip/_internal/__pycache__/pep425tags.cpython-38.pyc,, pip/_internal/__pycache__/pyproject.cpython-38.pyc,, pip/_internal/__pycache__/self_outdated_check.cpython-38.pyc,, pip/_internal/__pycache__/wheel.cpython-38.pyc,, pip/_internal/build_env.py,sha256=h8_IrSllK1aEv4cmoO9G7l5XDrwMeq2aADy6EWqbEKs,7517 pip/_internal/cache.py,sha256=tv5gyBOh4gx8my0ikxCEHzg92Y5D8mCbcE4lq9rQLhw,8371 pip/_internal/cli/__init__.py,sha256=FkHBgpxxb-_gd6r1FjnNhfMOzAUYyXoXKJ6abijfcFU,132 pip/_internal/cli/__pycache__/__init__.cpython-38.pyc,, pip/_internal/cli/__pycache__/autocompletion.cpython-38.pyc,, pip/_internal/cli/__pycache__/base_command.cpython-38.pyc,, pip/_internal/cli/__pycache__/cmdoptions.cpython-38.pyc,, pip/_internal/cli/__pycache__/command_context.cpython-38.pyc,, pip/_internal/cli/__pycache__/main_parser.cpython-38.pyc,, pip/_internal/cli/__pycache__/parser.cpython-38.pyc,, pip/_internal/cli/__pycache__/req_command.cpython-38.pyc,, pip/_internal/cli/__pycache__/status_codes.cpython-38.pyc,, pip/_internal/cli/autocompletion.py,sha256=KLBLfPkMKgkl6FS6RGlYlgRmdxFiwvdefv3Ywt5JqzM,6169 pip/_internal/cli/base_command.py,sha256=Xg0kP33AsjMYqXmFoZ-HX6-hRM5QaKDXAJRXLO03IVU,6504 pip/_internal/cli/cmdoptions.py,sha256=y6x_kbx-obAF2i6rO0cZR2HOi5SC31LKxl0ZjQV6uOs,26839 pip/_internal/cli/command_context.py,sha256=F0hZ0Xm8NZAgcOpl5J46OpjvAhIgx5P1nZHGaAAxlmc,796 pip/_internal/cli/main_parser.py,sha256=W9OWeryh7ZkqELohaFh0Ko9sB98ZkSeDmnYbOZ1imBc,2819 pip/_internal/cli/parser.py,sha256=O9djTuYQuSfObiY-NU6p4MJCfWsRUnDpE2YGA_fwols,9487 pip/_internal/cli/req_command.py,sha256=J1fTDjVZAExZDMBEqwJQ-NgNlab1g8zE1nZND-6TnP0,11363 pip/_internal/cli/status_codes.py,sha256=F6uDG6Gj7RNKQJUDnd87QKqI16Us-t-B0wPF_4QMpWc,156 pip/_internal/collector.py,sha256=zGSX67z_CN8uGIxxseKTIV2TC1QN1AE9mwBPvodTxak,18007 pip/_internal/commands/__init__.py,sha256=uTSj58QlrSKeXqCUSdL-eAf_APzx5BHy1ABxb0j5ZNE,3714 pip/_internal/commands/__pycache__/__init__.cpython-38.pyc,, pip/_internal/commands/__pycache__/check.cpython-38.pyc,, pip/_internal/commands/__pycache__/completion.cpython-38.pyc,, pip/_internal/commands/__pycache__/configuration.cpython-38.pyc,, pip/_internal/commands/__pycache__/debug.cpython-38.pyc,, pip/_internal/commands/__pycache__/download.cpython-38.pyc,, pip/_internal/commands/__pycache__/freeze.cpython-38.pyc,, pip/_internal/commands/__pycache__/hash.cpython-38.pyc,, pip/_internal/commands/__pycache__/help.cpython-38.pyc,, pip/_internal/commands/__pycache__/install.cpython-38.pyc,, pip/_internal/commands/__pycache__/list.cpython-38.pyc,, pip/_internal/commands/__pycache__/search.cpython-38.pyc,, pip/_internal/commands/__pycache__/show.cpython-38.pyc,, pip/_internal/commands/__pycache__/uninstall.cpython-38.pyc,, pip/_internal/commands/__pycache__/wheel.cpython-38.pyc,, pip/_internal/commands/check.py,sha256=mgLNYT3bd6Kmynwh4zzcBmVlFZ-urMo40jTgk6U405E,1505 pip/_internal/commands/completion.py,sha256=UFQvq0Q4_B96z1bvnQyMOq82aPSu05RejbLmqeTZjC0,2975 pip/_internal/commands/configuration.py,sha256=6riioZjMhsNSEct7dE-X8SobGodk3WERKJvuyjBje4Q,7226 pip/_internal/commands/debug.py,sha256=m9Ap_dd5jd7W6FbE9_t5495SAEvUC1vnDZ2m-TYXsQ4,3394 pip/_internal/commands/download.py,sha256=J6hxq7dO8Ld751WYueAdC18xiGhi27B6lrdOefFxwZo,5577 pip/_internal/commands/freeze.py,sha256=G9I_yoBHlpWLX1qItsWNOmmqc8ET7pekFybdbV333d4,3464 pip/_internal/commands/hash.py,sha256=47teimfAPhpkaVbSDaafck51BT3XXYuL83lAqc5lOcE,1735 pip/_internal/commands/help.py,sha256=Nhecq--ydFn80Gm1Zvbf9943EcRJfO0TnXUhsF0RO7s,1181 pip/_internal/commands/install.py,sha256=J6QUYQHD4XO-eTuC3wl3Mv1xYp3DxU1AWioHwnWIXwU,24559 pip/_internal/commands/list.py,sha256=wQijUCcLQmME_ryAk1EqjBM4CNfF8FXt5XGC5o3lQjc,10548 pip/_internal/commands/search.py,sha256=7Il8nKZ9mM7qF5jlnBoPvSIFY9f-0-5IbYoX3miTuZY,5148 pip/_internal/commands/show.py,sha256=Vzsj2oX0JBl94MPyF3LV8YoMcigl8B2UsMM8zp0pH2s,6792 pip/_internal/commands/uninstall.py,sha256=8mldFbrQecSoWDZRqxBgJkrlvx6Y9Iy7cs-2BIgtXt4,2983 pip/_internal/commands/wheel.py,sha256=cDwoMefUjOdLqE29rQGYwvdzwvsV0j5Ac7pCszbqaK8,6408 pip/_internal/configuration.py,sha256=ug70zCMs2OYEgef_6uebyXWDknbY-hz5cSaL9CpLYkI,14218 pip/_internal/distributions/__init__.py,sha256=plnByCzRAZJL98kuchZ4PoRKxseDWM7wBfjvb2_fMw8,967 pip/_internal/distributions/__pycache__/__init__.cpython-38.pyc,, pip/_internal/distributions/__pycache__/base.cpython-38.pyc,, pip/_internal/distributions/__pycache__/installed.cpython-38.pyc,, pip/_internal/distributions/__pycache__/wheel.cpython-38.pyc,, pip/_internal/distributions/base.py,sha256=91UqVXWnIyQPpIQed4nAuYWkGY665aYIV1EjS-CkBEg,1108 pip/_internal/distributions/installed.py,sha256=4bZq8NqgXgPqX-yTz94WzTRvobjwDnxVnFCL24FjAWA,542 pip/_internal/distributions/source/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_internal/distributions/source/__pycache__/__init__.cpython-38.pyc,, pip/_internal/distributions/source/__pycache__/legacy.cpython-38.pyc,, pip/_internal/distributions/source/legacy.py,sha256=8QbAup0CBA5iPsSAJgxZyz1tQxO1C6hvfFpTlCJnyKE,3941 pip/_internal/distributions/wheel.py,sha256=0hkgI9NpkT4bCwA9RTsX2NJcz9MJyPhlh3f7Z3529sY,616 pip/_internal/download.py,sha256=IdIOCKfLSqfAgq34SCOyMKJHnJOr81JF-EeGB_faLHQ,20805 pip/_internal/exceptions.py,sha256=6YRuwXAK6F1iyUWKIkCIpWWN2khkAn1sZOgrFA9S8Ro,10247 pip/_internal/index.py,sha256=P5WnrP8zyBK29EFDCcIsznQquTNn4ZL-cYP8L-NbfPU,36887 pip/_internal/legacy_resolve.py,sha256=X9Vh3sRLMKI7HnGZ2UuTO_liClsSPwRA4m9D7L5osxY,17234 pip/_internal/locations.py,sha256=n6KALLEJN-L99QFHPNdJl3nMktwcpmqMyyo6EXZoCL0,5414 pip/_internal/main.py,sha256=ikuDY1c82JZivOvnSXJZN31IeeZF_BkNpgZZ90p6iJ4,1359 pip/_internal/models/__init__.py,sha256=3DHUd_qxpPozfzouoqa9g9ts1Czr5qaHfFxbnxriepM,63 pip/_internal/models/__pycache__/__init__.cpython-38.pyc,, pip/_internal/models/__pycache__/candidate.cpython-38.pyc,, pip/_internal/models/__pycache__/format_control.cpython-38.pyc,, pip/_internal/models/__pycache__/index.cpython-38.pyc,, pip/_internal/models/__pycache__/link.cpython-38.pyc,, pip/_internal/models/__pycache__/search_scope.cpython-38.pyc,, pip/_internal/models/__pycache__/selection_prefs.cpython-38.pyc,, pip/_internal/models/__pycache__/target_python.cpython-38.pyc,, pip/_internal/models/candidate.py,sha256=S5ipn5ZvnWFX_A84OGrtc4DAZTJM7GSnIyDIA8PwZWw,1277 pip/_internal/models/format_control.py,sha256=bMcS0iTVLBXrYRDkGeGqYE3pTQUR96T7cGyMtLeH0Fs,2592 pip/_internal/models/index.py,sha256=K59A8-hVhBM20Xkahr4dTwP7OjkJyEqXH11UwHFVgqM,1060 pip/_internal/models/link.py,sha256=MFYBc2ko2B--zFyk6JBnorkAdFpJ8lwozz-64YBdDO0,6860 pip/_internal/models/search_scope.py,sha256=25LTMmqHvKxCpOsEH1qcb0Id-hqATsg2SVIjcxY702o,3971 pip/_internal/models/selection_prefs.py,sha256=rPeif2KKjhTPXeMoQYffjqh10oWpXhdkxRDaPT1HO8k,1908 pip/_internal/models/target_python.py,sha256=d66ljdpZZtAAQsuOytiZ7yq6spCa8GOmz5Vf7uoVZT0,3820 pip/_internal/network/__init__.py,sha256=jf6Tt5nV_7zkARBrKojIXItgejvoegVJVKUbhAa5Ioc,50 pip/_internal/network/__pycache__/__init__.cpython-38.pyc,, pip/_internal/network/__pycache__/auth.cpython-38.pyc,, pip/_internal/network/__pycache__/cache.cpython-38.pyc,, pip/_internal/network/__pycache__/session.cpython-38.pyc,, pip/_internal/network/__pycache__/xmlrpc.cpython-38.pyc,, pip/_internal/network/auth.py,sha256=K3G1ukKb3PiH8w_UnpXTz8qQsTULO-qdbfOE9zTo1fE,11119 pip/_internal/network/cache.py,sha256=WJafQhl9_gSxvEMvYN-ecsijGscKCVPsb09sEUbwL-E,2233 pip/_internal/network/session.py,sha256=m2E-pkpE3nWVswt2JzKP9NANwfhf38SjDOUor0cxqd8,15842 pip/_internal/network/xmlrpc.py,sha256=AL115M3vFJ8xiHVJneb8Hi0ZFeRvdPhblC89w25OG5s,1597 pip/_internal/operations/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_internal/operations/__pycache__/__init__.cpython-38.pyc,, pip/_internal/operations/__pycache__/check.cpython-38.pyc,, pip/_internal/operations/__pycache__/freeze.cpython-38.pyc,, pip/_internal/operations/__pycache__/generate_metadata.cpython-38.pyc,, pip/_internal/operations/__pycache__/prepare.cpython-38.pyc,, pip/_internal/operations/check.py,sha256=SG2AuZ4NFwUPAWo4OQOotKCn6OUxjkAgBdehNAo_2sM,5354 pip/_internal/operations/freeze.py,sha256=MeHazCBo6DF0B2ay0t0CqRRfR4bETqKB-ANv4yIq4Bg,9826 pip/_internal/operations/generate_metadata.py,sha256=oHJtWo0QpOt5xt5xQTfOEVNOpaPy9E0lhMCIE6Upidw,4699 pip/_internal/operations/prepare.py,sha256=_TNie4ORlITMh04dkEQcBVDFkcz5FW67VjUfeJDRUcM,11279 pip/_internal/pep425tags.py,sha256=mCZA3yvHjZfQGnQti2sNfzrinhLmDp0o9Fwb826Edj4,15941 pip/_internal/pyproject.py,sha256=98g1oHmpdqraOqAJuoCDr2XP3QQuN5oEuoD0fWDtVhE,6490 pip/_internal/req/__init__.py,sha256=iR73_MGlD4N2Dg86t2kooz7okpZgpAuZUSROrR9KkkE,2467 pip/_internal/req/__pycache__/__init__.cpython-38.pyc,, pip/_internal/req/__pycache__/constructors.cpython-38.pyc,, pip/_internal/req/__pycache__/req_file.cpython-38.pyc,, pip/_internal/req/__pycache__/req_install.cpython-38.pyc,, pip/_internal/req/__pycache__/req_set.cpython-38.pyc,, pip/_internal/req/__pycache__/req_tracker.cpython-38.pyc,, pip/_internal/req/__pycache__/req_uninstall.cpython-38.pyc,, pip/_internal/req/constructors.py,sha256=DcUoVBQabogW6FoTxo_7FApvYTBk5LmiTtLOSi9Wh7Y,14388 pip/_internal/req/req_file.py,sha256=Ip4KV67FLLexYq9xVGgzjbyjIQCP5W-RmkVZWFWV5AY,14294 pip/_internal/req/req_install.py,sha256=0Fkfk1ireEama-Ox3H7VJEW_ytMrb3g0zN_r4sEda-8,36635 pip/_internal/req/req_set.py,sha256=ykotCTON9_be3yzrkXs9yv986Tv-p-Sk5-q8tqgk6BY,8132 pip/_internal/req/req_tracker.py,sha256=M5VZxoKrvDmwH2G2kk1QthJmhF19zCmKSXLY-WDLKWY,3195 pip/_internal/req/req_uninstall.py,sha256=VyukcnrMTRob977YO6C5t9CVadmZIF24Yj30Hfz4lpc,23590 pip/_internal/self_outdated_check.py,sha256=0RuTI8jXNm_CK7ZV8huRUqtabE0SQgwJSogapEti_zo,7934 pip/_internal/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_internal/utils/__pycache__/__init__.cpython-38.pyc,, pip/_internal/utils/__pycache__/appdirs.cpython-38.pyc,, pip/_internal/utils/__pycache__/compat.cpython-38.pyc,, pip/_internal/utils/__pycache__/deprecation.cpython-38.pyc,, pip/_internal/utils/__pycache__/encoding.cpython-38.pyc,, pip/_internal/utils/__pycache__/filesystem.cpython-38.pyc,, pip/_internal/utils/__pycache__/filetypes.cpython-38.pyc,, pip/_internal/utils/__pycache__/glibc.cpython-38.pyc,, pip/_internal/utils/__pycache__/hashes.cpython-38.pyc,, pip/_internal/utils/__pycache__/inject_securetransport.cpython-38.pyc,, pip/_internal/utils/__pycache__/logging.cpython-38.pyc,, pip/_internal/utils/__pycache__/marker_files.cpython-38.pyc,, pip/_internal/utils/__pycache__/misc.cpython-38.pyc,, pip/_internal/utils/__pycache__/models.cpython-38.pyc,, pip/_internal/utils/__pycache__/packaging.cpython-38.pyc,, pip/_internal/utils/__pycache__/setuptools_build.cpython-38.pyc,, pip/_internal/utils/__pycache__/subprocess.cpython-38.pyc,, pip/_internal/utils/__pycache__/temp_dir.cpython-38.pyc,, pip/_internal/utils/__pycache__/typing.cpython-38.pyc,, pip/_internal/utils/__pycache__/ui.cpython-38.pyc,, pip/_internal/utils/__pycache__/unpacking.cpython-38.pyc,, pip/_internal/utils/__pycache__/urls.cpython-38.pyc,, pip/_internal/utils/__pycache__/virtualenv.cpython-38.pyc,, pip/_internal/utils/appdirs.py,sha256=hchP3vBMefwp3Jr05IqN6cssVlbzYPTj062hJzBDhA4,9766 pip/_internal/utils/compat.py,sha256=i9XoohQwMK73PwN0Nkan6DfiG47rdyllTXvX3WJO30c,9565 pip/_internal/utils/deprecation.py,sha256=pBnNogoA4UGTxa_JDnPXBRRYpKMbExAhXpBwAwklOBs,3318 pip/_internal/utils/encoding.py,sha256=hxZz0t3Whw3d4MHQEiofxalTlfKwxFdLc8fpeGfhKo8,1320 pip/_internal/utils/filesystem.py,sha256=TG_hHd0BZkYih17f4z4wEi0XXheo5TE-UnAQB964Pf0,3334 pip/_internal/utils/filetypes.py,sha256=R2FwzoeX7b-rZALOXx5cuO8VPPMhUQ4ne7wm3n3IcWA,571 pip/_internal/utils/glibc.py,sha256=HXnQNSGfrldTYbF6V6asT86h6QQKPwumYok1MFyQDnM,4397 pip/_internal/utils/hashes.py,sha256=XXj0SZfz4piN8XgqloGzPnfQYGTcSZZDtuqa1zWCts8,4020 pip/_internal/utils/inject_securetransport.py,sha256=M17ZlFVY66ApgeASVjKKLKNz0LAfk-SyU0HZ4ZB6MmI,810 pip/_internal/utils/logging.py,sha256=aJL7NldPhS5KGFof6Qt3o3MG5cjm5TOoo7bGRu9_wsg,13033 pip/_internal/utils/marker_files.py,sha256=ktYfV9ccPKzVREiWlmTveiKOztk0L7F2zXi2ob2lymM,823 pip/_internal/utils/misc.py,sha256=Ekpr2RWm4BwrccJT148FA7lvcJzJSB59qINEjHMre28,25579 pip/_internal/utils/models.py,sha256=IA0hw_T4awQzui0kqfIEASm5yLtgZAB08ag59Nip5G8,1148 pip/_internal/utils/packaging.py,sha256=VtiwcAAL7LBi7tGL2je7LeW4bE11KMHGCsJ1NZY5XtM,3035 pip/_internal/utils/setuptools_build.py,sha256=vHAmalU_IcDvU7ioRioVspEyVD2N4_NPDSbtHOoow8g,1631 pip/_internal/utils/subprocess.py,sha256=M3oBdbCSxDhmQSWYRZo0PTzGuyYvJaNIzt3DVCmtjJI,9911 pip/_internal/utils/temp_dir.py,sha256=fFNkfrE3MlyjZjoLzeX9A2AkjUWEEEoP-8ObyX3BSnE,5521 pip/_internal/utils/typing.py,sha256=bF73ImJzIaxLLEVwfEaSJzFGqV9LaxkQBvDULIyr1jI,1125 pip/_internal/utils/ui.py,sha256=SmwDgg45shGo0wVJSK3MUZUQtM4DNGdQntE3n2G97yk,13906 pip/_internal/utils/unpacking.py,sha256=7qOhuK60YvrXwqGGWYQ0rE7vBrMkAwYAtmsIsN0eQgc,9682 pip/_internal/utils/urls.py,sha256=aNV9wq5ClUmrz6sG-al7hEWJ4ToitOy7l82CmFGFNW8,1481 pip/_internal/utils/virtualenv.py,sha256=oSTrUMQUqmuXcDvQZGwV65w-hlvhBAqyQiWRxLf8fN0,891 pip/_internal/vcs/__init__.py,sha256=viJxJRqRE_mVScum85bgQIXAd6o0ozFt18VpC-qIJrM,617 pip/_internal/vcs/__pycache__/__init__.cpython-38.pyc,, pip/_internal/vcs/__pycache__/bazaar.cpython-38.pyc,, pip/_internal/vcs/__pycache__/git.cpython-38.pyc,, pip/_internal/vcs/__pycache__/mercurial.cpython-38.pyc,, pip/_internal/vcs/__pycache__/subversion.cpython-38.pyc,, pip/_internal/vcs/__pycache__/versioncontrol.cpython-38.pyc,, pip/_internal/vcs/bazaar.py,sha256=84q1-kj1_nJ9AMzMu8RmMp-riRZu81M7K9kowcYgi3U,3957 pip/_internal/vcs/git.py,sha256=PqhcEuFXQhutNNT4wI0uy6QcYnk9TOf8KGQNMBxL9hU,13568 pip/_internal/vcs/mercurial.py,sha256=2mg7BdYI_Fe00fF6omaNccFQLPHBsDBG5CAEzvqn5sA,5110 pip/_internal/vcs/subversion.py,sha256=Fpwy71AmuqXnoKi6h1SrXRtPjEMn8fieuM1O4j01IBg,12292 pip/_internal/vcs/versioncontrol.py,sha256=HiWHloaYAo4UWGxJGZhG0nO0PWS513H6ACiqT0z5Yh0,21374 pip/_internal/wheel.py,sha256=emOltdF4VmpsSabfIBbtlCcozviIOnUwIdB0D3_VMrQ,43080 pip/_vendor/__init__.py,sha256=gEJYEfJm7XGLslyjW3KBQyQxyTYxdvTEkRT5Bz28MDs,4657 pip/_vendor/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/__pycache__/appdirs.cpython-38.pyc,, pip/_vendor/__pycache__/contextlib2.cpython-38.pyc,, pip/_vendor/__pycache__/distro.cpython-38.pyc,, pip/_vendor/__pycache__/ipaddress.cpython-38.pyc,, pip/_vendor/__pycache__/pyparsing.cpython-38.pyc,, pip/_vendor/__pycache__/retrying.cpython-38.pyc,, pip/_vendor/__pycache__/six.cpython-38.pyc,, pip/_vendor/appdirs.py,sha256=BENKsvcA08IpccD9345-rMrg3aXWFA1q6BFEglnHg6I,24547 pip/_vendor/cachecontrol/__init__.py,sha256=6cRPchVqkAkeUtYTSW8qCetjSqJo-GxP-n4VMVDbvmc,302 pip/_vendor/cachecontrol/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/_cmd.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/adapter.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/cache.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/compat.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/controller.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/filewrapper.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/heuristics.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/serialize.cpython-38.pyc,, pip/_vendor/cachecontrol/__pycache__/wrapper.cpython-38.pyc,, pip/_vendor/cachecontrol/_cmd.py,sha256=URGE0KrA87QekCG3SGPatlSPT571dZTDjNa-ZXX3pDc,1295 pip/_vendor/cachecontrol/adapter.py,sha256=eBGAtVNRZgtl_Kj5JV54miqL9YND-D0JZPahwY8kFtY,4863 pip/_vendor/cachecontrol/cache.py,sha256=1fc4wJP8HYt1ycnJXeEw5pCpeBL2Cqxx6g9Fb0AYDWQ,805 pip/_vendor/cachecontrol/caches/__init__.py,sha256=-gHNKYvaeD0kOk5M74eOrsSgIKUtC6i6GfbmugGweEo,86 pip/_vendor/cachecontrol/caches/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/cachecontrol/caches/__pycache__/file_cache.cpython-38.pyc,, pip/_vendor/cachecontrol/caches/__pycache__/redis_cache.cpython-38.pyc,, pip/_vendor/cachecontrol/caches/file_cache.py,sha256=nYVKsJtXh6gJXvdn1iWyrhxvkwpQrK-eKoMRzuiwkKk,4153 pip/_vendor/cachecontrol/caches/redis_cache.py,sha256=HxelMpNCo-dYr2fiJDwM3hhhRmxUYtB5tXm1GpAAT4Y,856 pip/_vendor/cachecontrol/compat.py,sha256=kHNvMRdt6s_Xwqq_9qJmr9ou3wYMOMUMxPPcwNxT8Mc,695 pip/_vendor/cachecontrol/controller.py,sha256=U7g-YwizQ2O5NRgK_MZreF1ntM4E49C3PuF3od-Vwz4,13698 pip/_vendor/cachecontrol/filewrapper.py,sha256=vACKO8Llzu_ZWyjV1Fxn1MA4TGU60N5N3GSrAFdAY2Q,2533 pip/_vendor/cachecontrol/heuristics.py,sha256=BFGHJ3yQcxvZizfo90LLZ04T_Z5XSCXvFotrp7Us0sc,4070 pip/_vendor/cachecontrol/serialize.py,sha256=GebE34fgToyWwAsRPguh8hEPN6CqoG-5hRMXRsjVABQ,6954 pip/_vendor/cachecontrol/wrapper.py,sha256=sfr9YHWx-5TwNz1H5rT6QOo8ggII6v3vbEDjQFwR6wc,671 pip/_vendor/certifi/__init__.py,sha256=WFoavXHhpX-BZ5kbvyinZTbhLsqPJypLKIZu29nUsQg,52 pip/_vendor/certifi/__main__.py,sha256=NaCn6WtWME-zzVWQ2j4zFyl8cY4knDa9CwtHNIeFPhM,53 pip/_vendor/certifi/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/certifi/__pycache__/__main__.cpython-38.pyc,, pip/_vendor/certifi/__pycache__/core.cpython-38.pyc,, pip/_vendor/certifi/core.py,sha256=p7-3v1KMyEft0Z4vBDRxJ4TfjS7l7zALERrmncYjhSg,188 pip/_vendor/chardet/__init__.py,sha256=YsP5wQlsHJ2auF1RZJfypiSrCA7_bQiRm3ES_NI76-Y,1559 pip/_vendor/chardet/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/big5freq.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/big5prober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/chardistribution.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/charsetgroupprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/charsetprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/codingstatemachine.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/compat.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/cp949prober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/enums.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/escprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/escsm.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/eucjpprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/euckrfreq.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/euckrprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/euctwfreq.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/euctwprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/gb2312freq.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/gb2312prober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/hebrewprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/jisfreq.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/jpcntx.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langbulgarianmodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langcyrillicmodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langgreekmodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langhebrewmodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langhungarianmodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langthaimodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/langturkishmodel.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/latin1prober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/mbcharsetprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/mbcsgroupprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/mbcssm.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/sbcharsetprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/sbcsgroupprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/sjisprober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/universaldetector.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/utf8prober.cpython-38.pyc,, pip/_vendor/chardet/__pycache__/version.cpython-38.pyc,, pip/_vendor/chardet/big5freq.py,sha256=D_zK5GyzoVsRes0HkLJziltFQX0bKCLOrFe9_xDvO_8,31254 pip/_vendor/chardet/big5prober.py,sha256=kBxHbdetBpPe7xrlb-e990iot64g_eGSLd32lB7_h3M,1757 pip/_vendor/chardet/chardistribution.py,sha256=3woWS62KrGooKyqz4zQSnjFbJpa6V7g02daAibTwcl8,9411 pip/_vendor/chardet/charsetgroupprober.py,sha256=6bDu8YIiRuScX4ca9Igb0U69TA2PGXXDej6Cc4_9kO4,3787 pip/_vendor/chardet/charsetprober.py,sha256=KSmwJErjypyj0bRZmC5F5eM7c8YQgLYIjZXintZNstg,5110 pip/_vendor/chardet/cli/__init__.py,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1 pip/_vendor/chardet/cli/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/chardet/cli/__pycache__/chardetect.cpython-38.pyc,, pip/_vendor/chardet/cli/chardetect.py,sha256=DI8dlV3FBD0c0XA_y3sQ78z754DUv1J8n34RtDjOXNw,2774 pip/_vendor/chardet/codingstatemachine.py,sha256=VYp_6cyyki5sHgXDSZnXW4q1oelHc3cu9AyQTX7uug8,3590 pip/_vendor/chardet/compat.py,sha256=PKTzHkSbtbHDqS9PyujMbX74q1a8mMpeQTDVsQhZMRw,1134 pip/_vendor/chardet/cp949prober.py,sha256=TZ434QX8zzBsnUvL_8wm4AQVTZ2ZkqEEQL_lNw9f9ow,1855 pip/_vendor/chardet/enums.py,sha256=Aimwdb9as1dJKZaFNUH2OhWIVBVd6ZkJJ_WK5sNY8cU,1661 pip/_vendor/chardet/escprober.py,sha256=kkyqVg1Yw3DIOAMJ2bdlyQgUFQhuHAW8dUGskToNWSc,3950 pip/_vendor/chardet/escsm.py,sha256=RuXlgNvTIDarndvllNCk5WZBIpdCxQ0kcd9EAuxUh84,10510 pip/_vendor/chardet/eucjpprober.py,sha256=iD8Jdp0ISRjgjiVN7f0e8xGeQJ5GM2oeZ1dA8nbSeUw,3749 pip/_vendor/chardet/euckrfreq.py,sha256=-7GdmvgWez4-eO4SuXpa7tBiDi5vRXQ8WvdFAzVaSfo,13546 pip/_vendor/chardet/euckrprober.py,sha256=MqFMTQXxW4HbzIpZ9lKDHB3GN8SP4yiHenTmf8g_PxY,1748 pip/_vendor/chardet/euctwfreq.py,sha256=No1WyduFOgB5VITUA7PLyC5oJRNzRyMbBxaKI1l16MA,31621 pip/_vendor/chardet/euctwprober.py,sha256=13p6EP4yRaxqnP4iHtxHOJ6R2zxHq1_m8hTRjzVZ95c,1747 pip/_vendor/chardet/gb2312freq.py,sha256=JX8lsweKLmnCwmk8UHEQsLgkr_rP_kEbvivC4qPOrlc,20715 pip/_vendor/chardet/gb2312prober.py,sha256=gGvIWi9WhDjE-xQXHvNIyrnLvEbMAYgyUSZ65HUfylw,1754 pip/_vendor/chardet/hebrewprober.py,sha256=c3SZ-K7hvyzGY6JRAZxJgwJ_sUS9k0WYkvMY00YBYFo,13838 pip/_vendor/chardet/jisfreq.py,sha256=vpmJv2Bu0J8gnMVRPHMFefTRvo_ha1mryLig8CBwgOg,25777 pip/_vendor/chardet/jpcntx.py,sha256=PYlNqRUQT8LM3cT5FmHGP0iiscFlTWED92MALvBungo,19643 pip/_vendor/chardet/langbulgarianmodel.py,sha256=1HqQS9Pbtnj1xQgxitJMvw8X6kKr5OockNCZWfEQrPE,12839 pip/_vendor/chardet/langcyrillicmodel.py,sha256=LODajvsetH87yYDDQKA2CULXUH87tI223dhfjh9Zx9c,17948 pip/_vendor/chardet/langgreekmodel.py,sha256=8YAW7bU8YwSJap0kIJSbPMw1BEqzGjWzqcqf0WgUKAA,12688 pip/_vendor/chardet/langhebrewmodel.py,sha256=JSnqmE5E62tDLTPTvLpQsg5gOMO4PbdWRvV7Avkc0HA,11345 pip/_vendor/chardet/langhungarianmodel.py,sha256=RhapYSG5l0ZaO-VV4Fan5sW0WRGQqhwBM61yx3yxyOA,12592 pip/_vendor/chardet/langthaimodel.py,sha256=8l0173Gu_W6G8mxmQOTEF4ls2YdE7FxWf3QkSxEGXJQ,11290 pip/_vendor/chardet/langturkishmodel.py,sha256=W22eRNJsqI6uWAfwXSKVWWnCerYqrI8dZQTm_M0lRFk,11102 pip/_vendor/chardet/latin1prober.py,sha256=S2IoORhFk39FEFOlSFWtgVybRiP6h7BlLldHVclNkU8,5370 pip/_vendor/chardet/mbcharsetprober.py,sha256=AR95eFH9vuqSfvLQZN-L5ijea25NOBCoXqw8s5O9xLQ,3413 pip/_vendor/chardet/mbcsgroupprober.py,sha256=h6TRnnYq2OxG1WdD5JOyxcdVpn7dG0q-vB8nWr5mbh4,2012 pip/_vendor/chardet/mbcssm.py,sha256=SY32wVIF3HzcjY3BaEspy9metbNSKxIIB0RKPn7tjpI,25481 pip/_vendor/chardet/sbcharsetprober.py,sha256=LDSpCldDCFlYwUkGkwD2oFxLlPWIWXT09akH_2PiY74,5657 pip/_vendor/chardet/sbcsgroupprober.py,sha256=1IprcCB_k1qfmnxGC6MBbxELlKqD3scW6S8YIwdeyXA,3546 pip/_vendor/chardet/sjisprober.py,sha256=IIt-lZj0WJqK4rmUZzKZP4GJlE8KUEtFYVuY96ek5MQ,3774 pip/_vendor/chardet/universaldetector.py,sha256=qL0174lSZE442eB21nnktT9_VcAye07laFWUeUrjttY,12485 pip/_vendor/chardet/utf8prober.py,sha256=IdD8v3zWOsB8OLiyPi-y_fqwipRFxV9Nc1eKBLSuIEw,2766 pip/_vendor/chardet/version.py,sha256=sp3B08mrDXB-pf3K9fqJ_zeDHOCLC8RrngQyDFap_7g,242 pip/_vendor/colorama/__init__.py,sha256=lJdY6COz9uM_pXwuk9oLr0fp8H8q2RrUqN16GKabvq4,239 pip/_vendor/colorama/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/colorama/__pycache__/ansi.cpython-38.pyc,, pip/_vendor/colorama/__pycache__/ansitowin32.cpython-38.pyc,, pip/_vendor/colorama/__pycache__/initialise.cpython-38.pyc,, pip/_vendor/colorama/__pycache__/win32.cpython-38.pyc,, pip/_vendor/colorama/__pycache__/winterm.cpython-38.pyc,, pip/_vendor/colorama/ansi.py,sha256=Fi0un-QLqRm-v7o_nKiOqyC8PapBJK7DLV_q9LKtTO0,2524 pip/_vendor/colorama/ansitowin32.py,sha256=u8QaqdqS_xYSfNkPM1eRJLHz6JMWPodaJaP0mxgHCDc,10462 pip/_vendor/colorama/initialise.py,sha256=PprovDNxMTrvoNHFcL2NZjpH2XzDc8BLxLxiErfUl4k,1915 pip/_vendor/colorama/win32.py,sha256=bJ8Il9jwaBN5BJ8bmN6FoYZ1QYuMKv2j8fGrXh7TJjw,5404 pip/_vendor/colorama/winterm.py,sha256=2y_2b7Zsv34feAsP67mLOVc-Bgq51mdYGo571VprlrM,6438 pip/_vendor/contextlib2.py,sha256=5HjGflUzwWAUfcILhSmC2GqvoYdZZzFzVfIDztHigUs,16915 pip/_vendor/distlib/__init__.py,sha256=SkHYPuEQNQF2a2Cr18rfZ-LQyDqwwizn8tJE4seXPgU,587 pip/_vendor/distlib/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/compat.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/database.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/index.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/locators.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/manifest.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/markers.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/metadata.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/resources.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/scripts.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/util.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/version.cpython-38.pyc,, pip/_vendor/distlib/__pycache__/wheel.cpython-38.pyc,, pip/_vendor/distlib/_backport/__init__.py,sha256=bqS_dTOH6uW9iGgd0uzfpPjo6vZ4xpPZ7kyfZJ2vNaw,274 pip/_vendor/distlib/_backport/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/distlib/_backport/__pycache__/misc.cpython-38.pyc,, pip/_vendor/distlib/_backport/__pycache__/sysconfig.cpython-38.pyc,, pip/_vendor/distlib/_backport/misc.py,sha256=KWecINdbFNOxSOP1fGF680CJnaC6S4fBRgEtaYTw0ig,971 pip/_vendor/distlib/_backport/sysconfig.cfg,sha256=swZKxq9RY5e9r3PXCrlvQPMsvOdiWZBTHLEbqS8LJLU,2617 pip/_vendor/distlib/_backport/sysconfig.py,sha256=JdJ9ztRy4Hc-b5-VS74x3nUtdEIVr_OBvMsIb8O2sjc,26964 pip/_vendor/distlib/compat.py,sha256=xdNZmqFN5HwF30HjRn5M415pcC2kgXRBXn767xS8v-M,41404 pip/_vendor/distlib/database.py,sha256=-KJH63AJ7hqjLtGCwOTrionhKr2Vsytdwkjyo8UdEco,51029 pip/_vendor/distlib/index.py,sha256=SXKzpQCERctxYDMp_OLee2f0J0e19ZhGdCIoMlUfUQM,21066 pip/_vendor/distlib/locators.py,sha256=bqzEWP3Ad8UE3D1rmzW1pgzVTKkY4rDUA_EWIVYli54,51807 pip/_vendor/distlib/manifest.py,sha256=nQEhYmgoreaBZzyFzwYsXxJARu3fo4EkunU163U16iE,14811 pip/_vendor/distlib/markers.py,sha256=6Ac3cCfFBERexiESWIOXmg-apIP8l2esafNSX3KMy-8,4387 pip/_vendor/distlib/metadata.py,sha256=OhbCKmf5lswE8unWBopI1hj7tRpHp4ZbFvU4d6aAEMM,40234 pip/_vendor/distlib/resources.py,sha256=2FGv0ZHF14KXjLIlL0R991lyQQGcewOS4mJ-5n-JVnc,10766 pip/_vendor/distlib/scripts.py,sha256=W24OXnZUmgRX_XtDrVZdfc-Frf4X4_cybvhP87iR-QU,16290 pip/_vendor/distlib/util.py,sha256=vv8T7g_c8Ei9dRosWP1ypxKmaBpPpTwlMpPRKbLgWEY,60408 pip/_vendor/distlib/version.py,sha256=_n7F6juvQGAcn769E_SHa7fOcf5ERlEVymJ_EjPRwGw,23391 pip/_vendor/distlib/wheel.py,sha256=2lviV6L4IvTP5DRkKE0HGpClvdoTJQHZJLfTQ6dfn2A,40437 pip/_vendor/distro.py,sha256=X2So5kjrRKyMbQJ90Xgy93HU5eFtujCzKaYNeoy1k1c,43251 pip/_vendor/html5lib/__init__.py,sha256=Ztrn7UvF-wIFAgRBBa0ML-Gu5AffH3BPX_INJx4SaBI,1162 pip/_vendor/html5lib/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/_ihatexml.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/_inputstream.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/_tokenizer.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/_utils.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/constants.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/html5parser.cpython-38.pyc,, pip/_vendor/html5lib/__pycache__/serializer.cpython-38.pyc,, pip/_vendor/html5lib/_ihatexml.py,sha256=3LBtJMlzgwM8vpQiU1TvGmEEmNH72sV0yD8yS53y07A,16705 pip/_vendor/html5lib/_inputstream.py,sha256=bPUWcAfJScK4xkjQQaG_HsI2BvEVbFvI0AsodDYPQj0,32552 pip/_vendor/html5lib/_tokenizer.py,sha256=YAaOEBD6qc5ISq9Xt9Nif1OFgcybTTfMdwqBkZhpAq4,76580 pip/_vendor/html5lib/_trie/__init__.py,sha256=8VR1bcgD2OpeS2XExpu5yBhP_Q1K-lwKbBKICBPf1kU,289 pip/_vendor/html5lib/_trie/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/html5lib/_trie/__pycache__/_base.cpython-38.pyc,, pip/_vendor/html5lib/_trie/__pycache__/datrie.cpython-38.pyc,, pip/_vendor/html5lib/_trie/__pycache__/py.cpython-38.pyc,, pip/_vendor/html5lib/_trie/_base.py,sha256=CaybYyMro8uERQYjby2tTeSUatnWDfWroUN9N7ety5w,1013 pip/_vendor/html5lib/_trie/datrie.py,sha256=EQpqSfkZRuTbE-DuhW7xMdVDxdZNZ0CfmnYfHA_3zxM,1178 pip/_vendor/html5lib/_trie/py.py,sha256=wXmQLrZRf4MyWNyg0m3h81m9InhLR7GJ002mIIZh-8o,1775 pip/_vendor/html5lib/_utils.py,sha256=ismpASeqa2jqEPQjHUj8vReAf7yIoKnvLN5fuOw6nv0,4015 pip/_vendor/html5lib/constants.py,sha256=4lmZWLtEPRLnl8NzftOoYTJdo6jpeMtP6dqQC0g_bWQ,83518 pip/_vendor/html5lib/filters/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_vendor/html5lib/filters/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/alphabeticalattributes.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/base.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/inject_meta_charset.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/lint.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/optionaltags.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/sanitizer.cpython-38.pyc,, pip/_vendor/html5lib/filters/__pycache__/whitespace.cpython-38.pyc,, pip/_vendor/html5lib/filters/alphabeticalattributes.py,sha256=lViZc2JMCclXi_5gduvmdzrRxtO5Xo9ONnbHBVCsykU,919 pip/_vendor/html5lib/filters/base.py,sha256=z-IU9ZAYjpsVsqmVt7kuWC63jR11hDMr6CVrvuao8W0,286 pip/_vendor/html5lib/filters/inject_meta_charset.py,sha256=egDXUEHXmAG9504xz0K6ALDgYkvUrC2q15YUVeNlVQg,2945 pip/_vendor/html5lib/filters/lint.py,sha256=jk6q56xY0ojiYfvpdP-OZSm9eTqcAdRqhCoPItemPYA,3643 pip/_vendor/html5lib/filters/optionaltags.py,sha256=8lWT75J0aBOHmPgfmqTHSfPpPMp01T84NKu0CRedxcE,10588 pip/_vendor/html5lib/filters/sanitizer.py,sha256=4ON02KNjuqda1lCw5_JCUZxb0BzWR5M7ON84dtJ7dm0,26248 pip/_vendor/html5lib/filters/whitespace.py,sha256=8eWqZxd4UC4zlFGW6iyY6f-2uuT8pOCSALc3IZt7_t4,1214 pip/_vendor/html5lib/html5parser.py,sha256=g5g2ezkusHxhi7b23vK_-d6K6BfIJRbqIQmvQ9z4EgI,118963 pip/_vendor/html5lib/serializer.py,sha256=yfcfBHse2wDs6ojxn-kieJjLT5s1ipilQJ0gL3-rJis,15758 pip/_vendor/html5lib/treeadapters/__init__.py,sha256=A0rY5gXIe4bJOiSGRO_j_tFhngRBO8QZPzPtPw5dFzo,679 pip/_vendor/html5lib/treeadapters/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/html5lib/treeadapters/__pycache__/genshi.cpython-38.pyc,, pip/_vendor/html5lib/treeadapters/__pycache__/sax.cpython-38.pyc,, pip/_vendor/html5lib/treeadapters/genshi.py,sha256=CH27pAsDKmu4ZGkAUrwty7u0KauGLCZRLPMzaO3M5vo,1715 pip/_vendor/html5lib/treeadapters/sax.py,sha256=BKS8woQTnKiqeffHsxChUqL4q2ZR_wb5fc9MJ3zQC8s,1776 pip/_vendor/html5lib/treebuilders/__init__.py,sha256=AysSJyvPfikCMMsTVvaxwkgDieELD5dfR8FJIAuq7hY,3592 pip/_vendor/html5lib/treebuilders/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/html5lib/treebuilders/__pycache__/base.cpython-38.pyc,, pip/_vendor/html5lib/treebuilders/__pycache__/dom.cpython-38.pyc,, pip/_vendor/html5lib/treebuilders/__pycache__/etree.cpython-38.pyc,, pip/_vendor/html5lib/treebuilders/__pycache__/etree_lxml.cpython-38.pyc,, pip/_vendor/html5lib/treebuilders/base.py,sha256=wQGp5yy22TNG8tJ6aREe4UUeTR7A99dEz0BXVaedWb4,14579 pip/_vendor/html5lib/treebuilders/dom.py,sha256=22whb0C71zXIsai5mamg6qzBEiigcBIvaDy4Asw3at0,8925 pip/_vendor/html5lib/treebuilders/etree.py,sha256=aqIBOGj_dFYqBURIcTegGNBhAIJOw5iFDHb4jrkYH-8,12764 pip/_vendor/html5lib/treebuilders/etree_lxml.py,sha256=9V0dXxbJYYq-Skgb5-_OL2NkVYpjioEb4CHajo0e9yI,14122 pip/_vendor/html5lib/treewalkers/__init__.py,sha256=yhXxHpjlSqfQyUag3v8-vWjMPriFBU8YRAPNpDgBTn8,5714 pip/_vendor/html5lib/treewalkers/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/html5lib/treewalkers/__pycache__/base.cpython-38.pyc,, pip/_vendor/html5lib/treewalkers/__pycache__/dom.cpython-38.pyc,, pip/_vendor/html5lib/treewalkers/__pycache__/etree.cpython-38.pyc,, pip/_vendor/html5lib/treewalkers/__pycache__/etree_lxml.cpython-38.pyc,, pip/_vendor/html5lib/treewalkers/__pycache__/genshi.cpython-38.pyc,, pip/_vendor/html5lib/treewalkers/base.py,sha256=ouiOsuSzvI0KgzdWP8PlxIaSNs9falhbiinAEc_UIJY,7476 pip/_vendor/html5lib/treewalkers/dom.py,sha256=EHyFR8D8lYNnyDU9lx_IKigVJRyecUGua0mOi7HBukc,1413 pip/_vendor/html5lib/treewalkers/etree.py,sha256=sz1o6mmE93NQ53qJFDO7HKyDtuwgK-Ay3qSFZPC6u00,4550 pip/_vendor/html5lib/treewalkers/etree_lxml.py,sha256=sY6wfRshWTllu6n48TPWpKsQRPp-0CQrT0hj_AdzHSU,6309 pip/_vendor/html5lib/treewalkers/genshi.py,sha256=4D2PECZ5n3ZN3qu3jMl9yY7B81jnQApBQSVlfaIuYbA,2309 pip/_vendor/idna/__init__.py,sha256=9Nt7xpyet3DmOrPUGooDdAwmHZZu1qUAy2EaJ93kGiQ,58 pip/_vendor/idna/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/idna/__pycache__/codec.cpython-38.pyc,, pip/_vendor/idna/__pycache__/compat.cpython-38.pyc,, pip/_vendor/idna/__pycache__/core.cpython-38.pyc,, pip/_vendor/idna/__pycache__/idnadata.cpython-38.pyc,, pip/_vendor/idna/__pycache__/intranges.cpython-38.pyc,, pip/_vendor/idna/__pycache__/package_data.cpython-38.pyc,, pip/_vendor/idna/__pycache__/uts46data.cpython-38.pyc,, pip/_vendor/idna/codec.py,sha256=lvYb7yu7PhAqFaAIAdWcwgaWI2UmgseUua-1c0AsG0A,3299 pip/_vendor/idna/compat.py,sha256=R-h29D-6mrnJzbXxymrWUW7iZUvy-26TQwZ0ij57i4U,232 pip/_vendor/idna/core.py,sha256=JDCZZ_PLESqIgEbU8mPyoEufWwoOiIqygA17-QZIe3s,11733 pip/_vendor/idna/idnadata.py,sha256=HXaPFw6_YAJ0qppACPu0YLAULtRs3QovRM_CCZHGdY0,40899 pip/_vendor/idna/intranges.py,sha256=TY1lpxZIQWEP6tNqjZkFA5hgoMWOj1OBmnUG8ihT87E,1749 pip/_vendor/idna/package_data.py,sha256=kIzeKKXEouXLR4srqwf9Q3zv-NffKSOz5aSDOJARPB0,21 pip/_vendor/idna/uts46data.py,sha256=oLyNZ1pBaiBlj9zFzLFRd_P7J8MkRcgDisjExZR_4MY,198292 pip/_vendor/ipaddress.py,sha256=2OgbkeAD2rLkcXqbcvof3J5R7lRwjNLoBySyTkBtKnc,79852 pip/_vendor/msgpack/__init__.py,sha256=LnKzG5v0RyZgs7KlY2-SZYDBn-toylovXxKiXR6C-IQ,1535 pip/_vendor/msgpack/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/msgpack/__pycache__/_version.cpython-38.pyc,, pip/_vendor/msgpack/__pycache__/exceptions.cpython-38.pyc,, pip/_vendor/msgpack/__pycache__/fallback.cpython-38.pyc,, pip/_vendor/msgpack/_version.py,sha256=72BxB5FMl1q3Nz1hteHINzHhrFpXQ9nNtULaK52NLk8,20 pip/_vendor/msgpack/exceptions.py,sha256=dCTWei8dpkrMsQDcjQk74ATl9HsIBH0ybt8zOPNqMYc,1081 pip/_vendor/msgpack/fallback.py,sha256=vXo6S67Dmil9mz0PRBCLDu6znpv6CGKt9WPCEsdZx2A,37454 pip/_vendor/packaging/__about__.py,sha256=CpuMSyh1V7adw8QMjWKkY3LtdqRUkRX4MgJ6nF4stM0,744 pip/_vendor/packaging/__init__.py,sha256=6enbp5XgRfjBjsI9-bn00HjHf5TH21PDMOKkJW8xw-w,562 pip/_vendor/packaging/__pycache__/__about__.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/_compat.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/_structures.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/markers.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/requirements.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/specifiers.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/tags.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/utils.cpython-38.pyc,, pip/_vendor/packaging/__pycache__/version.cpython-38.pyc,, pip/_vendor/packaging/_compat.py,sha256=Ugdm-qcneSchW25JrtMIKgUxfEEBcCAz6WrEeXeqz9o,865 pip/_vendor/packaging/_structures.py,sha256=pVd90XcXRGwpZRB_qdFuVEibhCHpX_bL5zYr9-N0mc8,1416 pip/_vendor/packaging/markers.py,sha256=jRoHXMzT_7InY31pBB9Nkx66COZpQBAwa5scHla9uVQ,8250 pip/_vendor/packaging/requirements.py,sha256=grcnFU8x7KD230JaFLXtWl3VClLuOmsOy4c-m55tOWs,4700 pip/_vendor/packaging/specifiers.py,sha256=0ZzQpcUnvrQ6LjR-mQRLzMr8G6hdRv-mY0VSf_amFtI,27778 pip/_vendor/packaging/tags.py,sha256=EPLXhO6GTD7_oiWEO1U0l0PkfR8R_xivpMDHXnsTlts,12933 pip/_vendor/packaging/utils.py,sha256=VaTC0Ei7zO2xl9ARiWmz2YFLFt89PuuhLbAlXMyAGms,1520 pip/_vendor/packaging/version.py,sha256=Npdwnb8OHedj_2L86yiUqscujb7w_i5gmSK1PhOAFzg,11978 pip/_vendor/pep517/__init__.py,sha256=nCw8ZdLH4c19g8xP_Ndag1KPdQhlSDKaL9pg-X7uNWU,84 pip/_vendor/pep517/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/_in_process.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/build.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/check.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/colorlog.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/compat.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/dirtools.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/envbuild.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/meta.cpython-38.pyc,, pip/_vendor/pep517/__pycache__/wrappers.cpython-38.pyc,, pip/_vendor/pep517/_in_process.py,sha256=v1Viek27-MGCOFu8eSlLd2jGCrIqc1fISnutGFoRDps,7792 pip/_vendor/pep517/build.py,sha256=WqM0-X4KyzY566qxGf3FeaYc1hw95H7YP0ElZ1zuTb0,3318 pip/_vendor/pep517/check.py,sha256=ST02kRWBrRMOxgnRm9clw18Q2X7sJGaD4j3h6GmBhJ8,5949 pip/_vendor/pep517/colorlog.py,sha256=Tk9AuYm_cLF3BKTBoSTJt9bRryn0aFojIQOwbfVUTxQ,4098 pip/_vendor/pep517/compat.py,sha256=M-5s4VNp8rjyT76ZZ_ibnPD44DYVzSQlyCEHayjtDPw,780 pip/_vendor/pep517/dirtools.py,sha256=2mkAkAL0mRz_elYFjRKuekTJVipH1zTn4tbf1EDev84,1129 pip/_vendor/pep517/envbuild.py,sha256=K4dIGAbkXf3RoQX_9RFpZvMvPrVSHtcbH7o9VSrNnlM,6024 pip/_vendor/pep517/meta.py,sha256=8mnM5lDnT4zXQpBTliJbRGfesH7iioHwozbDxALPS9Y,2463 pip/_vendor/pep517/wrappers.py,sha256=QiQaEQlfCrhRpPBFQiGVM9QjrKSlj8AvM39haoyfPRk,10599 pip/_vendor/pkg_resources/__init__.py,sha256=hnT0Ph4iK40Ycr-OzSii_wZW5f7HCkP79E6Vf4cR3Vg,108237 pip/_vendor/pkg_resources/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/pkg_resources/__pycache__/py31compat.cpython-38.pyc,, pip/_vendor/pkg_resources/py31compat.py,sha256=CRk8fkiPRDLsbi5pZcKsHI__Pbmh_94L8mr9Qy9Ab2U,562 pip/_vendor/progress/__init__.py,sha256=fcbQQXo5np2CoQyhSH5XprkicwLZNLePR3uIahznSO0,4857 pip/_vendor/progress/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/progress/__pycache__/bar.cpython-38.pyc,, pip/_vendor/progress/__pycache__/counter.cpython-38.pyc,, pip/_vendor/progress/__pycache__/spinner.cpython-38.pyc,, pip/_vendor/progress/bar.py,sha256=QuDuVNcmXgpxtNtxO0Fq72xKigxABaVmxYGBw4J3Z_E,2854 pip/_vendor/progress/counter.py,sha256=MznyBrvPWrOlGe4MZAlGUb9q3aODe6_aNYeAE_VNoYA,1372 pip/_vendor/progress/spinner.py,sha256=k8JbDW94T0-WXuXfxZIFhdoNPYp3jfnpXqBnfRv5fGs,1380 pip/_vendor/pyparsing.py,sha256=uJuFv_UoLmKOHlCU-EhIiKjYqYS2OCj-Gmr-sRkCzmU,263468 pip/_vendor/pytoml/__init__.py,sha256=W_SKx36Hsew-Fty36BOpreLm4uF4V_Tgkm_z9rIoOE8,127 pip/_vendor/pytoml/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/pytoml/__pycache__/core.cpython-38.pyc,, pip/_vendor/pytoml/__pycache__/parser.cpython-38.pyc,, pip/_vendor/pytoml/__pycache__/test.cpython-38.pyc,, pip/_vendor/pytoml/__pycache__/utils.cpython-38.pyc,, pip/_vendor/pytoml/__pycache__/writer.cpython-38.pyc,, pip/_vendor/pytoml/core.py,sha256=9CrLLTs1PdWjEwRnYzt_i4dhHcZvGxs_GsMlYAX3iY4,509 pip/_vendor/pytoml/parser.py,sha256=qsc0NRnTgdFZgRp9gmr6D_KWFelrwxLkTj9dVxUcqS8,10309 pip/_vendor/pytoml/test.py,sha256=2nQs4aX3XQEaaQCx6x_OJTS2Hb0_IiTZRqNOeDmLCzo,1021 pip/_vendor/pytoml/utils.py,sha256=JCLHx77Hu1R3F-bRgiROIiKyCzLwyebnp5P35cRJxWs,1665 pip/_vendor/pytoml/writer.py,sha256=4QQky9JSuRv60uzuhVZASU8T3CuobSkLG1285X6bDW8,3369 pip/_vendor/requests/__init__.py,sha256=ONVsH6kJuPTV9nf-XVoubWsVX3qVtjCyju42kTW6Uug,4074 pip/_vendor/requests/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/requests/__pycache__/__version__.cpython-38.pyc,, pip/_vendor/requests/__pycache__/_internal_utils.cpython-38.pyc,, pip/_vendor/requests/__pycache__/adapters.cpython-38.pyc,, pip/_vendor/requests/__pycache__/api.cpython-38.pyc,, pip/_vendor/requests/__pycache__/auth.cpython-38.pyc,, pip/_vendor/requests/__pycache__/certs.cpython-38.pyc,, pip/_vendor/requests/__pycache__/compat.cpython-38.pyc,, pip/_vendor/requests/__pycache__/cookies.cpython-38.pyc,, pip/_vendor/requests/__pycache__/exceptions.cpython-38.pyc,, pip/_vendor/requests/__pycache__/help.cpython-38.pyc,, pip/_vendor/requests/__pycache__/hooks.cpython-38.pyc,, pip/_vendor/requests/__pycache__/models.cpython-38.pyc,, pip/_vendor/requests/__pycache__/packages.cpython-38.pyc,, pip/_vendor/requests/__pycache__/sessions.cpython-38.pyc,, pip/_vendor/requests/__pycache__/status_codes.cpython-38.pyc,, pip/_vendor/requests/__pycache__/structures.cpython-38.pyc,, pip/_vendor/requests/__pycache__/utils.cpython-38.pyc,, pip/_vendor/requests/__version__.py,sha256=Bm-GFstQaFezsFlnmEMrJDe8JNROz9n2XXYtODdvjjc,436 pip/_vendor/requests/_internal_utils.py,sha256=Zx3PnEUccyfsB-ie11nZVAW8qClJy0gx1qNME7rgT18,1096 pip/_vendor/requests/adapters.py,sha256=e-bmKEApNVqFdylxuMJJfiaHdlmS_zhWhIMEzlHvGuc,21548 pip/_vendor/requests/api.py,sha256=fbUo11QoLOoNgWU6FfvNz8vMj9bE_cMmICXBa7TZHJs,6271 pip/_vendor/requests/auth.py,sha256=QB2-cSUj1jrvWZfPXttsZpyAacQgtKLVk14vQW9TpSE,10206 pip/_vendor/requests/certs.py,sha256=nXRVq9DtGmv_1AYbwjTu9UrgAcdJv05ZvkNeaoLOZxY,465 pip/_vendor/requests/compat.py,sha256=FZX4Q_EMKiMnhZpZ3g_gOsT-j2ca9ij2gehDx1cwYeo,1941 pip/_vendor/requests/cookies.py,sha256=Y-bKX6TvW3FnYlE6Au0SXtVVWcaNdFvuAwQxw-G0iTI,18430 pip/_vendor/requests/exceptions.py,sha256=-mLam3TAx80V09EaH3H-ZxR61eAVuLRZ8zgBBSLjK44,3197 pip/_vendor/requests/help.py,sha256=SJPVcoXeo7KfK4AxJN5eFVQCjr0im87tU2n7ubLsksU,3578 pip/_vendor/requests/hooks.py,sha256=QReGyy0bRcr5rkwCuObNakbYsc7EkiKeBwG4qHekr2Q,757 pip/_vendor/requests/models.py,sha256=6s-37iAqXVptq8z7U_LoH_pbIPrCQUm_Z8QuIGE29Q0,34275 pip/_vendor/requests/packages.py,sha256=njJmVifY4aSctuW3PP5EFRCxjEwMRDO6J_feG2dKWsI,695 pip/_vendor/requests/sessions.py,sha256=DjbCotDW6xSAaBsjbW-L8l4N0UcwmrxVNgSrZgIjGWM,29332 pip/_vendor/requests/status_codes.py,sha256=XWlcpBjbCtq9sSqpH9_KKxgnLTf9Z__wCWolq21ySlg,4129 pip/_vendor/requests/structures.py,sha256=zoP8qly2Jak5e89HwpqjN1z2diztI-_gaqts1raJJBc,2981 pip/_vendor/requests/utils.py,sha256=LtPJ1db6mJff2TJSJWKi7rBpzjPS3mSOrjC9zRhoD3A,30049 pip/_vendor/retrying.py,sha256=k3fflf5_Mm0XcIJYhB7Tj34bqCCPhUDkYbx1NvW2FPE,9972 pip/_vendor/six.py,sha256=h9jch2pS86y4R36pKRS3LOYUCVFNIJMRwjZ4fJDtJ44,32452 pip/_vendor/urllib3/__init__.py,sha256=cedaGRiXnA8WzcI3WPbb9u2Al9l2ortwfXZSQ4yFHHU,2683 pip/_vendor/urllib3/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/_collections.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/connection.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/connectionpool.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/exceptions.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/fields.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/filepost.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/poolmanager.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/request.cpython-38.pyc,, pip/_vendor/urllib3/__pycache__/response.cpython-38.pyc,, pip/_vendor/urllib3/_collections.py,sha256=GouVsNzwg6jADZTmimMI6oqmwKSswnMo9dh5tGNVWO4,10792 pip/_vendor/urllib3/connection.py,sha256=p7uUWh3cP0hgja8fFlzfZvwVxviJa8-C5cx9G3wQ5-o,15170 pip/_vendor/urllib3/connectionpool.py,sha256=NYYFYX-L9XO4tQMU7ug5ABidsmGKVSE2_X4XMggzwxk,36446 pip/_vendor/urllib3/contrib/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_vendor/urllib3/contrib/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/contrib/__pycache__/_appengine_environ.cpython-38.pyc,, pip/_vendor/urllib3/contrib/__pycache__/appengine.cpython-38.pyc,, pip/_vendor/urllib3/contrib/__pycache__/ntlmpool.cpython-38.pyc,, pip/_vendor/urllib3/contrib/__pycache__/pyopenssl.cpython-38.pyc,, pip/_vendor/urllib3/contrib/__pycache__/securetransport.cpython-38.pyc,, pip/_vendor/urllib3/contrib/__pycache__/socks.cpython-38.pyc,, pip/_vendor/urllib3/contrib/_appengine_environ.py,sha256=tJvMXygi5UnFn4tmCHtrXOQIqy1FAfZoDDK36Q35F1I,707 pip/_vendor/urllib3/contrib/_securetransport/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_vendor/urllib3/contrib/_securetransport/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/contrib/_securetransport/__pycache__/bindings.cpython-38.pyc,, pip/_vendor/urllib3/contrib/_securetransport/__pycache__/low_level.cpython-38.pyc,, pip/_vendor/urllib3/contrib/_securetransport/bindings.py,sha256=ZuSEVZwiubk0oaVmkZa8bUoK9ACVJJhPVgRzPZN6KoQ,16805 pip/_vendor/urllib3/contrib/_securetransport/low_level.py,sha256=V7GnujxnWZh2N2sMsV5N4d9Imymokkm3zBwgt77_bSE,11956 pip/_vendor/urllib3/contrib/appengine.py,sha256=CosoKgcu5PE5COkPSGa7Q5AFzh9XWAf0PTBU7LSBE7A,11314 pip/_vendor/urllib3/contrib/ntlmpool.py,sha256=YnWc2-np7Rzi2LfCxJ2fEprhGkeZDSjQFdJuTQ5vuUE,4201 pip/_vendor/urllib3/contrib/pyopenssl.py,sha256=w35mWy_1POZUsbOhurVb_zhf0C1Jkd79AFlucLs6KuQ,16440 pip/_vendor/urllib3/contrib/securetransport.py,sha256=_vByA8KzFPxn9x67hilAPu9wpoKDS9fU3K1DVavEk74,32742 pip/_vendor/urllib3/contrib/socks.py,sha256=nzDMgDIFJWVubKHqvIn2-SKCO91hhJInP92WgHChGzA,7036 pip/_vendor/urllib3/exceptions.py,sha256=_tiSmwYQ8em6OSr5GPO5hpbCG8q0DTIuJ2F7NMEyDyc,6610 pip/_vendor/urllib3/fields.py,sha256=kroD76QK-GdHHW7f_AUN4XxDC3OQPI2FFrS9eSL4BCs,8553 pip/_vendor/urllib3/filepost.py,sha256=vj0qbrpT1AFzvvW4SuC8M5kJiw7wftHcSr-7b8UpPpw,2440 pip/_vendor/urllib3/packages/__init__.py,sha256=h4BLhD4tLaBx1adaDtKXfupsgqY0wWLXb_f1_yVlV6A,108 pip/_vendor/urllib3/packages/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/packages/__pycache__/six.cpython-38.pyc,, pip/_vendor/urllib3/packages/backports/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pip/_vendor/urllib3/packages/backports/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/packages/backports/__pycache__/makefile.cpython-38.pyc,, pip/_vendor/urllib3/packages/backports/makefile.py,sha256=005wrvH-_pWSnTFqQ2sdzzh4zVCtQUUQ4mR2Yyxwc0A,1418 pip/_vendor/urllib3/packages/six.py,sha256=adx4z-eM_D0Vvu0IIqVzFACQ_ux9l64y7DkSEfbxCDs,32536 pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.py,sha256=ywgKMtfHi1-DrXlzPfVAhzsLzzqcK7GT6eLgdode1Fg,688 pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/_implementation.cpython-38.pyc,, pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.py,sha256=EFrDbzAs-pV-Fm1HeuAqLl2mJ6MS23LwxYiukKrtf9U,5724 pip/_vendor/urllib3/poolmanager.py,sha256=JYUyBUN3IiEknUdjZ7VJrpCQr6SP7vi0WwSndrn8XpE,17053 pip/_vendor/urllib3/request.py,sha256=hhoHvEEatyd9Tn5EbGjQ0emn-ENMCyY591yNWTneINA,6018 pip/_vendor/urllib3/response.py,sha256=O2DVzBeWOzyxZDZ8k0EDFU3GW1jWXk_b03mS0O1ybxs,27836 pip/_vendor/urllib3/util/__init__.py,sha256=bWNaav_OT-1L7-sxm59cGb59rDORlbhb_4noduM5m0U,1038 pip/_vendor/urllib3/util/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/connection.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/queue.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/request.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/response.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/retry.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/ssl_.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/timeout.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/url.cpython-38.pyc,, pip/_vendor/urllib3/util/__pycache__/wait.cpython-38.pyc,, pip/_vendor/urllib3/util/connection.py,sha256=fOXAQ288KHaXmdTlDv5drwScDk9taQ9YzT42xegxzTg,4636 pip/_vendor/urllib3/util/queue.py,sha256=myTX3JDHntglKQNBf3b6dasHH-uF-W59vzGSQiFdAfI,497 pip/_vendor/urllib3/util/request.py,sha256=_LJPqQXTYA_9c0syijP8Bmj72BeKNO9PXBk62oM7HQY,3821 pip/_vendor/urllib3/util/response.py,sha256=_WbTQr8xRQuJuY2rTIZxVdJD6mnEOtQupjaK_bF_Vj8,2573 pip/_vendor/urllib3/util/retry.py,sha256=Ui74h44gLIIWkAxT9SK3A2mEvu55-odWgJMw3LiUNGk,15450 pip/_vendor/urllib3/util/ssl_.py,sha256=7mB3AsidIqLLq6gbeBL-7Ta0MyVOL5uZax8_5bH3y7c,14163 pip/_vendor/urllib3/util/timeout.py,sha256=2g39u7rU68ilOcGhP1sVzySm4yWDbiY1LxCjsrCrMk8,9874 pip/_vendor/urllib3/util/url.py,sha256=8rFC9RmQVDZuYlqBvlDUm7wcG7N4emdEWe0jQ101lzE,14313 pip/_vendor/urllib3/util/wait.py,sha256=k46KzqIYu3Vnzla5YW3EvtInNlU_QycFqQAghIOxoAg,5406 pip/_vendor/webencodings/__init__.py,sha256=qOBJIuPy_4ByYH6W_bNgJF-qYQ2DoU-dKsDu5yRWCXg,10579 pip/_vendor/webencodings/__pycache__/__init__.cpython-38.pyc,, pip/_vendor/webencodings/__pycache__/labels.cpython-38.pyc,, pip/_vendor/webencodings/__pycache__/mklabels.cpython-38.pyc,, pip/_vendor/webencodings/__pycache__/tests.cpython-38.pyc,, pip/_vendor/webencodings/__pycache__/x_user_defined.cpython-38.pyc,, pip/_vendor/webencodings/labels.py,sha256=4AO_KxTddqGtrL9ns7kAPjb0CcN6xsCIxbK37HY9r3E,8979 pip/_vendor/webencodings/mklabels.py,sha256=GYIeywnpaLnP0GSic8LFWgd0UVvO_l1Nc6YoF-87R_4,1305 pip/_vendor/webencodings/tests.py,sha256=OtGLyjhNY1fvkW1GvLJ_FV9ZoqC9Anyjr7q3kxTbzNs,6563 pip/_vendor/webencodings/x_user_defined.py,sha256=yOqWSdmpytGfUgh_Z6JYgDNhoc-BAHyyeeT15Fr42tM,4307 site-packages/pip-19.3.1.dist-info/top_level.txt000064400000000004151464452710015233 0ustar00pip site-packages/pip-19.3.1.dist-info/entry_points.txt000064400000000161151464452710016003 0ustar00[console_scripts] pip = pip._internal.main:main pip3 = pip._internal.main:main pip3.8 = pip._internal.main:main site-packages/pip-19.3.1.dist-info/INSTALLER000064400000000004151464452710013761 0ustar00rpm site-packages/setuptools-41.6.0.dist-info/zip-safe000064400000000001151464452710015565 0ustar00 site-packages/setuptools-41.6.0.dist-info/LICENSE000064400000002066151464452710015146 0ustar00Copyright (C) 2016 Jason R Coombs Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. site-packages/setuptools-41.6.0.dist-info/dependency_links.txt000064400000000357151464452710020221 0ustar00https://files.pythonhosted.org/packages/source/c/certifi/certifi-2016.9.26.tar.gz#md5=baa81e951a29958563689d868ef1064d https://files.pythonhosted.org/packages/source/w/wincertstore/wincertstore-0.2.zip#md5=ae728f2f007185648d0c7a8679b361e2 site-packages/setuptools-41.6.0.dist-info/WHEEL000064400000000156151464452710014726 0ustar00Wheel-Version: 1.0 Generator: bdist_wheel (0.33.6) Root-Is-Purelib: true Tag: py2-none-any Tag: py3-none-any site-packages/setuptools-41.6.0.dist-info/METADATA000064400000006433151464452710015246 0ustar00Metadata-Version: 2.1 Name: setuptools Version: 41.6.0 Summary: Easily download, build, install, upgrade, and uninstall Python packages Home-page: https://github.com/pypa/setuptools Author: Python Packaging Authority Author-email: distutils-sig@python.org License: UNKNOWN Project-URL: Documentation, https://setuptools.readthedocs.io/ Keywords: CPAN PyPI distutils eggs package management Platform: UNKNOWN Classifier: Development Status :: 5 - Production/Stable Classifier: Intended Audience :: Developers Classifier: License :: OSI Approved :: MIT License Classifier: Operating System :: OS Independent Classifier: Programming Language :: Python :: 2 Classifier: Programming Language :: Python :: 2.7 Classifier: Programming Language :: Python :: 3 Classifier: Programming Language :: Python :: 3.4 Classifier: Programming Language :: Python :: 3.5 Classifier: Programming Language :: Python :: 3.6 Classifier: Programming Language :: Python :: 3.7 Classifier: Programming Language :: Python :: 3.8 Classifier: Topic :: Software Development :: Libraries :: Python Modules Classifier: Topic :: System :: Archiving :: Packaging Classifier: Topic :: System :: Systems Administration Classifier: Topic :: Utilities Requires-Python: !=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,>=2.7 Description-Content-Type: text/x-rst; charset=UTF-8 Provides-Extra: certs Requires-Dist: certifi (==2016.9.26) ; extra == 'certs' Provides-Extra: ssl Requires-Dist: wincertstore (==0.2) ; (sys_platform == "win32") and extra == 'ssl' .. image:: https://img.shields.io/pypi/v/setuptools.svg :target: https://pypi.org/project/setuptools .. image:: https://img.shields.io/readthedocs/setuptools/latest.svg :target: https://setuptools.readthedocs.io .. image:: https://img.shields.io/travis/pypa/setuptools/master.svg?label=Linux%20CI&logo=travis&logoColor=white :target: https://travis-ci.org/pypa/setuptools .. image:: https://img.shields.io/appveyor/ci/pypa/setuptools/master.svg?label=Windows%20CI&logo=appveyor&logoColor=white :target: https://ci.appveyor.com/project/pypa/setuptools/branch/master .. image:: https://img.shields.io/codecov/c/github/pypa/setuptools/master.svg?logo=codecov&logoColor=white :target: https://codecov.io/gh/pypa/setuptools .. image:: https://tidelift.com/badges/github/pypa/setuptools?style=flat :target: https://tidelift.com/subscription/pkg/pypi-setuptools?utm_source=pypi-setuptools&utm_medium=readme .. image:: https://img.shields.io/pypi/pyversions/setuptools.svg See the `Installation Instructions `_ in the Python Packaging User's Guide for instructions on installing, upgrading, and uninstalling Setuptools. Questions and comments should be directed to the `distutils-sig mailing list `_. Bug reports and especially tested patches may be submitted directly to the `bug tracker `_. To report a security vulnerability, please use the `Tidelift security contact `_. Tidelift will coordinate the fix and disclosure. Code of Conduct --------------- Everyone interacting in the setuptools project's codebases, issue trackers, chat rooms, and mailing lists is expected to follow the `PyPA Code of Conduct `_. site-packages/setuptools-41.6.0.dist-info/RECORD000064400000032734151464452710015047 0ustar00../../../bin/easy_install,sha256=kyVHjd3KmOt_Hd-ehMBXo7pGpLnKgL4ShtrVFRGdt3E,232 ../../../bin/easy_install-3.8,sha256=kyVHjd3KmOt_Hd-ehMBXo7pGpLnKgL4ShtrVFRGdt3E,232 __pycache__/easy_install.cpython-38.pyc,, easy_install.py,sha256=MDC9vt5AxDsXX5qcKlBz2TnW6Tpuv_AobnfhCJ9X3PM,126 pkg_resources/__init__.py,sha256=8L9Xzt6qtgAwlbk95DVBMbFflwoxjIbnikZkRjdJU7U,108349 pkg_resources/__pycache__/__init__.cpython-38.pyc,, pkg_resources/__pycache__/py31compat.cpython-38.pyc,, pkg_resources/_vendor/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 pkg_resources/_vendor/__pycache__/__init__.cpython-38.pyc,, pkg_resources/_vendor/__pycache__/appdirs.cpython-38.pyc,, pkg_resources/_vendor/__pycache__/pyparsing.cpython-38.pyc,, pkg_resources/_vendor/__pycache__/six.cpython-38.pyc,, pkg_resources/_vendor/appdirs.py,sha256=C1zpkt179PgsJ4dspL09N-HhOxy6bkTSxUj2_vG6Hkw,24679 pkg_resources/_vendor/packaging/__about__.py,sha256=zkcCPTN_6TcLW0Nrlg0176-R1QQ_WVPTm8sz1R4-HjM,720 pkg_resources/_vendor/packaging/__init__.py,sha256=_vNac5TrzwsrzbOFIbF-5cHqc_Y2aPT2D7zrIR06BOo,513 pkg_resources/_vendor/packaging/__pycache__/__about__.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/__init__.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/_compat.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/_structures.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/markers.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/requirements.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/specifiers.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/utils.cpython-38.pyc,, pkg_resources/_vendor/packaging/__pycache__/version.cpython-38.pyc,, pkg_resources/_vendor/packaging/_compat.py,sha256=Vi_A0rAQeHbU-a9X0tt1yQm9RqkgQbDSxzRw8WlU9kA,860 pkg_resources/_vendor/packaging/_structures.py,sha256=RImECJ4c_wTlaTYYwZYLHEiebDMaAJmK1oPARhw1T5o,1416 pkg_resources/_vendor/packaging/markers.py,sha256=uEcBBtGvzqltgnArqb9c4RrcInXezDLos14zbBHhWJo,8248 pkg_resources/_vendor/packaging/requirements.py,sha256=SikL2UynbsT0qtY9ltqngndha_sfo0w6XGFhAhoSoaQ,4355 pkg_resources/_vendor/packaging/specifiers.py,sha256=SAMRerzO3fK2IkFZCaZkuwZaL_EGqHNOz4pni4vhnN0,28025 pkg_resources/_vendor/packaging/utils.py,sha256=3m6WvPm6NNxE8rkTGmn0r75B_GZSGg7ikafxHsBN1WA,421 pkg_resources/_vendor/packaging/version.py,sha256=OwGnxYfr2ghNzYx59qWIBkrK3SnB6n-Zfd1XaLpnnM0,11556 pkg_resources/_vendor/pyparsing.py,sha256=tmrp-lu-qO1i75ZzIN5A12nKRRD1Cm4Vpk-5LR9rims,232055 pkg_resources/_vendor/six.py,sha256=A6hdJZVjI3t_geebZ9BzUvwRrIXo0lfwzQlM2LcKyas,30098 pkg_resources/extern/__init__.py,sha256=cHiEfHuLmm6rs5Ve_ztBfMI7Lr31vss-D4wkqF5xzlI,2498 pkg_resources/extern/__pycache__/__init__.cpython-38.pyc,, pkg_resources/py31compat.py,sha256=-WQ0e4c3RG_acdhwC3gLiXhP_lg4G5q7XYkZkQg0gxU,558 setuptools-41.6.0.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4 setuptools-41.6.0.dist-info/LICENSE,sha256=wyo6w5WvYyHv0ovnPQagDw22q4h9HCHU_sRhKNIFbVo,1078 setuptools-41.6.0.dist-info/METADATA,sha256=C4M3emJvpj_5ir5cWHN7qwqzqSdoNn90zBzQzCZcfEw,3355 setuptools-41.6.0.dist-info/RECORD,, setuptools-41.6.0.dist-info/WHEEL,sha256=8zNYZbwQSXoB9IfXOjPfeNwvAsALAjffgk27FqvCWbo,110 setuptools-41.6.0.dist-info/dependency_links.txt,sha256=HlkCFkoK5TbZ5EMLbLKYhLcY_E31kBWD8TqW2EgmatQ,239 setuptools-41.6.0.dist-info/entry_points.txt,sha256=t51hYQSHTv2cDVO8fdyd3huPmuBPr2aDcevvakP7lbI,2990 setuptools-41.6.0.dist-info/top_level.txt,sha256=2HUXVVwA4Pff1xgTFr3GsTXXKaPaO6vlG6oNJ_4u4Tg,38 setuptools-41.6.0.dist-info/zip-safe,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1 setuptools/__init__.py,sha256=WBpCcn2lvdckotabeae1TTYonPOcgCIF3raD2zRWzBc,7283 setuptools/__pycache__/__init__.cpython-38.pyc,, setuptools/__pycache__/_deprecation_warning.cpython-38.pyc,, setuptools/__pycache__/_imp.cpython-38.pyc,, setuptools/__pycache__/archive_util.cpython-38.pyc,, setuptools/__pycache__/build_meta.cpython-38.pyc,, setuptools/__pycache__/config.cpython-38.pyc,, setuptools/__pycache__/dep_util.cpython-38.pyc,, setuptools/__pycache__/depends.cpython-38.pyc,, setuptools/__pycache__/dist.cpython-38.pyc,, setuptools/__pycache__/extension.cpython-38.pyc,, setuptools/__pycache__/glibc.cpython-38.pyc,, setuptools/__pycache__/glob.cpython-38.pyc,, setuptools/__pycache__/launch.cpython-38.pyc,, setuptools/__pycache__/lib2to3_ex.cpython-38.pyc,, setuptools/__pycache__/monkey.cpython-38.pyc,, setuptools/__pycache__/msvc.cpython-38.pyc,, setuptools/__pycache__/namespaces.cpython-38.pyc,, setuptools/__pycache__/package_index.cpython-38.pyc,, setuptools/__pycache__/pep425tags.cpython-38.pyc,, setuptools/__pycache__/py27compat.cpython-38.pyc,, setuptools/__pycache__/py31compat.cpython-38.pyc,, setuptools/__pycache__/py33compat.cpython-38.pyc,, setuptools/__pycache__/py34compat.cpython-38.pyc,, setuptools/__pycache__/sandbox.cpython-38.pyc,, setuptools/__pycache__/site-patch.cpython-38.pyc,, setuptools/__pycache__/ssl_support.cpython-38.pyc,, setuptools/__pycache__/unicode_utils.cpython-38.pyc,, setuptools/__pycache__/version.cpython-38.pyc,, setuptools/__pycache__/wheel.cpython-38.pyc,, setuptools/__pycache__/windows_support.cpython-38.pyc,, setuptools/_deprecation_warning.py,sha256=jU9-dtfv6cKmtQJOXN8nP1mm7gONw5kKEtiPtbwnZyI,218 setuptools/_imp.py,sha256=jloslOkxrTKbobgemfP94YII0nhqiJzE1bRmCTZ1a5I,2223 setuptools/_vendor/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 setuptools/_vendor/__pycache__/__init__.cpython-38.pyc,, setuptools/_vendor/__pycache__/ordered_set.cpython-38.pyc,, setuptools/_vendor/__pycache__/pyparsing.cpython-38.pyc,, setuptools/_vendor/__pycache__/six.cpython-38.pyc,, setuptools/_vendor/ordered_set.py,sha256=dbaCcs27dyN9gnMWGF5nA_BrVn6Q-NrjKYJpV9_fgBs,15130 setuptools/_vendor/packaging/__about__.py,sha256=zkcCPTN_6TcLW0Nrlg0176-R1QQ_WVPTm8sz1R4-HjM,720 setuptools/_vendor/packaging/__init__.py,sha256=_vNac5TrzwsrzbOFIbF-5cHqc_Y2aPT2D7zrIR06BOo,513 setuptools/_vendor/packaging/__pycache__/__about__.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/__init__.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/_compat.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/_structures.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/markers.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/requirements.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/specifiers.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/utils.cpython-38.pyc,, setuptools/_vendor/packaging/__pycache__/version.cpython-38.pyc,, setuptools/_vendor/packaging/_compat.py,sha256=Vi_A0rAQeHbU-a9X0tt1yQm9RqkgQbDSxzRw8WlU9kA,860 setuptools/_vendor/packaging/_structures.py,sha256=RImECJ4c_wTlaTYYwZYLHEiebDMaAJmK1oPARhw1T5o,1416 setuptools/_vendor/packaging/markers.py,sha256=Gvpk9EY20yKaMTiKgQZ8yFEEpodqVgVYtfekoic1Yts,8239 setuptools/_vendor/packaging/requirements.py,sha256=t44M2HVWtr8phIz2OhnILzuGT3rTATaovctV1dpnVIg,4343 setuptools/_vendor/packaging/specifiers.py,sha256=SAMRerzO3fK2IkFZCaZkuwZaL_EGqHNOz4pni4vhnN0,28025 setuptools/_vendor/packaging/utils.py,sha256=3m6WvPm6NNxE8rkTGmn0r75B_GZSGg7ikafxHsBN1WA,421 setuptools/_vendor/packaging/version.py,sha256=OwGnxYfr2ghNzYx59qWIBkrK3SnB6n-Zfd1XaLpnnM0,11556 setuptools/_vendor/pyparsing.py,sha256=tmrp-lu-qO1i75ZzIN5A12nKRRD1Cm4Vpk-5LR9rims,232055 setuptools/_vendor/six.py,sha256=A6hdJZVjI3t_geebZ9BzUvwRrIXo0lfwzQlM2LcKyas,30098 setuptools/archive_util.py,sha256=kw8Ib_lKjCcnPKNbS7h8HztRVK0d5RacU3r_KRdVnmM,6592 setuptools/build_meta.py,sha256=-9Nmj9YdbW4zX3TssPJZhsENrTa4fw3k86Jm1cdKMik,9597 setuptools/command/__init__.py,sha256=NWzJ0A1BEengZpVeqUyWLNm2bk4P3F4iL5QUErHy7kA,594 setuptools/command/__pycache__/__init__.cpython-38.pyc,, setuptools/command/__pycache__/alias.cpython-38.pyc,, setuptools/command/__pycache__/bdist_egg.cpython-38.pyc,, setuptools/command/__pycache__/bdist_rpm.cpython-38.pyc,, setuptools/command/__pycache__/bdist_wininst.cpython-38.pyc,, setuptools/command/__pycache__/build_clib.cpython-38.pyc,, setuptools/command/__pycache__/build_ext.cpython-38.pyc,, setuptools/command/__pycache__/build_py.cpython-38.pyc,, setuptools/command/__pycache__/develop.cpython-38.pyc,, setuptools/command/__pycache__/dist_info.cpython-38.pyc,, setuptools/command/__pycache__/easy_install.cpython-38.pyc,, setuptools/command/__pycache__/egg_info.cpython-38.pyc,, setuptools/command/__pycache__/install.cpython-38.pyc,, setuptools/command/__pycache__/install_egg_info.cpython-38.pyc,, setuptools/command/__pycache__/install_lib.cpython-38.pyc,, setuptools/command/__pycache__/install_scripts.cpython-38.pyc,, setuptools/command/__pycache__/py36compat.cpython-38.pyc,, setuptools/command/__pycache__/register.cpython-38.pyc,, setuptools/command/__pycache__/rotate.cpython-38.pyc,, setuptools/command/__pycache__/saveopts.cpython-38.pyc,, setuptools/command/__pycache__/sdist.cpython-38.pyc,, setuptools/command/__pycache__/setopt.cpython-38.pyc,, setuptools/command/__pycache__/test.cpython-38.pyc,, setuptools/command/__pycache__/upload.cpython-38.pyc,, setuptools/command/__pycache__/upload_docs.cpython-38.pyc,, setuptools/command/alias.py,sha256=KjpE0sz_SDIHv3fpZcIQK-sCkJz-SrC6Gmug6b9Nkc8,2426 setuptools/command/bdist_egg.py,sha256=nnfV8Ah8IRC_Ifv5Loa9FdxL66MVbyDXwy-foP810zM,18185 setuptools/command/bdist_rpm.py,sha256=B7l0TnzCGb-0nLlm6rS00jWLkojASwVmdhW2w5Qz_Ak,1508 setuptools/command/bdist_wininst.py,sha256=_6dz3lpB1tY200LxKPLM7qgwTCceOMgaWFF-jW2-pm0,637 setuptools/command/build_clib.py,sha256=bQ9aBr-5ZSO-9fGsGsDLz0mnnFteHUZnftVLkhvHDq0,4484 setuptools/command/build_ext.py,sha256=Ib42YUGksBswm2mL5xmQPF6NeTA6HcqrvAtEgFCv32A,13019 setuptools/command/build_py.py,sha256=yWyYaaS9F3o9JbIczn064A5g1C5_UiKRDxGaTqYbtLE,9596 setuptools/command/develop.py,sha256=MQlnGS6uP19erK2JCNOyQYoYyquk3PADrqrrinqqLtA,8184 setuptools/command/dist_info.py,sha256=5t6kOfrdgALT-P3ogss6PF9k-Leyesueycuk3dUyZnI,960 setuptools/command/easy_install.py,sha256=BLJBFm2-rtp-ZZj0SoN1KAd3GH-LGhKV5MrcgIuFx_8,87453 setuptools/command/egg_info.py,sha256=w73EdxYSOk2gsaAiHGL2dZrCldoPiuRr2eTfqcFvCds,25570 setuptools/command/install.py,sha256=a0EZpL_A866KEdhicTGbuyD_TYl1sykfzdrri-zazT4,4683 setuptools/command/install_egg_info.py,sha256=bMgeIeRiXzQ4DAGPV1328kcjwQjHjOWU4FngAWLV78Q,2203 setuptools/command/install_lib.py,sha256=r5NuasaSxvmIrjgZNj38Iq-1UJG1o1ms7CuHq6MCTbQ,3862 setuptools/command/install_scripts.py,sha256=UD0rEZ6861mTYhIdzcsqKnUl8PozocXWl9VBQ1VTWnc,2439 setuptools/command/launcher manifest.xml,sha256=xlLbjWrB01tKC0-hlVkOKkiSPbzMml2eOPtJ_ucCnbE,628 setuptools/command/py36compat.py,sha256=SzjZcOxF7zdFUT47Zv2n7AM3H8koDys_0OpS-n9gIfc,4986 setuptools/command/register.py,sha256=LO3MvYKPE8dN1m-KkrBRHC68ZFoPvA_vI8Xgp7vv6zI,534 setuptools/command/rotate.py,sha256=co5C1EkI7P0GGT6Tqz-T2SIj2LBJTZXYELpmao6d4KQ,2164 setuptools/command/saveopts.py,sha256=za7QCBcQimKKriWcoCcbhxPjUz30gSB74zuTL47xpP4,658 setuptools/command/sdist.py,sha256=gr5hFrDzUtGfp_0tu0sllzIyr3jMQegIkFmlDauQJxw,7388 setuptools/command/setopt.py,sha256=NTWDyx-gjDF-txf4dO577s7LOzHVoKR0Mq33rFxaRr8,5085 setuptools/command/test.py,sha256=u2kXngIIdSYqtvwFlHiN6Iye1IB4TU6uadB2uiV1szw,9602 setuptools/command/upload.py,sha256=GxtNkIl7SA0r8mACkbDcSCN1m2_WPppK9gZXJmQSiow,6811 setuptools/command/upload_docs.py,sha256=oXiGplM_cUKLwE4CWWw98RzCufAu8tBhMC97GegFcms,7311 setuptools/config.py,sha256=Rek2y8zh67yIc5Rs4biqo63jOwYrvsy5gy8SSDJbWU0,20534 setuptools/dep_util.py,sha256=fgixvC1R7sH3r13ktyf7N0FALoqEXL1cBarmNpSEoWg,935 setuptools/depends.py,sha256=qt2RWllArRvhnm8lxsyRpcthEZYp4GHQgREl1q0LkFw,5517 setuptools/dist.py,sha256=awhNuAZVxwnjntKLjGeUKhSWiCaBsetnPyYQ9LDUUls,50309 setuptools/extension.py,sha256=uc6nHI-MxwmNCNPbUiBnybSyqhpJqjbhvOQ-emdvt_E,1729 setuptools/extern/__init__.py,sha256=4q9gtShB1XFP6CisltsyPqtcfTO6ZM9Lu1QBl3l-qmo,2514 setuptools/extern/__pycache__/__init__.cpython-38.pyc,, setuptools/glibc.py,sha256=X64VvGPL2AbURKwYRsWJOXXGAYOiF_v2qixeTkAULuU,3146 setuptools/glob.py,sha256=o75cHrOxYsvn854thSxE0x9k8JrKDuhP_rRXlVB00Q4,5084 setuptools/launch.py,sha256=sd7ejwhBocCDx_wG9rIs0OaZ8HtmmFU8ZC6IR_S0Lvg,787 setuptools/lib2to3_ex.py,sha256=t5e12hbR2pi9V4ezWDTB4JM-AISUnGOkmcnYHek3xjg,2013 setuptools/monkey.py,sha256=FGc9fffh7gAxMLFmJs2DW_OYWpBjkdbNS2n14UAK4NA,5264 setuptools/msvc.py,sha256=8baJ6aYgCA4TRdWQQi185qB9dnU8FaP4wgpbmd7VODs,46751 setuptools/namespaces.py,sha256=F0Nrbv8KCT2OrO7rwa03om4N4GZKAlnce-rr-cgDQa8,3199 setuptools/package_index.py,sha256=6pb-B1POtHyLycAbkDETk4fO-Qv8_sY-rjTXhUOoh6k,40605 setuptools/pep425tags.py,sha256=o_D_WVeWcXZiI2xjPSg7pouGOvaWRgGRxEDK9DzAXIA,10861 setuptools/py27compat.py,sha256=tvmer0Tn-wk_JummCkoM22UIjpjL-AQ8uUiOaqTs8sI,1496 setuptools/py31compat.py,sha256=h2rtZghOfwoGYd8sQ0-auaKiF3TcL3qX0bX3VessqcE,838 setuptools/py33compat.py,sha256=SMF9Z8wnGicTOkU1uRNwZ_kz5Z_bj29PUBbqdqeeNsc,1330 setuptools/py34compat.py,sha256=KYOd6ybRxjBW8NJmYD8t_UyyVmysppFXqHpFLdslGXU,245 setuptools/sandbox.py,sha256=9UbwfEL5QY436oMI1LtFWohhoZ-UzwHvGyZjUH_qhkw,14276 setuptools/script (dev).tmpl,sha256=RUzQzCQUaXtwdLtYHWYbIQmOaES5Brqq1FvUA_tu-5I,218 setuptools/script.tmpl,sha256=WGTt5piezO27c-Dbx6l5Q4T3Ff20A5z7872hv3aAhYY,138 setuptools/site-patch.py,sha256=OumkIHMuoSenRSW1382kKWI1VAwxNE86E5W8iDd34FY,2302 setuptools/ssl_support.py,sha256=nLjPUBBw7RTTx6O4RJZ5eAMGgjJG8beiDbkFXDZpLuM,8493 setuptools/unicode_utils.py,sha256=NOiZ_5hD72A6w-4wVj8awHFM3n51Kmw1Ic_vx15XFqw,996 setuptools/version.py,sha256=og_cuZQb0QI6ukKZFfZWPlr1HgJBPPn2vO2m_bI9ZTE,144 setuptools/wheel.py,sha256=94uqXsOaKt91d9hW5z6ZppZmNSs_nO66R4uiwhcr4V0,8094 setuptools/windows_support.py,sha256=5GrfqSP2-dLGJoZTq2g6dCKkyQxxa2n5IQiXlJCoYEE,714 site-packages/setuptools-41.6.0.dist-info/top_level.txt000064400000000046151464452710016667 0ustar00easy_install pkg_resources setuptools site-packages/setuptools-41.6.0.dist-info/entry_points.txt000064400000005656151464452710017447 0ustar00[console_scripts] easy_install = setuptools.command.easy_install:main easy_install-3.8 = setuptools.command.easy_install:main [distutils.commands] alias = setuptools.command.alias:alias bdist_egg = setuptools.command.bdist_egg:bdist_egg bdist_rpm = setuptools.command.bdist_rpm:bdist_rpm bdist_wininst = setuptools.command.bdist_wininst:bdist_wininst build_clib = setuptools.command.build_clib:build_clib build_ext = setuptools.command.build_ext:build_ext build_py = setuptools.command.build_py:build_py develop = setuptools.command.develop:develop dist_info = setuptools.command.dist_info:dist_info easy_install = setuptools.command.easy_install:easy_install egg_info = setuptools.command.egg_info:egg_info install = setuptools.command.install:install install_egg_info = setuptools.command.install_egg_info:install_egg_info install_lib = setuptools.command.install_lib:install_lib install_scripts = setuptools.command.install_scripts:install_scripts register = setuptools.command.register:register rotate = setuptools.command.rotate:rotate saveopts = setuptools.command.saveopts:saveopts sdist = setuptools.command.sdist:sdist setopt = setuptools.command.setopt:setopt test = setuptools.command.test:test upload = setuptools.command.upload:upload upload_docs = setuptools.command.upload_docs:upload_docs [distutils.setup_keywords] convert_2to3_doctests = setuptools.dist:assert_string_list dependency_links = setuptools.dist:assert_string_list eager_resources = setuptools.dist:assert_string_list entry_points = setuptools.dist:check_entry_points exclude_package_data = setuptools.dist:check_package_data extras_require = setuptools.dist:check_extras include_package_data = setuptools.dist:assert_bool install_requires = setuptools.dist:check_requirements namespace_packages = setuptools.dist:check_nsp package_data = setuptools.dist:check_package_data packages = setuptools.dist:check_packages python_requires = setuptools.dist:check_specifier setup_requires = setuptools.dist:check_requirements test_loader = setuptools.dist:check_importable test_runner = setuptools.dist:check_importable test_suite = setuptools.dist:check_test_suite tests_require = setuptools.dist:check_requirements use_2to3 = setuptools.dist:assert_bool use_2to3_exclude_fixers = setuptools.dist:assert_string_list use_2to3_fixers = setuptools.dist:assert_string_list zip_safe = setuptools.dist:assert_bool [egg_info.writers] PKG-INFO = setuptools.command.egg_info:write_pkg_info dependency_links.txt = setuptools.command.egg_info:overwrite_arg depends.txt = setuptools.command.egg_info:warn_depends_obsolete eager_resources.txt = setuptools.command.egg_info:overwrite_arg entry_points.txt = setuptools.command.egg_info:write_entries namespace_packages.txt = setuptools.command.egg_info:overwrite_arg requires.txt = setuptools.command.egg_info:write_requirements top_level.txt = setuptools.command.egg_info:write_toplevel_names [setuptools.installation] eggsecutable = setuptools.command.easy_install:bootstrap site-packages/setuptools-41.6.0.dist-info/INSTALLER000064400000000004151464452710015407 0ustar00pip site-packages/setuptools/glibc.py000064400000006112151464452710013150 0ustar00# This file originally from pip: # https://github.com/pypa/pip/blob/8f4f15a5a95d7d5b511ceaee9ed261176c181970/src/pip/_internal/utils/glibc.py from __future__ import absolute_import import ctypes import re import warnings def glibc_version_string(): "Returns glibc version string, or None if not using glibc." # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen # manpage says, "If filename is NULL, then the returned handle is for the # main program". This way we can let the linker do the work to figure out # which libc our process is actually using. process_namespace = ctypes.CDLL(None) try: gnu_get_libc_version = process_namespace.gnu_get_libc_version except AttributeError: # Symbol doesn't exist -> therefore, we are not linked to # glibc. return None # Call gnu_get_libc_version, which returns a string like "2.5" gnu_get_libc_version.restype = ctypes.c_char_p version_str = gnu_get_libc_version() # py2 / py3 compatibility: if not isinstance(version_str, str): version_str = version_str.decode("ascii") return version_str # Separated out from have_compatible_glibc for easier unit testing def check_glibc_version(version_str, required_major, minimum_minor): # Parse string and check against requested version. # # We use a regexp instead of str.split because we want to discard any # random junk that might come after the minor version -- this might happen # in patched/forked versions of glibc (e.g. Linaro's version of glibc # uses version strings like "2.20-2014.11"). See gh-3588. m = re.match(r"(?P[0-9]+)\.(?P[0-9]+)", version_str) if not m: warnings.warn("Expected glibc version with 2 components major.minor," " got: %s" % version_str, RuntimeWarning) return False return (int(m.group("major")) == required_major and int(m.group("minor")) >= minimum_minor) def have_compatible_glibc(required_major, minimum_minor): version_str = glibc_version_string() if version_str is None: return False return check_glibc_version(version_str, required_major, minimum_minor) # platform.libc_ver regularly returns completely nonsensical glibc # versions. E.g. on my computer, platform says: # # ~$ python2.7 -c 'import platform; print(platform.libc_ver())' # ('glibc', '2.7') # ~$ python3.5 -c 'import platform; print(platform.libc_ver())' # ('glibc', '2.9') # # But the truth is: # # ~$ ldd --version # ldd (Debian GLIBC 2.22-11) 2.22 # # This is unfortunate, because it means that the linehaul data on libc # versions that was generated by pip 8.1.2 and earlier is useless and # misleading. Solution: instead of using platform, use our code that actually # works. def libc_ver(): """Try to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings in case the lookup fails. """ glibc_version = glibc_version_string() if glibc_version is None: return ("", "") else: return ("glibc", glibc_version) site-packages/setuptools/pep425tags.py000064400000025155151464452710013776 0ustar00# This file originally from pip: # https://github.com/pypa/pip/blob/8f4f15a5a95d7d5b511ceaee9ed261176c181970/src/pip/_internal/pep425tags.py """Generate and work with PEP 425 Compatibility Tags.""" from __future__ import absolute_import import distutils.util from distutils import log import platform import re import sys import sysconfig import warnings from collections import OrderedDict from .extern import six from . import glibc _osx_arch_pat = re.compile(r'(.+)_(\d+)_(\d+)_(.+)') def get_config_var(var): try: return sysconfig.get_config_var(var) except IOError as e: # Issue #1074 warnings.warn("{}".format(e), RuntimeWarning) return None def get_abbr_impl(): """Return abbreviated implementation name.""" if hasattr(sys, 'pypy_version_info'): pyimpl = 'pp' elif sys.platform.startswith('java'): pyimpl = 'jy' elif sys.platform == 'cli': pyimpl = 'ip' else: pyimpl = 'cp' return pyimpl def get_impl_ver(): """Return implementation version.""" impl_ver = get_config_var("py_version_nodot") if not impl_ver or get_abbr_impl() == 'pp': impl_ver = ''.join(map(str, get_impl_version_info())) return impl_ver def get_impl_version_info(): """Return sys.version_info-like tuple for use in decrementing the minor version.""" if get_abbr_impl() == 'pp': # as per https://github.com/pypa/pip/issues/2882 return (sys.version_info[0], sys.pypy_version_info.major, sys.pypy_version_info.minor) else: return sys.version_info[0], sys.version_info[1] def get_impl_tag(): """ Returns the Tag for this specific implementation. """ return "{}{}".format(get_abbr_impl(), get_impl_ver()) def get_flag(var, fallback, expected=True, warn=True): """Use a fallback method for determining SOABI flags if the needed config var is unset or unavailable.""" val = get_config_var(var) if val is None: if warn: log.debug("Config variable '%s' is unset, Python ABI tag may " "be incorrect", var) return fallback() return val == expected def get_abi_tag(): """Return the ABI tag based on SOABI (if available) or emulate SOABI (CPython 2, PyPy).""" soabi = get_config_var('SOABI') impl = get_abbr_impl() if not soabi and impl in {'cp', 'pp'} and hasattr(sys, 'maxunicode'): d = '' m = '' u = '' if get_flag('Py_DEBUG', lambda: hasattr(sys, 'gettotalrefcount'), warn=(impl == 'cp')): d = 'd' if get_flag('WITH_PYMALLOC', lambda: impl == 'cp', warn=(impl == 'cp')): m = 'm' if get_flag('Py_UNICODE_SIZE', lambda: sys.maxunicode == 0x10ffff, expected=4, warn=(impl == 'cp' and six.PY2)) \ and six.PY2: u = 'u' abi = '%s%s%s%s%s' % (impl, get_impl_ver(), d, m, u) elif soabi and soabi.startswith('cpython-'): abi = 'cp' + soabi.split('-')[1] elif soabi: abi = soabi.replace('.', '_').replace('-', '_') else: abi = None return abi def _is_running_32bit(): return sys.maxsize == 2147483647 def get_platform(): """Return our platform name 'win32', 'linux_x86_64'""" if sys.platform == 'darwin': # distutils.util.get_platform() returns the release based on the value # of MACOSX_DEPLOYMENT_TARGET on which Python was built, which may # be significantly older than the user's current machine. release, _, machine = platform.mac_ver() split_ver = release.split('.') if machine == "x86_64" and _is_running_32bit(): machine = "i386" elif machine == "ppc64" and _is_running_32bit(): machine = "ppc" return 'macosx_{}_{}_{}'.format(split_ver[0], split_ver[1], machine) # XXX remove distutils dependency result = distutils.util.get_platform().replace('.', '_').replace('-', '_') if result == "linux_x86_64" and _is_running_32bit(): # 32 bit Python program (running on a 64 bit Linux): pip should only # install and run 32 bit compiled extensions in that case. result = "linux_i686" return result def is_manylinux1_compatible(): # Only Linux, and only x86-64 / i686 if get_platform() not in {"linux_x86_64", "linux_i686"}: return False # Check for presence of _manylinux module try: import _manylinux return bool(_manylinux.manylinux1_compatible) except (ImportError, AttributeError): # Fall through to heuristic check below pass # Check glibc version. CentOS 5 uses glibc 2.5. return glibc.have_compatible_glibc(2, 5) def get_darwin_arches(major, minor, machine): """Return a list of supported arches (including group arches) for the given major, minor and machine architecture of a macOS machine. """ arches = [] def _supports_arch(major, minor, arch): # Looking at the application support for macOS versions in the chart # provided by https://en.wikipedia.org/wiki/OS_X#Versions it appears # our timeline looks roughly like: # # 10.0 - Introduces ppc support. # 10.4 - Introduces ppc64, i386, and x86_64 support, however the ppc64 # and x86_64 support is CLI only, and cannot be used for GUI # applications. # 10.5 - Extends ppc64 and x86_64 support to cover GUI applications. # 10.6 - Drops support for ppc64 # 10.7 - Drops support for ppc # # Given that we do not know if we're installing a CLI or a GUI # application, we must be conservative and assume it might be a GUI # application and behave as if ppc64 and x86_64 support did not occur # until 10.5. # # Note: The above information is taken from the "Application support" # column in the chart not the "Processor support" since I believe # that we care about what instruction sets an application can use # not which processors the OS supports. if arch == 'ppc': return (major, minor) <= (10, 5) if arch == 'ppc64': return (major, minor) == (10, 5) if arch == 'i386': return (major, minor) >= (10, 4) if arch == 'x86_64': return (major, minor) >= (10, 5) if arch in groups: for garch in groups[arch]: if _supports_arch(major, minor, garch): return True return False groups = OrderedDict([ ("fat", ("i386", "ppc")), ("intel", ("x86_64", "i386")), ("fat64", ("x86_64", "ppc64")), ("fat32", ("x86_64", "i386", "ppc")), ]) if _supports_arch(major, minor, machine): arches.append(machine) for garch in groups: if machine in groups[garch] and _supports_arch(major, minor, garch): arches.append(garch) arches.append('universal') return arches def get_supported(versions=None, noarch=False, platform=None, impl=None, abi=None): """Return a list of supported tags for each version specified in `versions`. :param versions: a list of string versions, of the form ["33", "32"], or None. The first version will be assumed to support our ABI. :param platform: specify the exact platform you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. :param abi: specify the exact abi you want valid tags for, or None. If None, use the local interpreter abi. """ supported = [] # Versions must be given with respect to the preference if versions is None: versions = [] version_info = get_impl_version_info() major = version_info[:-1] # Support all previous minor Python versions. for minor in range(version_info[-1], -1, -1): versions.append(''.join(map(str, major + (minor,)))) impl = impl or get_abbr_impl() abis = [] abi = abi or get_abi_tag() if abi: abis[0:0] = [abi] abi3s = set() import imp for suffix in imp.get_suffixes(): if suffix[0].startswith('.abi'): abi3s.add(suffix[0].split('.', 2)[1]) abis.extend(sorted(list(abi3s))) abis.append('none') if not noarch: arch = platform or get_platform() if arch.startswith('macosx'): # support macosx-10.6-intel on macosx-10.9-x86_64 match = _osx_arch_pat.match(arch) if match: name, major, minor, actual_arch = match.groups() tpl = '{}_{}_%i_%s'.format(name, major) arches = [] for m in reversed(range(int(minor) + 1)): for a in get_darwin_arches(int(major), m, actual_arch): arches.append(tpl % (m, a)) else: # arch pattern didn't match (?!) arches = [arch] elif platform is None and is_manylinux1_compatible(): arches = [arch.replace('linux', 'manylinux1'), arch] else: arches = [arch] # Current version, current API (built specifically for our Python): for abi in abis: for arch in arches: supported.append(('%s%s' % (impl, versions[0]), abi, arch)) # abi3 modules compatible with older version of Python for version in versions[1:]: # abi3 was introduced in Python 3.2 if version in {'31', '30'}: break for abi in abi3s: # empty set if not Python 3 for arch in arches: supported.append(("%s%s" % (impl, version), abi, arch)) # Has binaries, does not use the Python API: for arch in arches: supported.append(('py%s' % (versions[0][0]), 'none', arch)) # No abi / arch, but requires our implementation: supported.append(('%s%s' % (impl, versions[0]), 'none', 'any')) # Tagged specifically as being cross-version compatible # (with just the major version specified) supported.append(('%s%s' % (impl, versions[0][0]), 'none', 'any')) # No abi / arch, generic Python for i, version in enumerate(versions): supported.append(('py%s' % (version,), 'none', 'any')) if i == 0: supported.append(('py%s' % (version[0]), 'none', 'any')) return supported implementation_tag = get_impl_tag() site-packages/setuptools/command/saveopts.py000064400000001222151464452710015347 0ustar00from setuptools.command.setopt import edit_config, option_base class saveopts(option_base): """Save command-line options to a file""" description = "save supplied options to setup.cfg or other config file" def run(self): dist = self.distribution settings = {} for cmd in dist.command_options: if cmd == 'saveopts': continue # don't save our own options! for opt, (src, val) in dist.get_option_dict(cmd).items(): if src == "command line": settings.setdefault(cmd, {})[opt] = val edit_config(self.filename, settings, self.dry_run) site-packages/setuptools/command/alias.py000064400000004572151464452710014607 0ustar00from distutils.errors import DistutilsOptionError from setuptools.extern.six.moves import map from setuptools.command.setopt import edit_config, option_base, config_file def shquote(arg): """Quote an argument for later parsing by shlex.split()""" for c in '"', "'", "\\", "#": if c in arg: return repr(arg) if arg.split() != [arg]: return repr(arg) return arg class alias(option_base): """Define a shortcut that invokes one or more commands""" description = "define a shortcut to invoke one or more commands" command_consumes_arguments = True user_options = [ ('remove', 'r', 'remove (unset) the alias'), ] + option_base.user_options boolean_options = option_base.boolean_options + ['remove'] def initialize_options(self): option_base.initialize_options(self) self.args = None self.remove = None def finalize_options(self): option_base.finalize_options(self) if self.remove and len(self.args) != 1: raise DistutilsOptionError( "Must specify exactly one argument (the alias name) when " "using --remove" ) def run(self): aliases = self.distribution.get_option_dict('aliases') if not self.args: print("Command Aliases") print("---------------") for alias in aliases: print("setup.py alias", format_alias(alias, aliases)) return elif len(self.args) == 1: alias, = self.args if self.remove: command = None elif alias in aliases: print("setup.py alias", format_alias(alias, aliases)) return else: print("No alias definition found for %r" % alias) return else: alias = self.args[0] command = ' '.join(map(shquote, self.args[1:])) edit_config(self.filename, {'aliases': {alias: command}}, self.dry_run) def format_alias(name, aliases): source, command = aliases[name] if source == config_file('global'): source = '--global-config ' elif source == config_file('user'): source = '--user-config ' elif source == config_file('local'): source = '' else: source = '--filename=%r' % source return source + name + ' ' + command site-packages/setuptools/command/sdist.py000064400000016334151464452710014643 0ustar00from distutils import log import distutils.command.sdist as orig import os import sys import io import contextlib from setuptools.extern import six from .py36compat import sdist_add_defaults import pkg_resources _default_revctrl = list def walk_revctrl(dirname=''): """Find all files under revision control""" for ep in pkg_resources.iter_entry_points('setuptools.file_finders'): for item in ep.load()(dirname): yield item class sdist(sdist_add_defaults, orig.sdist): """Smart sdist that finds anything supported by revision control""" user_options = [ ('formats=', None, "formats for source distribution (comma-separated list)"), ('keep-temp', 'k', "keep the distribution tree around after creating " + "archive file(s)"), ('dist-dir=', 'd', "directory to put the source distribution archive(s) in " "[default: dist]"), ] negative_opt = {} README_EXTENSIONS = ['', '.rst', '.txt', '.md'] READMES = tuple('README{0}'.format(ext) for ext in README_EXTENSIONS) def run(self): self.run_command('egg_info') ei_cmd = self.get_finalized_command('egg_info') self.filelist = ei_cmd.filelist self.filelist.append(os.path.join(ei_cmd.egg_info, 'SOURCES.txt')) self.check_readme() # Run sub commands for cmd_name in self.get_sub_commands(): self.run_command(cmd_name) self.make_distribution() dist_files = getattr(self.distribution, 'dist_files', []) for file in self.archive_files: data = ('sdist', '', file) if data not in dist_files: dist_files.append(data) def initialize_options(self): orig.sdist.initialize_options(self) self._default_to_gztar() def _default_to_gztar(self): # only needed on Python prior to 3.6. if sys.version_info >= (3, 6, 0, 'beta', 1): return self.formats = ['gztar'] def make_distribution(self): """ Workaround for #516 """ with self._remove_os_link(): orig.sdist.make_distribution(self) @staticmethod @contextlib.contextmanager def _remove_os_link(): """ In a context, remove and restore os.link if it exists """ class NoValue: pass orig_val = getattr(os, 'link', NoValue) try: del os.link except Exception: pass try: yield finally: if orig_val is not NoValue: setattr(os, 'link', orig_val) def __read_template_hack(self): # This grody hack closes the template file (MANIFEST.in) if an # exception occurs during read_template. # Doing so prevents an error when easy_install attempts to delete the # file. try: orig.sdist.read_template(self) except Exception: _, _, tb = sys.exc_info() tb.tb_next.tb_frame.f_locals['template'].close() raise # Beginning with Python 2.7.2, 3.1.4, and 3.2.1, this leaky file handle # has been fixed, so only override the method if we're using an earlier # Python. has_leaky_handle = ( sys.version_info < (2, 7, 2) or (3, 0) <= sys.version_info < (3, 1, 4) or (3, 2) <= sys.version_info < (3, 2, 1) ) if has_leaky_handle: read_template = __read_template_hack def _add_defaults_python(self): """getting python files""" if self.distribution.has_pure_modules(): build_py = self.get_finalized_command('build_py') self.filelist.extend(build_py.get_source_files()) # This functionality is incompatible with include_package_data, and # will in fact create an infinite recursion if include_package_data # is True. Use of include_package_data will imply that # distutils-style automatic handling of package_data is disabled if not self.distribution.include_package_data: for _, src_dir, _, filenames in build_py.data_files: self.filelist.extend([os.path.join(src_dir, filename) for filename in filenames]) def _add_defaults_data_files(self): try: if six.PY2: sdist_add_defaults._add_defaults_data_files(self) else: super()._add_defaults_data_files() except TypeError: log.warn("data_files contains unexpected objects") def check_readme(self): for f in self.READMES: if os.path.exists(f): return else: self.warn( "standard file not found: should have one of " + ', '.join(self.READMES) ) def make_release_tree(self, base_dir, files): orig.sdist.make_release_tree(self, base_dir, files) # Save any egg_info command line options used to create this sdist dest = os.path.join(base_dir, 'setup.cfg') if hasattr(os, 'link') and os.path.exists(dest): # unlink and re-copy, since it might be hard-linked, and # we don't want to change the source version os.unlink(dest) self.copy_file('setup.cfg', dest) self.get_finalized_command('egg_info').save_version_info(dest) def _manifest_is_not_generated(self): # check for special comment used in 2.7.1 and higher if not os.path.isfile(self.manifest): return False with io.open(self.manifest, 'rb') as fp: first_line = fp.readline() return (first_line != '# file GENERATED by distutils, do NOT edit\n'.encode()) def read_manifest(self): """Read the manifest file (named by 'self.manifest') and use it to fill in 'self.filelist', the list of files to include in the source distribution. """ log.info("reading manifest file '%s'", self.manifest) manifest = open(self.manifest, 'rb') for line in manifest: # The manifest must contain UTF-8. See #303. if six.PY3: try: line = line.decode('UTF-8') except UnicodeDecodeError: log.warn("%r not UTF-8 decodable -- skipping" % line) continue # ignore comments and blank lines line = line.strip() if line.startswith('#') or not line: continue self.filelist.append(line) manifest.close() def check_license(self): """Checks if license_file' is configured and adds it to 'self.filelist' if the value contains a valid path. """ opts = self.distribution.get_option_dict('metadata') # ignore the source of the value _, license_file = opts.get('license_file', (None, None)) if license_file is None: log.debug("'license_file' option was not specified") return if not os.path.exists(license_file): log.warn("warning: Failed to find the configured license file '%s'", license_file) return self.filelist.append(license_file) site-packages/setuptools/command/develop.py000064400000017770151464452710015160 0ustar00from distutils.util import convert_path from distutils import log from distutils.errors import DistutilsError, DistutilsOptionError import os import glob import io from setuptools.extern import six import pkg_resources from setuptools.command.easy_install import easy_install from setuptools import namespaces import setuptools __metaclass__ = type class develop(namespaces.DevelopInstaller, easy_install): """Set up package for development""" description = "install package in 'development mode'" user_options = easy_install.user_options + [ ("uninstall", "u", "Uninstall this source package"), ("egg-path=", None, "Set the path to be used in the .egg-link file"), ] boolean_options = easy_install.boolean_options + ['uninstall'] command_consumes_arguments = False # override base def run(self): if self.uninstall: self.multi_version = True self.uninstall_link() self.uninstall_namespaces() else: self.install_for_development() self.warn_deprecated_options() def initialize_options(self): self.uninstall = None self.egg_path = None easy_install.initialize_options(self) self.setup_path = None self.always_copy_from = '.' # always copy eggs installed in curdir def finalize_options(self): ei = self.get_finalized_command("egg_info") if ei.broken_egg_info: template = "Please rename %r to %r before using 'develop'" args = ei.egg_info, ei.broken_egg_info raise DistutilsError(template % args) self.args = [ei.egg_name] easy_install.finalize_options(self) self.expand_basedirs() self.expand_dirs() # pick up setup-dir .egg files only: no .egg-info self.package_index.scan(glob.glob('*.egg')) egg_link_fn = ei.egg_name + '.egg-link' self.egg_link = os.path.join(self.install_dir, egg_link_fn) self.egg_base = ei.egg_base if self.egg_path is None: self.egg_path = os.path.abspath(ei.egg_base) target = pkg_resources.normalize_path(self.egg_base) egg_path = pkg_resources.normalize_path( os.path.join(self.install_dir, self.egg_path)) if egg_path != target: raise DistutilsOptionError( "--egg-path must be a relative path from the install" " directory to " + target ) # Make a distribution for the package's source self.dist = pkg_resources.Distribution( target, pkg_resources.PathMetadata(target, os.path.abspath(ei.egg_info)), project_name=ei.egg_name ) self.setup_path = self._resolve_setup_path( self.egg_base, self.install_dir, self.egg_path, ) @staticmethod def _resolve_setup_path(egg_base, install_dir, egg_path): """ Generate a path from egg_base back to '.' where the setup script resides and ensure that path points to the setup path from $install_dir/$egg_path. """ path_to_setup = egg_base.replace(os.sep, '/').rstrip('/') if path_to_setup != os.curdir: path_to_setup = '../' * (path_to_setup.count('/') + 1) resolved = pkg_resources.normalize_path( os.path.join(install_dir, egg_path, path_to_setup) ) if resolved != pkg_resources.normalize_path(os.curdir): raise DistutilsOptionError( "Can't get a consistent path to setup script from" " installation directory", resolved, pkg_resources.normalize_path(os.curdir)) return path_to_setup def install_for_development(self): if six.PY3 and getattr(self.distribution, 'use_2to3', False): # If we run 2to3 we can not do this inplace: # Ensure metadata is up-to-date self.reinitialize_command('build_py', inplace=0) self.run_command('build_py') bpy_cmd = self.get_finalized_command("build_py") build_path = pkg_resources.normalize_path(bpy_cmd.build_lib) # Build extensions self.reinitialize_command('egg_info', egg_base=build_path) self.run_command('egg_info') self.reinitialize_command('build_ext', inplace=0) self.run_command('build_ext') # Fixup egg-link and easy-install.pth ei_cmd = self.get_finalized_command("egg_info") self.egg_path = build_path self.dist.location = build_path # XXX self.dist._provider = pkg_resources.PathMetadata( build_path, ei_cmd.egg_info) else: # Without 2to3 inplace works fine: self.run_command('egg_info') # Build extensions in-place self.reinitialize_command('build_ext', inplace=1) self.run_command('build_ext') self.install_site_py() # ensure that target dir is site-safe if setuptools.bootstrap_install_from: self.easy_install(setuptools.bootstrap_install_from) setuptools.bootstrap_install_from = None self.install_namespaces() # create an .egg-link in the installation dir, pointing to our egg log.info("Creating %s (link to %s)", self.egg_link, self.egg_base) if not self.dry_run: with open(self.egg_link, "w") as f: f.write(self.egg_path + "\n" + self.setup_path) # postprocess the installed distro, fixing up .pth, installing scripts, # and handling requirements self.process_distribution(None, self.dist, not self.no_deps) def uninstall_link(self): if os.path.exists(self.egg_link): log.info("Removing %s (link to %s)", self.egg_link, self.egg_base) egg_link_file = open(self.egg_link) contents = [line.rstrip() for line in egg_link_file] egg_link_file.close() if contents not in ([self.egg_path], [self.egg_path, self.setup_path]): log.warn("Link points to %s: uninstall aborted", contents) return if not self.dry_run: os.unlink(self.egg_link) if not self.dry_run: self.update_pth(self.dist) # remove any .pth link to us if self.distribution.scripts: # XXX should also check for entry point scripts! log.warn("Note: you must uninstall or replace scripts manually!") def install_egg_scripts(self, dist): if dist is not self.dist: # Installing a dependency, so fall back to normal behavior return easy_install.install_egg_scripts(self, dist) # create wrapper scripts in the script dir, pointing to dist.scripts # new-style... self.install_wrapper_scripts(dist) # ...and old-style for script_name in self.distribution.scripts or []: script_path = os.path.abspath(convert_path(script_name)) script_name = os.path.basename(script_path) with io.open(script_path) as strm: script_text = strm.read() self.install_script(dist, script_name, script_text, script_path) def install_wrapper_scripts(self, dist): dist = VersionlessRequirement(dist) return easy_install.install_wrapper_scripts(self, dist) class VersionlessRequirement: """ Adapt a pkg_resources.Distribution to simply return the project name as the 'requirement' so that scripts will work across multiple versions. >>> from pkg_resources import Distribution >>> dist = Distribution(project_name='foo', version='1.0') >>> str(dist.as_requirement()) 'foo==1.0' >>> adapted_dist = VersionlessRequirement(dist) >>> str(adapted_dist.as_requirement()) 'foo' """ def __init__(self, dist): self.__dist = dist def __getattr__(self, name): return getattr(self.__dist, name) def as_requirement(self): return self.project_name site-packages/setuptools/command/register.py000064400000001026151464452710015331 0ustar00from distutils import log import distutils.command.register as orig class register(orig.register): __doc__ = orig.register.__doc__ def run(self): try: # Make sure that we are using valid current name/version info self.run_command('egg_info') orig.register.run(self) finally: self.announce( "WARNING: Registering is deprecated, use twine to " "upload instead (https://pypi.org/p/twine/)", log.WARN ) site-packages/setuptools/command/install_lib.py000064400000007426151464452710016013 0ustar00import os import sys from itertools import product, starmap import distutils.command.install_lib as orig class install_lib(orig.install_lib): """Don't add compiled flags to filenames of non-Python files""" def run(self): self.build() outfiles = self.install() if outfiles is not None: # always compile, in case we have any extension stubs to deal with self.byte_compile(outfiles) def get_exclusions(self): """ Return a collections.Sized collections.Container of paths to be excluded for single_version_externally_managed installations. """ all_packages = ( pkg for ns_pkg in self._get_SVEM_NSPs() for pkg in self._all_packages(ns_pkg) ) excl_specs = product(all_packages, self._gen_exclusion_paths()) return set(starmap(self._exclude_pkg_path, excl_specs)) def _exclude_pkg_path(self, pkg, exclusion_path): """ Given a package name and exclusion path within that package, compute the full exclusion path. """ parts = pkg.split('.') + [exclusion_path] return os.path.join(self.install_dir, *parts) @staticmethod def _all_packages(pkg_name): """ >>> list(install_lib._all_packages('foo.bar.baz')) ['foo.bar.baz', 'foo.bar', 'foo'] """ while pkg_name: yield pkg_name pkg_name, sep, child = pkg_name.rpartition('.') def _get_SVEM_NSPs(self): """ Get namespace packages (list) but only for single_version_externally_managed installations and empty otherwise. """ # TODO: is it necessary to short-circuit here? i.e. what's the cost # if get_finalized_command is called even when namespace_packages is # False? if not self.distribution.namespace_packages: return [] install_cmd = self.get_finalized_command('install') svem = install_cmd.single_version_externally_managed return self.distribution.namespace_packages if svem else [] @staticmethod def _gen_exclusion_paths(): """ Generate file paths to be excluded for namespace packages (bytecode cache files). """ # always exclude the package module itself yield '__init__.py' yield '__init__.pyc' yield '__init__.pyo' if not hasattr(sys, 'implementation'): return base = os.path.join('__pycache__', '__init__.' + sys.implementation.cache_tag) yield base + '.pyc' yield base + '.pyo' yield base + '.opt-1.pyc' yield base + '.opt-2.pyc' def copy_tree( self, infile, outfile, preserve_mode=1, preserve_times=1, preserve_symlinks=0, level=1 ): assert preserve_mode and preserve_times and not preserve_symlinks exclude = self.get_exclusions() if not exclude: return orig.install_lib.copy_tree(self, infile, outfile) # Exclude namespace package __init__.py* files from the output from setuptools.archive_util import unpack_directory from distutils import log outfiles = [] def pf(src, dst): if dst in exclude: log.warn("Skipping installation of %s (namespace package)", dst) return False log.info("copying %s -> %s", src, os.path.dirname(dst)) outfiles.append(dst) return dst unpack_directory(infile, outfile, pf) return outfiles def get_outputs(self): outputs = orig.install_lib.get_outputs(self) exclude = self.get_exclusions() if exclude: return [f for f in outputs if f not in exclude] return outputs site-packages/setuptools/command/bdist_rpm.py000064400000002744151464452710015500 0ustar00import distutils.command.bdist_rpm as orig class bdist_rpm(orig.bdist_rpm): """ Override the default bdist_rpm behavior to do the following: 1. Run egg_info to ensure the name and version are properly calculated. 2. Always run 'install' using --single-version-externally-managed to disable eggs in RPM distributions. 3. Replace dash with underscore in the version numbers for better RPM compatibility. """ def run(self): # ensure distro name is up-to-date self.run_command('egg_info') orig.bdist_rpm.run(self) def _make_spec_file(self): version = self.distribution.get_version() rpmversion = version.replace('-', '_') spec = orig.bdist_rpm._make_spec_file(self) line23 = '%define version ' + version line24 = '%define version ' + rpmversion spec = [ line.replace( "Source0: %{name}-%{version}.tar", "Source0: %{name}-%{unmangled_version}.tar" ).replace( "setup.py install ", "setup.py install --single-version-externally-managed " ).replace( "%setup", "%setup -n %{name}-%{unmangled_version}" ).replace(line23, line24) for line in spec ] insert_loc = spec.index(line24) + 1 unmangled_version = "%define unmangled_version " + version spec.insert(insert_loc, unmangled_version) return spec site-packages/setuptools/command/install.py000064400000011113151464452710015151 0ustar00from distutils.errors import DistutilsArgError import inspect import glob import warnings import platform import distutils.command.install as orig import setuptools # Prior to numpy 1.9, NumPy relies on the '_install' name, so provide it for # now. See https://github.com/pypa/setuptools/issues/199/ _install = orig.install class install(orig.install): """Use easy_install to install the package, w/dependencies""" user_options = orig.install.user_options + [ ('old-and-unmanageable', None, "Try not to use this!"), ('single-version-externally-managed', None, "used by system package builders to create 'flat' eggs"), ] boolean_options = orig.install.boolean_options + [ 'old-and-unmanageable', 'single-version-externally-managed', ] new_commands = [ ('install_egg_info', lambda self: True), ('install_scripts', lambda self: True), ] _nc = dict(new_commands) def initialize_options(self): orig.install.initialize_options(self) self.old_and_unmanageable = None self.single_version_externally_managed = None def finalize_options(self): orig.install.finalize_options(self) if self.root: self.single_version_externally_managed = True elif self.single_version_externally_managed: if not self.root and not self.record: raise DistutilsArgError( "You must specify --record or --root when building system" " packages" ) def handle_extra_path(self): if self.root or self.single_version_externally_managed: # explicit backward-compatibility mode, allow extra_path to work return orig.install.handle_extra_path(self) # Ignore extra_path when installing an egg (or being run by another # command without --root or --single-version-externally-managed self.path_file = None self.extra_dirs = '' def run(self): # Explicit request for old-style install? Just do it if self.old_and_unmanageable or self.single_version_externally_managed: return orig.install.run(self) if not self._called_from_setup(inspect.currentframe()): # Run in backward-compatibility mode to support bdist_* commands. orig.install.run(self) else: self.do_egg_install() @staticmethod def _called_from_setup(run_frame): """ Attempt to detect whether run() was called from setup() or by another command. If called by setup(), the parent caller will be the 'run_command' method in 'distutils.dist', and *its* caller will be the 'run_commands' method. If called any other way, the immediate caller *might* be 'run_command', but it won't have been called by 'run_commands'. Return True in that case or if a call stack is unavailable. Return False otherwise. """ if run_frame is None: msg = "Call stack not available. bdist_* commands may fail." warnings.warn(msg) if platform.python_implementation() == 'IronPython': msg = "For best results, pass -X:Frames to enable call stack." warnings.warn(msg) return True res = inspect.getouterframes(run_frame)[2] caller, = res[:1] info = inspect.getframeinfo(caller) caller_module = caller.f_globals.get('__name__', '') return ( caller_module == 'distutils.dist' and info.function == 'run_commands' ) def do_egg_install(self): easy_install = self.distribution.get_command_class('easy_install') cmd = easy_install( self.distribution, args="x", root=self.root, record=self.record, ) cmd.ensure_finalized() # finalize before bdist_egg munges install cmd cmd.always_copy_from = '.' # make sure local-dir eggs get installed # pick up setup-dir .egg files only: no .egg-info cmd.package_index.scan(glob.glob('*.egg')) self.run_command('bdist_egg') args = [self.distribution.get_command_obj('bdist_egg').egg_output] if setuptools.bootstrap_install_from: # Bootstrap self-installation of setuptools args.insert(0, setuptools.bootstrap_install_from) cmd.args = args cmd.run() setuptools.bootstrap_install_from = None # XXX Python 3.1 doesn't see _nc if this is inside the class install.sub_commands = ( [cmd for cmd in orig.install.sub_commands if cmd[0] not in install._nc] + install.new_commands ) site-packages/setuptools/command/test.py000064400000022602151464452710014467 0ustar00import os import operator import sys import contextlib import itertools import unittest from distutils.errors import DistutilsError, DistutilsOptionError from distutils import log from unittest import TestLoader from setuptools.extern import six from setuptools.extern.six.moves import map, filter from pkg_resources import (resource_listdir, resource_exists, normalize_path, working_set, _namespace_packages, evaluate_marker, add_activation_listener, require, EntryPoint) from setuptools import Command from .build_py import _unique_everseen __metaclass__ = type class ScanningLoader(TestLoader): def __init__(self): TestLoader.__init__(self) self._visited = set() def loadTestsFromModule(self, module, pattern=None): """Return a suite of all tests cases contained in the given module If the module is a package, load tests from all the modules in it. If the module has an ``additional_tests`` function, call it and add the return value to the tests. """ if module in self._visited: return None self._visited.add(module) tests = [] tests.append(TestLoader.loadTestsFromModule(self, module)) if hasattr(module, "additional_tests"): tests.append(module.additional_tests()) if hasattr(module, '__path__'): for file in resource_listdir(module.__name__, ''): if file.endswith('.py') and file != '__init__.py': submodule = module.__name__ + '.' + file[:-3] else: if resource_exists(module.__name__, file + '/__init__.py'): submodule = module.__name__ + '.' + file else: continue tests.append(self.loadTestsFromName(submodule)) if len(tests) != 1: return self.suiteClass(tests) else: return tests[0] # don't create a nested suite for only one return # adapted from jaraco.classes.properties:NonDataProperty class NonDataProperty: def __init__(self, fget): self.fget = fget def __get__(self, obj, objtype=None): if obj is None: return self return self.fget(obj) class test(Command): """Command to run unit tests after in-place build""" description = "run unit tests after in-place build (deprecated)" user_options = [ ('test-module=', 'm', "Run 'test_suite' in specified module"), ('test-suite=', 's', "Run single test, case or suite (e.g. 'module.test_suite')"), ('test-runner=', 'r', "Test runner to use"), ] def initialize_options(self): self.test_suite = None self.test_module = None self.test_loader = None self.test_runner = None def finalize_options(self): if self.test_suite and self.test_module: msg = "You may specify a module or a suite, but not both" raise DistutilsOptionError(msg) if self.test_suite is None: if self.test_module is None: self.test_suite = self.distribution.test_suite else: self.test_suite = self.test_module + ".test_suite" if self.test_loader is None: self.test_loader = getattr(self.distribution, 'test_loader', None) if self.test_loader is None: self.test_loader = "setuptools.command.test:ScanningLoader" if self.test_runner is None: self.test_runner = getattr(self.distribution, 'test_runner', None) @NonDataProperty def test_args(self): return list(self._test_args()) def _test_args(self): if not self.test_suite and sys.version_info >= (2, 7): yield 'discover' if self.verbose: yield '--verbose' if self.test_suite: yield self.test_suite def with_project_on_sys_path(self, func): """ Backward compatibility for project_on_sys_path context. """ with self.project_on_sys_path(): func() @contextlib.contextmanager def project_on_sys_path(self, include_dists=[]): with_2to3 = six.PY3 and getattr(self.distribution, 'use_2to3', False) if with_2to3: # If we run 2to3 we can not do this inplace: # Ensure metadata is up-to-date self.reinitialize_command('build_py', inplace=0) self.run_command('build_py') bpy_cmd = self.get_finalized_command("build_py") build_path = normalize_path(bpy_cmd.build_lib) # Build extensions self.reinitialize_command('egg_info', egg_base=build_path) self.run_command('egg_info') self.reinitialize_command('build_ext', inplace=0) self.run_command('build_ext') else: # Without 2to3 inplace works fine: self.run_command('egg_info') # Build extensions in-place self.reinitialize_command('build_ext', inplace=1) self.run_command('build_ext') ei_cmd = self.get_finalized_command("egg_info") old_path = sys.path[:] old_modules = sys.modules.copy() try: project_path = normalize_path(ei_cmd.egg_base) sys.path.insert(0, project_path) working_set.__init__() add_activation_listener(lambda dist: dist.activate()) require('%s==%s' % (ei_cmd.egg_name, ei_cmd.egg_version)) with self.paths_on_pythonpath([project_path]): yield finally: sys.path[:] = old_path sys.modules.clear() sys.modules.update(old_modules) working_set.__init__() @staticmethod @contextlib.contextmanager def paths_on_pythonpath(paths): """ Add the indicated paths to the head of the PYTHONPATH environment variable so that subprocesses will also see the packages at these paths. Do this in a context that restores the value on exit. """ nothing = object() orig_pythonpath = os.environ.get('PYTHONPATH', nothing) current_pythonpath = os.environ.get('PYTHONPATH', '') try: prefix = os.pathsep.join(_unique_everseen(paths)) to_join = filter(None, [prefix, current_pythonpath]) new_path = os.pathsep.join(to_join) if new_path: os.environ['PYTHONPATH'] = new_path yield finally: if orig_pythonpath is nothing: os.environ.pop('PYTHONPATH', None) else: os.environ['PYTHONPATH'] = orig_pythonpath @staticmethod def install_dists(dist): """ Install the requirements indicated by self.distribution and return an iterable of the dists that were built. """ ir_d = dist.fetch_build_eggs(dist.install_requires) tr_d = dist.fetch_build_eggs(dist.tests_require or []) er_d = dist.fetch_build_eggs( v for k, v in dist.extras_require.items() if k.startswith(':') and evaluate_marker(k[1:]) ) return itertools.chain(ir_d, tr_d, er_d) def run(self): self.announce( "WARNING: Testing via this command is deprecated and will be " "removed in a future version. Users looking for a generic test " "entry point independent of test runner are encouraged to use " "tox.", log.WARN, ) installed_dists = self.install_dists(self.distribution) cmd = ' '.join(self._argv) if self.dry_run: self.announce('skipping "%s" (dry run)' % cmd) return self.announce('running "%s"' % cmd) paths = map(operator.attrgetter('location'), installed_dists) with self.paths_on_pythonpath(paths): with self.project_on_sys_path(): self.run_tests() def run_tests(self): # Purge modules under test from sys.modules. The test loader will # re-import them from the build location. Required when 2to3 is used # with namespace packages. if six.PY3 and getattr(self.distribution, 'use_2to3', False): module = self.test_suite.split('.')[0] if module in _namespace_packages: del_modules = [] if module in sys.modules: del_modules.append(module) module += '.' for name in sys.modules: if name.startswith(module): del_modules.append(name) list(map(sys.modules.__delitem__, del_modules)) test = unittest.main( None, None, self._argv, testLoader=self._resolve_as_ep(self.test_loader), testRunner=self._resolve_as_ep(self.test_runner), exit=False, ) if not test.result.wasSuccessful(): msg = 'Test failed: %s' % test.result self.announce(msg, log.ERROR) raise DistutilsError(msg) @property def _argv(self): return ['unittest'] + self.test_args @staticmethod def _resolve_as_ep(val): """ Load the indicated attribute value, called, as a as if it were specified as an entry point. """ if val is None: return parsed = EntryPoint.parse("x=" + val) return parsed.resolve()() site-packages/setuptools/command/dist_info.py000064400000001700151464452710015462 0ustar00""" Create a dist_info directory As defined in the wheel specification """ import os from distutils.core import Command from distutils import log class dist_info(Command): description = 'create a .dist-info directory' user_options = [ ('egg-base=', 'e', "directory containing .egg-info directories" " (default: top of the source tree)"), ] def initialize_options(self): self.egg_base = None def finalize_options(self): pass def run(self): egg_info = self.get_finalized_command('egg_info') egg_info.egg_base = self.egg_base egg_info.finalize_options() egg_info.run() dist_info_dir = egg_info.egg_info[:-len('.egg-info')] + '.dist-info' log.info("creating '{}'".format(os.path.abspath(dist_info_dir))) bdist_wheel = self.get_finalized_command('bdist_wheel') bdist_wheel.egg2dist(egg_info.egg_info, dist_info_dir) site-packages/setuptools/command/launcher manifest.xml000064400000001164151464452710017250 0ustar00 site-packages/setuptools/command/__pycache__/register.cpython-38.opt-1.pyc000064400000001363151464452710022562 0ustar00U Qab@s4ddlmZddlmmZGdddejZdS))logNc@seZdZejjZddZdS)registerc Cs0z|dtj|W5|dtjXdS)Nz[WARNING: Registering is deprecated, use twine to upload instead (https://pypi.org/p/twine/)Zegg_info)ZannouncerZWARNZ run_commandorigrrun)selfr?/usr/lib/python3.8/site-packages/setuptools/command/register.pyrs z register.runN)__name__ __module__ __qualname__rr__doc__rrrrrrsr)Z distutilsrZdistutils.command.registerZcommandrrrrrrs site-packages/setuptools/command/__pycache__/egg_info.cpython-38.opt-1.pyc000064400000052404151464452710022515 0ustar00U Qabc@sdZddlmZddlmZddlmZddlm Z ddlZddlZddl Z ddl Z ddl Z ddl Z ddlZddlZddlZddlmZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZddlmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&ddl'm(Z(ddl)m*Z*ddlm+Z+ddlm,Z,ddZ-GdddZ.Gddde.eZ/GdddeZGdddeZ0ddZ1ddZ2d d!Z3d"d#Z4d$d%Z5d&d'Z6d(d)Z7d*d+Z8d5d-d.Z9d/d0Z:d1d2Z;Gd3d4d4e,Z|j|jkr>|j|_t|j|_ d|j_dS)Nz%s==%sz%s===%sz2Invalid distribution name or version syntax: %s-%srrP .egg-info-)!r:rQr?rRr isinstancerr>ZVersionlistr ValueError distutilserrorsZDistutilsOptionErrorrPr7Z package_dirgetr#curdirZensure_dirnamerrIr$joincheck_broken_egg_infometadataZ _patched_distkeylowerZ_versionZ_parsed_version)r9Zparsed_versionZ is_versionspecdirsZpdr2r2r3finalize_optionss>           zegg_info.finalize_optionsFcCsL|r||||n4tj|rH|dkr>|s>td||dS||dS)aWrite `data` to `filename` or delete if empty If `data` is non-empty, this routine is the same as ``write_file()``. If `data` is empty but not ``None``, this is the same as calling ``delete_file(filename)`. If `data` is ``None``, then this is a no-op unless `filename` exists, in which case a warning is issued about the orphaned file (if `force` is false), or deleted (if `force` is true). Nz$%s not set in setup(), but %s exists) write_filer#r$existsrwarn delete_file)r9whatrZdataforcer2r2r3write_or_delete_files   zegg_info.write_or_delete_filecCsDtd||tjr|d}|js@t|d}|||dS)zWrite `data` to `filename` (if not a dry run) after announcing it `what` is used in a log message to identify what is being written to the file. zwriting %s to %sutf-8wbN) rinforZPY3encodedry_runopenwriteclose)r9rqrZrrfr2r2r3rms   zegg_info.write_filecCs td||jst|dS)z8Delete `filename` (if not a dry run) after announcing itz deleting %sN)rrwryr#unlink)r9rZr2r2r3rps zegg_info.delete_filecCs||jt|jd|jj}tdD]4}|j|d|}|||j tj |j|j q*tj |jd}tj |r| ||dS)Nzegg_info.writers) installerznative_libs.txt)ZmkpathrIr#utimer7Zfetch_build_eggrZrequireZresolver:r$rernrp find_sources)r9repwriternlr2r2r3runs     z egg_info.runcCs4tj|jd}t|j}||_||j|_dS)z"Generate SOURCES.txt manifest filez SOURCES.txtN) r#r$rerImanifest_makerr7manifestrfilelist)r9Zmanifest_filenameZmmr2r2r3r*s  zegg_info.find_sourcescCsT|jd}|jtjkr&tj|j|}tj|rPtd||j |j |_ ||_ dS)Nr\aB------------------------------------------------------------------------------ Note: Your current .egg-info directory has a '-' in its name; this will not work correctly with "setup.py develop". Please rename %s to %s to correct this problem. ------------------------------------------------------------------------------) rQrPr#rdr$rernrrorIrS)r9Zbeir2r2r3rf2s   zegg_info.check_broken_egg_infoN)F)rErFrG descriptionZ user_optionsZboolean_optionsZ negative_optrTrHrUsetterr[rlrtrmrprrrfr2r2r2r3rIs.    1 rIc@s|eZdZddZddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ ddZddZddZdS)rcCs ||\}}}}|dkrR|dd||D]}||s2td|q2n|dkr|dd||D]}||srtd|qrn|dkr|d d||D]}||std |qnJ|d kr|d d||D]}||std |qn|dkr`|d|d|f|D]"}| ||s:td||q:n|dkr|d|d|f|D]"}| ||std||qnp|dkr|d|| |std|n>|dkr|d|| |std|n t d|dS)Nincludezinclude  z%warning: no files found matching '%s'excludezexclude z9warning: no previously-included files found matching '%s'zglobal-includezglobal-include z>warning: no files found matching '%s' anywhere in distributionzglobal-excludezglobal-exclude zRwarning: no previously-included files matching '%s' found anywhere in distributionzrecursive-includezrecursive-include %s %sz:warning: no files found matching '%s' under directory '%s'zrecursive-excludezrecursive-exclude %s %szNwarning: no previously-included files matching '%s' found under directory '%s'graftzgraft z+warning: no directories found matching '%s'prunezprune z6no previously-included directories found matching '%s'z'this cannot happen: invalid action '%s')Z_parse_template_line debug_printrerrrorglobal_includeglobal_excluderecursive_includerecursive_excluderrr)r9lineactionZpatternsdirZ dir_patternpatternr2r2r3process_template_lineEs|                  zFileList.process_template_linecCsRd}tt|jdddD]2}||j|r|d|j||j|=d}q|S)z Remove all files from the file list that match the predicate. Return True if any matching files were removed Frz removing T)ranger)filesr)r9Z predicatefoundr/r2r2r3 _remove_filesszFileList._remove_filescCs$ddt|D}||t|S)z#Include files that match 'pattern'.cSsg|]}tj|s|qSr2r#r$isdir.0r}r2r2r3 s z$FileList.include..rextendboolr9rrr2r2r3rs zFileList.includecCst|}||jS)z#Exclude files that match 'pattern'.)r4rmatchr9rrr2r2r3rszFileList.excludecCs8tj|d|}ddt|ddD}||t|S)zN Include all files anywhere in 'dir/' that match the pattern. rcSsg|]}tj|s|qSr2rrr2r2r3rs z.FileList.recursive_include..T) recursive)r#r$rerrr)r9rrZ full_patternrr2r2r3rs zFileList.recursive_includecCs ttj|d|}||jS)zM Exclude any file anywhere in 'dir/' that match the pattern. rr4r#r$rerr)r9rrrr2r2r3rszFileList.recursive_excludecCs$ddt|D}||t|S)zInclude all files from 'dir/'.cSs"g|]}tj|D]}|qqSr2)rarfindall)rZ match_diritemr2r2r3rsz"FileList.graft..r)r9rrr2r2r3rs  zFileList.graftcCsttj|d}||jS)zFilter out files from 'dir/'.rr)r9rrr2r2r3rszFileList.prunecsJ|jdkr|ttjd|fdd|jD}||t|S)z Include all files anywhere in the current directory that match the pattern. This is very inefficient on large file trees. Nrcsg|]}|r|qSr2rrrr2r3rs z+FileList.global_include..)Zallfilesrr4r#r$rerrrr2rr3rs   zFileList.global_includecCsttjd|}||jS)zD Exclude all files anywhere that match the pattern. rrrr2r2r3rszFileList.global_excludecCs8|dr|dd}t|}||r4|j|dS)N r)r<r _safe_pathrappend)r9rr$r2r2r3rs    zFileList.appendcCs|jt|j|dSr6)rrfilterr)r9pathsr2r2r3rszFileList.extendcCstt|j|j|_dS)z Replace self.files with only safe paths Because some owners of FileList manipulate the underlying ``files`` attribute directly, this method must be called to repair those paths. N)r_rrrr8r2r2r3_repairszFileList._repairc Csd}t|}|dkr(td|dSt|d}|dkrNt||ddSz"tj|shtj|rnWdSWn&tk rt||t YnXdS)Nz!'%s' not %s encodable -- skippingz''%s' in unexpected encoding -- skippingFruT) unicode_utilsfilesys_decoderroZ try_encoder#r$rnUnicodeEncodeErrorsysgetfilesystemencoding)r9r$Zenc_warnZu_pathZ utf8_pathr2r2r3rs   zFileList._safe_pathN)rErFrGrrrrrrrrrrrrrrr2r2r2r3rBsI     rc@s\eZdZdZddZddZddZdd Zd d Zd d Z e ddZ ddZ ddZ dS)rz MANIFEST.incCsd|_d|_d|_d|_dS)Nr)Z use_defaultsrZ manifest_onlyZforce_manifestr8r2r2r3rT sz!manifest_maker.initialize_optionscCsdSr6r2r8r2r2r3rlszmanifest_maker.finalize_optionscCsdt|_tj|js||tj|jr<| | |j |j |dSr6) rrr#r$rnrwrite_manifest add_defaultstemplateZ read_templateprune_file_listsortZremove_duplicatesr8r2r2r3rs  zmanifest_maker.runcCst|}|tjdS)N/)rrreplacer#r%)r9r$r2r2r3_manifest_normalizes z"manifest_maker._manifest_normalizecsBjfddjjD}dj}tj|f|dS)zo Write the file list in 'self.filelist' to the manifest file named by 'self.manifest'. csg|]}|qSr2)rrr8r2r3r*sz1manifest_maker.write_manifest..zwriting manifest file '%s'N)rrrrZexecuterm)r9rmsgr2r8r3r"s  zmanifest_maker.write_manifestcCs||st||dSr6)_should_suppress_warningr ro)r9rr2r2r3ro.s zmanifest_maker.warncCs td|S)z; suppress missing-file warnings from sdist zstandard file .*not found)r&r)rr2r2r3r2sz'manifest_maker._should_suppress_warningcCst|||j|j|j|jtt}|rJ|j |nt j |jr`| t j drx|jd|d}|j|jdS)Nzsetup.pyrI)r rZ check_licenserrrrr_r rr#r$rnZ read_manifestget_finalized_commandrrI)r9ZrcfilesZei_cmdr2r2r3r9s     zmanifest_maker.add_defaultscCsZ|d}|j}|j|j|j|ttj }|jj d|d|dddS)Nbuildz(^|z)(RCS|CVS|\.svn)r)Zis_regex) rr7Z get_fullnamerrZ build_baser&r'r#r%Zexclude_pattern)r9rZbase_dirr%r2r2r3rLs    zmanifest_maker.prune_file_listN)rErFrGrrTrlrrrro staticmethodrrrr2r2r2r3rs   rc Cs8d|}|d}t|d}||W5QRXdS)z{Create a file with the specified name and write 'contents' (a sequence of strings without line terminators) to it.  rurvN)rerxrzr{)rZcontentsr}r2r2r3rmVs   rmc Cs|td||jsx|jj}|j|j|_}|j|j|_}z| |j W5|||_|_Xt |jdd}t |j |dS)Nz writing %sZzip_safe)rrwryr7rgrRr>rQr:write_pkg_inforIgetattrr Zwrite_safety_flag)cmdbasenamerZrgZoldverZoldnameZsafer2r2r3rcs rcCstj|rtddS)NzsWARNING: 'depends.txt' is not used by setuptools 0.6! Use the install_requires/extras_require setup() args instead.)r#r$rnrrorrrZr2r2r3warn_depends_obsoletevs rcCs,t|pd}dd}t||}||dS)Nr2cSs|dS)Nrr2)rr2r2r3z%_write_requirements..)rr writelines)streamZreqslinesZ append_crr2r2r3_write_requirements~s  rcCsj|j}t}t||j|jp"i}t|D]&}|djft t|||q,| d|| dS)Nz [{extra}] Z requirements) r7rStringIOrZinstall_requiresextras_requiresortedr{formatvarsrtgetvalue)rrrZZdistrrrZextrar2r2r3write_requirementss   rcCs,t}t||jj|d||dS)Nzsetup-requirements)iorrr7Zsetup_requiresrtr)rrrZrrr2r2r3write_setup_requirementssrcCs:tdd|jD}|d|dt|ddS)NcSsg|]}|dddqS).rr)r")rkr2r2r3rsz(write_toplevel_names..ztop-level namesr)rYfromkeysr7Ziter_distribution_namesrmrer)rrrZZpkgsr2r2r3write_toplevel_namess rcCst|||ddS)NT) write_argrr2r2r3 overwrite_argsrFcCsHtj|d}t|j|d}|dk r4d|d}|||||dS)Nrr)r#r$splitextrr7rert)rrrZrsargnamerVr2r2r3rs rcCs|jj}t|tjs|dkr"|}nn|dk rg}t|D]J\}}t|tjsrt||}d tt t | }| d||fq:d |}|d||ddS)Nrz [%s] %s rz entry pointsT)r7Z entry_pointsr^rZ string_typesritemsrZ parse_grouprerstrvaluesrrt)rrrZrrrZsectionrr2r2r3 write_entriess   rc Csjtdttjdrftd>}|D]2}t d|}|r(t | dW5QRSq(W5QRXdS)zd Get a -r### off of PKG-INFO Version in case this is an sdist of a subversion revision. z$get_pkg_info_revision is deprecated.zPKG-INFOzVersion:.*-r(\d+)\s*$rr) warningsroEggInfoDeprecationWarningr#r$rnrrzr&rintgroup)r}rrr2r2r3get_pkg_info_revisions    *rc@seZdZdZdS)rzqClass for warning about deprecations in eggInfo in setupTools. Not ignored by default, unlike DeprecationWarning.N)rErFrG__doc__r2r2r2r3rsr)F)=rZdistutils.filelistrZ _FileListZdistutils.errorsrZdistutils.utilrrarr#r&rrrrBrWZsetuptools.externrZsetuptools.extern.six.movesrZ setuptoolsrZsetuptools.command.sdistr r Zsetuptools.command.setoptr Zsetuptools.commandr Z pkg_resourcesr rrrrrrrZsetuptools.unicode_utilsrZsetuptools.globrrrr4r5rIrrmrrrrrrrrrrrr2r2r2r3sX           (    S2EP    site-packages/setuptools/command/__pycache__/easy_install.cpython-38.opt-1.pyc000064400000177432151464452710023440 0ustar00U QabU@s4dZddlmZddlmZddlmZmZddlmZmZm Z m Z ddl m Z m Z ddlmZmZddlmZdd lmZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd l Z dd l!Z!dd l"Z"dd l#Z#dd l$Z$dd l%Z%dd l&m'Z'm(Z(dd l)m*Z*dd l+m,Z,ddl-m.Z.m/Z/ddl)m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6ddl7m8Z8ddl9m:Z:m;Z;mZ>ddl?m@Z@ddlAmBZBmCZCmDZDmEZEmFZFmGZGmHZHmIZImJZJmKZKmLZLmMZMmNZNmOZOmPZPdd lQZAeRZSejTdeAjUdddddddgZVd d!ZWd"dZXe,jYrBd#d$ZZd%d&Z[nd'd$ZZd(d&Z[d)d*Z\Gd+dde0Z]d,d-Z^d.d/Z_d0d1Z`d2dZad3dZbGd4ddeHZcGd5d6d6ecZdejefd7d8d9kredZcd:d;Zgdd?Zid@dAZjdsdBdCZkdDdEZldFdGZmdHejnkremZondIdJZodtdLdMZpdNdOZqdPdQZrdRdSZszddTlmtZuWnevk rndUdVZuYnXdWdXZtGdYdZdZewZxexyZzGd[d\d\exZ{Gd]d^d^Z|Gd_d`d`e|Z}Gdadbdbe}Z~e|jZe|jZdcddZdedfZdgehfdhdiZdjdkZdldmZdudndZe"jdodpZGdqdrdre*Zd S)va% Easy Install ------------ A tool for doing automatic download/extract/build of distutils-based Python packages. For detailed documentation, see the accompanying EasyInstall.txt file, or visit the `EasyInstall home page`__. __ https://setuptools.readthedocs.io/en/latest/easy_install.html )glob) get_platform) convert_path subst_vars)DistutilsArgErrorDistutilsOptionErrorDistutilsErrorDistutilsPlatformError)INSTALL_SCHEMES SCHEME_KEYS)logdir_util) first_line_re)find_executableN)get_config_varsget_path)SetuptoolsDeprecationWarning)six) configparsermap)Command) run_setup) rmtree_safe)setopt)unpack_archive) PackageIndexparse_requirement_arg URL_SCHEME) bdist_eggegg_info)Wheel) yield_linesnormalize_pathresource_stringensure_directoryget_distributionfind_distributions Environment Requirement Distribution PathMetadata EggMetadata WorkingSetDistributionNotFoundVersionConflict DEVELOP_DISTdefault)categorysamefile easy_installPthDistributionsextract_wininst_cfgmainget_exe_prefixescCstddkS)NP)structcalcsizer<r<C/usr/lib/python3.8/site-packages/setuptools/command/easy_install.pyis_64bitOsr>cCsjtj|otj|}ttjdo&|}|r:tj||Stjtj|}tjtj|}||kS)z Determine if two paths reference the same file. Augments os.path.samefile to work on Windows and suppresses errors if the path doesn't exist. r2)ospathexistshasattrr2normpathnormcase)Zp1Zp2Z both_existZ use_samefileZnorm_p1Znorm_p2r<r<r=r2SscCs|SNr<sr<r<r= _to_bytesesrHcCs.zt|dWdStk r(YdSXdSNasciiTF)rZ text_type UnicodeErrorrFr<r<r=isasciihs  rLcCs |dS)Nutf8)encoderFr<r<r=rHpscCs,z|dWdStk r&YdSXdSrI)rNrKrFr<r<r=rLss  cCst|ddS)N z; )textwrapdedentstripreplace)textr<r<r={rUc@seZdZdZdZdZdddddd d d d d dddddddddddgZdddddddd d!g Zej rd"ej Z e d#d$e fe d#d%diZ eZd&d'Zd(d)Zd*d+Zed,d-Zd.d/Zd0d1Zd2d3Zd4d5Zd6d7Zd8d9Zd:d;Zdd?Zed@ Z!edA Z"edB Z#dCdDZ$dEdFZ%dGdHZ&dIdJZ'dKdLZ(dMdNZ)e*j+dOdPZ,ddRdSZ-ddTdUZ.dVdWZ/ddXdYZ0dZd[Z1d\d]Z2d^d_Z3dd`daZ4edbdcZ5ddfdgZ6dhdiZ7djdkZ8dldmZ9dndoZ:dpdqZ;drdsZddwdxZ?edy Z@dzd{ZAd|d}ZBd~dZCddZDddZEddZFddZGddZHed ZIddZJddZKddZLeMeMddddZNeMdddZOddZPd$S)r3z'Manage a download/build/install processz Find/get/install Python packagesT)zprefix=Nzinstallation prefix)zip-okzzinstall package as a zipfile) multi-versionmz%make apps have to require() a version)upgradeUz1force upgrade (searches PyPI for latest versions))z install-dir=dzinstall package to DIR)z script-dir=rGzinstall scripts to DIR)exclude-scriptsxzDon't install scripts) always-copyaz'Copy all needed packages to install dir)z index-url=iz base URL of Python Package Index)z find-links=fz(additional URL(s) to search for packages)zbuild-directory=bz/download/extract/build in DIR; keep the results)z optimize=Ozlalso compile with optimization: -O1 for "python -O", -O2 for "python -OO", and -O0 to disable [default: -O0])zrecord=Nz3filename in which to record list of installed files) always-unzipZz*don't install as a zipfile, no matter what)z site-dirs=Sz)list of directories where .pth files work)editableez+Install specified packages in editable form)no-depsNzdon't install dependencies)z allow-hosts=Hz$pattern(s) that hostnames must match)local-snapshots-oklz(allow building eggs from local checkouts)versionNz"print version information and exit)z no-find-linksNz9Don't load find-links defined in packages being installedrWrYr^r[r`rirkrnrpz!install in user site-package '%s'userNrfcCs,d|_d|_|_d|_|_|_d|_d|_d|_d|_ d|_ |_ d|_ |_ |_d|_|_|_d|_|_|_d|_d|_d|_d|_d|_d|_d|_d|_d|_tjrtj |_!tj"|_#n d|_!d|_#d|_$d|_%d|_&|_'d|_(i|_)d|_*d|_+|j,j-|_-|j,.||j,/ddS)NrFr3)0rqzip_oklocal_snapshots_ok install_dir script_direxclude_scripts index_url find_linksbuild_directoryargsoptimizerecordr[ always_copy multi_versionrino_deps allow_hostsrootprefix no_reportrpinstall_purelibinstall_platlibinstall_headers install_libinstall_scripts install_data install_baseinstall_platbasesiteENABLE_USER_SITE USER_BASEinstall_userbase USER_SITEinstall_usersite no_find_links package_indexpth_filealways_copy_from site_dirsinstalled_projectssitepy_installedZ_dry_run distributionverboseZ_set_command_optionsget_option_dictselfr<r<r=initialize_optionssJ      zeasy_install.initialize_optionscCs"dd|D}tt|j|dS)Ncss*|]"}tj|stj|r|VqdSrE)r?r@rAislink).0filenamer<r<r= s z/easy_install.delete_blockers..)listr _delete_path)rblockersZextant_blockersr<r<r=delete_blockersszeasy_install.delete_blockerscCsJtd||jrdStj|o.tj| }|r8tntj}||dS)Nz Deleting %s) r infodry_runr?r@isdirrrmtreeunlink)rr@Zis_treeZremoverr<r<r=rs  zeasy_install._delete_pathcCs4djtj}td}d}t|jfttdS)zT Render the Setuptools version and installation details, then exit. {}.{} setuptoolsz=setuptools {dist.version} from {dist.location} (Python {ver})N)formatsys version_infor%printlocals SystemExit)Zverdisttmplr<r<r=_render_versions  zeasy_install._render_versionc Csh|jo |tjd}tdd\}}|j|j|j||dd|d|d||||t tddd |_ t j r|j |j d <|j|j d <||||d d d d|jdkr|j|_|jdkrd|_|dd|dd|jr|jr|j|_|j|_|ddtttj}t|_|jdk rdd|jdD}|D]N}t j!|s|t"#d|n,t||krt$|dn|j%t|q\|j&s|'|j(pd|_(|jdd|_)|jt|jfD] }||j)kr|j)*d|q|j+dk r0dd|j+dD}ndg}|j,dkrX|j-|j(|j)|d|_,t.|j)tj|_/|j0dk rt1|j0t2j3r|j0|_0ng|_0|j4r|j,5|j)tj|js|j,6|j0|ddt1|j7t8s6z0t8|j7|_7d|j7krdksnt9Wnt9k r4t$d YnX|j&rN|j:sNt;d!|j<s^t;d"g|_=dS)#Nrr exec_prefixabiflags) Z dist_nameZ dist_versionZ dist_fullname py_versionZpy_version_shortZpy_version_nodotZ sys_prefixrZsys_exec_prefixrruserbaseZusersitertruryrFr)rtrtrrtruinstall)r|r|cSsg|]}tj|qSr<)r?r@ expanduserrRrrGr<r<r= 8sz1easy_install.finalize_options..,z"%s (in --site-dirs) does not existz$ (in --site-dirs) is not on sys.pathzhttps://pypi.org/simple/cSsg|] }|qSr<)rRrr<r<r=rNs*)Z search_pathhosts)r{r{z--optimize must be 0, 1, or 2z9Must specify a build directory (-b) when using --editablez:No urls, filenames, or requirements specified (see --help))>rprrsplitrrZget_nameZ get_versionZ get_fullnamegetattr config_varsrrrr_fix_install_dir_for_user_siteexpand_basedirs expand_dirs_expandrurtrZset_undefined_optionsrqrrrr"r@ get_site_dirs all_site_dirsrr?rr warnrappendricheck_site_dirrw shadow_pathinsertrr create_indexr' local_indexrx isinstancerZ string_typesrsZscan_egg_linksadd_find_linksr{int ValueErrorryrrzoutputs) rrrrrCrr]Z path_itemrr<r<r=finalize_optionss                zeasy_install.finalize_optionscCs\|jr tjsdS||jdkr.d}t||j|_|_tj ddd}| |dS)z; Fix the install_dir if "--user" was used. Nz$User base directory is not specifiedposixZunixZ_user) rqrrcreate_home_pathrr rrr?namerS select_scheme)rmsgZ scheme_namer<r<r=rss  z+easy_install._fix_install_dir_for_user_sitecCsX|D]N}t||}|dk rtjdks.tjdkr:tj|}t||j}t|||qdS)Nrnt)rr?rr@rrrsetattr)rattrsattrvalr<r<r= _expand_attrss   zeasy_install._expand_attrscCs|dddgdS)zNCalls `os.path.expanduser` on install_base, install_platbase and root.rrrNrrr<r<r=rszeasy_install.expand_basedirscCsddddddg}||dS)z+Calls `os.path.expanduser` on install dirs.rrrrrrNr)rdirsr<r<r=rszeasy_install.expand_dirsc Cs|j|jjkrt|jz|jD]}|||j q"|jr|j}|j rzt |j }t t |D]}|||d||<q`ddl m }||j|j|fd|j|W5t|jjXdS)Nr) file_utilz'writing list of installed files to '%s')rrr set_verbosityrzr3rr|rrlenrange distutilsrexecuteZ write_filewarn_deprecated_options)rspecrZroot_lenZcounterrr<r<r=runs*     zeasy_install.runcCsDz t}Wn"tk r.tdtj}YnXtj|j d|S)zReturn a pseudo-tempname base in the install directory. This code is intentionally naive; if a malicious party can write to the target directory you're already in deep doodoo. rztest-easy-install-%s) r?getpid ExceptionrandomZrandintrmaxsizer@joinrt)rpidr<r<r=pseudo_tempnames  zeasy_install.pseudo_tempnamecCsdSrEr<rr<r<r=rsz$easy_install.warn_deprecated_optionsc CsZt|j}tj|d}tj|sTzt|Wn ttfk rR| YnX||j k}|sr|j sr| }nd| d}tj|}z*|rt|t|dt|Wn ttfk r| YnX|s|j st||r|jdkrt||j |_nd|_|tttkr.d|_n"|j rPtj|sPd|_d|_||_dS)z;Verify that self.install_dir is .pth-capable dir, if neededeasy-install.pthz .write-testwNT)r"rtr?r@rrAmakedirsOSErrorIOErrorcant_write_to_targetrr~check_pth_processingrropencloserno_default_version_msgrr4r _pythonpathr)rinstdirrZ is_site_dirZtestfileZ test_existsr<r<r=rs>           zeasy_install.check_site_diraS can't create or remove files in install directory The following error occurred while trying to add or remove files in the installation directory: %s The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s z This directory does not currently exist. Please create it and try again, or choose a different installation directory (using the -d or --install-dir option). a Perhaps your account does not have write access to this directory? If the installation directory is a system-owned directory, you may need to sign in as the administrator or "root" account. If you do not have administrative access to this machine, you may wish to choose a different installation directory, preferably one that is listed in your PYTHONPATH environment variable. For information on other options, you may wish to consult the documentation at: https://setuptools.readthedocs.io/en/latest/easy_install.html Please make the appropriate changes for your system and try again. cCsP|jtd|jf}tj|js6|d|j7}n|d|j7}t |dS)NrO) _easy_install__cant_write_msgrexc_infortr?r@rA_easy_install__not_exists_id_easy_install__access_msgr)rrr<r<r=rs z!easy_install.cant_write_to_targetc Cs|j}td||d}|d}tj|}tdd}z8|rNt|tj |}t j j |ddt |d}Wn ttfk r|YnXz||jft|d }tj}tjd krtj|\}} tj|d } | d kotj| } | r| }d dlm} | |dddgd tj|rNtd|WdSW5|r`|tj|rxt|tj|rt|X|jstd|dS)z@Empirically verify whether .pth files are supported in inst. dirz Checking .pth file support in %s.pthz.okzz import os f = open({ok_file!r}, 'w') f.write('OK') f.close() rOT)exist_okrNr pythonw.exe python.exer)spawnz-Ez-cpassz-TEST PASSED: %s appears to support .pth filesz+TEST FAILED: %s does NOT support .pth filesF)rtr rrr?r@rA _one_linerrdirname pkg_resourcesZ py31compatrrrrrrwriterrr executablerrrlowerdistutils.spawnrr~r) rrrZok_fileZ ok_existsrrrcrbasenameZaltZuse_altrr<r<r=rs\            z!easy_install.check_pth_processingc CsV|jsH|drH|dD],}|d|r.q||||d|q||dS)z=Write all the scripts for `dist`, unless scripts are excludedscriptszscripts/N)rvZmetadata_isdirZmetadata_listdirinstall_scriptZ get_metadatainstall_wrapper_scripts)rr script_namer<r<r=install_egg_scriptsYs z easy_install.install_egg_scriptscCsTtj|rDt|D]*\}}}|D]}|jtj||q$qn |j|dSrE)r?r@rwalkrrr)rr@baserfilesrr<r<r= add_outputgs  zeasy_install.add_outputcCs|jrtd|fdS)NzjInvalid argument %r: you can't use filenames or URLs with --editable (except via the --find-links option).)rirrrr<r<r= not_editableos zeasy_install.not_editablecCs<|js dStjtj|j|jr8td|j|jfdS)Nz2%r already exists in %s; can't do a checkout there)rir?r@rArrykeyrr!r<r<r=check_editablews zeasy_install.check_editablec cs:tjdd}zt|VW5tj|o2tt|XdS)Nz easy_install-)r)tempfilemkdtempr?r@rArrstr)rtmpdirr<r<r=_tmpdirs zeasy_install._tmpdirFc CsH|js||&}t|tst|rb|||j||}| d|||dW5QRSt j |r||| d|||dW5QRSt |}|||j|||j|j|j |j}|dkrd|}|jr|d7}t|nJ|jtkr||||d|W5QRS| ||j||W5QRSW5QRXdS)NTz+Could not find suitable distribution for %rz2 (--always-copy skips system and development eggs)Using)riinstall_site_pyr)rr(rr"rdownload install_itemr?r@rArr$Zfetch_distributionr[r}rrZ precedencer/process_distributionlocation)rrdepsr(dlrrr<r<r=r3s<        zeasy_install.easy_installcCs |p|j}|ptj||k}|p,|d }|pT|jdk oTtjt|t|jk}|r|s|j|jD]}|j |krjqqjd}t dtj ||r| |||}|D]}||||qn ||g}|||d|d|dk r|D]}||kr|SqdS)N.eggTz Processing %srr*)r}r?r@rendswithrr"r project_namer/r rr install_eggsr.egg_distribution)rrr,r(r0Zinstall_neededrZdistsr<r<r=r-s2     zeasy_install.install_itemcCs<t|}tD]*}d|}t||dkr t||||q dS)z=Sets the install directories by applying the install schemes.Zinstall_N)r r rr)rrschemer#attrnamer<r<r=rs zeasy_install.select_schemec Gs|||j|||j|jkr2|j||j|||||j|j<t |j ||f|| dr|j s|j |d|s|jsdS|dk r|j|jkrtd|dS|dks||kr|}tt|}t d|ztg|g|j|j}Wn^tk r<}ztt|W5d}~XYn0tk rj}zt|W5d}~XYnX|js||jr|D]"}|j|jkr||qt d|dS)Nzdependency_links.txtzSkipping dependencies for %szProcessing dependencies for %sz'Finished processing dependencies for %s) update_pthraddrr#removerrr rinstallation_report has_metadatarrZget_metadata_linesr}ras_requirementr(r'r,Zresolver3r-rr.Zreportr)rZ requirementrr0rZdistreqZdistrosrjr<r<r=r.sL           z!easy_install.process_distributioncCs2|jdk r|j S|dr dS|ds.dSdS)Nz not-zip-safeTzzip-safeF)rrr=rrr<r<r= should_unzips   zeasy_install.should_unzipcCstj|j|j}tj|r:d}t||j|j||Stj|rL|}nRtj ||krft |t |}t |dkrtj||d}tj|r|}t |t|||S)Nz<%r already exists in %s; build directory %s will not be keptrr)r?r@rryr#rAr rrrrlistdirrr$shutilmove)rr dist_filename setup_basedstrcontentsr<r<r= maybe_move s$       zeasy_install.maybe_movecCs,|jr dSt|D]}|j|qdSrE)rv ScriptWriterbestget_args write_script)rrrzr<r<r=r"sz$easy_install.install_wrapper_scriptscCsNt|}t||}|r8||t}t||}||t|ddS)z/Generate a legacy script wrapper and install itrdN) r'r>is_python_script_load_templaterrI get_headerrLrH)rrr script_textdev_pathrZ is_scriptZbodyr<r<r=r(s   zeasy_install.install_scriptcCs(d}|r|dd}td|}|dS)z There are a couple of template scripts in the package. This function loads one of them and prepares it for use. z script.tmplz.tmplz (dev).tmplrutf-8)rSr#decode)rQrZ raw_bytesr<r<r=rN2s   zeasy_install._load_templatetr<c sfdd|Dtd|jtjj|}|jrLdSt }t |tj |rpt |t |d|}||W5QRXt|d|dS)z1Write an executable file to the scripts directorycsg|]}tjj|qSr<)r?r@rrurr_rr<r=rDsz-easy_install.write_script..zInstalling %s script to %sNri)rr rrur?r@rr r current_umaskr$rArrrchmod)rrrGmodertargetmaskrcr<rr=rLAs   zeasy_install.write_scriptcCs^|dr|||gS|dr8|||gS|drT|||gS|}tj|r~|ds~t|||j ntj |rtj |}| |r|j r|dk r||||}tj|d}tj|s0ttj|dd}|stdtj |t|dkr(td tj ||d }|jrNt|||gS|||SdS) Nr2.exez.whl.pyzsetup.pyrz"Couldn't find a setup script in %srzMultiple setup scripts in %sr)rr3 install_egg install_exe install_wheelr?r@isfilerunpack_progressrabspath startswithryrHrrArrrrir rreport_editablebuild_and_install)rrrDr(rE setup_scriptZsetupsr<r<r=r5UsJ     zeasy_install.install_eggscCs>tj|r"t|tj|d}ntt|}tj ||dS)NEGG-INFO)metadata) r?r@rr*rr+ zipimport zipimporterr)Z from_filename)regg_pathrhr<r<r=r6s   zeasy_install.egg_distributionc Cstj|jtj|}tj|}|js2t|||}t ||sztj |rrtj |srt j ||jdn"tj|r|tj|fd|zd}tj |r||rtjd}}n tjd}}nL||r|||jd}}n*d}||r tjd}}n tjd}}||||f|dtj|tj|ft||d Wn$tk rxt|dd YnX||||S) Nr Removing FZMovingZCopyingZ ExtractingTz %s to %sfix_zipimporter_caches)r?r@rrtrrbrr$r6r2rrr remove_treerArrrcrBrCZcopytreer@Zmkpathunpack_and_compileZcopy2rupdate_dist_cachesrr )rrkr( destinationrZnew_dist_is_zippedrcrZr<r<r=r]s^                zeasy_install.install_eggc sPt|}|dkrtd|td|dd|ddtd}tj||d}||_ |d}tj|d}tj|d }t |t |||_ | ||tj|st|d } | d |dD].\} } | d kr| d | dd| fq| tj|d|fddt|Dtj|||j|jd|||S)Nz(%s is not a valid distutils Windows .exerhrrp)r4rpplatformr2z.tmprgPKG-INFOrzMetadata-Version: 1.0 target_versionz%s: %s _-rcsg|]}tj|dqS)r)r?r@r)rrzrur<r=rsz,easy_install.install_exe..)rr)r5rr)getrr?r@regg_namer/r$r*Z _provider exe_to_eggrArritemsrStitlerrrIrKrZ make_zipfilerrr]) rrDr(cfgrrkegg_tmpZ _egg_infoZpkg_infrckvr<ryr=r^sJ       zeasy_install.install_exec s6t|ggifdd}t||g}D]l}|dr<|d}|d}t|dd|d<tjj f|} || |t ||q<| t tj dt|dD]Z} t| rtj d| d } tj| st| d } | d t| d | qd S) z;Extract a bdist_wininst to the directories an egg would usecs|}D]\}}||r ||t|d}|d}tjjf|}|}|dsj|drt |d|d<dtj |dd< |n4|dr|dkrdtj |dd< ||Sq |d st d |dS) N/.pyd.dllrrr\SCRIPTS/r zWARNING: can't process %s)rrcrrr?r@rr3r strip_modulesplitextrr r)srcrFrGoldnewpartsr1r native_libsprefixes to_compile top_levelr<r=processs$        z(easy_install.exe_to_egg..processrrrr\rg)rrz.txtrrON)r7rrr3rrrr?r@rrZ write_stub byte_compileZwrite_safety_flagZ analyze_eggrrArrr) rrDrrZstubsresrresourceZpyfilerZtxtrcr<rr=r|s8          zeasy_install.exe_to_eggc Cst|}tj|j|}tj|}|js6t|tj |r`tj |s`t j ||jdn"tj |r|tj|fd|z.||j|fdtj|tj|fW5t|ddX||||S)NrlrmFrnzInstalling %s to %s)r r?r@rrtr{rbrr$rrr rprArrrrZinstall_as_eggrrr r6)rZ wheel_pathr(Zwheelrsr<r<r=r_!s2      zeasy_install.install_wheela( Because this distribution was installed --multi-version, before you can import modules from this package in an application, you will need to 'import pkg_resources' and then use a 'require()' call similar to one of these examples, in order to select the desired version: pkg_resources.require("%(name)s") # latest installed version pkg_resources.require("%(name)s==%(version)s") # this exact version pkg_resources.require("%(name)s>=%(version)s") # this version or higher z Note also that the installation directory must be on sys.path at runtime for this to work. (e.g. by being the application's script directory, by being on PYTHONPATH, or by being added to sys.path by your code.) Installedc Cs^d}|jr>|js>|d|j7}|jtttjkr>|d|j7}|j }|j }|j }d}|t S)z9Helpful installation message for display to package usersz %(what)s %(eggloc)s%(extras)srOr) r~r_easy_install__mv_warningrtrr"rr@_easy_install__id_warningr/r4rpr) rZreqrZwhatrZegglocrrpZextrasr<r<r=r<Os z easy_install.installation_reportaR Extracted editable version of %(spec)s to %(dirname)s If it uses setuptools in its setup script, you can activate it in "development" mode by going to that directory and running:: %(python)s setup.py develop See the setuptools documentation for the "develop" command for more info. cCs"tj|}tj}d|jtS)NrO)r?r@rrr_easy_install__editable_msgr)rrrfrpythonr<r<r=rdhs zeasy_install.report_editablec Cstjdttjdtt|}|jdkrNd|jd}|dd|n|jdkrd|dd|jrv|dd t d |t |ddd |zt ||Wn6tk r}ztd |jdfW5d}~XYnXdS) Nzdistutils.command.bdist_eggzdistutils.command.egg_inforrrrrx-qz-nz Running %s %s zSetup script exited with %s)rmodules setdefaultrrrrrrr rrrrrrrz)rrfrErzrr<r<r=rms&    zeasy_install.run_setupc Csddg}tjdtj|d}z| tj|| || |||t |g}g}|D]&}||D]}| | |j|qhq\|s|jstd||WSt|t|jXdS)Nrz --dist-dirz egg-dist-tmp-)rdirz+No eggs found in %s (setup script problem?))r%r&r?r@rrr rr_set_fetcher_optionsrrr'r]r/rr) rrfrErzZdist_dirZall_eggsZeggsr#rr<r<r=res*     zeasy_install.build_and_installc Csp|jd}d}i}|D]&\}}||kr2q |d||dd<q t|d}tj|d}t ||dS) a When easy_install is about to run bdist_egg on a source dist, that source dist might have 'setup_requires' directives, requiring additional fetching. Ensure the fetcher options given to easy_install are available to that command as well. r3)rxrrwr{rrrwrx)r3z setup.cfgN) rrcopyr}rSdictr?r@rrZ edit_config) rrZei_optsZfetch_directivesZ fetch_optionsr#rZsettingsZ cfg_filenamer<r<r=rs  z!easy_install._set_fetcher_optionscCs*|jdkrdS|j|jD]H}|js0|j|jkrtd||j||j|jkr|j|jq|js|j|jjkrtd|n2td||j ||j|jkr|j |j|j s&|j |jdkr&t j|jd}t j|rt |t|d}||j|jd|dS)Nz&Removing %s from easy-install.pth filez4%s is already the active version in easy-install.pthz"Adding %s to easy-install.pth filersetuptools.pthwtrO)rr#r~r/r rr;rpathsr:rrsaver?r@rrtrrrr make_relativer)rrr]rrcr<r<r=r9s6            zeasy_install.update_pthcCstd|||S)NzUnpacking %s to %s)r debug)rrrFr<r<r=raszeasy_install.unpack_progresscsdggfdd}t|||js`D]&}t|tjdBd@}t||q8dS)NcsZ|dr |ds |n|ds4|dr>|||j rV|pXdS)Nr\ EGG-INFO/rz.so)r3rcrrar)rrFrZto_chmodrr<r=pfs    z+easy_install.unpack_and_compile..pfimi)rrrr?statST_MODErW)rrkrsrrcrXr<rr=rqs  zeasy_install.unpack_and_compilec Csjtjr dSddlm}z@t|jd||dd|jd|jrT|||jd|jdW5t|jXdS)Nr)rr)r{forcer) rdont_write_bytecodedistutils.utilrr rrrr{)rrrr<r<r=rs  zeasy_install.byte_compilea bad install directory or PYTHONPATH You are attempting to install a package to a directory that is not on PYTHONPATH and which Python does not read ".pth" files from. The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s and your PYTHONPATH environment variable currently contains: %r Here are some of your options for correcting the problem: * You can choose a different installation directory, i.e., one that is on PYTHONPATH or supports .pth files * You can add the installation directory to the PYTHONPATH environment variable. (It must then also be on PYTHONPATH whenever you run Python and want to use the package(s) you are installing.) * You can set up the installation directory to support ".pth" files by using one of the approaches described here: https://setuptools.readthedocs.io/en/latest/easy_install.html#custom-installation-locations Please make the appropriate changes for your system and try again.cCs|j}||jtjddfSNZ PYTHONPATHr)_easy_install__no_default_msgrtr?environrz)rtemplater<r<r=rsz#easy_install.no_default_version_msgc Cs|jr dStj|jd}tdd}|d}d}tj|rt d|jt |}| }W5QRX| dstd |||krtd ||jst|t j |d dd }||W5QRX||gd |_dS)z8Make sure there's a site.py in the target dir, if neededNzsite.pyrz site-patch.pyrRrzChecking existing site.py in %sz def __boot():z;%s is not a setuptools-generated site.py; please remove it.z Creating %srencodingT)rr?r@rrtr#rSrAr riorreadrcrrrr$rr)rZsitepysourceZcurrentZstrmr<r<r=r+#s0       zeasy_install.install_site_pycCsd|js dSttjd}t|jD]8\}}||r&tj |s&| d|t |dq&dS)zCreate directories under ~.N~zos.makedirs('%s', 0o700)i) rqrr?r@rrZ iteritemsrrcrZ debug_printr)rhomerr@r<r<r=rCszeasy_install.create_home_pathz/$base/lib/python$py_version_short/site-packagesz $base/binrrz$base/Lib/site-packagesz $base/ScriptscGs|dj}|jrd|}|j|d<|jtj|j}| D]$\}}t ||ddkr>t |||q>ddl m }|D]B}t ||}|dk rt|||}tjdkrtj|}t |||qtdS)Nrrr)rr)Zget_finalized_commandrrrr rzr?rDEFAULT_SCHEMEr}rrrrr@r)rrrr7rrrr<r<r=rYs        zeasy_install._expand)F)F)T)N)rTr<)r)Q__name__ __module__ __qualname____doc__ descriptionZcommand_consumes_argumentsZ user_optionsZboolean_optionsrrrZhelp_msgrZ negative_optrrrrr staticmethodrrrrrrrrrrrPrQlstriprrr rrrr r"r$ contextlibcontextmanagerr)r3r-rr.r@rHrrrNrLr5r6r]r^r|r_rrr<rrdrrerr9rarqrrrr+rrr rrr<r<r<r=r3~s  0  z   0 ;   $ $ '    ,6-5   %  cCs tjddtj}td|Sr)r?rrzrpathsepfilter)r}r<r<r=rpsrc Cs~g}|ttjg}tjtjkr0|tj|D]}|r4tjdkr^|tj |ddnVtj dkr|tj |ddj tj dtj |ddgn||tj |ddgtjdkr4d |kr4tj d }|r4tj |d d d j tj d}||q4tdtdf}|D]}||kr||qtjrB|tjz|tWntk rjYnXttt|}|S)z& Return a list of 'site' dirs )Zos2emxZriscosZLibz site-packagesrlibz python{}.{}z site-pythondarwinzPython.frameworkHOMELibraryPythonrZpurelibZplatlib)extendrrrrrrtr?r@rseprrrrzrrrrgetsitepackagesAttributeErrorrrr")sitedirsrrrZhome_spZ lib_pathsZsite_libr<r<r=rus^             rccsi}|D]}t|}||krqd||<tj|s4qt|}||fV|D]}|ds\qL|dkrfqLttj||}tt |}| |D]H}| dst| }||krd||<tj|sq|t|fVqqLqdS)zBYield sys.path directories that might contain "old-style" packagesrr )rrimportN) r"r?r@rrAr3rrrr!rrcrstrip)Zinputsseenrrrrclinesliner<r<r= expand_pathss4        rc Cs@t|d}z$t|}|dkr*W dS|d|d|d}|dkrRWdS||dtd|d\}}}|dkrWdS||d|d d d }t |}z<||} | d d d } | t } |t| Wntjk rYW.dSX|dr"|ds*W dS|WS|XdS)znExtract configuration data from a bdist_wininst .exe Returns a configparser.RawConfigParser, or None rbN  zegg path translations for a given .exe file)zPURELIB/r)zPLATLIB/pywin32_system32r)zPLATLIB/r)rzEGG-INFO/scripts/)zDATA/lib/site-packagesrrrrrurz .egg-inforNrr z -nspkg.pth)ZPURELIBZPLATLIB\rz%s/%s/rcSsg|]\}}||fqSr<)r)rr_yr<r<r=r)sz$get_exe_prefixes..)rZZipFilerZinfolistrrrr3rrupperrrZPY3rSr!rRrSrcrsortreverse)Z exe_filenamerrXrrrrGZpthr<r<r=r7s@       " c@sReZdZdZdZdddZddZdd Zed d Z d d Z ddZ ddZ dS)r4z)A .pth file with Distribution paths in itFr<cCsl||_ttt||_ttj|j|_| t |gddt |j D]}tt|jt|dqLdS)NT)rrrr"rr?r@rbasedir_loadr'__init__r!rr:r&)rrrr@r<r<r=r4szPthDistributions.__init__cCsg|_d}t|j}tj|jrt|jd}|D]}| drHd}q4| }|j || r4| drtq4t tj|j|}|jd<tj|r||kr|jd|_q4d||<q4||jr|sd|_|jr|jd s|jqdS)NFZrtrT#rr)rrfromkeysrr?r@r`rrrcrrrRr"rrrApopdirtyr)rZ saw_importrrcrr@r<r<r=r=s4       zPthDistributions._loadc Cs|js dStt|j|j}|rtd|j||}d |d}t j |jr`t |jt|jd}||W5QRXn(t j |jrtd|jt |jd|_dS)z$Write changed .pth file back to diskNz Saving %srOrzDeleting empty %sF)rrrrrr rr _wrap_linesrr?r@rrrrrA)rZ rel_pathsrdatarcr<r<r=r\s   zPthDistributions.savecCs|SrEr<)rr<r<r=rrszPthDistributions._wrap_linescCsN|j|jko$|j|jkp$|jtk}|r>|j|jd|_t||dS)z"Add `dist` to the distribution mapTN) r/rrr?getcwdrrr'r:)rrnew_pathr<r<r=r:vs   zPthDistributions.addcCs2|j|jkr"|j|jd|_qt||dS)z'Remove `dist` from the distribution mapTN)r/rr;rr'r?r<r<r=r;s zPthDistributions.removecCstjt|\}}t|j}|g}tjdkr2dp6tj}t||kr||jkrl|tj | | |Stj|\}}||q8|S)Nr) r?r@rr"rraltseprrcurdirrr)rr@ZnpathZlastZbaselenrrr<r<r=rs      zPthDistributions.make_relativeN)r<) rrrrrrrrrrr:r;rr<r<r<r=r4/s  c@s(eZdZeddZedZedZdS)RewritePthDistributionsccs$|jV|D] }|Vq |jVdSrE)preludepostlude)clsrrr<r<r=rsz#RewritePthDistributions._wrap_linesz? import sys sys.__plen = len(sys.path) z import sys new = sys.path[sys.__plen:] del sys.path[sys.__plen:] p = getattr(sys, '__egginsert', 0) sys.path[p:p] = new sys.__egginsert = p + len(new) N)rrr classmethodrrrrr<r<r<r=rs rZSETUPTOOLS_SYS_PATH_TECHNIQUErawZrewritecCs ttjtrtSttjS)z_ Return a regular expression based on first_line_re suitable for matching strings. )rrpatternr'recompilerSr<r<r<r=_first_line_res rcCsd|tjtjfkr.tjdkr.t|tj||St\}}}t ||d|dd||ffdS)Nrrrz %s %s) r?rr;rrWrS_IWRITErrrZreraise)funcargexcZetZevrwr<r<r= auto_chmods  rcCs.t|}t|tj|r"t|nt|dS)aa Fix any globally cached `dist_path` related data `dist_path` should be a path of a newly installed egg distribution (zipped or unzipped). sys.path_importer_cache contains finder objects that have been cached when importing data from the original distribution. Any such finders need to be cleared since the replacement distribution might be packaged differently, e.g. a zipped egg distribution might get replaced with an unzipped egg folder or vice versa. Having the old finders cached may then cause Python to attempt loading modules from the replacement distribution using an incorrect loader. zipimport.zipimporter objects are Python loaders charged with importing data packaged inside zip archives. If stale loaders referencing the original distribution, are left behind, they can fail to load modules from the replacement distribution. E.g. if an old zipimport.zipimporter instance is used to load data from a new zipped egg archive, it may cause the operation to attempt to locate the requested data in the wrong location - one indicated by the original distribution's zip archive directory information. Such an operation may then fail outright, e.g. report having read a 'bad local file header', or even worse, it may fail silently & return invalid data. zipimport._zip_directory_cache contains cached zip archive directory information for all existing zipimport.zipimporter instances and all such instances connected to the same archive share the same cached directory information. If asked, and the underlying Python implementation allows it, we can fix all existing zipimport.zipimporter instances instead of having to track them down and remove them one by one, by updating their shared cached zip archive directory information. This, of course, assumes that the replacement distribution is packaged as a zipped egg. If not asked to fix existing zipimport.zipimporter instances, we still do our best to clear any remaining zipimport.zipimporter related cached data that might somehow later get used when attempting to load data from the new distribution and thus cause such load operations to fail. Note that when tracking down such remaining stale data, we can not catch every conceivable usage from here, and we clear only those that we know of and have found to cause problems if left alive. Any remaining caches should be updated by whomever is in charge of maintaining them, i.e. they should be ready to handle us replacing their zip archives with new distributions at runtime. N)r"_uncacherpath_importer_cache!_replace_zip_directory_cache_data*_remove_and_clear_zip_directory_cache_data)Z dist_pathronormalized_pathr<r<r=rrs <  rrcCsPg}t|}|D]:}t|}||r|||dtjdfkr||q|S)ap Return zipimporter cache entry keys related to a given normalized path. Alternative path spellings (e.g. those using different character case or those using alternative path separators) related to the same path are included. Any sub-path entries are included as well, i.e. those corresponding to zip archives embedded in other zip archives. rr)rr"rcr?rr)r cacheresult prefix_lenpZnpr<r<r="_collect_zipimporter_cache_entriess   rcCs@t||D]0}||}||=|o(|||}|dk r |||<q dS)a Update zipimporter cache data for a given normalized path. Any sub-path entries are processed as well, i.e. those corresponding to zip archives embedded in other zip archives. Given updater is a callable taking a cache entry key and the original entry (after already removing the entry from the cache), and expected to update the entry and possibly return a new one to be inserted in its place. Returning None indicates that the entry should not be replaced with a new one. If no updater is given, the cache entries are simply removed without any additional processing, the same as if the updater simply returned None. N)r)r rupdaterr old_entryZ new_entryr<r<r=_update_zipimporter_cache/s  rcCst||dSrE)r)r rr<r<r=r Osr cCsdd}t|tj|ddS)NcSs |dSrE)clearr@rr<r<r=2clear_and_remove_cached_zip_archive_directory_dataTszf_remove_and_clear_zip_directory_cache_data..clear_and_remove_cached_zip_archive_directory_datarrri_zip_directory_cache)r rr<r<r=r Ss r Z__pypy__cCsdd}t|tj|ddS)NcSs&|t||tj||SrE)rrirjupdaterrr<r<r=)replace_cached_zip_archive_directory_datajs zT_replace_zip_directory_cache_data..replace_cached_zip_archive_directory_datarr)r rr<r<r=r is  r c Cs4zt||dWnttfk r*YdSXdSdS)z%Is this string a valid Python script?execFTN)r SyntaxError TypeError)rTrr<r<r= is_python|s r"c CsNz(tj|dd}|d}W5QRXWnttfk rD|YSX|dkS)zCDetermine if the specified executable is a .sh (contains a #! line)zlatin-1rr#!)rrrrr)rfpmagicr<r<r=is_shs  r&cCs t|gS)z@Quote a command line argument according to Windows parsing rules subprocess list2cmdline)rr<r<r= nt_quote_argsr*cCsH|ds|drdSt||r&dS|drDd|dkSdS)zMIs this text, as a whole, a Python script? (as opposed to shell/bat/etc. r\.pywTr#rrF)r3r"rc splitlinesr)rPrr<r<r=rMs  rM)rWcGsdSrEr<)rzr<r<r=_chmodsr-c CsRtd||zt||Wn0tjk rL}ztd|W5d}~XYnXdS)Nzchanging mode of %s to %ozchmod failed: %s)r rr-r?error)r@rXrjr<r<r=rWs rWc@seZdZdZgZeZeddZeddZ eddZ edd Z ed d Z d d Z eddZddZeddZeddZdS) CommandSpeczm A command spec for a #! header, specified as a list of arguments akin to those passed to Popen. cCs|S)zV Choose the best CommandSpec class based on environmental conditions. r<rr<r<r=rJszCommandSpec.bestcCstjtj}tjd|S)N__PYVENV_LAUNCHER__)r?r@rCrrrrz)rZ_defaultr<r<r=_sys_executableszCommandSpec._sys_executablecCs:t||r|St|tr ||S|dkr0|S||S)zg Construct a CommandSpec from a parameter to build_scripts, which may be None. N)rrfrom_environment from_string)rZparamr<r<r= from_params  zCommandSpec.from_paramcCs||gSrE)r2r0r<r<r=r3szCommandSpec.from_environmentcCstj|f|j}||S)z} Construct a command spec from a simple string representing a command line parseable by shlex.split. )shlexr split_args)rstringr}r<r<r=r4szCommandSpec.from_stringcCs8t|||_t|}t|s4dg|jdd<dS)Nz-xr)r6r_extract_optionsoptionsr(r)rL)rrPcmdliner<r<r=install_optionss zCommandSpec.install_optionscCs:|dd}t|}|r.|dp0dnd}|S)zH Extract any options from the first line of the script. rOrrr)r,rmatchgrouprR)Z orig_scriptfirstr=r:r<r<r=r9s zCommandSpec._extract_optionscCs||t|jSrE)_renderrr:rr<r<r= as_headerszCommandSpec.as_headercCs6d}|D](}||r||r|ddSq|S)Nz"'rr)rcr3)itemZ_QUOTESqr<r<r= _strip_quotess zCommandSpec._strip_quotescCs tdd|D}d|dS)Ncss|]}t|VqdSrE)r/rDrR)rrBr<r<r=rsz&CommandSpec._render..r#rOr')r}r;r<r<r=r@s zCommandSpec._renderN)rrrrr:rr7rrJr2r5r3r4r<rr9rArDr@r<r<r<r=r/s*       r/c@seZdZeddZdS)WindowsCommandSpecFrN)rrrrr7r<r<r<r=rE srEc@seZdZdZedZeZ e dddZ e dddZ e dd d Z ed d Ze d dZe ddZe ddZe dddZdS)rIz` Encapsulates behavior around writing entry point scripts for console and gui apps. a # EASY-INSTALL-ENTRY-SCRIPT: %(spec)r,%(group)r,%(name)r __requires__ = %(spec)r import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point(%(spec)r, %(group)r, %(name)r)() ) NFcCs6tdt|rtnt}|d||}|||S)Nz Use get_argsr)warningsrEasyInstallDeprecationWarningWindowsScriptWriterrIrJget_script_headerrK)rrrwininstwriterheaderr<r<r=get_script_args$s zScriptWriter.get_script_argscCs$tjdtdd|rd}|||S)NzUse get_headerr) stacklevelr )rFrrGrO)rrPrrJr<r<r=rI,szScriptWriter.get_script_headerc cs|dkr|}t|}dD]Z}|d}||D]>\}}|||jt}|||||} | D] } | Vqlq:q dS)z Yield write_script() argument tuples for a distribution's console_scripts and gui_scripts entry points. NZconsoleguiZ_scripts) rOr'r>Z get_entry_mapr}_ensure_safe_namerr_get_script_args) rrrLrtype_r>rZeprPrzrr<r<r=rK4s   zScriptWriter.get_argscCstd|}|rtddS)z? Prevent paths in *_scripts entry point names. z[\\/]z+Path separators not allowed in script namesN)rsearchr)rZ has_path_sepr<r<r=rQFs zScriptWriter._ensure_safe_namecCs tdt|rtS|SNzUse best)rFrrGrHrJ)rZ force_windowsr<r<r= get_writerOs zScriptWriter.get_writercCs.tjdkstjdkr&tjdkr&tS|SdS)zD Select the best ScriptWriter for this environment. win32javarN)rrtr?r_namerHrJr0r<r<r=rJUszScriptWriter.bestccs|||fVdSrEr<)rrSrrLrPr<r<r=rR_szScriptWriter._get_script_argsrcCs"|j|}|||S)z;Create a #! line, getting options (if any) from script_text)command_spec_classrJr5r<rA)rrPrcmdr<r<r=rOds zScriptWriter.get_header)NF)NF)N)rN)rrrrrPrQrrr/rZrrMrIrKrrQrVrJrRrOr<r<r<r=rIs&       rIc@sLeZdZeZeddZeddZeddZeddZ e d d Z d S) rHcCstdt|SrU)rFrrGrJr0r<r<r=rVos zWindowsScriptWriter.get_writercCs"tt|d}tjdd}||S)zC Select the best ScriptWriter suitable for Windows )rZnaturalZSETUPTOOLS_LAUNCHERr)rWindowsExecutableLauncherWriterr?rrz)rZ writer_lookupZlauncherr<r<r=rJus zWindowsScriptWriter.bestc #stddd|}|tjddkrBdjft}t|t dddd d dd g}| || ||}fd d |D}|||d|fVdS)z For Windows, add a .py extensionz.pyar+rOZPATHEXT;zK{ext} not listed in PATHEXT; scripts will not be recognized as executables.r\ -script.py.pyc.pyor[csg|] }|qSr<r<rUrr<r=rsz8WindowsScriptWriter._get_script_args..rTN) rr?rrrrrrFr UserWarningr;_adjust_header) rrSrrLrPextrrrr<rar=rRs   z$WindowsScriptWriter._get_script_argscCsNd}d}|dkr||}}tt|tj}|j||d}||rJ|S|S)z Make sure 'pythonw' is used for gui and and 'python' is used for console (regardless of what sys.executable is). r r rP)r8repl)rrescape IGNORECASEsub _use_header)rrSZ orig_headerrreZ pattern_ob new_headerr<r<r=rcs z"WindowsScriptWriter._adjust_headercCs$|ddd}tjdkp"t|S)z Should _adjust_header use the replaced header? On non-windows systems, always use. On Windows systems, only use the replaced header if it resolves to an executable on the system. rr"rW)rRrrtr)rjZ clean_headerr<r<r=ris zWindowsScriptWriter._use_headerN) rrrrErZrrVrJrRrcrrir<r<r<r=rHls    rHc@seZdZeddZdS)r\c #s|dkrd}d}dg}nd}d}dddg}|||}fd d |D} |||d | fVd t|d fVtsd} | td fVdS)zG For Windows, add a .py extension and an .exe launcher rPz -script.pywr+Zclir^r\r_r`csg|] }|qSr<r<rUrar<r=rszDWindowsExecutableLauncherWriter._get_script_args..rTr[rdz .exe.manifestN)rcget_win_launcherr>load_launcher_manifest) rrSrrLrPZ launcher_typerdrZhdrrZm_namer<rar=rRs"  z0WindowsExecutableLauncherWriter._get_script_argsN)rrrrrRr<r<r<r=r\sr\cCs2d|}tr|dd}n |dd}td|S)z Load the Windows launcher (executable) suitable for launching a script. `type` should be either 'cli' or 'gui' Returns the executable as a byte string. z%s.exe.z-64.z-32.r)r>rSr#)typeZ launcher_fnr<r<r=rls  rlcCs0ttd}tjr|tS|dtSdS)Nzlauncher manifest.xmlrR)rr#rrPY2varsrS)rZmanifestr<r<r=rms  rmFcCst|||SrE)rBr)r@ ignore_errorsonerrorr<r<r=rsrcCstd}t||S)N)r?umask)Ztmpr<r<r=rVs  rVcCs:ddl}tj|jd}|tjd<tj|tdS)Nr) rr?r@r__path__rargvrr6)rZargv0r<r<r= bootstraps   rxc sddlm}ddlmGfddd}|dkrBtjdd}t0|fddd g|tjdpfd|d |W5QRXdS) Nr)setupr)cseZdZdZfddZdS)z-main..DistributionWithoutHelpCommandsrc s(tj|f||W5QRXdSrE) _patch_usage _show_help)rrzkwrzr<r=r| sz8main..DistributionWithoutHelpCommands._show_helpN)rrrZ common_usager|r<rzr<r=DistributionWithoutHelpCommands sr~rrr3z-v)Z script_argsrZ distclass)rryZsetuptools.distr)rrwr{)rwr}ryr~r<rzr=r6s    c#sLddl}tdfdd}|jj}||j_z dVW5||j_XdS)Nrze usage: %(script)s [options] requirement_or_url ... or: %(script)s --help csttj|dS)N)Zscript)rr?r@r)rZUSAGEr<r= gen_usage s z_patch_usage..gen_usage)Zdistutils.corerPrQrZcorer)rrZsavedr<rr=r{ s  r{c@seZdZdZdS)rGzuClass for warning about deprecations in EasyInstall in SetupTools. Not ignored by default, unlike DeprecationWarning.N)rrrrr<r<r<r=rG( srG)N)r)N)rrrrrrZdistutils.errorsrrrr Zdistutils.command.installr r rr r Zdistutils.command.build_scriptsrrrrr?rirBr%rrrrrPrFrr:rr(r6rZ sysconfigrrrrZsetuptools.externrZsetuptools.extern.six.movesrrrZsetuptools.sandboxrZsetuptools.py27compatrZsetuptools.commandrZsetuptools.archive_utilrZsetuptools.package_indexrrrrrZsetuptools.wheelr rr!r"r#r$r%r&r'r(r)r*r+r,r-r.r/Zpkg_resources.py31compatroZ __metaclass__filterwarningsZ PEP440Warning__all__r>r2rprHrLrr3rrrr5r7r4rrrzrrrrrrr r builtin_module_namesr r"r&r*rMrWr- ImportErrorrr/r2Zsys_executablerErIrHr\rMrIrlrmrrVrxr6rr{rGr<r<r<r=s            D {A))'l R    T^A   site-packages/setuptools/command/__pycache__/bdist_egg.cpython-38.pyc000064400000033546151464452710021736 0ustar00U Qab G@sxdZddlmZddlmZmZddlmZddlm Z ddl Z ddl Z ddl Z ddl Z ddlZddlmZddlmZmZmZdd lmZdd lmZdd lmZzdd lmZmZd dZWn,ek rddlm Z mZddZYnXddZ!ddZ"ddZ#GdddeZ$e%&d'Z(ddZ)ddZ*ddZ+d d!d"Z,d#d$Z-d%d&Z.d'd(Z/d)d*d+d,gZ0d1d/d0Z1dS)2z6setuptools.command.bdist_egg Build .egg distributions)DistutilsSetupError) remove_treemkpath)log)CodeTypeN)six)get_build_platform Distributionensure_directory) EntryPoint)Library)Command)get_pathget_python_versioncCstdS)NZpurelib)rrr@/usr/lib/python3.8/site-packages/setuptools/command/bdist_egg.py _get_purelibsr)get_python_librcCstdS)NF)rrrrrrscCs2d|krtj|d}|dr.|dd}|S)N.rmodulei)ospathsplitextendswith)filenamerrr strip_module#s   rccs6t|D]&\}}}|||||fVq dS)zbDo os.walk in a reproducible way, independent of indeterministic filesystem readdir order N)rwalksort)dirbasedirsfilesrrr sorted_walk+sr"c Cs6td}t|d}|||W5QRXdS)NaR def __bootstrap__(): global __bootstrap__, __loader__, __file__ import sys, pkg_resources, imp __file__ = pkg_resources.resource_filename(__name__, %r) __loader__ = None; del __bootstrap__, __loader__ imp.load_dynamic(__name__,__file__) __bootstrap__() w)textwrapdedentlstripopenwrite)resourcepyfileZ_stub_templatefrrr write_stub5s r,c@seZdZdZddddefdddd gZd d d gZd dZddZddZ ddZ ddZ ddZ ddZ ddZddZdd Zd!d"Zd#S)$ bdist_eggzcreate an "egg" distribution)z bdist-dir=bz1temporary directory for creating the distributionz plat-name=pz;platform name to embed in generated filenames (default: %s))exclude-source-filesNz+remove all .py files from the generated egg) keep-tempkzPkeep the pseudo-installation tree around after creating the distribution archive)z dist-dir=dz-directory to put final built distributions in) skip-buildNz2skip rebuilding everything (for testing/debugging)r1r4r0cCs.d|_d|_d|_d|_d|_d|_d|_dS)Nr) bdist_dir plat_name keep_tempdist_dir skip_build egg_outputexclude_source_filesselfrrrinitialize_optionsZszbdist_egg.initialize_optionscCs|d}|_|j|_|jdkr>|dj}tj|d|_|jdkrPt |_| dd|j dkrt dd|j |jt|jo|j }tj|j|d|_ dS)Negg_infoZbdistZegg)r8r8z.egg)get_finalized_commandei_cmdr?r5 bdist_baserrjoinr6rZset_undefined_optionsr:r Zegg_nameZ egg_versionr distributionhas_ext_modulesr8)r=rArBbasenamerrrfinalize_optionscs$     zbdist_egg.finalize_optionscCs|j|d_tjtjt}|jj g}|j_ |D]}t |t rt |dkrtj |drtj|d}tj|}||ks||tjr|t |dd|df}|jj |q:z"td|j|jddddW5||j_ XdS)Ninstallrzinstalling package data to %s install_data)forceroot)r5r@ install_librrnormcaserealpathrrD data_files isinstancetuplelenisabs startswithsepappendrinfo call_command)r=Z site_packagesolditemrPZ normalizedrrrdo_install_data{s"  zbdist_egg.do_install_datacCs|jgS)N)r:r<rrr get_outputsszbdist_egg.get_outputscKsPtD]}|||jq|d|j|d|j|j|f|}|||S)z8Invoke reinitialized command `cmdname` with keyword argsr9dry_run)INSTALL_DIRECTORY_ATTRS setdefaultr5r9r_Zreinitialize_command run_command)r=ZcmdnamekwdirnamecmdrrrrZs zbdist_egg.call_commandcCs|dtd|j|d}|j}d|_|jrH|jsH|d|j ddd}||_| \}}g|_ g}t |D]|\}}t j|\} } t j|jt| d} |j | td ||jstt j|| || |t jd ||<qz|r|||jjr||j} t j| d } || |jjrlt j| d }td ||j d|dd|| t j| d}|rtd||jst|t|d}| d|| d|!n,t j"|rtd||jst #|t$t j| d |%t j&t j|j'dr.+)\.(?P[^.]+)\.pycname.pyczRenaming file from [%s] to [%s])rrYwalk_eggr5rrrCrdebugrwrematchpardirgroupremoveOSErrorrename) r=rr r!rrZpath_oldpatternmZpath_newrrrr|s4       zbdist_egg.zap_pyfilescCs2t|jdd}|dk r|Stdt|j|jS)Nryz4zip_safe flag not set; analyzing archive contents...)rrDrr{ analyze_eggr5rq)r=saferrrry s  zbdist_egg.zip_safec Cst|jjpd}|did}|dkr0dS|jr<|jrJtd|fdjt j }|j }d |j}|jd}t j|j}d t}|jstt j|j|jd t|jd} | || d S) Nzsetuptools.installationZ eggsecutabler#zGeggsecutable entry point (%r) cannot have 'extras' or refer to a modulez{}.{}rraH#!/bin/sh if [ `basename $0` = "%(basename)s" ] then exec python%(pyver)s -c "import sys, os; sys.path.insert(0, os.path.abspath('$0')); from %(pkg)s import %(base)s; sys.exit(%(full)s())" "$@" else echo $0 is not the correct name for this egg file. echo Please rename it back to %(basename)s and try again. exec false fi roa)r Z parse_maprDZ entry_pointsgetZattrsZextrasrformatsys version_infoZ module_namerCrrrFr:localsr_rrdr'r(ru) r=ZepmZepZpyverpkgZfullrrFheaderr+rrrr~s0       zbdist_egg.gen_headercCshtj|j}tj|d}|jjjD]<}||r&tj||t |d}t || ||q&dS)z*Copy metadata (egg info) to the target_dirrN) rrnormpathr?rCrAZfilelistr!rVrTr Z copy_file)r=Z target_dirZ norm_egg_infoprefixrtargetrrrrt:s zbdist_egg.copy_metadata_toc Csg}g}|jdi}t|jD]f\}}}|D].}tj|dtkr*||||q*|D]"}|||d|tj||<q^q|j r| d}|j D]Z} t | trq|| j} || }tj|dstjtj|j|r||q||fS)zAGet a list of relative paths to C extensions in the output distrorrJrgZ build_extzdl-)r5r"rrrlowerNATIVE_EXTENSIONSrXrCrDrEr@ extensionsrRr Zget_ext_fullnamerZget_ext_filenamerFrVrz) r=rrpathsrr r!rZ build_cmdrfullnamerrrrpFs0        zbdist_egg.get_ext_outputsN)__name__ __module__ __qualname__ descriptionrZ user_optionsZboolean_optionsr>rGr]r^rZrr|ryr~rtrprrrrr-Cs2  Q' r-z.dll .so .dylib .pydccsHt|}t|\}}}d|kr(|d|||fV|D] }|Vq8dS)z@Walk an unpacked egg's contents, skipping the metadata directoryrhN)r"nextr)egg_dirZwalkerrr r!Zbdfrrrrfs  rc CstD](\}}tjtj|d|r|Sqtst||krzt|q|dk rt||krt|d}| d| qdS)Nrkrl) rrrrrCrzboolrwr'r(ru)rrrrr+rrrrxs    rxzzip-safez not-zip-safe)TFc Cstj||}|dd|kr"dS|t|ddtjd}||rJdpLdtj|d}tjrld}nt j d kr|d }nd }t |d }| |t |} |d} tt| } d D]} | | krtd|| d} qd| kr dD]} | | krtd|| d} q| S)z;Check whether module possibly uses unsafe-for-zipfile stuffNTrJrrr) rb)__file____path__z%s: module references %sFinspect) Z getsourceZ getabsfileZ getsourcefileZgetfilegetsourcelinesZ findsourceZ getcommentsZ getframeinfoZgetinnerframesZgetouterframesstackZtracez"%s: module MAY be using inspect.%s)rrrCrTrsrWrrZPY2rrr'readmarshalloadrudictfromkeys iter_symbolsrr{) rrrrqrrrskipr+coderZsymbolsZbadrrrrs4      rccsT|jD] }|Vq|jD]6}t|tjr0|Vqt|trt|D] }|VqBqdS)zBYield names and strings used by `code` and its nested code objectsN)co_names co_constsrRrZ string_typesrr)rrconstrrrrs     rcCs2tjdstjdkrdStdtddS)NjavaZcliTz1Unable to analyze compiled code on this platform.zfPlease ask the author to include a 'zip_safe' setting (either True or False) in the package's setup.py)rplatformrVrr{rrrrrs rrNrjrKZ install_baseTr#c sddl}ttj|dtd|fdd}|rB|jn|j}s|j |||d} t D]\} } } || | | qd| n t D]\} } } |d| | q|S)aqCreate a zip file from all the files under 'base_dir'. The output zip file will be named 'base_dir' + ".zip". Uses either the "zipfile" Python module (if available) or the InfoZIP "zip" utility (if installed and found on the default search path). If neither tool is available, raises DistutilsExecError. Returns the name of the output zip file. rNroz#creating '%s' and adding '%s' to itcs`|D]V}tjtj||}tj|r|tdd}sN|||td|qdS)NrJz adding '%s') rrrrCrvrTr(rr)zrdnamesrrr/base_dirr_rrvisits  zmake_zipfile..visit) compression) zipfilerrrrdrrYZ ZIP_DEFLATEDZ ZIP_STOREDZZipFiler"ru) Z zip_filenamerrmr_compressrnrrrrrdr r!rrrr}s  r})rrTr#)2__doc__Zdistutils.errorsrZdistutils.dir_utilrrZ distutilsrtypesrrrrr$rZsetuptools.externrZ pkg_resourcesrr r r Zsetuptools.extensionr Z setuptoolsr Z sysconfigrrr ImportErrorZdistutils.sysconfigrrr"r,r-rrsplitrrrrxrrrrr`r}rrrrsX         " $  site-packages/setuptools/command/__pycache__/develop.cpython-38.pyc000064400000014545151464452710021443 0ustar00U Qab@sddlmZddlmZddlmZmZddlZddlZddl Z ddl m Z ddl Z ddl mZddlmZddlZeZGdd d ejeZGd d d ZdS) ) convert_path)log)DistutilsErrorDistutilsOptionErrorN)six) easy_install) namespacesc@sveZdZdZdZejddgZejdgZdZddZ d d Z d d Z e d dZ ddZddZddZddZdS)developzSet up package for developmentz%install package in 'development mode') uninstalluzUninstall this source package)z egg-path=Nz-Set the path to be used in the .egg-link filer FcCs2|jrd|_||n||dS)NT)r Z multi_versionuninstall_linkZuninstall_namespacesinstall_for_developmentZwarn_deprecated_optionsselfr>/usr/lib/python3.8/site-packages/setuptools/command/develop.pyrun s  z develop.runcCs&d|_d|_t|d|_d|_dS)N.)r egg_pathrinitialize_options setup_pathZalways_copy_fromrrrrr)s  zdevelop.initialize_optionscCs|d}|jr,d}|j|jf}t|||jg|_t||| |j t d|jd}t j|j||_|j|_|jdkrt j|j|_t|j}tt j|j|j}||krtd|tj|t|t j|j|jd|_||j|j|j|_dS)Negg_infoz-Please rename %r to %r before using 'develop'z*.eggz .egg-linkzA--egg-path must be a relative path from the install directory to Z project_name)get_finalized_commandZbroken_egg_inforrZegg_nameargsrfinalize_optionsZexpand_basedirsZ expand_dirsZ package_indexscanglobospathjoin install_diregg_linkegg_baserabspath pkg_resourcesnormalize_pathrZ Distribution PathMetadatadist_resolve_setup_pathr)rZeitemplaterZ egg_link_fntargetrrrrr0sF        zdevelop.finalize_optionscCsn|tjdd}|tjkr0d|dd}ttj |||}|ttjkrjt d|ttj|S)z Generate a path from egg_base back to '.' where the setup script resides and ensure that path points to the setup path from $install_dir/$egg_path. /z../zGCan't get a consistent path to setup script from installation directory) replacerseprstripcurdircountr%r&rr r)r#r!rZ path_to_setupZresolvedrrrr)Zs  zdevelop._resolve_setup_pathc CsHtjrt|jddr|jddd|d|d}t|j }|jd|d|d|jddd|d|d}||_ ||j _ t ||j|j _n"|d|jdd d|d|tjr|tjdt_|td |j|j|js0t|jd }||j d |jW5QRX|d|j |j dS) NZuse_2to3FZbuild_pyr)Zinplacer)r#Z build_extr-zCreating %s (link to %s)w )rZPY3getattr distributionZreinitialize_commandZ run_commandrr%r&Z build_librr(locationr'rZ _providerZinstall_site_py setuptoolsZbootstrap_install_fromrZinstall_namespacesrinfor"r#dry_runopenwriterZprocess_distributionZno_deps)rZbpy_cmdZ build_pathZei_cmdfrrrr ns:           zdevelop.install_for_developmentcCstj|jrztd|j|jt|j}dd|D}|||j g|j |j gfkrht d|dS|j szt |j|j s||j|jjrt ddS)NzRemoving %s (link to %s)cSsg|] }|qSr)r0).0linerrr sz*develop.uninstall_link..z$Link points to %s: uninstall abortedz5Note: you must uninstall or replace scripts manually!)rrexistsr"rr9r#r;closerrwarnr:unlinkZ update_pthr(r6scripts)rZ egg_link_filecontentsrrrr s      zdevelop.uninstall_linkc Cs||jk rt||S|||jjp*gD]N}tjt |}tj |}t |}| }W5QRX|||||q,dSN)r(rinstall_egg_scriptsinstall_wrapper_scriptsr6rErrr$rbasenameior;readZinstall_script)rr(Z script_nameZ script_pathZstrmZ script_textrrrrHs     zdevelop.install_egg_scriptscCst|}t||SrG)VersionlessRequirementrrIrr(rrrrIszdevelop.install_wrapper_scriptsN)__name__ __module__ __qualname____doc__ descriptionrZ user_optionsZboolean_optionsZcommand_consumes_argumentsrrr staticmethodr)r r rHrIrrrrr s"  * 0r c@s(eZdZdZddZddZddZdS) rMa Adapt a pkg_resources.Distribution to simply return the project name as the 'requirement' so that scripts will work across multiple versions. >>> from pkg_resources import Distribution >>> dist = Distribution(project_name='foo', version='1.0') >>> str(dist.as_requirement()) 'foo==1.0' >>> adapted_dist = VersionlessRequirement(dist) >>> str(adapted_dist.as_requirement()) 'foo' cCs ||_dSrG)_VersionlessRequirement__distrNrrr__init__szVersionlessRequirement.__init__cCs t|j|SrG)r5rU)rnamerrr __getattr__sz"VersionlessRequirement.__getattr__cCs|jSrGrrrrras_requirementsz%VersionlessRequirement.as_requirementN)rOrPrQrRrVrXrYrrrrrMsrM)Zdistutils.utilrZ distutilsrZdistutils.errorsrrrrrKZsetuptools.externrr%Zsetuptools.command.easy_installrr8rtypeZ __metaclass__ZDevelopInstallerr rMrrrrs     6site-packages/setuptools/command/__pycache__/test.cpython-38.opt-1.pyc000064400000020433151464452710021714 0ustar00U Qab%@sddlZddlZddlZddlZddlZddlZddlmZmZddl m Z ddlm Z ddl m Z ddlmZmZddlmZmZmZmZmZmZmZmZmZddlmZd d lmZeZ Gd d d e Z!Gd ddZ"GdddeZ#dS)N)DistutilsErrorDistutilsOptionError)log) TestLoader)six)mapfilter) resource_listdirresource_existsnormalize_path working_set_namespace_packagesevaluate_markeradd_activation_listenerrequire EntryPoint)Command)_unique_everseenc@seZdZddZdddZdS)ScanningLoadercCst|t|_dSN)r__init__set_visitedselfr;/usr/lib/python3.8/site-packages/setuptools/command/test.pyrs zScanningLoader.__init__NcCs||jkrdS|j|g}|t||t|drH||t|drt|jdD]`}| dr|dkr|jd|dd}n"t |j|d r^|jd|}nq^|| |q^t |d kr| |S|d SdS) aReturn a suite of all tests cases contained in the given module If the module is a package, load tests from all the modules in it. If the module has an ``additional_tests`` function, call it and add the return value to the tests. Nadditional_tests__path__z.pyz __init__.py.z /__init__.pyrr)raddappendrloadTestsFromModulehasattrrr __name__endswithr ZloadTestsFromNamelenZ suiteClass)rmodulepatternZtestsfileZ submodulerrrr%s$      z"ScanningLoader.loadTestsFromModule)N)r' __module__ __qualname__rr%rrrrrsrc@seZdZddZdddZdS)NonDataPropertycCs ||_dSrfget)rr1rrrrAszNonDataProperty.__init__NcCs|dkr |S||Srr0)robjZobjtyperrr__get__DszNonDataProperty.__get__)N)r'r-r.rr3rrrrr/@sr/c@seZdZdZdZdddgZddZdd Zed d Z d d Z ddZ e j gfddZee j ddZeddZddZddZeddZeddZdS)testz.Command to run unit tests after in-place buildz0run unit tests after in-place build (deprecated))z test-module=mz$Run 'test_suite' in specified module)z test-suite=sz9Run single test, case or suite (e.g. 'module.test_suite'))z test-runner=rzTest runner to usecCsd|_d|_d|_d|_dSr) test_suite test_module test_loader test_runnerrrrrinitialize_optionsVsztest.initialize_optionscCs|jr|jrd}t||jdkrD|jdkr8|jj|_n |jd|_|jdkr^t|jdd|_|jdkrnd|_|jdkrt|jdd|_dS)Nz1You may specify a module or a suite, but not bothz .test_suiter:z&setuptools.command.test:ScanningLoaderr;)r8r9r distributionr:getattrr;)rmsgrrrfinalize_options\s        ztest.finalize_optionscCs t|Sr)list _test_argsrrrr test_argsosztest.test_argsccs4|jstjdkrdV|jr"dV|jr0|jVdS)N)Zdiscoverz --verbose)r8sys version_infoverboserrrrrBss ztest._test_argsc Cs| |W5QRXdS)zI Backward compatibility for project_on_sys_path context. N)project_on_sys_path)rfuncrrrwith_project_on_sys_path{s ztest.with_project_on_sys_pathc csPtjot|jdd}|rv|jddd|d|d}t|j}|jd|d|d|jddd|dn"|d|jdd d|d|d}t j dd}t j }zbt|j}t j d|ttd d td |j|jf||g dVW5QRXW5|t j dd<t j t j |tXdS) Nuse_2to3Fbuild_pyr)ZinplaceZegg_info)egg_baseZ build_extrcSs|Sr)Zactivate)distrrrz*test.project_on_sys_path..z%s==%s)rPY3r>r=Zreinitialize_commandZ run_commandZget_finalized_commandr Z build_librFpathmodulescopyclearupdater rrNinsertrrZegg_nameZ egg_versionpaths_on_pythonpath) rZ include_distsZ with_2to3Zbpy_cmdZ build_pathZei_cmdZold_pathZ old_modulesZ project_pathrrrrIs8             ztest.project_on_sys_pathc cst}tjd|}tjdd}zBtjt|}td||g}tj|}|r\|tjd<dVW5||kr~tjddn |tjd<XdS)z Add the indicated paths to the head of the PYTHONPATH environment variable so that subprocesses will also see the packages at these paths. Do this in a context that restores the value on exit. Z PYTHONPATHr N) objectosenvirongetpoppathsepjoinrr)pathsZnothingZorig_pythonpathZcurrent_pythonpathprefixZto_joinnew_pathrrrrYs    ztest.paths_on_pythonpathcCsD||j}||jpg}|dd|jD}t|||S)z Install the requirements indicated by self.distribution and return an iterable of the dists that were built. css0|](\}}|drt|ddr|VqdS):rN) startswithr).0kvrrr s z%test.install_dists..)Zfetch_build_eggsZinstall_requiresZ tests_requireZextras_requireitems itertoolschain)rOZir_dZtr_dZer_drrr install_distss   ztest.install_distsc Cs|dtj||j}d|j}|jr>|d|dS|d|tt d|}| |"| | W5QRXW5QRXdS)NzWARNING: Testing via this command is deprecated and will be removed in a future version. Users looking for a generic test entry point independent of test runner are encouraged to use tox. zskipping "%s" (dry run)z running "%s"location)announcerZWARNrmr=r`_argvZdry_runroperator attrgetterrYrI run_tests)rZinstalled_distscmdrarrrruns    ztest.runcCstjr~t|jddr~|jdd}|tkr~g}|tjkrD| ||d7}tjD]}| |rR| |qRt t tjj |tjdd|j||j||jdd}|jsd|j}||tjt|dS)NrLFr!r)Z testLoaderZ testRunnerexitzTest failed: %s)rrRr>r=r8splitr rFrTr$rerAr __delitem__unittestmainrq_resolve_as_epr:r;resultZ wasSuccessfulrprZERRORr)rr*Z del_modulesnamer4r?rrrrts.         ztest.run_testscCs dg|jS)Nrz)rCrrrrrq sz test._argvcCs$|dkr dStd|}|S)zu Load the indicated attribute value, called, as a as if it were specified as an entry point. Nzx=)rparseZresolve)valZparsedrrrr|sztest._resolve_as_epN)r'r-r.__doc__ descriptionZ user_optionsr<r@r/rCrBrK contextlibcontextmanagerrI staticmethodrYrmrvrtpropertyrqr|rrrrr4Js2 -   r4)$r[rrrFrrkrzZdistutils.errorsrrZ distutilsrrZsetuptools.externrZsetuptools.extern.six.movesrrZ pkg_resourcesr r r r r rrrrZ setuptoolsrrMrtypeZ __metaclass__rr/r4rrrrs"   ,  ) site-packages/setuptools/command/__pycache__/upload.cpython-38.pyc000064400000012123151464452710021257 0ustar00U Qab@sddlZddlZddlZddlZddlmZddlmZddlm Z ddl m Z ddl mZddlmZmZddlmZdd lmZGd d d e j Z dS) N)standard_b64encode)log)upload)spawn)DistutilsError)urlopenRequest) HTTPError)urlparsec@s8eZdZdZddZddZddZdd Zd d Zd S) rza Override default upload behavior to obtain password in a variety of different ways. c Cs&ztj|W5|dtjXdS)NzjWARNING: Uploading via this command is deprecated, use twine to upload instead (https://pypi.org/p/twine/))announcerZWARNorigrrunselfr=/usr/lib/python3.8/site-packages/setuptools/command/upload.pyr s z upload.runcCs8tj||jpt|_|jp0|p0||_dSN) r rfinalize_optionsusernamegetpassZgetuserpassword_load_password_from_keyring_prompt_for_passwordrrrrr"s zupload.finalize_optionsc Cst|j\}}}}}} |s"|s"| r0td|j|dkrDtd||jr|ddd|g} |jrnd|jg| dd<t| |jd t|d } | } W5QRX|j j } d d | | t j|| f||t| t| | | | | | | | | | | | | | !d }d|d<|jrdt j|dt|dd f|d<|j"d|j#$d}dt%|&d}d}d|$d}|d}t'(}|)D]\}}d|}t*|t+s|g}|D]j}t,|t-kr|d|d7}|d}nt|$d}|.||.|$d|.d|.|qܐq|.||/}d||jf}|0|t1j2d|tt3||d }t4|j||d!}zt5|}|6}|j7}Wnft8k r}z|j9}|j7}W5d}~XYn8t:k r"}z|0t|t1j;W5d}~XYnX|d"kr|0d#||ft1j2|j<rt=|d$d%d&|}|dk rd'>d(|d(f}|0|t1j2n"d)||f}|0|t1j;t?|dS)*NzIncompatible url %s)ZhttpZhttpszunsupported schema Zgpgz --detach-signz-az --local-user)dry_runrbZ file_upload1)z:actionZprotocol_versionnameversioncontentZfiletype pyversionZ md5_digestZmetadata_versionZsummaryZ home_pageZauthorZ author_emaillicense descriptionkeywordsplatformZ classifiersZ download_urlZprovidesZrequiresZ obsoletesZcommentz.ascZ gpg_signature:asciizBasic z3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254s --s-- z+ Content-Disposition: form-data; name="%s"z; filename="%s"rzutf-8s zSubmitting %s to %sz multipart/form-data; boundary=%s)z Content-typezContent-lengthZ Authorization)dataheaderszServer response (%s): %sZ_read_pypi_responsecSsdSrr)xrrrz$upload.upload_file.. zK---------------------------------------------------------------------------zUpload failed (%s): %s)@r repositoryAssertionErrorZsignZidentityrropenreadZ distributionZmetadataZget_nameZ get_versionospathbasenamehashlibZmd5Z hexdigeststrZget_metadata_versionZget_descriptionZget_urlZ get_contactZget_contact_emailZ get_licenceZget_long_descriptionZ get_keywordsZ get_platformsZget_classifiersZget_download_urlZ get_providesZ get_requiresZ get_obsoletesrrencoderdecodeioBytesIOitems isinstancelisttypetuplewritegetvaluer rINFOlenrrZgetcodemsgr codeOSErrorZERRORZ show_responsegetattrjoinr) rZcommandr filenameZschemaZnetlocZurlZparamsZqueryZ fragmentsZgpg_argsfrmetar)Z user_passZauthboundaryZ sep_boundaryZ end_boundaryZbodykeyvaluetitlerFr*ZrequestresultZstatusreasonetextrrr upload_file0s      !          zupload.upload_filecCs4ztd}||j|jWStk r.YnXdS)zM Attempt to load password from keyring. Suppress Exceptions. keyringN) __import__Z get_passwordr0r Exception)rrWrrrrs z"upload._load_password_from_keyringc Cs(z tWSttfk r"YnXdS)zH Prompt for a password on the tty. Suppress Exceptions. N)rrYKeyboardInterruptrrrrrs zupload._prompt_for_passwordN) __name__ __module__ __qualname____doc__r rrVrrrrrrrs  r)r;r4r7rbase64rZ distutilsrZdistutils.commandrr Zdistutils.spawnrZdistutils.errorsrZ*setuptools.extern.six.moves.urllib.requestrrZ(setuptools.extern.six.moves.urllib.errorr Z(setuptools.extern.six.moves.urllib.parser rrrrs       site-packages/setuptools/command/__pycache__/egg_info.cpython-38.pyc000064400000052404151464452710021556 0ustar00U Qabc@sdZddlmZddlmZddlmZddlm Z ddlZddlZddl Z ddl Z ddl Z ddl Z ddlZddlZddlZddlmZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZddlmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&ddl'm(Z(ddl)m*Z*ddlm+Z+ddlm,Z,ddZ-GdddZ.Gddde.eZ/GdddeZGdddeZ0ddZ1ddZ2d d!Z3d"d#Z4d$d%Z5d&d'Z6d(d)Z7d*d+Z8d5d-d.Z9d/d0Z:d1d2Z;Gd3d4d4e,Z|j|jkr>|j|_t|j|_ d|j_dS)Nz%s==%sz%s===%sz2Invalid distribution name or version syntax: %s-%srrP .egg-info-)!r:rQr?rRr isinstancerr>ZVersionlistr ValueError distutilserrorsZDistutilsOptionErrorrPr7Z package_dirgetr#curdirZensure_dirnamerrIr$joincheck_broken_egg_infometadataZ _patched_distkeylowerZ_versionZ_parsed_version)r9Zparsed_versionZ is_versionspecdirsZpdr2r2r3finalize_optionss>           zegg_info.finalize_optionsFcCsL|r||||n4tj|rH|dkr>|s>td||dS||dS)aWrite `data` to `filename` or delete if empty If `data` is non-empty, this routine is the same as ``write_file()``. If `data` is empty but not ``None``, this is the same as calling ``delete_file(filename)`. If `data` is ``None``, then this is a no-op unless `filename` exists, in which case a warning is issued about the orphaned file (if `force` is false), or deleted (if `force` is true). Nz$%s not set in setup(), but %s exists) write_filer#r$existsrwarn delete_file)r9whatrZdataforcer2r2r3write_or_delete_files   zegg_info.write_or_delete_filecCsDtd||tjr|d}|js@t|d}|||dS)zWrite `data` to `filename` (if not a dry run) after announcing it `what` is used in a log message to identify what is being written to the file. zwriting %s to %sutf-8wbN) rinforZPY3encodedry_runopenwriteclose)r9rqrZrrfr2r2r3rms   zegg_info.write_filecCs td||jst|dS)z8Delete `filename` (if not a dry run) after announcing itz deleting %sN)rrwryr#unlink)r9rZr2r2r3rps zegg_info.delete_filecCs||jt|jd|jj}tdD]4}|j|d|}|||j tj |j|j q*tj |jd}tj |r| ||dS)Nzegg_info.writers) installerznative_libs.txt)ZmkpathrIr#utimer7Zfetch_build_eggrZrequireZresolver:r$rernrp find_sources)r9repwriternlr2r2r3runs     z egg_info.runcCs4tj|jd}t|j}||_||j|_dS)z"Generate SOURCES.txt manifest filez SOURCES.txtN) r#r$rerImanifest_makerr7manifestrfilelist)r9Zmanifest_filenameZmmr2r2r3r*s  zegg_info.find_sourcescCsT|jd}|jtjkr&tj|j|}tj|rPtd||j |j |_ ||_ dS)Nr\aB------------------------------------------------------------------------------ Note: Your current .egg-info directory has a '-' in its name; this will not work correctly with "setup.py develop". Please rename %s to %s to correct this problem. ------------------------------------------------------------------------------) rQrPr#rdr$rernrrorIrS)r9Zbeir2r2r3rf2s   zegg_info.check_broken_egg_infoN)F)rErFrG descriptionZ user_optionsZboolean_optionsZ negative_optrTrHrUsetterr[rlrtrmrprrrfr2r2r2r3rIs.    1 rIc@s|eZdZddZddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ ddZddZddZdS)rcCs ||\}}}}|dkrR|dd||D]}||s2td|q2n|dkr|dd||D]}||srtd|qrn|dkr|d d||D]}||std |qnJ|d kr|d d||D]}||std |qn|dkr`|d|d|f|D]"}| ||s:td||q:n|dkr|d|d|f|D]"}| ||std||qnp|dkr|d|| |std|n>|dkr|d|| |std|n t d|dS)Nincludezinclude  z%warning: no files found matching '%s'excludezexclude z9warning: no previously-included files found matching '%s'zglobal-includezglobal-include z>warning: no files found matching '%s' anywhere in distributionzglobal-excludezglobal-exclude zRwarning: no previously-included files matching '%s' found anywhere in distributionzrecursive-includezrecursive-include %s %sz:warning: no files found matching '%s' under directory '%s'zrecursive-excludezrecursive-exclude %s %szNwarning: no previously-included files matching '%s' found under directory '%s'graftzgraft z+warning: no directories found matching '%s'prunezprune z6no previously-included directories found matching '%s'z'this cannot happen: invalid action '%s')Z_parse_template_line debug_printrerrrorglobal_includeglobal_excluderecursive_includerecursive_excluderrr)r9lineactionZpatternsdirZ dir_patternpatternr2r2r3process_template_lineEs|                  zFileList.process_template_linecCsRd}tt|jdddD]2}||j|r|d|j||j|=d}q|S)z Remove all files from the file list that match the predicate. Return True if any matching files were removed Frz removing T)ranger)filesr)r9Z predicatefoundr/r2r2r3 _remove_filesszFileList._remove_filescCs$ddt|D}||t|S)z#Include files that match 'pattern'.cSsg|]}tj|s|qSr2r#r$isdir.0r}r2r2r3 s z$FileList.include..rextendboolr9rrr2r2r3rs zFileList.includecCst|}||jS)z#Exclude files that match 'pattern'.)r4rmatchr9rrr2r2r3rszFileList.excludecCs8tj|d|}ddt|ddD}||t|S)zN Include all files anywhere in 'dir/' that match the pattern. rcSsg|]}tj|s|qSr2rrr2r2r3rs z.FileList.recursive_include..T) recursive)r#r$rerrr)r9rrZ full_patternrr2r2r3rs zFileList.recursive_includecCs ttj|d|}||jS)zM Exclude any file anywhere in 'dir/' that match the pattern. rr4r#r$rerr)r9rrrr2r2r3rszFileList.recursive_excludecCs$ddt|D}||t|S)zInclude all files from 'dir/'.cSs"g|]}tj|D]}|qqSr2)rarfindall)rZ match_diritemr2r2r3rsz"FileList.graft..r)r9rrr2r2r3rs  zFileList.graftcCsttj|d}||jS)zFilter out files from 'dir/'.rr)r9rrr2r2r3rszFileList.prunecsJ|jdkr|ttjd|fdd|jD}||t|S)z Include all files anywhere in the current directory that match the pattern. This is very inefficient on large file trees. Nrcsg|]}|r|qSr2rrrr2r3rs z+FileList.global_include..)Zallfilesrr4r#r$rerrrr2rr3rs   zFileList.global_includecCsttjd|}||jS)zD Exclude all files anywhere that match the pattern. rrrr2r2r3rszFileList.global_excludecCs8|dr|dd}t|}||r4|j|dS)N r)r<r _safe_pathrappend)r9rr$r2r2r3rs    zFileList.appendcCs|jt|j|dSr6)rrfilterr)r9pathsr2r2r3rszFileList.extendcCstt|j|j|_dS)z Replace self.files with only safe paths Because some owners of FileList manipulate the underlying ``files`` attribute directly, this method must be called to repair those paths. N)r_rrrr8r2r2r3_repairszFileList._repairc Csd}t|}|dkr(td|dSt|d}|dkrNt||ddSz"tj|shtj|rnWdSWn&tk rt||t YnXdS)Nz!'%s' not %s encodable -- skippingz''%s' in unexpected encoding -- skippingFruT) unicode_utilsfilesys_decoderroZ try_encoder#r$rnUnicodeEncodeErrorsysgetfilesystemencoding)r9r$Zenc_warnZu_pathZ utf8_pathr2r2r3rs   zFileList._safe_pathN)rErFrGrrrrrrrrrrrrrrr2r2r2r3rBsI     rc@s\eZdZdZddZddZddZdd Zd d Zd d Z e ddZ ddZ ddZ dS)rz MANIFEST.incCsd|_d|_d|_d|_dS)Nr)Z use_defaultsrZ manifest_onlyZforce_manifestr8r2r2r3rT sz!manifest_maker.initialize_optionscCsdSr6r2r8r2r2r3rlszmanifest_maker.finalize_optionscCsdt|_tj|js||tj|jr<| | |j |j |dSr6) rrr#r$rnrwrite_manifest add_defaultstemplateZ read_templateprune_file_listsortZremove_duplicatesr8r2r2r3rs  zmanifest_maker.runcCst|}|tjdS)N/)rrreplacer#r%)r9r$r2r2r3_manifest_normalizes z"manifest_maker._manifest_normalizecsBjfddjjD}dj}tj|f|dS)zo Write the file list in 'self.filelist' to the manifest file named by 'self.manifest'. csg|]}|qSr2)rrr8r2r3r*sz1manifest_maker.write_manifest..zwriting manifest file '%s'N)rrrrZexecuterm)r9rmsgr2r8r3r"s  zmanifest_maker.write_manifestcCs||st||dSr6)_should_suppress_warningr ro)r9rr2r2r3ro.s zmanifest_maker.warncCs td|S)z; suppress missing-file warnings from sdist zstandard file .*not found)r&r)rr2r2r3r2sz'manifest_maker._should_suppress_warningcCst|||j|j|j|jtt}|rJ|j |nt j |jr`| t j drx|jd|d}|j|jdS)Nzsetup.pyrI)r rZ check_licenserrrrr_r rr#r$rnZ read_manifestget_finalized_commandrrI)r9ZrcfilesZei_cmdr2r2r3r9s     zmanifest_maker.add_defaultscCsZ|d}|j}|j|j|j|ttj }|jj d|d|dddS)Nbuildz(^|z)(RCS|CVS|\.svn)r)Zis_regex) rr7Z get_fullnamerrZ build_baser&r'r#r%Zexclude_pattern)r9rZbase_dirr%r2r2r3rLs    zmanifest_maker.prune_file_listN)rErFrGrrTrlrrrro staticmethodrrrr2r2r2r3rs   rc Cs8d|}|d}t|d}||W5QRXdS)z{Create a file with the specified name and write 'contents' (a sequence of strings without line terminators) to it.  rurvN)rerxrzr{)rZcontentsr}r2r2r3rmVs   rmc Cs|td||jsx|jj}|j|j|_}|j|j|_}z| |j W5|||_|_Xt |jdd}t |j |dS)Nz writing %sZzip_safe)rrwryr7rgrRr>rQr:write_pkg_inforIgetattrr Zwrite_safety_flag)cmdbasenamerZrgZoldverZoldnameZsafer2r2r3rcs rcCstj|rtddS)NzsWARNING: 'depends.txt' is not used by setuptools 0.6! Use the install_requires/extras_require setup() args instead.)r#r$rnrrorrrZr2r2r3warn_depends_obsoletevs rcCs,t|pd}dd}t||}||dS)Nr2cSs|dS)Nrr2)rr2r2r3z%_write_requirements..)rr writelines)streamZreqslinesZ append_crr2r2r3_write_requirements~s  rcCsj|j}t}t||j|jp"i}t|D]&}|djft t|||q,| d|| dS)Nz [{extra}] Z requirements) r7rStringIOrZinstall_requiresextras_requiresortedr{formatvarsrtgetvalue)rrrZZdistrrrZextrar2r2r3write_requirementss   rcCs,t}t||jj|d||dS)Nzsetup-requirements)iorrr7Zsetup_requiresrtr)rrrZrrr2r2r3write_setup_requirementssrcCs:tdd|jD}|d|dt|ddS)NcSsg|]}|dddqS).rr)r")rkr2r2r3rsz(write_toplevel_names..ztop-level namesr)rYfromkeysr7Ziter_distribution_namesrmrer)rrrZZpkgsr2r2r3write_toplevel_namess rcCst|||ddS)NT) write_argrr2r2r3 overwrite_argsrFcCsHtj|d}t|j|d}|dk r4d|d}|||||dS)Nrr)r#r$splitextrr7rert)rrrZrsargnamerVr2r2r3rs rcCs|jj}t|tjs|dkr"|}nn|dk rg}t|D]J\}}t|tjsrt||}d tt t | }| d||fq:d |}|d||ddS)Nrz [%s] %s rz entry pointsT)r7Z entry_pointsr^rZ string_typesritemsrZ parse_grouprerstrvaluesrrt)rrrZrrrZsectionrr2r2r3 write_entriess   rc Csjtdttjdrftd>}|D]2}t d|}|r(t | dW5QRSq(W5QRXdS)zd Get a -r### off of PKG-INFO Version in case this is an sdist of a subversion revision. z$get_pkg_info_revision is deprecated.zPKG-INFOzVersion:.*-r(\d+)\s*$rr) warningsroEggInfoDeprecationWarningr#r$rnrrzr&rintgroup)r}rrr2r2r3get_pkg_info_revisions    *rc@seZdZdZdS)rzqClass for warning about deprecations in eggInfo in setupTools. Not ignored by default, unlike DeprecationWarning.N)rErFrG__doc__r2r2r2r3rsr)F)=rZdistutils.filelistrZ _FileListZdistutils.errorsrZdistutils.utilrrarr#r&rrrrBrWZsetuptools.externrZsetuptools.extern.six.movesrZ setuptoolsrZsetuptools.command.sdistr r Zsetuptools.command.setoptr Zsetuptools.commandr Z pkg_resourcesr rrrrrrrZsetuptools.unicode_utilsrZsetuptools.globrrrr4r5rIrrmrrrrrrrrrrrr2r2r2r3sX           (    S2EP    site-packages/setuptools/command/__pycache__/install_lib.cpython-38.opt-1.pyc000064400000007772151464452710023244 0ustar00U Qab@sHddlZddlZddlmZmZddlmmZGdddejZdS)N)productstarmapc@sZeZdZdZddZddZddZedd Zd d Z ed d Z dddZ ddZ dS) install_libz9Don't add compiled flags to filenames of non-Python filescCs&||}|dk r"||dSN)ZbuildinstallZ byte_compile)selfoutfilesr B/usr/lib/python3.8/site-packages/setuptools/command/install_lib.pyrun szinstall_lib.runcs4fddD}t|}ttj|S)z Return a collections.Sized collections.Container of paths to be excluded for single_version_externally_managed installations. c3s"|]}|D] }|VqqdSr) _all_packages).0Zns_pkgpkgrr r s z-install_lib.get_exclusions..)_get_SVEM_NSPsr_gen_exclusion_pathssetr_exclude_pkg_path)rZ all_packagesZ excl_specsr rr get_exclusionss  zinstall_lib.get_exclusionscCs$|d|g}tjj|jf|S)zw Given a package name and exclusion path within that package, compute the full exclusion path. .)splitospathjoinZ install_dir)rrZexclusion_pathpartsr r r rszinstall_lib._exclude_pkg_pathccs |r|V|d\}}}qdS)zn >>> list(install_lib._all_packages('foo.bar.baz')) ['foo.bar.baz', 'foo.bar', 'foo'] rN) rpartition)Zpkg_namesepZchildr r r r 'szinstall_lib._all_packagescCs,|jjs gS|d}|j}|r(|jjSgS)z Get namespace packages (list) but only for single_version_externally_managed installations and empty otherwise. r)Z distributionZnamespace_packagesZget_finalized_commandZ!single_version_externally_managed)rZ install_cmdZsvemr r r r1s  zinstall_lib._get_SVEM_NSPsccsbdVdVdVttds dStjddtjj}|dV|d V|d V|d VdS) zk Generate file paths to be excluded for namespace packages (bytecode cache files). z __init__.pyz __init__.pycz __init__.pyoimplementationN __pycache__z __init__.z.pycz.pyoz .opt-1.pycz .opt-2.pyc)hasattrsysrrrr cache_tag)baser r r rAs    z install_lib._gen_exclusion_pathsrc sX|stj|||Sddlm}ddlmgfdd}||||S)Nr)unpack_directory)logcs<|krd|dSd|tj|||S)Nz/Skipping installation of %s (namespace package)Fzcopying %s -> %s)warninforrdirnameappend)srcZdstexcluder&rr r pfgs z!install_lib.copy_tree..pf)rorigr copy_treeZsetuptools.archive_utilr%Z distutilsr&) rZinfileZoutfileZ preserve_modeZpreserve_timesZpreserve_symlinkslevelr%r.r r,r r0Vs   zinstall_lib.copy_treecs.tj|}|r*fdd|DS|S)Ncsg|]}|kr|qSr r )r fr-r r xsz+install_lib.get_outputs..)r/r get_outputsr)rZoutputsr r3r r5ts  zinstall_lib.get_outputsN)r$r$rr$) __name__ __module__ __qualname____doc__r rr staticmethodr rrr0r5r r r r rs   r) rr! itertoolsrrZdistutils.command.install_libZcommandrr/r r r r ssite-packages/setuptools/command/__pycache__/build_clib.cpython-38.opt-1.pyc000064400000004605151464452710023030 0ustar00U Qab@sLddlmmZddlmZddlmZddlm Z GdddejZdS)N)DistutilsSetupError)log)newer_pairwise_groupc@seZdZdZddZdS) build_clibav Override the default build_clib behaviour to do the following: 1. Implement a rudimentary timestamp-based dependency system so 'compile()' doesn't run every time. 2. Add more keys to the 'build_info' dictionary: * obj_deps - specify dependencies for each object compiled. this should be a dictionary mapping a key with the source filename to a list of dependencies. Use an empty string for global dependencies. * cflags - specify a list of additional flags to pass to the compiler. c Csr|D]f\}}|d}|dks.t|ttfs:td|t|}td||dt}t|tsrtd|g}|dt}t|ttfstd||D]P}|g} | |||t} t| ttfstd|| | | | q|j j ||j d} t || ggfkrT|d} |d } |d }|j j||j | | ||jd }|j j| ||j|jd qdS) Nsourceszfin 'libraries' option (library '%s'), 'sources' must be present and must be a list of source filenameszbuilding '%s' libraryobj_depsz\in 'libraries' option (library '%s'), 'obj_deps' must be a dictionary of type 'source: list') output_dirmacros include_dirscflags)r r r Zextra_postargsdebug)r r )get isinstancelisttuplerrinfodictextendappendZcompilerZobject_filenamesZ build_temprcompiler Zcreate_static_libr)selfZ librariesZlib_nameZ build_inforrZ dependenciesZ global_depssourceZsrc_depsZ extra_depsZexpected_objectsr r r ZobjectsrA/usr/lib/python3.8/site-packages/setuptools/command/build_clib.pybuild_librariessv          zbuild_clib.build_librariesN)__name__ __module__ __qualname____doc__rrrrrrsr) Zdistutils.command.build_clibZcommandrZorigZdistutils.errorsrZ distutilsrZsetuptools.dep_utilrrrrrs   site-packages/setuptools/command/__pycache__/bdist_wininst.cpython-38.opt-1.pyc000064400000001672151464452710023621 0ustar00U Qab}@s(ddlmmZGdddejZdS)Nc@seZdZdddZddZdS) bdist_wininstrcCs |j||}|dkrd|_|S)zj Supplement reinitialize_command to work around http://bugs.python.org/issue20819 )Zinstall install_libN)Z distributionreinitialize_commandr)selfcommandZreinit_subcommandscmdrD/usr/lib/python3.8/site-packages/setuptools/command/bdist_wininst.pyrsz"bdist_wininst.reinitialize_commandcCs$d|_ztj|W5d|_XdS)NTF)Z _is_runningorigrrun)rrrr r szbdist_wininst.runN)r)__name__ __module__ __qualname__rr rrrr rs r)Zdistutils.command.bdist_wininstrrr rrrr ssite-packages/setuptools/command/__pycache__/install_lib.cpython-38.pyc000064400000010034151464452710022266 0ustar00U Qab@sHddlZddlZddlmZmZddlmmZGdddejZdS)N)productstarmapc@sZeZdZdZddZddZddZedd Zd d Z ed d Z dddZ ddZ dS) install_libz9Don't add compiled flags to filenames of non-Python filescCs&||}|dk r"||dSN)ZbuildinstallZ byte_compile)selfoutfilesr B/usr/lib/python3.8/site-packages/setuptools/command/install_lib.pyrun szinstall_lib.runcs4fddD}t|}ttj|S)z Return a collections.Sized collections.Container of paths to be excluded for single_version_externally_managed installations. c3s"|]}|D] }|VqqdSr) _all_packages).0Zns_pkgpkgrr r s z-install_lib.get_exclusions..)_get_SVEM_NSPsr_gen_exclusion_pathssetr_exclude_pkg_path)rZ all_packagesZ excl_specsr rr get_exclusionss  zinstall_lib.get_exclusionscCs$|d|g}tjj|jf|S)zw Given a package name and exclusion path within that package, compute the full exclusion path. .)splitospathjoinZ install_dir)rrZexclusion_pathpartsr r r rszinstall_lib._exclude_pkg_pathccs |r|V|d\}}}qdS)zn >>> list(install_lib._all_packages('foo.bar.baz')) ['foo.bar.baz', 'foo.bar', 'foo'] rN) rpartition)Zpkg_namesepZchildr r r r 'szinstall_lib._all_packagescCs,|jjs gS|d}|j}|r(|jjSgS)z Get namespace packages (list) but only for single_version_externally_managed installations and empty otherwise. r)Z distributionZnamespace_packagesZget_finalized_commandZ!single_version_externally_managed)rZ install_cmdZsvemr r r r1s  zinstall_lib._get_SVEM_NSPsccsbdVdVdVttds dStjddtjj}|dV|d V|d V|d VdS) zk Generate file paths to be excluded for namespace packages (bytecode cache files). z __init__.pyz __init__.pycz __init__.pyoimplementationN __pycache__z __init__.z.pycz.pyoz .opt-1.pycz .opt-2.pyc)hasattrsysrrrr cache_tag)baser r r rAs    z install_lib._gen_exclusion_pathsrc sh|r |r |rt|s,tj|||Sddlm}ddlmgfdd}||||S)Nr)unpack_directory)logcs<|krd|dSd|tj|||S)Nz/Skipping installation of %s (namespace package)Fzcopying %s -> %s)warninforrdirnameappend)srcZdstexcluder&rr r pfgs z!install_lib.copy_tree..pf) AssertionErrorrorigr copy_treeZsetuptools.archive_utilr%Z distutilsr&) rZinfileZoutfileZ preserve_modeZpreserve_timesZpreserve_symlinkslevelr%r.r r,r r1Vs   zinstall_lib.copy_treecs.tj|}|r*fdd|DS|S)Ncsg|]}|kr|qSr r )r fr-r r xsz+install_lib.get_outputs..)r0r get_outputsr)rZoutputsr r4r r6ts  zinstall_lib.get_outputsN)r$r$rr$) __name__ __module__ __qualname____doc__r rr staticmethodr rrr1r6r r r r rs   r) rr! itertoolsrrZdistutils.command.install_libZcommandrr0r r r r ssite-packages/setuptools/command/__pycache__/build_py.cpython-38.opt-1.pyc000064400000020701151464452710022542 0ustar00U Qab|%@sddlmZddlmZddlmmZddlZddlZddl Z ddl Z ddl Z ddl Z ddlmZddlmZmZmZzddlmZWn"ek rGdddZYnXGd d d ejeZdd d Zd dZdS))glob) convert_pathN)six)mapfilter filterfalse) Mixin2to3c@seZdZdddZdS)rTcCsdS)z do nothingN)selffilesZdoctestsr r ?/usr/lib/python3.8/site-packages/setuptools/command/build_py.pyrun_2to3szMixin2to3.run_2to3N)T)__name__ __module__ __qualname__r r r r r rsrc@seZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ ddZ ddZddZddZddZeddZd S)!build_pyaXEnhanced 'build_py' command that includes data files with packages The data files are specified via a 'package_data' argument to 'setup()'. See 'setuptools.dist.Distribution' for more details. Also, this version of the 'build_py' command allows you to specify both 'py_modules' and 'packages' in the same setup operation. cCsFtj||jj|_|jjp i|_d|jkr6|jd=g|_g|_dS)N data_files) origrfinalize_options distribution package_dataexclude_package_data__dict___build_py__updated_files_build_py__doctests_2to3r r r r r!s   zbuild_py.finalize_optionscCsx|js|jsdS|jr||jr4||||jd||jd||jd|t j j |dddS)z?Build modules, packages, and copy data files to build directoryNFTr)Zinclude_bytecode) Z py_modulespackagesZ build_modulesZbuild_packagesbuild_package_datar rrZ byte_compilerrZ get_outputsrr r r run+s z build_py.runcCs&|dkr||_|jStj||S)zlazily compute data filesr)_get_data_filesrrr __getattr__)r attrr r r r ?s zbuild_py.__getattr__cCsJtjrt|tjr|d}tj||||\}}|rB|j |||fS)N.) rZPY2 isinstanceZ string_typessplitrr build_modulerappend)r moduleZ module_filepackageZoutfilecopiedr r r r%Fs   zbuild_py.build_modulecCs|tt|j|jpdS)z?Generate list of '(package,src_dir,build_dir,filenames)' tuplesr )analyze_manifestlistr_get_pkg_data_filesrrr r r rPszbuild_py._get_data_filescsJ||tjj|jg|d}fdd||D}|||fS)Nr"csg|]}tj|qSr )ospathrelpath).0filesrc_dirr r ]sz0build_py._get_pkg_data_files..)get_package_dirr-r.joinZ build_libr$find_data_files)r r( build_dir filenamesr r2r r,Us    zbuild_py._get_pkg_data_filescCsX||j||}tt|}tj|}ttj j |}t|j |g|}| |||S)z6Return filenames for package's data files in 'src_dir')_get_platform_patternsrrr itertoolschain from_iterablerr-r.isfilemanifest_filesgetexclude_data_files)r r(r3patternsZglobs_expandedZ globs_matchesZ glob_filesr r r r r7cs   zbuild_py.find_data_filesc Cs|jD]|\}}}}|D]j}tj||}|tj|tj||}|||\}} tj|}| r||jj kr|j |qqdS)z$Copy data files into build directoryN) rr-r.r6ZmkpathdirnameZ copy_fileabspathrZconvert_2to3_doctestsrr&) r r(r3r8r9filenametargetZsrcfileZoutfr)r r r rts  zbuild_py.build_package_datac Csi|_}|jjsdSi}|jp"dD]}||t||<q$|d|d}|jj D]}t j t|\}}d}|} |r||kr||kr|}t j |\}} t j | |}qx||krX|dr|| krqX|||g|qXdS)Nr Zegg_infoz.py)r?rZinclude_package_datarassert_relativer5Z run_commandZget_finalized_commandZfilelistr r-r.r$r6endswith setdefaultr&) r ZmfZsrc_dirsr(Zei_cmdr.dfprevZoldfZdfr r r r*s(    zbuild_py.analyze_manifestcCsdSNr rr r r get_data_filesszbuild_py.get_data_filesc Csz |j|WStk r YnXtj|||}||j|<|rH|jjsL|S|jjD]}||ksn||drTqxqT|St |d}| }W5QRXd|krt j d|f|S)z8Check namespace packages' __init__ for declare_namespacer"rbsdeclare_namespacezNamespace package problem: %s is a namespace package, but its __init__.py does not call declare_namespace()! Please fix it. (See the setuptools manual under "Namespace Packages" for details.) ")packages_checkedKeyErrorrr check_packagerZnamespace_packages startswithioopenread distutilserrorsZDistutilsError)r r(Z package_dirZinit_pyZpkgrKcontentsr r r rRs*    zbuild_py.check_packagecCsi|_tj|dSrM)rPrrinitialize_optionsrr r r rZszbuild_py.initialize_optionscCs0tj||}|jjdk r,tj|jj|S|SrM)rrr5rZsrc_rootr-r.r6)r r(resr r r r5s zbuild_py.get_package_dircs\t||j||}fdd|D}tj|}t|fddD}tt|S)z6Filter filenames for package's data files in 'src_dir'c3s|]}t|VqdSrM)fnmatchrr0pattern)r r r sz.build_py.exclude_data_files..c3s|]}|kr|VqdSrMr )r0fn)badr r r_s)r+r:rr;r<r=set_unique_everseen)r r(r3r rBZ match_groupsZmatchesZkeepersr )rar r rAs   zbuild_py.exclude_data_filescs.t|dg||g}fdd|DS)z yield platform-specific path patterns (suitable for glob or fn_match) from a glob-based spec (such as self.package_data or self.exclude_package_data) matching package in src_dir. c3s |]}tjt|VqdSrM)r-r.r6rr]r2r r r_sz2build_py._get_platform_patterns..)r;r<r@)specr(r3Z raw_patternsr r2r r:s   zbuild_py._get_platform_patternsN)rrr__doc__rrr r%rr,r7rr*rNrRrZr5rA staticmethodr:r r r r rs"    rccsbt}|j}|dkr6t|j|D]}|||Vq n(|D]"}||}||kr:|||Vq:dS)zHList unique elements, preserving order. Remember all elements ever seen.N)rbaddr __contains__)iterablekeyseenZseen_addZelementkr r r rcs rccCs:tj|s|Sddlm}td|}||dS)Nr)DistutilsSetupErrorz Error: setup script specifies an absolute path: %s setup() arguments must *always* be /-separated paths relative to the setup.py directory, *never* absolute paths. )r-r.isabsdistutils.errorsrntextwrapdedentlstrip)r.rnmsgr r r rGs   rG)N)rZdistutils.utilrZdistutils.command.build_pyZcommandrrr-r\rqrTrprWr;Zsetuptools.externrZsetuptools.extern.six.movesrrrZsetuptools.lib2to3_exr ImportErrorrcrGr r r r s$   Y site-packages/setuptools/command/__pycache__/register.cpython-38.pyc000064400000001363151464452710021623 0ustar00U Qab@s4ddlmZddlmmZGdddejZdS))logNc@seZdZejjZddZdS)registerc Cs0z|dtj|W5|dtjXdS)Nz[WARNING: Registering is deprecated, use twine to upload instead (https://pypi.org/p/twine/)Zegg_info)ZannouncerZWARNZ run_commandorigrrun)selfr?/usr/lib/python3.8/site-packages/setuptools/command/register.pyrs z register.runN)__name__ __module__ __qualname__rr__doc__rrrrrrsr)Z distutilsrZdistutils.command.registerZcommandrrrrrrs site-packages/setuptools/command/__pycache__/sdist.cpython-38.pyc000064400000015227151464452710021131 0ustar00U Qab@sddlmZddlmmZddlZddlZddlZddl Z ddl m Z ddl m Z ddlZeZd ddZGd d d e ejZdS) )logN)six)sdist_add_defaultsccs,tdD]}||D] }|Vqq dS)z%Find all files under revision controlzsetuptools.file_findersN) pkg_resourcesZiter_entry_pointsload)dirnameZepitemr )szsdist.cCs|d|d}|j|_|jtj|jd|| D]}||qD| t |j dg}|j D] }dd|f}||krp||qpdS)Negg_infoz SOURCES.txt dist_filesrr)Z run_commandget_finalized_commandfilelistappendospathjoinr check_readmeZget_sub_commandsmake_distributiongetattr distributionZ archive_files)selfZei_cmdZcmd_namerfiledatar r r run+s      z sdist.runcCstj||dS)N)origrinitialize_options_default_to_gztarr r r r r%>s zsdist.initialize_optionscCstjdkrdSdg|_dS)N)rZbetarZgztar)sys version_infoZformatsr'r r r r&Cs zsdist._default_to_gztarc Cs$|tj|W5QRXdS)z% Workaround for #516 N)_remove_os_linkr$rrr'r r r rIs zsdist.make_distributionc cs^Gddd}ttd|}zt`Wntk r6YnXz dVW5||k rXttd|XdS)zG In a context, remove and restore os.link if it exists c@s eZdZdS)z&sdist._remove_os_link..NoValueN)__name__ __module__ __qualname__r r r r NoValueWsr0linkN)rrr1 Exceptionsetattr)r0Zorig_valr r r r,Ps  zsdist._remove_os_linkcCsLztj|Wn6tk rFt\}}}|jjjd YnXdS)Ntemplate) r$r read_templater2r*exc_infotb_nexttb_framef_localsclose)r _tbr r r Z__read_template_hackes zsdist.__read_template_hack)r=)r(r)r(r)r(r=)r(r=rcs^|jrZ|d}|j||jjsZ|jD]&\}}}|jfdd|Dq2dS)zgetting python filesbuild_pycsg|]}tj|qSr )rrr)rfilenameZsrc_dirr r sz.sdist._add_defaults_python..N)rZhas_pure_modulesrrextendZget_source_filesZinclude_package_dataZ data_files)r r@r; filenamesr rBr _add_defaults_python|s  zsdist._add_defaults_pythoncsDz tjrt|n tWntk r>tdYnXdS)Nz&data_files contains unexpected objects)rZPY2r_add_defaults_data_filessuper TypeErrorrwarnr' __class__r r rGs  zsdist._add_defaults_data_filescCs8|jD]}tj|rdSq|dd|jdS)Nz,standard file not found: should have one of z, )READMESrrexistsrJr)r fr r r rs   zsdist.check_readmecCs^tj|||tj|d}ttdrJtj|rJt|| d|| d |dS)Nz setup.cfgr1r) r$rmake_release_treerrrhasattrrNunlinkZ copy_filerZsave_version_info)r Zbase_dirfilesdestr r r rPs   zsdist.make_release_treec Cs@tj|jsdSt|jd}|}W5QRX|dkS)NFrbz+# file GENERATED by distutils, do NOT edit )rrisfilemanifestioopenreadlineencode)r fpZ first_liner r r _manifest_is_not_generatedsz sdist._manifest_is_not_generatedc Cstd|jt|jd}|D]d}tjr^z|d}Wn&tk r\td|YqYnX| }| ds|svq|j |q| dS)zRead the manifest file (named by 'self.manifest') and use it to fill in 'self.filelist', the list of files to include in the source distribution. zreading manifest file '%s'rUzUTF-8z"%r not UTF-8 decodable -- skipping#N)rinforWrYrZPY3decodeUnicodeDecodeErrorrJstrip startswithrrr:)r rWliner r r read_manifests  zsdist.read_manifestcCs^|jd}|dd\}}|dkr2tddStj|sNtd|dS|j |dS)zyChecks if license_file' is configured and adds it to 'self.filelist' if the value contains a valid path. Zmetadata license_file)NNNz''license_file' option was not specifiedz8warning: Failed to find the configured license file '%s') rZget_option_dictgetrdebugrrrNrJrr)r Zoptsr;rfr r r check_licenses   zsdist.check_license)r-r.r/__doc__Z user_optionsZ negative_optZREADME_EXTENSIONStuplerMr#r%r&r staticmethod contextlibcontextmanagerr,Z_sdist__read_template_hackr*r+Zhas_leaky_handler5rFrGrrPr]reri __classcell__r r rKr rs<        r)r)Z distutilsrZdistutils.command.sdistZcommandrr$rr*rXrmZsetuptools.externrZ py36compatrrlistZ_default_revctrlr r r r r s    site-packages/setuptools/command/__pycache__/__init__.cpython-38.pyc000064400000001342151464452710021533 0ustar00U QabR@szdddddddddd d d d d dddddddddgZddlmZddlZddlmZdejkrrdejd<ejd[[dS)alias bdist_eggZ bdist_rpmZ build_extZbuild_pyZdevelopZ easy_installZegg_infoZinstallZ install_librotateZsaveoptsZsdistZsetoptZtestZinstall_egg_infoinstall_scriptsregisterZ bdist_wininstZ upload_docsZuploadZ build_clibZ dist_info)bdistN)rZegg)rzPython .egg file) __all__Zdistutils.command.bdistrsysZsetuptools.commandrZformat_commandsZformat_commandappendr r ?/usr/lib/python3.8/site-packages/setuptools/command/__init__.pys<     site-packages/setuptools/command/__pycache__/easy_install.cpython-38.pyc000064400000177507151464452710022504 0ustar00U QabU@s4dZddlmZddlmZddlmZmZddlmZmZm Z m Z ddl m Z m Z ddlmZmZddlmZdd lmZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd lZdd l Z dd l!Z!dd l"Z"dd l#Z#dd l$Z$dd l%Z%dd l&m'Z'm(Z(dd l)m*Z*dd l+m,Z,ddl-m.Z.m/Z/ddl)m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6ddl7m8Z8ddl9m:Z:m;Z;mZ>ddl?m@Z@ddlAmBZBmCZCmDZDmEZEmFZFmGZGmHZHmIZImJZJmKZKmLZLmMZMmNZNmOZOmPZPdd lQZAeRZSejTdeAjUdddddddgZVd d!ZWd"dZXe,jYrBd#d$ZZd%d&Z[nd'd$ZZd(d&Z[d)d*Z\Gd+dde0Z]d,d-Z^d.d/Z_d0d1Z`d2dZad3dZbGd4ddeHZcGd5d6d6ecZdejefd7d8d9kredZcd:d;Zgdd?Zid@dAZjdsdBdCZkdDdEZldFdGZmdHejnkremZondIdJZodtdLdMZpdNdOZqdPdQZrdRdSZszddTlmtZuWnevk rndUdVZuYnXdWdXZtGdYdZdZewZxexyZzGd[d\d\exZ{Gd]d^d^Z|Gd_d`d`e|Z}Gdadbdbe}Z~e|jZe|jZdcddZdedfZdgehfdhdiZdjdkZdldmZdudndZe"jdodpZGdqdrdre*Zd S)va% Easy Install ------------ A tool for doing automatic download/extract/build of distutils-based Python packages. For detailed documentation, see the accompanying EasyInstall.txt file, or visit the `EasyInstall home page`__. __ https://setuptools.readthedocs.io/en/latest/easy_install.html )glob) get_platform) convert_path subst_vars)DistutilsArgErrorDistutilsOptionErrorDistutilsErrorDistutilsPlatformError)INSTALL_SCHEMES SCHEME_KEYS)logdir_util) first_line_re)find_executableN)get_config_varsget_path)SetuptoolsDeprecationWarning)six) configparsermap)Command) run_setup) rmtree_safe)setopt)unpack_archive) PackageIndexparse_requirement_arg URL_SCHEME) bdist_eggegg_info)Wheel) yield_linesnormalize_pathresource_stringensure_directoryget_distributionfind_distributions Environment Requirement Distribution PathMetadata EggMetadata WorkingSetDistributionNotFoundVersionConflict DEVELOP_DISTdefault)categorysamefile easy_installPthDistributionsextract_wininst_cfgmainget_exe_prefixescCstddkS)NP)structcalcsizer<r<C/usr/lib/python3.8/site-packages/setuptools/command/easy_install.pyis_64bitOsr>cCsjtj|otj|}ttjdo&|}|r:tj||Stjtj|}tjtj|}||kS)z Determine if two paths reference the same file. Augments os.path.samefile to work on Windows and suppresses errors if the path doesn't exist. r2)ospathexistshasattrr2normpathnormcase)Zp1Zp2Z both_existZ use_samefileZnorm_p1Znorm_p2r<r<r=r2SscCs|SNr<sr<r<r= _to_bytesesrHcCs.zt|dWdStk r(YdSXdSNasciiTF)rZ text_type UnicodeErrorrFr<r<r=isasciihs  rLcCs |dS)Nutf8)encoderFr<r<r=rHpscCs,z|dWdStk r&YdSXdSrI)rNrKrFr<r<r=rLss  cCst|ddS)N z; )textwrapdedentstripreplace)textr<r<r={rUc@seZdZdZdZdZdddddd d d d d dddddddddddgZdddddddd d!g Zej rd"ej Z e d#d$e fe d#d%diZ eZd&d'Zd(d)Zd*d+Zed,d-Zd.d/Zd0d1Zd2d3Zd4d5Zd6d7Zd8d9Zd:d;Zdd?Zed@ Z!edA Z"edB Z#dCdDZ$dEdFZ%dGdHZ&dIdJZ'dKdLZ(dMdNZ)e*j+dOdPZ,ddRdSZ-ddTdUZ.dVdWZ/ddXdYZ0dZd[Z1d\d]Z2d^d_Z3dd`daZ4edbdcZ5ddfdgZ6dhdiZ7djdkZ8dldmZ9dndoZ:dpdqZ;drdsZddwdxZ?edy Z@dzd{ZAd|d}ZBd~dZCddZDddZEddZFddZGddZHed ZIddZJddZKddZLeMeMddddZNeMdddZOddZPd$S)r3z'Manage a download/build/install processz Find/get/install Python packagesT)zprefix=Nzinstallation prefix)zip-okzzinstall package as a zipfile) multi-versionmz%make apps have to require() a version)upgradeUz1force upgrade (searches PyPI for latest versions))z install-dir=dzinstall package to DIR)z script-dir=rGzinstall scripts to DIR)exclude-scriptsxzDon't install scripts) always-copyaz'Copy all needed packages to install dir)z index-url=iz base URL of Python Package Index)z find-links=fz(additional URL(s) to search for packages)zbuild-directory=bz/download/extract/build in DIR; keep the results)z optimize=Ozlalso compile with optimization: -O1 for "python -O", -O2 for "python -OO", and -O0 to disable [default: -O0])zrecord=Nz3filename in which to record list of installed files) always-unzipZz*don't install as a zipfile, no matter what)z site-dirs=Sz)list of directories where .pth files work)editableez+Install specified packages in editable form)no-depsNzdon't install dependencies)z allow-hosts=Hz$pattern(s) that hostnames must match)local-snapshots-oklz(allow building eggs from local checkouts)versionNz"print version information and exit)z no-find-linksNz9Don't load find-links defined in packages being installedrWrYr^r[r`rirkrnrpz!install in user site-package '%s'userNrfcCs,d|_d|_|_d|_|_|_d|_d|_d|_d|_ d|_ |_ d|_ |_ |_d|_|_|_d|_|_|_d|_d|_d|_d|_d|_d|_d|_d|_d|_tjrtj |_!tj"|_#n d|_!d|_#d|_$d|_%d|_&|_'d|_(i|_)d|_*d|_+|j,j-|_-|j,.||j,/ddS)NrFr3)0rqzip_oklocal_snapshots_ok install_dir script_direxclude_scripts index_url find_linksbuild_directoryargsoptimizerecordr[ always_copy multi_versionrino_deps allow_hostsrootprefix no_reportrpinstall_purelibinstall_platlibinstall_headers install_libinstall_scripts install_data install_baseinstall_platbasesiteENABLE_USER_SITE USER_BASEinstall_userbase USER_SITEinstall_usersite no_find_links package_indexpth_filealways_copy_from site_dirsinstalled_projectssitepy_installedZ_dry_run distributionverboseZ_set_command_optionsget_option_dictselfr<r<r=initialize_optionssJ      zeasy_install.initialize_optionscCs"dd|D}tt|j|dS)Ncss*|]"}tj|stj|r|VqdSrE)r?r@rAislink).0filenamer<r<r= s z/easy_install.delete_blockers..)listr _delete_path)rblockersZextant_blockersr<r<r=delete_blockersszeasy_install.delete_blockerscCsJtd||jrdStj|o.tj| }|r8tntj}||dS)Nz Deleting %s) r infodry_runr?r@isdirrrmtreeunlink)rr@Zis_treeZremoverr<r<r=rs  zeasy_install._delete_pathcCs4djtj}td}d}t|jfttdS)zT Render the Setuptools version and installation details, then exit. {}.{} setuptoolsz=setuptools {dist.version} from {dist.location} (Python {ver})N)formatsys version_infor%printlocals SystemExit)Zverdisttmplr<r<r=_render_versions  zeasy_install._render_versionc Csh|jo |tjd}tdd\}}|j|j|j||dd|d|d||||t tddd |_ t j r|j |j d <|j|j d <||||d d d d|jdkr|j|_|jdkrd|_|dd|dd|jr|jr|j|_|j|_|ddtttj}t|_|jdk rdd|jdD}|D]N}t j!|s|t"#d|n,t||krt$|dn|j%t|q\|j&s|'|j(pd|_(|jdd|_)|jt|jfD] }||j)kr|j)*d|q|j+dk r0dd|j+dD}ndg}|j,dkrX|j-|j(|j)|d|_,t.|j)tj|_/|j0dk rt1|j0t2j3r|j0|_0ng|_0|j4r|j,5|j)tj|js|j,6|j0|ddt1|j7t8s6z0t8|j7|_7d|j7krdksnt9Wnt9k r4t$d YnX|j&rN|j:sNt;d!|j<s^t;d"g|_=dS)#Nrr exec_prefixabiflags) Z dist_nameZ dist_versionZ dist_fullname py_versionZpy_version_shortZpy_version_nodotZ sys_prefixrZsys_exec_prefixrruserbaseZusersitertruryrFr)rtrtrrtruinstall)r|r|cSsg|]}tj|qSr<)r?r@ expanduserrRrrGr<r<r= 8sz1easy_install.finalize_options..,z"%s (in --site-dirs) does not existz$ (in --site-dirs) is not on sys.pathzhttps://pypi.org/simple/cSsg|] }|qSr<)rRrr<r<r=rNs*)Z search_pathhosts)r{r{z--optimize must be 0, 1, or 2z9Must specify a build directory (-b) when using --editablez:No urls, filenames, or requirements specified (see --help))>rprrsplitrrZget_nameZ get_versionZ get_fullnamegetattr config_varsrrrr_fix_install_dir_for_user_siteexpand_basedirs expand_dirs_expandrurtrZset_undefined_optionsrqrrrr"r@ get_site_dirs all_site_dirsrr?rr warnrappendricheck_site_dirrw shadow_pathinsertrr create_indexr' local_indexrx isinstancerZ string_typesrsZscan_egg_linksadd_find_linksr{int ValueErrorryrrzoutputs) rrrrrCrr]Z path_itemrr<r<r=finalize_optionss                zeasy_install.finalize_optionscCs\|jr tjsdS||jdkr.d}t||j|_|_tj ddd}| |dS)z; Fix the install_dir if "--user" was used. Nz$User base directory is not specifiedposixZunixZ_user) rqrrcreate_home_pathrr rrr?namerS select_scheme)rmsgZ scheme_namer<r<r=rss  z+easy_install._fix_install_dir_for_user_sitecCsX|D]N}t||}|dk rtjdks.tjdkr:tj|}t||j}t|||qdS)Nrnt)rr?rr@rrrsetattr)rattrsattrvalr<r<r= _expand_attrss   zeasy_install._expand_attrscCs|dddgdS)zNCalls `os.path.expanduser` on install_base, install_platbase and root.rrrNrrr<r<r=rszeasy_install.expand_basedirscCsddddddg}||dS)z+Calls `os.path.expanduser` on install dirs.rrrrrrNr)rdirsr<r<r=rszeasy_install.expand_dirsc Cs|j|jjkrt|jz|jD]}|||j q"|jr|j}|j rzt |j }t t |D]}|||d||<q`ddl m }||j|j|fd|j|W5t|jjXdS)Nr) file_utilz'writing list of installed files to '%s')rrr set_verbosityrzr3rr|rrlenrange distutilsrexecuteZ write_filewarn_deprecated_options)rspecrZroot_lenZcounterrr<r<r=runs*     zeasy_install.runcCsDz t}Wn"tk r.tdtj}YnXtj|j d|S)zReturn a pseudo-tempname base in the install directory. This code is intentionally naive; if a malicious party can write to the target directory you're already in deep doodoo. rztest-easy-install-%s) r?getpid ExceptionrandomZrandintrmaxsizer@joinrt)rpidr<r<r=pseudo_tempnames  zeasy_install.pseudo_tempnamecCsdSrEr<rr<r<r=rsz$easy_install.warn_deprecated_optionsc CsZt|j}tj|d}tj|sTzt|Wn ttfk rR| YnX||j k}|sr|j sr| }nd| d}tj|}z*|rt|t|dt|Wn ttfk r| YnX|s|j st||r|jdkrt||j |_nd|_|tttkr.d|_n"|j rPtj|sPd|_d|_||_dS)z;Verify that self.install_dir is .pth-capable dir, if neededeasy-install.pthz .write-testwNT)r"rtr?r@rrAmakedirsOSErrorIOErrorcant_write_to_targetrr~check_pth_processingrropencloserno_default_version_msgrr4r _pythonpathr)rinstdirrZ is_site_dirZtestfileZ test_existsr<r<r=rs>           zeasy_install.check_site_diraS can't create or remove files in install directory The following error occurred while trying to add or remove files in the installation directory: %s The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s z This directory does not currently exist. Please create it and try again, or choose a different installation directory (using the -d or --install-dir option). a Perhaps your account does not have write access to this directory? If the installation directory is a system-owned directory, you may need to sign in as the administrator or "root" account. If you do not have administrative access to this machine, you may wish to choose a different installation directory, preferably one that is listed in your PYTHONPATH environment variable. For information on other options, you may wish to consult the documentation at: https://setuptools.readthedocs.io/en/latest/easy_install.html Please make the appropriate changes for your system and try again. cCsP|jtd|jf}tj|js6|d|j7}n|d|j7}t |dS)NrO) _easy_install__cant_write_msgrexc_infortr?r@rA_easy_install__not_exists_id_easy_install__access_msgr)rrr<r<r=rs z!easy_install.cant_write_to_targetc Cs|j}td||d}|d}tj|}tdd}z8|rNt|tj |}t j j |ddt |d}Wn ttfk r|YnXz||jft|d }tj}tjd krtj|\}} tj|d } | d kotj| } | r| }d dlm} | |dddgd tj|rNtd|WdSW5|r`|tj|rxt|tj|rt|X|jstd|dS)z@Empirically verify whether .pth files are supported in inst. dirz Checking .pth file support in %s.pthz.okzz import os f = open({ok_file!r}, 'w') f.write('OK') f.close() rOT)exist_okrNr pythonw.exe python.exer)spawnz-Ez-cpassz-TEST PASSED: %s appears to support .pth filesz+TEST FAILED: %s does NOT support .pth filesF)rtr rrr?r@rA _one_linerrdirname pkg_resourcesZ py31compatrrrrrrwriterrr executablerrrlowerdistutils.spawnrr~r) rrrZok_fileZ ok_existsrrrcrbasenameZaltZuse_altrr<r<r=rs\            z!easy_install.check_pth_processingc CsV|jsH|drH|dD],}|d|r.q||||d|q||dS)z=Write all the scripts for `dist`, unless scripts are excludedscriptszscripts/N)rvZmetadata_isdirZmetadata_listdirinstall_scriptZ get_metadatainstall_wrapper_scripts)rr script_namer<r<r=install_egg_scriptsYs z easy_install.install_egg_scriptscCsTtj|rDt|D]*\}}}|D]}|jtj||q$qn |j|dSrE)r?r@rwalkrrr)rr@baserfilesrr<r<r= add_outputgs  zeasy_install.add_outputcCs|jrtd|fdS)NzjInvalid argument %r: you can't use filenames or URLs with --editable (except via the --find-links option).)rirrrr<r<r= not_editableos zeasy_install.not_editablecCs<|js dStjtj|j|jr8td|j|jfdS)Nz2%r already exists in %s; can't do a checkout there)rir?r@rArrykeyrr!r<r<r=check_editablews zeasy_install.check_editablec cs:tjdd}zt|VW5tj|o2tt|XdS)Nz easy_install-)r)tempfilemkdtempr?r@rArrstr)rtmpdirr<r<r=_tmpdirs zeasy_install._tmpdirFc CsH|js||&}t|tst|rb|||j||}| d|||dW5QRSt j |r||| d|||dW5QRSt |}|||j|||j|j|j |j}|dkrd|}|jr|d7}t|nJ|jtkr||||d|W5QRS| ||j||W5QRSW5QRXdS)NTz+Could not find suitable distribution for %rz2 (--always-copy skips system and development eggs)Using)riinstall_site_pyr)rr(rr"rdownload install_itemr?r@rArr$Zfetch_distributionr[r}rrZ precedencer/process_distributionlocation)rrdepsr(dlrrr<r<r=r3s<        zeasy_install.easy_installcCs |p|j}|ptj||k}|p,|d }|pT|jdk oTtjt|t|jk}|r|s|j|jD]}|j |krjqqjd}t dtj ||r| |||}|D]}||||qn ||g}|||d|d|dk r|D]}||kr|SqdS)N.eggTz Processing %srr*)r}r?r@rendswithrr"r project_namer/r rr install_eggsr.egg_distribution)rrr,r(r0Zinstall_neededrZdistsr<r<r=r-s2     zeasy_install.install_itemcCs<t|}tD]*}d|}t||dkr t||||q dS)z=Sets the install directories by applying the install schemes.Zinstall_N)r r rr)rrschemer#attrnamer<r<r=rs zeasy_install.select_schemec Gs|||j|||j|jkr2|j||j|||||j|j<t |j ||f|| dr|j s|j |d|s|jsdS|dk r|j|jkrtd|dS|dks||kr|}tt|}t d|ztg|g|j|j}Wn^tk r<}ztt|W5d}~XYn0tk rj}zt|W5d}~XYnX|js||jr|D]"}|j|jkr||qt d|dS)Nzdependency_links.txtzSkipping dependencies for %szProcessing dependencies for %sz'Finished processing dependencies for %s) update_pthraddrr#removerrr rinstallation_report has_metadatarrZget_metadata_linesr}ras_requirementr(r'r,Zresolver3r-rr.Zreportr)rZ requirementrr0rZdistreqZdistrosrjr<r<r=r.sL           z!easy_install.process_distributioncCs2|jdk r|j S|dr dS|ds.dSdS)Nz not-zip-safeTzzip-safeF)rrr=rrr<r<r= should_unzips   zeasy_install.should_unzipcCstj|j|j}tj|r:d}t||j|j||Stj|rL|}nRtj ||krft |t |}t |dkrtj||d}tj|r|}t |t|||S)Nz<%r already exists in %s; build directory %s will not be keptrr)r?r@rryr#rAr rrrrlistdirrr$shutilmove)rr dist_filename setup_basedstrcontentsr<r<r= maybe_move s$       zeasy_install.maybe_movecCs,|jr dSt|D]}|j|qdSrE)rv ScriptWriterbestget_args write_script)rrrzr<r<r=r"sz$easy_install.install_wrapper_scriptscCsNt|}t||}|r8||t}t||}||t|ddS)z/Generate a legacy script wrapper and install itrdN) r'r>is_python_script_load_templaterrI get_headerrLrH)rrr script_textdev_pathrZ is_scriptZbodyr<r<r=r(s   zeasy_install.install_scriptcCs(d}|r|dd}td|}|dS)z There are a couple of template scripts in the package. This function loads one of them and prepares it for use. z script.tmplz.tmplz (dev).tmplrutf-8)rSr#decode)rQrZ raw_bytesr<r<r=rN2s   zeasy_install._load_templatetr<c sfdd|Dtd|jtjj|}|jrLdSt }t |tj |rpt |t |d|}||W5QRXt|d|dS)z1Write an executable file to the scripts directorycsg|]}tjj|qSr<)r?r@rrurr_rr<r=rDsz-easy_install.write_script..zInstalling %s script to %sNri)rr rrur?r@rr r current_umaskr$rArrrchmod)rrrGmodertargetmaskrcr<rr=rLAs   zeasy_install.write_scriptcCs^|dr|||gS|dr8|||gS|drT|||gS|}tj|r~|ds~t|||j ntj |rtj |}| |r|j r|dk r||||}tj|d}tj|s0ttj|dd}|stdtj |t|dkr(td tj ||d }|jrNt|||gS|||SdS) Nr2.exez.whl.pyzsetup.pyrz"Couldn't find a setup script in %srzMultiple setup scripts in %sr)rr3 install_egg install_exe install_wheelr?r@isfilerunpack_progressrabspath startswithryrHrrArrrrir rreport_editablebuild_and_install)rrrDr(rE setup_scriptZsetupsr<r<r=r5UsJ     zeasy_install.install_eggscCs>tj|r"t|tj|d}ntt|}tj ||dS)NEGG-INFO)metadata) r?r@rr*rr+ zipimport zipimporterr)Z from_filename)regg_pathrhr<r<r=r6s   zeasy_install.egg_distributionc Cstj|jtj|}tj|}|js2t|||}t ||sztj |rrtj |srt j ||jdn"tj|r|tj|fd|zd}tj |r||rtjd}}n tjd}}nL||r|||jd}}n*d}||r tjd}}n tjd}}||||f|dtj|tj|ft||d Wn$tk rxt|dd YnX||||S) Nr Removing FZMovingZCopyingZ ExtractingTz %s to %sfix_zipimporter_caches)r?r@rrtrrbrr$r6r2rrr remove_treerArrrcrBrCZcopytreer@Zmkpathunpack_and_compileZcopy2rupdate_dist_cachesrr )rrkr( destinationrZnew_dist_is_zippedrcrZr<r<r=r]s^                zeasy_install.install_eggc sPt|}|dkrtd|td|dd|ddtd}tj||d}||_ |d}tj|d}tj|d }t |t |||_ | ||tj|st|d } | d |dD].\} } | d kr| d | dd| fq| tj|d|fddt|Dtj|||j|jd|||S)Nz(%s is not a valid distutils Windows .exerhrrp)r4rpplatformr2z.tmprgPKG-INFOrzMetadata-Version: 1.0 target_versionz%s: %s _-rcsg|]}tj|dqS)r)r?r@r)rrzrur<r=rsz,easy_install.install_exe..)rr)r5rr)getrr?r@regg_namer/r$r*Z _provider exe_to_eggrArritemsrStitlerrrIrKrZ make_zipfilerrr]) rrDr(cfgrrkegg_tmpZ _egg_infoZpkg_infrckvr<ryr=r^sJ       zeasy_install.install_exec s6t|ggifdd}t||g}D]l}|dr<|d}|d}t|dd|d<tjj f|} || |t ||q<| t tj dt|dD]Z} t| rtj d| d } tj| st| d } | d t| d | qd S) z;Extract a bdist_wininst to the directories an egg would usecs|}D]\}}||r ||t|d}|d}tjjf|}|}|dsj|drt |d|d<dtj |dd< |n4|dr|dkrdtj |dd< ||Sq |d st d |dS) N/.pyd.dllrrr\SCRIPTS/r zWARNING: can't process %s)rrcrrr?r@rr3r strip_modulesplitextrr r)srcrFrGoldnewpartsr1r native_libsprefixes to_compile top_levelr<r=processs$        z(easy_install.exe_to_egg..processrrrr\rg)rrz.txtrrON)r7rrr3rrrr?r@rrZ write_stub byte_compileZwrite_safety_flagZ analyze_eggrrArrr) rrDrrZstubsresrresourceZpyfilerZtxtrcr<rr=r|s8          zeasy_install.exe_to_eggc Cst|}|sttj|j|}tj|}|j sBt |tj |rltj |slt j||j dn"tj|r|tj|fd|z.||j|fdtj|tj|fW5t|ddX||||S)NrlrmFrnzInstalling %s to %s)r Z is_compatibleAssertionErrorr?r@rrtr{rbrr$rrr rprArrrrZinstall_as_eggrrr r6)rZ wheel_pathr(Zwheelrsr<r<r=r_!s4       zeasy_install.install_wheela( Because this distribution was installed --multi-version, before you can import modules from this package in an application, you will need to 'import pkg_resources' and then use a 'require()' call similar to one of these examples, in order to select the desired version: pkg_resources.require("%(name)s") # latest installed version pkg_resources.require("%(name)s==%(version)s") # this exact version pkg_resources.require("%(name)s>=%(version)s") # this version or higher z Note also that the installation directory must be on sys.path at runtime for this to work. (e.g. by being the application's script directory, by being on PYTHONPATH, or by being added to sys.path by your code.) Installedc Cs^d}|jr>|js>|d|j7}|jtttjkr>|d|j7}|j }|j }|j }d}|t S)z9Helpful installation message for display to package usersz %(what)s %(eggloc)s%(extras)srOr) r~r_easy_install__mv_warningrtrr"rr@_easy_install__id_warningr/r4rpr) rZreqrZwhatrZegglocrrpZextrasr<r<r=r<Os z easy_install.installation_reportaR Extracted editable version of %(spec)s to %(dirname)s If it uses setuptools in its setup script, you can activate it in "development" mode by going to that directory and running:: %(python)s setup.py develop See the setuptools documentation for the "develop" command for more info. cCs"tj|}tj}d|jtS)NrO)r?r@rrr_easy_install__editable_msgr)rrrfrpythonr<r<r=rdhs zeasy_install.report_editablec Cstjdttjdtt|}|jdkrNd|jd}|dd|n|jdkrd|dd|jrv|dd t d |t |ddd |zt ||Wn6tk r}ztd |jdfW5d}~XYnXdS) Nzdistutils.command.bdist_eggzdistutils.command.egg_inforrrrrx-qz-nz Running %s %s zSetup script exited with %s)rmodules setdefaultrrrrrrr rrrrrrrz)rrfrErzrr<r<r=rms&    zeasy_install.run_setupc Csddg}tjdtj|d}z| tj|| || |||t |g}g}|D]&}||D]}| | |j|qhq\|s|jstd||WSt|t|jXdS)Nrz --dist-dirz egg-dist-tmp-)rdirz+No eggs found in %s (setup script problem?))r%r&r?r@rrr rr_set_fetcher_optionsrrr'r]r/rr) rrfrErzZdist_dirZall_eggsZeggsr#rr<r<r=res*     zeasy_install.build_and_installc Csp|jd}d}i}|D]&\}}||kr2q |d||dd<q t|d}tj|d}t ||dS) a When easy_install is about to run bdist_egg on a source dist, that source dist might have 'setup_requires' directives, requiring additional fetching. Ensure the fetcher options given to easy_install are available to that command as well. r3)rxrrwr{rrrwrx)r3z setup.cfgN) rrcopyr}rSdictr?r@rrZ edit_config) rrZei_optsZfetch_directivesZ fetch_optionsr#rZsettingsZ cfg_filenamer<r<r=rs  z!easy_install._set_fetcher_optionscCs*|jdkrdS|j|jD]H}|js0|j|jkrtd||j||j|jkr|j|jq|js|j|jjkrtd|n2td||j ||j|jkr|j |j|j s&|j |jdkr&t j|jd}t j|rt |t|d}||j|jd|dS)Nz&Removing %s from easy-install.pth filez4%s is already the active version in easy-install.pthz"Adding %s to easy-install.pth filersetuptools.pthwtrO)rr#r~r/r rr;rpathsr:rrsaver?r@rrtrrrr make_relativer)rrr]rrcr<r<r=r9s6            zeasy_install.update_pthcCstd|||S)NzUnpacking %s to %s)r debug)rrrFr<r<r=raszeasy_install.unpack_progresscsdggfdd}t|||js`D]&}t|tjdBd@}t||q8dS)NcsZ|dr |ds |n|ds4|dr>|||j rV|pXdS)Nr\ EGG-INFO/rz.so)r3rcrrar)rrFrZto_chmodrr<r=pfs    z+easy_install.unpack_and_compile..pfimi)rrrr?statST_MODErW)rrkrsrrcrXr<rr=rqs  zeasy_install.unpack_and_compilec Csjtjr dSddlm}z@t|jd||dd|jd|jrT|||jd|jdW5t|jXdS)Nr)rr)r{forcer) rdont_write_bytecodedistutils.utilrr rrrr{)rrrr<r<r=rs  zeasy_install.byte_compilea bad install directory or PYTHONPATH You are attempting to install a package to a directory that is not on PYTHONPATH and which Python does not read ".pth" files from. The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s and your PYTHONPATH environment variable currently contains: %r Here are some of your options for correcting the problem: * You can choose a different installation directory, i.e., one that is on PYTHONPATH or supports .pth files * You can add the installation directory to the PYTHONPATH environment variable. (It must then also be on PYTHONPATH whenever you run Python and want to use the package(s) you are installing.) * You can set up the installation directory to support ".pth" files by using one of the approaches described here: https://setuptools.readthedocs.io/en/latest/easy_install.html#custom-installation-locations Please make the appropriate changes for your system and try again.cCs|j}||jtjddfSNZ PYTHONPATHr)_easy_install__no_default_msgrtr?environrz)rtemplater<r<r=rsz#easy_install.no_default_version_msgc Cs|jr dStj|jd}tdd}|d}d}tj|rt d|jt |}| }W5QRX| dstd |||krtd ||jst|t j |d dd }||W5QRX||gd |_dS)z8Make sure there's a site.py in the target dir, if neededNzsite.pyrz site-patch.pyrRrzChecking existing site.py in %sz def __boot():z;%s is not a setuptools-generated site.py; please remove it.z Creating %srencodingT)rr?r@rrtr#rSrAr riorreadrcrrrr$rr)rZsitepysourceZcurrentZstrmr<r<r=r+#s0       zeasy_install.install_site_pycCsd|js dSttjd}t|jD]8\}}||r&tj |s&| d|t |dq&dS)zCreate directories under ~.N~zos.makedirs('%s', 0o700)i) rqrr?r@rrZ iteritemsrrcrZ debug_printr)rhomerr@r<r<r=rCszeasy_install.create_home_pathz/$base/lib/python$py_version_short/site-packagesz $base/binrrz$base/Lib/site-packagesz $base/ScriptscGs|dj}|jrd|}|j|d<|jtj|j}| D]$\}}t ||ddkr>t |||q>ddl m }|D]B}t ||}|dk rt|||}tjdkrtj|}t |||qtdS)Nrrr)rr)Zget_finalized_commandrrrr rzr?rDEFAULT_SCHEMEr}rrrrr@r)rrrr7rrrr<r<r=rYs        zeasy_install._expand)F)F)T)N)rTr<)r)Q__name__ __module__ __qualname____doc__ descriptionZcommand_consumes_argumentsZ user_optionsZboolean_optionsrrrZhelp_msgrZ negative_optrrrrr staticmethodrrrrrrrrrrrPrQlstriprrr rrrr r"r$ contextlibcontextmanagerr)r3r-rr.r@rHrrrNrLr5r6r]r^r|r_rrr<rrdrrerr9rarqrrrr+rrr rrr<r<r<r=r3~s  0  z   0 ;   $ $ '    ,6-5   %  cCs tjddtj}td|Sr)r?rrzrpathsepfilter)r}r<r<r=rpsrc Cs~g}|ttjg}tjtjkr0|tj|D]}|r4tjdkr^|tj |ddnVtj dkr|tj |ddj tj dtj |ddgn||tj |ddgtjdkr4d |kr4tj d }|r4tj |d d d j tj d}||q4tdtdf}|D]}||kr||qtjrB|tjz|tWntk rjYnXttt|}|S)z& Return a list of 'site' dirs )Zos2emxZriscosZLibz site-packagesrlibz python{}.{}z site-pythondarwinzPython.frameworkHOMELibraryPythonrZpurelibZplatlib)extendrrrrrrtr?r@rseprrrrzrrrrgetsitepackagesAttributeErrorrrr")sitedirsrrrZhome_spZ lib_pathsZsite_libr<r<r=rus^             rccsi}|D]}t|}||krqd||<tj|s4qt|}||fV|D]}|ds\qL|dkrfqLttj||}tt |}| |D]H}| dst| }||krd||<tj|sq|t|fVqqLqdS)zBYield sys.path directories that might contain "old-style" packagesrr )rrimportN) r"r?r@rrAr3rrrr!rrcrstrip)Zinputsseenrrrrclinesliner<r<r= expand_pathss4        rc Cs@t|d}z$t|}|dkr*W dS|d|d|d}|dkrRWdS||dtd|d\}}}|dkrWdS||d|d d d }t |}z<||} | d d d } | t } |t| Wntjk rYW.dSX|dr"|ds*W dS|WS|XdS)znExtract configuration data from a bdist_wininst .exe Returns a configparser.RawConfigParser, or None rbN  zegg path translations for a given .exe file)zPURELIB/r)zPLATLIB/pywin32_system32r)zPLATLIB/r)rzEGG-INFO/scripts/)zDATA/lib/site-packagesrrrrrurz .egg-inforNrr z -nspkg.pth)ZPURELIBZPLATLIB\rz%s/%s/rcSsg|]\}}||fqSr<)r)rr_yr<r<r=r)sz$get_exe_prefixes..)rZZipFilerZinfolistrrrr3rrupperrrZPY3rSr!rRrSrcrsortreverse)Z exe_filenamerrXrrrrGZpthr<r<r=r7s@       " c@sReZdZdZdZdddZddZdd Zed d Z d d Z ddZ ddZ dS)r4z)A .pth file with Distribution paths in itFr<cCsl||_ttt||_ttj|j|_| t |gddt |j D]}tt|jt|dqLdS)NT)rrrr"rr?r@rbasedir_loadr'__init__r!rr:r&)rrrr@r<r<r=r4szPthDistributions.__init__cCsg|_d}t|j}tj|jrt|jd}|D]}| drHd}q4| }|j || r4| drtq4t tj|j|}|jd<tj|r||kr|jd|_q4d||<q4||jr|sd|_|jr|jd s|jqdS)NFZrtrT#rr)rrfromkeysrr?r@r`rrrcrrrRr"rrrApopdirtyr)rZ saw_importrrcrr@r<r<r=r=s4       zPthDistributions._loadc Cs|js dStt|j|j}|rtd|j||}d |d}t j |jr`t |jt|jd}||W5QRXn(t j |jrtd|jt |jd|_dS)z$Write changed .pth file back to diskNz Saving %srOrzDeleting empty %sF)rrrrrr rr _wrap_linesrr?r@rrrrrA)rZ rel_pathsrdatarcr<r<r=r\s   zPthDistributions.savecCs|SrEr<)rr<r<r=rrszPthDistributions._wrap_linescCsN|j|jko$|j|jkp$|jtk}|r>|j|jd|_t||dS)z"Add `dist` to the distribution mapTN) r/rrr?getcwdrrr'r:)rrnew_pathr<r<r=r:vs   zPthDistributions.addcCs2|j|jkr"|j|jd|_qt||dS)z'Remove `dist` from the distribution mapTN)r/rr;rr'r?r<r<r=r;s zPthDistributions.removecCstjt|\}}t|j}|g}tjdkr2dp6tj}t||kr||jkrl|tj | | |Stj|\}}||q8|S)Nr) r?r@rr"rraltseprrcurdirrr)rr@ZnpathZlastZbaselenrrr<r<r=rs      zPthDistributions.make_relativeN)r<) rrrrrrrrrrr:r;rr<r<r<r=r4/s  c@s(eZdZeddZedZedZdS)RewritePthDistributionsccs$|jV|D] }|Vq |jVdSrE)preludepostlude)clsrrr<r<r=rsz#RewritePthDistributions._wrap_linesz? import sys sys.__plen = len(sys.path) z import sys new = sys.path[sys.__plen:] del sys.path[sys.__plen:] p = getattr(sys, '__egginsert', 0) sys.path[p:p] = new sys.__egginsert = p + len(new) N)rrr classmethodrrrrr<r<r<r=rs rZSETUPTOOLS_SYS_PATH_TECHNIQUErawZrewritecCs ttjtrtSttjS)z_ Return a regular expression based on first_line_re suitable for matching strings. )rrpatternr'recompilerSr<r<r<r=_first_line_res rcCsd|tjtjfkr.tjdkr.t|tj||St\}}}t ||d|dd||ffdS)Nrrrz %s %s) r?rr;rrWrS_IWRITErrrZreraise)funcargexcZetZevrwr<r<r= auto_chmods  r cCs.t|}t|tj|r"t|nt|dS)aa Fix any globally cached `dist_path` related data `dist_path` should be a path of a newly installed egg distribution (zipped or unzipped). sys.path_importer_cache contains finder objects that have been cached when importing data from the original distribution. Any such finders need to be cleared since the replacement distribution might be packaged differently, e.g. a zipped egg distribution might get replaced with an unzipped egg folder or vice versa. Having the old finders cached may then cause Python to attempt loading modules from the replacement distribution using an incorrect loader. zipimport.zipimporter objects are Python loaders charged with importing data packaged inside zip archives. If stale loaders referencing the original distribution, are left behind, they can fail to load modules from the replacement distribution. E.g. if an old zipimport.zipimporter instance is used to load data from a new zipped egg archive, it may cause the operation to attempt to locate the requested data in the wrong location - one indicated by the original distribution's zip archive directory information. Such an operation may then fail outright, e.g. report having read a 'bad local file header', or even worse, it may fail silently & return invalid data. zipimport._zip_directory_cache contains cached zip archive directory information for all existing zipimport.zipimporter instances and all such instances connected to the same archive share the same cached directory information. If asked, and the underlying Python implementation allows it, we can fix all existing zipimport.zipimporter instances instead of having to track them down and remove them one by one, by updating their shared cached zip archive directory information. This, of course, assumes that the replacement distribution is packaged as a zipped egg. If not asked to fix existing zipimport.zipimporter instances, we still do our best to clear any remaining zipimport.zipimporter related cached data that might somehow later get used when attempting to load data from the new distribution and thus cause such load operations to fail. Note that when tracking down such remaining stale data, we can not catch every conceivable usage from here, and we clear only those that we know of and have found to cause problems if left alive. Any remaining caches should be updated by whomever is in charge of maintaining them, i.e. they should be ready to handle us replacing their zip archives with new distributions at runtime. N)r"_uncacherpath_importer_cache!_replace_zip_directory_cache_data*_remove_and_clear_zip_directory_cache_data)Z dist_pathronormalized_pathr<r<r=rrs <  rrcCsPg}t|}|D]:}t|}||r|||dtjdfkr||q|S)ap Return zipimporter cache entry keys related to a given normalized path. Alternative path spellings (e.g. those using different character case or those using alternative path separators) related to the same path are included. Any sub-path entries are included as well, i.e. those corresponding to zip archives embedded in other zip archives. rr)rr"rcr?rr)rcacheresult prefix_lenpZnpr<r<r="_collect_zipimporter_cache_entriess   rcCs@t||D]0}||}||=|o(|||}|dk r |||<q dS)a Update zipimporter cache data for a given normalized path. Any sub-path entries are processed as well, i.e. those corresponding to zip archives embedded in other zip archives. Given updater is a callable taking a cache entry key and the original entry (after already removing the entry from the cache), and expected to update the entry and possibly return a new one to be inserted in its place. Returning None indicates that the entry should not be replaced with a new one. If no updater is given, the cache entries are simply removed without any additional processing, the same as if the updater simply returned None. N)r)rrupdaterr old_entryZ new_entryr<r<r=_update_zipimporter_cache/s  rcCst||dSrE)r)rrr<r<r=r Osr cCsdd}t|tj|ddS)NcSs |dSrE)clearr@rr<r<r=2clear_and_remove_cached_zip_archive_directory_dataTszf_remove_and_clear_zip_directory_cache_data..clear_and_remove_cached_zip_archive_directory_datarrri_zip_directory_cache)rrr<r<r=r Ss r Z__pypy__cCsdd}t|tj|ddS)NcSs&|t||tj||SrE)rrirjupdaterrr<r<r=)replace_cached_zip_archive_directory_datajs zT_replace_zip_directory_cache_data..replace_cached_zip_archive_directory_datarr)rrr<r<r=r is  r c Cs4zt||dWnttfk r*YdSXdSdS)z%Is this string a valid Python script?execFTN)r SyntaxError TypeError)rTrr<r<r= is_python|s r#c CsNz(tj|dd}|d}W5QRXWnttfk rD|YSX|dkS)zCDetermine if the specified executable is a .sh (contains a #! line)zlatin-1rr#!)rrrrr)rfpmagicr<r<r=is_shs  r'cCs t|gS)z@Quote a command line argument according to Windows parsing rules subprocess list2cmdline)rr<r<r= nt_quote_argsr+cCsH|ds|drdSt||r&dS|drDd|dkSdS)zMIs this text, as a whole, a Python script? (as opposed to shell/bat/etc. r\.pywTr$rrF)r3r#rc splitlinesr)rPrr<r<r=rMs  rM)rWcGsdSrEr<)rzr<r<r=_chmodsr.c CsRtd||zt||Wn0tjk rL}ztd|W5d}~XYnXdS)Nzchanging mode of %s to %ozchmod failed: %s)r rr.r?error)r@rXrjr<r<r=rWs rWc@seZdZdZgZeZeddZeddZ eddZ edd Z ed d Z d d Z eddZddZeddZeddZdS) CommandSpeczm A command spec for a #! header, specified as a list of arguments akin to those passed to Popen. cCs|S)zV Choose the best CommandSpec class based on environmental conditions. r<rr<r<r=rJszCommandSpec.bestcCstjtj}tjd|S)N__PYVENV_LAUNCHER__)r?r@rCrrrrz)rZ_defaultr<r<r=_sys_executableszCommandSpec._sys_executablecCs:t||r|St|tr ||S|dkr0|S||S)zg Construct a CommandSpec from a parameter to build_scripts, which may be None. N)rrfrom_environment from_string)rZparamr<r<r= from_params  zCommandSpec.from_paramcCs||gSrE)r3r1r<r<r=r4szCommandSpec.from_environmentcCstj|f|j}||S)z} Construct a command spec from a simple string representing a command line parseable by shlex.split. )shlexr split_args)rstringr}r<r<r=r5szCommandSpec.from_stringcCs8t|||_t|}t|s4dg|jdd<dS)Nz-xr)r7r_extract_optionsoptionsr)r*rL)rrPcmdliner<r<r=install_optionss zCommandSpec.install_optionscCs:|dd}t|}|r.|dp0dnd}|S)zH Extract any options from the first line of the script. rOrrr)r-rmatchgrouprR)Z orig_scriptfirstr>r;r<r<r=r:s zCommandSpec._extract_optionscCs||t|jSrE)_renderrr;rr<r<r= as_headerszCommandSpec.as_headercCs6d}|D](}||r||r|ddSq|S)Nz"'rr)rcr3)itemZ_QUOTESqr<r<r= _strip_quotess zCommandSpec._strip_quotescCs tdd|D}d|dS)Ncss|]}t|VqdSrE)r0rErR)rrCr<r<r=rsz&CommandSpec._render..r$rOr()r}r<r<r<r=rAs zCommandSpec._renderN)rrrrr;rr8rrJr3r6r4r5r=rr:rBrErAr<r<r<r=r0s*       r0c@seZdZeddZdS)WindowsCommandSpecFrN)rrrrr8r<r<r<r=rF srFc@seZdZdZedZeZ e dddZ e dddZ e dd d Z ed d Ze d dZe ddZe ddZe dddZdS)rIz` Encapsulates behavior around writing entry point scripts for console and gui apps. a # EASY-INSTALL-ENTRY-SCRIPT: %(spec)r,%(group)r,%(name)r __requires__ = %(spec)r import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point(%(spec)r, %(group)r, %(name)r)() ) NFcCs6tdt|rtnt}|d||}|||S)Nz Use get_argsr)warningsrEasyInstallDeprecationWarningWindowsScriptWriterrIrJget_script_headerrK)rrrwininstwriterheaderr<r<r=get_script_args$s zScriptWriter.get_script_argscCs$tjdtdd|rd}|||S)NzUse get_headerr) stacklevelr )rGrrHrO)rrPrrKr<r<r=rJ,szScriptWriter.get_script_headerc cs|dkr|}t|}dD]Z}|d}||D]>\}}|||jt}|||||} | D] } | Vqlq:q dS)z Yield write_script() argument tuples for a distribution's console_scripts and gui_scripts entry points. NZconsoleguiZ_scripts) rOr'r>Z get_entry_mapr}_ensure_safe_namerr_get_script_args) rrrMrtype_r?rZeprPrzrr<r<r=rK4s   zScriptWriter.get_argscCstd|}|rtddS)z? Prevent paths in *_scripts entry point names. z[\\/]z+Path separators not allowed in script namesN)rsearchr)rZ has_path_sepr<r<r=rRFs zScriptWriter._ensure_safe_namecCs tdt|rtS|SNzUse best)rGrrHrIrJ)rZ force_windowsr<r<r= get_writerOs zScriptWriter.get_writercCs.tjdkstjdkr&tjdkr&tS|SdS)zD Select the best ScriptWriter for this environment. win32javarN)rrtr?r_namerIrJr1r<r<r=rJUszScriptWriter.bestccs|||fVdSrEr<)rrTrrMrPr<r<r=rS_szScriptWriter._get_script_argsrcCs"|j|}|||S)z;Create a #! line, getting options (if any) from script_text)command_spec_classrJr6r=rB)rrPrcmdr<r<r=rOds zScriptWriter.get_header)NF)NF)N)rN)rrrrrPrQrrr0r[rrNrJrKrrRrWrJrSrOr<r<r<r=rIs&       rIc@sLeZdZeZeddZeddZeddZeddZ e d d Z d S) rIcCstdt|SrV)rGrrHrJr1r<r<r=rWos zWindowsScriptWriter.get_writercCs"tt|d}tjdd}||S)zC Select the best ScriptWriter suitable for Windows )rZnaturalZSETUPTOOLS_LAUNCHERr)rWindowsExecutableLauncherWriterr?rrz)rZ writer_lookupZlauncherr<r<r=rJus zWindowsScriptWriter.bestc #stddd|}|tjddkrBdjft}t|t dddd d dd g}| || ||}fd d |D}|||d|fVdS)z For Windows, add a .py extensionz.pyar,rPZPATHEXT;zK{ext} not listed in PATHEXT; scripts will not be recognized as executables.r\ -script.py.pyc.pyor[csg|] }|qSr<r<rUrr<r=rsz8WindowsScriptWriter._get_script_args..rTN) rr?rrrrrrGr UserWarningr;_adjust_header) rrTrrMrPextrrrr<rbr=rSs   z$WindowsScriptWriter._get_script_argscCsNd}d}|dkr||}}tt|tj}|j||d}||rJ|S|S)z Make sure 'pythonw' is used for gui and and 'python' is used for console (regardless of what sys.executable is). r r rQ)r9repl)rrescape IGNORECASEsub _use_header)rrTZ orig_headerrrfZ pattern_ob new_headerr<r<r=rds z"WindowsScriptWriter._adjust_headercCs$|ddd}tjdkp"t|S)z Should _adjust_header use the replaced header? On non-windows systems, always use. On Windows systems, only use the replaced header if it resolves to an executable on the system. rr"rX)rRrrtr)rkZ clean_headerr<r<r=rjs zWindowsScriptWriter._use_headerN) rrrrFr[rrWrJrSrdrrjr<r<r<r=rIls    rIc@seZdZeddZdS)r]c #s|dkrd}d}dg}nd}d}dddg}|||}fd d |D} |||d | fVd t|d fVtsd} | td fVdS)zG For Windows, add a .py extension and an .exe launcher rQz -script.pywr,Zclir_r\r`racsg|] }|qSr<r<rUrbr<r=rszDWindowsExecutableLauncherWriter._get_script_args..rTr[rdz .exe.manifestN)rdget_win_launcherr>load_launcher_manifest) rrTrrMrPZ launcher_typererZhdrrZm_namer<rbr=rSs"  z0WindowsExecutableLauncherWriter._get_script_argsN)rrrrrSr<r<r<r=r]sr]cCs2d|}tr|dd}n |dd}td|S)z Load the Windows launcher (executable) suitable for launching a script. `type` should be either 'cli' or 'gui' Returns the executable as a byte string. z%s.exe.z-64.z-32.r)r>rSr#)typeZ launcher_fnr<r<r=rms  rmcCs0ttd}tjr|tS|dtSdS)Nzlauncher manifest.xmlrR)rr#rrPY2varsrS)rZmanifestr<r<r=rns  rnFcCst|||SrE)rBr)r@ ignore_errorsonerrorr<r<r=rsrcCstd}t||S)N)r?umask)Ztmpr<r<r=rVs  rVcCs:ddl}tj|jd}|tjd<tj|tdS)Nr) rr?r@r__path__rargvrr6)rZargv0r<r<r= bootstraps   ryc sddlm}ddlmGfddd}|dkrBtjdd}t0|fddd g|tjdpfd|d |W5QRXdS) Nr)setupr)cseZdZdZfddZdS)z-main..DistributionWithoutHelpCommandsrc s(tj|f||W5QRXdSrE) _patch_usage _show_help)rrzkwr{r<r=r} sz8main..DistributionWithoutHelpCommands._show_helpN)rrrZ common_usager}r<r{r<r=DistributionWithoutHelpCommands srrrr3z-v)Z script_argsrZ distclass)rrzZsetuptools.distr)rrxr|)rxr~rzrr<r{r=r6s    c#sLddl}tdfdd}|jj}||j_z dVW5||j_XdS)Nrze usage: %(script)s [options] requirement_or_url ... or: %(script)s --help csttj|dS)N)Zscript)rr?r@r)rZUSAGEr<r= gen_usage s z_patch_usage..gen_usage)Zdistutils.corerPrQrZcorer)rrZsavedr<rr=r| s  r|c@seZdZdZdS)rHzuClass for warning about deprecations in EasyInstall in SetupTools. Not ignored by default, unlike DeprecationWarning.N)rrrrr<r<r<r=rH( srH)N)r)N)rrrrrrZdistutils.errorsrrrr Zdistutils.command.installr r rr r Zdistutils.command.build_scriptsrrrrr?rirBr%rrrrrPrGrr:rr)r7rZ sysconfigrrrrZsetuptools.externrZsetuptools.extern.six.movesrrrZsetuptools.sandboxrZsetuptools.py27compatrZsetuptools.commandrZsetuptools.archive_utilrZsetuptools.package_indexrrrrrZsetuptools.wheelr rr!r"r#r$r%r&r'r(r)r*r+r,r-r.r/Zpkg_resources.py31compatrpZ __metaclass__filterwarningsZ PEP440Warning__all__r>r2rqrHrLrr3rrrr5r7r4rrrzrr rrrrr r builtin_module_namesr r#r'r+rMrWr. ImportErrorrr0r3Zsys_executablerFrIrIr]rNrJrmrnrrVryr6rr|rHr<r<r<r=s            D {A))'l R    T^A   site-packages/setuptools/command/__pycache__/alias.cpython-38.opt-1.pyc000064400000004530151464452710022026 0ustar00U Qabz @sPddlmZddlmZddlmZmZmZddZGdddeZ dd Z d S) )DistutilsOptionError)map) edit_config option_base config_filecCs8dD]}||krt|Sq||gkr4t|S|S)z4Quote an argument for later parsing by shlex.split())"'\#)reprsplit)argcrs   4site-packages/setuptools/command/__pycache__/upload_docs.cpython-38.opt-1.pyc000064400000013752151464452710023237 0ustar00U Qab@sdZddlmZddlmZddlmZddlZddlZddl Z ddl Z ddl Z ddl Z ddl Z ddlmZddlmZmZddlmZd d lmZd d ZGd ddeZdS)zpupload_docs Implements a Distutils 'upload_docs' subcommand (upload documentation to PyPI's pythonhosted.org). )standard_b64encode)log)DistutilsOptionErrorN)six) http_clienturllib)iter_entry_points)uploadcCstjr dnd}|d|S)Nsurrogateescapestrictzutf-8)rPY3encode)serrorsrB/usr/lib/python3.8/site-packages/setuptools/command/upload_docs.py_encodesrc@seZdZdZdZdddejfddgZejZdd Zd efgZ d d Z d dZ ddZ ddZ eddZeddZddZdS) upload_docszhttps://pypi.python.org/pypi/zUpload documentation to PyPIz repository=rzurl of repository [default: %s])z show-responseNz&display full response text from server)z upload-dir=Nzdirectory to uploadcCs"|jdkrtddD]}dSdS)Nzdistutils.commands build_sphinxT) upload_dirr)selfZeprrr has_sphinx/s zupload_docs.has_sphinxrcCst|d|_d|_dS)N)r initialize_optionsr target_dir)rrrrr6s zupload_docs.initialize_optionscCst||jdkrN|r0|d}|j|_q`|d}tj |j d|_n| d|j|_d|j krtt d|d|jdS)NrbuildZdocsrzpypi.python.orgz3Upload_docs command is deprecated. Use RTD instead.zUsing upload directory %s)r finalize_optionsrrZget_finalized_commandZbuilder_target_dirrospathjoinZ build_baseZensure_dirname repositoryrwarnannounce)rrrrrrr;s        zupload_docs.finalize_optionsc Cst|d}z||jt|jD]x\}}}||jkrP|sPd}t||j|D]H}tj ||}|t |jd tjj } tj | |} | || qTq&W5|XdS)Nwz'no files found in upload directory '%s')zipfileZZipFilecloseZmkpathrrwalkrrr lenlstripsepwrite) rfilenamezip_filerootdirsfilesZtmplnameZfullrelativedestrrrcreate_zipfileKs  zupload_docs.create_zipfilec Csh|D]}||qt}|jj}tj |d|}z| || |W5t |XdS)Nz%s.zip)Zget_sub_commandsZ run_commandtempfileZmkdtemp distributionmetadataget_namerrr shutilZrmtreer4 upload_file)rZcmd_nameZtmp_dirr1r-rrrrun[s    zupload_docs.runccs|\}}d|}t|ts |g}|D]f}t|trL|d|d7}|d}nt|}|Vt|VdV|V|r$|dddkr$dVq$dS) Nz* Content-Disposition: form-data; name="%s"z; filename="%s"rr s   ) isinstancelisttupler)item sep_boundarykeyvaluestitlevaluerrr _build_partis    zupload_docs._build_partc Csnd}d|}|d}|df}tj|j|d}t||}tj|}t||} d|d} d | | fS) z= Build up the MIME payload for the POST data s3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254s --s--r>)rCz multipart/form-data; boundary=%sascii) functoolspartialrHmapitems itertoolschain from_iterabledecoder ) clsdataboundaryrCZ end_boundaryZ end_itemsZbuilderZ part_groupspartsZ body_items content_typerrr_build_multipart}s  zupload_docs._build_multipartc Cs:t|d}|}W5QRX|jj}d|tj||fd}t|j d|j }t |}t j rn|d}d|}||\}} d|j} || tjtj|j\} } } }}}| dkrt| }n | d krt| }n td | d }zZ||d | | }|d ||dtt||d|| |!|Wn>t"j#k r}z|t|tj$WYdSd}~XYnX|%}|j&dkrd|j&|j'f} || tjnb|j&dkr|(d}|dkrd|}d|} || tjnd|j&|j'f} || tj$|j)r6t*d|ddS)NrbZ doc_upload)z:actionr1content:rIzBasic zSubmitting documentation to %sZhttpZhttpszunsupported schema ZPOSTz Content-typezContent-lengthZ AuthorizationzServer response (%s): %si-ZLocationzhttps://pythonhosted.org/%s/zUpload successful. Visit %szUpload failed (%s): %szK---------------------------------------------------------------------------)+openreadr6r7r8rrbasenamerZusernameZpasswordrrr rRrXr!r#rINFOrparseZurlparserZHTTPConnectionZHTTPSConnectionAssertionErrorZconnectZ putrequestZ putheaderstrr(Z endheaderssendsocketerrorZERRORZ getresponseZstatusreasonZ getheaderZ show_responseprint)rr,frZmetarTZ credentialsZauthZbodyZctmsgZschemaZnetlocZurlZparamsZqueryZ fragmentsZconnrWerlocationrrrr:sb               zupload_docs.upload_fileN)__name__ __module__ __qualname__ZDEFAULT_REPOSITORY descriptionr Z user_optionsZboolean_optionsrZ sub_commandsrrr4r; staticmethodrH classmethodrXr:rrrrrs(   r)__doc__base64rZ distutilsrZdistutils.errorsrrrfr%r5r9rOrKZsetuptools.externrZsetuptools.extern.six.movesrrZ pkg_resourcesrr rrrrrrs       site-packages/setuptools/command/__pycache__/py36compat.cpython-38.opt-1.pyc000064400000011002151464452710022732 0ustar00U Qabz@sdddlZddlmZddlmZddlmZddlmZGdddZe ejdr`Gd ddZdS) N)glob) convert_path)sdist)filterc@s\eZdZdZddZeddZddZdd Zd d Z d d Z ddZ ddZ ddZ dS)sdist_add_defaultsz Mix-in providing forward-compatibility for functionality as found in distutils on Python 3.7. Do not edit the code in this class except to update functionality as implemented in distutils. Instead, override in the subclass. cCs<|||||||dS)a9Add all the default files to self.filelist: - README or README.txt - setup.py - test/test*.py - all pure Python modules mentioned in setup script - all files pointed by package_data (build_py) - all files defined in data_files. - all files defined as scripts. - all C sources listed as part of extensions or C libraries in the setup script (doesn't catch C headers!) Warns if (README or README.txt) or setup.py are missing; everything else is optional. N)_add_defaults_standards_add_defaults_optional_add_defaults_python_add_defaults_data_files_add_defaults_ext_add_defaults_c_libs_add_defaults_scripts)selfrA/usr/lib/python3.8/site-packages/setuptools/command/py36compat.py add_defaultsszsdist_add_defaults.add_defaultscCs:tj|sdStj|}tj|\}}|t|kS)z Case-sensitive path existence check >>> sdist_add_defaults._cs_path_exists(__file__) True >>> sdist_add_defaults._cs_path_exists(__file__.upper()) False F)ospathexistsabspathsplitlistdir)fspathrZ directoryfilenamerrr_cs_path_exists(s  z"sdist_add_defaults._cs_path_existscCs|j|jjg}|D]~}t|trj|}d}|D]"}||r,d}|j|qPq,|s|dd |q||r|j|q|d|qdS)NFTz,standard file not found: should have one of z, zstandard file '%s' not found) ZREADMES distributionZ script_name isinstancetuplerfilelistappendwarnjoin)rZ standardsfnZaltsZgot_itrrrr9s"    z*sdist_add_defaults._add_defaults_standardscCs4ddg}|D]"}ttjjt|}|j|q dS)Nz test/test*.pyz setup.cfg)rrrisfilerrextend)rZoptionalpatternfilesrrrrNsz)sdist_add_defaults._add_defaults_optionalcCs\|d}|jr$|j||jD],\}}}}|D]}|jtj ||q:q*dS)Nbuild_py) get_finalized_commandrZhas_pure_modulesrr$get_source_files data_filesrrrr!)rr'ZpkgZsrc_dirZ build_dir filenamesrrrrr Ts   z'sdist_add_defaults._add_defaults_pythoncCsz|jrv|jjD]b}t|trBt|}tj|rt|j |q|\}}|D]$}t|}tj|rN|j |qNqdS)N) rZhas_data_filesr*rstrrrrr#rr)ritemdirnamer+frrrr ds     z+sdist_add_defaults._add_defaults_data_filescCs(|jr$|d}|j|dS)N build_ext)rZhas_ext_modulesr(rr$r))rr0rrrr us  z$sdist_add_defaults._add_defaults_extcCs(|jr$|d}|j|dS)N build_clib)rZhas_c_librariesr(rr$r))rr1rrrr zs  z'sdist_add_defaults._add_defaults_c_libscCs(|jr$|d}|j|dS)N build_scripts)rZ has_scriptsr(rr$r))rr2rrrr s  z(sdist_add_defaults._add_defaults_scriptsN)__name__ __module__ __qualname____doc__r staticmethodrrrr r r r r rrrrr s rrc@s eZdZdS)rN)r3r4r5rrrrrs) rrZdistutils.utilrZdistutils.commandrZsetuptools.extern.six.movesrrhasattrrrrrs    | site-packages/setuptools/command/__pycache__/rotate.cpython-38.pyc000064400000004730151464452710021276 0ustar00U Qabt@s`ddlmZddlmZddlmZddlZddlZddlm Z ddl m Z Gddde Z dS) ) convert_path)log)DistutilsOptionErrorN)six)Commandc@s:eZdZdZdZdddgZgZddZdd Zd d Z d S) rotatezDelete older distributionsz2delete older distributions, keeping N newest files)zmatch=mzpatterns to match (required))z dist-dir=dz%directory where the distributions are)zkeep=kz(number of matching distributions to keepcCsd|_d|_d|_dS)N)matchdist_dirkeepselfr=/usr/lib/python3.8/site-packages/setuptools/command/rotate.pyinitialize_optionsszrotate.initialize_optionscCs|jdkrtd|jdkr$tdzt|j|_Wntk rPtdYnXt|jtjrxdd|jdD|_| dddS) NzQMust specify one or more (comma-separated) match patterns (e.g. '.zip' or '.egg')z$Must specify number of files to keepz--keep must be an integercSsg|]}t|qSr)rstrip).0prrr *sz+rotate.finalize_options..,Zbdist)r r ) r rr int ValueError isinstancerZ string_typessplitZset_undefined_optionsrrrrfinalize_optionss   zrotate.finalize_optionscCs|dddlm}|jD]}|jd|}|tj|j|}dd|D}| | t dt ||||jd}|D]<\}}t d||jstj|rt|qt|qqdS) NZegg_infor)glob*cSsg|]}tj||fqSr)ospathgetmtime)rfrrrr6szrotate.run..z%d file(s) matching %sz Deleting %s)Z run_commandrr Z distributionZget_namerr joinr sortreverserinfolenr Zdry_runisdirshutilZrmtreeunlink)rrpatternfilestr"rrrrun/s        z rotate.runN) __name__ __module__ __qualname____doc__ descriptionZ user_optionsZboolean_optionsrrr.rrrrr sr) Zdistutils.utilrZ distutilsrZdistutils.errorsrrr)Zsetuptools.externrZ setuptoolsrrrrrrs     site-packages/setuptools/command/__pycache__/bdist_wininst.cpython-38.pyc000064400000001672151464452710022662 0ustar00U Qab}@s(ddlmmZGdddejZdS)Nc@seZdZdddZddZdS) bdist_wininstrcCs |j||}|dkrd|_|S)zj Supplement reinitialize_command to work around http://bugs.python.org/issue20819 )Zinstall install_libN)Z distributionreinitialize_commandr)selfcommandZreinit_subcommandscmdrD/usr/lib/python3.8/site-packages/setuptools/command/bdist_wininst.pyrsz"bdist_wininst.reinitialize_commandcCs$d|_ztj|W5d|_XdS)NTF)Z _is_runningorigrrun)rrrr r szbdist_wininst.runN)r)__name__ __module__ __qualname__rr rrrr rs r)Zdistutils.command.bdist_wininstrrr rrrr ssite-packages/setuptools/command/__pycache__/build_ext.cpython-38.pyc000064400000023247151464452710021763 0ustar00U Qab2 @sddlZddlZddlZddlmZddlmZddlm Z ddl m Z m Z ddl mZddlmZddlmZdd lmZejrddlZd d eDZn dd lmZzddlmZed Wnek reZYnXe dddl mZddZ dZ!dZ"dZ#ej$dkrdZ"n>ej%dkrTzddl&Z&e'e&dZ"Z!Wnek rRYnXddZ(ddZ)GdddeZe"sej%dkrd"ddZ*nd Z#d#d!dZ*dS)$N) build_ext) copy_file) new_compiler)customize_compilerget_config_var)DistutilsError)log)Library)sixcCs g|]\}}}|tjkr|qS)impZ C_EXTENSION).0s_tpr r @/usr/lib/python3.8/site-packages/setuptools/command/build_ext.py s rEXTENSION_SUFFIXESzCython.Compiler.MainLDSHARED) _config_varsc CsZtjdkrNt}z$dtd<dtd<dtd<t|W5tt|Xnt|dS)Ndarwinz0gcc -Wl,-x -dynamiclib -undefined dynamic_lookuprz -dynamiclibZCCSHAREDz.dylibZSO)sysplatform _CONFIG_VARScopyclearupdater)compilerZtmpr r r_customize_compiler_for_shlib#s  rFZsharedrTntRTLD_NOWcCs tr|SdS)N) have_rtld)rr r rDr$cCs.tD]$}d|kr|S|dkr|SqdS)z;Return the file extension for an abi3-compliant Extension()z.abi3z.pydNr)suffixr r rget_abi3_suffixGs r'c@sveZdZddZddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ ddZdddZdS)rcCs.|jd}|_t|||_|r*|dS)z;Build extensions in build directory, then copy if --inplacerN)Zinplace _build_extruncopy_extensions_to_source)selfZ old_inplacer r rr)Qs  z build_ext.runc Cs|d}|jD]}||j}||}|d}d|dd}||}tj |tj |}tj |j |} t | ||j |jd|jr||ptj|dqdS)Nbuild_py.)verbosedry_runT)get_finalized_command extensionsget_ext_fullnamenameget_ext_filenamesplitjoinZget_package_dirospathbasename build_librr/r0 _needs_stub write_stubcurdir) r+r,extfullnamefilenamemodpathpackageZ package_dirZ dest_filenameZ src_filenamer r rr*Ys&       z#build_ext.copy_extensions_to_sourcecCst||}||jkr|j|}tjo4t|do4t}|r^td}|dt| }|t}t |t rt j |\}}|j|tStr|jrt j |\}}t j |d|S|S)NZpy_limited_apiZ EXT_SUFFIXzdl-)r(r5ext_mapr ZPY3getattrr'rlen isinstancer r8r9splitextshlib_compilerlibrary_filenamelibtype use_stubs_links_to_dynamicr6r7)r+r@rAr?Zuse_abi3Zso_extfndr r rr5os&      zbuild_ext.get_ext_filenamecCs t|d|_g|_i|_dSN)r(initialize_optionsrIshlibsrDr+r r rrQs zbuild_ext.initialize_optionscCs,t||jpg|_||jdd|jD|_|jrB||jD]}||j|_qH|jD]}|j}||j |<||j | dd<|jr| |pd}|ot ot |t }||_||_||}|_tjtj|j|}|r||jkr|j||rbt rbtj|jkrb|jtjqbdS)NcSsg|]}t|tr|qSr )rGr r r?r r rrs z.build_ext.finalize_options..r-r.F)r(finalize_optionsr2Zcheck_extensions_listrRsetup_shlib_compilerr3r4 _full_namerDr6links_to_dynamicrLrGr rMr<r5 _file_namer8r9dirnamer7r; library_dirsappendr>runtime_library_dirs)r+r?r@ZltdnsrAZlibdirr r rrUs,       zbuild_ext.finalize_optionscCst|j|j|jd}|_t||jdk r8||j|jdk r^|jD]\}}| ||qH|j dk r~|j D]}| |qn|j dk r| |j |jdk r||j|jdk r||j|jdk r||jt||_dS)N)rr0force)rrr0r_rIrZ include_dirsZset_include_dirsZdefineZ define_macroZundefZundefine_macro librariesZ set_librariesr[Zset_library_dirsZrpathZset_runtime_library_dirsZ link_objectsZset_link_objectslink_shared_object__get__)r+rr4valueZmacror r rrVs.               zbuild_ext.setup_shlib_compilercCst|tr|jSt||SrP)rGr export_symbolsr(get_export_symbolsr+r?r r rres zbuild_ext.get_export_symbolscCs\||j}z@t|tr"|j|_t|||jrL|dj }| ||W5||_XdS)Nr,) Z_convert_pyx_sources_to_langrrGr rIr(build_extensionr<r1r;r=)r+r?Z _compilercmdr r rrgs   zbuild_ext.build_extensioncsPtdd|jDd|jddddgtfdd|jDS) z?Return true if 'ext' links to a dynamic lib in the same packagecSsg|] }|jqSr )rW)r libr r rrsz.build_ext.links_to_dynamic..r-Nr.r"c3s|]}|kVqdSrPr )r ZlibnameZlibnamesZpkgr r sz-build_ext.links_to_dynamic..)dictfromkeysrRr7rWr6anyr`rfr rjrrXs zbuild_ext.links_to_dynamiccCst||SrP)r( get_outputs_build_ext__get_stubs_outputsrSr r rroszbuild_ext.get_outputscs6fddjD}t|}tdd|DS)Nc3s0|](}|jrtjjjf|jdVqdS)r-N)r<r8r9r7r;rWr6rTrSr rrksz0build_ext.__get_stubs_outputs..css|]\}}||VqdSrPr )r baseZfnextr r rrks)r2 itertoolsproduct!_build_ext__get_output_extensionslist)r+Z ns_ext_basesZpairsr rSrZ__get_stubs_outputss  zbuild_ext.__get_stubs_outputsccs"dVdV|djrdVdS)N.pyz.pycr,z.pyo)r1optimizerSr r rZ__get_output_extensionss z!build_ext.__get_output_extensionsFcCs,td|j|tjj|f|jdd}|rJtj|rJt|d|j st |d}| dddd t d d tj |jd d dt ddddt dddt ddddg||r(ddlm}||gdd|j d|dj}|dkr||g|d|j dtj|r(|j s(t|dS)Nz writing stub loader for %s to %sr-rvz already exists! Please delete.w zdef __bootstrap__():z- global __bootstrap__, __file__, __loader__z% import sys, os, pkg_resources, impz, dlz: __file__ = pkg_resources.resource_filename(__name__,%r)z del __bootstrap__z if '__loader__' in globals():z del __loader__z# old_flags = sys.getdlopenflags()z old_dir = os.getcwd()z try:z( os.chdir(os.path.dirname(__file__))z$ sys.setdlopenflags(dl.RTLD_NOW)z( imp.load_dynamic(__name__,__file__)z finally:z" sys.setdlopenflags(old_flags)z os.chdir(old_dir)z__bootstrap__()r"r) byte_compileT)rwr_r0Z install_lib)rinforWr8r9r7r6existsrr0openwriteif_dlr:rYcloseZdistutils.utilrzr1rwunlink)r+ output_dirr?compileZ stub_filefrzrwr r rr=sb        zbuild_ext.write_stubN)F)__name__ __module__ __qualname__r)r*r5rQrUrVrergrXrorprtr=r r r rrPs   rc Cs(||j||||||||| | | | dSrP)linkZSHARED_LIBRARY) r+objectsoutput_libnamerr`r[r]rddebug extra_preargsextra_postargs build_temp target_langr r rra#sraZstaticc Cs^|dks ttj|\}} tj| \}}|ddrH|dd}|||||| dS)Nxri)AssertionErrorr8r9r6rHrJ startswithZcreate_static_lib)r+rrrr`r[r]rdrrrrrrAr:r?r r rra2s  ) NNNNNrNNNN) NNNNNrNNNN)+r8rrrZdistutils.command.build_extrZ _du_build_extZdistutils.file_utilrZdistutils.ccompilerrZdistutils.sysconfigrrZdistutils.errorsrZ distutilsrZsetuptools.extensionr Zsetuptools.externr ZPY2r Z get_suffixesrZimportlib.machineryZCython.Distutils.build_extr( __import__ ImportErrorrrrr#rLrKrr4Zdlhasattrrr'rar r r rsz               Q site-packages/setuptools/command/__pycache__/rotate.cpython-38.opt-1.pyc000064400000004730151464452710022235 0ustar00U Qabt@s`ddlmZddlmZddlmZddlZddlZddlm Z ddl m Z Gddde Z dS) ) convert_path)log)DistutilsOptionErrorN)six)Commandc@s:eZdZdZdZdddgZgZddZdd Zd d Z d S) rotatezDelete older distributionsz2delete older distributions, keeping N newest files)zmatch=mzpatterns to match (required))z dist-dir=dz%directory where the distributions are)zkeep=kz(number of matching distributions to keepcCsd|_d|_d|_dS)N)matchdist_dirkeepselfr=/usr/lib/python3.8/site-packages/setuptools/command/rotate.pyinitialize_optionsszrotate.initialize_optionscCs|jdkrtd|jdkr$tdzt|j|_Wntk rPtdYnXt|jtjrxdd|jdD|_| dddS) NzQMust specify one or more (comma-separated) match patterns (e.g. '.zip' or '.egg')z$Must specify number of files to keepz--keep must be an integercSsg|]}t|qSr)rstrip).0prrr *sz+rotate.finalize_options..,Zbdist)r r ) r rr int ValueError isinstancerZ string_typessplitZset_undefined_optionsrrrrfinalize_optionss   zrotate.finalize_optionscCs|dddlm}|jD]}|jd|}|tj|j|}dd|D}| | t dt ||||jd}|D]<\}}t d||jstj|rt|qt|qqdS) NZegg_infor)glob*cSsg|]}tj||fqSr)ospathgetmtime)rfrrrr6szrotate.run..z%d file(s) matching %sz Deleting %s)Z run_commandrr Z distributionZget_namerr joinr sortreverserinfolenr Zdry_runisdirshutilZrmtreeunlink)rrpatternfilestr"rrrrun/s        z rotate.runN) __name__ __module__ __qualname____doc__ descriptionZ user_optionsZboolean_optionsrrr.rrrrr sr) Zdistutils.utilrZ distutilsrZdistutils.errorsrrr)Zsetuptools.externrZ setuptoolsrrrrrrs     site-packages/setuptools/command/__pycache__/bdist_egg.cpython-38.opt-1.pyc000064400000033546151464452710022675 0ustar00U Qab G@sxdZddlmZddlmZmZddlmZddlm Z ddl Z ddl Z ddl Z ddl Z ddlZddlmZddlmZmZmZdd lmZdd lmZdd lmZzdd lmZmZd dZWn,ek rddlm Z mZddZYnXddZ!ddZ"ddZ#GdddeZ$e%&d'Z(ddZ)ddZ*ddZ+d d!d"Z,d#d$Z-d%d&Z.d'd(Z/d)d*d+d,gZ0d1d/d0Z1dS)2z6setuptools.command.bdist_egg Build .egg distributions)DistutilsSetupError) remove_treemkpath)log)CodeTypeN)six)get_build_platform Distributionensure_directory) EntryPoint)Library)Command)get_pathget_python_versioncCstdS)NZpurelib)rrr@/usr/lib/python3.8/site-packages/setuptools/command/bdist_egg.py _get_purelibsr)get_python_librcCstdS)NF)rrrrrrscCs2d|krtj|d}|dr.|dd}|S)N.rmodulei)ospathsplitextendswith)filenamerrr strip_module#s   rccs6t|D]&\}}}|||||fVq dS)zbDo os.walk in a reproducible way, independent of indeterministic filesystem readdir order N)rwalksort)dirbasedirsfilesrrr sorted_walk+sr"c Cs6td}t|d}|||W5QRXdS)NaR def __bootstrap__(): global __bootstrap__, __loader__, __file__ import sys, pkg_resources, imp __file__ = pkg_resources.resource_filename(__name__, %r) __loader__ = None; del __bootstrap__, __loader__ imp.load_dynamic(__name__,__file__) __bootstrap__() w)textwrapdedentlstripopenwrite)resourcepyfileZ_stub_templatefrrr write_stub5s r,c@seZdZdZddddefdddd gZd d d gZd dZddZddZ ddZ ddZ ddZ ddZ ddZddZdd Zd!d"Zd#S)$ bdist_eggzcreate an "egg" distribution)z bdist-dir=bz1temporary directory for creating the distributionz plat-name=pz;platform name to embed in generated filenames (default: %s))exclude-source-filesNz+remove all .py files from the generated egg) keep-tempkzPkeep the pseudo-installation tree around after creating the distribution archive)z dist-dir=dz-directory to put final built distributions in) skip-buildNz2skip rebuilding everything (for testing/debugging)r1r4r0cCs.d|_d|_d|_d|_d|_d|_d|_dS)Nr) bdist_dir plat_name keep_tempdist_dir skip_build egg_outputexclude_source_filesselfrrrinitialize_optionsZszbdist_egg.initialize_optionscCs|d}|_|j|_|jdkr>|dj}tj|d|_|jdkrPt |_| dd|j dkrt dd|j |jt|jo|j }tj|j|d|_ dS)Negg_infoZbdistZegg)r8r8z.egg)get_finalized_commandei_cmdr?r5 bdist_baserrjoinr6rZset_undefined_optionsr:r Zegg_nameZ egg_versionr distributionhas_ext_modulesr8)r=rArBbasenamerrrfinalize_optionscs$     zbdist_egg.finalize_optionscCs|j|d_tjtjt}|jj g}|j_ |D]}t |t rt |dkrtj |drtj|d}tj|}||ks||tjr|t |dd|df}|jj |q:z"td|j|jddddW5||j_ XdS)Ninstallrzinstalling package data to %s install_data)forceroot)r5r@ install_librrnormcaserealpathrrD data_files isinstancetuplelenisabs startswithsepappendrinfo call_command)r=Z site_packagesolditemrPZ normalizedrrrdo_install_data{s"  zbdist_egg.do_install_datacCs|jgS)N)r:r<rrr get_outputsszbdist_egg.get_outputscKsPtD]}|||jq|d|j|d|j|j|f|}|||S)z8Invoke reinitialized command `cmdname` with keyword argsr9dry_run)INSTALL_DIRECTORY_ATTRS setdefaultr5r9r_Zreinitialize_command run_command)r=ZcmdnamekwdirnamecmdrrrrZs zbdist_egg.call_commandcCs|dtd|j|d}|j}d|_|jrH|jsH|d|j ddd}||_| \}}g|_ g}t |D]|\}}t j|\} } t j|jt| d} |j | td ||jstt j|| || |t jd ||<qz|r|||jjr||j} t j| d } || |jjrlt j| d }td ||j d|dd|| t j| d}|rtd||jst|t|d}| d|| d|!n,t j"|rtd||jst #|t$t j| d |%t j&t j|j'dr.+)\.(?P[^.]+)\.pycname.pyczRenaming file from [%s] to [%s])rrYwalk_eggr5rrrCrdebugrwrematchpardirgroupremoveOSErrorrename) r=rr r!rrZpath_oldpatternmZpath_newrrrr|s4       zbdist_egg.zap_pyfilescCs2t|jdd}|dk r|Stdt|j|jS)Nryz4zip_safe flag not set; analyzing archive contents...)rrDrr{ analyze_eggr5rq)r=saferrrry s  zbdist_egg.zip_safec Cst|jjpd}|did}|dkr0dS|jr<|jrJtd|fdjt j }|j }d |j}|jd}t j|j}d t}|jstt j|j|jd t|jd} | || d S) Nzsetuptools.installationZ eggsecutabler#zGeggsecutable entry point (%r) cannot have 'extras' or refer to a modulez{}.{}rraH#!/bin/sh if [ `basename $0` = "%(basename)s" ] then exec python%(pyver)s -c "import sys, os; sys.path.insert(0, os.path.abspath('$0')); from %(pkg)s import %(base)s; sys.exit(%(full)s())" "$@" else echo $0 is not the correct name for this egg file. echo Please rename it back to %(basename)s and try again. exec false fi roa)r Z parse_maprDZ entry_pointsgetZattrsZextrasrformatsys version_infoZ module_namerCrrrFr:localsr_rrdr'r(ru) r=ZepmZepZpyverpkgZfullrrFheaderr+rrrr~s0       zbdist_egg.gen_headercCshtj|j}tj|d}|jjjD]<}||r&tj||t |d}t || ||q&dS)z*Copy metadata (egg info) to the target_dirrN) rrnormpathr?rCrAZfilelistr!rVrTr Z copy_file)r=Z target_dirZ norm_egg_infoprefixrtargetrrrrt:s zbdist_egg.copy_metadata_toc Csg}g}|jdi}t|jD]f\}}}|D].}tj|dtkr*||||q*|D]"}|||d|tj||<q^q|j r| d}|j D]Z} t | trq|| j} || }tj|dstjtj|j|r||q||fS)zAGet a list of relative paths to C extensions in the output distrorrJrgZ build_extzdl-)r5r"rrrlowerNATIVE_EXTENSIONSrXrCrDrEr@ extensionsrRr Zget_ext_fullnamerZget_ext_filenamerFrVrz) r=rrpathsrr r!rZ build_cmdrfullnamerrrrpFs0        zbdist_egg.get_ext_outputsN)__name__ __module__ __qualname__ descriptionrZ user_optionsZboolean_optionsr>rGr]r^rZrr|ryr~rtrprrrrr-Cs2  Q' r-z.dll .so .dylib .pydccsHt|}t|\}}}d|kr(|d|||fV|D] }|Vq8dS)z@Walk an unpacked egg's contents, skipping the metadata directoryrhN)r"nextr)egg_dirZwalkerrr r!Zbdfrrrrfs  rc CstD](\}}tjtj|d|r|Sqtst||krzt|q|dk rt||krt|d}| d| qdS)Nrkrl) rrrrrCrzboolrwr'r(ru)rrrrr+rrrrxs    rxzzip-safez not-zip-safe)TFc Cstj||}|dd|kr"dS|t|ddtjd}||rJdpLdtj|d}tjrld}nt j d kr|d }nd }t |d }| |t |} |d} tt| } d D]} | | krtd|| d} qd| kr dD]} | | krtd|| d} q| S)z;Check whether module possibly uses unsafe-for-zipfile stuffNTrJrrr) rb)__file____path__z%s: module references %sFinspect) Z getsourceZ getabsfileZ getsourcefileZgetfilegetsourcelinesZ findsourceZ getcommentsZ getframeinfoZgetinnerframesZgetouterframesstackZtracez"%s: module MAY be using inspect.%s)rrrCrTrsrWrrZPY2rrr'readmarshalloadrudictfromkeys iter_symbolsrr{) rrrrqrrrskipr+coderZsymbolsZbadrrrrs4      rccsT|jD] }|Vq|jD]6}t|tjr0|Vqt|trt|D] }|VqBqdS)zBYield names and strings used by `code` and its nested code objectsN)co_names co_constsrRrZ string_typesrr)rrconstrrrrs     rcCs2tjdstjdkrdStdtddS)NjavaZcliTz1Unable to analyze compiled code on this platform.zfPlease ask the author to include a 'zip_safe' setting (either True or False) in the package's setup.py)rplatformrVrr{rrrrrs rrNrjrKZ install_baseTr#c sddl}ttj|dtd|fdd}|rB|jn|j}s|j |||d} t D]\} } } || | | qd| n t D]\} } } |d| | q|S)aqCreate a zip file from all the files under 'base_dir'. The output zip file will be named 'base_dir' + ".zip". Uses either the "zipfile" Python module (if available) or the InfoZIP "zip" utility (if installed and found on the default search path). If neither tool is available, raises DistutilsExecError. Returns the name of the output zip file. rNroz#creating '%s' and adding '%s' to itcs`|D]V}tjtj||}tj|r|tdd}sN|||td|qdS)NrJz adding '%s') rrrrCrvrTr(rr)zrdnamesrrr/base_dirr_rrvisits  zmake_zipfile..visit) compression) zipfilerrrrdrrYZ ZIP_DEFLATEDZ ZIP_STOREDZZipFiler"ru) Z zip_filenamerrmr_compressrnrrrrrdr r!rrrr}s  r})rrTr#)2__doc__Zdistutils.errorsrZdistutils.dir_utilrrZ distutilsrtypesrrrrr$rZsetuptools.externrZ pkg_resourcesrr r r Zsetuptools.extensionr Z setuptoolsr Z sysconfigrrr ImportErrorZdistutils.sysconfigrrr"r,r-rrsplitrrrrxrrrrr`r}rrrrsX         " $  site-packages/setuptools/command/__pycache__/dist_info.cpython-38.pyc000064400000002521151464452710021752 0ustar00U Qab@s8dZddlZddlmZddlmZGdddeZdS)zD Create a dist_info directory As defined in the wheel specification N)Command)logc@s.eZdZdZdgZddZddZddZd S) dist_infozcreate a .dist-info directory)z egg-base=ezLdirectory containing .egg-info directories (default: top of the source tree)cCs d|_dSN)egg_baseselfr @/usr/lib/python3.8/site-packages/setuptools/command/dist_info.pyinitialize_optionsszdist_info.initialize_optionscCsdSrr rr r r finalize_optionsszdist_info.finalize_optionscCsn|d}|j|_|||jdtd d}tdt j ||d}| |j|dS)Negg_infoz .egg-infoz .dist-infoz creating '{}' bdist_wheel) Zget_finalized_commandrr runrlenrinfoformatospathabspathZegg2dist)r rZ dist_info_dirrr r r rs  z dist_info.runN)__name__ __module__ __qualname__ descriptionZ user_optionsr r rr r r r r s r)__doc__rZdistutils.corerZ distutilsrrr r r r s  site-packages/setuptools/command/__pycache__/test.cpython-38.pyc000064400000020433151464452710020755 0ustar00U Qab%@sddlZddlZddlZddlZddlZddlZddlmZmZddl m Z ddlm Z ddl m Z ddlmZmZddlmZmZmZmZmZmZmZmZmZddlmZd d lmZeZ Gd d d e Z!Gd ddZ"GdddeZ#dS)N)DistutilsErrorDistutilsOptionError)log) TestLoader)six)mapfilter) resource_listdirresource_existsnormalize_path working_set_namespace_packagesevaluate_markeradd_activation_listenerrequire EntryPoint)Command)_unique_everseenc@seZdZddZdddZdS)ScanningLoadercCst|t|_dSN)r__init__set_visitedselfr;/usr/lib/python3.8/site-packages/setuptools/command/test.pyrs zScanningLoader.__init__NcCs||jkrdS|j|g}|t||t|drH||t|drt|jdD]`}| dr|dkr|jd|dd}n"t |j|d r^|jd|}nq^|| |q^t |d kr| |S|d SdS) aReturn a suite of all tests cases contained in the given module If the module is a package, load tests from all the modules in it. If the module has an ``additional_tests`` function, call it and add the return value to the tests. Nadditional_tests__path__z.pyz __init__.py.z /__init__.pyrr)raddappendrloadTestsFromModulehasattrrr __name__endswithr ZloadTestsFromNamelenZ suiteClass)rmodulepatternZtestsfileZ submodulerrrr%s$      z"ScanningLoader.loadTestsFromModule)N)r' __module__ __qualname__rr%rrrrrsrc@seZdZddZdddZdS)NonDataPropertycCs ||_dSrfget)rr1rrrrAszNonDataProperty.__init__NcCs|dkr |S||Srr0)robjZobjtyperrr__get__DszNonDataProperty.__get__)N)r'r-r.rr3rrrrr/@sr/c@seZdZdZdZdddgZddZdd Zed d Z d d Z ddZ e j gfddZee j ddZeddZddZddZeddZeddZdS)testz.Command to run unit tests after in-place buildz0run unit tests after in-place build (deprecated))z test-module=mz$Run 'test_suite' in specified module)z test-suite=sz9Run single test, case or suite (e.g. 'module.test_suite'))z test-runner=rzTest runner to usecCsd|_d|_d|_d|_dSr) test_suite test_module test_loader test_runnerrrrrinitialize_optionsVsztest.initialize_optionscCs|jr|jrd}t||jdkrD|jdkr8|jj|_n |jd|_|jdkr^t|jdd|_|jdkrnd|_|jdkrt|jdd|_dS)Nz1You may specify a module or a suite, but not bothz .test_suiter:z&setuptools.command.test:ScanningLoaderr;)r8r9r distributionr:getattrr;)rmsgrrrfinalize_options\s        ztest.finalize_optionscCs t|Sr)list _test_argsrrrr test_argsosztest.test_argsccs4|jstjdkrdV|jr"dV|jr0|jVdS)N)Zdiscoverz --verbose)r8sys version_infoverboserrrrrBss ztest._test_argsc Cs| |W5QRXdS)zI Backward compatibility for project_on_sys_path context. N)project_on_sys_path)rfuncrrrwith_project_on_sys_path{s ztest.with_project_on_sys_pathc csPtjot|jdd}|rv|jddd|d|d}t|j}|jd|d|d|jddd|dn"|d|jdd d|d|d}t j dd}t j }zbt|j}t j d|ttd d td |j|jf||g dVW5QRXW5|t j dd<t j t j |tXdS) Nuse_2to3Fbuild_pyr)ZinplaceZegg_info)egg_baseZ build_extrcSs|Sr)Zactivate)distrrrz*test.project_on_sys_path..z%s==%s)rPY3r>r=Zreinitialize_commandZ run_commandZget_finalized_commandr Z build_librFpathmodulescopyclearupdater rrNinsertrrZegg_nameZ egg_versionpaths_on_pythonpath) rZ include_distsZ with_2to3Zbpy_cmdZ build_pathZei_cmdZold_pathZ old_modulesZ project_pathrrrrIs8             ztest.project_on_sys_pathc cst}tjd|}tjdd}zBtjt|}td||g}tj|}|r\|tjd<dVW5||kr~tjddn |tjd<XdS)z Add the indicated paths to the head of the PYTHONPATH environment variable so that subprocesses will also see the packages at these paths. Do this in a context that restores the value on exit. Z PYTHONPATHr N) objectosenvirongetpoppathsepjoinrr)pathsZnothingZorig_pythonpathZcurrent_pythonpathprefixZto_joinnew_pathrrrrYs    ztest.paths_on_pythonpathcCsD||j}||jpg}|dd|jD}t|||S)z Install the requirements indicated by self.distribution and return an iterable of the dists that were built. css0|](\}}|drt|ddr|VqdS):rN) startswithr).0kvrrr s z%test.install_dists..)Zfetch_build_eggsZinstall_requiresZ tests_requireZextras_requireitems itertoolschain)rOZir_dZtr_dZer_drrr install_distss   ztest.install_distsc Cs|dtj||j}d|j}|jr>|d|dS|d|tt d|}| |"| | W5QRXW5QRXdS)NzWARNING: Testing via this command is deprecated and will be removed in a future version. Users looking for a generic test entry point independent of test runner are encouraged to use tox. zskipping "%s" (dry run)z running "%s"location)announcerZWARNrmr=r`_argvZdry_runroperator attrgetterrYrI run_tests)rZinstalled_distscmdrarrrruns    ztest.runcCstjr~t|jddr~|jdd}|tkr~g}|tjkrD| ||d7}tjD]}| |rR| |qRt t tjj |tjdd|j||j||jdd}|jsd|j}||tjt|dS)NrLFr!r)Z testLoaderZ testRunnerexitzTest failed: %s)rrRr>r=r8splitr rFrTr$rerAr __delitem__unittestmainrq_resolve_as_epr:r;resultZ wasSuccessfulrprZERRORr)rr*Z del_modulesnamer4r?rrrrts.         ztest.run_testscCs dg|jS)Nrz)rCrrrrrq sz test._argvcCs$|dkr dStd|}|S)zu Load the indicated attribute value, called, as a as if it were specified as an entry point. Nzx=)rparseZresolve)valZparsedrrrr|sztest._resolve_as_epN)r'r-r.__doc__ descriptionZ user_optionsr<r@r/rCrBrK contextlibcontextmanagerrI staticmethodrYrmrvrtpropertyrqr|rrrrr4Js2 -   r4)$r[rrrFrrkrzZdistutils.errorsrrZ distutilsrrZsetuptools.externrZsetuptools.extern.six.movesrrZ pkg_resourcesr r r r r rrrrZ setuptoolsrrMrtypeZ __metaclass__rr/r4rrrrs"   ,  ) site-packages/setuptools/command/__pycache__/upload_docs.cpython-38.pyc000064400000013774151464452710022304 0ustar00U Qab@sdZddlmZddlmZddlmZddlZddlZddl Z ddl Z ddl Z ddl Z ddl Z ddlmZddlmZmZddlmZd d lmZd d ZGd ddeZdS)zpupload_docs Implements a Distutils 'upload_docs' subcommand (upload documentation to PyPI's pythonhosted.org). )standard_b64encode)log)DistutilsOptionErrorN)six) http_clienturllib)iter_entry_points)uploadcCstjr dnd}|d|S)Nsurrogateescapestrictzutf-8)rPY3encode)serrorsrB/usr/lib/python3.8/site-packages/setuptools/command/upload_docs.py_encodesrc@seZdZdZdZdddejfddgZejZdd Zd efgZ d d Z d dZ ddZ ddZ eddZeddZddZdS) upload_docszhttps://pypi.python.org/pypi/zUpload documentation to PyPIz repository=rzurl of repository [default: %s])z show-responseNz&display full response text from server)z upload-dir=Nzdirectory to uploadcCs"|jdkrtddD]}dSdS)Nzdistutils.commands build_sphinxT) upload_dirr)selfZeprrr has_sphinx/s zupload_docs.has_sphinxrcCst|d|_d|_dS)N)r initialize_optionsr target_dir)rrrrr6s zupload_docs.initialize_optionscCst||jdkrN|r0|d}|j|_q`|d}tj |j d|_n| d|j|_d|j krtt d|d|jdS)NrbuildZdocsrzpypi.python.orgz3Upload_docs command is deprecated. Use RTD instead.zUsing upload directory %s)r finalize_optionsrrZget_finalized_commandZbuilder_target_dirrospathjoinZ build_baseZensure_dirname repositoryrwarnannounce)rrrrrrr;s        zupload_docs.finalize_optionsc Cst|d}z||jt|jD]x\}}}||jkrP|sPd}t||j|D]H}tj ||}|t |jd tjj } tj | |} | || qTq&W5|XdS)Nwz'no files found in upload directory '%s')zipfileZZipFilecloseZmkpathrrwalkrrr lenlstripsepwrite) rfilenamezip_filerootdirsfilesZtmplnameZfullrelativedestrrrcreate_zipfileKs  zupload_docs.create_zipfilec Csh|D]}||qt}|jj}tj |d|}z| || |W5t |XdS)Nz%s.zip)Zget_sub_commandsZ run_commandtempfileZmkdtemp distributionmetadataget_namerrr shutilZrmtreer4 upload_file)rZcmd_nameZtmp_dirr1r-rrrrun[s    zupload_docs.runccs|\}}d|}t|ts |g}|D]f}t|trL|d|d7}|d}nt|}|Vt|VdV|V|r$|dddkr$dVq$dS) Nz* Content-Disposition: form-data; name="%s"z; filename="%s"rr s   ) isinstancelisttupler)item sep_boundarykeyvaluestitlevaluerrr _build_partis    zupload_docs._build_partc Csnd}d|}|d}|df}tj|j|d}t||}tj|}t||} d|d} d | | fS) z= Build up the MIME payload for the POST data s3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254s --s--r>)rCz multipart/form-data; boundary=%sascii) functoolspartialrHmapitems itertoolschain from_iterabledecoder ) clsdataboundaryrCZ end_boundaryZ end_itemsZbuilderZ part_groupspartsZ body_items content_typerrr_build_multipart}s  zupload_docs._build_multipartc CsJt|d}|}W5QRX|jj}d|tj||fd}t|j d|j }t |}t j rn|d}d|}||\}} d|j} || tjtj|j\} } } }}}|s|s|rt| dkrt| }n | d krt| }n td | d }zZ||d | | }|d ||dtt||d|| |!|Wn>t"j#k r}z|t|tj$WYdSd}~XYnX|%}|j&dkrd|j&|j'f} || tjnb|j&dkr|(d}|dkrd|}d|} || tjnd|j&|j'f} || tj$|j)rFt*d|ddS)NrbZ doc_upload)z:actionr1content:rIzBasic zSubmitting documentation to %sZhttpZhttpszunsupported schema ZPOSTz Content-typezContent-lengthZ AuthorizationzServer response (%s): %si-ZLocationzhttps://pythonhosted.org/%s/zUpload successful. Visit %szUpload failed (%s): %szK---------------------------------------------------------------------------)+openreadr6r7r8rrbasenamerZusernameZpasswordrrr rRrXr!r#rINFOrparseZurlparseAssertionErrorrZHTTPConnectionZHTTPSConnectionZconnectZ putrequestZ putheaderstrr(Z endheaderssendsocketerrorZERRORZ getresponseZstatusreasonZ getheaderZ show_responseprint)rr,frZmetarTZ credentialsZauthZbodyZctmsgZschemaZnetlocZurlZparamsZqueryZ fragmentsZconnrWerlocationrrrr:sd               zupload_docs.upload_fileN)__name__ __module__ __qualname__ZDEFAULT_REPOSITORY descriptionr Z user_optionsZboolean_optionsrZ sub_commandsrrr4r; staticmethodrH classmethodrXr:rrrrrs(   r)__doc__base64rZ distutilsrZdistutils.errorsrrrfr%r5r9rOrKZsetuptools.externrZsetuptools.extern.six.movesrrZ pkg_resourcesrr rrrrrrs       site-packages/setuptools/command/__pycache__/bdist_rpm.cpython-38.opt-1.pyc000064400000003370151464452710022721 0ustar00U Qab@s(ddlmmZGdddejZdS)Nc@s eZdZdZddZddZdS) bdist_rpmaf Override the default bdist_rpm behavior to do the following: 1. Run egg_info to ensure the name and version are properly calculated. 2. Always run 'install' using --single-version-externally-managed to disable eggs in RPM distributions. 3. Replace dash with underscore in the version numbers for better RPM compatibility. cCs|dtj|dS)NZegg_info)Z run_commandorigrrun)selfr@/usr/lib/python3.8/site-packages/setuptools/command/bdist_rpm.pyrs z bdist_rpm.runcsl|j}|dd}tj|}d|d|fdd|D}|d}d|}||||S)N-_z%define version cs0g|](}|ddddddqS)zSource0: %{name}-%{version}.tarz)Source0: %{name}-%{unmangled_version}.tarzsetup.py install z5setup.py install --single-version-externally-managed z%setupz&%setup -n %{name}-%{unmangled_version})replace).0lineZline23Zline24rr s  z-bdist_rpm._make_spec_file..z%define unmangled_version )Z distributionZ get_versionr rr_make_spec_fileindexinsert)rversionZ rpmversionspecZ insert_locZunmangled_versionrr rrs      zbdist_rpm._make_spec_fileN)__name__ __module__ __qualname____doc__rrrrrrrs r)Zdistutils.command.bdist_rpmZcommandrrrrrrssite-packages/setuptools/command/__pycache__/saveopts.cpython-38.pyc000064400000001571151464452710021644 0ustar00U Qab@s$ddlmZmZGdddeZdS)) edit_config option_basec@seZdZdZdZddZdS)saveoptsz#Save command-line options to a filez7save supplied options to setup.cfg or other config filecCsh|j}i}|jD]B}|dkrq||D]$\}\}}|dkr,|||i|<q,qt|j||jdS)Nrz command line)Z distributionZcommand_optionsZget_option_dictitems setdefaultrfilenameZdry_run)selfZdistZsettingscmdoptsrcvalr ?/usr/lib/python3.8/site-packages/setuptools/command/saveopts.pyrun s z saveopts.runN)__name__ __module__ __qualname____doc__ descriptionrr r r rrsrN)Zsetuptools.command.setoptrrrr r r rssite-packages/setuptools/command/__pycache__/install_egg_info.cpython-38.opt-1.pyc000064400000004543151464452710024244 0ustar00U Qab@s\ddlmZmZddlZddlmZddlmZddlmZddl Z Gdddej eZ dS))logdir_utilN)Command) namespaces)unpack_archivec@sBeZdZdZdZdgZddZddZddZd d Z d d Z d S)install_egg_infoz.Install an .egg-info directory for the package)z install-dir=dzdirectory to install tocCs d|_dSN) install_dirselfr G/usr/lib/python3.8/site-packages/setuptools/command/install_egg_info.pyinitialize_optionssz#install_egg_info.initialize_optionscCsV|dd|d}tdd|j|jd}|j|_tj |j ||_ g|_ dS)NZ install_lib)r r egg_infoz .egg-info)Zset_undefined_optionsZget_finalized_command pkg_resourcesZ DistributionZegg_nameZ egg_versionrsourceospathjoinr targetoutputs)r Zei_cmdbasenamer r rfinalize_optionss z!install_egg_info.finalize_optionscCs|dtj|jr:tj|js:tj|j|jdn(tj |jrb| tj |jfd|j|jstt |j| |jdd|j|jf|dS)Nr)dry_runz Removing r Copying %s to %s)Z run_commandrrisdirrislinkrZ remove_treerexistsZexecuteunlinkrZensure_directorycopytreerZinstall_namespacesr r r rrun!s  zinstall_egg_info.runcCs|jSr )rr r r r get_outputs.szinstall_egg_info.get_outputscs fdd}tjj|dS)NcsDdD] }||sd||krdSqj|td|||S)N)z.svn/zCVS//r) startswithrappendrdebug)srcZdstskipr r rskimmer3s  z*install_egg_info.copytree..skimmer)rrr)r r)r r rr 1s zinstall_egg_info.copytreeN) __name__ __module__ __qualname____doc__ descriptionZ user_optionsrrr!r"r r r r rr s  r) Z distutilsrrrZ setuptoolsrrZsetuptools.archive_utilrrZ Installerrr r r rs    site-packages/setuptools/command/__pycache__/upload.cpython-38.opt-1.pyc000064400000012123151464452710022216 0ustar00U Qab@sddlZddlZddlZddlZddlmZddlmZddlm Z ddl m Z ddl mZddlmZmZddlmZdd lmZGd d d e j Z dS) N)standard_b64encode)log)upload)spawn)DistutilsError)urlopenRequest) HTTPError)urlparsec@s8eZdZdZddZddZddZdd Zd d Zd S) rza Override default upload behavior to obtain password in a variety of different ways. c Cs&ztj|W5|dtjXdS)NzjWARNING: Uploading via this command is deprecated, use twine to upload instead (https://pypi.org/p/twine/))announcerZWARNorigrrunselfr=/usr/lib/python3.8/site-packages/setuptools/command/upload.pyr s z upload.runcCs8tj||jpt|_|jp0|p0||_dSN) r rfinalize_optionsusernamegetpassZgetuserpassword_load_password_from_keyring_prompt_for_passwordrrrrr"s zupload.finalize_optionsc Cst|j\}}}}}} |s"|s"| r0td|j|dkrDtd||jr|ddd|g} |jrnd|jg| dd<t| |jd t|d } | } W5QRX|j j } d d | | t j|| f||t| t| | | | | | | | | | | | | | !d }d|d<|jrdt j|dt|dd f|d<|j"d|j#$d}dt%|&d}d}d|$d}|d}t'(}|)D]\}}d|}t*|t+s|g}|D]j}t,|t-kr|d|d7}|d}nt|$d}|.||.|$d|.d|.|qܐq|.||/}d||jf}|0|t1j2d|tt3||d }t4|j||d!}zt5|}|6}|j7}Wnft8k r}z|j9}|j7}W5d}~XYn8t:k r"}z|0t|t1j;W5d}~XYnX|d"kr|0d#||ft1j2|j<rt=|d$d%d&|}|dk rd'>d(|d(f}|0|t1j2n"d)||f}|0|t1j;t?|dS)*NzIncompatible url %s)ZhttpZhttpszunsupported schema Zgpgz --detach-signz-az --local-user)dry_runrbZ file_upload1)z:actionZprotocol_versionnameversioncontentZfiletype pyversionZ md5_digestZmetadata_versionZsummaryZ home_pageZauthorZ author_emaillicense descriptionkeywordsplatformZ classifiersZ download_urlZprovidesZrequiresZ obsoletesZcommentz.ascZ gpg_signature:asciizBasic z3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254s --s-- z+ Content-Disposition: form-data; name="%s"z; filename="%s"rzutf-8s zSubmitting %s to %sz multipart/form-data; boundary=%s)z Content-typezContent-lengthZ Authorization)dataheaderszServer response (%s): %sZ_read_pypi_responsecSsdSrr)xrrrz$upload.upload_file.. zK---------------------------------------------------------------------------zUpload failed (%s): %s)@r repositoryAssertionErrorZsignZidentityrropenreadZ distributionZmetadataZget_nameZ get_versionospathbasenamehashlibZmd5Z hexdigeststrZget_metadata_versionZget_descriptionZget_urlZ get_contactZget_contact_emailZ get_licenceZget_long_descriptionZ get_keywordsZ get_platformsZget_classifiersZget_download_urlZ get_providesZ get_requiresZ get_obsoletesrrencoderdecodeioBytesIOitems isinstancelisttypetuplewritegetvaluer rINFOlenrrZgetcodemsgr codeOSErrorZERRORZ show_responsegetattrjoinr) rZcommandr filenameZschemaZnetlocZurlZparamsZqueryZ fragmentsZgpg_argsfrmetar)Z user_passZauthboundaryZ sep_boundaryZ end_boundaryZbodykeyvaluetitlerFr*ZrequestresultZstatusreasonetextrrr upload_file0s      !          zupload.upload_filecCs4ztd}||j|jWStk r.YnXdS)zM Attempt to load password from keyring. Suppress Exceptions. keyringN) __import__Z get_passwordr0r Exception)rrWrrrrs z"upload._load_password_from_keyringc Cs(z tWSttfk r"YnXdS)zH Prompt for a password on the tty. Suppress Exceptions. N)rrYKeyboardInterruptrrrrrs zupload._prompt_for_passwordN) __name__ __module__ __qualname____doc__r rrVrrrrrrrs  r)r;r4r7rbase64rZ distutilsrZdistutils.commandrr Zdistutils.spawnrZdistutils.errorsrZ*setuptools.extern.six.moves.urllib.requestrrZ(setuptools.extern.six.moves.urllib.errorr Z(setuptools.extern.six.moves.urllib.parser rrrrs       site-packages/setuptools/command/__pycache__/setopt.cpython-38.opt-1.pyc000064400000010671151464452710022256 0ustar00U Qab@sddlmZddlmZddlmZddlZddlZddlmZddl m Z ddd d gZ dd dZ dddZ Gdd d e ZGdd d eZdS)) convert_path)log)DistutilsOptionErrorN) configparser)Command config_file edit_config option_basesetoptlocalcCsh|dkr dS|dkr,tjtjtjdS|dkrZtjdkrBdpDd}tjtd |St d |d S) zGet the filename of the distutils, local, global, or per-user config `kind` must be one of "local", "global", or "user" r z setup.cfgglobalz distutils.cfguserposix.z~/%spydistutils.cfgz7config_file() type must be 'local', 'global', or 'user'N) ospathjoindirname distutils__file__name expanduserr ValueError)Zkinddotr=/usr/lib/python3.8/site-packages/setuptools/command/setopt.pyrs Fc Cs&td|t}||g|D]\}}|dkrRtd||||q(||sttd||| ||D]p\}}|dkrtd|||| ||| |std||||q|td||||| |||q|q(td||s"t |d }||W5QRXdS) aYEdit a configuration file to include `settings` `settings` is a dictionary of dictionaries or ``None`` values, keyed by command/section name. A ``None`` value means to delete the entire section, while a dictionary lists settings to be changed or deleted in that section. A setting of ``None`` means to delete that setting. zReading configuration from %sNzDeleting section [%s] from %szAdding new section [%s] to %szDeleting %s.%s from %sz#Deleting empty [%s] section from %szSetting %s.%s to %r in %sz Writing %sw)rdebugrZRawConfigParserreaditemsinfoZremove_sectionZ has_sectionZ add_sectionZ remove_optionoptionssetopenwrite) filenameZsettingsdry_runZoptsZsectionr"optionvaluefrrrr!sJ          c@s2eZdZdZdddgZddgZddZd d Zd S) r z|js>tddS)Nz%Must specify --command *and* --optionz$Must specify --set-value or --remove)r r9rEr(rrFrCr2rrrr9s  zsetopt.finalize_optionscCs*t|j|j|jdd|jii|jdS)N-_)rr&rEr(replacerFr'r2rrrrunsz setopt.runN) r:r;r<r= descriptionr r>r?r4r9rJrrrrr ss )r )F)Zdistutils.utilrrrZdistutils.errorsrrZsetuptools.extern.six.movesrZ setuptoolsr__all__rrr r rrrrs        +'site-packages/setuptools/command/__pycache__/build_clib.cpython-38.pyc000064400000004605151464452710022071 0ustar00U Qab@sLddlmmZddlmZddlmZddlm Z GdddejZdS)N)DistutilsSetupError)log)newer_pairwise_groupc@seZdZdZddZdS) build_clibav Override the default build_clib behaviour to do the following: 1. Implement a rudimentary timestamp-based dependency system so 'compile()' doesn't run every time. 2. Add more keys to the 'build_info' dictionary: * obj_deps - specify dependencies for each object compiled. this should be a dictionary mapping a key with the source filename to a list of dependencies. Use an empty string for global dependencies. * cflags - specify a list of additional flags to pass to the compiler. c Csr|D]f\}}|d}|dks.t|ttfs:td|t|}td||dt}t|tsrtd|g}|dt}t|ttfstd||D]P}|g} | |||t} t| ttfstd|| | | | q|j j ||j d} t || ggfkrT|d} |d } |d }|j j||j | | ||jd }|j j| ||j|jd qdS) Nsourceszfin 'libraries' option (library '%s'), 'sources' must be present and must be a list of source filenameszbuilding '%s' libraryobj_depsz\in 'libraries' option (library '%s'), 'obj_deps' must be a dictionary of type 'source: list') output_dirmacros include_dirscflags)r r r Zextra_postargsdebug)r r )get isinstancelisttuplerrinfodictextendappendZcompilerZobject_filenamesZ build_temprcompiler Zcreate_static_libr)selfZ librariesZlib_nameZ build_inforrZ dependenciesZ global_depssourceZsrc_depsZ extra_depsZexpected_objectsr r r ZobjectsrA/usr/lib/python3.8/site-packages/setuptools/command/build_clib.pybuild_librariessv          zbuild_clib.build_librariesN)__name__ __module__ __qualname____doc__rrrrrrsr) Zdistutils.command.build_clibZcommandrZorigZdistutils.errorsrZ distutilsrZsetuptools.dep_utilrrrrrs   site-packages/setuptools/command/__pycache__/install.cpython-38.pyc000064400000007627151464452710021456 0ustar00U QabK@s|ddlmZddlZddlZddlZddlZddlmmZ ddl Z e jZ Gddde jZdde jj Dej e_ dS))DistutilsArgErrorNc@seZdZdZejjddgZejjddgZdddfd d dfgZe eZ d d Z d dZ ddZ ddZeddZddZdS)installz7Use easy_install to install the package, w/dependencies)old-and-unmanageableNzTry not to use this!)!single-version-externally-managedNz5used by system package builders to create 'flat' eggsrrZinstall_egg_infocCsdSNTselfrr>/usr/lib/python3.8/site-packages/setuptools/command/install.pyzinstall.Zinstall_scriptscCsdSrrrrrr r r cCstj|d|_d|_dSN)origrinitialize_optionsold_and_unmanageable!single_version_externally_managedrrrr r s zinstall.initialize_optionscCs8tj||jrd|_n|jr4|js4|js4tddS)NTzAYou must specify --record or --root when building system packages)rrfinalize_optionsrootrrecordrrrrr r%s  zinstall.finalize_optionscCs(|js |jrtj|Sd|_d|_dS)N)rrrrhandle_extra_pathZ path_fileZ extra_dirsrrrr r0s  zinstall.handle_extra_pathcCs@|js |jrtj|S|ts4tj|n|dSr ) rrrrrun_called_from_setupinspectZ currentframedo_egg_installrrrr r:s   z install.runcCsz|dkr4d}t|tdkr0d}t|dSt|d}|dd\}t|}|jdd }|d kox|j d kS) a Attempt to detect whether run() was called from setup() or by another command. If called by setup(), the parent caller will be the 'run_command' method in 'distutils.dist', and *its* caller will be the 'run_commands' method. If called any other way, the immediate caller *might* be 'run_command', but it won't have been called by 'run_commands'. Return True in that case or if a call stack is unavailable. Return False otherwise. Nz4Call stack not available. bdist_* commands may fail.Z IronPythonz6For best results, pass -X:Frames to enable call stack.T__name__rzdistutils.distZ run_commands) warningswarnplatformZpython_implementationrZgetouterframesZ getframeinfo f_globalsgetZfunction)Z run_framemsgresZcallerinfoZ caller_modulerrr rEs     zinstall._called_from_setupcCs|jd}||jd|j|jd}|d|_|jtd| d|j dj g}t j rp|dt j ||_|dt _ dS)N easy_installx)argsrr.z*.eggZ bdist_eggr)Z distributionZget_command_classrrZensure_finalizedZalways_copy_fromZ package_indexscanglobZ run_commandZget_command_objZ egg_output setuptoolsZbootstrap_install_frominsertr(r)r r&cmdr(rrr r`s"  zinstall.do_egg_installN)r __module__ __qualname____doc__rrZ user_optionsZboolean_options new_commandsdict_ncrrrr staticmethodrrrrrr rs(      rcCsg|]}|dtjkr|qS)r)rr4).0r.rrr {sr7)Zdistutils.errorsrrr+rr Zdistutils.command.installZcommandrrr,_installZ sub_commandsr2rrrr s lsite-packages/setuptools/command/__pycache__/install_scripts.cpython-38.opt-1.pyc000064400000004341151464452710024152 0ustar00U Qab @sXddlmZddlmmZddlZddlZddlm Z m Z m Z GdddejZdS))logN) Distribution PathMetadataensure_directoryc@s*eZdZdZddZddZd ddZd S) install_scriptsz;Do normal script install, plus any egg_info wrapper scriptscCstj|d|_dS)NF)origrinitialize_optionsno_ep)selfr F/usr/lib/python3.8/site-packages/setuptools/command/install_scripts.pyr s z"install_scripts.initialize_optionsc Csddlmm}|d|jjr2tj|ng|_ |j rBdS| d}t |j t|j |j|j|j}| d}t|dd}| d}t|dd}|j}|rd}|j}|tjkr|g}|}|j|} ||| D]} |j| qdS) Nregg_infoZ build_scripts executableZ bdist_wininstZ _is_runningFz python.exe)setuptools.command.easy_installcommandZ easy_installZ run_commandZ distributionZscriptsrrrunoutfilesr Zget_finalized_commandrZegg_baserr Zegg_nameZ egg_versiongetattrZ ScriptWriterZWindowsScriptWritersysrZbestZcommand_spec_classZ from_paramZget_argsZ as_header write_script) r ZeiZei_cmdZdistZbs_cmdZ exec_paramZbw_cmdZ is_wininstwritercmdargsr r r rs8        zinstall_scripts.runtc Gsddlm}m}td||jtj|j|}|j ||}|j s~t |t |d|} | || ||d|dS)z1Write an executable file to the scripts directoryr)chmod current_umaskzInstalling %s script to %swiN)rrrrinfoZ install_dirospathjoinrappendZdry_runropenwriteclose) r Z script_namecontentsmodeZignoredrrtargetmaskfr r r r3s  zinstall_scripts.write_scriptN)r)__name__ __module__ __qualname____doc__rrrr r r r r s#r) Z distutilsrZ!distutils.command.install_scriptsrrrrrZ pkg_resourcesrrrr r r r s site-packages/setuptools/command/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001342151464452710022472 0ustar00U QabR@szdddddddddd d d d d dddddddddgZddlmZddlZddlmZdejkrrdejd<ejd[[dS)alias bdist_eggZ bdist_rpmZ build_extZbuild_pyZdevelopZ easy_installZegg_infoZinstallZ install_librotateZsaveoptsZsdistZsetoptZtestZinstall_egg_infoinstall_scriptsregisterZ bdist_wininstZ upload_docsZuploadZ build_clibZ dist_info)bdistN)rZegg)rzPython .egg file) __all__Zdistutils.command.bdistrsysZsetuptools.commandrZformat_commandsZformat_commandappendr r ?/usr/lib/python3.8/site-packages/setuptools/command/__init__.pys<     site-packages/setuptools/command/__pycache__/alias.cpython-38.pyc000064400000004530151464452710021067 0ustar00U Qabz @sPddlmZddlmZddlmZmZmZddZGdddeZ dd Z d S) )DistutilsOptionError)map) edit_config option_base config_filecCs8dD]}||krt|Sq||gkr4t|S|S)z4Quote an argument for later parsing by shlex.split())"'\#)reprsplit)argcrs   4site-packages/setuptools/command/__pycache__/install_egg_info.cpython-38.pyc000064400000004543151464452710023305 0ustar00U Qab@s\ddlmZmZddlZddlmZddlmZddlmZddl Z Gdddej eZ dS))logdir_utilN)Command) namespaces)unpack_archivec@sBeZdZdZdZdgZddZddZddZd d Z d d Z d S)install_egg_infoz.Install an .egg-info directory for the package)z install-dir=dzdirectory to install tocCs d|_dSN) install_dirselfr G/usr/lib/python3.8/site-packages/setuptools/command/install_egg_info.pyinitialize_optionssz#install_egg_info.initialize_optionscCsV|dd|d}tdd|j|jd}|j|_tj |j ||_ g|_ dS)NZ install_lib)r r egg_infoz .egg-info)Zset_undefined_optionsZget_finalized_command pkg_resourcesZ DistributionZegg_nameZ egg_versionrsourceospathjoinr targetoutputs)r Zei_cmdbasenamer r rfinalize_optionss z!install_egg_info.finalize_optionscCs|dtj|jr:tj|js:tj|j|jdn(tj |jrb| tj |jfd|j|jstt |j| |jdd|j|jf|dS)Nr)dry_runz Removing r Copying %s to %s)Z run_commandrrisdirrislinkrZ remove_treerexistsZexecuteunlinkrZensure_directorycopytreerZinstall_namespacesr r r rrun!s  zinstall_egg_info.runcCs|jSr )rr r r r get_outputs.szinstall_egg_info.get_outputscs fdd}tjj|dS)NcsDdD] }||sd||krdSqj|td|||S)N)z.svn/zCVS//r) startswithrappendrdebug)srcZdstskipr r rskimmer3s  z*install_egg_info.copytree..skimmer)rrr)r r)r r rr 1s zinstall_egg_info.copytreeN) __name__ __module__ __qualname____doc__ descriptionZ user_optionsrrr!r"r r r r rr s  r) Z distutilsrrrZ setuptoolsrrZsetuptools.archive_utilrrZ Installerrr r r rs    site-packages/setuptools/command/__pycache__/install_scripts.cpython-38.pyc000064400000004341151464452710023213 0ustar00U Qab @sXddlmZddlmmZddlZddlZddlm Z m Z m Z GdddejZdS))logN) Distribution PathMetadataensure_directoryc@s*eZdZdZddZddZd ddZd S) install_scriptsz;Do normal script install, plus any egg_info wrapper scriptscCstj|d|_dS)NF)origrinitialize_optionsno_ep)selfr F/usr/lib/python3.8/site-packages/setuptools/command/install_scripts.pyr s z"install_scripts.initialize_optionsc Csddlmm}|d|jjr2tj|ng|_ |j rBdS| d}t |j t|j |j|j|j}| d}t|dd}| d}t|dd}|j}|rd}|j}|tjkr|g}|}|j|} ||| D]} |j| qdS) Nregg_infoZ build_scripts executableZ bdist_wininstZ _is_runningFz python.exe)setuptools.command.easy_installcommandZ easy_installZ run_commandZ distributionZscriptsrrrunoutfilesr Zget_finalized_commandrZegg_baserr Zegg_nameZ egg_versiongetattrZ ScriptWriterZWindowsScriptWritersysrZbestZcommand_spec_classZ from_paramZget_argsZ as_header write_script) r ZeiZei_cmdZdistZbs_cmdZ exec_paramZbw_cmdZ is_wininstwritercmdargsr r r rs8        zinstall_scripts.runtc Gsddlm}m}td||jtj|j|}|j ||}|j s~t |t |d|} | || ||d|dS)z1Write an executable file to the scripts directoryr)chmod current_umaskzInstalling %s script to %swiN)rrrrinfoZ install_dirospathjoinrappendZdry_runropenwriteclose) r Z script_namecontentsmodeZignoredrrtargetmaskfr r r r3s  zinstall_scripts.write_scriptN)r)__name__ __module__ __qualname____doc__rrrr r r r r s#r) Z distutilsrZ!distutils.command.install_scriptsrrrrrZ pkg_resourcesrrrr r r r s site-packages/setuptools/command/__pycache__/py36compat.cpython-38.pyc000064400000011002151464452710021773 0ustar00U Qabz@sdddlZddlmZddlmZddlmZddlmZGdddZe ejdr`Gd ddZdS) N)glob) convert_path)sdist)filterc@s\eZdZdZddZeddZddZdd Zd d Z d d Z ddZ ddZ ddZ dS)sdist_add_defaultsz Mix-in providing forward-compatibility for functionality as found in distutils on Python 3.7. Do not edit the code in this class except to update functionality as implemented in distutils. Instead, override in the subclass. cCs<|||||||dS)a9Add all the default files to self.filelist: - README or README.txt - setup.py - test/test*.py - all pure Python modules mentioned in setup script - all files pointed by package_data (build_py) - all files defined in data_files. - all files defined as scripts. - all C sources listed as part of extensions or C libraries in the setup script (doesn't catch C headers!) Warns if (README or README.txt) or setup.py are missing; everything else is optional. N)_add_defaults_standards_add_defaults_optional_add_defaults_python_add_defaults_data_files_add_defaults_ext_add_defaults_c_libs_add_defaults_scripts)selfrA/usr/lib/python3.8/site-packages/setuptools/command/py36compat.py add_defaultsszsdist_add_defaults.add_defaultscCs:tj|sdStj|}tj|\}}|t|kS)z Case-sensitive path existence check >>> sdist_add_defaults._cs_path_exists(__file__) True >>> sdist_add_defaults._cs_path_exists(__file__.upper()) False F)ospathexistsabspathsplitlistdir)fspathrZ directoryfilenamerrr_cs_path_exists(s  z"sdist_add_defaults._cs_path_existscCs|j|jjg}|D]~}t|trj|}d}|D]"}||r,d}|j|qPq,|s|dd |q||r|j|q|d|qdS)NFTz,standard file not found: should have one of z, zstandard file '%s' not found) ZREADMES distributionZ script_name isinstancetuplerfilelistappendwarnjoin)rZ standardsfnZaltsZgot_itrrrr9s"    z*sdist_add_defaults._add_defaults_standardscCs4ddg}|D]"}ttjjt|}|j|q dS)Nz test/test*.pyz setup.cfg)rrrisfilerrextend)rZoptionalpatternfilesrrrrNsz)sdist_add_defaults._add_defaults_optionalcCs\|d}|jr$|j||jD],\}}}}|D]}|jtj ||q:q*dS)Nbuild_py) get_finalized_commandrZhas_pure_modulesrr$get_source_files data_filesrrrr!)rr'ZpkgZsrc_dirZ build_dir filenamesrrrrr Ts   z'sdist_add_defaults._add_defaults_pythoncCsz|jrv|jjD]b}t|trBt|}tj|rt|j |q|\}}|D]$}t|}tj|rN|j |qNqdS)N) rZhas_data_filesr*rstrrrrr#rr)ritemdirnamer+frrrr ds     z+sdist_add_defaults._add_defaults_data_filescCs(|jr$|d}|j|dS)N build_ext)rZhas_ext_modulesr(rr$r))rr0rrrr us  z$sdist_add_defaults._add_defaults_extcCs(|jr$|d}|j|dS)N build_clib)rZhas_c_librariesr(rr$r))rr1rrrr zs  z'sdist_add_defaults._add_defaults_c_libscCs(|jr$|d}|j|dS)N build_scripts)rZ has_scriptsr(rr$r))rr2rrrr s  z(sdist_add_defaults._add_defaults_scriptsN)__name__ __module__ __qualname____doc__r staticmethodrrrr r r r r rrrrr s rrc@s eZdZdS)rN)r3r4r5rrrrrs) rrZdistutils.utilrZdistutils.commandrZsetuptools.extern.six.movesrrhasattrrrrrs    | site-packages/setuptools/command/__pycache__/dist_info.cpython-38.opt-1.pyc000064400000002521151464452710022711 0ustar00U Qab@s8dZddlZddlmZddlmZGdddeZdS)zD Create a dist_info directory As defined in the wheel specification N)Command)logc@s.eZdZdZdgZddZddZddZd S) dist_infozcreate a .dist-info directory)z egg-base=ezLdirectory containing .egg-info directories (default: top of the source tree)cCs d|_dSN)egg_baseselfr @/usr/lib/python3.8/site-packages/setuptools/command/dist_info.pyinitialize_optionsszdist_info.initialize_optionscCsdSrr rr r r finalize_optionsszdist_info.finalize_optionscCsn|d}|j|_|||jdtd d}tdt j ||d}| |j|dS)Negg_infoz .egg-infoz .dist-infoz creating '{}' bdist_wheel) Zget_finalized_commandrr runrlenrinfoformatospathabspathZegg2dist)r rZ dist_info_dirrr r r rs  z dist_info.runN)__name__ __module__ __qualname__ descriptionZ user_optionsr r rr r r r r s r)__doc__rZdistutils.corerZ distutilsrrr r r r s  site-packages/setuptools/command/__pycache__/build_py.cpython-38.pyc000064400000020701151464452710021603 0ustar00U Qab|%@sddlmZddlmZddlmmZddlZddlZddl Z ddl Z ddl Z ddl Z ddlmZddlmZmZmZzddlmZWn"ek rGdddZYnXGd d d ejeZdd d Zd dZdS))glob) convert_pathN)six)mapfilter filterfalse) Mixin2to3c@seZdZdddZdS)rTcCsdS)z do nothingN)selffilesZdoctestsr r ?/usr/lib/python3.8/site-packages/setuptools/command/build_py.pyrun_2to3szMixin2to3.run_2to3N)T)__name__ __module__ __qualname__r r r r r rsrc@seZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ ddZ ddZddZddZddZeddZd S)!build_pyaXEnhanced 'build_py' command that includes data files with packages The data files are specified via a 'package_data' argument to 'setup()'. See 'setuptools.dist.Distribution' for more details. Also, this version of the 'build_py' command allows you to specify both 'py_modules' and 'packages' in the same setup operation. cCsFtj||jj|_|jjp i|_d|jkr6|jd=g|_g|_dS)N data_files) origrfinalize_options distribution package_dataexclude_package_data__dict___build_py__updated_files_build_py__doctests_2to3r r r r r!s   zbuild_py.finalize_optionscCsx|js|jsdS|jr||jr4||||jd||jd||jd|t j j |dddS)z?Build modules, packages, and copy data files to build directoryNFTr)Zinclude_bytecode) Z py_modulespackagesZ build_modulesZbuild_packagesbuild_package_datar rrZ byte_compilerrZ get_outputsrr r r run+s z build_py.runcCs&|dkr||_|jStj||S)zlazily compute data filesr)_get_data_filesrrr __getattr__)r attrr r r r ?s zbuild_py.__getattr__cCsJtjrt|tjr|d}tj||||\}}|rB|j |||fS)N.) rZPY2 isinstanceZ string_typessplitrr build_modulerappend)r moduleZ module_filepackageZoutfilecopiedr r r r%Fs   zbuild_py.build_modulecCs|tt|j|jpdS)z?Generate list of '(package,src_dir,build_dir,filenames)' tuplesr )analyze_manifestlistr_get_pkg_data_filesrrr r r rPszbuild_py._get_data_filescsJ||tjj|jg|d}fdd||D}|||fS)Nr"csg|]}tj|qSr )ospathrelpath).0filesrc_dirr r ]sz0build_py._get_pkg_data_files..)get_package_dirr-r.joinZ build_libr$find_data_files)r r( build_dir filenamesr r2r r,Us    zbuild_py._get_pkg_data_filescCsX||j||}tt|}tj|}ttj j |}t|j |g|}| |||S)z6Return filenames for package's data files in 'src_dir')_get_platform_patternsrrr itertoolschain from_iterablerr-r.isfilemanifest_filesgetexclude_data_files)r r(r3patternsZglobs_expandedZ globs_matchesZ glob_filesr r r r r7cs   zbuild_py.find_data_filesc Cs|jD]|\}}}}|D]j}tj||}|tj|tj||}|||\}} tj|}| r||jj kr|j |qqdS)z$Copy data files into build directoryN) rr-r.r6ZmkpathdirnameZ copy_fileabspathrZconvert_2to3_doctestsrr&) r r(r3r8r9filenametargetZsrcfileZoutfr)r r r rts  zbuild_py.build_package_datac Csi|_}|jjsdSi}|jp"dD]}||t||<q$|d|d}|jj D]}t j t|\}}d}|} |r||kr||kr|}t j |\}} t j | |}qx||krX|dr|| krqX|||g|qXdS)Nr Zegg_infoz.py)r?rZinclude_package_datarassert_relativer5Z run_commandZget_finalized_commandZfilelistr r-r.r$r6endswith setdefaultr&) r ZmfZsrc_dirsr(Zei_cmdr.dfprevZoldfZdfr r r r*s(    zbuild_py.analyze_manifestcCsdSNr rr r r get_data_filesszbuild_py.get_data_filesc Csz |j|WStk r YnXtj|||}||j|<|rH|jjsL|S|jjD]}||ksn||drTqxqT|St |d}| }W5QRXd|krt j d|f|S)z8Check namespace packages' __init__ for declare_namespacer"rbsdeclare_namespacezNamespace package problem: %s is a namespace package, but its __init__.py does not call declare_namespace()! Please fix it. (See the setuptools manual under "Namespace Packages" for details.) ")packages_checkedKeyErrorrr check_packagerZnamespace_packages startswithioopenread distutilserrorsZDistutilsError)r r(Z package_dirZinit_pyZpkgrKcontentsr r r rRs*    zbuild_py.check_packagecCsi|_tj|dSrM)rPrrinitialize_optionsrr r r rZszbuild_py.initialize_optionscCs0tj||}|jjdk r,tj|jj|S|SrM)rrr5rZsrc_rootr-r.r6)r r(resr r r r5s zbuild_py.get_package_dircs\t||j||}fdd|D}tj|}t|fddD}tt|S)z6Filter filenames for package's data files in 'src_dir'c3s|]}t|VqdSrM)fnmatchrr0pattern)r r r sz.build_py.exclude_data_files..c3s|]}|kr|VqdSrMr )r0fn)badr r r_s)r+r:rr;r<r=set_unique_everseen)r r(r3r rBZ match_groupsZmatchesZkeepersr )rar r rAs   zbuild_py.exclude_data_filescs.t|dg||g}fdd|DS)z yield platform-specific path patterns (suitable for glob or fn_match) from a glob-based spec (such as self.package_data or self.exclude_package_data) matching package in src_dir. c3s |]}tjt|VqdSrM)r-r.r6rr]r2r r r_sz2build_py._get_platform_patterns..)r;r<r@)specr(r3Z raw_patternsr r2r r:s   zbuild_py._get_platform_patternsN)rrr__doc__rrr r%rr,r7rr*rNrRrZr5rA staticmethodr:r r r r rs"    rccsbt}|j}|dkr6t|j|D]}|||Vq n(|D]"}||}||kr:|||Vq:dS)zHList unique elements, preserving order. Remember all elements ever seen.N)rbaddr __contains__)iterablekeyseenZseen_addZelementkr r r rcs rccCs:tj|s|Sddlm}td|}||dS)Nr)DistutilsSetupErrorz Error: setup script specifies an absolute path: %s setup() arguments must *always* be /-separated paths relative to the setup.py directory, *never* absolute paths. )r-r.isabsdistutils.errorsrntextwrapdedentlstrip)r.rnmsgr r r rGs   rG)N)rZdistutils.utilrZdistutils.command.build_pyZcommandrrr-r\rqrTrprWr;Zsetuptools.externrZsetuptools.extern.six.movesrrrZsetuptools.lib2to3_exr ImportErrorrcrGr r r r s$   Y site-packages/setuptools/command/__pycache__/bdist_rpm.cpython-38.pyc000064400000003370151464452710021762 0ustar00U Qab@s(ddlmmZGdddejZdS)Nc@s eZdZdZddZddZdS) bdist_rpmaf Override the default bdist_rpm behavior to do the following: 1. Run egg_info to ensure the name and version are properly calculated. 2. Always run 'install' using --single-version-externally-managed to disable eggs in RPM distributions. 3. Replace dash with underscore in the version numbers for better RPM compatibility. cCs|dtj|dS)NZegg_info)Z run_commandorigrrun)selfr@/usr/lib/python3.8/site-packages/setuptools/command/bdist_rpm.pyrs z bdist_rpm.runcsl|j}|dd}tj|}d|d|fdd|D}|d}d|}||||S)N-_z%define version cs0g|](}|ddddddqS)zSource0: %{name}-%{version}.tarz)Source0: %{name}-%{unmangled_version}.tarzsetup.py install z5setup.py install --single-version-externally-managed z%setupz&%setup -n %{name}-%{unmangled_version})replace).0lineZline23Zline24rr s  z-bdist_rpm._make_spec_file..z%define unmangled_version )Z distributionZ get_versionr rr_make_spec_fileindexinsert)rversionZ rpmversionspecZ insert_locZunmangled_versionrr rrs      zbdist_rpm._make_spec_fileN)__name__ __module__ __qualname____doc__rrrrrrrs r)Zdistutils.command.bdist_rpmZcommandrrrrrrssite-packages/setuptools/command/__pycache__/sdist.cpython-38.opt-1.pyc000064400000015227151464452710022070 0ustar00U Qab@sddlmZddlmmZddlZddlZddlZddl Z ddl m Z ddl m Z ddlZeZd ddZGd d d e ejZdS) )logN)six)sdist_add_defaultsccs,tdD]}||D] }|Vqq dS)z%Find all files under revision controlzsetuptools.file_findersN) pkg_resourcesZiter_entry_pointsload)dirnameZepitemr )szsdist.cCs|d|d}|j|_|jtj|jd|| D]}||qD| t |j dg}|j D] }dd|f}||krp||qpdS)Negg_infoz SOURCES.txt dist_filesrr)Z run_commandget_finalized_commandfilelistappendospathjoinr check_readmeZget_sub_commandsmake_distributiongetattr distributionZ archive_files)selfZei_cmdZcmd_namerfiledatar r r run+s      z sdist.runcCstj||dS)N)origrinitialize_options_default_to_gztarr r r r r%>s zsdist.initialize_optionscCstjdkrdSdg|_dS)N)rZbetarZgztar)sys version_infoZformatsr'r r r r&Cs zsdist._default_to_gztarc Cs$|tj|W5QRXdS)z% Workaround for #516 N)_remove_os_linkr$rrr'r r r rIs zsdist.make_distributionc cs^Gddd}ttd|}zt`Wntk r6YnXz dVW5||k rXttd|XdS)zG In a context, remove and restore os.link if it exists c@s eZdZdS)z&sdist._remove_os_link..NoValueN)__name__ __module__ __qualname__r r r r NoValueWsr0linkN)rrr1 Exceptionsetattr)r0Zorig_valr r r r,Ps  zsdist._remove_os_linkcCsLztj|Wn6tk rFt\}}}|jjjd YnXdS)Ntemplate) r$r read_templater2r*exc_infotb_nexttb_framef_localsclose)r _tbr r r Z__read_template_hackes zsdist.__read_template_hack)r=)r(r)r(r)r(r=)r(r=rcs^|jrZ|d}|j||jjsZ|jD]&\}}}|jfdd|Dq2dS)zgetting python filesbuild_pycsg|]}tj|qSr )rrr)rfilenameZsrc_dirr r sz.sdist._add_defaults_python..N)rZhas_pure_modulesrrextendZget_source_filesZinclude_package_dataZ data_files)r r@r; filenamesr rBr _add_defaults_python|s  zsdist._add_defaults_pythoncsDz tjrt|n tWntk r>tdYnXdS)Nz&data_files contains unexpected objects)rZPY2r_add_defaults_data_filessuper TypeErrorrwarnr' __class__r r rGs  zsdist._add_defaults_data_filescCs8|jD]}tj|rdSq|dd|jdS)Nz,standard file not found: should have one of z, )READMESrrexistsrJr)r fr r r rs   zsdist.check_readmecCs^tj|||tj|d}ttdrJtj|rJt|| d|| d |dS)Nz setup.cfgr1r) r$rmake_release_treerrrhasattrrNunlinkZ copy_filerZsave_version_info)r Zbase_dirfilesdestr r r rPs   zsdist.make_release_treec Cs@tj|jsdSt|jd}|}W5QRX|dkS)NFrbz+# file GENERATED by distutils, do NOT edit )rrisfilemanifestioopenreadlineencode)r fpZ first_liner r r _manifest_is_not_generatedsz sdist._manifest_is_not_generatedc Cstd|jt|jd}|D]d}tjr^z|d}Wn&tk r\td|YqYnX| }| ds|svq|j |q| dS)zRead the manifest file (named by 'self.manifest') and use it to fill in 'self.filelist', the list of files to include in the source distribution. zreading manifest file '%s'rUzUTF-8z"%r not UTF-8 decodable -- skipping#N)rinforWrYrZPY3decodeUnicodeDecodeErrorrJstrip startswithrrr:)r rWliner r r read_manifests  zsdist.read_manifestcCs^|jd}|dd\}}|dkr2tddStj|sNtd|dS|j |dS)zyChecks if license_file' is configured and adds it to 'self.filelist' if the value contains a valid path. Zmetadata license_file)NNNz''license_file' option was not specifiedz8warning: Failed to find the configured license file '%s') rZget_option_dictgetrdebugrrrNrJrr)r Zoptsr;rfr r r check_licenses   zsdist.check_license)r-r.r/__doc__Z user_optionsZ negative_optZREADME_EXTENSIONStuplerMr#r%r&r staticmethod contextlibcontextmanagerr,Z_sdist__read_template_hackr*r+Zhas_leaky_handler5rFrGrrPr]reri __classcell__r r rKr rs<        r)r)Z distutilsrZdistutils.command.sdistZcommandrr$rr*rXrmZsetuptools.externrZ py36compatrrlistZ_default_revctrlr r r r r s    site-packages/setuptools/command/__pycache__/setopt.cpython-38.pyc000064400000010671151464452710021317 0ustar00U Qab@sddlmZddlmZddlmZddlZddlZddlmZddl m Z ddd d gZ dd dZ dddZ Gdd d e ZGdd d eZdS)) convert_path)log)DistutilsOptionErrorN) configparser)Command config_file edit_config option_basesetoptlocalcCsh|dkr dS|dkr,tjtjtjdS|dkrZtjdkrBdpDd}tjtd |St d |d S) zGet the filename of the distutils, local, global, or per-user config `kind` must be one of "local", "global", or "user" r z setup.cfgglobalz distutils.cfguserposix.z~/%spydistutils.cfgz7config_file() type must be 'local', 'global', or 'user'N) ospathjoindirname distutils__file__name expanduserr ValueError)Zkinddotr=/usr/lib/python3.8/site-packages/setuptools/command/setopt.pyrs Fc Cs&td|t}||g|D]\}}|dkrRtd||||q(||sttd||| ||D]p\}}|dkrtd|||| ||| |std||||q|td||||| |||q|q(td||s"t |d }||W5QRXdS) aYEdit a configuration file to include `settings` `settings` is a dictionary of dictionaries or ``None`` values, keyed by command/section name. A ``None`` value means to delete the entire section, while a dictionary lists settings to be changed or deleted in that section. A setting of ``None`` means to delete that setting. zReading configuration from %sNzDeleting section [%s] from %szAdding new section [%s] to %szDeleting %s.%s from %sz#Deleting empty [%s] section from %szSetting %s.%s to %r in %sz Writing %sw)rdebugrZRawConfigParserreaditemsinfoZremove_sectionZ has_sectionZ add_sectionZ remove_optionoptionssetopenwrite) filenameZsettingsdry_runZoptsZsectionr"optionvaluefrrrr!sJ          c@s2eZdZdZdddgZddgZddZd d Zd S) r z|js>tddS)Nz%Must specify --command *and* --optionz$Must specify --set-value or --remove)r r9rEr(rrFrCr2rrrr9s  zsetopt.finalize_optionscCs*t|j|j|jdd|jii|jdS)N-_)rr&rEr(replacerFr'r2rrrrunsz setopt.runN) r:r;r<r= descriptionr r>r?r4r9rJrrrrr ss )r )F)Zdistutils.utilrrrZdistutils.errorsrrZsetuptools.extern.six.movesrZ setuptoolsr__all__rrr r rrrrs        +'site-packages/setuptools/command/__pycache__/build_ext.cpython-38.opt-1.pyc000064400000023211151464452710022711 0ustar00U Qab2 @sddlZddlZddlZddlmZddlmZddlm Z ddl m Z m Z ddl mZddlmZddlmZdd lmZejrddlZd d eDZn dd lmZzddlmZed Wnek reZYnXe dddl mZddZ dZ!dZ"dZ#ej$dkrdZ"n>ej%dkrTzddl&Z&e'e&dZ"Z!Wnek rRYnXddZ(ddZ)GdddeZe"sej%dkrd"ddZ*nd Z#d#d!dZ*dS)$N) build_ext) copy_file) new_compiler)customize_compilerget_config_var)DistutilsError)log)Library)sixcCs g|]\}}}|tjkr|qS)impZ C_EXTENSION).0s_tpr r @/usr/lib/python3.8/site-packages/setuptools/command/build_ext.py s rEXTENSION_SUFFIXESzCython.Compiler.MainLDSHARED) _config_varsc CsZtjdkrNt}z$dtd<dtd<dtd<t|W5tt|Xnt|dS)Ndarwinz0gcc -Wl,-x -dynamiclib -undefined dynamic_lookuprz -dynamiclibZCCSHAREDz.dylibZSO)sysplatform _CONFIG_VARScopyclearupdater)compilerZtmpr r r_customize_compiler_for_shlib#s  rFZsharedrTntRTLD_NOWcCs tr|SdS)N) have_rtld)rr r rDr$cCs.tD]$}d|kr|S|dkr|SqdS)z;Return the file extension for an abi3-compliant Extension()z.abi3z.pydNr)suffixr r rget_abi3_suffixGs r'c@sveZdZddZddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ ddZdddZdS)rcCs.|jd}|_t|||_|r*|dS)z;Build extensions in build directory, then copy if --inplacerN)Zinplace _build_extruncopy_extensions_to_source)selfZ old_inplacer r rr)Qs  z build_ext.runc Cs|d}|jD]}||j}||}|d}d|dd}||}tj |tj |}tj |j |} t | ||j |jd|jr||ptj|dqdS)Nbuild_py.)verbosedry_runT)get_finalized_command extensionsget_ext_fullnamenameget_ext_filenamesplitjoinZget_package_dirospathbasename build_librr/r0 _needs_stub write_stubcurdir) r+r,extfullnamefilenamemodpathpackageZ package_dirZ dest_filenameZ src_filenamer r rr*Ys&       z#build_ext.copy_extensions_to_sourcecCst||}||jkr|j|}tjo4t|do4t}|r^td}|dt| }|t}t |t rt j |\}}|j|tStr|jrt j |\}}t j |d|S|S)NZpy_limited_apiZ EXT_SUFFIXzdl-)r(r5ext_mapr ZPY3getattrr'rlen isinstancer r8r9splitextshlib_compilerlibrary_filenamelibtype use_stubs_links_to_dynamicr6r7)r+r@rAr?Zuse_abi3Zso_extfndr r rr5os&      zbuild_ext.get_ext_filenamecCs t|d|_g|_i|_dSN)r(initialize_optionsrIshlibsrDr+r r rrQs zbuild_ext.initialize_optionscCs,t||jpg|_||jdd|jD|_|jrB||jD]}||j|_qH|jD]}|j}||j |<||j | dd<|jr| |pd}|ot ot |t }||_||_||}|_tjtj|j|}|r||jkr|j||rbt rbtj|jkrb|jtjqbdS)NcSsg|]}t|tr|qSr )rGr r r?r r rrs z.build_ext.finalize_options..r-r.F)r(finalize_optionsr2Zcheck_extensions_listrRsetup_shlib_compilerr3r4 _full_namerDr6links_to_dynamicrLrGr rMr<r5 _file_namer8r9dirnamer7r; library_dirsappendr>runtime_library_dirs)r+r?r@ZltdnsrAZlibdirr r rrUs,       zbuild_ext.finalize_optionscCst|j|j|jd}|_t||jdk r8||j|jdk r^|jD]\}}| ||qH|j dk r~|j D]}| |qn|j dk r| |j |jdk r||j|jdk r||j|jdk r||jt||_dS)N)rr0force)rrr0r_rIrZ include_dirsZset_include_dirsZdefineZ define_macroZundefZundefine_macro librariesZ set_librariesr[Zset_library_dirsZrpathZset_runtime_library_dirsZ link_objectsZset_link_objectslink_shared_object__get__)r+rr4valueZmacror r rrVs.               zbuild_ext.setup_shlib_compilercCst|tr|jSt||SrP)rGr export_symbolsr(get_export_symbolsr+r?r r rres zbuild_ext.get_export_symbolscCs\||j}z@t|tr"|j|_t|||jrL|dj }| ||W5||_XdS)Nr,) Z_convert_pyx_sources_to_langrrGr rIr(build_extensionr<r1r;r=)r+r?Z _compilercmdr r rrgs   zbuild_ext.build_extensioncsPtdd|jDd|jddddgtfdd|jDS) z?Return true if 'ext' links to a dynamic lib in the same packagecSsg|] }|jqSr )rW)r libr r rrsz.build_ext.links_to_dynamic..r-Nr.r"c3s|]}|kVqdSrPr )r ZlibnameZlibnamesZpkgr r sz-build_ext.links_to_dynamic..)dictfromkeysrRr7rWr6anyr`rfr rjrrXs zbuild_ext.links_to_dynamiccCst||SrP)r( get_outputs_build_ext__get_stubs_outputsrSr r rroszbuild_ext.get_outputscs6fddjD}t|}tdd|DS)Nc3s0|](}|jrtjjjf|jdVqdS)r-N)r<r8r9r7r;rWr6rTrSr rrksz0build_ext.__get_stubs_outputs..css|]\}}||VqdSrPr )r baseZfnextr r rrks)r2 itertoolsproduct!_build_ext__get_output_extensionslist)r+Z ns_ext_basesZpairsr rSrZ__get_stubs_outputss  zbuild_ext.__get_stubs_outputsccs"dVdV|djrdVdS)N.pyz.pycr,z.pyo)r1optimizerSr r rZ__get_output_extensionss z!build_ext.__get_output_extensionsFcCs,td|j|tjj|f|jdd}|rJtj|rJt|d|j st |d}| dddd t d d tj |jd d dt ddddt dddt ddddg||r(ddlm}||gdd|j d|dj}|dkr||g|d|j dtj|r(|j s(t|dS)Nz writing stub loader for %s to %sr-rvz already exists! Please delete.w zdef __bootstrap__():z- global __bootstrap__, __file__, __loader__z% import sys, os, pkg_resources, impz, dlz: __file__ = pkg_resources.resource_filename(__name__,%r)z del __bootstrap__z if '__loader__' in globals():z del __loader__z# old_flags = sys.getdlopenflags()z old_dir = os.getcwd()z try:z( os.chdir(os.path.dirname(__file__))z$ sys.setdlopenflags(dl.RTLD_NOW)z( imp.load_dynamic(__name__,__file__)z finally:z" sys.setdlopenflags(old_flags)z os.chdir(old_dir)z__bootstrap__()r"r) byte_compileT)rwr_r0Z install_lib)rinforWr8r9r7r6existsrr0openwriteif_dlr:rYcloseZdistutils.utilrzr1rwunlink)r+ output_dirr?compileZ stub_filefrzrwr r rr=sb        zbuild_ext.write_stubN)F)__name__ __module__ __qualname__r)r*r5rQrUrVrergrXrorprtr=r r r rrPs   rc Cs(||j||||||||| | | | dSrP)linkZSHARED_LIBRARY) r+objectsoutput_libnamerr`r[r]rddebug extra_preargsextra_postargs build_temp target_langr r rra#sraZstaticc CsRtj|\}} tj| \}}|ddr<|dd}|||||| dS)Nxri)r8r9r6rHrJ startswithZcreate_static_lib)r+rrrr`r[r]rdrrrrrrAr:r?r r rra2s  ) NNNNNrNNNN) NNNNNrNNNN)+r8rrrZdistutils.command.build_extrZ _du_build_extZdistutils.file_utilrZdistutils.ccompilerrZdistutils.sysconfigrrZdistutils.errorsrZ distutilsrZsetuptools.extensionr Zsetuptools.externr ZPY2r Z get_suffixesrZimportlib.machineryZCython.Distutils.build_extr( __import__ ImportErrorrrrr#rLrKrr4Zdlhasattrrr'rar r r rsz               Q site-packages/setuptools/command/__pycache__/install.cpython-38.opt-1.pyc000064400000007627151464452710022415 0ustar00U QabK@s|ddlmZddlZddlZddlZddlZddlmmZ ddl Z e jZ Gddde jZdde jj Dej e_ dS))DistutilsArgErrorNc@seZdZdZejjddgZejjddgZdddfd d dfgZe eZ d d Z d dZ ddZ ddZeddZddZdS)installz7Use easy_install to install the package, w/dependencies)old-and-unmanageableNzTry not to use this!)!single-version-externally-managedNz5used by system package builders to create 'flat' eggsrrZinstall_egg_infocCsdSNTselfrr>/usr/lib/python3.8/site-packages/setuptools/command/install.pyzinstall.Zinstall_scriptscCsdSrrrrrr r r cCstj|d|_d|_dSN)origrinitialize_optionsold_and_unmanageable!single_version_externally_managedrrrr r s zinstall.initialize_optionscCs8tj||jrd|_n|jr4|js4|js4tddS)NTzAYou must specify --record or --root when building system packages)rrfinalize_optionsrootrrecordrrrrr r%s  zinstall.finalize_optionscCs(|js |jrtj|Sd|_d|_dS)N)rrrrhandle_extra_pathZ path_fileZ extra_dirsrrrr r0s  zinstall.handle_extra_pathcCs@|js |jrtj|S|ts4tj|n|dSr ) rrrrrun_called_from_setupinspectZ currentframedo_egg_installrrrr r:s   z install.runcCsz|dkr4d}t|tdkr0d}t|dSt|d}|dd\}t|}|jdd }|d kox|j d kS) a Attempt to detect whether run() was called from setup() or by another command. If called by setup(), the parent caller will be the 'run_command' method in 'distutils.dist', and *its* caller will be the 'run_commands' method. If called any other way, the immediate caller *might* be 'run_command', but it won't have been called by 'run_commands'. Return True in that case or if a call stack is unavailable. Return False otherwise. Nz4Call stack not available. bdist_* commands may fail.Z IronPythonz6For best results, pass -X:Frames to enable call stack.T__name__rzdistutils.distZ run_commands) warningswarnplatformZpython_implementationrZgetouterframesZ getframeinfo f_globalsgetZfunction)Z run_framemsgresZcallerinfoZ caller_modulerrr rEs     zinstall._called_from_setupcCs|jd}||jd|j|jd}|d|_|jtd| d|j dj g}t j rp|dt j ||_|dt _ dS)N easy_installx)argsrr.z*.eggZ bdist_eggr)Z distributionZget_command_classrrZensure_finalizedZalways_copy_fromZ package_indexscanglobZ run_commandZget_command_objZ egg_output setuptoolsZbootstrap_install_frominsertr(r)r r&cmdr(rrr r`s"  zinstall.do_egg_installN)r __module__ __qualname____doc__rrZ user_optionsZboolean_options new_commandsdict_ncrrrr staticmethodrrrrrr rs(      rcCsg|]}|dtjkr|qS)r)rr4).0r.rrr {sr7)Zdistutils.errorsrrr+rr Zdistutils.command.installZcommandrrr,_installZ sub_commandsr2rrrr s lsite-packages/setuptools/command/__pycache__/saveopts.cpython-38.opt-1.pyc000064400000001571151464452710022603 0ustar00U Qab@s$ddlmZmZGdddeZdS)) edit_config option_basec@seZdZdZdZddZdS)saveoptsz#Save command-line options to a filez7save supplied options to setup.cfg or other config filecCsh|j}i}|jD]B}|dkrq||D]$\}\}}|dkr,|||i|<q,qt|j||jdS)Nrz command line)Z distributionZcommand_optionsZget_option_dictitems setdefaultrfilenameZdry_run)selfZdistZsettingscmdoptsrcvalr ?/usr/lib/python3.8/site-packages/setuptools/command/saveopts.pyrun s z saveopts.runN)__name__ __module__ __qualname____doc__ descriptionrr r r rrsrN)Zsetuptools.command.setoptrrrr r r rssite-packages/setuptools/command/__pycache__/develop.cpython-38.opt-1.pyc000064400000014545151464452710022402 0ustar00U Qab@sddlmZddlmZddlmZmZddlZddlZddl Z ddl m Z ddl Z ddl mZddlmZddlZeZGdd d ejeZGd d d ZdS) ) convert_path)log)DistutilsErrorDistutilsOptionErrorN)six) easy_install) namespacesc@sveZdZdZdZejddgZejdgZdZddZ d d Z d d Z e d dZ ddZddZddZddZdS)developzSet up package for developmentz%install package in 'development mode') uninstalluzUninstall this source package)z egg-path=Nz-Set the path to be used in the .egg-link filer FcCs2|jrd|_||n||dS)NT)r Z multi_versionuninstall_linkZuninstall_namespacesinstall_for_developmentZwarn_deprecated_optionsselfr>/usr/lib/python3.8/site-packages/setuptools/command/develop.pyrun s  z develop.runcCs&d|_d|_t|d|_d|_dS)N.)r egg_pathrinitialize_options setup_pathZalways_copy_fromrrrrr)s  zdevelop.initialize_optionscCs|d}|jr,d}|j|jf}t|||jg|_t||| |j t d|jd}t j|j||_|j|_|jdkrt j|j|_t|j}tt j|j|j}||krtd|tj|t|t j|j|jd|_||j|j|j|_dS)Negg_infoz-Please rename %r to %r before using 'develop'z*.eggz .egg-linkzA--egg-path must be a relative path from the install directory to Z project_name)get_finalized_commandZbroken_egg_inforrZegg_nameargsrfinalize_optionsZexpand_basedirsZ expand_dirsZ package_indexscanglobospathjoin install_diregg_linkegg_baserabspath pkg_resourcesnormalize_pathrZ Distribution PathMetadatadist_resolve_setup_pathr)rZeitemplaterZ egg_link_fntargetrrrrr0sF        zdevelop.finalize_optionscCsn|tjdd}|tjkr0d|dd}ttj |||}|ttjkrjt d|ttj|S)z Generate a path from egg_base back to '.' where the setup script resides and ensure that path points to the setup path from $install_dir/$egg_path. /z../zGCan't get a consistent path to setup script from installation directory) replacerseprstripcurdircountr%r&rr r)r#r!rZ path_to_setupZresolvedrrrr)Zs  zdevelop._resolve_setup_pathc CsHtjrt|jddr|jddd|d|d}t|j }|jd|d|d|jddd|d|d}||_ ||j _ t ||j|j _n"|d|jdd d|d|tjr|tjdt_|td |j|j|js0t|jd }||j d |jW5QRX|d|j |j dS) NZuse_2to3FZbuild_pyr)Zinplacer)r#Z build_extr-zCreating %s (link to %s)w )rZPY3getattr distributionZreinitialize_commandZ run_commandrr%r&Z build_librr(locationr'rZ _providerZinstall_site_py setuptoolsZbootstrap_install_fromrZinstall_namespacesrinfor"r#dry_runopenwriterZprocess_distributionZno_deps)rZbpy_cmdZ build_pathZei_cmdfrrrr ns:           zdevelop.install_for_developmentcCstj|jrztd|j|jt|j}dd|D}|||j g|j |j gfkrht d|dS|j szt |j|j s||j|jjrt ddS)NzRemoving %s (link to %s)cSsg|] }|qSr)r0).0linerrr sz*develop.uninstall_link..z$Link points to %s: uninstall abortedz5Note: you must uninstall or replace scripts manually!)rrexistsr"rr9r#r;closerrwarnr:unlinkZ update_pthr(r6scripts)rZ egg_link_filecontentsrrrr s      zdevelop.uninstall_linkc Cs||jk rt||S|||jjp*gD]N}tjt |}tj |}t |}| }W5QRX|||||q,dSN)r(rinstall_egg_scriptsinstall_wrapper_scriptsr6rErrr$rbasenameior;readZinstall_script)rr(Z script_nameZ script_pathZstrmZ script_textrrrrHs     zdevelop.install_egg_scriptscCst|}t||SrG)VersionlessRequirementrrIrr(rrrrIszdevelop.install_wrapper_scriptsN)__name__ __module__ __qualname____doc__ descriptionrZ user_optionsZboolean_optionsZcommand_consumes_argumentsrrr staticmethodr)r r rHrIrrrrr s"  * 0r c@s(eZdZdZddZddZddZdS) rMa Adapt a pkg_resources.Distribution to simply return the project name as the 'requirement' so that scripts will work across multiple versions. >>> from pkg_resources import Distribution >>> dist = Distribution(project_name='foo', version='1.0') >>> str(dist.as_requirement()) 'foo==1.0' >>> adapted_dist = VersionlessRequirement(dist) >>> str(adapted_dist.as_requirement()) 'foo' cCs ||_dSrG)_VersionlessRequirement__distrNrrr__init__szVersionlessRequirement.__init__cCs t|j|SrG)r5rU)rnamerrr __getattr__sz"VersionlessRequirement.__getattr__cCs|jSrGrrrrras_requirementsz%VersionlessRequirement.as_requirementN)rOrPrQrRrVrXrYrrrrrMsrM)Zdistutils.utilrZ distutilsrZdistutils.errorsrrrrrKZsetuptools.externrr%Zsetuptools.command.easy_installrr8rtypeZ __metaclass__ZDevelopInstallerr rMrrrrs     6site-packages/setuptools/command/bdist_wininst.py000064400000001175151464452710016372 0ustar00import distutils.command.bdist_wininst as orig class bdist_wininst(orig.bdist_wininst): def reinitialize_command(self, command, reinit_subcommands=0): """ Supplement reinitialize_command to work around http://bugs.python.org/issue20819 """ cmd = self.distribution.reinitialize_command( command, reinit_subcommands) if command in ('install', 'install_lib'): cmd.install_lib = None return cmd def run(self): self._is_running = True try: orig.bdist_wininst.run(self) finally: self._is_running = False site-packages/setuptools/command/upload_docs.py000064400000016217151464452710016011 0ustar00# -*- coding: utf-8 -*- """upload_docs Implements a Distutils 'upload_docs' subcommand (upload documentation to PyPI's pythonhosted.org). """ from base64 import standard_b64encode from distutils import log from distutils.errors import DistutilsOptionError import os import socket import zipfile import tempfile import shutil import itertools import functools from setuptools.extern import six from setuptools.extern.six.moves import http_client, urllib from pkg_resources import iter_entry_points from .upload import upload def _encode(s): errors = 'surrogateescape' if six.PY3 else 'strict' return s.encode('utf-8', errors) class upload_docs(upload): # override the default repository as upload_docs isn't # supported by Warehouse (and won't be). DEFAULT_REPOSITORY = 'https://pypi.python.org/pypi/' description = 'Upload documentation to PyPI' user_options = [ ('repository=', 'r', "url of repository [default: %s]" % upload.DEFAULT_REPOSITORY), ('show-response', None, 'display full response text from server'), ('upload-dir=', None, 'directory to upload'), ] boolean_options = upload.boolean_options def has_sphinx(self): if self.upload_dir is None: for ep in iter_entry_points('distutils.commands', 'build_sphinx'): return True sub_commands = [('build_sphinx', has_sphinx)] def initialize_options(self): upload.initialize_options(self) self.upload_dir = None self.target_dir = None def finalize_options(self): upload.finalize_options(self) if self.upload_dir is None: if self.has_sphinx(): build_sphinx = self.get_finalized_command('build_sphinx') self.target_dir = build_sphinx.builder_target_dir else: build = self.get_finalized_command('build') self.target_dir = os.path.join(build.build_base, 'docs') else: self.ensure_dirname('upload_dir') self.target_dir = self.upload_dir if 'pypi.python.org' in self.repository: log.warn("Upload_docs command is deprecated. Use RTD instead.") self.announce('Using upload directory %s' % self.target_dir) def create_zipfile(self, filename): zip_file = zipfile.ZipFile(filename, "w") try: self.mkpath(self.target_dir) # just in case for root, dirs, files in os.walk(self.target_dir): if root == self.target_dir and not files: tmpl = "no files found in upload directory '%s'" raise DistutilsOptionError(tmpl % self.target_dir) for name in files: full = os.path.join(root, name) relative = root[len(self.target_dir):].lstrip(os.path.sep) dest = os.path.join(relative, name) zip_file.write(full, dest) finally: zip_file.close() def run(self): # Run sub commands for cmd_name in self.get_sub_commands(): self.run_command(cmd_name) tmp_dir = tempfile.mkdtemp() name = self.distribution.metadata.get_name() zip_file = os.path.join(tmp_dir, "%s.zip" % name) try: self.create_zipfile(zip_file) self.upload_file(zip_file) finally: shutil.rmtree(tmp_dir) @staticmethod def _build_part(item, sep_boundary): key, values = item title = '\nContent-Disposition: form-data; name="%s"' % key # handle multiple entries for the same name if not isinstance(values, list): values = [values] for value in values: if isinstance(value, tuple): title += '; filename="%s"' % value[0] value = value[1] else: value = _encode(value) yield sep_boundary yield _encode(title) yield b"\n\n" yield value if value and value[-1:] == b'\r': yield b'\n' # write an extra newline (lurve Macs) @classmethod def _build_multipart(cls, data): """ Build up the MIME payload for the POST data """ boundary = b'--------------GHSKFJDLGDS7543FJKLFHRE75642756743254' sep_boundary = b'\n--' + boundary end_boundary = sep_boundary + b'--' end_items = end_boundary, b"\n", builder = functools.partial( cls._build_part, sep_boundary=sep_boundary, ) part_groups = map(builder, data.items()) parts = itertools.chain.from_iterable(part_groups) body_items = itertools.chain(parts, end_items) content_type = 'multipart/form-data; boundary=%s' % boundary.decode('ascii') return b''.join(body_items), content_type def upload_file(self, filename): with open(filename, 'rb') as f: content = f.read() meta = self.distribution.metadata data = { ':action': 'doc_upload', 'name': meta.get_name(), 'content': (os.path.basename(filename), content), } # set up the authentication credentials = _encode(self.username + ':' + self.password) credentials = standard_b64encode(credentials) if six.PY3: credentials = credentials.decode('ascii') auth = "Basic " + credentials body, ct = self._build_multipart(data) msg = "Submitting documentation to %s" % (self.repository) self.announce(msg, log.INFO) # build the Request # We can't use urllib2 since we need to send the Basic # auth right with the first request schema, netloc, url, params, query, fragments = \ urllib.parse.urlparse(self.repository) assert not params and not query and not fragments if schema == 'http': conn = http_client.HTTPConnection(netloc) elif schema == 'https': conn = http_client.HTTPSConnection(netloc) else: raise AssertionError("unsupported schema " + schema) data = '' try: conn.connect() conn.putrequest("POST", url) content_type = ct conn.putheader('Content-type', content_type) conn.putheader('Content-length', str(len(body))) conn.putheader('Authorization', auth) conn.endheaders() conn.send(body) except socket.error as e: self.announce(str(e), log.ERROR) return r = conn.getresponse() if r.status == 200: msg = 'Server response (%s): %s' % (r.status, r.reason) self.announce(msg, log.INFO) elif r.status == 301: location = r.getheader('Location') if location is None: location = 'https://pythonhosted.org/%s/' % meta.get_name() msg = 'Upload successful. Visit %s' % location self.announce(msg, log.INFO) else: msg = 'Upload failed (%s): %s' % (r.status, r.reason) self.announce(msg, log.ERROR) if self.show_response: print('-' * 75, r.read(), '-' * 75) site-packages/setuptools/command/py36compat.py000064400000011572151464452710015521 0ustar00import os from glob import glob from distutils.util import convert_path from distutils.command import sdist from setuptools.extern.six.moves import filter class sdist_add_defaults: """ Mix-in providing forward-compatibility for functionality as found in distutils on Python 3.7. Do not edit the code in this class except to update functionality as implemented in distutils. Instead, override in the subclass. """ def add_defaults(self): """Add all the default files to self.filelist: - README or README.txt - setup.py - test/test*.py - all pure Python modules mentioned in setup script - all files pointed by package_data (build_py) - all files defined in data_files. - all files defined as scripts. - all C sources listed as part of extensions or C libraries in the setup script (doesn't catch C headers!) Warns if (README or README.txt) or setup.py are missing; everything else is optional. """ self._add_defaults_standards() self._add_defaults_optional() self._add_defaults_python() self._add_defaults_data_files() self._add_defaults_ext() self._add_defaults_c_libs() self._add_defaults_scripts() @staticmethod def _cs_path_exists(fspath): """ Case-sensitive path existence check >>> sdist_add_defaults._cs_path_exists(__file__) True >>> sdist_add_defaults._cs_path_exists(__file__.upper()) False """ if not os.path.exists(fspath): return False # make absolute so we always have a directory abspath = os.path.abspath(fspath) directory, filename = os.path.split(abspath) return filename in os.listdir(directory) def _add_defaults_standards(self): standards = [self.READMES, self.distribution.script_name] for fn in standards: if isinstance(fn, tuple): alts = fn got_it = False for fn in alts: if self._cs_path_exists(fn): got_it = True self.filelist.append(fn) break if not got_it: self.warn("standard file not found: should have one of " + ', '.join(alts)) else: if self._cs_path_exists(fn): self.filelist.append(fn) else: self.warn("standard file '%s' not found" % fn) def _add_defaults_optional(self): optional = ['test/test*.py', 'setup.cfg'] for pattern in optional: files = filter(os.path.isfile, glob(pattern)) self.filelist.extend(files) def _add_defaults_python(self): # build_py is used to get: # - python modules # - files defined in package_data build_py = self.get_finalized_command('build_py') # getting python files if self.distribution.has_pure_modules(): self.filelist.extend(build_py.get_source_files()) # getting package_data files # (computed in build_py.data_files by build_py.finalize_options) for pkg, src_dir, build_dir, filenames in build_py.data_files: for filename in filenames: self.filelist.append(os.path.join(src_dir, filename)) def _add_defaults_data_files(self): # getting distribution.data_files if self.distribution.has_data_files(): for item in self.distribution.data_files: if isinstance(item, str): # plain file item = convert_path(item) if os.path.isfile(item): self.filelist.append(item) else: # a (dirname, filenames) tuple dirname, filenames = item for f in filenames: f = convert_path(f) if os.path.isfile(f): self.filelist.append(f) def _add_defaults_ext(self): if self.distribution.has_ext_modules(): build_ext = self.get_finalized_command('build_ext') self.filelist.extend(build_ext.get_source_files()) def _add_defaults_c_libs(self): if self.distribution.has_c_libraries(): build_clib = self.get_finalized_command('build_clib') self.filelist.extend(build_clib.get_source_files()) def _add_defaults_scripts(self): if self.distribution.has_scripts(): build_scripts = self.get_finalized_command('build_scripts') self.filelist.extend(build_scripts.get_source_files()) if hasattr(sdist.sdist, '_add_defaults_standards'): # disable the functionality already available upstream class sdist_add_defaults: pass site-packages/setuptools/command/setopt.py000064400000011735151464452710015033 0ustar00from distutils.util import convert_path from distutils import log from distutils.errors import DistutilsOptionError import distutils import os from setuptools.extern.six.moves import configparser from setuptools import Command __all__ = ['config_file', 'edit_config', 'option_base', 'setopt'] def config_file(kind="local"): """Get the filename of the distutils, local, global, or per-user config `kind` must be one of "local", "global", or "user" """ if kind == 'local': return 'setup.cfg' if kind == 'global': return os.path.join( os.path.dirname(distutils.__file__), 'distutils.cfg' ) if kind == 'user': dot = os.name == 'posix' and '.' or '' return os.path.expanduser(convert_path("~/%spydistutils.cfg" % dot)) raise ValueError( "config_file() type must be 'local', 'global', or 'user'", kind ) def edit_config(filename, settings, dry_run=False): """Edit a configuration file to include `settings` `settings` is a dictionary of dictionaries or ``None`` values, keyed by command/section name. A ``None`` value means to delete the entire section, while a dictionary lists settings to be changed or deleted in that section. A setting of ``None`` means to delete that setting. """ log.debug("Reading configuration from %s", filename) opts = configparser.RawConfigParser() opts.read([filename]) for section, options in settings.items(): if options is None: log.info("Deleting section [%s] from %s", section, filename) opts.remove_section(section) else: if not opts.has_section(section): log.debug("Adding new section [%s] to %s", section, filename) opts.add_section(section) for option, value in options.items(): if value is None: log.debug( "Deleting %s.%s from %s", section, option, filename ) opts.remove_option(section, option) if not opts.options(section): log.info("Deleting empty [%s] section from %s", section, filename) opts.remove_section(section) else: log.debug( "Setting %s.%s to %r in %s", section, option, value, filename ) opts.set(section, option, value) log.info("Writing %s", filename) if not dry_run: with open(filename, 'w') as f: opts.write(f) class option_base(Command): """Abstract base class for commands that mess with config files""" user_options = [ ('global-config', 'g', "save options to the site-wide distutils.cfg file"), ('user-config', 'u', "save options to the current user's pydistutils.cfg file"), ('filename=', 'f', "configuration file to use (default=setup.cfg)"), ] boolean_options = [ 'global-config', 'user-config', ] def initialize_options(self): self.global_config = None self.user_config = None self.filename = None def finalize_options(self): filenames = [] if self.global_config: filenames.append(config_file('global')) if self.user_config: filenames.append(config_file('user')) if self.filename is not None: filenames.append(self.filename) if not filenames: filenames.append(config_file('local')) if len(filenames) > 1: raise DistutilsOptionError( "Must specify only one configuration file option", filenames ) self.filename, = filenames class setopt(option_base): """Save command-line options to a file""" description = "set an option in setup.cfg or another config file" user_options = [ ('command=', 'c', 'command to set an option for'), ('option=', 'o', 'option to set'), ('set-value=', 's', 'value of the option'), ('remove', 'r', 'remove (unset) the value'), ] + option_base.user_options boolean_options = option_base.boolean_options + ['remove'] def initialize_options(self): option_base.initialize_options(self) self.command = None self.option = None self.set_value = None self.remove = None def finalize_options(self): option_base.finalize_options(self) if self.command is None or self.option is None: raise DistutilsOptionError("Must specify --command *and* --option") if self.set_value is None and not self.remove: raise DistutilsOptionError("Must specify --set-value or --remove") def run(self): edit_config( self.filename, { self.command: {self.option.replace('-', '_'): self.set_value} }, self.dry_run ) site-packages/setuptools/command/egg_info.py000064400000061742151464452710015275 0ustar00"""setuptools.command.egg_info Create a distribution's .egg-info directory and contents""" from distutils.filelist import FileList as _FileList from distutils.errors import DistutilsInternalError from distutils.util import convert_path from distutils import log import distutils.errors import distutils.filelist import os import re import sys import io import warnings import time import collections from setuptools.extern import six from setuptools.extern.six.moves import map from setuptools import Command from setuptools.command.sdist import sdist from setuptools.command.sdist import walk_revctrl from setuptools.command.setopt import edit_config from setuptools.command import bdist_egg from pkg_resources import ( parse_requirements, safe_name, parse_version, safe_version, yield_lines, EntryPoint, iter_entry_points, to_filename) import setuptools.unicode_utils as unicode_utils from setuptools.glob import glob from setuptools.extern import packaging from setuptools import SetuptoolsDeprecationWarning def translate_pattern(glob): """ Translate a file path glob like '*.txt' in to a regular expression. This differs from fnmatch.translate which allows wildcards to match directory separators. It also knows about '**/' which matches any number of directories. """ pat = '' # This will split on '/' within [character classes]. This is deliberate. chunks = glob.split(os.path.sep) sep = re.escape(os.sep) valid_char = '[^%s]' % (sep,) for c, chunk in enumerate(chunks): last_chunk = c == len(chunks) - 1 # Chunks that are a literal ** are globstars. They match anything. if chunk == '**': if last_chunk: # Match anything if this is the last component pat += '.*' else: # Match '(name/)*' pat += '(?:%s+%s)*' % (valid_char, sep) continue # Break here as the whole path component has been handled # Find any special characters in the remainder i = 0 chunk_len = len(chunk) while i < chunk_len: char = chunk[i] if char == '*': # Match any number of name characters pat += valid_char + '*' elif char == '?': # Match a name character pat += valid_char elif char == '[': # Character class inner_i = i + 1 # Skip initial !/] chars if inner_i < chunk_len and chunk[inner_i] == '!': inner_i = inner_i + 1 if inner_i < chunk_len and chunk[inner_i] == ']': inner_i = inner_i + 1 # Loop till the closing ] is found while inner_i < chunk_len and chunk[inner_i] != ']': inner_i = inner_i + 1 if inner_i >= chunk_len: # Got to the end of the string without finding a closing ] # Do not treat this as a matching group, but as a literal [ pat += re.escape(char) else: # Grab the insides of the [brackets] inner = chunk[i + 1:inner_i] char_class = '' # Class negation if inner[0] == '!': char_class = '^' inner = inner[1:] char_class += re.escape(inner) pat += '[%s]' % (char_class,) # Skip to the end ] i = inner_i else: pat += re.escape(char) i += 1 # Join each chunk with the dir separator if not last_chunk: pat += sep pat += r'\Z' return re.compile(pat, flags=re.MULTILINE|re.DOTALL) class InfoCommon: tag_build = None tag_date = None @property def name(self): return safe_name(self.distribution.get_name()) def tagged_version(self): version = self.distribution.get_version() # egg_info may be called more than once for a distribution, # in which case the version string already contains all tags. if self.vtags and version.endswith(self.vtags): return safe_version(version) return safe_version(version + self.vtags) def tags(self): version = '' if self.tag_build: version += self.tag_build if self.tag_date: version += time.strftime("-%Y%m%d") return version vtags = property(tags) class egg_info(InfoCommon, Command): description = "create a distribution's .egg-info directory" user_options = [ ('egg-base=', 'e', "directory containing .egg-info directories" " (default: top of the source tree)"), ('tag-date', 'd', "Add date stamp (e.g. 20050528) to version number"), ('tag-build=', 'b', "Specify explicit tag to add to version number"), ('no-date', 'D', "Don't include date stamp [default]"), ] boolean_options = ['tag-date'] negative_opt = { 'no-date': 'tag-date', } def initialize_options(self): self.egg_base = None self.egg_name = None self.egg_info = None self.egg_version = None self.broken_egg_info = False #################################### # allow the 'tag_svn_revision' to be detected and # set, supporting sdists built on older Setuptools. @property def tag_svn_revision(self): pass @tag_svn_revision.setter def tag_svn_revision(self, value): pass #################################### def save_version_info(self, filename): """ Materialize the value of date into the build tag. Install build keys in a deterministic order to avoid arbitrary reordering on subsequent builds. """ egg_info = collections.OrderedDict() # follow the order these keys would have been added # when PYTHONHASHSEED=0 egg_info['tag_build'] = self.tags() egg_info['tag_date'] = 0 edit_config(filename, dict(egg_info=egg_info)) def finalize_options(self): # Note: we need to capture the current value returned # by `self.tagged_version()`, so we can later update # `self.distribution.metadata.version` without # repercussions. self.egg_name = self.name self.egg_version = self.tagged_version() parsed_version = parse_version(self.egg_version) try: is_version = isinstance(parsed_version, packaging.version.Version) spec = ( "%s==%s" if is_version else "%s===%s" ) list( parse_requirements(spec % (self.egg_name, self.egg_version)) ) except ValueError: raise distutils.errors.DistutilsOptionError( "Invalid distribution name or version syntax: %s-%s" % (self.egg_name, self.egg_version) ) if self.egg_base is None: dirs = self.distribution.package_dir self.egg_base = (dirs or {}).get('', os.curdir) self.ensure_dirname('egg_base') self.egg_info = to_filename(self.egg_name) + '.egg-info' if self.egg_base != os.curdir: self.egg_info = os.path.join(self.egg_base, self.egg_info) if '-' in self.egg_name: self.check_broken_egg_info() # Set package version for the benefit of dumber commands # (e.g. sdist, bdist_wininst, etc.) # self.distribution.metadata.version = self.egg_version # If we bootstrapped around the lack of a PKG-INFO, as might be the # case in a fresh checkout, make sure that any special tags get added # to the version info # pd = self.distribution._patched_dist if pd is not None and pd.key == self.egg_name.lower(): pd._version = self.egg_version pd._parsed_version = parse_version(self.egg_version) self.distribution._patched_dist = None def write_or_delete_file(self, what, filename, data, force=False): """Write `data` to `filename` or delete if empty If `data` is non-empty, this routine is the same as ``write_file()``. If `data` is empty but not ``None``, this is the same as calling ``delete_file(filename)`. If `data` is ``None``, then this is a no-op unless `filename` exists, in which case a warning is issued about the orphaned file (if `force` is false), or deleted (if `force` is true). """ if data: self.write_file(what, filename, data) elif os.path.exists(filename): if data is None and not force: log.warn( "%s not set in setup(), but %s exists", what, filename ) return else: self.delete_file(filename) def write_file(self, what, filename, data): """Write `data` to `filename` (if not a dry run) after announcing it `what` is used in a log message to identify what is being written to the file. """ log.info("writing %s to %s", what, filename) if six.PY3: data = data.encode("utf-8") if not self.dry_run: f = open(filename, 'wb') f.write(data) f.close() def delete_file(self, filename): """Delete `filename` (if not a dry run) after announcing it""" log.info("deleting %s", filename) if not self.dry_run: os.unlink(filename) def run(self): self.mkpath(self.egg_info) os.utime(self.egg_info, None) installer = self.distribution.fetch_build_egg for ep in iter_entry_points('egg_info.writers'): ep.require(installer=installer) writer = ep.resolve() writer(self, ep.name, os.path.join(self.egg_info, ep.name)) # Get rid of native_libs.txt if it was put there by older bdist_egg nl = os.path.join(self.egg_info, "native_libs.txt") if os.path.exists(nl): self.delete_file(nl) self.find_sources() def find_sources(self): """Generate SOURCES.txt manifest file""" manifest_filename = os.path.join(self.egg_info, "SOURCES.txt") mm = manifest_maker(self.distribution) mm.manifest = manifest_filename mm.run() self.filelist = mm.filelist def check_broken_egg_info(self): bei = self.egg_name + '.egg-info' if self.egg_base != os.curdir: bei = os.path.join(self.egg_base, bei) if os.path.exists(bei): log.warn( "-" * 78 + '\n' "Note: Your current .egg-info directory has a '-' in its name;" '\nthis will not work correctly with "setup.py develop".\n\n' 'Please rename %s to %s to correct this problem.\n' + '-' * 78, bei, self.egg_info ) self.broken_egg_info = self.egg_info self.egg_info = bei # make it work for now class FileList(_FileList): # Implementations of the various MANIFEST.in commands def process_template_line(self, line): # Parse the line: split it up, make sure the right number of words # is there, and return the relevant words. 'action' is always # defined: it's the first word of the line. Which of the other # three are defined depends on the action; it'll be either # patterns, (dir and patterns), or (dir_pattern). (action, patterns, dir, dir_pattern) = self._parse_template_line(line) # OK, now we know that the action is valid and we have the # right number of words on the line for that action -- so we # can proceed with minimal error-checking. if action == 'include': self.debug_print("include " + ' '.join(patterns)) for pattern in patterns: if not self.include(pattern): log.warn("warning: no files found matching '%s'", pattern) elif action == 'exclude': self.debug_print("exclude " + ' '.join(patterns)) for pattern in patterns: if not self.exclude(pattern): log.warn(("warning: no previously-included files " "found matching '%s'"), pattern) elif action == 'global-include': self.debug_print("global-include " + ' '.join(patterns)) for pattern in patterns: if not self.global_include(pattern): log.warn(("warning: no files found matching '%s' " "anywhere in distribution"), pattern) elif action == 'global-exclude': self.debug_print("global-exclude " + ' '.join(patterns)) for pattern in patterns: if not self.global_exclude(pattern): log.warn(("warning: no previously-included files matching " "'%s' found anywhere in distribution"), pattern) elif action == 'recursive-include': self.debug_print("recursive-include %s %s" % (dir, ' '.join(patterns))) for pattern in patterns: if not self.recursive_include(dir, pattern): log.warn(("warning: no files found matching '%s' " "under directory '%s'"), pattern, dir) elif action == 'recursive-exclude': self.debug_print("recursive-exclude %s %s" % (dir, ' '.join(patterns))) for pattern in patterns: if not self.recursive_exclude(dir, pattern): log.warn(("warning: no previously-included files matching " "'%s' found under directory '%s'"), pattern, dir) elif action == 'graft': self.debug_print("graft " + dir_pattern) if not self.graft(dir_pattern): log.warn("warning: no directories found matching '%s'", dir_pattern) elif action == 'prune': self.debug_print("prune " + dir_pattern) if not self.prune(dir_pattern): log.warn(("no previously-included directories found " "matching '%s'"), dir_pattern) else: raise DistutilsInternalError( "this cannot happen: invalid action '%s'" % action) def _remove_files(self, predicate): """ Remove all files from the file list that match the predicate. Return True if any matching files were removed """ found = False for i in range(len(self.files) - 1, -1, -1): if predicate(self.files[i]): self.debug_print(" removing " + self.files[i]) del self.files[i] found = True return found def include(self, pattern): """Include files that match 'pattern'.""" found = [f for f in glob(pattern) if not os.path.isdir(f)] self.extend(found) return bool(found) def exclude(self, pattern): """Exclude files that match 'pattern'.""" match = translate_pattern(pattern) return self._remove_files(match.match) def recursive_include(self, dir, pattern): """ Include all files anywhere in 'dir/' that match the pattern. """ full_pattern = os.path.join(dir, '**', pattern) found = [f for f in glob(full_pattern, recursive=True) if not os.path.isdir(f)] self.extend(found) return bool(found) def recursive_exclude(self, dir, pattern): """ Exclude any file anywhere in 'dir/' that match the pattern. """ match = translate_pattern(os.path.join(dir, '**', pattern)) return self._remove_files(match.match) def graft(self, dir): """Include all files from 'dir/'.""" found = [ item for match_dir in glob(dir) for item in distutils.filelist.findall(match_dir) ] self.extend(found) return bool(found) def prune(self, dir): """Filter out files from 'dir/'.""" match = translate_pattern(os.path.join(dir, '**')) return self._remove_files(match.match) def global_include(self, pattern): """ Include all files anywhere in the current directory that match the pattern. This is very inefficient on large file trees. """ if self.allfiles is None: self.findall() match = translate_pattern(os.path.join('**', pattern)) found = [f for f in self.allfiles if match.match(f)] self.extend(found) return bool(found) def global_exclude(self, pattern): """ Exclude all files anywhere that match the pattern. """ match = translate_pattern(os.path.join('**', pattern)) return self._remove_files(match.match) def append(self, item): if item.endswith('\r'): # Fix older sdists built on Windows item = item[:-1] path = convert_path(item) if self._safe_path(path): self.files.append(path) def extend(self, paths): self.files.extend(filter(self._safe_path, paths)) def _repair(self): """ Replace self.files with only safe paths Because some owners of FileList manipulate the underlying ``files`` attribute directly, this method must be called to repair those paths. """ self.files = list(filter(self._safe_path, self.files)) def _safe_path(self, path): enc_warn = "'%s' not %s encodable -- skipping" # To avoid accidental trans-codings errors, first to unicode u_path = unicode_utils.filesys_decode(path) if u_path is None: log.warn("'%s' in unexpected encoding -- skipping" % path) return False # Must ensure utf-8 encodability utf8_path = unicode_utils.try_encode(u_path, "utf-8") if utf8_path is None: log.warn(enc_warn, path, 'utf-8') return False try: # accept is either way checks out if os.path.exists(u_path) or os.path.exists(utf8_path): return True # this will catch any encode errors decoding u_path except UnicodeEncodeError: log.warn(enc_warn, path, sys.getfilesystemencoding()) class manifest_maker(sdist): template = "MANIFEST.in" def initialize_options(self): self.use_defaults = 1 self.prune = 1 self.manifest_only = 1 self.force_manifest = 1 def finalize_options(self): pass def run(self): self.filelist = FileList() if not os.path.exists(self.manifest): self.write_manifest() # it must exist so it'll get in the list self.add_defaults() if os.path.exists(self.template): self.read_template() self.prune_file_list() self.filelist.sort() self.filelist.remove_duplicates() self.write_manifest() def _manifest_normalize(self, path): path = unicode_utils.filesys_decode(path) return path.replace(os.sep, '/') def write_manifest(self): """ Write the file list in 'self.filelist' to the manifest file named by 'self.manifest'. """ self.filelist._repair() # Now _repairs should encodability, but not unicode files = [self._manifest_normalize(f) for f in self.filelist.files] msg = "writing manifest file '%s'" % self.manifest self.execute(write_file, (self.manifest, files), msg) def warn(self, msg): if not self._should_suppress_warning(msg): sdist.warn(self, msg) @staticmethod def _should_suppress_warning(msg): """ suppress missing-file warnings from sdist """ return re.match(r"standard file .*not found", msg) def add_defaults(self): sdist.add_defaults(self) self.check_license() self.filelist.append(self.template) self.filelist.append(self.manifest) rcfiles = list(walk_revctrl()) if rcfiles: self.filelist.extend(rcfiles) elif os.path.exists(self.manifest): self.read_manifest() if os.path.exists("setup.py"): # setup.py should be included by default, even if it's not # the script called to create the sdist self.filelist.append("setup.py") ei_cmd = self.get_finalized_command('egg_info') self.filelist.graft(ei_cmd.egg_info) def prune_file_list(self): build = self.get_finalized_command('build') base_dir = self.distribution.get_fullname() self.filelist.prune(build.build_base) self.filelist.prune(base_dir) sep = re.escape(os.sep) self.filelist.exclude_pattern(r'(^|' + sep + r')(RCS|CVS|\.svn)' + sep, is_regex=1) def write_file(filename, contents): """Create a file with the specified name and write 'contents' (a sequence of strings without line terminators) to it. """ contents = "\n".join(contents) # assuming the contents has been vetted for utf-8 encoding contents = contents.encode("utf-8") with open(filename, "wb") as f: # always write POSIX-style manifest f.write(contents) def write_pkg_info(cmd, basename, filename): log.info("writing %s", filename) if not cmd.dry_run: metadata = cmd.distribution.metadata metadata.version, oldver = cmd.egg_version, metadata.version metadata.name, oldname = cmd.egg_name, metadata.name try: # write unescaped data to PKG-INFO, so older pkg_resources # can still parse it metadata.write_pkg_info(cmd.egg_info) finally: metadata.name, metadata.version = oldname, oldver safe = getattr(cmd.distribution, 'zip_safe', None) bdist_egg.write_safety_flag(cmd.egg_info, safe) def warn_depends_obsolete(cmd, basename, filename): if os.path.exists(filename): log.warn( "WARNING: 'depends.txt' is not used by setuptools 0.6!\n" "Use the install_requires/extras_require setup() args instead." ) def _write_requirements(stream, reqs): lines = yield_lines(reqs or ()) append_cr = lambda line: line + '\n' lines = map(append_cr, lines) stream.writelines(lines) def write_requirements(cmd, basename, filename): dist = cmd.distribution data = six.StringIO() _write_requirements(data, dist.install_requires) extras_require = dist.extras_require or {} for extra in sorted(extras_require): data.write('\n[{extra}]\n'.format(**vars())) _write_requirements(data, extras_require[extra]) cmd.write_or_delete_file("requirements", filename, data.getvalue()) def write_setup_requirements(cmd, basename, filename): data = io.StringIO() _write_requirements(data, cmd.distribution.setup_requires) cmd.write_or_delete_file("setup-requirements", filename, data.getvalue()) def write_toplevel_names(cmd, basename, filename): pkgs = dict.fromkeys( [ k.split('.', 1)[0] for k in cmd.distribution.iter_distribution_names() ] ) cmd.write_file("top-level names", filename, '\n'.join(sorted(pkgs)) + '\n') def overwrite_arg(cmd, basename, filename): write_arg(cmd, basename, filename, True) def write_arg(cmd, basename, filename, force=False): argname = os.path.splitext(basename)[0] value = getattr(cmd.distribution, argname, None) if value is not None: value = '\n'.join(value) + '\n' cmd.write_or_delete_file(argname, filename, value, force) def write_entries(cmd, basename, filename): ep = cmd.distribution.entry_points if isinstance(ep, six.string_types) or ep is None: data = ep elif ep is not None: data = [] for section, contents in sorted(ep.items()): if not isinstance(contents, six.string_types): contents = EntryPoint.parse_group(section, contents) contents = '\n'.join(sorted(map(str, contents.values()))) data.append('[%s]\n%s\n\n' % (section, contents)) data = ''.join(data) cmd.write_or_delete_file('entry points', filename, data, True) def get_pkg_info_revision(): """ Get a -r### off of PKG-INFO Version in case this is an sdist of a subversion revision. """ warnings.warn("get_pkg_info_revision is deprecated.", EggInfoDeprecationWarning) if os.path.exists('PKG-INFO'): with io.open('PKG-INFO') as f: for line in f: match = re.match(r"Version:.*-r(\d+)\s*$", line) if match: return int(match.group(1)) return 0 class EggInfoDeprecationWarning(SetuptoolsDeprecationWarning): """Class for warning about deprecations in eggInfo in setupTools. Not ignored by default, unlike DeprecationWarning.""" site-packages/setuptools/command/upload.py000064400000015233151464452710014776 0ustar00import io import os import hashlib import getpass from base64 import standard_b64encode from distutils import log from distutils.command import upload as orig from distutils.spawn import spawn from distutils.errors import DistutilsError from setuptools.extern.six.moves.urllib.request import urlopen, Request from setuptools.extern.six.moves.urllib.error import HTTPError from setuptools.extern.six.moves.urllib.parse import urlparse class upload(orig.upload): """ Override default upload behavior to obtain password in a variety of different ways. """ def run(self): try: orig.upload.run(self) finally: self.announce( "WARNING: Uploading via this command is deprecated, use twine " "to upload instead (https://pypi.org/p/twine/)", log.WARN ) def finalize_options(self): orig.upload.finalize_options(self) self.username = ( self.username or getpass.getuser() ) # Attempt to obtain password. Short circuit evaluation at the first # sign of success. self.password = ( self.password or self._load_password_from_keyring() or self._prompt_for_password() ) def upload_file(self, command, pyversion, filename): # Makes sure the repository URL is compliant schema, netloc, url, params, query, fragments = \ urlparse(self.repository) if params or query or fragments: raise AssertionError("Incompatible url %s" % self.repository) if schema not in ('http', 'https'): raise AssertionError("unsupported schema " + schema) # Sign if requested if self.sign: gpg_args = ["gpg", "--detach-sign", "-a", filename] if self.identity: gpg_args[2:2] = ["--local-user", self.identity] spawn(gpg_args, dry_run=self.dry_run) # Fill in the data - send all the meta-data in case we need to # register a new release with open(filename, 'rb') as f: content = f.read() meta = self.distribution.metadata data = { # action ':action': 'file_upload', 'protocol_version': '1', # identify release 'name': meta.get_name(), 'version': meta.get_version(), # file content 'content': (os.path.basename(filename), content), 'filetype': command, 'pyversion': pyversion, 'md5_digest': hashlib.md5(content).hexdigest(), # additional meta-data 'metadata_version': str(meta.get_metadata_version()), 'summary': meta.get_description(), 'home_page': meta.get_url(), 'author': meta.get_contact(), 'author_email': meta.get_contact_email(), 'license': meta.get_licence(), 'description': meta.get_long_description(), 'keywords': meta.get_keywords(), 'platform': meta.get_platforms(), 'classifiers': meta.get_classifiers(), 'download_url': meta.get_download_url(), # PEP 314 'provides': meta.get_provides(), 'requires': meta.get_requires(), 'obsoletes': meta.get_obsoletes(), } data['comment'] = '' if self.sign: data['gpg_signature'] = (os.path.basename(filename) + ".asc", open(filename+".asc", "rb").read()) # set up the authentication user_pass = (self.username + ":" + self.password).encode('ascii') # The exact encoding of the authentication string is debated. # Anyway PyPI only accepts ascii for both username or password. auth = "Basic " + standard_b64encode(user_pass).decode('ascii') # Build up the MIME payload for the POST data boundary = '--------------GHSKFJDLGDS7543FJKLFHRE75642756743254' sep_boundary = b'\r\n--' + boundary.encode('ascii') end_boundary = sep_boundary + b'--\r\n' body = io.BytesIO() for key, value in data.items(): title = '\r\nContent-Disposition: form-data; name="%s"' % key # handle multiple entries for the same name if not isinstance(value, list): value = [value] for value in value: if type(value) is tuple: title += '; filename="%s"' % value[0] value = value[1] else: value = str(value).encode('utf-8') body.write(sep_boundary) body.write(title.encode('utf-8')) body.write(b"\r\n\r\n") body.write(value) body.write(end_boundary) body = body.getvalue() msg = "Submitting %s to %s" % (filename, self.repository) self.announce(msg, log.INFO) # build the Request headers = { 'Content-type': 'multipart/form-data; boundary=%s' % boundary, 'Content-length': str(len(body)), 'Authorization': auth, } request = Request(self.repository, data=body, headers=headers) # send the data try: result = urlopen(request) status = result.getcode() reason = result.msg except HTTPError as e: status = e.code reason = e.msg except OSError as e: self.announce(str(e), log.ERROR) raise if status == 200: self.announce('Server response (%s): %s' % (status, reason), log.INFO) if self.show_response: text = getattr(self, '_read_pypi_response', lambda x: None)(result) if text is not None: msg = '\n'.join(('-' * 75, text, '-' * 75)) self.announce(msg, log.INFO) else: msg = 'Upload failed (%s): %s' % (status, reason) self.announce(msg, log.ERROR) raise DistutilsError(msg) def _load_password_from_keyring(self): """ Attempt to load password from keyring. Suppress Exceptions. """ try: keyring = __import__('keyring') return keyring.get_password(self.repository, self.username) except Exception: pass def _prompt_for_password(self): """ Prompt for a password on the tty. Suppress Exceptions. """ try: return getpass.getpass() except (Exception, KeyboardInterrupt): pass site-packages/setuptools/command/install_egg_info.py000064400000004233151464452710017013 0ustar00from distutils import log, dir_util import os from setuptools import Command from setuptools import namespaces from setuptools.archive_util import unpack_archive import pkg_resources class install_egg_info(namespaces.Installer, Command): """Install an .egg-info directory for the package""" description = "Install an .egg-info directory for the package" user_options = [ ('install-dir=', 'd', "directory to install to"), ] def initialize_options(self): self.install_dir = None def finalize_options(self): self.set_undefined_options('install_lib', ('install_dir', 'install_dir')) ei_cmd = self.get_finalized_command("egg_info") basename = pkg_resources.Distribution( None, None, ei_cmd.egg_name, ei_cmd.egg_version ).egg_name() + '.egg-info' self.source = ei_cmd.egg_info self.target = os.path.join(self.install_dir, basename) self.outputs = [] def run(self): self.run_command('egg_info') if os.path.isdir(self.target) and not os.path.islink(self.target): dir_util.remove_tree(self.target, dry_run=self.dry_run) elif os.path.exists(self.target): self.execute(os.unlink, (self.target,), "Removing " + self.target) if not self.dry_run: pkg_resources.ensure_directory(self.target) self.execute( self.copytree, (), "Copying %s to %s" % (self.source, self.target) ) self.install_namespaces() def get_outputs(self): return self.outputs def copytree(self): # Copy the .egg-info tree to site-packages def skimmer(src, dst): # filter out source-control directories; note that 'src' is always # a '/'-separated path, regardless of platform. 'dst' is a # platform-specific path. for skip in '.svn/', 'CVS/': if src.startswith(skip) or '/' + skip in src: return None self.outputs.append(dst) log.debug("Copying %s to %s", src, dst) return dst unpack_archive(self.source, self.target, skimmer) site-packages/setuptools/command/easy_install.py000064400000252635151464452710016212 0ustar00""" Easy Install ------------ A tool for doing automatic download/extract/build of distutils-based Python packages. For detailed documentation, see the accompanying EasyInstall.txt file, or visit the `EasyInstall home page`__. __ https://setuptools.readthedocs.io/en/latest/easy_install.html """ from glob import glob from distutils.util import get_platform from distutils.util import convert_path, subst_vars from distutils.errors import ( DistutilsArgError, DistutilsOptionError, DistutilsError, DistutilsPlatformError, ) from distutils.command.install import INSTALL_SCHEMES, SCHEME_KEYS from distutils import log, dir_util from distutils.command.build_scripts import first_line_re from distutils.spawn import find_executable import sys import os import zipimport import shutil import tempfile import zipfile import re import stat import random import textwrap import warnings import site import struct import contextlib import subprocess import shlex import io from sysconfig import get_config_vars, get_path from setuptools import SetuptoolsDeprecationWarning from setuptools.extern import six from setuptools.extern.six.moves import configparser, map from setuptools import Command from setuptools.sandbox import run_setup from setuptools.py27compat import rmtree_safe from setuptools.command import setopt from setuptools.archive_util import unpack_archive from setuptools.package_index import ( PackageIndex, parse_requirement_arg, URL_SCHEME, ) from setuptools.command import bdist_egg, egg_info from setuptools.wheel import Wheel from pkg_resources import ( yield_lines, normalize_path, resource_string, ensure_directory, get_distribution, find_distributions, Environment, Requirement, Distribution, PathMetadata, EggMetadata, WorkingSet, DistributionNotFound, VersionConflict, DEVELOP_DIST, ) import pkg_resources.py31compat __metaclass__ = type # Turn on PEP440Warnings warnings.filterwarnings("default", category=pkg_resources.PEP440Warning) __all__ = [ 'samefile', 'easy_install', 'PthDistributions', 'extract_wininst_cfg', 'main', 'get_exe_prefixes', ] def is_64bit(): return struct.calcsize("P") == 8 def samefile(p1, p2): """ Determine if two paths reference the same file. Augments os.path.samefile to work on Windows and suppresses errors if the path doesn't exist. """ both_exist = os.path.exists(p1) and os.path.exists(p2) use_samefile = hasattr(os.path, 'samefile') and both_exist if use_samefile: return os.path.samefile(p1, p2) norm_p1 = os.path.normpath(os.path.normcase(p1)) norm_p2 = os.path.normpath(os.path.normcase(p2)) return norm_p1 == norm_p2 if six.PY2: def _to_bytes(s): return s def isascii(s): try: six.text_type(s, 'ascii') return True except UnicodeError: return False else: def _to_bytes(s): return s.encode('utf8') def isascii(s): try: s.encode('ascii') return True except UnicodeError: return False _one_liner = lambda text: textwrap.dedent(text).strip().replace('\n', '; ') class easy_install(Command): """Manage a download/build/install process""" description = "Find/get/install Python packages" command_consumes_arguments = True user_options = [ ('prefix=', None, "installation prefix"), ("zip-ok", "z", "install package as a zipfile"), ("multi-version", "m", "make apps have to require() a version"), ("upgrade", "U", "force upgrade (searches PyPI for latest versions)"), ("install-dir=", "d", "install package to DIR"), ("script-dir=", "s", "install scripts to DIR"), ("exclude-scripts", "x", "Don't install scripts"), ("always-copy", "a", "Copy all needed packages to install dir"), ("index-url=", "i", "base URL of Python Package Index"), ("find-links=", "f", "additional URL(s) to search for packages"), ("build-directory=", "b", "download/extract/build in DIR; keep the results"), ('optimize=', 'O', "also compile with optimization: -O1 for \"python -O\", " "-O2 for \"python -OO\", and -O0 to disable [default: -O0]"), ('record=', None, "filename in which to record list of installed files"), ('always-unzip', 'Z', "don't install as a zipfile, no matter what"), ('site-dirs=', 'S', "list of directories where .pth files work"), ('editable', 'e', "Install specified packages in editable form"), ('no-deps', 'N', "don't install dependencies"), ('allow-hosts=', 'H', "pattern(s) that hostnames must match"), ('local-snapshots-ok', 'l', "allow building eggs from local checkouts"), ('version', None, "print version information and exit"), ('no-find-links', None, "Don't load find-links defined in packages being installed") ] boolean_options = [ 'zip-ok', 'multi-version', 'exclude-scripts', 'upgrade', 'always-copy', 'editable', 'no-deps', 'local-snapshots-ok', 'version' ] if site.ENABLE_USER_SITE: help_msg = "install in user site-package '%s'" % site.USER_SITE user_options.append(('user', None, help_msg)) boolean_options.append('user') negative_opt = {'always-unzip': 'zip-ok'} create_index = PackageIndex def initialize_options(self): # the --user option seems to be an opt-in one, # so the default should be False. self.user = 0 self.zip_ok = self.local_snapshots_ok = None self.install_dir = self.script_dir = self.exclude_scripts = None self.index_url = None self.find_links = None self.build_directory = None self.args = None self.optimize = self.record = None self.upgrade = self.always_copy = self.multi_version = None self.editable = self.no_deps = self.allow_hosts = None self.root = self.prefix = self.no_report = None self.version = None self.install_purelib = None # for pure module distributions self.install_platlib = None # non-pure (dists w/ extensions) self.install_headers = None # for C/C++ headers self.install_lib = None # set to either purelib or platlib self.install_scripts = None self.install_data = None self.install_base = None self.install_platbase = None if site.ENABLE_USER_SITE: self.install_userbase = site.USER_BASE self.install_usersite = site.USER_SITE else: self.install_userbase = None self.install_usersite = None self.no_find_links = None # Options not specifiable via command line self.package_index = None self.pth_file = self.always_copy_from = None self.site_dirs = None self.installed_projects = {} self.sitepy_installed = False # Always read easy_install options, even if we are subclassed, or have # an independent instance created. This ensures that defaults will # always come from the standard configuration file(s)' "easy_install" # section, even if this is a "develop" or "install" command, or some # other embedding. self._dry_run = None self.verbose = self.distribution.verbose self.distribution._set_command_options( self, self.distribution.get_option_dict('easy_install') ) def delete_blockers(self, blockers): extant_blockers = ( filename for filename in blockers if os.path.exists(filename) or os.path.islink(filename) ) list(map(self._delete_path, extant_blockers)) def _delete_path(self, path): log.info("Deleting %s", path) if self.dry_run: return is_tree = os.path.isdir(path) and not os.path.islink(path) remover = rmtree if is_tree else os.unlink remover(path) @staticmethod def _render_version(): """ Render the Setuptools version and installation details, then exit. """ ver = '{}.{}'.format(*sys.version_info) dist = get_distribution('setuptools') tmpl = 'setuptools {dist.version} from {dist.location} (Python {ver})' print(tmpl.format(**locals())) raise SystemExit() def finalize_options(self): self.version and self._render_version() py_version = sys.version.split()[0] prefix, exec_prefix = get_config_vars('prefix', 'exec_prefix') self.config_vars = { 'dist_name': self.distribution.get_name(), 'dist_version': self.distribution.get_version(), 'dist_fullname': self.distribution.get_fullname(), 'py_version': py_version, 'py_version_short': py_version[0:3], 'py_version_nodot': py_version[0] + py_version[2], 'sys_prefix': prefix, 'prefix': prefix, 'sys_exec_prefix': exec_prefix, 'exec_prefix': exec_prefix, # Only python 3.2+ has abiflags 'abiflags': getattr(sys, 'abiflags', ''), } if site.ENABLE_USER_SITE: self.config_vars['userbase'] = self.install_userbase self.config_vars['usersite'] = self.install_usersite self._fix_install_dir_for_user_site() self.expand_basedirs() self.expand_dirs() self._expand( 'install_dir', 'script_dir', 'build_directory', 'site_dirs', ) # If a non-default installation directory was specified, default the # script directory to match it. if self.script_dir is None: self.script_dir = self.install_dir if self.no_find_links is None: self.no_find_links = False # Let install_dir get set by install_lib command, which in turn # gets its info from the install command, and takes into account # --prefix and --home and all that other crud. self.set_undefined_options( 'install_lib', ('install_dir', 'install_dir') ) # Likewise, set default script_dir from 'install_scripts.install_dir' self.set_undefined_options( 'install_scripts', ('install_dir', 'script_dir') ) if self.user and self.install_purelib: self.install_dir = self.install_purelib self.script_dir = self.install_scripts # default --record from the install command self.set_undefined_options('install', ('record', 'record')) # Should this be moved to the if statement below? It's not used # elsewhere normpath = map(normalize_path, sys.path) self.all_site_dirs = get_site_dirs() if self.site_dirs is not None: site_dirs = [ os.path.expanduser(s.strip()) for s in self.site_dirs.split(',') ] for d in site_dirs: if not os.path.isdir(d): log.warn("%s (in --site-dirs) does not exist", d) elif normalize_path(d) not in normpath: raise DistutilsOptionError( d + " (in --site-dirs) is not on sys.path" ) else: self.all_site_dirs.append(normalize_path(d)) if not self.editable: self.check_site_dir() self.index_url = self.index_url or "https://pypi.org/simple/" self.shadow_path = self.all_site_dirs[:] for path_item in self.install_dir, normalize_path(self.script_dir): if path_item not in self.shadow_path: self.shadow_path.insert(0, path_item) if self.allow_hosts is not None: hosts = [s.strip() for s in self.allow_hosts.split(',')] else: hosts = ['*'] if self.package_index is None: self.package_index = self.create_index( self.index_url, search_path=self.shadow_path, hosts=hosts, ) self.local_index = Environment(self.shadow_path + sys.path) if self.find_links is not None: if isinstance(self.find_links, six.string_types): self.find_links = self.find_links.split() else: self.find_links = [] if self.local_snapshots_ok: self.package_index.scan_egg_links(self.shadow_path + sys.path) if not self.no_find_links: self.package_index.add_find_links(self.find_links) self.set_undefined_options('install_lib', ('optimize', 'optimize')) if not isinstance(self.optimize, int): try: self.optimize = int(self.optimize) if not (0 <= self.optimize <= 2): raise ValueError except ValueError: raise DistutilsOptionError("--optimize must be 0, 1, or 2") if self.editable and not self.build_directory: raise DistutilsArgError( "Must specify a build directory (-b) when using --editable" ) if not self.args: raise DistutilsArgError( "No urls, filenames, or requirements specified (see --help)") self.outputs = [] def _fix_install_dir_for_user_site(self): """ Fix the install_dir if "--user" was used. """ if not self.user or not site.ENABLE_USER_SITE: return self.create_home_path() if self.install_userbase is None: msg = "User base directory is not specified" raise DistutilsPlatformError(msg) self.install_base = self.install_platbase = self.install_userbase scheme_name = os.name.replace('posix', 'unix') + '_user' self.select_scheme(scheme_name) def _expand_attrs(self, attrs): for attr in attrs: val = getattr(self, attr) if val is not None: if os.name == 'posix' or os.name == 'nt': val = os.path.expanduser(val) val = subst_vars(val, self.config_vars) setattr(self, attr, val) def expand_basedirs(self): """Calls `os.path.expanduser` on install_base, install_platbase and root.""" self._expand_attrs(['install_base', 'install_platbase', 'root']) def expand_dirs(self): """Calls `os.path.expanduser` on install dirs.""" dirs = [ 'install_purelib', 'install_platlib', 'install_lib', 'install_headers', 'install_scripts', 'install_data', ] self._expand_attrs(dirs) def run(self): if self.verbose != self.distribution.verbose: log.set_verbosity(self.verbose) try: for spec in self.args: self.easy_install(spec, not self.no_deps) if self.record: outputs = self.outputs if self.root: # strip any package prefix root_len = len(self.root) for counter in range(len(outputs)): outputs[counter] = outputs[counter][root_len:] from distutils import file_util self.execute( file_util.write_file, (self.record, outputs), "writing list of installed files to '%s'" % self.record ) self.warn_deprecated_options() finally: log.set_verbosity(self.distribution.verbose) def pseudo_tempname(self): """Return a pseudo-tempname base in the install directory. This code is intentionally naive; if a malicious party can write to the target directory you're already in deep doodoo. """ try: pid = os.getpid() except Exception: pid = random.randint(0, sys.maxsize) return os.path.join(self.install_dir, "test-easy-install-%s" % pid) def warn_deprecated_options(self): pass def check_site_dir(self): """Verify that self.install_dir is .pth-capable dir, if needed""" instdir = normalize_path(self.install_dir) pth_file = os.path.join(instdir, 'easy-install.pth') if not os.path.exists(instdir): try: os.makedirs(instdir) except (OSError, IOError): self.cant_write_to_target() # Is it a configured, PYTHONPATH, implicit, or explicit site dir? is_site_dir = instdir in self.all_site_dirs if not is_site_dir and not self.multi_version: # No? Then directly test whether it does .pth file processing is_site_dir = self.check_pth_processing() else: # make sure we can write to target dir testfile = self.pseudo_tempname() + '.write-test' test_exists = os.path.exists(testfile) try: if test_exists: os.unlink(testfile) open(testfile, 'w').close() os.unlink(testfile) except (OSError, IOError): self.cant_write_to_target() if not is_site_dir and not self.multi_version: # Can't install non-multi to non-site dir raise DistutilsError(self.no_default_version_msg()) if is_site_dir: if self.pth_file is None: self.pth_file = PthDistributions(pth_file, self.all_site_dirs) else: self.pth_file = None if instdir not in map(normalize_path, _pythonpath()): # only PYTHONPATH dirs need a site.py, so pretend it's there self.sitepy_installed = True elif self.multi_version and not os.path.exists(pth_file): self.sitepy_installed = True # don't need site.py in this case self.pth_file = None # and don't create a .pth file self.install_dir = instdir __cant_write_msg = textwrap.dedent(""" can't create or remove files in install directory The following error occurred while trying to add or remove files in the installation directory: %s The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s """).lstrip() __not_exists_id = textwrap.dedent(""" This directory does not currently exist. Please create it and try again, or choose a different installation directory (using the -d or --install-dir option). """).lstrip() __access_msg = textwrap.dedent(""" Perhaps your account does not have write access to this directory? If the installation directory is a system-owned directory, you may need to sign in as the administrator or "root" account. If you do not have administrative access to this machine, you may wish to choose a different installation directory, preferably one that is listed in your PYTHONPATH environment variable. For information on other options, you may wish to consult the documentation at: https://setuptools.readthedocs.io/en/latest/easy_install.html Please make the appropriate changes for your system and try again. """).lstrip() def cant_write_to_target(self): msg = self.__cant_write_msg % (sys.exc_info()[1], self.install_dir,) if not os.path.exists(self.install_dir): msg += '\n' + self.__not_exists_id else: msg += '\n' + self.__access_msg raise DistutilsError(msg) def check_pth_processing(self): """Empirically verify whether .pth files are supported in inst. dir""" instdir = self.install_dir log.info("Checking .pth file support in %s", instdir) pth_file = self.pseudo_tempname() + ".pth" ok_file = pth_file + '.ok' ok_exists = os.path.exists(ok_file) tmpl = _one_liner(""" import os f = open({ok_file!r}, 'w') f.write('OK') f.close() """) + '\n' try: if ok_exists: os.unlink(ok_file) dirname = os.path.dirname(ok_file) pkg_resources.py31compat.makedirs(dirname, exist_ok=True) f = open(pth_file, 'w') except (OSError, IOError): self.cant_write_to_target() else: try: f.write(tmpl.format(**locals())) f.close() f = None executable = sys.executable if os.name == 'nt': dirname, basename = os.path.split(executable) alt = os.path.join(dirname, 'pythonw.exe') use_alt = ( basename.lower() == 'python.exe' and os.path.exists(alt) ) if use_alt: # use pythonw.exe to avoid opening a console window executable = alt from distutils.spawn import spawn spawn([executable, '-E', '-c', 'pass'], 0) if os.path.exists(ok_file): log.info( "TEST PASSED: %s appears to support .pth files", instdir ) return True finally: if f: f.close() if os.path.exists(ok_file): os.unlink(ok_file) if os.path.exists(pth_file): os.unlink(pth_file) if not self.multi_version: log.warn("TEST FAILED: %s does NOT support .pth files", instdir) return False def install_egg_scripts(self, dist): """Write all the scripts for `dist`, unless scripts are excluded""" if not self.exclude_scripts and dist.metadata_isdir('scripts'): for script_name in dist.metadata_listdir('scripts'): if dist.metadata_isdir('scripts/' + script_name): # The "script" is a directory, likely a Python 3 # __pycache__ directory, so skip it. continue self.install_script( dist, script_name, dist.get_metadata('scripts/' + script_name) ) self.install_wrapper_scripts(dist) def add_output(self, path): if os.path.isdir(path): for base, dirs, files in os.walk(path): for filename in files: self.outputs.append(os.path.join(base, filename)) else: self.outputs.append(path) def not_editable(self, spec): if self.editable: raise DistutilsArgError( "Invalid argument %r: you can't use filenames or URLs " "with --editable (except via the --find-links option)." % (spec,) ) def check_editable(self, spec): if not self.editable: return if os.path.exists(os.path.join(self.build_directory, spec.key)): raise DistutilsArgError( "%r already exists in %s; can't do a checkout there" % (spec.key, self.build_directory) ) @contextlib.contextmanager def _tmpdir(self): tmpdir = tempfile.mkdtemp(prefix=u"easy_install-") try: # cast to str as workaround for #709 and #710 and #712 yield str(tmpdir) finally: os.path.exists(tmpdir) and rmtree(rmtree_safe(tmpdir)) def easy_install(self, spec, deps=False): if not self.editable: self.install_site_py() with self._tmpdir() as tmpdir: if not isinstance(spec, Requirement): if URL_SCHEME(spec): # It's a url, download it to tmpdir and process self.not_editable(spec) dl = self.package_index.download(spec, tmpdir) return self.install_item(None, dl, tmpdir, deps, True) elif os.path.exists(spec): # Existing file or directory, just process it directly self.not_editable(spec) return self.install_item(None, spec, tmpdir, deps, True) else: spec = parse_requirement_arg(spec) self.check_editable(spec) dist = self.package_index.fetch_distribution( spec, tmpdir, self.upgrade, self.editable, not self.always_copy, self.local_index ) if dist is None: msg = "Could not find suitable distribution for %r" % spec if self.always_copy: msg += " (--always-copy skips system and development eggs)" raise DistutilsError(msg) elif dist.precedence == DEVELOP_DIST: # .egg-info dists don't need installing, just process deps self.process_distribution(spec, dist, deps, "Using") return dist else: return self.install_item(spec, dist.location, tmpdir, deps) def install_item(self, spec, download, tmpdir, deps, install_needed=False): # Installation is also needed if file in tmpdir or is not an egg install_needed = install_needed or self.always_copy install_needed = install_needed or os.path.dirname(download) == tmpdir install_needed = install_needed or not download.endswith('.egg') install_needed = install_needed or ( self.always_copy_from is not None and os.path.dirname(normalize_path(download)) == normalize_path(self.always_copy_from) ) if spec and not install_needed: # at this point, we know it's a local .egg, we just don't know if # it's already installed. for dist in self.local_index[spec.project_name]: if dist.location == download: break else: install_needed = True # it's not in the local index log.info("Processing %s", os.path.basename(download)) if install_needed: dists = self.install_eggs(spec, download, tmpdir) for dist in dists: self.process_distribution(spec, dist, deps) else: dists = [self.egg_distribution(download)] self.process_distribution(spec, dists[0], deps, "Using") if spec is not None: for dist in dists: if dist in spec: return dist def select_scheme(self, name): """Sets the install directories by applying the install schemes.""" # it's the caller's problem if they supply a bad name! scheme = INSTALL_SCHEMES[name] for key in SCHEME_KEYS: attrname = 'install_' + key if getattr(self, attrname) is None: setattr(self, attrname, scheme[key]) def process_distribution(self, requirement, dist, deps=True, *info): self.update_pth(dist) self.package_index.add(dist) if dist in self.local_index[dist.key]: self.local_index.remove(dist) self.local_index.add(dist) self.install_egg_scripts(dist) self.installed_projects[dist.key] = dist log.info(self.installation_report(requirement, dist, *info)) if (dist.has_metadata('dependency_links.txt') and not self.no_find_links): self.package_index.add_find_links( dist.get_metadata_lines('dependency_links.txt') ) if not deps and not self.always_copy: return elif requirement is not None and dist.key != requirement.key: log.warn("Skipping dependencies for %s", dist) return # XXX this is not the distribution we were looking for elif requirement is None or dist not in requirement: # if we wound up with a different version, resolve what we've got distreq = dist.as_requirement() requirement = Requirement(str(distreq)) log.info("Processing dependencies for %s", requirement) try: distros = WorkingSet([]).resolve( [requirement], self.local_index, self.easy_install ) except DistributionNotFound as e: raise DistutilsError(str(e)) except VersionConflict as e: raise DistutilsError(e.report()) if self.always_copy or self.always_copy_from: # Force all the relevant distros to be copied or activated for dist in distros: if dist.key not in self.installed_projects: self.easy_install(dist.as_requirement()) log.info("Finished processing dependencies for %s", requirement) def should_unzip(self, dist): if self.zip_ok is not None: return not self.zip_ok if dist.has_metadata('not-zip-safe'): return True if not dist.has_metadata('zip-safe'): return True return False def maybe_move(self, spec, dist_filename, setup_base): dst = os.path.join(self.build_directory, spec.key) if os.path.exists(dst): msg = ( "%r already exists in %s; build directory %s will not be kept" ) log.warn(msg, spec.key, self.build_directory, setup_base) return setup_base if os.path.isdir(dist_filename): setup_base = dist_filename else: if os.path.dirname(dist_filename) == setup_base: os.unlink(dist_filename) # get it out of the tmp dir contents = os.listdir(setup_base) if len(contents) == 1: dist_filename = os.path.join(setup_base, contents[0]) if os.path.isdir(dist_filename): # if the only thing there is a directory, move it instead setup_base = dist_filename ensure_directory(dst) shutil.move(setup_base, dst) return dst def install_wrapper_scripts(self, dist): if self.exclude_scripts: return for args in ScriptWriter.best().get_args(dist): self.write_script(*args) def install_script(self, dist, script_name, script_text, dev_path=None): """Generate a legacy script wrapper and install it""" spec = str(dist.as_requirement()) is_script = is_python_script(script_text, script_name) if is_script: body = self._load_template(dev_path) % locals() script_text = ScriptWriter.get_header(script_text) + body self.write_script(script_name, _to_bytes(script_text), 'b') @staticmethod def _load_template(dev_path): """ There are a couple of template scripts in the package. This function loads one of them and prepares it for use. """ # See https://github.com/pypa/setuptools/issues/134 for info # on script file naming and downstream issues with SVR4 name = 'script.tmpl' if dev_path: name = name.replace('.tmpl', ' (dev).tmpl') raw_bytes = resource_string('setuptools', name) return raw_bytes.decode('utf-8') def write_script(self, script_name, contents, mode="t", blockers=()): """Write an executable file to the scripts directory""" self.delete_blockers( # clean up old .py/.pyw w/o a script [os.path.join(self.script_dir, x) for x in blockers] ) log.info("Installing %s script to %s", script_name, self.script_dir) target = os.path.join(self.script_dir, script_name) self.add_output(target) if self.dry_run: return mask = current_umask() ensure_directory(target) if os.path.exists(target): os.unlink(target) with open(target, "w" + mode) as f: f.write(contents) chmod(target, 0o777 - mask) def install_eggs(self, spec, dist_filename, tmpdir): # .egg dirs or files are already built, so just return them if dist_filename.lower().endswith('.egg'): return [self.install_egg(dist_filename, tmpdir)] elif dist_filename.lower().endswith('.exe'): return [self.install_exe(dist_filename, tmpdir)] elif dist_filename.lower().endswith('.whl'): return [self.install_wheel(dist_filename, tmpdir)] # Anything else, try to extract and build setup_base = tmpdir if os.path.isfile(dist_filename) and not dist_filename.endswith('.py'): unpack_archive(dist_filename, tmpdir, self.unpack_progress) elif os.path.isdir(dist_filename): setup_base = os.path.abspath(dist_filename) if (setup_base.startswith(tmpdir) # something we downloaded and self.build_directory and spec is not None): setup_base = self.maybe_move(spec, dist_filename, setup_base) # Find the setup.py file setup_script = os.path.join(setup_base, 'setup.py') if not os.path.exists(setup_script): setups = glob(os.path.join(setup_base, '*', 'setup.py')) if not setups: raise DistutilsError( "Couldn't find a setup script in %s" % os.path.abspath(dist_filename) ) if len(setups) > 1: raise DistutilsError( "Multiple setup scripts in %s" % os.path.abspath(dist_filename) ) setup_script = setups[0] # Now run it, and return the result if self.editable: log.info(self.report_editable(spec, setup_script)) return [] else: return self.build_and_install(setup_script, setup_base) def egg_distribution(self, egg_path): if os.path.isdir(egg_path): metadata = PathMetadata(egg_path, os.path.join(egg_path, 'EGG-INFO')) else: metadata = EggMetadata(zipimport.zipimporter(egg_path)) return Distribution.from_filename(egg_path, metadata=metadata) def install_egg(self, egg_path, tmpdir): destination = os.path.join( self.install_dir, os.path.basename(egg_path), ) destination = os.path.abspath(destination) if not self.dry_run: ensure_directory(destination) dist = self.egg_distribution(egg_path) if not samefile(egg_path, destination): if os.path.isdir(destination) and not os.path.islink(destination): dir_util.remove_tree(destination, dry_run=self.dry_run) elif os.path.exists(destination): self.execute( os.unlink, (destination,), "Removing " + destination, ) try: new_dist_is_zipped = False if os.path.isdir(egg_path): if egg_path.startswith(tmpdir): f, m = shutil.move, "Moving" else: f, m = shutil.copytree, "Copying" elif self.should_unzip(dist): self.mkpath(destination) f, m = self.unpack_and_compile, "Extracting" else: new_dist_is_zipped = True if egg_path.startswith(tmpdir): f, m = shutil.move, "Moving" else: f, m = shutil.copy2, "Copying" self.execute( f, (egg_path, destination), (m + " %s to %s") % ( os.path.basename(egg_path), os.path.dirname(destination) ), ) update_dist_caches( destination, fix_zipimporter_caches=new_dist_is_zipped, ) except Exception: update_dist_caches(destination, fix_zipimporter_caches=False) raise self.add_output(destination) return self.egg_distribution(destination) def install_exe(self, dist_filename, tmpdir): # See if it's valid, get data cfg = extract_wininst_cfg(dist_filename) if cfg is None: raise DistutilsError( "%s is not a valid distutils Windows .exe" % dist_filename ) # Create a dummy distribution object until we build the real distro dist = Distribution( None, project_name=cfg.get('metadata', 'name'), version=cfg.get('metadata', 'version'), platform=get_platform(), ) # Convert the .exe to an unpacked egg egg_path = os.path.join(tmpdir, dist.egg_name() + '.egg') dist.location = egg_path egg_tmp = egg_path + '.tmp' _egg_info = os.path.join(egg_tmp, 'EGG-INFO') pkg_inf = os.path.join(_egg_info, 'PKG-INFO') ensure_directory(pkg_inf) # make sure EGG-INFO dir exists dist._provider = PathMetadata(egg_tmp, _egg_info) # XXX self.exe_to_egg(dist_filename, egg_tmp) # Write EGG-INFO/PKG-INFO if not os.path.exists(pkg_inf): f = open(pkg_inf, 'w') f.write('Metadata-Version: 1.0\n') for k, v in cfg.items('metadata'): if k != 'target_version': f.write('%s: %s\n' % (k.replace('_', '-').title(), v)) f.close() script_dir = os.path.join(_egg_info, 'scripts') # delete entry-point scripts to avoid duping self.delete_blockers([ os.path.join(script_dir, args[0]) for args in ScriptWriter.get_args(dist) ]) # Build .egg file from tmpdir bdist_egg.make_zipfile( egg_path, egg_tmp, verbose=self.verbose, dry_run=self.dry_run, ) # install the .egg return self.install_egg(egg_path, tmpdir) def exe_to_egg(self, dist_filename, egg_tmp): """Extract a bdist_wininst to the directories an egg would use""" # Check for .pth file and set up prefix translations prefixes = get_exe_prefixes(dist_filename) to_compile = [] native_libs = [] top_level = {} def process(src, dst): s = src.lower() for old, new in prefixes: if s.startswith(old): src = new + src[len(old):] parts = src.split('/') dst = os.path.join(egg_tmp, *parts) dl = dst.lower() if dl.endswith('.pyd') or dl.endswith('.dll'): parts[-1] = bdist_egg.strip_module(parts[-1]) top_level[os.path.splitext(parts[0])[0]] = 1 native_libs.append(src) elif dl.endswith('.py') and old != 'SCRIPTS/': top_level[os.path.splitext(parts[0])[0]] = 1 to_compile.append(dst) return dst if not src.endswith('.pth'): log.warn("WARNING: can't process %s", src) return None # extract, tracking .pyd/.dll->native_libs and .py -> to_compile unpack_archive(dist_filename, egg_tmp, process) stubs = [] for res in native_libs: if res.lower().endswith('.pyd'): # create stubs for .pyd's parts = res.split('/') resource = parts[-1] parts[-1] = bdist_egg.strip_module(parts[-1]) + '.py' pyfile = os.path.join(egg_tmp, *parts) to_compile.append(pyfile) stubs.append(pyfile) bdist_egg.write_stub(resource, pyfile) self.byte_compile(to_compile) # compile .py's bdist_egg.write_safety_flag( os.path.join(egg_tmp, 'EGG-INFO'), bdist_egg.analyze_egg(egg_tmp, stubs)) # write zip-safety flag for name in 'top_level', 'native_libs': if locals()[name]: txt = os.path.join(egg_tmp, 'EGG-INFO', name + '.txt') if not os.path.exists(txt): f = open(txt, 'w') f.write('\n'.join(locals()[name]) + '\n') f.close() def install_wheel(self, wheel_path, tmpdir): wheel = Wheel(wheel_path) assert wheel.is_compatible() destination = os.path.join(self.install_dir, wheel.egg_name()) destination = os.path.abspath(destination) if not self.dry_run: ensure_directory(destination) if os.path.isdir(destination) and not os.path.islink(destination): dir_util.remove_tree(destination, dry_run=self.dry_run) elif os.path.exists(destination): self.execute( os.unlink, (destination,), "Removing " + destination, ) try: self.execute( wheel.install_as_egg, (destination,), ("Installing %s to %s") % ( os.path.basename(wheel_path), os.path.dirname(destination) ), ) finally: update_dist_caches(destination, fix_zipimporter_caches=False) self.add_output(destination) return self.egg_distribution(destination) __mv_warning = textwrap.dedent(""" Because this distribution was installed --multi-version, before you can import modules from this package in an application, you will need to 'import pkg_resources' and then use a 'require()' call similar to one of these examples, in order to select the desired version: pkg_resources.require("%(name)s") # latest installed version pkg_resources.require("%(name)s==%(version)s") # this exact version pkg_resources.require("%(name)s>=%(version)s") # this version or higher """).lstrip() __id_warning = textwrap.dedent(""" Note also that the installation directory must be on sys.path at runtime for this to work. (e.g. by being the application's script directory, by being on PYTHONPATH, or by being added to sys.path by your code.) """) def installation_report(self, req, dist, what="Installed"): """Helpful installation message for display to package users""" msg = "\n%(what)s %(eggloc)s%(extras)s" if self.multi_version and not self.no_report: msg += '\n' + self.__mv_warning if self.install_dir not in map(normalize_path, sys.path): msg += '\n' + self.__id_warning eggloc = dist.location name = dist.project_name version = dist.version extras = '' # TODO: self.report_extras(req, dist) return msg % locals() __editable_msg = textwrap.dedent(""" Extracted editable version of %(spec)s to %(dirname)s If it uses setuptools in its setup script, you can activate it in "development" mode by going to that directory and running:: %(python)s setup.py develop See the setuptools documentation for the "develop" command for more info. """).lstrip() def report_editable(self, spec, setup_script): dirname = os.path.dirname(setup_script) python = sys.executable return '\n' + self.__editable_msg % locals() def run_setup(self, setup_script, setup_base, args): sys.modules.setdefault('distutils.command.bdist_egg', bdist_egg) sys.modules.setdefault('distutils.command.egg_info', egg_info) args = list(args) if self.verbose > 2: v = 'v' * (self.verbose - 1) args.insert(0, '-' + v) elif self.verbose < 2: args.insert(0, '-q') if self.dry_run: args.insert(0, '-n') log.info( "Running %s %s", setup_script[len(setup_base) + 1:], ' '.join(args) ) try: run_setup(setup_script, args) except SystemExit as v: raise DistutilsError("Setup script exited with %s" % (v.args[0],)) def build_and_install(self, setup_script, setup_base): args = ['bdist_egg', '--dist-dir'] dist_dir = tempfile.mkdtemp( prefix='egg-dist-tmp-', dir=os.path.dirname(setup_script) ) try: self._set_fetcher_options(os.path.dirname(setup_script)) args.append(dist_dir) self.run_setup(setup_script, setup_base, args) all_eggs = Environment([dist_dir]) eggs = [] for key in all_eggs: for dist in all_eggs[key]: eggs.append(self.install_egg(dist.location, setup_base)) if not eggs and not self.dry_run: log.warn("No eggs found in %s (setup script problem?)", dist_dir) return eggs finally: rmtree(dist_dir) log.set_verbosity(self.verbose) # restore our log verbosity def _set_fetcher_options(self, base): """ When easy_install is about to run bdist_egg on a source dist, that source dist might have 'setup_requires' directives, requiring additional fetching. Ensure the fetcher options given to easy_install are available to that command as well. """ # find the fetch options from easy_install and write them out # to the setup.cfg file. ei_opts = self.distribution.get_option_dict('easy_install').copy() fetch_directives = ( 'find_links', 'site_dirs', 'index_url', 'optimize', 'allow_hosts', ) fetch_options = {} for key, val in ei_opts.items(): if key not in fetch_directives: continue fetch_options[key.replace('_', '-')] = val[1] # create a settings dictionary suitable for `edit_config` settings = dict(easy_install=fetch_options) cfg_filename = os.path.join(base, 'setup.cfg') setopt.edit_config(cfg_filename, settings) def update_pth(self, dist): if self.pth_file is None: return for d in self.pth_file[dist.key]: # drop old entries if self.multi_version or d.location != dist.location: log.info("Removing %s from easy-install.pth file", d) self.pth_file.remove(d) if d.location in self.shadow_path: self.shadow_path.remove(d.location) if not self.multi_version: if dist.location in self.pth_file.paths: log.info( "%s is already the active version in easy-install.pth", dist, ) else: log.info("Adding %s to easy-install.pth file", dist) self.pth_file.add(dist) # add new entry if dist.location not in self.shadow_path: self.shadow_path.append(dist.location) if not self.dry_run: self.pth_file.save() if dist.key == 'setuptools': # Ensure that setuptools itself never becomes unavailable! # XXX should this check for latest version? filename = os.path.join(self.install_dir, 'setuptools.pth') if os.path.islink(filename): os.unlink(filename) f = open(filename, 'wt') f.write(self.pth_file.make_relative(dist.location) + '\n') f.close() def unpack_progress(self, src, dst): # Progress filter for unpacking log.debug("Unpacking %s to %s", src, dst) return dst # only unpack-and-compile skips files for dry run def unpack_and_compile(self, egg_path, destination): to_compile = [] to_chmod = [] def pf(src, dst): if dst.endswith('.py') and not src.startswith('EGG-INFO/'): to_compile.append(dst) elif dst.endswith('.dll') or dst.endswith('.so'): to_chmod.append(dst) self.unpack_progress(src, dst) return not self.dry_run and dst or None unpack_archive(egg_path, destination, pf) self.byte_compile(to_compile) if not self.dry_run: for f in to_chmod: mode = ((os.stat(f)[stat.ST_MODE]) | 0o555) & 0o7755 chmod(f, mode) def byte_compile(self, to_compile): if sys.dont_write_bytecode: return from distutils.util import byte_compile try: # try to make the byte compile messages quieter log.set_verbosity(self.verbose - 1) byte_compile(to_compile, optimize=0, force=1, dry_run=self.dry_run) if self.optimize: byte_compile( to_compile, optimize=self.optimize, force=1, dry_run=self.dry_run, ) finally: log.set_verbosity(self.verbose) # restore original verbosity __no_default_msg = textwrap.dedent(""" bad install directory or PYTHONPATH You are attempting to install a package to a directory that is not on PYTHONPATH and which Python does not read ".pth" files from. The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s and your PYTHONPATH environment variable currently contains: %r Here are some of your options for correcting the problem: * You can choose a different installation directory, i.e., one that is on PYTHONPATH or supports .pth files * You can add the installation directory to the PYTHONPATH environment variable. (It must then also be on PYTHONPATH whenever you run Python and want to use the package(s) you are installing.) * You can set up the installation directory to support ".pth" files by using one of the approaches described here: https://setuptools.readthedocs.io/en/latest/easy_install.html#custom-installation-locations Please make the appropriate changes for your system and try again.""").lstrip() def no_default_version_msg(self): template = self.__no_default_msg return template % (self.install_dir, os.environ.get('PYTHONPATH', '')) def install_site_py(self): """Make sure there's a site.py in the target dir, if needed""" if self.sitepy_installed: return # already did it, or don't need to sitepy = os.path.join(self.install_dir, "site.py") source = resource_string("setuptools", "site-patch.py") source = source.decode('utf-8') current = "" if os.path.exists(sitepy): log.debug("Checking existing site.py in %s", self.install_dir) with io.open(sitepy) as strm: current = strm.read() if not current.startswith('def __boot():'): raise DistutilsError( "%s is not a setuptools-generated site.py; please" " remove it." % sitepy ) if current != source: log.info("Creating %s", sitepy) if not self.dry_run: ensure_directory(sitepy) with io.open(sitepy, 'w', encoding='utf-8') as strm: strm.write(source) self.byte_compile([sitepy]) self.sitepy_installed = True def create_home_path(self): """Create directories under ~.""" if not self.user: return home = convert_path(os.path.expanduser("~")) for name, path in six.iteritems(self.config_vars): if path.startswith(home) and not os.path.isdir(path): self.debug_print("os.makedirs('%s', 0o700)" % path) os.makedirs(path, 0o700) INSTALL_SCHEMES = dict( posix=dict( install_dir='$base/lib/python$py_version_short/site-packages', script_dir='$base/bin', ), ) DEFAULT_SCHEME = dict( install_dir='$base/Lib/site-packages', script_dir='$base/Scripts', ) def _expand(self, *attrs): config_vars = self.get_finalized_command('install').config_vars if self.prefix: # Set default install_dir/scripts from --prefix config_vars = config_vars.copy() config_vars['base'] = self.prefix scheme = self.INSTALL_SCHEMES.get(os.name, self.DEFAULT_SCHEME) for attr, val in scheme.items(): if getattr(self, attr, None) is None: setattr(self, attr, val) from distutils.util import subst_vars for attr in attrs: val = getattr(self, attr) if val is not None: val = subst_vars(val, config_vars) if os.name == 'posix': val = os.path.expanduser(val) setattr(self, attr, val) def _pythonpath(): items = os.environ.get('PYTHONPATH', '').split(os.pathsep) return filter(None, items) def get_site_dirs(): """ Return a list of 'site' dirs """ sitedirs = [] # start with PYTHONPATH sitedirs.extend(_pythonpath()) prefixes = [sys.prefix] if sys.exec_prefix != sys.prefix: prefixes.append(sys.exec_prefix) for prefix in prefixes: if prefix: if sys.platform in ('os2emx', 'riscos'): sitedirs.append(os.path.join(prefix, "Lib", "site-packages")) elif os.sep == '/': sitedirs.extend([ os.path.join( prefix, "lib", "python{}.{}".format(*sys.version_info), "site-packages", ), os.path.join(prefix, "lib", "site-python"), ]) else: sitedirs.extend([ prefix, os.path.join(prefix, "lib", "site-packages"), ]) if sys.platform == 'darwin': # for framework builds *only* we add the standard Apple # locations. Currently only per-user, but /Library and # /Network/Library could be added too if 'Python.framework' in prefix: home = os.environ.get('HOME') if home: home_sp = os.path.join( home, 'Library', 'Python', '{}.{}'.format(*sys.version_info), 'site-packages', ) sitedirs.append(home_sp) lib_paths = get_path('purelib'), get_path('platlib') for site_lib in lib_paths: if site_lib not in sitedirs: sitedirs.append(site_lib) if site.ENABLE_USER_SITE: sitedirs.append(site.USER_SITE) try: sitedirs.extend(site.getsitepackages()) except AttributeError: pass sitedirs = list(map(normalize_path, sitedirs)) return sitedirs def expand_paths(inputs): """Yield sys.path directories that might contain "old-style" packages""" seen = {} for dirname in inputs: dirname = normalize_path(dirname) if dirname in seen: continue seen[dirname] = 1 if not os.path.isdir(dirname): continue files = os.listdir(dirname) yield dirname, files for name in files: if not name.endswith('.pth'): # We only care about the .pth files continue if name in ('easy-install.pth', 'setuptools.pth'): # Ignore .pth files that we control continue # Read the .pth file f = open(os.path.join(dirname, name)) lines = list(yield_lines(f)) f.close() # Yield existing non-dupe, non-import directory lines from it for line in lines: if not line.startswith("import"): line = normalize_path(line.rstrip()) if line not in seen: seen[line] = 1 if not os.path.isdir(line): continue yield line, os.listdir(line) def extract_wininst_cfg(dist_filename): """Extract configuration data from a bdist_wininst .exe Returns a configparser.RawConfigParser, or None """ f = open(dist_filename, 'rb') try: endrec = zipfile._EndRecData(f) if endrec is None: return None prepended = (endrec[9] - endrec[5]) - endrec[6] if prepended < 12: # no wininst data here return None f.seek(prepended - 12) tag, cfglen, bmlen = struct.unpack("egg path translations for a given .exe file""" prefixes = [ ('PURELIB/', ''), ('PLATLIB/pywin32_system32', ''), ('PLATLIB/', ''), ('SCRIPTS/', 'EGG-INFO/scripts/'), ('DATA/lib/site-packages', ''), ] z = zipfile.ZipFile(exe_filename) try: for info in z.infolist(): name = info.filename parts = name.split('/') if len(parts) == 3 and parts[2] == 'PKG-INFO': if parts[1].endswith('.egg-info'): prefixes.insert(0, ('/'.join(parts[:2]), 'EGG-INFO/')) break if len(parts) != 2 or not name.endswith('.pth'): continue if name.endswith('-nspkg.pth'): continue if parts[0].upper() in ('PURELIB', 'PLATLIB'): contents = z.read(name) if six.PY3: contents = contents.decode() for pth in yield_lines(contents): pth = pth.strip().replace('\\', '/') if not pth.startswith('import'): prefixes.append((('%s/%s/' % (parts[0], pth)), '')) finally: z.close() prefixes = [(x.lower(), y) for x, y in prefixes] prefixes.sort() prefixes.reverse() return prefixes class PthDistributions(Environment): """A .pth file with Distribution paths in it""" dirty = False def __init__(self, filename, sitedirs=()): self.filename = filename self.sitedirs = list(map(normalize_path, sitedirs)) self.basedir = normalize_path(os.path.dirname(self.filename)) self._load() Environment.__init__(self, [], None, None) for path in yield_lines(self.paths): list(map(self.add, find_distributions(path, True))) def _load(self): self.paths = [] saw_import = False seen = dict.fromkeys(self.sitedirs) if os.path.isfile(self.filename): f = open(self.filename, 'rt') for line in f: if line.startswith('import'): saw_import = True continue path = line.rstrip() self.paths.append(path) if not path.strip() or path.strip().startswith('#'): continue # skip non-existent paths, in case somebody deleted a package # manually, and duplicate paths as well path = self.paths[-1] = normalize_path( os.path.join(self.basedir, path) ) if not os.path.exists(path) or path in seen: self.paths.pop() # skip it self.dirty = True # we cleaned up, so we're dirty now :) continue seen[path] = 1 f.close() if self.paths and not saw_import: self.dirty = True # ensure anything we touch has import wrappers while self.paths and not self.paths[-1].strip(): self.paths.pop() def save(self): """Write changed .pth file back to disk""" if not self.dirty: return rel_paths = list(map(self.make_relative, self.paths)) if rel_paths: log.debug("Saving %s", self.filename) lines = self._wrap_lines(rel_paths) data = '\n'.join(lines) + '\n' if os.path.islink(self.filename): os.unlink(self.filename) with open(self.filename, 'wt') as f: f.write(data) elif os.path.exists(self.filename): log.debug("Deleting empty %s", self.filename) os.unlink(self.filename) self.dirty = False @staticmethod def _wrap_lines(lines): return lines def add(self, dist): """Add `dist` to the distribution map""" new_path = ( dist.location not in self.paths and ( dist.location not in self.sitedirs or # account for '.' being in PYTHONPATH dist.location == os.getcwd() ) ) if new_path: self.paths.append(dist.location) self.dirty = True Environment.add(self, dist) def remove(self, dist): """Remove `dist` from the distribution map""" while dist.location in self.paths: self.paths.remove(dist.location) self.dirty = True Environment.remove(self, dist) def make_relative(self, path): npath, last = os.path.split(normalize_path(path)) baselen = len(self.basedir) parts = [last] sep = os.altsep == '/' and '/' or os.sep while len(npath) >= baselen: if npath == self.basedir: parts.append(os.curdir) parts.reverse() return sep.join(parts) npath, last = os.path.split(npath) parts.append(last) else: return path class RewritePthDistributions(PthDistributions): @classmethod def _wrap_lines(cls, lines): yield cls.prelude for line in lines: yield line yield cls.postlude prelude = _one_liner(""" import sys sys.__plen = len(sys.path) """) postlude = _one_liner(""" import sys new = sys.path[sys.__plen:] del sys.path[sys.__plen:] p = getattr(sys, '__egginsert', 0) sys.path[p:p] = new sys.__egginsert = p + len(new) """) if os.environ.get('SETUPTOOLS_SYS_PATH_TECHNIQUE', 'raw') == 'rewrite': PthDistributions = RewritePthDistributions def _first_line_re(): """ Return a regular expression based on first_line_re suitable for matching strings. """ if isinstance(first_line_re.pattern, str): return first_line_re # first_line_re in Python >=3.1.4 and >=3.2.1 is a bytes pattern. return re.compile(first_line_re.pattern.decode()) def auto_chmod(func, arg, exc): if func in [os.unlink, os.remove] and os.name == 'nt': chmod(arg, stat.S_IWRITE) return func(arg) et, ev, _ = sys.exc_info() six.reraise(et, (ev[0], ev[1] + (" %s %s" % (func, arg)))) def update_dist_caches(dist_path, fix_zipimporter_caches): """ Fix any globally cached `dist_path` related data `dist_path` should be a path of a newly installed egg distribution (zipped or unzipped). sys.path_importer_cache contains finder objects that have been cached when importing data from the original distribution. Any such finders need to be cleared since the replacement distribution might be packaged differently, e.g. a zipped egg distribution might get replaced with an unzipped egg folder or vice versa. Having the old finders cached may then cause Python to attempt loading modules from the replacement distribution using an incorrect loader. zipimport.zipimporter objects are Python loaders charged with importing data packaged inside zip archives. If stale loaders referencing the original distribution, are left behind, they can fail to load modules from the replacement distribution. E.g. if an old zipimport.zipimporter instance is used to load data from a new zipped egg archive, it may cause the operation to attempt to locate the requested data in the wrong location - one indicated by the original distribution's zip archive directory information. Such an operation may then fail outright, e.g. report having read a 'bad local file header', or even worse, it may fail silently & return invalid data. zipimport._zip_directory_cache contains cached zip archive directory information for all existing zipimport.zipimporter instances and all such instances connected to the same archive share the same cached directory information. If asked, and the underlying Python implementation allows it, we can fix all existing zipimport.zipimporter instances instead of having to track them down and remove them one by one, by updating their shared cached zip archive directory information. This, of course, assumes that the replacement distribution is packaged as a zipped egg. If not asked to fix existing zipimport.zipimporter instances, we still do our best to clear any remaining zipimport.zipimporter related cached data that might somehow later get used when attempting to load data from the new distribution and thus cause such load operations to fail. Note that when tracking down such remaining stale data, we can not catch every conceivable usage from here, and we clear only those that we know of and have found to cause problems if left alive. Any remaining caches should be updated by whomever is in charge of maintaining them, i.e. they should be ready to handle us replacing their zip archives with new distributions at runtime. """ # There are several other known sources of stale zipimport.zipimporter # instances that we do not clear here, but might if ever given a reason to # do so: # * Global setuptools pkg_resources.working_set (a.k.a. 'master working # set') may contain distributions which may in turn contain their # zipimport.zipimporter loaders. # * Several zipimport.zipimporter loaders held by local variables further # up the function call stack when running the setuptools installation. # * Already loaded modules may have their __loader__ attribute set to the # exact loader instance used when importing them. Python 3.4 docs state # that this information is intended mostly for introspection and so is # not expected to cause us problems. normalized_path = normalize_path(dist_path) _uncache(normalized_path, sys.path_importer_cache) if fix_zipimporter_caches: _replace_zip_directory_cache_data(normalized_path) else: # Here, even though we do not want to fix existing and now stale # zipimporter cache information, we still want to remove it. Related to # Python's zip archive directory information cache, we clear each of # its stale entries in two phases: # 1. Clear the entry so attempting to access zip archive information # via any existing stale zipimport.zipimporter instances fails. # 2. Remove the entry from the cache so any newly constructed # zipimport.zipimporter instances do not end up using old stale # zip archive directory information. # This whole stale data removal step does not seem strictly necessary, # but has been left in because it was done before we started replacing # the zip archive directory information cache content if possible, and # there are no relevant unit tests that we can depend on to tell us if # this is really needed. _remove_and_clear_zip_directory_cache_data(normalized_path) def _collect_zipimporter_cache_entries(normalized_path, cache): """ Return zipimporter cache entry keys related to a given normalized path. Alternative path spellings (e.g. those using different character case or those using alternative path separators) related to the same path are included. Any sub-path entries are included as well, i.e. those corresponding to zip archives embedded in other zip archives. """ result = [] prefix_len = len(normalized_path) for p in cache: np = normalize_path(p) if (np.startswith(normalized_path) and np[prefix_len:prefix_len + 1] in (os.sep, '')): result.append(p) return result def _update_zipimporter_cache(normalized_path, cache, updater=None): """ Update zipimporter cache data for a given normalized path. Any sub-path entries are processed as well, i.e. those corresponding to zip archives embedded in other zip archives. Given updater is a callable taking a cache entry key and the original entry (after already removing the entry from the cache), and expected to update the entry and possibly return a new one to be inserted in its place. Returning None indicates that the entry should not be replaced with a new one. If no updater is given, the cache entries are simply removed without any additional processing, the same as if the updater simply returned None. """ for p in _collect_zipimporter_cache_entries(normalized_path, cache): # N.B. pypy's custom zipimport._zip_directory_cache implementation does # not support the complete dict interface: # * Does not support item assignment, thus not allowing this function # to be used only for removing existing cache entries. # * Does not support the dict.pop() method, forcing us to use the # get/del patterns instead. For more detailed information see the # following links: # https://github.com/pypa/setuptools/issues/202#issuecomment-202913420 # http://bit.ly/2h9itJX old_entry = cache[p] del cache[p] new_entry = updater and updater(p, old_entry) if new_entry is not None: cache[p] = new_entry def _uncache(normalized_path, cache): _update_zipimporter_cache(normalized_path, cache) def _remove_and_clear_zip_directory_cache_data(normalized_path): def clear_and_remove_cached_zip_archive_directory_data(path, old_entry): old_entry.clear() _update_zipimporter_cache( normalized_path, zipimport._zip_directory_cache, updater=clear_and_remove_cached_zip_archive_directory_data) # PyPy Python implementation does not allow directly writing to the # zipimport._zip_directory_cache and so prevents us from attempting to correct # its content. The best we can do there is clear the problematic cache content # and have PyPy repopulate it as needed. The downside is that if there are any # stale zipimport.zipimporter instances laying around, attempting to use them # will fail due to not having its zip archive directory information available # instead of being automatically corrected to use the new correct zip archive # directory information. if '__pypy__' in sys.builtin_module_names: _replace_zip_directory_cache_data = \ _remove_and_clear_zip_directory_cache_data else: def _replace_zip_directory_cache_data(normalized_path): def replace_cached_zip_archive_directory_data(path, old_entry): # N.B. In theory, we could load the zip directory information just # once for all updated path spellings, and then copy it locally and # update its contained path strings to contain the correct # spelling, but that seems like a way too invasive move (this cache # structure is not officially documented anywhere and could in # theory change with new Python releases) for no significant # benefit. old_entry.clear() zipimport.zipimporter(path) old_entry.update(zipimport._zip_directory_cache[path]) return old_entry _update_zipimporter_cache( normalized_path, zipimport._zip_directory_cache, updater=replace_cached_zip_archive_directory_data) def is_python(text, filename=''): "Is this string a valid Python script?" try: compile(text, filename, 'exec') except (SyntaxError, TypeError): return False else: return True def is_sh(executable): """Determine if the specified executable is a .sh (contains a #! line)""" try: with io.open(executable, encoding='latin-1') as fp: magic = fp.read(2) except (OSError, IOError): return executable return magic == '#!' def nt_quote_arg(arg): """Quote a command line argument according to Windows parsing rules""" return subprocess.list2cmdline([arg]) def is_python_script(script_text, filename): """Is this text, as a whole, a Python script? (as opposed to shell/bat/etc. """ if filename.endswith('.py') or filename.endswith('.pyw'): return True # extension says it's Python if is_python(script_text, filename): return True # it's syntactically valid Python if script_text.startswith('#!'): # It begins with a '#!' line, so check if 'python' is in it somewhere return 'python' in script_text.splitlines()[0].lower() return False # Not any Python I can recognize try: from os import chmod as _chmod except ImportError: # Jython compatibility def _chmod(*args): pass def chmod(path, mode): log.debug("changing mode of %s to %o", path, mode) try: _chmod(path, mode) except os.error as e: log.debug("chmod failed: %s", e) class CommandSpec(list): """ A command spec for a #! header, specified as a list of arguments akin to those passed to Popen. """ options = [] split_args = dict() @classmethod def best(cls): """ Choose the best CommandSpec class based on environmental conditions. """ return cls @classmethod def _sys_executable(cls): _default = os.path.normpath(sys.executable) return os.environ.get('__PYVENV_LAUNCHER__', _default) @classmethod def from_param(cls, param): """ Construct a CommandSpec from a parameter to build_scripts, which may be None. """ if isinstance(param, cls): return param if isinstance(param, list): return cls(param) if param is None: return cls.from_environment() # otherwise, assume it's a string. return cls.from_string(param) @classmethod def from_environment(cls): return cls([cls._sys_executable()]) @classmethod def from_string(cls, string): """ Construct a command spec from a simple string representing a command line parseable by shlex.split. """ items = shlex.split(string, **cls.split_args) return cls(items) def install_options(self, script_text): self.options = shlex.split(self._extract_options(script_text)) cmdline = subprocess.list2cmdline(self) if not isascii(cmdline): self.options[:0] = ['-x'] @staticmethod def _extract_options(orig_script): """ Extract any options from the first line of the script. """ first = (orig_script + '\n').splitlines()[0] match = _first_line_re().match(first) options = match.group(1) or '' if match else '' return options.strip() def as_header(self): return self._render(self + list(self.options)) @staticmethod def _strip_quotes(item): _QUOTES = '"\'' for q in _QUOTES: if item.startswith(q) and item.endswith(q): return item[1:-1] return item @staticmethod def _render(items): cmdline = subprocess.list2cmdline( CommandSpec._strip_quotes(item.strip()) for item in items) return '#!' + cmdline + '\n' # For pbr compat; will be removed in a future version. sys_executable = CommandSpec._sys_executable() class WindowsCommandSpec(CommandSpec): split_args = dict(posix=False) class ScriptWriter: """ Encapsulates behavior around writing entry point scripts for console and gui apps. """ template = textwrap.dedent(r""" # EASY-INSTALL-ENTRY-SCRIPT: %(spec)r,%(group)r,%(name)r __requires__ = %(spec)r import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point(%(spec)r, %(group)r, %(name)r)() ) """).lstrip() command_spec_class = CommandSpec @classmethod def get_script_args(cls, dist, executable=None, wininst=False): # for backward compatibility warnings.warn("Use get_args", EasyInstallDeprecationWarning) writer = (WindowsScriptWriter if wininst else ScriptWriter).best() header = cls.get_script_header("", executable, wininst) return writer.get_args(dist, header) @classmethod def get_script_header(cls, script_text, executable=None, wininst=False): # for backward compatibility warnings.warn("Use get_header", EasyInstallDeprecationWarning, stacklevel=2) if wininst: executable = "python.exe" return cls.get_header(script_text, executable) @classmethod def get_args(cls, dist, header=None): """ Yield write_script() argument tuples for a distribution's console_scripts and gui_scripts entry points. """ if header is None: header = cls.get_header() spec = str(dist.as_requirement()) for type_ in 'console', 'gui': group = type_ + '_scripts' for name, ep in dist.get_entry_map(group).items(): cls._ensure_safe_name(name) script_text = cls.template % locals() args = cls._get_script_args(type_, name, header, script_text) for res in args: yield res @staticmethod def _ensure_safe_name(name): """ Prevent paths in *_scripts entry point names. """ has_path_sep = re.search(r'[\\/]', name) if has_path_sep: raise ValueError("Path separators not allowed in script names") @classmethod def get_writer(cls, force_windows): # for backward compatibility warnings.warn("Use best", EasyInstallDeprecationWarning) return WindowsScriptWriter.best() if force_windows else cls.best() @classmethod def best(cls): """ Select the best ScriptWriter for this environment. """ if sys.platform == 'win32' or (os.name == 'java' and os._name == 'nt'): return WindowsScriptWriter.best() else: return cls @classmethod def _get_script_args(cls, type_, name, header, script_text): # Simply write the stub with no extension. yield (name, header + script_text) @classmethod def get_header(cls, script_text="", executable=None): """Create a #! line, getting options (if any) from script_text""" cmd = cls.command_spec_class.best().from_param(executable) cmd.install_options(script_text) return cmd.as_header() class WindowsScriptWriter(ScriptWriter): command_spec_class = WindowsCommandSpec @classmethod def get_writer(cls): # for backward compatibility warnings.warn("Use best", EasyInstallDeprecationWarning) return cls.best() @classmethod def best(cls): """ Select the best ScriptWriter suitable for Windows """ writer_lookup = dict( executable=WindowsExecutableLauncherWriter, natural=cls, ) # for compatibility, use the executable launcher by default launcher = os.environ.get('SETUPTOOLS_LAUNCHER', 'executable') return writer_lookup[launcher] @classmethod def _get_script_args(cls, type_, name, header, script_text): "For Windows, add a .py extension" ext = dict(console='.pya', gui='.pyw')[type_] if ext not in os.environ['PATHEXT'].lower().split(';'): msg = ( "{ext} not listed in PATHEXT; scripts will not be " "recognized as executables." ).format(**locals()) warnings.warn(msg, UserWarning) old = ['.pya', '.py', '-script.py', '.pyc', '.pyo', '.pyw', '.exe'] old.remove(ext) header = cls._adjust_header(type_, header) blockers = [name + x for x in old] yield name + ext, header + script_text, 't', blockers @classmethod def _adjust_header(cls, type_, orig_header): """ Make sure 'pythonw' is used for gui and and 'python' is used for console (regardless of what sys.executable is). """ pattern = 'pythonw.exe' repl = 'python.exe' if type_ == 'gui': pattern, repl = repl, pattern pattern_ob = re.compile(re.escape(pattern), re.IGNORECASE) new_header = pattern_ob.sub(string=orig_header, repl=repl) return new_header if cls._use_header(new_header) else orig_header @staticmethod def _use_header(new_header): """ Should _adjust_header use the replaced header? On non-windows systems, always use. On Windows systems, only use the replaced header if it resolves to an executable on the system. """ clean_header = new_header[2:-1].strip('"') return sys.platform != 'win32' or find_executable(clean_header) class WindowsExecutableLauncherWriter(WindowsScriptWriter): @classmethod def _get_script_args(cls, type_, name, header, script_text): """ For Windows, add a .py extension and an .exe launcher """ if type_ == 'gui': launcher_type = 'gui' ext = '-script.pyw' old = ['.pyw'] else: launcher_type = 'cli' ext = '-script.py' old = ['.py', '.pyc', '.pyo'] hdr = cls._adjust_header(type_, header) blockers = [name + x for x in old] yield (name + ext, hdr + script_text, 't', blockers) yield ( name + '.exe', get_win_launcher(launcher_type), 'b' # write in binary mode ) if not is_64bit(): # install a manifest for the launcher to prevent Windows # from detecting it as an installer (which it will for # launchers like easy_install.exe). Consider only # adding a manifest for launchers detected as installers. # See Distribute #143 for details. m_name = name + '.exe.manifest' yield (m_name, load_launcher_manifest(name), 't') # for backward-compatibility get_script_args = ScriptWriter.get_script_args get_script_header = ScriptWriter.get_script_header def get_win_launcher(type): """ Load the Windows launcher (executable) suitable for launching a script. `type` should be either 'cli' or 'gui' Returns the executable as a byte string. """ launcher_fn = '%s.exe' % type if is_64bit(): launcher_fn = launcher_fn.replace(".", "-64.") else: launcher_fn = launcher_fn.replace(".", "-32.") return resource_string('setuptools', launcher_fn) def load_launcher_manifest(name): manifest = pkg_resources.resource_string(__name__, 'launcher manifest.xml') if six.PY2: return manifest % vars() else: return manifest.decode('utf-8') % vars() def rmtree(path, ignore_errors=False, onerror=auto_chmod): return shutil.rmtree(path, ignore_errors, onerror) def current_umask(): tmp = os.umask(0o022) os.umask(tmp) return tmp def bootstrap(): # This function is called when setuptools*.egg is run using /bin/sh import setuptools argv0 = os.path.dirname(setuptools.__path__[0]) sys.argv[0] = argv0 sys.argv.append(argv0) main() def main(argv=None, **kw): from setuptools import setup from setuptools.dist import Distribution class DistributionWithoutHelpCommands(Distribution): common_usage = "" def _show_help(self, *args, **kw): with _patch_usage(): Distribution._show_help(self, *args, **kw) if argv is None: argv = sys.argv[1:] with _patch_usage(): setup( script_args=['-q', 'easy_install', '-v'] + argv, script_name=sys.argv[0] or 'easy_install', distclass=DistributionWithoutHelpCommands, **kw ) @contextlib.contextmanager def _patch_usage(): import distutils.core USAGE = textwrap.dedent(""" usage: %(script)s [options] requirement_or_url ... or: %(script)s --help """).lstrip() def gen_usage(script_name): return USAGE % dict( script=os.path.basename(script_name), ) saved = distutils.core.gen_usage distutils.core.gen_usage = gen_usage try: yield finally: distutils.core.gen_usage = saved class EasyInstallDeprecationWarning(SetuptoolsDeprecationWarning): """Class for warning about deprecations in EasyInstall in SetupTools. Not ignored by default, unlike DeprecationWarning.""" site-packages/setuptools/command/rotate.py000064400000004164151464452710015011 0ustar00from distutils.util import convert_path from distutils import log from distutils.errors import DistutilsOptionError import os import shutil from setuptools.extern import six from setuptools import Command class rotate(Command): """Delete older distributions""" description = "delete older distributions, keeping N newest files" user_options = [ ('match=', 'm', "patterns to match (required)"), ('dist-dir=', 'd', "directory where the distributions are"), ('keep=', 'k', "number of matching distributions to keep"), ] boolean_options = [] def initialize_options(self): self.match = None self.dist_dir = None self.keep = None def finalize_options(self): if self.match is None: raise DistutilsOptionError( "Must specify one or more (comma-separated) match patterns " "(e.g. '.zip' or '.egg')" ) if self.keep is None: raise DistutilsOptionError("Must specify number of files to keep") try: self.keep = int(self.keep) except ValueError: raise DistutilsOptionError("--keep must be an integer") if isinstance(self.match, six.string_types): self.match = [ convert_path(p.strip()) for p in self.match.split(',') ] self.set_undefined_options('bdist', ('dist_dir', 'dist_dir')) def run(self): self.run_command("egg_info") from glob import glob for pattern in self.match: pattern = self.distribution.get_name() + '*' + pattern files = glob(os.path.join(self.dist_dir, pattern)) files = [(os.path.getmtime(f), f) for f in files] files.sort() files.reverse() log.info("%d file(s) matching %s", len(files), pattern) files = files[self.keep:] for (t, f) in files: log.info("Deleting %s", f) if not self.dry_run: if os.path.isdir(f): shutil.rmtree(f) else: os.unlink(f) site-packages/setuptools/command/install_scripts.py000064400000004607151464452710016732 0ustar00from distutils import log import distutils.command.install_scripts as orig import os import sys from pkg_resources import Distribution, PathMetadata, ensure_directory class install_scripts(orig.install_scripts): """Do normal script install, plus any egg_info wrapper scripts""" def initialize_options(self): orig.install_scripts.initialize_options(self) self.no_ep = False def run(self): import setuptools.command.easy_install as ei self.run_command("egg_info") if self.distribution.scripts: orig.install_scripts.run(self) # run first to set up self.outfiles else: self.outfiles = [] if self.no_ep: # don't install entry point scripts into .egg file! return ei_cmd = self.get_finalized_command("egg_info") dist = Distribution( ei_cmd.egg_base, PathMetadata(ei_cmd.egg_base, ei_cmd.egg_info), ei_cmd.egg_name, ei_cmd.egg_version, ) bs_cmd = self.get_finalized_command('build_scripts') exec_param = getattr(bs_cmd, 'executable', None) bw_cmd = self.get_finalized_command("bdist_wininst") is_wininst = getattr(bw_cmd, '_is_running', False) writer = ei.ScriptWriter if is_wininst: exec_param = "python.exe" writer = ei.WindowsScriptWriter if exec_param == sys.executable: # In case the path to the Python executable contains a space, wrap # it so it's not split up. exec_param = [exec_param] # resolve the writer to the environment writer = writer.best() cmd = writer.command_spec_class.best().from_param(exec_param) for args in writer.get_args(dist, cmd.as_header()): self.write_script(*args) def write_script(self, script_name, contents, mode="t", *ignored): """Write an executable file to the scripts directory""" from setuptools.command.easy_install import chmod, current_umask log.info("Installing %s script to %s", script_name, self.install_dir) target = os.path.join(self.install_dir, script_name) self.outfiles.append(target) mask = current_umask() if not self.dry_run: ensure_directory(target) f = open(target, "w" + mode) f.write(contents) f.close() chmod(target, 0o777 - mask) site-packages/setuptools/command/__init__.py000064400000001122151464452710015241 0ustar00__all__ = [ 'alias', 'bdist_egg', 'bdist_rpm', 'build_ext', 'build_py', 'develop', 'easy_install', 'egg_info', 'install', 'install_lib', 'rotate', 'saveopts', 'sdist', 'setopt', 'test', 'install_egg_info', 'install_scripts', 'register', 'bdist_wininst', 'upload_docs', 'upload', 'build_clib', 'dist_info', ] from distutils.command.bdist import bdist import sys from setuptools.command import install_scripts if 'egg' not in bdist.format_commands: bdist.format_command['egg'] = ('bdist_egg', "Python .egg file") bdist.format_commands.append('egg') del bdist, sys site-packages/setuptools/command/build_ext.py000064400000031333151464452710015470 0ustar00import os import sys import itertools from distutils.command.build_ext import build_ext as _du_build_ext from distutils.file_util import copy_file from distutils.ccompiler import new_compiler from distutils.sysconfig import customize_compiler, get_config_var from distutils.errors import DistutilsError from distutils import log from setuptools.extension import Library from setuptools.extern import six if six.PY2: import imp EXTENSION_SUFFIXES = [s for s, _, tp in imp.get_suffixes() if tp == imp.C_EXTENSION] else: from importlib.machinery import EXTENSION_SUFFIXES try: # Attempt to use Cython for building extensions, if available from Cython.Distutils.build_ext import build_ext as _build_ext # Additionally, assert that the compiler module will load # also. Ref #1229. __import__('Cython.Compiler.Main') except ImportError: _build_ext = _du_build_ext # make sure _config_vars is initialized get_config_var("LDSHARED") from distutils.sysconfig import _config_vars as _CONFIG_VARS def _customize_compiler_for_shlib(compiler): if sys.platform == "darwin": # building .dylib requires additional compiler flags on OSX; here we # temporarily substitute the pyconfig.h variables so that distutils' # 'customize_compiler' uses them before we build the shared libraries. tmp = _CONFIG_VARS.copy() try: # XXX Help! I don't have any idea whether these are right... _CONFIG_VARS['LDSHARED'] = ( "gcc -Wl,-x -dynamiclib -undefined dynamic_lookup") _CONFIG_VARS['CCSHARED'] = " -dynamiclib" _CONFIG_VARS['SO'] = ".dylib" customize_compiler(compiler) finally: _CONFIG_VARS.clear() _CONFIG_VARS.update(tmp) else: customize_compiler(compiler) have_rtld = False use_stubs = False libtype = 'shared' if sys.platform == "darwin": use_stubs = True elif os.name != 'nt': try: import dl use_stubs = have_rtld = hasattr(dl, 'RTLD_NOW') except ImportError: pass if_dl = lambda s: s if have_rtld else '' def get_abi3_suffix(): """Return the file extension for an abi3-compliant Extension()""" for suffix in EXTENSION_SUFFIXES: if '.abi3' in suffix: # Unix return suffix elif suffix == '.pyd': # Windows return suffix class build_ext(_build_ext): def run(self): """Build extensions in build directory, then copy if --inplace""" old_inplace, self.inplace = self.inplace, 0 _build_ext.run(self) self.inplace = old_inplace if old_inplace: self.copy_extensions_to_source() def copy_extensions_to_source(self): build_py = self.get_finalized_command('build_py') for ext in self.extensions: fullname = self.get_ext_fullname(ext.name) filename = self.get_ext_filename(fullname) modpath = fullname.split('.') package = '.'.join(modpath[:-1]) package_dir = build_py.get_package_dir(package) dest_filename = os.path.join(package_dir, os.path.basename(filename)) src_filename = os.path.join(self.build_lib, filename) # Always copy, even if source is older than destination, to ensure # that the right extensions for the current Python/platform are # used. copy_file( src_filename, dest_filename, verbose=self.verbose, dry_run=self.dry_run ) if ext._needs_stub: self.write_stub(package_dir or os.curdir, ext, True) def get_ext_filename(self, fullname): filename = _build_ext.get_ext_filename(self, fullname) if fullname in self.ext_map: ext = self.ext_map[fullname] use_abi3 = ( six.PY3 and getattr(ext, 'py_limited_api') and get_abi3_suffix() ) if use_abi3: so_ext = get_config_var('EXT_SUFFIX') filename = filename[:-len(so_ext)] filename = filename + get_abi3_suffix() if isinstance(ext, Library): fn, ext = os.path.splitext(filename) return self.shlib_compiler.library_filename(fn, libtype) elif use_stubs and ext._links_to_dynamic: d, fn = os.path.split(filename) return os.path.join(d, 'dl-' + fn) return filename def initialize_options(self): _build_ext.initialize_options(self) self.shlib_compiler = None self.shlibs = [] self.ext_map = {} def finalize_options(self): _build_ext.finalize_options(self) self.extensions = self.extensions or [] self.check_extensions_list(self.extensions) self.shlibs = [ext for ext in self.extensions if isinstance(ext, Library)] if self.shlibs: self.setup_shlib_compiler() for ext in self.extensions: ext._full_name = self.get_ext_fullname(ext.name) for ext in self.extensions: fullname = ext._full_name self.ext_map[fullname] = ext # distutils 3.1 will also ask for module names # XXX what to do with conflicts? self.ext_map[fullname.split('.')[-1]] = ext ltd = self.shlibs and self.links_to_dynamic(ext) or False ns = ltd and use_stubs and not isinstance(ext, Library) ext._links_to_dynamic = ltd ext._needs_stub = ns filename = ext._file_name = self.get_ext_filename(fullname) libdir = os.path.dirname(os.path.join(self.build_lib, filename)) if ltd and libdir not in ext.library_dirs: ext.library_dirs.append(libdir) if ltd and use_stubs and os.curdir not in ext.runtime_library_dirs: ext.runtime_library_dirs.append(os.curdir) def setup_shlib_compiler(self): compiler = self.shlib_compiler = new_compiler( compiler=self.compiler, dry_run=self.dry_run, force=self.force ) _customize_compiler_for_shlib(compiler) if self.include_dirs is not None: compiler.set_include_dirs(self.include_dirs) if self.define is not None: # 'define' option is a list of (name,value) tuples for (name, value) in self.define: compiler.define_macro(name, value) if self.undef is not None: for macro in self.undef: compiler.undefine_macro(macro) if self.libraries is not None: compiler.set_libraries(self.libraries) if self.library_dirs is not None: compiler.set_library_dirs(self.library_dirs) if self.rpath is not None: compiler.set_runtime_library_dirs(self.rpath) if self.link_objects is not None: compiler.set_link_objects(self.link_objects) # hack so distutils' build_extension() builds a library instead compiler.link_shared_object = link_shared_object.__get__(compiler) def get_export_symbols(self, ext): if isinstance(ext, Library): return ext.export_symbols return _build_ext.get_export_symbols(self, ext) def build_extension(self, ext): ext._convert_pyx_sources_to_lang() _compiler = self.compiler try: if isinstance(ext, Library): self.compiler = self.shlib_compiler _build_ext.build_extension(self, ext) if ext._needs_stub: cmd = self.get_finalized_command('build_py').build_lib self.write_stub(cmd, ext) finally: self.compiler = _compiler def links_to_dynamic(self, ext): """Return true if 'ext' links to a dynamic lib in the same package""" # XXX this should check to ensure the lib is actually being built # XXX as dynamic, and not just using a locally-found version or a # XXX static-compiled version libnames = dict.fromkeys([lib._full_name for lib in self.shlibs]) pkg = '.'.join(ext._full_name.split('.')[:-1] + ['']) return any(pkg + libname in libnames for libname in ext.libraries) def get_outputs(self): return _build_ext.get_outputs(self) + self.__get_stubs_outputs() def __get_stubs_outputs(self): # assemble the base name for each extension that needs a stub ns_ext_bases = ( os.path.join(self.build_lib, *ext._full_name.split('.')) for ext in self.extensions if ext._needs_stub ) # pair each base with the extension pairs = itertools.product(ns_ext_bases, self.__get_output_extensions()) return list(base + fnext for base, fnext in pairs) def __get_output_extensions(self): yield '.py' yield '.pyc' if self.get_finalized_command('build_py').optimize: yield '.pyo' def write_stub(self, output_dir, ext, compile=False): log.info("writing stub loader for %s to %s", ext._full_name, output_dir) stub_file = (os.path.join(output_dir, *ext._full_name.split('.')) + '.py') if compile and os.path.exists(stub_file): raise DistutilsError(stub_file + " already exists! Please delete.") if not self.dry_run: f = open(stub_file, 'w') f.write( '\n'.join([ "def __bootstrap__():", " global __bootstrap__, __file__, __loader__", " import sys, os, pkg_resources, imp" + if_dl(", dl"), " __file__ = pkg_resources.resource_filename" "(__name__,%r)" % os.path.basename(ext._file_name), " del __bootstrap__", " if '__loader__' in globals():", " del __loader__", if_dl(" old_flags = sys.getdlopenflags()"), " old_dir = os.getcwd()", " try:", " os.chdir(os.path.dirname(__file__))", if_dl(" sys.setdlopenflags(dl.RTLD_NOW)"), " imp.load_dynamic(__name__,__file__)", " finally:", if_dl(" sys.setdlopenflags(old_flags)"), " os.chdir(old_dir)", "__bootstrap__()", "" # terminal \n ]) ) f.close() if compile: from distutils.util import byte_compile byte_compile([stub_file], optimize=0, force=True, dry_run=self.dry_run) optimize = self.get_finalized_command('install_lib').optimize if optimize > 0: byte_compile([stub_file], optimize=optimize, force=True, dry_run=self.dry_run) if os.path.exists(stub_file) and not self.dry_run: os.unlink(stub_file) if use_stubs or os.name == 'nt': # Build shared libraries # def link_shared_object( self, objects, output_libname, output_dir=None, libraries=None, library_dirs=None, runtime_library_dirs=None, export_symbols=None, debug=0, extra_preargs=None, extra_postargs=None, build_temp=None, target_lang=None): self.link( self.SHARED_LIBRARY, objects, output_libname, output_dir, libraries, library_dirs, runtime_library_dirs, export_symbols, debug, extra_preargs, extra_postargs, build_temp, target_lang ) else: # Build static libraries everywhere else libtype = 'static' def link_shared_object( self, objects, output_libname, output_dir=None, libraries=None, library_dirs=None, runtime_library_dirs=None, export_symbols=None, debug=0, extra_preargs=None, extra_postargs=None, build_temp=None, target_lang=None): # XXX we need to either disallow these attrs on Library instances, # or warn/abort here if set, or something... # libraries=None, library_dirs=None, runtime_library_dirs=None, # export_symbols=None, extra_preargs=None, extra_postargs=None, # build_temp=None assert output_dir is None # distutils build_ext doesn't pass this output_dir, filename = os.path.split(output_libname) basename, ext = os.path.splitext(filename) if self.library_filename("x").startswith('lib'): # strip 'lib' prefix; this is kludgy if some platform uses # a different prefix basename = basename[3:] self.create_static_lib( objects, basename, output_dir, debug, target_lang ) site-packages/setuptools/command/build_clib.py000064400000010604151464452710015577 0ustar00import distutils.command.build_clib as orig from distutils.errors import DistutilsSetupError from distutils import log from setuptools.dep_util import newer_pairwise_group class build_clib(orig.build_clib): """ Override the default build_clib behaviour to do the following: 1. Implement a rudimentary timestamp-based dependency system so 'compile()' doesn't run every time. 2. Add more keys to the 'build_info' dictionary: * obj_deps - specify dependencies for each object compiled. this should be a dictionary mapping a key with the source filename to a list of dependencies. Use an empty string for global dependencies. * cflags - specify a list of additional flags to pass to the compiler. """ def build_libraries(self, libraries): for (lib_name, build_info) in libraries: sources = build_info.get('sources') if sources is None or not isinstance(sources, (list, tuple)): raise DistutilsSetupError( "in 'libraries' option (library '%s'), " "'sources' must be present and must be " "a list of source filenames" % lib_name) sources = list(sources) log.info("building '%s' library", lib_name) # Make sure everything is the correct type. # obj_deps should be a dictionary of keys as sources # and a list/tuple of files that are its dependencies. obj_deps = build_info.get('obj_deps', dict()) if not isinstance(obj_deps, dict): raise DistutilsSetupError( "in 'libraries' option (library '%s'), " "'obj_deps' must be a dictionary of " "type 'source: list'" % lib_name) dependencies = [] # Get the global dependencies that are specified by the '' key. # These will go into every source's dependency list. global_deps = obj_deps.get('', list()) if not isinstance(global_deps, (list, tuple)): raise DistutilsSetupError( "in 'libraries' option (library '%s'), " "'obj_deps' must be a dictionary of " "type 'source: list'" % lib_name) # Build the list to be used by newer_pairwise_group # each source will be auto-added to its dependencies. for source in sources: src_deps = [source] src_deps.extend(global_deps) extra_deps = obj_deps.get(source, list()) if not isinstance(extra_deps, (list, tuple)): raise DistutilsSetupError( "in 'libraries' option (library '%s'), " "'obj_deps' must be a dictionary of " "type 'source: list'" % lib_name) src_deps.extend(extra_deps) dependencies.append(src_deps) expected_objects = self.compiler.object_filenames( sources, output_dir=self.build_temp ) if newer_pairwise_group(dependencies, expected_objects) != ([], []): # First, compile the source code to object files in the library # directory. (This should probably change to putting object # files in a temporary build directory.) macros = build_info.get('macros') include_dirs = build_info.get('include_dirs') cflags = build_info.get('cflags') objects = self.compiler.compile( sources, output_dir=self.build_temp, macros=macros, include_dirs=include_dirs, extra_postargs=cflags, debug=self.debug ) # Now "link" the object files together into a static library. # (On Unix at least, this isn't really linking -- it just # builds an archive. Whatever.) self.compiler.create_static_lib( expected_objects, lib_name, output_dir=self.build_clib, debug=self.debug ) site-packages/setuptools/command/bdist_egg.py000064400000043411151464452710015440 0ustar00"""setuptools.command.bdist_egg Build .egg distributions""" from distutils.errors import DistutilsSetupError from distutils.dir_util import remove_tree, mkpath from distutils import log from types import CodeType import sys import os import re import textwrap import marshal from setuptools.extern import six from pkg_resources import get_build_platform, Distribution, ensure_directory from pkg_resources import EntryPoint from setuptools.extension import Library from setuptools import Command try: # Python 2.7 or >=3.2 from sysconfig import get_path, get_python_version def _get_purelib(): return get_path("purelib") except ImportError: from distutils.sysconfig import get_python_lib, get_python_version def _get_purelib(): return get_python_lib(False) def strip_module(filename): if '.' in filename: filename = os.path.splitext(filename)[0] if filename.endswith('module'): filename = filename[:-6] return filename def sorted_walk(dir): """Do os.walk in a reproducible way, independent of indeterministic filesystem readdir order """ for base, dirs, files in os.walk(dir): dirs.sort() files.sort() yield base, dirs, files def write_stub(resource, pyfile): _stub_template = textwrap.dedent(""" def __bootstrap__(): global __bootstrap__, __loader__, __file__ import sys, pkg_resources, imp __file__ = pkg_resources.resource_filename(__name__, %r) __loader__ = None; del __bootstrap__, __loader__ imp.load_dynamic(__name__,__file__) __bootstrap__() """).lstrip() with open(pyfile, 'w') as f: f.write(_stub_template % resource) class bdist_egg(Command): description = "create an \"egg\" distribution" user_options = [ ('bdist-dir=', 'b', "temporary directory for creating the distribution"), ('plat-name=', 'p', "platform name to embed in generated filenames " "(default: %s)" % get_build_platform()), ('exclude-source-files', None, "remove all .py files from the generated egg"), ('keep-temp', 'k', "keep the pseudo-installation tree around after " + "creating the distribution archive"), ('dist-dir=', 'd', "directory to put final built distributions in"), ('skip-build', None, "skip rebuilding everything (for testing/debugging)"), ] boolean_options = [ 'keep-temp', 'skip-build', 'exclude-source-files' ] def initialize_options(self): self.bdist_dir = None self.plat_name = None self.keep_temp = 0 self.dist_dir = None self.skip_build = 0 self.egg_output = None self.exclude_source_files = None def finalize_options(self): ei_cmd = self.ei_cmd = self.get_finalized_command("egg_info") self.egg_info = ei_cmd.egg_info if self.bdist_dir is None: bdist_base = self.get_finalized_command('bdist').bdist_base self.bdist_dir = os.path.join(bdist_base, 'egg') if self.plat_name is None: self.plat_name = get_build_platform() self.set_undefined_options('bdist', ('dist_dir', 'dist_dir')) if self.egg_output is None: # Compute filename of the output egg basename = Distribution( None, None, ei_cmd.egg_name, ei_cmd.egg_version, get_python_version(), self.distribution.has_ext_modules() and self.plat_name ).egg_name() self.egg_output = os.path.join(self.dist_dir, basename + '.egg') def do_install_data(self): # Hack for packages that install data to install's --install-lib self.get_finalized_command('install').install_lib = self.bdist_dir site_packages = os.path.normcase(os.path.realpath(_get_purelib())) old, self.distribution.data_files = self.distribution.data_files, [] for item in old: if isinstance(item, tuple) and len(item) == 2: if os.path.isabs(item[0]): realpath = os.path.realpath(item[0]) normalized = os.path.normcase(realpath) if normalized == site_packages or normalized.startswith( site_packages + os.sep ): item = realpath[len(site_packages) + 1:], item[1] # XXX else: raise ??? self.distribution.data_files.append(item) try: log.info("installing package data to %s", self.bdist_dir) self.call_command('install_data', force=0, root=None) finally: self.distribution.data_files = old def get_outputs(self): return [self.egg_output] def call_command(self, cmdname, **kw): """Invoke reinitialized command `cmdname` with keyword args""" for dirname in INSTALL_DIRECTORY_ATTRS: kw.setdefault(dirname, self.bdist_dir) kw.setdefault('skip_build', self.skip_build) kw.setdefault('dry_run', self.dry_run) cmd = self.reinitialize_command(cmdname, **kw) self.run_command(cmdname) return cmd def run(self): # Generate metadata first self.run_command("egg_info") # We run install_lib before install_data, because some data hacks # pull their data path from the install_lib command. log.info("installing library code to %s", self.bdist_dir) instcmd = self.get_finalized_command('install') old_root = instcmd.root instcmd.root = None if self.distribution.has_c_libraries() and not self.skip_build: self.run_command('build_clib') cmd = self.call_command('install_lib', warn_dir=0) instcmd.root = old_root all_outputs, ext_outputs = self.get_ext_outputs() self.stubs = [] to_compile = [] for (p, ext_name) in enumerate(ext_outputs): filename, ext = os.path.splitext(ext_name) pyfile = os.path.join(self.bdist_dir, strip_module(filename) + '.py') self.stubs.append(pyfile) log.info("creating stub loader for %s", ext_name) if not self.dry_run: write_stub(os.path.basename(ext_name), pyfile) to_compile.append(pyfile) ext_outputs[p] = ext_name.replace(os.sep, '/') if to_compile: cmd.byte_compile(to_compile) if self.distribution.data_files: self.do_install_data() # Make the EGG-INFO directory archive_root = self.bdist_dir egg_info = os.path.join(archive_root, 'EGG-INFO') self.mkpath(egg_info) if self.distribution.scripts: script_dir = os.path.join(egg_info, 'scripts') log.info("installing scripts to %s", script_dir) self.call_command('install_scripts', install_dir=script_dir, no_ep=1) self.copy_metadata_to(egg_info) native_libs = os.path.join(egg_info, "native_libs.txt") if all_outputs: log.info("writing %s", native_libs) if not self.dry_run: ensure_directory(native_libs) libs_file = open(native_libs, 'wt') libs_file.write('\n'.join(all_outputs)) libs_file.write('\n') libs_file.close() elif os.path.isfile(native_libs): log.info("removing %s", native_libs) if not self.dry_run: os.unlink(native_libs) write_safety_flag( os.path.join(archive_root, 'EGG-INFO'), self.zip_safe() ) if os.path.exists(os.path.join(self.egg_info, 'depends.txt')): log.warn( "WARNING: 'depends.txt' will not be used by setuptools 0.6!\n" "Use the install_requires/extras_require setup() args instead." ) if self.exclude_source_files: self.zap_pyfiles() # Make the archive make_zipfile(self.egg_output, archive_root, verbose=self.verbose, dry_run=self.dry_run, mode=self.gen_header()) if not self.keep_temp: remove_tree(self.bdist_dir, dry_run=self.dry_run) # Add to 'Distribution.dist_files' so that the "upload" command works getattr(self.distribution, 'dist_files', []).append( ('bdist_egg', get_python_version(), self.egg_output)) def zap_pyfiles(self): log.info("Removing .py files from temporary directory") for base, dirs, files in walk_egg(self.bdist_dir): for name in files: path = os.path.join(base, name) if name.endswith('.py'): log.debug("Deleting %s", path) os.unlink(path) if base.endswith('__pycache__'): path_old = path pattern = r'(?P.+)\.(?P[^.]+)\.pyc' m = re.match(pattern, name) path_new = os.path.join( base, os.pardir, m.group('name') + '.pyc') log.info( "Renaming file from [%s] to [%s]" % (path_old, path_new)) try: os.remove(path_new) except OSError: pass os.rename(path_old, path_new) def zip_safe(self): safe = getattr(self.distribution, 'zip_safe', None) if safe is not None: return safe log.warn("zip_safe flag not set; analyzing archive contents...") return analyze_egg(self.bdist_dir, self.stubs) def gen_header(self): epm = EntryPoint.parse_map(self.distribution.entry_points or '') ep = epm.get('setuptools.installation', {}).get('eggsecutable') if ep is None: return 'w' # not an eggsecutable, do it the usual way. if not ep.attrs or ep.extras: raise DistutilsSetupError( "eggsecutable entry point (%r) cannot have 'extras' " "or refer to a module" % (ep,) ) pyver = '{}.{}'.format(*sys.version_info) pkg = ep.module_name full = '.'.join(ep.attrs) base = ep.attrs[0] basename = os.path.basename(self.egg_output) header = ( "#!/bin/sh\n" 'if [ `basename $0` = "%(basename)s" ]\n' 'then exec python%(pyver)s -c "' "import sys, os; sys.path.insert(0, os.path.abspath('$0')); " "from %(pkg)s import %(base)s; sys.exit(%(full)s())" '" "$@"\n' 'else\n' ' echo $0 is not the correct name for this egg file.\n' ' echo Please rename it back to %(basename)s and try again.\n' ' exec false\n' 'fi\n' ) % locals() if not self.dry_run: mkpath(os.path.dirname(self.egg_output), dry_run=self.dry_run) f = open(self.egg_output, 'w') f.write(header) f.close() return 'a' def copy_metadata_to(self, target_dir): "Copy metadata (egg info) to the target_dir" # normalize the path (so that a forward-slash in egg_info will # match using startswith below) norm_egg_info = os.path.normpath(self.egg_info) prefix = os.path.join(norm_egg_info, '') for path in self.ei_cmd.filelist.files: if path.startswith(prefix): target = os.path.join(target_dir, path[len(prefix):]) ensure_directory(target) self.copy_file(path, target) def get_ext_outputs(self): """Get a list of relative paths to C extensions in the output distro""" all_outputs = [] ext_outputs = [] paths = {self.bdist_dir: ''} for base, dirs, files in sorted_walk(self.bdist_dir): for filename in files: if os.path.splitext(filename)[1].lower() in NATIVE_EXTENSIONS: all_outputs.append(paths[base] + filename) for filename in dirs: paths[os.path.join(base, filename)] = (paths[base] + filename + '/') if self.distribution.has_ext_modules(): build_cmd = self.get_finalized_command('build_ext') for ext in build_cmd.extensions: if isinstance(ext, Library): continue fullname = build_cmd.get_ext_fullname(ext.name) filename = build_cmd.get_ext_filename(fullname) if not os.path.basename(filename).startswith('dl-'): if os.path.exists(os.path.join(self.bdist_dir, filename)): ext_outputs.append(filename) return all_outputs, ext_outputs NATIVE_EXTENSIONS = dict.fromkeys('.dll .so .dylib .pyd'.split()) def walk_egg(egg_dir): """Walk an unpacked egg's contents, skipping the metadata directory""" walker = sorted_walk(egg_dir) base, dirs, files = next(walker) if 'EGG-INFO' in dirs: dirs.remove('EGG-INFO') yield base, dirs, files for bdf in walker: yield bdf def analyze_egg(egg_dir, stubs): # check for existing flag in EGG-INFO for flag, fn in safety_flags.items(): if os.path.exists(os.path.join(egg_dir, 'EGG-INFO', fn)): return flag if not can_scan(): return False safe = True for base, dirs, files in walk_egg(egg_dir): for name in files: if name.endswith('.py') or name.endswith('.pyw'): continue elif name.endswith('.pyc') or name.endswith('.pyo'): # always scan, even if we already know we're not safe safe = scan_module(egg_dir, base, name, stubs) and safe return safe def write_safety_flag(egg_dir, safe): # Write or remove zip safety flag file(s) for flag, fn in safety_flags.items(): fn = os.path.join(egg_dir, fn) if os.path.exists(fn): if safe is None or bool(safe) != flag: os.unlink(fn) elif safe is not None and bool(safe) == flag: f = open(fn, 'wt') f.write('\n') f.close() safety_flags = { True: 'zip-safe', False: 'not-zip-safe', } def scan_module(egg_dir, base, name, stubs): """Check whether module possibly uses unsafe-for-zipfile stuff""" filename = os.path.join(base, name) if filename[:-1] in stubs: return True # Extension module pkg = base[len(egg_dir) + 1:].replace(os.sep, '.') module = pkg + (pkg and '.' or '') + os.path.splitext(name)[0] if six.PY2: skip = 8 # skip magic & date elif sys.version_info < (3, 7): skip = 12 # skip magic & date & file size else: skip = 16 # skip magic & reserved? & date & file size f = open(filename, 'rb') f.read(skip) code = marshal.load(f) f.close() safe = True symbols = dict.fromkeys(iter_symbols(code)) for bad in ['__file__', '__path__']: if bad in symbols: log.warn("%s: module references %s", module, bad) safe = False if 'inspect' in symbols: for bad in [ 'getsource', 'getabsfile', 'getsourcefile', 'getfile' 'getsourcelines', 'findsource', 'getcomments', 'getframeinfo', 'getinnerframes', 'getouterframes', 'stack', 'trace' ]: if bad in symbols: log.warn("%s: module MAY be using inspect.%s", module, bad) safe = False return safe def iter_symbols(code): """Yield names and strings used by `code` and its nested code objects""" for name in code.co_names: yield name for const in code.co_consts: if isinstance(const, six.string_types): yield const elif isinstance(const, CodeType): for name in iter_symbols(const): yield name def can_scan(): if not sys.platform.startswith('java') and sys.platform != 'cli': # CPython, PyPy, etc. return True log.warn("Unable to analyze compiled code on this platform.") log.warn("Please ask the author to include a 'zip_safe'" " setting (either True or False) in the package's setup.py") # Attribute names of options for commands that might need to be convinced to # install to the egg build directory INSTALL_DIRECTORY_ATTRS = [ 'install_lib', 'install_dir', 'install_data', 'install_base' ] def make_zipfile(zip_filename, base_dir, verbose=0, dry_run=0, compress=True, mode='w'): """Create a zip file from all the files under 'base_dir'. The output zip file will be named 'base_dir' + ".zip". Uses either the "zipfile" Python module (if available) or the InfoZIP "zip" utility (if installed and found on the default search path). If neither tool is available, raises DistutilsExecError. Returns the name of the output zip file. """ import zipfile mkpath(os.path.dirname(zip_filename), dry_run=dry_run) log.info("creating '%s' and adding '%s' to it", zip_filename, base_dir) def visit(z, dirname, names): for name in names: path = os.path.normpath(os.path.join(dirname, name)) if os.path.isfile(path): p = path[len(base_dir) + 1:] if not dry_run: z.write(path, p) log.debug("adding '%s'", p) compression = zipfile.ZIP_DEFLATED if compress else zipfile.ZIP_STORED if not dry_run: z = zipfile.ZipFile(zip_filename, mode, compression=compression) for dirname, dirs, files in sorted_walk(base_dir): visit(z, dirname, files) z.close() else: for dirname, dirs, files in sorted_walk(base_dir): visit(None, dirname, files) return zip_filename site-packages/setuptools/command/build_py.py000064400000022574151464452710015327 0ustar00from glob import glob from distutils.util import convert_path import distutils.command.build_py as orig import os import fnmatch import textwrap import io import distutils.errors import itertools from setuptools.extern import six from setuptools.extern.six.moves import map, filter, filterfalse try: from setuptools.lib2to3_ex import Mixin2to3 except ImportError: class Mixin2to3: def run_2to3(self, files, doctests=True): "do nothing" class build_py(orig.build_py, Mixin2to3): """Enhanced 'build_py' command that includes data files with packages The data files are specified via a 'package_data' argument to 'setup()'. See 'setuptools.dist.Distribution' for more details. Also, this version of the 'build_py' command allows you to specify both 'py_modules' and 'packages' in the same setup operation. """ def finalize_options(self): orig.build_py.finalize_options(self) self.package_data = self.distribution.package_data self.exclude_package_data = (self.distribution.exclude_package_data or {}) if 'data_files' in self.__dict__: del self.__dict__['data_files'] self.__updated_files = [] self.__doctests_2to3 = [] def run(self): """Build modules, packages, and copy data files to build directory""" if not self.py_modules and not self.packages: return if self.py_modules: self.build_modules() if self.packages: self.build_packages() self.build_package_data() self.run_2to3(self.__updated_files, False) self.run_2to3(self.__updated_files, True) self.run_2to3(self.__doctests_2to3, True) # Only compile actual .py files, using our base class' idea of what our # output files are. self.byte_compile(orig.build_py.get_outputs(self, include_bytecode=0)) def __getattr__(self, attr): "lazily compute data files" if attr == 'data_files': self.data_files = self._get_data_files() return self.data_files return orig.build_py.__getattr__(self, attr) def build_module(self, module, module_file, package): if six.PY2 and isinstance(package, six.string_types): # avoid errors on Python 2 when unicode is passed (#190) package = package.split('.') outfile, copied = orig.build_py.build_module(self, module, module_file, package) if copied: self.__updated_files.append(outfile) return outfile, copied def _get_data_files(self): """Generate list of '(package,src_dir,build_dir,filenames)' tuples""" self.analyze_manifest() return list(map(self._get_pkg_data_files, self.packages or ())) def _get_pkg_data_files(self, package): # Locate package source directory src_dir = self.get_package_dir(package) # Compute package build directory build_dir = os.path.join(*([self.build_lib] + package.split('.'))) # Strip directory from globbed filenames filenames = [ os.path.relpath(file, src_dir) for file in self.find_data_files(package, src_dir) ] return package, src_dir, build_dir, filenames def find_data_files(self, package, src_dir): """Return filenames for package's data files in 'src_dir'""" patterns = self._get_platform_patterns( self.package_data, package, src_dir, ) globs_expanded = map(glob, patterns) # flatten the expanded globs into an iterable of matches globs_matches = itertools.chain.from_iterable(globs_expanded) glob_files = filter(os.path.isfile, globs_matches) files = itertools.chain( self.manifest_files.get(package, []), glob_files, ) return self.exclude_data_files(package, src_dir, files) def build_package_data(self): """Copy data files into build directory""" for package, src_dir, build_dir, filenames in self.data_files: for filename in filenames: target = os.path.join(build_dir, filename) self.mkpath(os.path.dirname(target)) srcfile = os.path.join(src_dir, filename) outf, copied = self.copy_file(srcfile, target) srcfile = os.path.abspath(srcfile) if (copied and srcfile in self.distribution.convert_2to3_doctests): self.__doctests_2to3.append(outf) def analyze_manifest(self): self.manifest_files = mf = {} if not self.distribution.include_package_data: return src_dirs = {} for package in self.packages or (): # Locate package source directory src_dirs[assert_relative(self.get_package_dir(package))] = package self.run_command('egg_info') ei_cmd = self.get_finalized_command('egg_info') for path in ei_cmd.filelist.files: d, f = os.path.split(assert_relative(path)) prev = None oldf = f while d and d != prev and d not in src_dirs: prev = d d, df = os.path.split(d) f = os.path.join(df, f) if d in src_dirs: if path.endswith('.py') and f == oldf: continue # it's a module, not data mf.setdefault(src_dirs[d], []).append(path) def get_data_files(self): pass # Lazily compute data files in _get_data_files() function. def check_package(self, package, package_dir): """Check namespace packages' __init__ for declare_namespace""" try: return self.packages_checked[package] except KeyError: pass init_py = orig.build_py.check_package(self, package, package_dir) self.packages_checked[package] = init_py if not init_py or not self.distribution.namespace_packages: return init_py for pkg in self.distribution.namespace_packages: if pkg == package or pkg.startswith(package + '.'): break else: return init_py with io.open(init_py, 'rb') as f: contents = f.read() if b'declare_namespace' not in contents: raise distutils.errors.DistutilsError( "Namespace package problem: %s is a namespace package, but " "its\n__init__.py does not call declare_namespace()! Please " 'fix it.\n(See the setuptools manual under ' '"Namespace Packages" for details.)\n"' % (package,) ) return init_py def initialize_options(self): self.packages_checked = {} orig.build_py.initialize_options(self) def get_package_dir(self, package): res = orig.build_py.get_package_dir(self, package) if self.distribution.src_root is not None: return os.path.join(self.distribution.src_root, res) return res def exclude_data_files(self, package, src_dir, files): """Filter filenames for package's data files in 'src_dir'""" files = list(files) patterns = self._get_platform_patterns( self.exclude_package_data, package, src_dir, ) match_groups = ( fnmatch.filter(files, pattern) for pattern in patterns ) # flatten the groups of matches into an iterable of matches matches = itertools.chain.from_iterable(match_groups) bad = set(matches) keepers = ( fn for fn in files if fn not in bad ) # ditch dupes return list(_unique_everseen(keepers)) @staticmethod def _get_platform_patterns(spec, package, src_dir): """ yield platform-specific path patterns (suitable for glob or fn_match) from a glob-based spec (such as self.package_data or self.exclude_package_data) matching package in src_dir. """ raw_patterns = itertools.chain( spec.get('', []), spec.get(package, []), ) return ( # Each pattern has to be converted to a platform-specific path os.path.join(src_dir, convert_path(pattern)) for pattern in raw_patterns ) # from Python docs def _unique_everseen(iterable, key=None): "List unique elements, preserving order. Remember all elements ever seen." # unique_everseen('AAAABBBCCDAABBB') --> A B C D # unique_everseen('ABBCcAD', str.lower) --> A B C D seen = set() seen_add = seen.add if key is None: for element in filterfalse(seen.__contains__, iterable): seen_add(element) yield element else: for element in iterable: k = key(element) if k not in seen: seen_add(k) yield element def assert_relative(path): if not os.path.isabs(path): return path from distutils.errors import DistutilsSetupError msg = textwrap.dedent(""" Error: setup script specifies an absolute path: %s setup() arguments must *always* be /-separated paths relative to the setup.py directory, *never* absolute paths. """).lstrip() % path raise DistutilsSetupError(msg) site-packages/setuptools/glob.py000064400000011734151464452710013021 0ustar00""" Filename globbing utility. Mostly a copy of `glob` from Python 3.5. Changes include: * `yield from` and PEP3102 `*` removed. * Hidden files are not ignored. """ import os import re import fnmatch __all__ = ["glob", "iglob", "escape"] def glob(pathname, recursive=False): """Return a list of paths matching a pathname pattern. The pattern may contain simple shell-style wildcards a la fnmatch. However, unlike fnmatch, filenames starting with a dot are special cases that are not matched by '*' and '?' patterns. If recursive is true, the pattern '**' will match any files and zero or more directories and subdirectories. """ return list(iglob(pathname, recursive=recursive)) def iglob(pathname, recursive=False): """Return an iterator which yields the paths matching a pathname pattern. The pattern may contain simple shell-style wildcards a la fnmatch. However, unlike fnmatch, filenames starting with a dot are special cases that are not matched by '*' and '?' patterns. If recursive is true, the pattern '**' will match any files and zero or more directories and subdirectories. """ it = _iglob(pathname, recursive) if recursive and _isrecursive(pathname): s = next(it) # skip empty string assert not s return it def _iglob(pathname, recursive): dirname, basename = os.path.split(pathname) if not has_magic(pathname): if basename: if os.path.lexists(pathname): yield pathname else: # Patterns ending with a slash should match only directories if os.path.isdir(dirname): yield pathname return if not dirname: if recursive and _isrecursive(basename): for x in glob2(dirname, basename): yield x else: for x in glob1(dirname, basename): yield x return # `os.path.split()` returns the argument itself as a dirname if it is a # drive or UNC path. Prevent an infinite recursion if a drive or UNC path # contains magic characters (i.e. r'\\?\C:'). if dirname != pathname and has_magic(dirname): dirs = _iglob(dirname, recursive) else: dirs = [dirname] if has_magic(basename): if recursive and _isrecursive(basename): glob_in_dir = glob2 else: glob_in_dir = glob1 else: glob_in_dir = glob0 for dirname in dirs: for name in glob_in_dir(dirname, basename): yield os.path.join(dirname, name) # These 2 helper functions non-recursively glob inside a literal directory. # They return a list of basenames. `glob1` accepts a pattern while `glob0` # takes a literal basename (so it only has to check for its existence). def glob1(dirname, pattern): if not dirname: if isinstance(pattern, bytes): dirname = os.curdir.encode('ASCII') else: dirname = os.curdir try: names = os.listdir(dirname) except OSError: return [] return fnmatch.filter(names, pattern) def glob0(dirname, basename): if not basename: # `os.path.split()` returns an empty basename for paths ending with a # directory separator. 'q*x/' should match only directories. if os.path.isdir(dirname): return [basename] else: if os.path.lexists(os.path.join(dirname, basename)): return [basename] return [] # This helper function recursively yields relative pathnames inside a literal # directory. def glob2(dirname, pattern): assert _isrecursive(pattern) yield pattern[:0] for x in _rlistdir(dirname): yield x # Recursively yields relative pathnames inside a literal directory. def _rlistdir(dirname): if not dirname: if isinstance(dirname, bytes): dirname = os.curdir.encode('ASCII') else: dirname = os.curdir try: names = os.listdir(dirname) except os.error: return for x in names: yield x path = os.path.join(dirname, x) if dirname else x for y in _rlistdir(path): yield os.path.join(x, y) magic_check = re.compile('([*?[])') magic_check_bytes = re.compile(b'([*?[])') def has_magic(s): if isinstance(s, bytes): match = magic_check_bytes.search(s) else: match = magic_check.search(s) return match is not None def _isrecursive(pattern): if isinstance(pattern, bytes): return pattern == b'**' else: return pattern == '**' def escape(pathname): """Escape all special characters. """ # Escaping is done by wrapping any of "*?[" between square brackets. # Metacharacters do not work in the drive part and shouldn't be escaped. drive, pathname = os.path.splitdrive(pathname) if isinstance(pathname, bytes): pathname = magic_check_bytes.sub(br'[\1]', pathname) else: pathname = magic_check.sub(r'[\1]', pathname) return drive + pathname site-packages/setuptools/unicode_utils.py000064400000001744151464452710014744 0ustar00import unicodedata import sys from setuptools.extern import six # HFS Plus uses decomposed UTF-8 def decompose(path): if isinstance(path, six.text_type): return unicodedata.normalize('NFD', path) try: path = path.decode('utf-8') path = unicodedata.normalize('NFD', path) path = path.encode('utf-8') except UnicodeError: pass # Not UTF-8 return path def filesys_decode(path): """ Ensure that the given path is decoded, NONE when no expected encoding works """ if isinstance(path, six.text_type): return path fs_enc = sys.getfilesystemencoding() or 'utf-8' candidates = fs_enc, 'utf-8' for enc in candidates: try: return path.decode(enc) except UnicodeDecodeError: continue def try_encode(string, enc): "turn unicode encoding into a functional routine" try: return string.encode(enc) except UnicodeEncodeError: return None site-packages/setuptools/dep_util.py000064400000001647151464452710013705 0ustar00from distutils.dep_util import newer_group # yes, this is was almost entirely copy-pasted from # 'newer_pairwise()', this is just another convenience # function. def newer_pairwise_group(sources_groups, targets): """Walk both arguments in parallel, testing if each source group is newer than its corresponding target. Returns a pair of lists (sources_groups, targets) where sources is newer than target, according to the semantics of 'newer_group()'. """ if len(sources_groups) != len(targets): raise ValueError("'sources_group' and 'targets' must be the same length") # build a pair of lists (sources_groups, targets) where source is newer n_sources = [] n_targets = [] for i in range(len(sources_groups)): if newer_group(sources_groups[i], targets[i]): n_sources.append(sources_groups[i]) n_targets.append(targets[i]) return n_sources, n_targets site-packages/setuptools/py31compat.py000064400000001506151464452710014072 0ustar00__all__ = [] __metaclass__ = type try: # Python >=3.2 from tempfile import TemporaryDirectory except ImportError: import shutil import tempfile class TemporaryDirectory: """ Very simple temporary directory context manager. Will try to delete afterward, but will also ignore OS and similar errors on deletion. """ def __init__(self, **kwargs): self.name = None # Handle mkdtemp raising an exception self.name = tempfile.mkdtemp(**kwargs) def __enter__(self): return self.name def __exit__(self, exctype, excvalue, exctrace): try: shutil.rmtree(self.name, True) except OSError: # removal errors are not the only possible pass self.name = None site-packages/setuptools/msvc.py000064400000133237151464452710013051 0ustar00""" Improved support for Microsoft Visual C++ compilers. Known supported compilers: -------------------------- Microsoft Visual C++ 9.0: Microsoft Visual C++ Compiler for Python 2.7 (x86, amd64) Microsoft Windows SDK 6.1 (x86, x64, ia64) Microsoft Windows SDK 7.0 (x86, x64, ia64) Microsoft Visual C++ 10.0: Microsoft Windows SDK 7.1 (x86, x64, ia64) Microsoft Visual C++ 14.X: Microsoft Visual C++ Build Tools 2015 (x86, x64, arm) Microsoft Visual Studio Build Tools 2017 (x86, x64, arm, arm64) Microsoft Visual Studio Build Tools 2019 (x86, x64, arm, arm64) This may also support compilers shipped with compatible Visual Studio versions. """ import json from io import open from os import listdir, pathsep from os.path import join, isfile, isdir, dirname import sys import platform import itertools import distutils.errors from setuptools.extern.packaging.version import LegacyVersion from setuptools.extern.six.moves import filterfalse from .monkey import get_unpatched if platform.system() == 'Windows': from setuptools.extern.six.moves import winreg from os import environ else: # Mock winreg and environ so the module can be imported on this platform. class winreg: HKEY_USERS = None HKEY_CURRENT_USER = None HKEY_LOCAL_MACHINE = None HKEY_CLASSES_ROOT = None environ = dict() _msvc9_suppress_errors = ( # msvc9compiler isn't available on some platforms ImportError, # msvc9compiler raises DistutilsPlatformError in some # environments. See #1118. distutils.errors.DistutilsPlatformError, ) try: from distutils.msvc9compiler import Reg except _msvc9_suppress_errors: pass def msvc9_find_vcvarsall(version): """ Patched "distutils.msvc9compiler.find_vcvarsall" to use the standalone compiler build for Python (VCForPython / Microsoft Visual C++ Compiler for Python 2.7). Fall back to original behavior when the standalone compiler is not available. Redirect the path of "vcvarsall.bat". Parameters ---------- version: float Required Microsoft Visual C++ version. Return ------ str vcvarsall.bat path """ vc_base = r'Software\%sMicrosoft\DevDiv\VCForPython\%0.1f' key = vc_base % ('', version) try: # Per-user installs register the compiler path here productdir = Reg.get_value(key, "installdir") except KeyError: try: # All-user installs on a 64-bit system register here key = vc_base % ('Wow6432Node\\', version) productdir = Reg.get_value(key, "installdir") except KeyError: productdir = None if productdir: vcvarsall = join(productdir, "vcvarsall.bat") if isfile(vcvarsall): return vcvarsall return get_unpatched(msvc9_find_vcvarsall)(version) def msvc9_query_vcvarsall(ver, arch='x86', *args, **kwargs): """ Patched "distutils.msvc9compiler.query_vcvarsall" for support extra Microsoft Visual C++ 9.0 and 10.0 compilers. Set environment without use of "vcvarsall.bat". Parameters ---------- ver: float Required Microsoft Visual C++ version. arch: str Target architecture. Return ------ dict environment """ # Try to get environment from vcvarsall.bat (Classical way) try: orig = get_unpatched(msvc9_query_vcvarsall) return orig(ver, arch, *args, **kwargs) except distutils.errors.DistutilsPlatformError: # Pass error if Vcvarsall.bat is missing pass except ValueError: # Pass error if environment not set after executing vcvarsall.bat pass # If error, try to set environment directly try: return EnvironmentInfo(arch, ver).return_env() except distutils.errors.DistutilsPlatformError as exc: _augment_exception(exc, ver, arch) raise def msvc14_get_vc_env(plat_spec): """ Patched "distutils._msvccompiler._get_vc_env" for support extra Microsoft Visual C++ 14.X compilers. Set environment without use of "vcvarsall.bat". Parameters ---------- plat_spec: str Target architecture. Return ------ dict environment """ # Try to get environment from vcvarsall.bat (Classical way) try: return get_unpatched(msvc14_get_vc_env)(plat_spec) except distutils.errors.DistutilsPlatformError: # Pass error Vcvarsall.bat is missing pass # If error, try to set environment directly try: return EnvironmentInfo(plat_spec, vc_min_ver=14.0).return_env() except distutils.errors.DistutilsPlatformError as exc: _augment_exception(exc, 14.0) raise def msvc14_gen_lib_options(*args, **kwargs): """ Patched "distutils._msvccompiler.gen_lib_options" for fix compatibility between "numpy.distutils" and "distutils._msvccompiler" (for Numpy < 1.11.2) """ if "numpy.distutils" in sys.modules: import numpy as np if LegacyVersion(np.__version__) < LegacyVersion('1.11.2'): return np.distutils.ccompiler.gen_lib_options(*args, **kwargs) return get_unpatched(msvc14_gen_lib_options)(*args, **kwargs) def _augment_exception(exc, version, arch=''): """ Add details to the exception message to help guide the user as to what action will resolve it. """ # Error if MSVC++ directory not found or environment not set message = exc.args[0] if "vcvarsall" in message.lower() or "visual c" in message.lower(): # Special error message if MSVC++ not installed tmpl = 'Microsoft Visual C++ {version:0.1f} is required.' message = tmpl.format(**locals()) msdownload = 'www.microsoft.com/download/details.aspx?id=%d' if version == 9.0: if arch.lower().find('ia64') > -1: # For VC++ 9.0, if IA64 support is needed, redirect user # to Windows SDK 7.0. # Note: No download link available from Microsoft. message += ' Get it with "Microsoft Windows SDK 7.0"' else: # For VC++ 9.0 redirect user to Vc++ for Python 2.7 : # This redirection link is maintained by Microsoft. # Contact vspython@microsoft.com if it needs updating. message += ' Get it from http://aka.ms/vcpython27' elif version == 10.0: # For VC++ 10.0 Redirect user to Windows SDK 7.1 message += ' Get it with "Microsoft Windows SDK 7.1": ' message += msdownload % 8279 elif version >= 14.0: # For VC++ 14.X Redirect user to latest Visual C++ Build Tools message += (' Get it with "Build Tools for Visual Studio": ' r'https://visualstudio.microsoft.com/downloads/') exc.args = (message, ) class PlatformInfo: """ Current and Target Architectures information. Parameters ---------- arch: str Target architecture. """ current_cpu = environ.get('processor_architecture', '').lower() def __init__(self, arch): self.arch = arch.lower().replace('x64', 'amd64') @property def target_cpu(self): """ Return Target CPU architecture. Return ------ str Target CPU """ return self.arch[self.arch.find('_') + 1:] def target_is_x86(self): """ Return True if target CPU is x86 32 bits.. Return ------ bool CPU is x86 32 bits """ return self.target_cpu == 'x86' def current_is_x86(self): """ Return True if current CPU is x86 32 bits.. Return ------ bool CPU is x86 32 bits """ return self.current_cpu == 'x86' def current_dir(self, hidex86=False, x64=False): """ Current platform specific subfolder. Parameters ---------- hidex86: bool return '' and not '\x86' if architecture is x86. x64: bool return '\x64' and not '\amd64' if architecture is amd64. Return ------ str subfolder: '\target', or '' (see hidex86 parameter) """ return ( '' if (self.current_cpu == 'x86' and hidex86) else r'\x64' if (self.current_cpu == 'amd64' and x64) else r'\%s' % self.current_cpu ) def target_dir(self, hidex86=False, x64=False): r""" Target platform specific subfolder. Parameters ---------- hidex86: bool return '' and not '\x86' if architecture is x86. x64: bool return '\x64' and not '\amd64' if architecture is amd64. Return ------ str subfolder: '\current', or '' (see hidex86 parameter) """ return ( '' if (self.target_cpu == 'x86' and hidex86) else r'\x64' if (self.target_cpu == 'amd64' and x64) else r'\%s' % self.target_cpu ) def cross_dir(self, forcex86=False): r""" Cross platform specific subfolder. Parameters ---------- forcex86: bool Use 'x86' as current architecture even if current architecture is not x86. Return ------ str subfolder: '' if target architecture is current architecture, '\current_target' if not. """ current = 'x86' if forcex86 else self.current_cpu return ( '' if self.target_cpu == current else self.target_dir().replace('\\', '\\%s_' % current) ) class RegistryInfo: """ Microsoft Visual Studio related registry information. Parameters ---------- platform_info: PlatformInfo "PlatformInfo" instance. """ HKEYS = (winreg.HKEY_USERS, winreg.HKEY_CURRENT_USER, winreg.HKEY_LOCAL_MACHINE, winreg.HKEY_CLASSES_ROOT) def __init__(self, platform_info): self.pi = platform_info @property def visualstudio(self): """ Microsoft Visual Studio root registry key. Return ------ str Registry key """ return 'VisualStudio' @property def sxs(self): """ Microsoft Visual Studio SxS registry key. Return ------ str Registry key """ return join(self.visualstudio, 'SxS') @property def vc(self): """ Microsoft Visual C++ VC7 registry key. Return ------ str Registry key """ return join(self.sxs, 'VC7') @property def vs(self): """ Microsoft Visual Studio VS7 registry key. Return ------ str Registry key """ return join(self.sxs, 'VS7') @property def vc_for_python(self): """ Microsoft Visual C++ for Python registry key. Return ------ str Registry key """ return r'DevDiv\VCForPython' @property def microsoft_sdk(self): """ Microsoft SDK registry key. Return ------ str Registry key """ return 'Microsoft SDKs' @property def windows_sdk(self): """ Microsoft Windows/Platform SDK registry key. Return ------ str Registry key """ return join(self.microsoft_sdk, 'Windows') @property def netfx_sdk(self): """ Microsoft .NET Framework SDK registry key. Return ------ str Registry key """ return join(self.microsoft_sdk, 'NETFXSDK') @property def windows_kits_roots(self): """ Microsoft Windows Kits Roots registry key. Return ------ str Registry key """ return r'Windows Kits\Installed Roots' def microsoft(self, key, x86=False): """ Return key in Microsoft software registry. Parameters ---------- key: str Registry key path where look. x86: str Force x86 software registry. Return ------ str Registry key """ node64 = '' if self.pi.current_is_x86() or x86 else 'Wow6432Node' return join('Software', node64, 'Microsoft', key) def lookup(self, key, name): """ Look for values in registry in Microsoft software registry. Parameters ---------- key: str Registry key path where look. name: str Value name to find. Return ------ str value """ key_read = winreg.KEY_READ openkey = winreg.OpenKey ms = self.microsoft for hkey in self.HKEYS: try: bkey = openkey(hkey, ms(key), 0, key_read) except (OSError, IOError): if not self.pi.current_is_x86(): try: bkey = openkey(hkey, ms(key, True), 0, key_read) except (OSError, IOError): continue else: continue try: return winreg.QueryValueEx(bkey, name)[0] except (OSError, IOError): pass class SystemInfo: """ Microsoft Windows and Visual Studio related system information. Parameters ---------- registry_info: RegistryInfo "RegistryInfo" instance. vc_ver: float Required Microsoft Visual C++ version. """ # Variables and properties in this class use originals CamelCase variables # names from Microsoft source files for more easy comparison. WinDir = environ.get('WinDir', '') ProgramFiles = environ.get('ProgramFiles', '') ProgramFilesx86 = environ.get('ProgramFiles(x86)', ProgramFiles) def __init__(self, registry_info, vc_ver=None): self.ri = registry_info self.pi = self.ri.pi self.known_vs_paths = self.find_programdata_vs_vers() # Except for VS15+, VC version is aligned with VS version self.vs_ver = self.vc_ver = ( vc_ver or self._find_latest_available_vs_ver()) def _find_latest_available_vs_ver(self): """ Find the latest VC version Return ------ float version """ reg_vc_vers = self.find_reg_vs_vers() if not (reg_vc_vers or self.known_vs_paths): raise distutils.errors.DistutilsPlatformError( 'No Microsoft Visual C++ version found') vc_vers = set(reg_vc_vers) vc_vers.update(self.known_vs_paths) return sorted(vc_vers)[-1] def find_reg_vs_vers(self): """ Find Microsoft Visual Studio versions available in registry. Return ------ list of float Versions """ ms = self.ri.microsoft vckeys = (self.ri.vc, self.ri.vc_for_python, self.ri.vs) vs_vers = [] for hkey in self.ri.HKEYS: for key in vckeys: try: bkey = winreg.OpenKey(hkey, ms(key), 0, winreg.KEY_READ) except (OSError, IOError): continue subkeys, values, _ = winreg.QueryInfoKey(bkey) for i in range(values): try: ver = float(winreg.EnumValue(bkey, i)[0]) if ver not in vs_vers: vs_vers.append(ver) except ValueError: pass for i in range(subkeys): try: ver = float(winreg.EnumKey(bkey, i)) if ver not in vs_vers: vs_vers.append(ver) except ValueError: pass return sorted(vs_vers) def find_programdata_vs_vers(self): r""" Find Visual studio 2017+ versions from information in "C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances". Return ------ dict float version as key, path as value. """ vs_versions = {} instances_dir = \ r'C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances' try: hashed_names = listdir(instances_dir) except (OSError, IOError): # Directory not exists with all Visual Studio versions return vs_versions for name in hashed_names: try: # Get VS installation path from "state.json" file state_path = join(instances_dir, name, 'state.json') with open(state_path, 'rt', encoding='utf-8') as state_file: state = json.load(state_file) vs_path = state['installationPath'] # Raises OSError if this VS installation does not contain VC listdir(join(vs_path, r'VC\Tools\MSVC')) # Store version and path vs_versions[self._as_float_version( state['installationVersion'])] = vs_path except (OSError, IOError, KeyError): # Skip if "state.json" file is missing or bad format continue return vs_versions @staticmethod def _as_float_version(version): """ Return a string version as a simplified float version (major.minor) Parameters ---------- version: str Version. Return ------ float version """ return float('.'.join(version.split('.')[:2])) @property def VSInstallDir(self): """ Microsoft Visual Studio directory. Return ------ str path """ # Default path default = join(self.ProgramFilesx86, 'Microsoft Visual Studio %0.1f' % self.vs_ver) # Try to get path from registry, if fail use default path return self.ri.lookup(self.ri.vs, '%0.1f' % self.vs_ver) or default @property def VCInstallDir(self): """ Microsoft Visual C++ directory. Return ------ str path """ path = self._guess_vc() or self._guess_vc_legacy() if not isdir(path): msg = 'Microsoft Visual C++ directory not found' raise distutils.errors.DistutilsPlatformError(msg) return path def _guess_vc(self): """ Locate Visual C++ for VS2017+. Return ------ str path """ if self.vs_ver <= 14.0: return '' try: # First search in known VS paths vs_dir = self.known_vs_paths[self.vs_ver] except KeyError: # Else, search with path from registry vs_dir = self.VSInstallDir guess_vc = join(vs_dir, r'VC\Tools\MSVC') # Subdir with VC exact version as name try: # Update the VC version with real one instead of VS version vc_ver = listdir(guess_vc)[-1] self.vc_ver = self._as_float_version(vc_ver) return join(guess_vc, vc_ver) except (OSError, IOError, IndexError): return '' def _guess_vc_legacy(self): """ Locate Visual C++ for versions prior to 2017. Return ------ str path """ default = join(self.ProgramFilesx86, r'Microsoft Visual Studio %0.1f\VC' % self.vs_ver) # Try to get "VC++ for Python" path from registry as default path reg_path = join(self.ri.vc_for_python, '%0.1f' % self.vs_ver) python_vc = self.ri.lookup(reg_path, 'installdir') default_vc = join(python_vc, 'VC') if python_vc else default # Try to get path from registry, if fail use default path return self.ri.lookup(self.ri.vc, '%0.1f' % self.vs_ver) or default_vc @property def WindowsSdkVersion(self): """ Microsoft Windows SDK versions for specified MSVC++ version. Return ------ tuple of str versions """ if self.vs_ver <= 9.0: return '7.0', '6.1', '6.0a' elif self.vs_ver == 10.0: return '7.1', '7.0a' elif self.vs_ver == 11.0: return '8.0', '8.0a' elif self.vs_ver == 12.0: return '8.1', '8.1a' elif self.vs_ver >= 14.0: return '10.0', '8.1' @property def WindowsSdkLastVersion(self): """ Microsoft Windows SDK last version. Return ------ str version """ return self._use_last_dir_name(join(self.WindowsSdkDir, 'lib')) @property def WindowsSdkDir(self): """ Microsoft Windows SDK directory. Return ------ str path """ sdkdir = '' for ver in self.WindowsSdkVersion: # Try to get it from registry loc = join(self.ri.windows_sdk, 'v%s' % ver) sdkdir = self.ri.lookup(loc, 'installationfolder') if sdkdir: break if not sdkdir or not isdir(sdkdir): # Try to get "VC++ for Python" version from registry path = join(self.ri.vc_for_python, '%0.1f' % self.vc_ver) install_base = self.ri.lookup(path, 'installdir') if install_base: sdkdir = join(install_base, 'WinSDK') if not sdkdir or not isdir(sdkdir): # If fail, use default new path for ver in self.WindowsSdkVersion: intver = ver[:ver.rfind('.')] path = r'Microsoft SDKs\Windows Kits\%s' % intver d = join(self.ProgramFiles, path) if isdir(d): sdkdir = d if not sdkdir or not isdir(sdkdir): # If fail, use default old path for ver in self.WindowsSdkVersion: path = r'Microsoft SDKs\Windows\v%s' % ver d = join(self.ProgramFiles, path) if isdir(d): sdkdir = d if not sdkdir: # If fail, use Platform SDK sdkdir = join(self.VCInstallDir, 'PlatformSDK') return sdkdir @property def WindowsSDKExecutablePath(self): """ Microsoft Windows SDK executable directory. Return ------ str path """ # Find WinSDK NetFx Tools registry dir name if self.vs_ver <= 11.0: netfxver = 35 arch = '' else: netfxver = 40 hidex86 = True if self.vs_ver <= 12.0 else False arch = self.pi.current_dir(x64=True, hidex86=hidex86) fx = 'WinSDK-NetFx%dTools%s' % (netfxver, arch.replace('\\', '-')) # list all possibles registry paths regpaths = [] if self.vs_ver >= 14.0: for ver in self.NetFxSdkVersion: regpaths += [join(self.ri.netfx_sdk, ver, fx)] for ver in self.WindowsSdkVersion: regpaths += [join(self.ri.windows_sdk, 'v%sA' % ver, fx)] # Return installation folder from the more recent path for path in regpaths: execpath = self.ri.lookup(path, 'installationfolder') if execpath: return execpath @property def FSharpInstallDir(self): """ Microsoft Visual F# directory. Return ------ str path """ path = join(self.ri.visualstudio, r'%0.1f\Setup\F#' % self.vs_ver) return self.ri.lookup(path, 'productdir') or '' @property def UniversalCRTSdkDir(self): """ Microsoft Universal CRT SDK directory. Return ------ str path """ # Set Kit Roots versions for specified MSVC++ version vers = ('10', '81') if self.vs_ver >= 14.0 else () # Find path of the more recent Kit for ver in vers: sdkdir = self.ri.lookup(self.ri.windows_kits_roots, 'kitsroot%s' % ver) if sdkdir: return sdkdir or '' @property def UniversalCRTSdkLastVersion(self): """ Microsoft Universal C Runtime SDK last version. Return ------ str version """ return self._use_last_dir_name(join(self.UniversalCRTSdkDir, 'lib')) @property def NetFxSdkVersion(self): """ Microsoft .NET Framework SDK versions. Return ------ tuple of str versions """ # Set FxSdk versions for specified VS version return (('4.7.2', '4.7.1', '4.7', '4.6.2', '4.6.1', '4.6', '4.5.2', '4.5.1', '4.5') if self.vs_ver >= 14.0 else ()) @property def NetFxSdkDir(self): """ Microsoft .NET Framework SDK directory. Return ------ str path """ sdkdir = '' for ver in self.NetFxSdkVersion: loc = join(self.ri.netfx_sdk, ver) sdkdir = self.ri.lookup(loc, 'kitsinstallationfolder') if sdkdir: break return sdkdir @property def FrameworkDir32(self): """ Microsoft .NET Framework 32bit directory. Return ------ str path """ # Default path guess_fw = join(self.WinDir, r'Microsoft.NET\Framework') # Try to get path from registry, if fail use default path return self.ri.lookup(self.ri.vc, 'frameworkdir32') or guess_fw @property def FrameworkDir64(self): """ Microsoft .NET Framework 64bit directory. Return ------ str path """ # Default path guess_fw = join(self.WinDir, r'Microsoft.NET\Framework64') # Try to get path from registry, if fail use default path return self.ri.lookup(self.ri.vc, 'frameworkdir64') or guess_fw @property def FrameworkVersion32(self): """ Microsoft .NET Framework 32bit versions. Return ------ tuple of str versions """ return self._find_dot_net_versions(32) @property def FrameworkVersion64(self): """ Microsoft .NET Framework 64bit versions. Return ------ tuple of str versions """ return self._find_dot_net_versions(64) def _find_dot_net_versions(self, bits): """ Find Microsoft .NET Framework versions. Parameters ---------- bits: int Platform number of bits: 32 or 64. Return ------ tuple of str versions """ # Find actual .NET version in registry reg_ver = self.ri.lookup(self.ri.vc, 'frameworkver%d' % bits) dot_net_dir = getattr(self, 'FrameworkDir%d' % bits) ver = reg_ver or self._use_last_dir_name(dot_net_dir, 'v') or '' # Set .NET versions for specified MSVC++ version if self.vs_ver >= 12.0: return ver, 'v4.0' elif self.vs_ver >= 10.0: return 'v4.0.30319' if ver.lower()[:2] != 'v4' else ver, 'v3.5' elif self.vs_ver == 9.0: return 'v3.5', 'v2.0.50727' elif self.vs_ver == 8.0: return 'v3.0', 'v2.0.50727' @staticmethod def _use_last_dir_name(path, prefix=''): """ Return name of the last dir in path or '' if no dir found. Parameters ---------- path: str Use dirs in this path prefix: str Use only dirs starting by this prefix Return ------ str name """ matching_dirs = ( dir_name for dir_name in reversed(listdir(path)) if isdir(join(path, dir_name)) and dir_name.startswith(prefix) ) return next(matching_dirs, None) or '' class EnvironmentInfo: """ Return environment variables for specified Microsoft Visual C++ version and platform : Lib, Include, Path and libpath. This function is compatible with Microsoft Visual C++ 9.0 to 14.X. Script created by analysing Microsoft environment configuration files like "vcvars[...].bat", "SetEnv.Cmd", "vcbuildtools.bat", ... Parameters ---------- arch: str Target architecture. vc_ver: float Required Microsoft Visual C++ version. If not set, autodetect the last version. vc_min_ver: float Minimum Microsoft Visual C++ version. """ # Variables and properties in this class use originals CamelCase variables # names from Microsoft source files for more easy comparison. def __init__(self, arch, vc_ver=None, vc_min_ver=0): self.pi = PlatformInfo(arch) self.ri = RegistryInfo(self.pi) self.si = SystemInfo(self.ri, vc_ver) if self.vc_ver < vc_min_ver: err = 'No suitable Microsoft Visual C++ version found' raise distutils.errors.DistutilsPlatformError(err) @property def vs_ver(self): """ Microsoft Visual Studio. Return ------ float version """ return self.si.vs_ver @property def vc_ver(self): """ Microsoft Visual C++ version. Return ------ float version """ return self.si.vc_ver @property def VSTools(self): """ Microsoft Visual Studio Tools. Return ------ list of str paths """ paths = [r'Common7\IDE', r'Common7\Tools'] if self.vs_ver >= 14.0: arch_subdir = self.pi.current_dir(hidex86=True, x64=True) paths += [r'Common7\IDE\CommonExtensions\Microsoft\TestWindow'] paths += [r'Team Tools\Performance Tools'] paths += [r'Team Tools\Performance Tools%s' % arch_subdir] return [join(self.si.VSInstallDir, path) for path in paths] @property def VCIncludes(self): """ Microsoft Visual C++ & Microsoft Foundation Class Includes. Return ------ list of str paths """ return [join(self.si.VCInstallDir, 'Include'), join(self.si.VCInstallDir, r'ATLMFC\Include')] @property def VCLibraries(self): """ Microsoft Visual C++ & Microsoft Foundation Class Libraries. Return ------ list of str paths """ if self.vs_ver >= 15.0: arch_subdir = self.pi.target_dir(x64=True) else: arch_subdir = self.pi.target_dir(hidex86=True) paths = ['Lib%s' % arch_subdir, r'ATLMFC\Lib%s' % arch_subdir] if self.vs_ver >= 14.0: paths += [r'Lib\store%s' % arch_subdir] return [join(self.si.VCInstallDir, path) for path in paths] @property def VCStoreRefs(self): """ Microsoft Visual C++ store references Libraries. Return ------ list of str paths """ if self.vs_ver < 14.0: return [] return [join(self.si.VCInstallDir, r'Lib\store\references')] @property def VCTools(self): """ Microsoft Visual C++ Tools. Return ------ list of str paths """ si = self.si tools = [join(si.VCInstallDir, 'VCPackages')] forcex86 = True if self.vs_ver <= 10.0 else False arch_subdir = self.pi.cross_dir(forcex86) if arch_subdir: tools += [join(si.VCInstallDir, 'Bin%s' % arch_subdir)] if self.vs_ver == 14.0: path = 'Bin%s' % self.pi.current_dir(hidex86=True) tools += [join(si.VCInstallDir, path)] elif self.vs_ver >= 15.0: host_dir = (r'bin\HostX86%s' if self.pi.current_is_x86() else r'bin\HostX64%s') tools += [join( si.VCInstallDir, host_dir % self.pi.target_dir(x64=True))] if self.pi.current_cpu != self.pi.target_cpu: tools += [join( si.VCInstallDir, host_dir % self.pi.current_dir(x64=True))] else: tools += [join(si.VCInstallDir, 'Bin')] return tools @property def OSLibraries(self): """ Microsoft Windows SDK Libraries. Return ------ list of str paths """ if self.vs_ver <= 10.0: arch_subdir = self.pi.target_dir(hidex86=True, x64=True) return [join(self.si.WindowsSdkDir, 'Lib%s' % arch_subdir)] else: arch_subdir = self.pi.target_dir(x64=True) lib = join(self.si.WindowsSdkDir, 'lib') libver = self._sdk_subdir return [join(lib, '%sum%s' % (libver , arch_subdir))] @property def OSIncludes(self): """ Microsoft Windows SDK Include. Return ------ list of str paths """ include = join(self.si.WindowsSdkDir, 'include') if self.vs_ver <= 10.0: return [include, join(include, 'gl')] else: if self.vs_ver >= 14.0: sdkver = self._sdk_subdir else: sdkver = '' return [join(include, '%sshared' % sdkver), join(include, '%sum' % sdkver), join(include, '%swinrt' % sdkver)] @property def OSLibpath(self): """ Microsoft Windows SDK Libraries Paths. Return ------ list of str paths """ ref = join(self.si.WindowsSdkDir, 'References') libpath = [] if self.vs_ver <= 9.0: libpath += self.OSLibraries if self.vs_ver >= 11.0: libpath += [join(ref, r'CommonConfiguration\Neutral')] if self.vs_ver >= 14.0: libpath += [ ref, join(self.si.WindowsSdkDir, 'UnionMetadata'), join(ref, 'Windows.Foundation.UniversalApiContract', '1.0.0.0'), join(ref, 'Windows.Foundation.FoundationContract', '1.0.0.0'), join(ref,'Windows.Networking.Connectivity.WwanContract', '1.0.0.0'), join(self.si.WindowsSdkDir, 'ExtensionSDKs', 'Microsoft.VCLibs', '%0.1f' % self.vs_ver, 'References', 'CommonConfiguration', 'neutral'), ] return libpath @property def SdkTools(self): """ Microsoft Windows SDK Tools. Return ------ list of str paths """ return list(self._sdk_tools()) def _sdk_tools(self): """ Microsoft Windows SDK Tools paths generator. Return ------ generator of str paths """ if self.vs_ver < 15.0: bin_dir = 'Bin' if self.vs_ver <= 11.0 else r'Bin\x86' yield join(self.si.WindowsSdkDir, bin_dir) if not self.pi.current_is_x86(): arch_subdir = self.pi.current_dir(x64=True) path = 'Bin%s' % arch_subdir yield join(self.si.WindowsSdkDir, path) if self.vs_ver in (10.0, 11.0): if self.pi.target_is_x86(): arch_subdir = '' else: arch_subdir = self.pi.current_dir(hidex86=True, x64=True) path = r'Bin\NETFX 4.0 Tools%s' % arch_subdir yield join(self.si.WindowsSdkDir, path) elif self.vs_ver >= 15.0: path = join(self.si.WindowsSdkDir, 'Bin') arch_subdir = self.pi.current_dir(x64=True) sdkver = self.si.WindowsSdkLastVersion yield join(path, '%s%s' % (sdkver, arch_subdir)) if self.si.WindowsSDKExecutablePath: yield self.si.WindowsSDKExecutablePath @property def _sdk_subdir(self): """ Microsoft Windows SDK version subdir. Return ------ str subdir """ ucrtver = self.si.WindowsSdkLastVersion return ('%s\\' % ucrtver) if ucrtver else '' @property def SdkSetup(self): """ Microsoft Windows SDK Setup. Return ------ list of str paths """ if self.vs_ver > 9.0: return [] return [join(self.si.WindowsSdkDir, 'Setup')] @property def FxTools(self): """ Microsoft .NET Framework Tools. Return ------ list of str paths """ pi = self.pi si = self.si if self.vs_ver <= 10.0: include32 = True include64 = not pi.target_is_x86() and not pi.current_is_x86() else: include32 = pi.target_is_x86() or pi.current_is_x86() include64 = pi.current_cpu == 'amd64' or pi.target_cpu == 'amd64' tools = [] if include32: tools += [join(si.FrameworkDir32, ver) for ver in si.FrameworkVersion32] if include64: tools += [join(si.FrameworkDir64, ver) for ver in si.FrameworkVersion64] return tools @property def NetFxSDKLibraries(self): """ Microsoft .Net Framework SDK Libraries. Return ------ list of str paths """ if self.vs_ver < 14.0 or not self.si.NetFxSdkDir: return [] arch_subdir = self.pi.target_dir(x64=True) return [join(self.si.NetFxSdkDir, r'lib\um%s' % arch_subdir)] @property def NetFxSDKIncludes(self): """ Microsoft .Net Framework SDK Includes. Return ------ list of str paths """ if self.vs_ver < 14.0 or not self.si.NetFxSdkDir: return [] return [join(self.si.NetFxSdkDir, r'include\um')] @property def VsTDb(self): """ Microsoft Visual Studio Team System Database. Return ------ list of str paths """ return [join(self.si.VSInstallDir, r'VSTSDB\Deploy')] @property def MSBuild(self): """ Microsoft Build Engine. Return ------ list of str paths """ if self.vs_ver < 12.0: return [] elif self.vs_ver < 15.0: base_path = self.si.ProgramFilesx86 arch_subdir = self.pi.current_dir(hidex86=True) else: base_path = self.si.VSInstallDir arch_subdir = '' path = r'MSBuild\%0.1f\bin%s' % (self.vs_ver, arch_subdir) build = [join(base_path, path)] if self.vs_ver >= 15.0: # Add Roslyn C# & Visual Basic Compiler build += [join(base_path, path, 'Roslyn')] return build @property def HTMLHelpWorkshop(self): """ Microsoft HTML Help Workshop. Return ------ list of str paths """ if self.vs_ver < 11.0: return [] return [join(self.si.ProgramFilesx86, 'HTML Help Workshop')] @property def UCRTLibraries(self): """ Microsoft Universal C Runtime SDK Libraries. Return ------ list of str paths """ if self.vs_ver < 14.0: return [] arch_subdir = self.pi.target_dir(x64=True) lib = join(self.si.UniversalCRTSdkDir, 'lib') ucrtver = self._ucrt_subdir return [join(lib, '%sucrt%s' % (ucrtver, arch_subdir))] @property def UCRTIncludes(self): """ Microsoft Universal C Runtime SDK Include. Return ------ list of str paths """ if self.vs_ver < 14.0: return [] include = join(self.si.UniversalCRTSdkDir, 'include') return [join(include, '%sucrt' % self._ucrt_subdir)] @property def _ucrt_subdir(self): """ Microsoft Universal C Runtime SDK version subdir. Return ------ str subdir """ ucrtver = self.si.UniversalCRTSdkLastVersion return ('%s\\' % ucrtver) if ucrtver else '' @property def FSharp(self): """ Microsoft Visual F#. Return ------ list of str paths """ if 11.0 > self.vs_ver > 12.0: return [] return [self.si.FSharpInstallDir] @property def VCRuntimeRedist(self): """ Microsoft Visual C++ runtime redistributable dll. Return ------ str path """ vcruntime = 'vcruntime%d0.dll' % self.vc_ver arch_subdir = self.pi.target_dir(x64=True).strip('\\') # Installation prefixes candidates prefixes = [] tools_path = self.si.VCInstallDir redist_path = dirname(tools_path.replace(r'\Tools', r'\Redist')) if isdir(redist_path): # Redist version may not be exactly the same as tools redist_path = join(redist_path, listdir(redist_path)[-1]) prefixes += [redist_path, join(redist_path, 'onecore')] prefixes += [join(tools_path, 'redist')] # VS14 legacy path # CRT directory crt_dirs = ('Microsoft.VC%d.CRT' % (self.vc_ver * 10), # Sometime store in directory with VS version instead of VC 'Microsoft.VC%d.CRT' % (int(self.vs_ver) * 10)) # vcruntime path for prefix, crt_dir in itertools.product(prefixes, crt_dirs): path = join(prefix, arch_subdir, crt_dir, vcruntime) if isfile(path): return path def return_env(self, exists=True): """ Return environment dict. Parameters ---------- exists: bool It True, only return existing paths. Return ------ dict environment """ env = dict( include=self._build_paths('include', [self.VCIncludes, self.OSIncludes, self.UCRTIncludes, self.NetFxSDKIncludes], exists), lib=self._build_paths('lib', [self.VCLibraries, self.OSLibraries, self.FxTools, self.UCRTLibraries, self.NetFxSDKLibraries], exists), libpath=self._build_paths('libpath', [self.VCLibraries, self.FxTools, self.VCStoreRefs, self.OSLibpath], exists), path=self._build_paths('path', [self.VCTools, self.VSTools, self.VsTDb, self.SdkTools, self.SdkSetup, self.FxTools, self.MSBuild, self.HTMLHelpWorkshop, self.FSharp], exists), ) if self.vs_ver >= 14 and isfile(self.VCRuntimeRedist): env['py_vcruntime_redist'] = self.VCRuntimeRedist return env def _build_paths(self, name, spec_path_lists, exists): """ Given an environment variable name and specified paths, return a pathsep-separated string of paths containing unique, extant, directories from those paths and from the environment variable. Raise an error if no paths are resolved. Parameters ---------- name: str Environment variable name spec_path_lists: list of str Paths exists: bool It True, only return existing paths. Return ------ str Pathsep-separated paths """ # flatten spec_path_lists spec_paths = itertools.chain.from_iterable(spec_path_lists) env_paths = environ.get(name, '').split(pathsep) paths = itertools.chain(spec_paths, env_paths) extant_paths = list(filter(isdir, paths)) if exists else paths if not extant_paths: msg = "%s environment variable is empty" % name.upper() raise distutils.errors.DistutilsPlatformError(msg) unique_paths = self._unique_everseen(extant_paths) return pathsep.join(unique_paths) # from Python docs @staticmethod def _unique_everseen(iterable, key=None): """ List unique elements, preserving order. Remember all elements ever seen. _unique_everseen('AAAABBBCCDAABBB') --> A B C D _unique_everseen('ABBCcAD', str.lower) --> A B C D """ seen = set() seen_add = seen.add if key is None: for element in filterfalse(seen.__contains__, iterable): seen_add(element) yield element else: for element in iterable: k = key(element) if k not in seen: seen_add(k) yield element site-packages/setuptools/py33compat.py000064400000002462151464452710014076 0ustar00import dis import array import collections try: import html except ImportError: html = None from setuptools.extern import six from setuptools.extern.six.moves import html_parser __metaclass__ = type OpArg = collections.namedtuple('OpArg', 'opcode arg') class Bytecode_compat: def __init__(self, code): self.code = code def __iter__(self): """Yield '(op,arg)' pair for each operation in code object 'code'""" bytes = array.array('b', self.code.co_code) eof = len(self.code.co_code) ptr = 0 extended_arg = 0 while ptr < eof: op = bytes[ptr] if op >= dis.HAVE_ARGUMENT: arg = bytes[ptr + 1] + bytes[ptr + 2] * 256 + extended_arg ptr += 3 if op == dis.EXTENDED_ARG: long_type = six.integer_types[-1] extended_arg = arg * long_type(65536) continue else: arg = None ptr += 1 yield OpArg(op, arg) Bytecode = getattr(dis, 'Bytecode', Bytecode_compat) unescape = getattr(html, 'unescape', None) if unescape is None: # HTMLParser.unescape is deprecated since Python 3.4, and will be removed # from 3.9. unescape = html_parser.HTMLParser().unescape site-packages/setuptools/config.py000064400000050066151464452710013344 0ustar00from __future__ import absolute_import, unicode_literals import io import os import sys import warnings import functools from collections import defaultdict from functools import partial from functools import wraps from importlib import import_module from distutils.errors import DistutilsOptionError, DistutilsFileError from setuptools.extern.packaging.version import LegacyVersion, parse from setuptools.extern.packaging.specifiers import SpecifierSet from setuptools.extern.six import string_types, PY3 __metaclass__ = type def read_configuration( filepath, find_others=False, ignore_option_errors=False): """Read given configuration file and returns options from it as a dict. :param str|unicode filepath: Path to configuration file to get options from. :param bool find_others: Whether to search for other configuration files which could be on in various places. :param bool ignore_option_errors: Whether to silently ignore options, values of which could not be resolved (e.g. due to exceptions in directives such as file:, attr:, etc.). If False exceptions are propagated as expected. :rtype: dict """ from setuptools.dist import Distribution, _Distribution filepath = os.path.abspath(filepath) if not os.path.isfile(filepath): raise DistutilsFileError( 'Configuration file %s does not exist.' % filepath) current_directory = os.getcwd() os.chdir(os.path.dirname(filepath)) try: dist = Distribution() filenames = dist.find_config_files() if find_others else [] if filepath not in filenames: filenames.append(filepath) _Distribution.parse_config_files(dist, filenames=filenames) handlers = parse_configuration( dist, dist.command_options, ignore_option_errors=ignore_option_errors) finally: os.chdir(current_directory) return configuration_to_dict(handlers) def _get_option(target_obj, key): """ Given a target object and option key, get that option from the target object, either through a get_{key} method or from an attribute directly. """ getter_name = 'get_{key}'.format(**locals()) by_attribute = functools.partial(getattr, target_obj, key) getter = getattr(target_obj, getter_name, by_attribute) return getter() def configuration_to_dict(handlers): """Returns configuration data gathered by given handlers as a dict. :param list[ConfigHandler] handlers: Handlers list, usually from parse_configuration() :rtype: dict """ config_dict = defaultdict(dict) for handler in handlers: for option in handler.set_options: value = _get_option(handler.target_obj, option) config_dict[handler.section_prefix][option] = value return config_dict def parse_configuration( distribution, command_options, ignore_option_errors=False): """Performs additional parsing of configuration options for a distribution. Returns a list of used option handlers. :param Distribution distribution: :param dict command_options: :param bool ignore_option_errors: Whether to silently ignore options, values of which could not be resolved (e.g. due to exceptions in directives such as file:, attr:, etc.). If False exceptions are propagated as expected. :rtype: list """ options = ConfigOptionsHandler( distribution, command_options, ignore_option_errors) options.parse() meta = ConfigMetadataHandler( distribution.metadata, command_options, ignore_option_errors, distribution.package_dir) meta.parse() return meta, options class ConfigHandler: """Handles metadata supplied in configuration files.""" section_prefix = None """Prefix for config sections handled by this handler. Must be provided by class heirs. """ aliases = {} """Options aliases. For compatibility with various packages. E.g.: d2to1 and pbr. Note: `-` in keys is replaced with `_` by config parser. """ def __init__(self, target_obj, options, ignore_option_errors=False): sections = {} section_prefix = self.section_prefix for section_name, section_options in options.items(): if not section_name.startswith(section_prefix): continue section_name = section_name.replace(section_prefix, '').strip('.') sections[section_name] = section_options self.ignore_option_errors = ignore_option_errors self.target_obj = target_obj self.sections = sections self.set_options = [] @property def parsers(self): """Metadata item name to parser function mapping.""" raise NotImplementedError( '%s must provide .parsers property' % self.__class__.__name__) def __setitem__(self, option_name, value): unknown = tuple() target_obj = self.target_obj # Translate alias into real name. option_name = self.aliases.get(option_name, option_name) current_value = getattr(target_obj, option_name, unknown) if current_value is unknown: raise KeyError(option_name) if current_value: # Already inhabited. Skipping. return skip_option = False parser = self.parsers.get(option_name) if parser: try: value = parser(value) except Exception: skip_option = True if not self.ignore_option_errors: raise if skip_option: return setter = getattr(target_obj, 'set_%s' % option_name, None) if setter is None: setattr(target_obj, option_name, value) else: setter(value) self.set_options.append(option_name) @classmethod def _parse_list(cls, value, separator=','): """Represents value as a list. Value is split either by separator (defaults to comma) or by lines. :param value: :param separator: List items separator character. :rtype: list """ if isinstance(value, list): # _get_parser_compound case return value if '\n' in value: value = value.splitlines() else: value = value.split(separator) return [chunk.strip() for chunk in value if chunk.strip()] @classmethod def _parse_dict(cls, value): """Represents value as a dict. :param value: :rtype: dict """ separator = '=' result = {} for line in cls._parse_list(value): key, sep, val = line.partition(separator) if sep != separator: raise DistutilsOptionError( 'Unable to parse option value to dict: %s' % value) result[key.strip()] = val.strip() return result @classmethod def _parse_bool(cls, value): """Represents value as boolean. :param value: :rtype: bool """ value = value.lower() return value in ('1', 'true', 'yes') @classmethod def _exclude_files_parser(cls, key): """Returns a parser function to make sure field inputs are not files. Parses a value after getting the key so error messages are more informative. :param key: :rtype: callable """ def parser(value): exclude_directive = 'file:' if value.startswith(exclude_directive): raise ValueError( 'Only strings are accepted for the {0} field, ' 'files are not accepted'.format(key)) return value return parser @classmethod def _parse_file(cls, value): """Represents value as a string, allowing including text from nearest files using `file:` directive. Directive is sandboxed and won't reach anything outside directory with setup.py. Examples: file: README.rst, CHANGELOG.md, src/file.txt :param str value: :rtype: str """ include_directive = 'file:' if not isinstance(value, string_types): return value if not value.startswith(include_directive): return value spec = value[len(include_directive):] filepaths = (os.path.abspath(path.strip()) for path in spec.split(',')) return '\n'.join( cls._read_file(path) for path in filepaths if (cls._assert_local(path) or True) and os.path.isfile(path) ) @staticmethod def _assert_local(filepath): if not filepath.startswith(os.getcwd()): raise DistutilsOptionError( '`file:` directive can not access %s' % filepath) @staticmethod def _read_file(filepath): with io.open(filepath, encoding='utf-8') as f: return f.read() @classmethod def _parse_attr(cls, value, package_dir=None): """Represents value as a module attribute. Examples: attr: package.attr attr: package.module.attr :param str value: :rtype: str """ attr_directive = 'attr:' if not value.startswith(attr_directive): return value attrs_path = value.replace(attr_directive, '').strip().split('.') attr_name = attrs_path.pop() module_name = '.'.join(attrs_path) module_name = module_name or '__init__' parent_path = os.getcwd() if package_dir: if attrs_path[0] in package_dir: # A custom path was specified for the module we want to import custom_path = package_dir[attrs_path[0]] parts = custom_path.rsplit('/', 1) if len(parts) > 1: parent_path = os.path.join(os.getcwd(), parts[0]) module_name = parts[1] else: module_name = custom_path elif '' in package_dir: # A custom parent directory was specified for all root modules parent_path = os.path.join(os.getcwd(), package_dir['']) sys.path.insert(0, parent_path) try: module = import_module(module_name) value = getattr(module, attr_name) finally: sys.path = sys.path[1:] return value @classmethod def _get_parser_compound(cls, *parse_methods): """Returns parser function to represents value as a list. Parses a value applying given methods one after another. :param parse_methods: :rtype: callable """ def parse(value): parsed = value for method in parse_methods: parsed = method(parsed) return parsed return parse @classmethod def _parse_section_to_dict(cls, section_options, values_parser=None): """Parses section options into a dictionary. Optionally applies a given parser to values. :param dict section_options: :param callable values_parser: :rtype: dict """ value = {} values_parser = values_parser or (lambda val: val) for key, (_, val) in section_options.items(): value[key] = values_parser(val) return value def parse_section(self, section_options): """Parses configuration file section. :param dict section_options: """ for (name, (_, value)) in section_options.items(): try: self[name] = value except KeyError: pass # Keep silent for a new option may appear anytime. def parse(self): """Parses configuration file items from one or more related sections. """ for section_name, section_options in self.sections.items(): method_postfix = '' if section_name: # [section.option] variant method_postfix = '_%s' % section_name section_parser_method = getattr( self, # Dots in section names are translated into dunderscores. ('parse_section%s' % method_postfix).replace('.', '__'), None) if section_parser_method is None: raise DistutilsOptionError( 'Unsupported distribution option section: [%s.%s]' % ( self.section_prefix, section_name)) section_parser_method(section_options) def _deprecated_config_handler(self, func, msg, warning_class): """ this function will wrap around parameters that are deprecated :param msg: deprecation message :param warning_class: class of warning exception to be raised :param func: function to be wrapped around """ @wraps(func) def config_handler(*args, **kwargs): warnings.warn(msg, warning_class) return func(*args, **kwargs) return config_handler class ConfigMetadataHandler(ConfigHandler): section_prefix = 'metadata' aliases = { 'home_page': 'url', 'summary': 'description', 'classifier': 'classifiers', 'platform': 'platforms', } strict_mode = False """We need to keep it loose, to be partially compatible with `pbr` and `d2to1` packages which also uses `metadata` section. """ def __init__(self, target_obj, options, ignore_option_errors=False, package_dir=None): super(ConfigMetadataHandler, self).__init__(target_obj, options, ignore_option_errors) self.package_dir = package_dir @property def parsers(self): """Metadata item name to parser function mapping.""" parse_list = self._parse_list parse_file = self._parse_file parse_dict = self._parse_dict exclude_files_parser = self._exclude_files_parser return { 'platforms': parse_list, 'keywords': parse_list, 'provides': parse_list, 'requires': self._deprecated_config_handler( parse_list, "The requires parameter is deprecated, please use " "install_requires for runtime dependencies.", DeprecationWarning), 'obsoletes': parse_list, 'classifiers': self._get_parser_compound(parse_file, parse_list), 'license': exclude_files_parser('license'), 'description': parse_file, 'long_description': parse_file, 'version': self._parse_version, 'project_urls': parse_dict, } def _parse_version(self, value): """Parses `version` option value. :param value: :rtype: str """ version = self._parse_file(value) if version != value: version = version.strip() # Be strict about versions loaded from file because it's easy to # accidentally include newlines and other unintended content if isinstance(parse(version), LegacyVersion): tmpl = ( 'Version loaded from {value} does not ' 'comply with PEP 440: {version}' ) raise DistutilsOptionError(tmpl.format(**locals())) return version version = self._parse_attr(value, self.package_dir) if callable(version): version = version() if not isinstance(version, string_types): if hasattr(version, '__iter__'): version = '.'.join(map(str, version)) else: version = '%s' % version return version class ConfigOptionsHandler(ConfigHandler): section_prefix = 'options' @property def parsers(self): """Metadata item name to parser function mapping.""" parse_list = self._parse_list parse_list_semicolon = partial(self._parse_list, separator=';') parse_bool = self._parse_bool parse_dict = self._parse_dict return { 'zip_safe': parse_bool, 'use_2to3': parse_bool, 'include_package_data': parse_bool, 'package_dir': parse_dict, 'use_2to3_fixers': parse_list, 'use_2to3_exclude_fixers': parse_list, 'convert_2to3_doctests': parse_list, 'scripts': parse_list, 'eager_resources': parse_list, 'dependency_links': parse_list, 'namespace_packages': parse_list, 'install_requires': parse_list_semicolon, 'setup_requires': parse_list_semicolon, 'tests_require': parse_list_semicolon, 'packages': self._parse_packages, 'entry_points': self._parse_file, 'py_modules': parse_list, 'python_requires': SpecifierSet, } def _parse_packages(self, value): """Parses `packages` option value. :param value: :rtype: list """ find_directives = ['find:', 'find_namespace:'] trimmed_value = value.strip() if trimmed_value not in find_directives: return self._parse_list(value) findns = trimmed_value == find_directives[1] if findns and not PY3: raise DistutilsOptionError( 'find_namespace: directive is unsupported on Python < 3.3') # Read function arguments from a dedicated section. find_kwargs = self.parse_section_packages__find( self.sections.get('packages.find', {})) if findns: from setuptools import find_namespace_packages as find_packages else: from setuptools import find_packages return find_packages(**find_kwargs) def parse_section_packages__find(self, section_options): """Parses `packages.find` configuration file section. To be used in conjunction with _parse_packages(). :param dict section_options: """ section_data = self._parse_section_to_dict( section_options, self._parse_list) valid_keys = ['where', 'include', 'exclude'] find_kwargs = dict( [(k, v) for k, v in section_data.items() if k in valid_keys and v]) where = find_kwargs.get('where') if where is not None: find_kwargs['where'] = where[0] # cast list to single val return find_kwargs def parse_section_entry_points(self, section_options): """Parses `entry_points` configuration file section. :param dict section_options: """ parsed = self._parse_section_to_dict(section_options, self._parse_list) self['entry_points'] = parsed def _parse_package_data(self, section_options): parsed = self._parse_section_to_dict(section_options, self._parse_list) root = parsed.get('*') if root: parsed[''] = root del parsed['*'] return parsed def parse_section_package_data(self, section_options): """Parses `package_data` configuration file section. :param dict section_options: """ self['package_data'] = self._parse_package_data(section_options) def parse_section_exclude_package_data(self, section_options): """Parses `exclude_package_data` configuration file section. :param dict section_options: """ self['exclude_package_data'] = self._parse_package_data( section_options) def parse_section_extras_require(self, section_options): """Parses `extras_require` configuration file section. :param dict section_options: """ parse_list = partial(self._parse_list, separator=';') self['extras_require'] = self._parse_section_to_dict( section_options, parse_list) def parse_section_data_files(self, section_options): """Parses `data_files` configuration file section. :param dict section_options: """ parsed = self._parse_section_to_dict(section_options, self._parse_list) self['data_files'] = [(k, v) for k, v in parsed.items()] site-packages/setuptools/py27compat.py000064400000002730151464452710014077 0ustar00""" Compatibility Support for Python 2.7 and earlier """ import sys import platform from setuptools.extern import six def get_all_headers(message, key): """ Given an HTTPMessage, return all headers matching a given key. """ return message.get_all(key) if six.PY2: def get_all_headers(message, key): return message.getheaders(key) linux_py2_ascii = ( platform.system() == 'Linux' and six.PY2 ) rmtree_safe = str if linux_py2_ascii else lambda x: x """Workaround for http://bugs.python.org/issue24672""" try: from ._imp import find_module, PY_COMPILED, PY_FROZEN, PY_SOURCE from ._imp import get_frozen_object, get_module except ImportError: import imp from imp import PY_COMPILED, PY_FROZEN, PY_SOURCE # noqa def find_module(module, paths=None): """Just like 'imp.find_module()', but with package support""" parts = module.split('.') while parts: part = parts.pop(0) f, path, (suffix, mode, kind) = info = imp.find_module(part, paths) if kind == imp.PKG_DIRECTORY: parts = parts or ['__init__'] paths = [path] elif parts: raise ImportError("Can't find %r in %s" % (parts, module)) return info def get_frozen_object(module, paths): return imp.get_frozen_object(module) def get_module(module, paths, info): imp.load_module(module, *info) return sys.modules[module] site-packages/setuptools/_deprecation_warning.py000064400000000332151464452710016247 0ustar00class SetuptoolsDeprecationWarning(Warning): """ Base class for warning deprecations in ``setuptools`` This class is not derived from ``DeprecationWarning``, and as such is visible by default. """ site-packages/setuptools/monkey.py000064400000012220151464452710013367 0ustar00""" Monkey patching of distutils. """ import sys import distutils.filelist import platform import types import functools from importlib import import_module import inspect from setuptools.extern import six import setuptools __all__ = [] """ Everything is private. Contact the project team if you think you need this functionality. """ def _get_mro(cls): """ Returns the bases classes for cls sorted by the MRO. Works around an issue on Jython where inspect.getmro will not return all base classes if multiple classes share the same name. Instead, this function will return a tuple containing the class itself, and the contents of cls.__bases__. See https://github.com/pypa/setuptools/issues/1024. """ if platform.python_implementation() == "Jython": return (cls,) + cls.__bases__ return inspect.getmro(cls) def get_unpatched(item): lookup = ( get_unpatched_class if isinstance(item, six.class_types) else get_unpatched_function if isinstance(item, types.FunctionType) else lambda item: None ) return lookup(item) def get_unpatched_class(cls): """Protect against re-patching the distutils if reloaded Also ensures that no other distutils extension monkeypatched the distutils first. """ external_bases = ( cls for cls in _get_mro(cls) if not cls.__module__.startswith('setuptools') ) base = next(external_bases) if not base.__module__.startswith('distutils'): msg = "distutils has already been patched by %r" % cls raise AssertionError(msg) return base def patch_all(): # we can't patch distutils.cmd, alas distutils.core.Command = setuptools.Command has_issue_12885 = sys.version_info <= (3, 5, 3) if has_issue_12885: # fix findall bug in distutils (http://bugs.python.org/issue12885) distutils.filelist.findall = setuptools.findall needs_warehouse = ( sys.version_info < (2, 7, 13) or (3, 4) < sys.version_info < (3, 4, 6) or (3, 5) < sys.version_info <= (3, 5, 3) ) if needs_warehouse: warehouse = 'https://upload.pypi.org/legacy/' distutils.config.PyPIRCCommand.DEFAULT_REPOSITORY = warehouse _patch_distribution_metadata() # Install Distribution throughout the distutils for module in distutils.dist, distutils.core, distutils.cmd: module.Distribution = setuptools.dist.Distribution # Install the patched Extension distutils.core.Extension = setuptools.extension.Extension distutils.extension.Extension = setuptools.extension.Extension if 'distutils.command.build_ext' in sys.modules: sys.modules['distutils.command.build_ext'].Extension = ( setuptools.extension.Extension ) patch_for_msvc_specialized_compiler() def _patch_distribution_metadata(): """Patch write_pkg_file and read_pkg_file for higher metadata standards""" for attr in ('write_pkg_file', 'read_pkg_file', 'get_metadata_version'): new_val = getattr(setuptools.dist, attr) setattr(distutils.dist.DistributionMetadata, attr, new_val) def patch_func(replacement, target_mod, func_name): """ Patch func_name in target_mod with replacement Important - original must be resolved by name to avoid patching an already patched function. """ original = getattr(target_mod, func_name) # set the 'unpatched' attribute on the replacement to # point to the original. vars(replacement).setdefault('unpatched', original) # replace the function in the original module setattr(target_mod, func_name, replacement) def get_unpatched_function(candidate): return getattr(candidate, 'unpatched') def patch_for_msvc_specialized_compiler(): """ Patch functions in distutils to use standalone Microsoft Visual C++ compilers. """ # import late to avoid circular imports on Python < 3.5 msvc = import_module('setuptools.msvc') if platform.system() != 'Windows': # Compilers only availables on Microsoft Windows return def patch_params(mod_name, func_name): """ Prepare the parameters for patch_func to patch indicated function. """ repl_prefix = 'msvc9_' if 'msvc9' in mod_name else 'msvc14_' repl_name = repl_prefix + func_name.lstrip('_') repl = getattr(msvc, repl_name) mod = import_module(mod_name) if not hasattr(mod, func_name): raise ImportError(func_name) return repl, mod, func_name # Python 2.7 to 3.4 msvc9 = functools.partial(patch_params, 'distutils.msvc9compiler') # Python 3.5+ msvc14 = functools.partial(patch_params, 'distutils._msvccompiler') try: # Patch distutils.msvc9compiler patch_func(*msvc9('find_vcvarsall')) patch_func(*msvc9('query_vcvarsall')) except ImportError: pass try: # Patch distutils._msvccompiler._get_vc_env patch_func(*msvc14('_get_vc_env')) except ImportError: pass try: # Patch distutils._msvccompiler.gen_lib_options for Numpy patch_func(*msvc14('gen_lib_options')) except ImportError: pass site-packages/setuptools/version.py000064400000000220151464452710013547 0ustar00import pkg_resources try: __version__ = pkg_resources.get_distribution('setuptools').version except Exception: __version__ = 'unknown' site-packages/setuptools/package_index.py000064400000117235151464452710014663 0ustar00"""PyPI and direct package downloading""" import sys import os import re import shutil import socket import base64 import hashlib import itertools import warnings from functools import wraps from setuptools.extern import six from setuptools.extern.six.moves import urllib, http_client, configparser, map import setuptools from pkg_resources import ( CHECKOUT_DIST, Distribution, BINARY_DIST, normalize_path, SOURCE_DIST, Environment, find_distributions, safe_name, safe_version, to_filename, Requirement, DEVELOP_DIST, EGG_DIST, ) from setuptools import ssl_support from distutils import log from distutils.errors import DistutilsError from fnmatch import translate from setuptools.py27compat import get_all_headers from setuptools.py33compat import unescape from setuptools.wheel import Wheel __metaclass__ = type EGG_FRAGMENT = re.compile(r'^egg=([-A-Za-z0-9_.+!]+)$') HREF = re.compile(r"""href\s*=\s*['"]?([^'"> ]+)""", re.I) PYPI_MD5 = re.compile( r'([^<]+)\n\s+\(md5\)' ) URL_SCHEME = re.compile('([-+.a-z0-9]{2,}):', re.I).match EXTENSIONS = ".tar.gz .tar.bz2 .tar .zip .tgz".split() __all__ = [ 'PackageIndex', 'distros_for_url', 'parse_bdist_wininst', 'interpret_distro_name', ] _SOCKET_TIMEOUT = 15 _tmpl = "setuptools/{setuptools.__version__} Python-urllib/{py_major}" user_agent = _tmpl.format(py_major='{}.{}'.format(*sys.version_info), setuptools=setuptools) def parse_requirement_arg(spec): try: return Requirement.parse(spec) except ValueError: raise DistutilsError( "Not a URL, existing file, or requirement spec: %r" % (spec,) ) def parse_bdist_wininst(name): """Return (base,pyversion) or (None,None) for possible .exe name""" lower = name.lower() base, py_ver, plat = None, None, None if lower.endswith('.exe'): if lower.endswith('.win32.exe'): base = name[:-10] plat = 'win32' elif lower.startswith('.win32-py', -16): py_ver = name[-7:-4] base = name[:-16] plat = 'win32' elif lower.endswith('.win-amd64.exe'): base = name[:-14] plat = 'win-amd64' elif lower.startswith('.win-amd64-py', -20): py_ver = name[-7:-4] base = name[:-20] plat = 'win-amd64' return base, py_ver, plat def egg_info_for_url(url): parts = urllib.parse.urlparse(url) scheme, server, path, parameters, query, fragment = parts base = urllib.parse.unquote(path.split('/')[-1]) if server == 'sourceforge.net' and base == 'download': # XXX Yuck base = urllib.parse.unquote(path.split('/')[-2]) if '#' in base: base, fragment = base.split('#', 1) return base, fragment def distros_for_url(url, metadata=None): """Yield egg or source distribution objects that might be found at a URL""" base, fragment = egg_info_for_url(url) for dist in distros_for_location(url, base, metadata): yield dist if fragment: match = EGG_FRAGMENT.match(fragment) if match: for dist in interpret_distro_name( url, match.group(1), metadata, precedence=CHECKOUT_DIST ): yield dist def distros_for_location(location, basename, metadata=None): """Yield egg or source distribution objects based on basename""" if basename.endswith('.egg.zip'): basename = basename[:-4] # strip the .zip if basename.endswith('.egg') and '-' in basename: # only one, unambiguous interpretation return [Distribution.from_location(location, basename, metadata)] if basename.endswith('.whl') and '-' in basename: wheel = Wheel(basename) if not wheel.is_compatible(): return [] return [Distribution( location=location, project_name=wheel.project_name, version=wheel.version, # Increase priority over eggs. precedence=EGG_DIST + 1, )] if basename.endswith('.exe'): win_base, py_ver, platform = parse_bdist_wininst(basename) if win_base is not None: return interpret_distro_name( location, win_base, metadata, py_ver, BINARY_DIST, platform ) # Try source distro extensions (.zip, .tgz, etc.) # for ext in EXTENSIONS: if basename.endswith(ext): basename = basename[:-len(ext)] return interpret_distro_name(location, basename, metadata) return [] # no extension matched def distros_for_filename(filename, metadata=None): """Yield possible egg or source distribution objects based on a filename""" return distros_for_location( normalize_path(filename), os.path.basename(filename), metadata ) def interpret_distro_name( location, basename, metadata, py_version=None, precedence=SOURCE_DIST, platform=None ): """Generate alternative interpretations of a source distro name Note: if `location` is a filesystem filename, you should call ``pkg_resources.normalize_path()`` on it before passing it to this routine! """ # Generate alternative interpretations of a source distro name # Because some packages are ambiguous as to name/versions split # e.g. "adns-python-1.1.0", "egenix-mx-commercial", etc. # So, we generate each possible interepretation (e.g. "adns, python-1.1.0" # "adns-python, 1.1.0", and "adns-python-1.1.0, no version"). In practice, # the spurious interpretations should be ignored, because in the event # there's also an "adns" package, the spurious "python-1.1.0" version will # compare lower than any numeric version number, and is therefore unlikely # to match a request for it. It's still a potential problem, though, and # in the long run PyPI and the distutils should go for "safe" names and # versions in distribution archive names (sdist and bdist). parts = basename.split('-') if not py_version and any(re.match(r'py\d\.\d$', p) for p in parts[2:]): # it is a bdist_dumb, not an sdist -- bail out return for p in range(1, len(parts) + 1): yield Distribution( location, metadata, '-'.join(parts[:p]), '-'.join(parts[p:]), py_version=py_version, precedence=precedence, platform=platform ) # From Python 2.7 docs def unique_everseen(iterable, key=None): "List unique elements, preserving order. Remember all elements ever seen." # unique_everseen('AAAABBBCCDAABBB') --> A B C D # unique_everseen('ABBCcAD', str.lower) --> A B C D seen = set() seen_add = seen.add if key is None: for element in six.moves.filterfalse(seen.__contains__, iterable): seen_add(element) yield element else: for element in iterable: k = key(element) if k not in seen: seen_add(k) yield element def unique_values(func): """ Wrap a function returning an iterable such that the resulting iterable only ever yields unique items. """ @wraps(func) def wrapper(*args, **kwargs): return unique_everseen(func(*args, **kwargs)) return wrapper REL = re.compile(r"""<([^>]*\srel\s*=\s*['"]?([^'">]+)[^>]*)>""", re.I) # this line is here to fix emacs' cruddy broken syntax highlighting @unique_values def find_external_links(url, page): """Find rel="homepage" and rel="download" links in `page`, yielding URLs""" for match in REL.finditer(page): tag, rel = match.groups() rels = set(map(str.strip, rel.lower().split(','))) if 'homepage' in rels or 'download' in rels: for match in HREF.finditer(tag): yield urllib.parse.urljoin(url, htmldecode(match.group(1))) for tag in ("Home Page", "Download URL"): pos = page.find(tag) if pos != -1: match = HREF.search(page, pos) if match: yield urllib.parse.urljoin(url, htmldecode(match.group(1))) class ContentChecker: """ A null content checker that defines the interface for checking content """ def feed(self, block): """ Feed a block of data to the hash. """ return def is_valid(self): """ Check the hash. Return False if validation fails. """ return True def report(self, reporter, template): """ Call reporter with information about the checker (hash name) substituted into the template. """ return class HashChecker(ContentChecker): pattern = re.compile( r'(?Psha1|sha224|sha384|sha256|sha512|md5)=' r'(?P[a-f0-9]+)' ) def __init__(self, hash_name, expected): self.hash_name = hash_name self.hash = hashlib.new(hash_name) self.expected = expected @classmethod def from_url(cls, url): "Construct a (possibly null) ContentChecker from a URL" fragment = urllib.parse.urlparse(url)[-1] if not fragment: return ContentChecker() match = cls.pattern.search(fragment) if not match: return ContentChecker() return cls(**match.groupdict()) def feed(self, block): self.hash.update(block) def is_valid(self): return self.hash.hexdigest() == self.expected def report(self, reporter, template): msg = template % self.hash_name return reporter(msg) class PackageIndex(Environment): """A distribution index that scans web pages for download URLs""" def __init__( self, index_url="https://pypi.org/simple/", hosts=('*',), ca_bundle=None, verify_ssl=True, *args, **kw ): Environment.__init__(self, *args, **kw) self.index_url = index_url + "/" [:not index_url.endswith('/')] self.scanned_urls = {} self.fetched_urls = {} self.package_pages = {} self.allows = re.compile('|'.join(map(translate, hosts))).match self.to_scan = [] use_ssl = ( verify_ssl and ssl_support.is_available and (ca_bundle or ssl_support.find_ca_bundle()) ) if use_ssl: self.opener = ssl_support.opener_for(ca_bundle) else: self.opener = urllib.request.urlopen def process_url(self, url, retrieve=False): """Evaluate a URL as a possible download, and maybe retrieve it""" if url in self.scanned_urls and not retrieve: return self.scanned_urls[url] = True if not URL_SCHEME(url): self.process_filename(url) return else: dists = list(distros_for_url(url)) if dists: if not self.url_ok(url): return self.debug("Found link: %s", url) if dists or not retrieve or url in self.fetched_urls: list(map(self.add, dists)) return # don't need the actual page if not self.url_ok(url): self.fetched_urls[url] = True return self.info("Reading %s", url) self.fetched_urls[url] = True # prevent multiple fetch attempts tmpl = "Download error on %s: %%s -- Some packages may not be found!" f = self.open_url(url, tmpl % url) if f is None: return self.fetched_urls[f.url] = True if 'html' not in f.headers.get('content-type', '').lower(): f.close() # not html, we can't process it return base = f.url # handle redirects page = f.read() if not isinstance(page, str): # In Python 3 and got bytes but want str. if isinstance(f, urllib.error.HTTPError): # Errors have no charset, assume latin1: charset = 'latin-1' else: charset = f.headers.get_param('charset') or 'latin-1' page = page.decode(charset, "ignore") f.close() for match in HREF.finditer(page): link = urllib.parse.urljoin(base, htmldecode(match.group(1))) self.process_url(link) if url.startswith(self.index_url) and getattr(f, 'code', None) != 404: page = self.process_index(url, page) def process_filename(self, fn, nested=False): # process filenames or directories if not os.path.exists(fn): self.warn("Not found: %s", fn) return if os.path.isdir(fn) and not nested: path = os.path.realpath(fn) for item in os.listdir(path): self.process_filename(os.path.join(path, item), True) dists = distros_for_filename(fn) if dists: self.debug("Found: %s", fn) list(map(self.add, dists)) def url_ok(self, url, fatal=False): s = URL_SCHEME(url) is_file = s and s.group(1).lower() == 'file' if is_file or self.allows(urllib.parse.urlparse(url)[1]): return True msg = ( "\nNote: Bypassing %s (disallowed host; see " "http://bit.ly/2hrImnY for details).\n") if fatal: raise DistutilsError(msg % url) else: self.warn(msg, url) def scan_egg_links(self, search_path): dirs = filter(os.path.isdir, search_path) egg_links = ( (path, entry) for path in dirs for entry in os.listdir(path) if entry.endswith('.egg-link') ) list(itertools.starmap(self.scan_egg_link, egg_links)) def scan_egg_link(self, path, entry): with open(os.path.join(path, entry)) as raw_lines: # filter non-empty lines lines = list(filter(None, map(str.strip, raw_lines))) if len(lines) != 2: # format is not recognized; punt return egg_path, setup_path = lines for dist in find_distributions(os.path.join(path, egg_path)): dist.location = os.path.join(path, *lines) dist.precedence = SOURCE_DIST self.add(dist) def process_index(self, url, page): """Process the contents of a PyPI page""" def scan(link): # Process a URL to see if it's for a package page if link.startswith(self.index_url): parts = list(map( urllib.parse.unquote, link[len(self.index_url):].split('/') )) if len(parts) == 2 and '#' not in parts[1]: # it's a package page, sanitize and index it pkg = safe_name(parts[0]) ver = safe_version(parts[1]) self.package_pages.setdefault(pkg.lower(), {})[link] = True return to_filename(pkg), to_filename(ver) return None, None # process an index page into the package-page index for match in HREF.finditer(page): try: scan(urllib.parse.urljoin(url, htmldecode(match.group(1)))) except ValueError: pass pkg, ver = scan(url) # ensure this page is in the page index if pkg: # process individual package page for new_url in find_external_links(url, page): # Process the found URL base, frag = egg_info_for_url(new_url) if base.endswith('.py') and not frag: if ver: new_url += '#egg=%s-%s' % (pkg, ver) else: self.need_version_info(url) self.scan_url(new_url) return PYPI_MD5.sub( lambda m: '%s' % m.group(1, 3, 2), page ) else: return "" # no sense double-scanning non-package pages def need_version_info(self, url): self.scan_all( "Page at %s links to .py file(s) without version info; an index " "scan is required.", url ) def scan_all(self, msg=None, *args): if self.index_url not in self.fetched_urls: if msg: self.warn(msg, *args) self.info( "Scanning index of all packages (this may take a while)" ) self.scan_url(self.index_url) def find_packages(self, requirement): self.scan_url(self.index_url + requirement.unsafe_name + '/') if not self.package_pages.get(requirement.key): # Fall back to safe version of the name self.scan_url(self.index_url + requirement.project_name + '/') if not self.package_pages.get(requirement.key): # We couldn't find the target package, so search the index page too self.not_found_in_index(requirement) for url in list(self.package_pages.get(requirement.key, ())): # scan each page that might be related to the desired package self.scan_url(url) def obtain(self, requirement, installer=None): self.prescan() self.find_packages(requirement) for dist in self[requirement.key]: if dist in requirement: return dist self.debug("%s does not match %s", requirement, dist) return super(PackageIndex, self).obtain(requirement, installer) def check_hash(self, checker, filename, tfp): """ checker is a ContentChecker """ checker.report( self.debug, "Validating %%s checksum for %s" % filename) if not checker.is_valid(): tfp.close() os.unlink(filename) raise DistutilsError( "%s validation failed for %s; " "possible download problem?" % (checker.hash.name, os.path.basename(filename)) ) def add_find_links(self, urls): """Add `urls` to the list that will be prescanned for searches""" for url in urls: if ( self.to_scan is None # if we have already "gone online" or not URL_SCHEME(url) # or it's a local file/directory or url.startswith('file:') or list(distros_for_url(url)) # or a direct package link ): # then go ahead and process it now self.scan_url(url) else: # otherwise, defer retrieval till later self.to_scan.append(url) def prescan(self): """Scan urls scheduled for prescanning (e.g. --find-links)""" if self.to_scan: list(map(self.scan_url, self.to_scan)) self.to_scan = None # from now on, go ahead and process immediately def not_found_in_index(self, requirement): if self[requirement.key]: # we've seen at least one distro meth, msg = self.info, "Couldn't retrieve index page for %r" else: # no distros seen for this name, might be misspelled meth, msg = ( self.warn, "Couldn't find index page for %r (maybe misspelled?)") meth(msg, requirement.unsafe_name) self.scan_all() def download(self, spec, tmpdir): """Locate and/or download `spec` to `tmpdir`, returning a local path `spec` may be a ``Requirement`` object, or a string containing a URL, an existing local filename, or a project/version requirement spec (i.e. the string form of a ``Requirement`` object). If it is the URL of a .py file with an unambiguous ``#egg=name-version`` tag (i.e., one that escapes ``-`` as ``_`` throughout), a trivial ``setup.py`` is automatically created alongside the downloaded file. If `spec` is a ``Requirement`` object or a string containing a project/version requirement spec, this method returns the location of a matching distribution (possibly after downloading it to `tmpdir`). If `spec` is a locally existing file or directory name, it is simply returned unchanged. If `spec` is a URL, it is downloaded to a subpath of `tmpdir`, and the local filename is returned. Various errors may be raised if a problem occurs during downloading. """ if not isinstance(spec, Requirement): scheme = URL_SCHEME(spec) if scheme: # It's a url, download it to tmpdir found = self._download_url(scheme.group(1), spec, tmpdir) base, fragment = egg_info_for_url(spec) if base.endswith('.py'): found = self.gen_setup(found, fragment, tmpdir) return found elif os.path.exists(spec): # Existing file or directory, just return it return spec else: spec = parse_requirement_arg(spec) return getattr(self.fetch_distribution(spec, tmpdir), 'location', None) def fetch_distribution( self, requirement, tmpdir, force_scan=False, source=False, develop_ok=False, local_index=None): """Obtain a distribution suitable for fulfilling `requirement` `requirement` must be a ``pkg_resources.Requirement`` instance. If necessary, or if the `force_scan` flag is set, the requirement is searched for in the (online) package index as well as the locally installed packages. If a distribution matching `requirement` is found, the returned distribution's ``location`` is the value you would have gotten from calling the ``download()`` method with the matching distribution's URL or filename. If no matching distribution is found, ``None`` is returned. If the `source` flag is set, only source distributions and source checkout links will be considered. Unless the `develop_ok` flag is set, development and system eggs (i.e., those using the ``.egg-info`` format) will be ignored. """ # process a Requirement self.info("Searching for %s", requirement) skipped = {} dist = None def find(req, env=None): if env is None: env = self # Find a matching distribution; may be called more than once for dist in env[req.key]: if dist.precedence == DEVELOP_DIST and not develop_ok: if dist not in skipped: self.warn( "Skipping development or system egg: %s", dist, ) skipped[dist] = 1 continue test = ( dist in req and (dist.precedence <= SOURCE_DIST or not source) ) if test: loc = self.download(dist.location, tmpdir) dist.download_location = loc if os.path.exists(dist.download_location): return dist if force_scan: self.prescan() self.find_packages(requirement) dist = find(requirement) if not dist and local_index is not None: dist = find(requirement, local_index) if dist is None: if self.to_scan is not None: self.prescan() dist = find(requirement) if dist is None and not force_scan: self.find_packages(requirement) dist = find(requirement) if dist is None: self.warn( "No local packages or working download links found for %s%s", (source and "a source distribution of " or ""), requirement, ) else: self.info("Best match: %s", dist) return dist.clone(location=dist.download_location) def fetch(self, requirement, tmpdir, force_scan=False, source=False): """Obtain a file suitable for fulfilling `requirement` DEPRECATED; use the ``fetch_distribution()`` method now instead. For backward compatibility, this routine is identical but returns the ``location`` of the downloaded distribution instead of a distribution object. """ dist = self.fetch_distribution(requirement, tmpdir, force_scan, source) if dist is not None: return dist.location return None def gen_setup(self, filename, fragment, tmpdir): match = EGG_FRAGMENT.match(fragment) dists = match and [ d for d in interpret_distro_name(filename, match.group(1), None) if d.version ] or [] if len(dists) == 1: # unambiguous ``#egg`` fragment basename = os.path.basename(filename) # Make sure the file has been downloaded to the temp dir. if os.path.dirname(filename) != tmpdir: dst = os.path.join(tmpdir, basename) from setuptools.command.easy_install import samefile if not samefile(filename, dst): shutil.copy2(filename, dst) filename = dst with open(os.path.join(tmpdir, 'setup.py'), 'w') as file: file.write( "from setuptools import setup\n" "setup(name=%r, version=%r, py_modules=[%r])\n" % ( dists[0].project_name, dists[0].version, os.path.splitext(basename)[0] ) ) return filename elif match: raise DistutilsError( "Can't unambiguously interpret project/version identifier %r; " "any dashes in the name or version should be escaped using " "underscores. %r" % (fragment, dists) ) else: raise DistutilsError( "Can't process plain .py files without an '#egg=name-version'" " suffix to enable automatic setup script generation." ) dl_blocksize = 8192 def _download_to(self, url, filename): self.info("Downloading %s", url) # Download the file fp = None try: checker = HashChecker.from_url(url) fp = self.open_url(url) if isinstance(fp, urllib.error.HTTPError): raise DistutilsError( "Can't download %s: %s %s" % (url, fp.code, fp.msg) ) headers = fp.info() blocknum = 0 bs = self.dl_blocksize size = -1 if "content-length" in headers: # Some servers return multiple Content-Length headers :( sizes = get_all_headers(headers, 'Content-Length') size = max(map(int, sizes)) self.reporthook(url, filename, blocknum, bs, size) with open(filename, 'wb') as tfp: while True: block = fp.read(bs) if block: checker.feed(block) tfp.write(block) blocknum += 1 self.reporthook(url, filename, blocknum, bs, size) else: break self.check_hash(checker, filename, tfp) return headers finally: if fp: fp.close() def reporthook(self, url, filename, blocknum, blksize, size): pass # no-op def open_url(self, url, warning=None): if url.startswith('file:'): return local_open(url) try: return open_with_auth(url, self.opener) except (ValueError, http_client.InvalidURL) as v: msg = ' '.join([str(arg) for arg in v.args]) if warning: self.warn(warning, msg) else: raise DistutilsError('%s %s' % (url, msg)) except urllib.error.HTTPError as v: return v except urllib.error.URLError as v: if warning: self.warn(warning, v.reason) else: raise DistutilsError("Download error for %s: %s" % (url, v.reason)) except http_client.BadStatusLine as v: if warning: self.warn(warning, v.line) else: raise DistutilsError( '%s returned a bad status line. The server might be ' 'down, %s' % (url, v.line) ) except (http_client.HTTPException, socket.error) as v: if warning: self.warn(warning, v) else: raise DistutilsError("Download error for %s: %s" % (url, v)) def _download_url(self, scheme, url, tmpdir): # Determine download filename # name, fragment = egg_info_for_url(url) if name: while '..' in name: name = name.replace('..', '.').replace('\\', '_') else: name = "__downloaded__" # default if URL has no path contents if name.endswith('.egg.zip'): name = name[:-4] # strip the extra .zip before download filename = os.path.join(tmpdir, name) # Download the file # if scheme == 'svn' or scheme.startswith('svn+'): return self._download_svn(url, filename) elif scheme == 'git' or scheme.startswith('git+'): return self._download_git(url, filename) elif scheme.startswith('hg+'): return self._download_hg(url, filename) elif scheme == 'file': return urllib.request.url2pathname(urllib.parse.urlparse(url)[2]) else: self.url_ok(url, True) # raises error if not allowed return self._attempt_download(url, filename) def scan_url(self, url): self.process_url(url, True) def _attempt_download(self, url, filename): headers = self._download_to(url, filename) if 'html' in headers.get('content-type', '').lower(): return self._download_html(url, headers, filename) else: return filename def _download_html(self, url, headers, filename): file = open(filename) for line in file: if line.strip(): # Check for a subversion index page if re.search(r'([^- ]+ - )?Revision \d+:', line): # it's a subversion index page: file.close() os.unlink(filename) return self._download_svn(url, filename) break # not an index page file.close() os.unlink(filename) raise DistutilsError("Unexpected HTML page found at " + url) def _download_svn(self, url, filename): warnings.warn("SVN download support is deprecated", UserWarning) url = url.split('#', 1)[0] # remove any fragment for svn's sake creds = '' if url.lower().startswith('svn:') and '@' in url: scheme, netloc, path, p, q, f = urllib.parse.urlparse(url) if not netloc and path.startswith('//') and '/' in path[2:]: netloc, path = path[2:].split('/', 1) auth, host = _splituser(netloc) if auth: if ':' in auth: user, pw = auth.split(':', 1) creds = " --username=%s --password=%s" % (user, pw) else: creds = " --username=" + auth netloc = host parts = scheme, netloc, url, p, q, f url = urllib.parse.urlunparse(parts) self.info("Doing subversion checkout from %s to %s", url, filename) os.system("svn checkout%s -q %s %s" % (creds, url, filename)) return filename @staticmethod def _vcs_split_rev_from_url(url, pop_prefix=False): scheme, netloc, path, query, frag = urllib.parse.urlsplit(url) scheme = scheme.split('+', 1)[-1] # Some fragment identification fails path = path.split('#', 1)[0] rev = None if '@' in path: path, rev = path.rsplit('@', 1) # Also, discard fragment url = urllib.parse.urlunsplit((scheme, netloc, path, query, '')) return url, rev def _download_git(self, url, filename): filename = filename.split('#', 1)[0] url, rev = self._vcs_split_rev_from_url(url, pop_prefix=True) self.info("Doing git clone from %s to %s", url, filename) os.system("git clone --quiet %s %s" % (url, filename)) if rev is not None: self.info("Checking out %s", rev) os.system("git -C %s checkout --quiet %s" % ( filename, rev, )) return filename def _download_hg(self, url, filename): filename = filename.split('#', 1)[0] url, rev = self._vcs_split_rev_from_url(url, pop_prefix=True) self.info("Doing hg clone from %s to %s", url, filename) os.system("hg clone --quiet %s %s" % (url, filename)) if rev is not None: self.info("Updating to %s", rev) os.system("hg --cwd %s up -C -r %s -q" % ( filename, rev, )) return filename def debug(self, msg, *args): log.debug(msg, *args) def info(self, msg, *args): log.info(msg, *args) def warn(self, msg, *args): log.warn(msg, *args) # This pattern matches a character entity reference (a decimal numeric # references, a hexadecimal numeric reference, or a named reference). entity_sub = re.compile(r'&(#(\d+|x[\da-fA-F]+)|[\w.:-]+);?').sub def decode_entity(match): what = match.group(0) return unescape(what) def htmldecode(text): """ Decode HTML entities in the given text. >>> htmldecode( ... 'https://../package_name-0.1.2.tar.gz' ... '?tokena=A&tokenb=B">package_name-0.1.2.tar.gz') 'https://../package_name-0.1.2.tar.gz?tokena=A&tokenb=B">package_name-0.1.2.tar.gz' """ return entity_sub(decode_entity, text) def socket_timeout(timeout=15): def _socket_timeout(func): def _socket_timeout(*args, **kwargs): old_timeout = socket.getdefaulttimeout() socket.setdefaulttimeout(timeout) try: return func(*args, **kwargs) finally: socket.setdefaulttimeout(old_timeout) return _socket_timeout return _socket_timeout def _encode_auth(auth): """ A function compatible with Python 2.3-3.3 that will encode auth from a URL suitable for an HTTP header. >>> str(_encode_auth('username%3Apassword')) 'dXNlcm5hbWU6cGFzc3dvcmQ=' Long auth strings should not cause a newline to be inserted. >>> long_auth = 'username:' + 'password'*10 >>> chr(10) in str(_encode_auth(long_auth)) False """ auth_s = urllib.parse.unquote(auth) # convert to bytes auth_bytes = auth_s.encode() encoded_bytes = base64.b64encode(auth_bytes) # convert back to a string encoded = encoded_bytes.decode() # strip the trailing carriage return return encoded.replace('\n', '') class Credential: """ A username/password pair. Use like a namedtuple. """ def __init__(self, username, password): self.username = username self.password = password def __iter__(self): yield self.username yield self.password def __str__(self): return '%(username)s:%(password)s' % vars(self) class PyPIConfig(configparser.RawConfigParser): def __init__(self): """ Load from ~/.pypirc """ defaults = dict.fromkeys(['username', 'password', 'repository'], '') configparser.RawConfigParser.__init__(self, defaults) rc = os.path.join(os.path.expanduser('~'), '.pypirc') if os.path.exists(rc): self.read(rc) @property def creds_by_repository(self): sections_with_repositories = [ section for section in self.sections() if self.get(section, 'repository').strip() ] return dict(map(self._get_repo_cred, sections_with_repositories)) def _get_repo_cred(self, section): repo = self.get(section, 'repository').strip() return repo, Credential( self.get(section, 'username').strip(), self.get(section, 'password').strip(), ) def find_credential(self, url): """ If the URL indicated appears to be a repository defined in this config, return the credential for that repository. """ for repository, cred in self.creds_by_repository.items(): if url.startswith(repository): return cred def open_with_auth(url, opener=urllib.request.urlopen): """Open a urllib2 request, handling HTTP authentication""" parsed = urllib.parse.urlparse(url) scheme, netloc, path, params, query, frag = parsed # Double scheme does not raise on Mac OS X as revealed by a # failing test. We would expect "nonnumeric port". Refs #20. if netloc.endswith(':'): raise http_client.InvalidURL("nonnumeric port: ''") if scheme in ('http', 'https'): auth, address = _splituser(netloc) else: auth = None if not auth: cred = PyPIConfig().find_credential(url) if cred: auth = str(cred) info = cred.username, url log.info('Authenticating as %s for %s (from .pypirc)', *info) if auth: auth = "Basic " + _encode_auth(auth) parts = scheme, address, path, params, query, frag new_url = urllib.parse.urlunparse(parts) request = urllib.request.Request(new_url) request.add_header("Authorization", auth) else: request = urllib.request.Request(url) request.add_header('User-Agent', user_agent) fp = opener(request) if auth: # Put authentication info back into request URL if same host, # so that links found on the page will work s2, h2, path2, param2, query2, frag2 = urllib.parse.urlparse(fp.url) if s2 == scheme and h2 == address: parts = s2, netloc, path2, param2, query2, frag2 fp.url = urllib.parse.urlunparse(parts) return fp # copy of urllib.parse._splituser from Python 3.8 def _splituser(host): """splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.""" user, delim, host = host.rpartition('@') return (user if delim else None), host # adding a timeout to avoid freezing package_index open_with_auth = socket_timeout(_SOCKET_TIMEOUT)(open_with_auth) def fix_sf_url(url): return url # backward compatibility def local_open(url): """Read a local path, with special support for directories""" scheme, server, path, param, query, frag = urllib.parse.urlparse(url) filename = urllib.request.url2pathname(path) if os.path.isfile(filename): return urllib.request.urlopen(url) elif path.endswith('/') and os.path.isdir(filename): files = [] for f in os.listdir(filename): filepath = os.path.join(filename, f) if f == 'index.html': with open(filepath, 'r') as fp: body = fp.read() break elif os.path.isdir(filepath): f += '/' files.append('<a href="{name}">{name}</a>'.format(name=f)) else: tmpl = ( "<html><head><title>{url}" "{files}") body = tmpl.format(url=url, files='\n'.join(files)) status, message = 200, "OK" else: status, message, body = 404, "Path not found", "Not found" headers = {'content-type': 'text/html'} body_stream = six.StringIO(body) return urllib.error.HTTPError(url, status, message, headers, body_stream) site-packages/setuptools/site-patch.py000064400000004376151464452710014143 0ustar00def __boot(): import sys import os PYTHONPATH = os.environ.get('PYTHONPATH') if PYTHONPATH is None or (sys.platform == 'win32' and not PYTHONPATH): PYTHONPATH = [] else: PYTHONPATH = PYTHONPATH.split(os.pathsep) pic = getattr(sys, 'path_importer_cache', {}) stdpath = sys.path[len(PYTHONPATH):] mydir = os.path.dirname(__file__) for item in stdpath: if item == mydir or not item: continue # skip if current dir. on Windows, or my own directory importer = pic.get(item) if importer is not None: loader = importer.find_module('site') if loader is not None: # This should actually reload the current module loader.load_module('site') break else: try: import imp # Avoid import loop in Python 3 stream, path, descr = imp.find_module('site', [item]) except ImportError: continue if stream is None: continue try: # This should actually reload the current module imp.load_module('site', stream, path, descr) finally: stream.close() break else: raise ImportError("Couldn't find the real 'site' module") known_paths = dict([(makepath(item)[1], 1) for item in sys.path]) # 2.2 comp oldpos = getattr(sys, '__egginsert', 0) # save old insertion position sys.__egginsert = 0 # and reset the current one for item in PYTHONPATH: addsitedir(item) sys.__egginsert += oldpos # restore effective old position d, nd = makepath(stdpath[0]) insert_at = None new_path = [] for item in sys.path: p, np = makepath(item) if np == nd and insert_at is None: # We've hit the first 'system' path entry, so added entries go here insert_at = len(new_path) if np in known_paths or insert_at is None: new_path.append(item) else: # new path after the insert point, back-insert it new_path.insert(insert_at, item) insert_at += 1 sys.path[:] = new_path if __name__ == 'site': __boot() del __boot site-packages/setuptools/_imp.py000064400000004257151464452710013024 0ustar00""" Re-implementation of find_module and get_frozen_object from the deprecated imp module. """ import os import importlib.util import importlib.machinery from .py34compat import module_from_spec PY_SOURCE = 1 PY_COMPILED = 2 C_EXTENSION = 3 C_BUILTIN = 6 PY_FROZEN = 7 def find_module(module, paths=None): """Just like 'imp.find_module()', but with package support""" spec = importlib.util.find_spec(module, paths) if spec is None: raise ImportError("Can't find %s" % module) if not spec.has_location and hasattr(spec, 'submodule_search_locations'): spec = importlib.util.spec_from_loader('__init__.py', spec.loader) kind = -1 file = None static = isinstance(spec.loader, type) if spec.origin == 'frozen' or static and issubclass( spec.loader, importlib.machinery.FrozenImporter): kind = PY_FROZEN path = None # imp compabilty suffix = mode = '' # imp compability elif spec.origin == 'built-in' or static and issubclass( spec.loader, importlib.machinery.BuiltinImporter): kind = C_BUILTIN path = None # imp compabilty suffix = mode = '' # imp compability elif spec.has_location: path = spec.origin suffix = os.path.splitext(path)[1] mode = 'r' if suffix in importlib.machinery.SOURCE_SUFFIXES else 'rb' if suffix in importlib.machinery.SOURCE_SUFFIXES: kind = PY_SOURCE elif suffix in importlib.machinery.BYTECODE_SUFFIXES: kind = PY_COMPILED elif suffix in importlib.machinery.EXTENSION_SUFFIXES: kind = C_EXTENSION if kind in {PY_SOURCE, PY_COMPILED}: file = open(path, mode) else: path = None suffix = mode = '' return file, path, (suffix, mode, kind) def get_frozen_object(module, paths=None): spec = importlib.util.find_spec(module, paths) if not spec: raise ImportError("Can't find %s" % module) return spec.loader.get_code(module) def get_module(module, paths, info): spec = importlib.util.find_spec(module, paths) if not spec: raise ImportError("Can't find %s" % module) return module_from_spec(spec) site-packages/setuptools/__pycache__/ssl_support.cpython-38.pyc000064400000015307151464452710020761 0ustar00U Qab-! @sddlZddlZddlZddlZddlZddlmZmZmZm Z ddl m Z m Z z ddl Z Wnek rtdZ YnXdddddgZd ZzejjZejZWnek reZZYnXe dk oeeefkZzdd l mZmZWnRek r:zdd lmZdd lmZWnek r4dZdZYnXYnXesRGd ddeZesjdddZddZGdddeZGdddeZd ddZ ddZ!e!ddZ"ddZ#ddZ$dS)!N)urllib http_clientmapfilter)ResolutionErrorExtractionErrorVerifyingHTTPSHandlerfind_ca_bundle is_available cert_paths opener_fora /etc/pki/tls/certs/ca-bundle.crt /etc/ssl/certs/ca-certificates.crt /usr/share/ssl/certs/ca-bundle.crt /usr/local/share/certs/ca-root.crt /etc/ssl/cert.pem /System/Library/OpenSSL/certs/cert.pem /usr/local/share/certs/ca-root-nss.crt /etc/ssl/ca-bundle.pem )CertificateErrormatch_hostname)r )rc@s eZdZdS)r N)__name__ __module__ __qualname__rr:/usr/lib/python3.8/site-packages/setuptools/ssl_support.pyr 5sr c Csg}|s dS|d}|d}|dd}|d}||krLtdt||s`||kS|dkrt|dn>|d s|d r|t|n|t| d d |D]}|t|qt d d |dtj } | |S)zqMatching according to RFC 6125, section 6.4.3 https://tools.ietf.org/html/rfc6125#section-6.4.3 F.rrN*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)splitcountr reprlowerappend startswithreescapereplacecompilejoin IGNORECASEmatch) ZdnhostnameZ max_wildcardsZpatspartsZleftmostZ remainderZ wildcardsZfragZpatrrr_dnsname_match;s,     r&cCs|s tdg}|dd}|D]*\}}|dkr t||r@dS||q |s|ddD]6}|D],\}}|dkrdt||rdS||qdq\t|dkrtd |d tt|fn*t|dkrtd ||d fntd dS)a=Verify that *cert* (in decoded format as returned by SSLSocket.getpeercert()) matches the *hostname*. RFC 2818 and RFC 6125 rules are followed, but IP addresses are not accepted for *hostname*. CertificateError is raised on failure. On success, the function returns nothing. zempty or no certificateZsubjectAltNamerZDNSNZsubjectZ commonNamerz&hostname %r doesn't match either of %sz, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found) ValueErrorgetr&rlenr r!rr)Zcertr$ZdnsnamesZsankeyvaluesubrrrros2         rc@s eZdZdZddZddZdS)rz=Simple verifying handler: no auth, subclasses, timeouts, etc.cCs||_t|dSN) ca_bundle HTTPSHandler__init__)selfr.rrrr0szVerifyingHTTPSHandler.__init__csfdd|S)Ncst|jf|Sr-)VerifyingHTTPSConnr.)hostkwr1rrz2VerifyingHTTPSHandler.https_open..)Zdo_open)r1Zreqrr5r https_opens z VerifyingHTTPSHandler.https_openN)rrr__doc__r0r8rrrrrsc@s eZdZdZddZddZdS)r2z@Simple verifying connection: no auth, subclasses, timeouts, etc.cKstj||f|||_dSr-)HTTPSConnectionr0r.)r1r3r.r4rrrr0szVerifyingHTTPSConn.__init__cCst|j|jft|dd}t|drHt|ddrH||_||j}n|j}tt drxt j |j d}|j ||d|_nt j |t j |j d|_zt|j|Wn.tk r|jtj|jYnXdS)NZsource_address_tunnel _tunnel_hostcreate_default_context)Zcafile)Zserver_hostname)Z cert_reqsZca_certs)socketZcreate_connectionr3Zportgetattrhasattrsockr;r<sslr=r.Z wrap_socketZ CERT_REQUIREDrZ getpeercertr ZshutdownZ SHUT_RDWRclose)r1rAZ actual_hostZctxrrrconnects.   zVerifyingHTTPSConn.connectN)rrrr9r0rDrrrrr2sr2cCstjt|ptjS)z@Get a urlopen() replacement that uses ca_bundle for verification)rrequestZ build_openerrr open)r.rrrr s cstfdd}|S)Ncstds||_jS)Nalways_returns)r@rG)argskwargsfuncrrwrappers  zonce..wrapper) functoolswraps)rKrLrrJroncesrOcsZz ddl}Wntk r"YdSXGfddd|j}|d|d|jS)Nrcs,eZdZfddZfddZZS)z"get_win_certfile..CertFilecst|t|jdSr-)superr0atexitregisterrCr5CertFile __class__rrr0sz+get_win_certfile..CertFile.__init__cs,zt|Wntk r&YnXdSr-)rPrCOSErrorr5rSrrrCsz(get_win_certfile..CertFile.close)rrrr0rC __classcell__rrT)rUrrTsrTZCAZROOT) wincertstore ImportErrorrTZaddstorename)rYZ _wincertsrrXrget_win_certfiles    r\cCs$ttjjt}tp"t|dp"tS)z*Return an existing CA bundle path, or NoneN)rospathisfiler r\next_certifi_where)Zextant_cert_pathsrrrr s c Cs.ztdWStttfk r(YnXdS)NZcertifi) __import__whererZrrrrrrrasra)r)N)%r]r>rQrrMZsetuptools.extern.six.movesrrrrZ pkg_resourcesrrrBrZ__all__striprr rEr/r:AttributeErrorobjectr r rZbackports.ssl_match_hostnamer'r&rr2r rOr\r rarrrrsZ      4) (    site-packages/setuptools/__pycache__/unicode_utils.cpython-38.pyc000064400000002177151464452710021233 0ustar00U Qab@s8ddlZddlZddlmZddZddZddZdS) N)sixcCsVt|tjrtd|Sz$|d}td|}|d}Wntk rPYnX|S)NZNFDutf-8) isinstancer text_type unicodedataZ normalizedecodeencode UnicodeError)pathr s   site-packages/setuptools/__pycache__/namespaces.cpython-38.pyc000064400000007040151464452710020476 0ustar00U Qab @sRddlZddlmZddlZddlmZejjZGdddZ Gddde Z dS)N)log)mapc@sTeZdZdZddZddZddZdZd Zd d Z d d Z ddZ e ddZ dS) Installerz -nspkg.pthc Cs|}|sdStj|\}}||j7}|j|t d|t |j |}|j rdt |dSt|d}||W5QRXdS)Nz Installing %sZwt)_get_all_ns_packagesospathsplitext _get_target nspkg_extZoutputsappendrinfor_gen_nspkg_lineZdry_runlistopen writelines)selfZnspfilenameextlinesfr9/usr/lib/python3.8/site-packages/setuptools/namespaces.pyinstall_namespacess     zInstaller.install_namespacescCsHtj|\}}||j7}tj|s.dStd|t|dS)Nz Removing %s) rrrr r existsrr remove)rrrrrruninstall_namespaces!s    zInstaller.uninstall_namespacescCs|jSN)targetrrrrr )szInstaller._get_target) zimport sys, types, osz#has_mfs = sys.version_info > (3, 5)z$p = os.path.join(%(root)s, *%(pth)r)z4importlib = has_mfs and __import__('importlib.util')z-has_mfs and __import__('importlib.machinery')zm = has_mfs and sys.modules.setdefault(%(pkg)r, importlib.util.module_from_spec(importlib.machinery.PathFinder.find_spec(%(pkg)r, [os.path.dirname(p)])))zCm = m or sys.modules.setdefault(%(pkg)r, types.ModuleType(%(pkg)r))z7mp = (m or []) and m.__dict__.setdefault('__path__',[])z(p not in mp) and mp.append(p))z4m and setattr(sys.modules[%(parent)r], %(child)r, m)cCsdS)Nz$sys._getframe(1).f_locals['sitedir']rrrrr _get_rootCszInstaller._get_rootcCsVt|}t|d}|}|j}|d\}}}|rB||j7}d|tdS)N.; ) strtuplesplitr _nspkg_tmpl rpartition_nspkg_tmpl_multijoinlocals)rpkgZpthrootZ tmpl_linesparentsepZchildrrrr Fs zInstaller._gen_nspkg_linecCs |jjp g}ttt|j|S)z,Return sorted list of all package namespaces)Z distributionZnamespace_packagessortedflattenr _pkg_names)rZpkgsrrrrQs zInstaller._get_all_ns_packagesccs(|d}|r$d|V|q dS)z Given a namespace package, yield the components of that package. >>> names = Installer._pkg_names('a.b.c') >>> set(names) == set(['a', 'a.b', 'a.b.c']) True r N)r%r)pop)r+partsrrrr1Vs  zInstaller._pkg_namesN)__name__ __module__ __qualname__r rrr r&r(rr r staticmethodr1rrrrr s rc@seZdZddZddZdS)DevelopInstallercCstt|jSr)reprr#Zegg_pathrrrrrgszDevelopInstaller._get_rootcCs|jSr)Zegg_linkrrrrr jszDevelopInstaller._get_targetN)r4r5r6rr rrrrr8fsr8) rZ distutilsr itertoolsZsetuptools.extern.six.movesrchain from_iterabler0rr8rrrrs   [site-packages/setuptools/__pycache__/sandbox.cpython-38.pyc000064400000036260151464452710020023 0ustar00U Qab7@s ddlZddlZddlZddlZddlZddlZddlZddlZddlZddl Z ddl m Z ddl m Z mZddlZejdrddlmmmmZn ejejZzeZWnek rdZYnXeZddlm Z ddlm!Z!ddd d gZ"d-d d Z#ej$d.d dZ%ej$ddZ&ej$ddZ'ej$ddZ(Gddde)Z*GdddZ+ej$ddZ,ddZ-ej$ddZ.ej$dd Z/d!d"Z0d#d$Z1d%d Z2Gd&ddZ3e4ed'rej5gZ6ngZ6Gd(dde3Z7e8ej9d)d*d+:DZ;Gd,d d e Zr4r5tbrrr__exit__{s zExceptionSaver.__exit__cCs6dt|krdSttj|j\}}t|||jdS)z"restore and re-raise any exceptionrAN)varsrr.loadsrArZreraiserB)r>r4r5rrrresumes zExceptionSaver.resumeN)r7r8r9r:r?rDrGrrrrr<rs r<c #sVtjt }VW5QRXtjfddtjD}t||dS)z Context in which imported modules are saved. Translates exceptions internal to the context into the equivalent exception outside the context. c3s$|]}|kr|ds|VqdS)z encodings.N startswith).0mod_namer!rr s zsave_modules..N)rmodulescopyr<update_clear_modulesrG) saved_excZ del_modulesrr!r save_moduless   rRcCst|D] }tj|=qdSr)listrrM)Z module_namesrKrrrrPs rPc cs$t}z |VW5t|XdSr)r$ __getstate__ __setstate__r!rrrsave_pkg_resources_states rVccstj|d}txtfttNt<t|(t |t ddVW5QRXW5QRXW5QRXW5QRXW5QRXW5QRXdS)NZtempZ setuptools) r(r joinrVrRhide_setuptoolsr"rr'r, __import__) setup_dirZtemp_dirrrr setup_contexts  r[cCstd}t||S)aH >>> _needs_hiding('setuptools') True >>> _needs_hiding('pkg_resources') True >>> _needs_hiding('setuptools_plugin') False >>> _needs_hiding('setuptools.__init__') True >>> _needs_hiding('distutils') True >>> _needs_hiding('os') False >>> _needs_hiding('Cython') True z1(setuptools|pkg_resources|distutils|Cython)(\.|$))rerboolmatch)rKpatternrrr _needs_hidings r`cCstttj}t|dS)a% Remove references to setuptools' modules from sys.modules to allow the invocation to import the most appropriate setuptools. This technique is necessary to avoid issues such as #315 where setuptools upgrading itself would fail to find a function declared in the metadata. N)filterr`rrMrP)rMrrrrXs rXc Cstjtj|}t|z|gt|tjdd<tjd|t t j ddt |trl|n |t}t|t|dd}t||W5QRXWn4tk r}z|jr|jdrʂW5d}~XYnXW5QRXdS)z8Run a distutils setup script, sandboxed in its directoryNrcSs|Sr)Zactivate)Zdistrrrzrun_setup..__main__)__file__r7)r(r abspathdirnamer[rSrrinsertr__init__Z callbacksappend isinstancestrencodegetfilesystemencodingr dictr SystemExitargs)Z setup_scriptrqrZZ dunder_filensvrrrr s"    c@seZdZdZdZddZddZddZd d Zd d Z d dZ dD]Z e e e rDe e ee <qDd$ddZerzedeZedeZdD]Z e e e ree ee <qddZdD]Z e e e ree ee <qddZdD]Z e e e ree ee <qddZddZd d!Zd"d#ZdS)%rzDWrap 'os' module and 'open()' builtin for virtualizing setup scriptsFcsfddttD_dS)Ncs$g|]}|dst|r|qS)_)rIhasattr)rJnamer=rr s z,AbstractSandbox.__init__..)dir_os_attrsr=rr=rris zAbstractSandbox.__init__cCs"|jD]}tt|t||qdSr)rzsetattrr(getattr)r>sourcervrrr_copy s zAbstractSandbox._copycCs(||tr|jt_|jt_d|_dSr@)r~_filerfile_openr_activer=rrrr?s  zAbstractSandbox.__enter__cCs$d|_trtt_tt_|tdSNF)rrrrrrr~ry)r>exc_type exc_value tracebackrrrrDs zAbstractSandbox.__exit__c Cs"||W5QRSQRXdS)zRun 'func' under os sandboxingNr)r>funcrrrrunszAbstractSandbox.runcsttfdd}|S)Ncs2|jr |j||f||\}}||f||Sr)r _remap_pair)r>srcdstrqkwrvoriginalrrwrap&sz3AbstractSandbox._mk_dual_path_wrapper..wrapr|ryrvrrrr_mk_dual_path_wrapper#s z%AbstractSandbox._mk_dual_path_wrapper)renamelinksymlinkNcs p ttfdd}|S)Ncs*|jr|j|f||}|f||Sr)r _remap_inputr>r rqrrrrr4sz5AbstractSandbox._mk_single_path_wrapper..wrapr)rvrrrrr_mk_single_path_wrapper1sz'AbstractSandbox._mk_single_path_wrapperrr)statlistdirr*rchmodchownmkdirremoveunlinkrmdirutimelchownchrootlstatZ startfilemkfifomknodpathconfaccesscsttfdd}|S)NcsB|jr2|j|f||}||f||S|f||Sr)rr _remap_outputrrrrrIsz4AbstractSandbox._mk_single_with_return..wraprrrrr_mk_single_with_returnFs z&AbstractSandbox._mk_single_with_return)readlinktempnamcsttfdd}|S)Ncs ||}|jr||S|Sr)rr)r>rqrZretvalrrrrXs  z'AbstractSandbox._mk_query..wraprrrrr _mk_queryUs zAbstractSandbox._mk_query)r)tmpnamcCs|S)z=Called to remap or validate any path, whether input or outputr)r>r rrr_validate_pathdszAbstractSandbox._validate_pathcOs ||SzCalled for path inputsrr> operationr rqrrrrrhszAbstractSandbox._remap_inputcCs ||S)zCalled for path outputsr)r>rr rrrrlszAbstractSandbox._remap_outputcOs0|j|d|f|||j|d|f||fS)?Called for path pairs like rename, link, and symlink operationsz-fromz-to)rr>rrrrqrrrrrpszAbstractSandbox._remap_pair)N)r7r8r9r:rrir~r?rDrrrvruryrrrrrrrrrrrrrrrs<          devnullc@seZdZdZedddddddd d d d d dg ZdgZefddZ ddZ e rXd'ddZ d(ddZ ddZ ddZddZdd Zd!d"Zd)d$d%Zd&S)*r z.) r(r rr_sandboxrW_prefix _exceptionsrri)r>Zsandbox exceptionsrrrris zDirectorySandbox.__init__cOsddlm}||||dS)Nr)r )r1r )r>rrqrr rrr _violations zDirectorySandbox._violationrcOs:|dkr(||s(|jd||f||t||f||S)NrZrtr ZrUUr)_okrrr>r rrqrrrrrszDirectorySandbox._filecOs:|dkr(||s(|jd||f||t||f||S)Nrr)rrrrrrrrszDirectorySandbox._opencCs|ddS)Nr)rr=rrrrszDirectorySandbox.tmpnamcCsR|j}z>d|_tjtj|}||p@||jkp@||jWS||_XdSr) rr(r rr _exemptedrrIr)r>r Zactiverrrrrs  zDirectorySandbox._okcs<fdd|jD}fdd|jD}t||}t|S)Nc3s|]}|VqdSrrH)rJZ exceptionfilepathrrrLsz-DirectorySandbox._exempted..c3s|]}t|VqdSr)r\r^)rJr_rrrrLs)r_exception_patterns itertoolschainany)r>rZ start_matchesZpattern_matchesZ candidatesrrrrs   zDirectorySandbox._exemptedcOs4||jkr0||s0|j|tj|f|||Sr) write_opsrrr(r rrrrrrszDirectorySandbox._remap_inputcOs2||r||s*|j|||f||||fS)r)rrrrrrrszDirectorySandbox._remap_paircOs@|t@r*||s*|jd|||f||tj|||f||S)zCalled for low-level os.open()zos.open) WRITE_FLAGSrrryr)r>rflagsrrqrrrrrszDirectorySandbox.openN)r)r)r)r7r8r9r:rofromkeysrr _EXCEPTIONSrirrrrrrrrrrrrrr ~s:     cCsg|]}tt|dqS)rr)rJarrrrwsrwz4O_WRONLY O_RDWR O_APPEND O_CREAT O_TRUNC O_TEMPORARYc@s&eZdZdZedZddZdS)r zEA setup script attempted to modify the filesystem outside the sandboxa SandboxViolation: {cmd}{args!r} {kwargs} The package setup script has attempted to modify files on your system that are not within the EasyInstall build area, and has been aborted. This package cannot be safely installed by EasyInstall, and may not support alternate installation locations even if you run its setup script by hand. Please inform the package's author and the EasyInstall maintainers to find out if a fix or workaround is available. cCs|j\}}}|jjftSr)rqtmplformatr)r>cmdrqkwargsrrr__str__s zSandboxViolation.__str__N) r7r8r9r:textwrapdedentlstriprrrrrrr s )N)N)=r(rr&operator functoolsrr\ contextlibr.rZsetuptools.externrZsetuptools.extern.six.movesrrZpkg_resources.py31compatr$platformrIZ$org.python.modules.posix.PosixModulepythonrMposixZ PosixModuleryrvrr NameErrorrrZdistutils.errorsrr__all__rcontextmanagerrr"r'r,r0r-r<rRrPrVr[r`rXr rrurrr reduceor_splitrr rrrrsx                w  Vsite-packages/setuptools/__pycache__/version.cpython-38.opt-1.pyc000064400000000444151464452710021004 0ustar00U Qab@s6ddlZzedjZWnek r0dZYnXdS)NZ setuptoolsunknown)Z pkg_resourcesZget_distributionversion __version__ Exceptionrr6/usr/lib/python3.8/site-packages/setuptools/version.pyssite-packages/setuptools/__pycache__/dist.cpython-38.pyc000064400000122323151464452710017324 0ustar00U Qab@sdgZddlZddlZddlZddlZddlZddlZddlZddl Zddl Zddl Zddl m Z ddlmZddlmZddlZddlmZddlmZddlmZmZmZdd l mZdd lmZdd lmZdd lm Z dd lm!Z!ddl"m#Z#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/ddl0Z0e1de1dddZ2ddZ3ddZ4ddZ5e6e7fZ8dd Z9d!d"Z:d#d$Z;d%d&Zd+d,Z?d-d.Z@d/d0ZAd1d2ZBd3d4ZCd5d6ZDe-ejEjFZGGd7ddeGZFGd8d9d9ZHGd:d;d;e'ZIdS)< DistributionN) strtobool)DEBUGtranslate_longopt) defaultdict)message_from_file)DistutilsOptionErrorDistutilsPlatformErrorDistutilsSetupError) rfc822_escape) StrictVersion)six) packaging) ordered_set)mapfilter filterfalse)SetuptoolsDeprecationWarning)Require)windows_support) get_unpatched)parse_configurationz&setuptools.extern.packaging.specifiersz#setuptools.extern.packaging.versioncCstdtt|S)NzDo not call this function)warningswarnDistDeprecationWarningr)clsr3/usr/lib/python3.8/site-packages/setuptools/dist.py_get_unpatched-s r cCst|dd}|dkr|js |jr*td}nd|jdk sT|jdk sTt|dddk sT|jr^td}n0|js||js||j s||j s||j rtd}ntd}||_ |S)Nmetadata_versionz2.1python_requires1.21.1z1.0) getattrlong_description_content_typeprovides_extrasr maintainermaintainer_email project_urlsprovidesrequires obsoletes classifiers download_urlr!)selfZmvrrrget_metadata_version2s*      r1cs t|fdd}fdd}td|_|d|_|d|_|d|_|d |_d |_|d |_d |_ |d |_ |d |_ dkr|d|_ nd |_ |d|_ |d|_dkr|dd|_|d|_|d|_|jtdkr |d|_|d|_|d|_nd |_d |_d |_d S)z-Reads the metadata values from a file object.cs|}|dkrdS|S)NZUNKNOWNr)namevaluemsgrr _read_fieldLsz"read_pkg_file.._read_fieldcs|d}|gkrdS|SN)Zget_all)r2valuesr4rr _read_listRs z!read_pkg_file.._read_listzmetadata-versionr2versionZsummaryauthorNz author-emailz home-pagelicensez download-url descriptionkeywords,platformZ classifierr$r,r+r-)rr r!r2r:r=r;r( author_emailr)Zurlr<r/Zlong_descriptionsplitr>Z platformsr.r,r+r-)r0filer6r9rr4r read_pkg_fileHs:                 rDc s}tjrfdd}n fdd}|dt||d|d|d|d|td kr|d  |d  n.d }|D]$\}}t |}|d k r|||q|d j r|dj jD]}|dd|qt} |d| d} | r:|d| |td krdD]} |d| qPndddddtdr|djjr|djjr jD]} |d| qd S)z5Write the PKG-INFO format data to a file object. csd||fdSNz%s: %s )writeZ _encode_fieldkeyr3rCr0rr write_fieldsz#write_pkg_file..write_fieldcsd||fdSrE)rFrG)rCrrrJszMetadata-VersionNameVersionZSummaryz Home-pager#Author Author-email))rMr;)rNrA)Z Maintainerr()zMaintainer-emailr)NZLicensez Download-URLz Project-URLz%s, %sZ Descriptionr?ZKeywordsZPlatformZ ClassifierZRequiresZProvidesZ Obsoletesr"zRequires-PythonzDescription-Content-TypezProvides-Extra)r1rPY2strZget_nameZ get_versionZget_descriptionZget_urlr Z get_contactZget_contact_emailr%Z get_licenser/r*itemsr Zget_long_descriptionjoinZ get_keywordsZ get_platformsZ _write_listZget_classifiersZ get_requiresZ get_providesZ get_obsoleteshasattrr"r&r') r0rCr:rJZoptional_fieldsZfieldattrZattr_valZ project_urlZ long_descr>r@extrarrIrwrite_pkg_file~sZ             rVc CsPztjd|}|jrtWn,ttttfk rJtd||fYnXdS)Nzx=z4%r must be importable 'module:attrs' string (got %r)) pkg_resources EntryPointparseextrasAssertionError TypeError ValueErrorAttributeErrorr )distrTr3eprrrcheck_importablesrac CsZz(t|ttfstd||ks&tWn,ttttfk rTtd||fYnXdS)z"Verify that value is a string listz%%r must be a list of strings (got %r)N) isinstancelisttupler[rRr\r]r^r r_rTr3rrrassert_string_lists rgcCsd|}t||||D]J}||s2tdd||d\}}}|r||krtjd||qdS)z(Verify that namespace packages are validz1Distribution contains no modules or packages for znamespace package %r.z^WARNING: %r is declared as a package namespace, but %r is not: please correct this in setup.pyN)rghas_contents_forr rpartition distutilslogr)r_rTr3Z ns_packagesZnspparentsepZchildrrr check_nsps    roc Cs@zttt|Wn"tttfk r:tdYnXdS)z+Verify that extras_require mapping is validz'extras_require' must be a dictionary whose values are strings or lists of strings containing valid project/version requirement specifiers.N) rd itertoolsstarmap _check_extrarQr\r]r^r rfrrr check_extrass rscCs<|d\}}}|r*t|r*td|tt|dS)N:zInvalid environment marker: ) partitionrWZinvalid_markerr rdparse_requirements)rUZreqsr2rnmarkerrrrrrs rrcCs&t||kr"d}t|j||ddS)z)Verify that value is True, False, 0, or 1z0{attr!r} must be a boolean value (got {value!r}))rTr3N)boolr format)r_rTr3tmplrrr assert_bool s r{c Csjz(tt|t|ttfr&tdWn<ttfk rd}zd}t|j ||dW5d}~XYnXdS)z9Verify that install_requires is a valid requirements listzUnordered types are not allowedzm{attr!r} must be a string or list of strings containing valid project/version requirement specifiers; {error}rTerrorN) rdrWrvrcdictsetr\r]r ryr_rTr3r}rzrrrcheck_requirementss rc CsRztj|Wn<tjjk rL}zd}t|j||dW5d}~XYnXdS)z.Verify that value is a valid version specifierzF{attr!r} must be a string containing valid version specifiers; {error}r|N)rZ specifiersZ SpecifierSetZInvalidSpecifierr ryrrrrcheck_specifier s rc Cs@ztj|Wn*tk r:}z t|W5d}~XYnXdS)z)Verify that entry_points map is parseableN)rWrXZ parse_mapr]r )r_rTr3errrcheck_entry_points,srcCst|tjstddS)Nztest_suite must be a string)rcr string_typesr rfrrrcheck_test_suite4s rcCs\t|tstd||D]6\}}t|tjsDtd||t|d||q dS)z@Verify that value is a dictionary of package names to glob listszT{!r} must be a dictionary mapping package names to lists of string wildcard patternsz,keys of {!r} dict must be strings (got {!r})zvalues of {!r} dictN)rcr~r ryrQrrrg)r_rTr3kvrrrcheck_package_data9s  rcCs(|D]}td|stjd|qdS)Nz \w+(\.\w+)*z[WARNING: %r not a valid package name; please use only .-separated package names in setup.py)rematchrkrlr)r_rTr3Zpkgnamerrrcheck_packagesHs  rc@sReZdZdZdeejdZdZddZ dMddZ dd Z d d Z e d d ZddZddZdNddZe ddZdOddZdPddZddZddZdd Zd!d"Zd#d$Zd%d&Zd'd(Zd)d*Zd+d,Zd-d.Zd/d0Zd1d2Z d3d4Z!d5d6Z"d7d8Z#d9d:Z$d;d<Z%d=d>Z&d?d@Z'dAdBZ(dCdDZ)dEdFZ*dGdHZ+dIdJZ,dKdLZ-dS)QraDistribution with support for features, tests, and package data This is an enhanced version of 'distutils.dist.Distribution' that effectively adds the following new optional keyword arguments to 'setup()': 'install_requires' -- a string or sequence of strings specifying project versions that the distribution requires when installed, in the format used by 'pkg_resources.require()'. They will be installed automatically when the package is installed. If you wish to use packages that are not available in PyPI, or want to give your users an alternate download location, you can add a 'find_links' option to the '[easy_install]' section of your project's 'setup.cfg' file, and then setuptools will scan the listed web pages for links that satisfy the requirements. 'extras_require' -- a dictionary mapping names of optional "extras" to the additional requirement(s) that using those extras incurs. For example, this:: extras_require = dict(reST = ["docutils>=0.3", "reSTedit"]) indicates that the distribution can optionally provide an extra capability called "reST", but it can only be used if docutils and reSTedit are installed. If the user installs your package using EasyInstall and requests one of your extras, the corresponding additional requirements will be installed if needed. 'features' **deprecated** -- a dictionary mapping option names to 'setuptools.Feature' objects. Features are a portion of the distribution that can be included or excluded based on user options, inter-feature dependencies, and availability on the current system. Excluded features are omitted from all setup commands, including source and binary distributions, so you can create multiple distributions from the same source tree. Feature names should be valid Python identifiers, except that they may contain the '-' (minus) sign. Features can be included or excluded via the command line options '--with-X' and '--without-X', where 'X' is the name of the feature. Whether a feature is included by default, and whether you are allowed to control this from the command line, is determined by the Feature object. See the 'Feature' class for more information. 'test_suite' -- the name of a test suite to run for the 'test' command. If the user runs 'python setup.py test', the package will be installed, and the named test suite will be run. The format is the same as would be used on a 'unittest.py' command line. That is, it is the dotted name of an object to import and call to generate a test suite. 'package_data' -- a dictionary mapping package names to lists of filenames or globs to use to find data files contained in the named packages. If the dictionary has filenames or globs listed under '""' (the empty string), those names will be searched for in every package, in addition to any names for the specific package. Data files found using these names/globs will be installed along with the package, in the same location as the package. Note that globs are allowed to reference the contents of non-package subdirectories, as long as you use '/' as a path separator. (Globs are automatically converted to platform-specific paths at runtime.) In addition to these new keywords, this class also has several new methods for manipulating the distribution's contents. For example, the 'include()' and 'exclude()' methods can be thought of as in-place add and subtract commands that add or remove packages, modules, extensions, and so on from the distribution. They are used by the feature subsystem to configure the distribution for the included and excluded features. N)r&r*r'cCsl|rd|ksd|krdStt|d}tjj|}|dk rh|dshtt|d|_ ||_ dS)Nr2r:zPKG-INFO) rWZ safe_namerPlower working_setZby_keygetZ has_metadataZ safe_versionZ_version _patched_dist)r0attrsrHr_rrrpatch_missing_pkg_infosz#Distribution.patch_missing_pkg_infoc std}|si_|pi}d|ks,d|kr4tg_i_g_|dd_ ||dg_ |dg_ t dD]}t|jdqtfdd |DjD]L\}}jj|fD]}||kr||}qq|r|nd}tj||qtjjtjr4tjjj_jjdk rzHtjjj}t|} jj| krt d jj| f| j_Wn0tjj!t"fk rt d jjYnX#dS) N package_datafeaturesrequire_featuressrc_rootdependency_linkssetup_requiresdistutils.setup_keywordscs i|]\}}|jkr||qSr)_DISTUTILS_UNSUPPORTED_METADATA.0rrr0rr s z)Distribution.__init__..zNormalizing '%s' to '%s'zThe version specified (%r) is an invalid version, this may not work as expected with newer versions of setuptools, pip, and PyPI. Please see PEP 440 for more details.)$rSrFeaturewarn_deprecatedrrZ dist_filespoprrrrrWiter_entry_pointsvars setdefaultr2 _Distribution__init__rQrmetadata__dict__setattrrcr:numbersNumberrPrrLrrZInvalidVersionr\_finalize_requires) r0rZhave_package_datar`optiondefaultsourcer3ZverZnormalized_versionrrrrs\    zDistribution.__init__cCsft|ddr|j|j_t|ddrR|jD]$}|dd}|r,|jj|q,|| dS)z Set `metadata.python_requires` and fix environment markers in `install_requires` and `extras_require`. r"Nextras_requirertr) r%r"rrkeysrBr'add_convert_extras_requirements"_move_install_requirements_markers)r0rUrrrrs   zDistribution._finalize_requirescCsht|ddpi}tt|_|D]@\}}|j|t|D]"}||}|j|||q>q"dS)z Convert requirements in `extras_require` of the form `"extra": ["barbazquux; {marker}"]` to `"extra:{marker}": ["barbazquux"]`. rN) r%rrd_tmp_extras_requirerQrWrv _suffix_forappend)r0Z spec_ext_reqssectionrrsuffixrrrrs   z)Distribution._convert_extras_requirementscCs|jrdt|jSdS)ze For a requirement, return the 'extras_require' suffix for that requirement. rtrb)rwrPreqrrrr szDistribution._suffix_forcsdd}tddpd}tt|}t||}t||}ttt|_|D]}j dt|j  |qNt fddj D_dS) zv Move requirements in `install_requires` that are using environment markers `extras_require`. cSs|j Sr7rwrrrr is_simple_reqszFDistribution._move_install_requirements_markers..is_simple_reqinstall_requiresNrrtc3s,|]$\}}|ddtj|DfVqdS)cSsg|] }t|qSr)rPrrrrr )szMDistribution._move_install_requirements_markers...N)r _clean_reqrrrr (szBDistribution._move_install_requirements_markers..)r%rdrWrvrrrrPrrrwrr~rQr)r0rZspec_inst_reqsZ inst_reqsZ simple_reqsZ complex_reqsrrrrrs    z/Distribution._move_install_requirements_markerscCs d|_|S)zP Given a Requirement, remove environment markers and return it. Nr)r0rrrrr-szDistribution._clean_reqc Csddlm}tjr>tjtjkr>ddddddd d d d d ddg }ng}t|}|dkrZ|}t rh| d|}|D]}t j |dd4}t r| dj fttjr|jn|j|W5QRX|D]\}||}||} |D]>} | dkr| |kr|||| } | dd} || f| | <qq|qrd|jkr|jdD]\} \} } |j| } zF| r|t|| t|  n(| dkrt|| t| n t|| | Wn,tk r}z t|W5d}~XYnXqHdS)z Adapted from distutils.dist.Distribution.parse_config_files, this method provides the same functionality in subtly-improved ways. r) ConfigParserz install-basezinstall-platbasez install-libzinstall-platlibzinstall-purelibzinstall-headerszinstall-scriptsz install-dataprefixz exec-prefixhomeuserrootNz"Distribution.parse_config_files():utf-8)encodingz reading {filename}__name__-_global)verboseZdry_run)Z(setuptools.extern.six.moves.configparserrrPY3sysr base_prefix frozensetZfind_config_filesrannounceioopenrylocalsZ read_fileZreadfpZsectionsoptionsget_option_dict_try_strrreplacercommand_optionsrQ negative_optrrr]r )r0 filenamesrZignore_optionsparserfilenamereaderrrZopt_dictoptvalsrcaliasr5rrr_parse_config_files4s`           z Distribution._parse_config_filescCs.tjr |Sz |WStk r(YnX|S)ab On Python 2, much of distutils relies on string values being of type 'str' (bytes) and not unicode text. If the value can be safely encoded to bytes using the default encoding, prefer that. Why the default encoding? Because that value can be implicitly decoded back to text if needed. Ref #1653 )rrencodeUnicodeEncodeError)rrrrrrs  zDistribution._try_strc Cs^|}|dkr||}tr,|d||D]"\}\}}trZ|d|||fzdd|jD}Wntk rg}YnXz |j}Wntk ri}YnXz~t|t j } ||kr| rt |||t | nJ||kr| rt ||t |n,t ||rt |||ntd|||fWq4tk rV} z t| W5d} ~ XYq4Xq4dS)a Set the options for 'command_obj' from 'option_dict'. Basically this means copying elements of a dictionary ('option_dict') to attributes of an instance ('command'). 'command_obj' must be a Command instance. If 'option_dict' is not supplied, uses the standard option dictionary for this command (from 'self.command_options'). (Adopted from distutils.dist.Distribution._set_command_options) Nz# setting options for '%s' command:z %s = %s (from %s)cSsg|] }t|qSrr)rorrrrsz5Distribution._set_command_options..z1error in %s: command '%s' has no such option '%s')Zget_command_namerrrrQZboolean_optionsr^rrcrrrrrSr r]) r0Z command_objZ option_dictZ command_namerrr3Z bool_optsneg_optZ is_stringr5rrr_set_command_optionssF           z!Distribution._set_command_optionsFcCs(|j|dt||j|d|dS)zYParses configuration files from various levels and loads configuration. )r)ignore_option_errorsN)rrrr)r0rrrrrparse_config_filess  zDistribution.parse_config_filescCst|}|jr||S)z3Process features after parsing command line options)rparse_command_liner_finalize_features)r0resultrrrrs zDistribution.parse_command_linecCsd|ddS)z;Convert feature name to corresponding option attribute nameZwith_rrrr0r2rrr_feature_attrnameszDistribution._feature_attrnamecCs8tjjt||jdd}|D]}tjj|ddq|S)zResolve pre-setup requirementsT) installerZreplace_conflictingr)rWrresolvervfetch_build_eggr)r0r,Zresolved_distsr_rrrfetch_build_eggsszDistribution.fetch_build_eggscCst||jr|tdD]:}t||jd}|dk r"|j|j d| ||j|q"t|ddr~dd|j D|_ ng|_ dS)Nrrconvert_2to3_doctestscSsg|]}tj|qSr)ospathabspathrprrrrsz1Distribution.finalize_options..) rfinalize_optionsr_set_global_opts_from_featuresrWrr%r2requirerloadr)r0r`r3rrrrs   zDistribution.finalize_optionsc Csvtjtjd}tj|srt|t|tj|d}t|d$}| d| d| dW5QRX|S)Nz.eggsz README.txtwzcThis directory contains eggs that were downloaded by setuptools to build, test, and run plug-ins. zAThis directory caches those eggs to prevent repeated downloads. z/However, it is safe to delete this directory. ) rrrRcurdirexistsmkdirrZ hide_filerrF)r0Z egg_cache_dirZreadme_txt_filenamefrrrget_egg_cache_dirs      zDistribution.get_egg_cache_dirc Csddlm}|ddgi}|d}||dd|dD|jr|jdd}d|krx|dd |}d |f|d<|}||d g|d d dd d d d d d }| ||S)z Fetch an egg needed for buildingr) easy_installZ script_argsr css"|]\}}|dkr||fVqdS)) find_links site_dirsZ index_urloptimizer Z allow_hostsNrrrrrrsz/Distribution.fetch_build_egg..Nr rZsetupxTF) args install_dirZexclude_scriptsZ always_copyZbuild_directoryZeditableZupgradeZ multi_versionZ no_reportr) Zsetuptools.command.easy_installr  __class__rclearupdaterQrrZensure_finalized)r0rr r_optsZlinksrcmdrrrrs8     zDistribution.fetch_build_eggc Csg}|j}|jD]\}}||d|||jr|j}d}d}|s\||}}d|dd||fd|dd||ff}| |d||d|<q||j |_ |_ ||_|_ dS)z;Add --with-X/--without-X options based on optional featuresNz (default)rbzwith-zinclude zwithout-zexclude ) rcopyrrQ _set_featurevalidateoptionalr=include_by_defaultextendglobal_optionsZfeature_optionsZfeature_negopt) r0ZgoZnor2featuredescrZincdefZexcdefnewrrrrs$     z+Distribution._set_global_opts_from_featurescCs|jD]<\}}||}|s0|dkr |r ||||dq |jD](\}}||sR||||dqRdS)z9Add/remove features and resolve dependencies between themNrr)rrQfeature_is_includedr include_inr exclude_from)r0r2rZenabledrrrr0s    zDistribution._finalize_featurescCs\||jkr|j|Std|}|D]*}|j|jd||j|<}|St||S)z(Pluggable version of get_command_class()distutils.commandsrN)cmdclassrWrrrrrget_command_class)r0commandZepsr`r#rrrr$As   zDistribution.get_command_classcCs:tdD]$}|j|jkr |}||j|j<q t|SNr")rWrr2r#rrprint_commandsr0r`r#rrrr'Ns  zDistribution.print_commandscCs:tdD]$}|j|jkr |}||j|j<q t|Sr&)rWrr2r#rrget_command_listr(rrrr)Vs  zDistribution.get_command_listcCst||||dS)zSet feature's inclusion statusN)rr)r0r2Zstatusrrrr^szDistribution._set_featurecCst|||S)zAReturn 1 if feature is included, 0 if excluded, 'None' if unknown)r%rrrrrrbsz Distribution.feature_is_includedcCsF||dkr&|j|j}t|d|j||||ddS)z)Request inclusion of feature named 'name'rz2 is required, but was excluded or is not availablerN)rrr=r r r)r0r2rrrrinclude_featurefs zDistribution.include_featurecKs@|D]2\}}t|d|d}|r.||q|||qdS)aAdd items to distribution that are named in keyword arguments For example, 'dist.include(py_modules=["x"])' would add 'x' to the distribution's 'py_modules' attribute, if it was not already there. Currently, this method only supports inclusion for attributes that are lists or tuples. If you need to add support for adding to other attributes in this or a subclass, you can add an '_include_X' method, where 'X' is the name of the attribute. The method will be called with the value passed to 'include()'. So, 'dist.include(foo={"bar":"baz"})' will try to call 'dist._include_foo({"bar":"baz"})', which can then handle whatever special inclusion logic is needed. Z _include_N)rQr% _include_misc)r0rrrincluderrrr,qs  zDistribution.includecsfd|jr&fdd|jD|_|jrDfdd|jD|_|jrbfdd|jD|_dS)z9Remove packages, modules, and extensions in named packagerhcs"g|]}|kr|s|qSr startswithrpackagepfxrrrs z0Distribution.exclude_package..cs"g|]}|kr|s|qSrr-rr/rrrs cs&g|]}|jkr|js|qSr)r2r.rr/rrrs N)packages py_modules ext_modules)r0r0rr/rexclude_packages   zDistribution.exclude_packagecCs2|d}|D]}||ks&||rdSqdS)z.rcsequencer r%r^r)r0r2r3oldrr;r _exclude_miscs    zDistribution._exclude_misccst|tstd||fzt||Wn tk rHtd|YnXdkr`t|||n:ttsxt|dn"fdd|D}t|||dS)zAHandle 'include()' for list/tuple attrs without a special handlerz%s: setting must be a list (%r)r7Nr8csg|]}|kr|qSrrr9r>rrrsz.Distribution._include_misc..r<)r0r2r3rrr@rr+s$    zDistribution._include_misccKs@|D]2\}}t|d|d}|r.||q|||qdS)aRemove items from distribution that are named in keyword arguments For example, 'dist.exclude(py_modules=["x"])' would remove 'x' from the distribution's 'py_modules' attribute. Excluding packages uses the 'exclude_package()' method, so all of the package's contained packages, modules, and extensions are also excluded. Currently, this method only supports exclusion from attributes that are lists or tuples. If you need to add support for excluding from other attributes in this or a subclass, you can add an '_exclude_X' method, where 'X' is the name of the attribute. The method will be called with the value passed to 'exclude()'. So, 'dist.exclude(foo={"bar":"baz"})' will try to call 'dist._exclude_foo({"bar":"baz"})', which can then handle whatever special exclusion logic is needed. Z _exclude_N)rQr%r?)r0rrrexcluderrrrAs  zDistribution.excludecCs,t|tstd|ftt|j|dS)Nz.packages: setting must be a list or tuple (%r))rcr=r rdrr5)r0r2rrr_exclude_packagess  zDistribution._exclude_packagesc Cs|jj|_|jj|_|d}|d}||krf||\}}||=ddl}||d|dd<|d}q&t|||}||} t | ddrd|f||d<|dk rgS|S)NraliasesTrZcommand_consumes_arguments command liner) rrrrshlexrBr_parse_command_optsr$r%) r0rrr%rCrrrEnargsZ cmd_classrrrrFs"       z Distribution._parse_command_optsc Csi}|jD]\}}|D]\}\}}|dkr4q|dd}|dkr||}|j}|t|di|D]\} } | |krv| }d}qqvtdn |dkrd}|| |i|<qq|S) ahReturn a '{cmd: {opt:val}}' map of all command-line options Option names are all long, but do not include the leading '--', and contain dashes rather than underscores. If the option doesn't take an argument (e.g. '--quiet'), the 'val' is 'None'. Note that options provided by config files are intentionally excluded. rDrrrrNzShouldn't be able to get herer) rrQrZget_command_objrrrr%r[r) r0drrrrrZcmdobjrnegposrrrget_cmdline_optionss(     z Distribution.get_cmdline_optionsccsv|jpdD] }|Vq |jpdD] }|Vq |jp4dD]:}t|trN|\}}n|j}|drj|dd}|Vq6dS)z@Yield all packages, modules, and extension names in distributionrmoduleNi)r2r3r4rcrer2endswith)r0ZpkgrLZextr2Z buildinforrrr60s    z$Distribution.iter_distribution_namesc Csddl}tjs|jr t||St|jtj s:t||S|jj dkrVt||S|jj }|jj }|j dkrtdpvd}|jj}t |jd||||_zt||WSt |j|||||_XdS)zIf there were any non-global "display-only" options (--help-commands or the metadata display options) on the command line, display the requested info and return true; else return false. rN)rutf8Zwin32 r)rrrOZ help_commandsrhandle_display_optionsrcstdoutr TextIOWrapperrrerrorsr@line_bufferingdetach)r0Z option_orderrrrSnewlinerTrrrrPBs6    z#Distribution.handle_display_options)N)N)N)NF).r __module__ __qualname____doc__r~rZ OrderedSetrrrrrr staticmethodrrrrrrrrrrrrrrrr$r'r)rrr*r,r5rir?r+rArBrFrKr6rPrrrrrTsXD ;  >  /     (c@sFeZdZdZeddZdddZd d Zd d Zd dZ ddZ dS)ra **deprecated** -- The `Feature` facility was never completely implemented or supported, `has reported issues `_ and will be removed in a future version. A subset of the distribution that can be excluded if unneeded/wanted Features are created using these keyword arguments: 'description' -- a short, human readable description of the feature, to be used in error messages, and option help messages. 'standard' -- if true, the feature is included by default if it is available on the current system. Otherwise, the feature is only included if requested via a command line '--with-X' option, or if another included feature requires it. The default setting is 'False'. 'available' -- if true, the feature is available for installation on the current system. The default setting is 'True'. 'optional' -- if true, the feature's inclusion can be controlled from the command line, using the '--with-X' or '--without-X' options. If false, the feature's inclusion status is determined automatically, based on 'availabile', 'standard', and whether any other feature requires it. The default setting is 'True'. 'require_features' -- a string or sequence of strings naming features that should also be included if this feature is included. Defaults to empty list. May also contain 'Require' objects that should be added/removed from the distribution. 'remove' -- a string or list of strings naming packages to be removed from the distribution if this feature is *not* included. If the feature *is* included, this argument is ignored. This argument exists to support removing features that "crosscut" a distribution, such as defining a 'tests' feature that removes all the 'tests' subpackages provided by other features. The default for this argument is an empty list. (Note: the named package(s) or modules must exist in the base distribution when the 'setup()' function is initially called.) other keywords -- any other keyword arguments are saved, and passed to the distribution's 'include()' and 'exclude()' methods when the feature is included or excluded, respectively. So, for example, you could pass 'packages=["a","b"]' to cause packages 'a' and 'b' to be added or removed from the distribution as appropriate. A feature must include at least one 'requires', 'remove', or other keyword argument. Otherwise, it can't affect the distribution in any way. Note also that you can subclass 'Feature' to create your own specialized feature types that modify the distribution in other ways when included or excluded. See the docstrings for the various methods here for more detail. Aside from the methods, the only feature attributes that distributions look at are 'description' and 'optional'. cCsd}tj|tdddS)NzrFeatures are deprecated and will be removed in a future version. See https://github.com/pypa/setuptools/issues/65.) stacklevel)rrrr4rrrrszFeature.warn_deprecatedFTrc Ks|||_||_||_||_t|ttfr4|f}dd|D|_dd|D}|r^||d<t|trn|f}||_ ||_ |s|s|st ddS)NcSsg|]}t|tr|qSrrcrPrrrrrs z$Feature.__init__..cSsg|]}t|ts|qSrr]rrrrrs rzgFeature %s: must define 'require_features', 'remove', or at least one of 'packages', 'py_modules', etc.) rr=standard availablerrcrPrrremoverZr ) r0r=r^r_rrr`rZZerrrrrs*  zFeature.__init__cCs |jo |jS)z+Should this feature be included by default?)r_r^rrrrrszFeature.include_by_defaultcCs<|jst|jd|jf|j|jD]}||q(dS)aEnsure feature and its requirements are included in distribution You may override this in a subclass to perform additional operations on the distribution. Note that this method may be called more than once per feature, and so should be idempotent. z3 is required, but is not available on this platformN)r_r r=r,rZrr*)r0r_rrrrr s  zFeature.include_incCs.|jf|j|jr*|jD]}||qdS)a2Ensure feature is excluded from distribution You may override this in a subclass to perform additional operations on the distribution. This method will be called at most once per feature, and only after all included features have been asked to include themselves. N)rArZr`r5r0r_r:rrrr!s  zFeature.exclude_fromcCs.|jD]"}||std|j||fqdS)aVerify that feature makes sense in context of distribution This method is called by the distribution just before it parses its command line. It checks to ensure that the 'remove' attribute, if any, contains only valid package/module names that are present in the base distribution when 'setup()' is called. You may override it in a subclass to perform any other required validation of the feature against a target distribution. zg%s wants to be able to remove %s, but the distribution doesn't contain any packages or modules under %sN)r`rir r=rarrrrs   zFeature.validateN)FTTrr) rrWrXrYrZrrrr r!rrrrrres8  rc@seZdZdZdS)rzrClass for warning about deprecations in dist in setuptools. Not ignored by default, unlike DeprecationWarning.N)rrWrXrYrrrrrsr)J__all__rrrrrrZ distutils.logrkZdistutils.coreZ distutils.cmdZdistutils.distZdistutils.utilrZdistutils.debugrZdistutils.fancy_getoptrrp collectionsrZemailrZdistutils.errorsr r r r Zdistutils.versionr Zsetuptools.externrrrZsetuptools.extern.six.movesrrrrbrZsetuptools.dependsrZ setuptoolsrZsetuptools.monkeyrZsetuptools.configrrW __import__r r1rDrVrerdr=rargrorsrrr{rrrrrrZcorerrrrrrrrsv               6L    site-packages/setuptools/__pycache__/sandbox.cpython-38.opt-1.pyc000064400000036260151464452710020762 0ustar00U Qab7@s ddlZddlZddlZddlZddlZddlZddlZddlZddlZddl Z ddl m Z ddl m Z mZddlZejdrddlmmmmZn ejejZzeZWnek rdZYnXeZddlm Z ddlm!Z!ddd d gZ"d-d d Z#ej$d.d dZ%ej$ddZ&ej$ddZ'ej$ddZ(Gddde)Z*GdddZ+ej$ddZ,ddZ-ej$ddZ.ej$dd Z/d!d"Z0d#d$Z1d%d Z2Gd&ddZ3e4ed'rej5gZ6ngZ6Gd(dde3Z7e8ej9d)d*d+:DZ;Gd,d d e Zr4r5tbrrr__exit__{s zExceptionSaver.__exit__cCs6dt|krdSttj|j\}}t|||jdS)z"restore and re-raise any exceptionrAN)varsrr.loadsrArZreraiserB)r>r4r5rrrresumes zExceptionSaver.resumeN)r7r8r9r:r?rDrGrrrrr<rs r<c #sVtjt }VW5QRXtjfddtjD}t||dS)z Context in which imported modules are saved. Translates exceptions internal to the context into the equivalent exception outside the context. c3s$|]}|kr|ds|VqdS)z encodings.N startswith).0mod_namer!rr s zsave_modules..N)rmodulescopyr<update_clear_modulesrG) saved_excZ del_modulesrr!r save_moduless   rRcCst|D] }tj|=qdSr)listrrM)Z module_namesrKrrrrPs rPc cs$t}z |VW5t|XdSr)r$ __getstate__ __setstate__r!rrrsave_pkg_resources_states rVccstj|d}txtfttNt<t|(t |t ddVW5QRXW5QRXW5QRXW5QRXW5QRXW5QRXdS)NZtempZ setuptools) r(r joinrVrRhide_setuptoolsr"rr'r, __import__) setup_dirZtemp_dirrrr setup_contexts  r[cCstd}t||S)aH >>> _needs_hiding('setuptools') True >>> _needs_hiding('pkg_resources') True >>> _needs_hiding('setuptools_plugin') False >>> _needs_hiding('setuptools.__init__') True >>> _needs_hiding('distutils') True >>> _needs_hiding('os') False >>> _needs_hiding('Cython') True z1(setuptools|pkg_resources|distutils|Cython)(\.|$))rerboolmatch)rKpatternrrr _needs_hidings r`cCstttj}t|dS)a% Remove references to setuptools' modules from sys.modules to allow the invocation to import the most appropriate setuptools. This technique is necessary to avoid issues such as #315 where setuptools upgrading itself would fail to find a function declared in the metadata. N)filterr`rrMrP)rMrrrrXs rXc Cstjtj|}t|z|gt|tjdd<tjd|t t j ddt |trl|n |t}t|t|dd}t||W5QRXWn4tk r}z|jr|jdrʂW5d}~XYnXW5QRXdS)z8Run a distutils setup script, sandboxed in its directoryNrcSs|Sr)Zactivate)Zdistrrrzrun_setup..__main__)__file__r7)r(r abspathdirnamer[rSrrinsertr__init__Z callbacksappend isinstancestrencodegetfilesystemencodingr dictr SystemExitargs)Z setup_scriptrqrZZ dunder_filensvrrrr s"    c@seZdZdZdZddZddZddZd d Zd d Z d dZ dD]Z e e e rDe e ee <qDd$ddZerzedeZedeZdD]Z e e e ree ee <qddZdD]Z e e e ree ee <qddZdD]Z e e e ree ee <qddZddZd d!Zd"d#ZdS)%rzDWrap 'os' module and 'open()' builtin for virtualizing setup scriptsFcsfddttD_dS)Ncs$g|]}|dst|r|qS)_)rIhasattr)rJnamer=rr s z,AbstractSandbox.__init__..)dir_os_attrsr=rr=rris zAbstractSandbox.__init__cCs"|jD]}tt|t||qdSr)rzsetattrr(getattr)r>sourcervrrr_copy s zAbstractSandbox._copycCs(||tr|jt_|jt_d|_dSr@)r~_filerfile_openr_activer=rrrr?s  zAbstractSandbox.__enter__cCs$d|_trtt_tt_|tdSNF)rrrrrrr~ry)r>exc_type exc_value tracebackrrrrDs zAbstractSandbox.__exit__c Cs"||W5QRSQRXdS)zRun 'func' under os sandboxingNr)r>funcrrrrunszAbstractSandbox.runcsttfdd}|S)Ncs2|jr |j||f||\}}||f||Sr)r _remap_pair)r>srcdstrqkwrvoriginalrrwrap&sz3AbstractSandbox._mk_dual_path_wrapper..wrapr|ryrvrrrr_mk_dual_path_wrapper#s z%AbstractSandbox._mk_dual_path_wrapper)renamelinksymlinkNcs p ttfdd}|S)Ncs*|jr|j|f||}|f||Sr)r _remap_inputr>r rqrrrrr4sz5AbstractSandbox._mk_single_path_wrapper..wrapr)rvrrrrr_mk_single_path_wrapper1sz'AbstractSandbox._mk_single_path_wrapperrr)statlistdirr*rchmodchownmkdirremoveunlinkrmdirutimelchownchrootlstatZ startfilemkfifomknodpathconfaccesscsttfdd}|S)NcsB|jr2|j|f||}||f||S|f||Sr)rr _remap_outputrrrrrIsz4AbstractSandbox._mk_single_with_return..wraprrrrr_mk_single_with_returnFs z&AbstractSandbox._mk_single_with_return)readlinktempnamcsttfdd}|S)Ncs ||}|jr||S|Sr)rr)r>rqrZretvalrrrrXs  z'AbstractSandbox._mk_query..wraprrrrr _mk_queryUs zAbstractSandbox._mk_query)r)tmpnamcCs|S)z=Called to remap or validate any path, whether input or outputr)r>r rrr_validate_pathdszAbstractSandbox._validate_pathcOs ||SzCalled for path inputsrr> operationr rqrrrrrhszAbstractSandbox._remap_inputcCs ||S)zCalled for path outputsr)r>rr rrrrlszAbstractSandbox._remap_outputcOs0|j|d|f|||j|d|f||fS)?Called for path pairs like rename, link, and symlink operationsz-fromz-to)rr>rrrrqrrrrrpszAbstractSandbox._remap_pair)N)r7r8r9r:rrir~r?rDrrrvruryrrrrrrrrrrrrrrrs<          devnullc@seZdZdZedddddddd d d d d dg ZdgZefddZ ddZ e rXd'ddZ d(ddZ ddZ ddZddZdd Zd!d"Zd)d$d%Zd&S)*r z.) r(r rr_sandboxrW_prefix _exceptionsrri)r>Zsandbox exceptionsrrrris zDirectorySandbox.__init__cOsddlm}||||dS)Nr)r )r1r )r>rrqrr rrr _violations zDirectorySandbox._violationrcOs:|dkr(||s(|jd||f||t||f||S)NrZrtr ZrUUr)_okrrr>r rrqrrrrrszDirectorySandbox._filecOs:|dkr(||s(|jd||f||t||f||S)Nrr)rrrrrrrrszDirectorySandbox._opencCs|ddS)Nr)rr=rrrrszDirectorySandbox.tmpnamcCsR|j}z>d|_tjtj|}||p@||jkp@||jWS||_XdSr) rr(r rr _exemptedrrIr)r>r Zactiverrrrrs  zDirectorySandbox._okcs<fdd|jD}fdd|jD}t||}t|S)Nc3s|]}|VqdSrrH)rJZ exceptionfilepathrrrLsz-DirectorySandbox._exempted..c3s|]}t|VqdSr)r\r^)rJr_rrrrLs)r_exception_patterns itertoolschainany)r>rZ start_matchesZpattern_matchesZ candidatesrrrrs   zDirectorySandbox._exemptedcOs4||jkr0||s0|j|tj|f|||Sr) write_opsrrr(r rrrrrrszDirectorySandbox._remap_inputcOs2||r||s*|j|||f||||fS)r)rrrrrrrszDirectorySandbox._remap_paircOs@|t@r*||s*|jd|||f||tj|||f||S)zCalled for low-level os.open()zos.open) WRITE_FLAGSrrryr)r>rflagsrrqrrrrrszDirectorySandbox.openN)r)r)r)r7r8r9r:rofromkeysrr _EXCEPTIONSrirrrrrrrrrrrrrr ~s:     cCsg|]}tt|dqS)rr)rJarrrrwsrwz4O_WRONLY O_RDWR O_APPEND O_CREAT O_TRUNC O_TEMPORARYc@s&eZdZdZedZddZdS)r zEA setup script attempted to modify the filesystem outside the sandboxa SandboxViolation: {cmd}{args!r} {kwargs} The package setup script has attempted to modify files on your system that are not within the EasyInstall build area, and has been aborted. This package cannot be safely installed by EasyInstall, and may not support alternate installation locations even if you run its setup script by hand. Please inform the package's author and the EasyInstall maintainers to find out if a fix or workaround is available. cCs|j\}}}|jjftSr)rqtmplformatr)r>cmdrqkwargsrrr__str__s zSandboxViolation.__str__N) r7r8r9r:textwrapdedentlstriprrrrrrr s )N)N)=r(rr&operator functoolsrr\ contextlibr.rZsetuptools.externrZsetuptools.extern.six.movesrrZpkg_resources.py31compatr$platformrIZ$org.python.modules.posix.PosixModulepythonrMposixZ PosixModuleryrvrr NameErrorrrZdistutils.errorsrr__all__rcontextmanagerrr"r'r,r0r-r<rRrPrVr[r`rXr rrurrr reduceor_splitrr rrrrsx                w  Vsite-packages/setuptools/__pycache__/launch.cpython-38.pyc000064400000001470151464452710017632 0ustar00U Qab@s.dZddlZddlZddZedkr*edS)z[ Launch the Python script on the command line after setuptools is bootstrapped via import. NcCsrttjd}t|ddd}tjddtjdd<ttdt}||}|dd}t ||d}t ||dS) zP Run the script in sys.argv[1] as if it had been invoked naturally. __main__N)__file____name____doc__openz\r\nz\nexec) __builtins__sysargvdictgetattrtokenizerreadreplacecompiler)Z script_name namespaceZopen_ZscriptZ norm_scriptcoder5/usr/lib/python3.8/site-packages/setuptools/launch.pyrun s     rr)rrr rrrrrrs site-packages/setuptools/__pycache__/depends.cpython-38.pyc000064400000012140151464452710017776 0ustar00U Qab@sddlZddlZddlZddlmZddlmZddlmZm Z m Z m Z ddl mZddd d gZ Gd ddZd d Zddd Zddd ZddZedS)N) StrictVersion)Bytecode) find_module PY_COMPILED PY_FROZEN PY_SOURCE) py27compatRequirerget_module_constantextract_constantc@sHeZdZdZdddZddZdd Zdd d Zdd dZdddZ dS)r z7A prerequisite to building or installing a distributionNcCsF|dkr|dk rt}|dk r0||}|dkr0d}|jt|`dS)N __version__)r__dict__updatelocalsself)rnamerequested_versionmoduleZhomepage attributeformatr6/usr/lib/python3.8/site-packages/setuptools/depends.py__init__szRequire.__init__cCs |jdk rd|j|jfS|jS)z0Return full package/distribution name, w/versionNz%s-%s)rr)rrrr full_name#s zRequire.full_namecCs*|jdkp(|jdkp(t|dko(||jkS)z%Is 'version' sufficiently up-to-date?Nunknown)rrstrr)rversionrrr version_ok)szRequire.version_okrcCs|jdkrFz$t|j|\}}}|r*||WStk rDYdSXt|j|j||}|dk r|||k r||jdk r|||S|S)aGet version number of installed module, 'None', or 'default' Search 'paths' for module. If not found, return 'None'. If found, return the extracted version attribute, or 'default' if no version attribute was specified, or the value cannot be determined without importing the module. The version is formatted according to the requirement's version format (if any), unless it is 'None' or the supplied 'default'. N)rrrclose ImportErrorr r)rpathsdefaultfpivrrr get_version.s  zRequire.get_versioncCs||dk S)z/Return true if dependency is present on 'paths'N)r()rr"rrr is_presentIszRequire.is_presentcCs ||}|dkrdS||S)z>Return true if dependency is present and up-to-date on 'paths'NF)r(r)rr"rrrr is_currentMs zRequire.is_current)r NN)Nr)N)N) __name__ __module__ __qualname____doc__rrrr(r)r*rrrrr s   cCs"tjdd}|s|St|S)Ncss dVdS)NrrrrremptyVszmaybe_close..empty) contextlibcontextmanagerclosing)r$r/rrr maybe_closeUs  r3c Cszt||\}}\}}}} Wntk r4YdSXt|z|tkr^|dt|} nV|tkrtt ||} n@|t krt ||d} n&t ||| } t | |dW5QRSW5QRXt| ||S)zFind 'module' by searching 'paths', and extract 'symbol' Return 'None' if 'module' does not exist on 'paths', or it does not define 'symbol'. If the module defines 'symbol' as a constant, return the constant. Otherwise, return 'default'.Nexec)rr!r3rreadmarshalloadrr get_frozen_objectrcompileZ get_modulegetattrr ) rsymbolr#r"r$pathsuffixmodeZkindinfocodeZimportedrrrr `s   "c Cs||jkrdSt|j|}d}d}d}|}t|D]H}|j} |j} | |krZ|j| }q6| |krz| |ksr| |krz|S|}q6dS)aExtract the constant value of 'symbol' from 'code' If the name 'symbol' is bound to a constant value by the Python code object 'code', return that value. If 'symbol' is bound to an expression, return 'default'. Otherwise, return 'None'. Return value is based on the first assignment to 'symbol'. 'symbol' must be a global, or at least a non-"fast" local in the code block. That is, only 'STORE_NAME' and 'STORE_GLOBAL' opcodes are checked, and 'symbol' must be present in 'code.co_names'. NZad)co_nameslistindexrZopcodearg co_consts) rBr=r#Zname_idxZ STORE_NAMEZ STORE_GLOBALZ LOAD_CONSTconstZ byte_codeoprIrrrr }s   cCs>tjdstjdkrdSd}|D]}t|=t|q"dS)z Patch the globals to remove the objects not available on some platforms. XXX it'd be better to test assertions about bytecode instead. javaZcliN)r r )sysplatform startswithglobals__all__remove)Z incompatiblerrrr_update_globalss rT)r4N)r4)rNr8r0Zdistutils.versionrZ py33compatrr rrrrr rRr r3r r rTrrrrs"   D  $site-packages/setuptools/__pycache__/py34compat.cpython-38.opt-1.pyc000064400000000702151464452710021317 0ustar00U Qab@sXddlZz ddlZWnek r(YnXz ejjZWnek rRddZYnXdS)NcCs|j|jS)N)loader load_modulename)specr9/usr/lib/python3.8/site-packages/setuptools/py34compat.pymodule_from_spec sr) importlibimportlib.util ImportErrorutilrAttributeErrorrrrrs  site-packages/setuptools/__pycache__/package_index.cpython-38.opt-1.pyc000064400000100312151464452710022074 0ustar00U Qab@sdZddlZddlZddlZddlZddlZddlZddlZddlZddl Z ddl m Z ddl m Z ddlmZmZmZmZddlZddlmZmZmZmZmZmZmZmZmZmZmZm Z m!Z!ddlm"Z"ddl#m$Z$dd l%m&Z&dd l'm(Z(dd l)m*Z*dd l+m,Z,dd l-m.Z.e/Z0e1dZ2e1dej3Z4e1dZ5e1dej3j6Z7d8Z9ddddgZ:dZ;dZedZ?ddZ@ddZAddZBdGd dZCdHd!d"ZDdId#d$ZEdedfd%dZFdJd&d'ZGd(d)ZHe1d*ej3ZIeHd+d,ZJGd-d.d.ZKGd/d0d0eKZLGd1ddeZMe1d2jNZOd3d4ZPd5d6ZQdKd7d8ZRd9d:ZSGd;d<d<ZTGd=d>d>ejUZVejWjXfd?d@ZYdAdBZZeRe;eYZYdCdDZ[dEdFZ\dS)Lz#PyPI and direct package downloadingNwraps)six)urllib http_client configparsermap) CHECKOUT_DIST Distribution BINARY_DISTnormalize_path SOURCE_DIST Environmentfind_distributions safe_name safe_version to_filename Requirement DEVELOP_DISTEGG_DIST) ssl_support)log)DistutilsError) translate)get_all_headers)unescape)Wheelz^egg=([-A-Za-z0-9_.+!]+)$zhref\s*=\s*['"]?([^'"> ]+)z([^<]+)\n\s+\(md5\)z([-+.a-z0-9]{2,}):z.tar.gz .tar.bz2 .tar .zip .tgz PackageIndexdistros_for_urlparse_bdist_wininstinterpret_distro_namezrP)filenamerGr&r&r'distros_for_filenames  rTc cs||d}|s,tdd|ddDr,dStdt|dD]8}t||d|d|d||d|||dVq>dS)zGenerate alternative interpretations of a source distro name Note: if `location` is a filesystem filename, you should call ``pkg_resources.normalize_path()`` on it before passing it to this routine! rJcss|]}td|VqdS)z py\d\.\d$N)rerE).0pr&r&r' sz(interpret_distro_name..Nr6) py_versionrBrQ)r9anyrangerOr join)rKrPrGrZrBrQr;rWr&r&r'r s ccsft}|j}|dkr:tj|j|D]}|||Vq$n(|D]"}||}||kr>|||Vq>dS)zHList unique elements, preserving order. Remember all elements ever seen.N)setaddrZmoves filterfalse __contains__)iterablekeyseenZseen_addZelementkr&r&r'unique_everseens rfcstfdd}|S)zs Wrap a function returning an iterable such that the resulting iterable only ever yields unique items. cst||SN)rf)argskwargsfuncr&r'wrapperszunique_values..wrapperr)rkrlr&rjr' unique_valuessrmz(<([^>]*\srel\s*=\s*['"]?([^'">]+)[^>]*)>c cst|D]d}|\}}tttj|d}d|ksDd|kr t |D]}t j |t |dVqNq dD]@}||}|dkrtt ||}|rtt j |t |dVqtdS)zEFind rel="homepage" and rel="download" links in `page`, yielding URLs,Zhomepager3r6)z Home PagezDownload URLr2N)RELfinditergroupsr^rstrstripr+r9HREFrr#urljoin htmldecoderFfindsearch)r:pagerEtagZrelZrelsposr&r&r'find_external_linkss   r|c@s(eZdZdZddZddZddZdS) ContentCheckerzP A null content checker that defines the interface for checking content cCsdS)z3 Feed a block of data to the hash. Nr&selfblockr&r&r'feedszContentChecker.feedcCsdS)zC Check the hash. Return False if validation fails. Tr&rr&r&r'is_validszContentChecker.is_validcCsdS)zu Call reporter with information about the checker (hash name) substituted into the template. Nr&)rreportertemplater&r&r'reportszContentChecker.reportN)__name__ __module__ __qualname____doc__rrrr&r&r&r'r}sr}c@sBeZdZedZddZeddZddZ dd Z d d Z d S) HashCheckerzK(?Psha1|sha224|sha384|sha256|sha512|md5)=(?P[a-f0-9]+)cCs||_t||_||_dSrg) hash_namehashlibnewhashexpected)rrrr&r&r'__init__s zHashChecker.__init__cCs>tj|d}|stS|j|}|s0tS|f|S)z5Construct a (possibly null) ContentChecker from a URLr2)rr#r7r}patternrx groupdict)clsr:r@rEr&r&r'from_urls zHashChecker.from_urlcCs|j|dSrg)rupdater~r&r&r'rszHashChecker.feedcCs|j|jkSrg)rZ hexdigestrrr&r&r'r"szHashChecker.is_validcCs||j}||Srg)r)rrrmsgr&r&r'r%s zHashChecker.reportN) rrrrUcompilerr classmethodrrrrr&r&r&r'r s rcs<eZdZdZdJddZdKd d ZdLd d ZdMd dZddZddZ ddZ ddZ dNddZ ddZ dOfdd ZddZdd Zd!d"Zd#d$Zd%d&ZdPd'd(ZdQd)d*Zd+d,Zd-Zd.d/Zd0d1ZdRd2d3Zd4d5Zd6d7Zd8d9Zd:d;Zdd?Z!d@dAZ"dBdCZ#dDdEZ$dFdGZ%dHdIZ&Z'S)Trz;A distribution index that scans web pages for download URLshttps://pypi.org/simple/*NTcOstj|f|||dd|d |_i|_i|_i|_td t t |j |_ g|_|ortjor|prt}|rt||_n tjj|_dS)Nr1|)rrr, index_url scanned_urls fetched_urls package_pagesrUrr]rrrEallowsto_scanrZ is_availableZfind_ca_bundleZ opener_foropenerrrequesturlopen)rrZhostsZ ca_bundleZ verify_sslrhkwZuse_sslr&r&r'r-s zPackageIndex.__init__Fc Cs||jkr|sdSd|j|<t|s2||dStt|}|r\||sPdS|d||sn|rn||jkrtt|j |dS||sd|j|<dS| d|d|j|<d}| |||}|dkrdSd|j|j <d|j ddkr|dS|j }|}t|tsNt|tjjr0d }n|j d p@d }||d }|t|D](} tj|t| d } || q`| |j!rt"|d ddkr|#||}dS)zexistswarnisdirrealpathlistdirrr]rTrrrr_)rfnnestedr>itemrr&r&r'rus    zPackageIndex.process_filenamecCsbt|}|o|ddk}|s8|tj|drentryr&r&r'rXs   z.PackageIndex.scan_egg_links..)filterrRr>rr itertoolsstarmap scan_egg_link)rZ search_pathdirsZ egg_linksr&r&r'scan_egg_linkss zPackageIndex.scan_egg_linksc Csttj||}ttdttj|}W5QRXt |dkrDdS|\}}t tj||D](}tjj|f||_ t |_ ||q^dS)NrY)openrRr>r]rrrrrrsrOrrKr rBr_)rr>rZ raw_lineslinesZegg_pathZ setup_pathrHr&r&r'rs  zPackageIndex.scan_egg_linkc sfdd}t|D]:}z |tj|t|dWqtk rNYqXq||\}}|rt||D]H}t |\}} | dr| s|r|d||f7}n | |qlt dd|SdSd S) z#Process the contents of a PyPI pagecs|jrtttjj|tjdd}t|dkrd|dkrt |d}t |d}dj | i|<t|t|fSdS)Nr1rYr5r6rT)NN)r-rrrrr#r8rOr9rrr setdefaultr+r)rr;pkgverrr&r'scans   z(PackageIndex.process_index..scanr6.pyz #egg=%s-%scSsd|dddS)Nz%sr6rY)rF)mr&r&r'z,PackageIndex.process_index..rN)rtrprr#rurvrFr$r|rAr,need_version_infoscan_urlPYPI_MD5sub) rr:ryrrErrnew_urlr/fragr&rr'rs(      zPackageIndex.process_indexcCs|d|dS)NzPPage at %s links to .py file(s) without version info; an index scan is required.)scan_allrr:r&r&r'rszPackageIndex.need_version_infocGs:|j|jkr*|r |j|f||d||jdS)Nz6Scanning index of all packages (this may take a while))rrrrrrrrhr&r&r'rs zPackageIndex.scan_allcCsz||j|jd|j|js:||j|jd|j|jsR||t|j|jdD]}||qfdS)Nr1r&) rr unsafe_namerrrcrLnot_found_in_indexr)r requirementr:r&r&r' find_packagess zPackageIndex.find_packagescsR|||||jD]"}||kr0|S|d||qtt|||S)Nz%s does not match %s)prescanrrcrsuperrobtain)rrZ installerrH __class__r&r'rs zPackageIndex.obtaincCsL||jd||sH|t|td|jjtj |fdS)z- checker is a ContentChecker zValidating %%s checksum for %sz7%s validation failed for %s; possible download problem?N) rrrrrRunlinkrrr.r>rP)rcheckerrStfpr&r&r' check_hashs zPackageIndex.check_hashcCsN|D]D}|jdks0t|r0|ds0tt|r<||q|j|qdS)z;Add `urls` to the list that will be prescanned for searchesNfile:)rrr-rrrappend)rZurlsr:r&r&r'add_find_linkss  zPackageIndex.add_find_linkscCs"|jrtt|j|jd|_dS)z7Scan urls scheduled for prescanning (e.g. --find-links)N)rrrrrr&r&r'rszPackageIndex.prescancCs<||jr|jd}}n |jd}}|||j|dS)Nz#Couldn't retrieve index page for %rz3Couldn't find index page for %r (maybe misspelled?))rcrrrr)rrZmethrr&r&r'r#s  zPackageIndex.not_found_in_indexcCs~t|tsjt|}|rR||d||}t|\}}|drN||||}|Stj |rb|St |}t | ||ddS)aLocate and/or download `spec` to `tmpdir`, returning a local path `spec` may be a ``Requirement`` object, or a string containing a URL, an existing local filename, or a project/version requirement spec (i.e. the string form of a ``Requirement`` object). If it is the URL of a .py file with an unambiguous ``#egg=name-version`` tag (i.e., one that escapes ``-`` as ``_`` throughout), a trivial ``setup.py`` is automatically created alongside the downloaded file. If `spec` is a ``Requirement`` object or a string containing a project/version requirement spec, this method returns the location of a matching distribution (possibly after downloading it to `tmpdir`). If `spec` is a locally existing file or directory name, it is simply returned unchanged. If `spec` is a URL, it is downloaded to a subpath of `tmpdir`, and the local filename is returned. Various errors may be raised if a problem occurs during downloading. r6rrKN)rrr _download_urlrFrAr, gen_setuprRr>rr(rfetch_distribution)rr%tmpdirr<foundr/r@r&r&r'r3-s    zPackageIndex.downloadc sd|id}d fdd }|rH|||}|s^|dk r^|||}|dkrjdk rx||}|dkr|s|||}|dkrdrdpd|nd||j|jd SdS) a|Obtain a distribution suitable for fulfilling `requirement` `requirement` must be a ``pkg_resources.Requirement`` instance. If necessary, or if the `force_scan` flag is set, the requirement is searched for in the (online) package index as well as the locally installed packages. If a distribution matching `requirement` is found, the returned distribution's ``location`` is the value you would have gotten from calling the ``download()`` method with the matching distribution's URL or filename. If no matching distribution is found, ``None`` is returned. If the `source` flag is set, only source distributions and source checkout links will be considered. Unless the `develop_ok` flag is set, development and system eggs (i.e., those using the ``.egg-info`` format) will be ignored. zSearching for %sNcs|dkr }||jD]v}|jtkrFsF|krd|d|<q||ko\|jtkp\ }|r|j}||_tj |jr|SqdS)Nz&Skipping development or system egg: %sr6) rcrBrrr r3rKdownload_locationrRr>r)ZreqenvrHZtestZloc develop_okrZskippedsourcerr&r'rwgs&z-PackageIndex.fetch_distribution..findz:No local packages or working download links found for %s%sza source distribution of rzBest match: %s)rK)N)rrrrrZcloner) rrr force_scanrrZ local_indexrHrwr&rr'rOs2         zPackageIndex.fetch_distributioncCs"|||||}|dk r|jSdS)a3Obtain a file suitable for fulfilling `requirement` DEPRECATED; use the ``fetch_distribution()`` method now instead. For backward compatibility, this routine is identical but returns the ``location`` of the downloaded distribution instead of a distribution object. N)rrK)rrrrrrHr&r&r'fetchszPackageIndex.fetchc Cst|}|r*ddt||ddDp,g}t|dkrtj|}tj||krtj ||}ddl m }|||st |||}ttj |dd2} | d|dj|djtj|dfW5QRX|S|rtd ||fntd dS) NcSsg|]}|jr|qSr&)rM)rVdr&r&r' sz*PackageIndex.gen_setup..r6r)samefilezsetup.pywzIfrom setuptools import setup setup(name=%r, version=%r, py_modules=[%r]) zCan't unambiguously interpret project/version identifier %r; any dashes in the name or version should be escaped using underscores. %rzpCan't process plain .py files without an '#egg=name-version' suffix to enable automatic setup script generation.)rDrEr rFrOrRr>rPdirnamer]Zsetuptools.command.easy_installr shutilZcopy2rwriterLrMsplitextr) rrSr@rrErrPZdstr rr&r&r'rsB       zPackageIndex.gen_setupi c Cs|d|d}zt|}||}t|tjjrJt d||j |j f|}d}|j }d}d|krt |d} ttt| }||||||t|dV} ||} | r|| | | |d7}||||||qqq|||| W5QRX|WS|r|XdS) NzDownloading %szCan't download %s: %s %srr2zcontent-lengthzContent-Lengthwbr6)rrrrrrrrrrrr dl_blocksizermaxrint reporthookrrrr r) rr:rSfprrblocknumZbssizeZsizesrrr&r&r' _download_tos:        zPackageIndex._download_tocCsdSrgr&)rr:rSrZblksizerr&r&r'rszPackageIndex.reporthookc Cs|drt|Szt||jWSttjfk r}z.z%s %szDownload error for %s: %sz;%s returned a bad status line. The server might be down, %s)r- local_openopen_with_authrr$r InvalidURLr]rhrrrrrZURLErrorreasonZ BadStatusLinelineZ HTTPExceptionsocket)rr:Zwarningvrr&r&r'rs> "zPackageIndex.open_urlcCst|\}}|r0d|kr4|dddd}qnd}|drJ|dd}tj||}|dksj|d rv|||S|d ks|d r|||S|d r| ||S|d krt j t j |dS||d|||SdS)Nz...\_Z__downloaded__rIr*Zsvnzsvn+Zgitzgit+zhg+rrYT)rAreplacer,rRr>r]r- _download_svn _download_git _download_hgrr url2pathnamer#r7r_attempt_download)rr<r:rr.r@rSr&r&r'rs$        zPackageIndex._download_urlcCs||ddS)NT)rrr&r&r'r:szPackageIndex.scan_urlcCs6|||}d|ddkr.||||S|SdS)Nrrr)rrr+_download_html)rr:rSrr&r&r'r)=s zPackageIndex._attempt_downloadcCsnt|}|D]>}|r td|rF|t||||SqLq |t|td|dS)Nz ([^- ]+ - )?Revision \d+:zUnexpected HTML page found at ) r���rs���rU���rx���r���rR���r���r%��r���)r���r:���r���rS���r���r��r&���r&���r'���r*��D��s����   zPackageIndex._download_htmlc�����������������C���s��t�dt�|ddd�}d}|�drd|krtj|\}}}}}} |s|drd |d d��kr|d d��d d\}}t |\} } | rd | kr| d d\} } d | | f�}nd | �}| }|||||| f}tj |}|� d||�t d|||f��|S�)Nz"SVN download support is deprecatedr5���r6���r���r���zsvn:@z//r1���rY���:z --username=%s --password=%sz --username=z'Doing subversion checkout from %s to %szsvn checkout%s -q %s %s)warningsr��� UserWarningr9���r+���r-���r���r#���r7��� _splituser urlunparser���rR���system)r���r:���rS���Zcredsr<���netlocr>���rW���qr���authhostuserZpwr;���r&���r&���r'���r%��S��s&����   zPackageIndex._download_svnc�����������������C���sp���t�j|�\}}}}}|ddd�}|ddd�}d�}d|krR|dd\}}t�j||||df}�|�|fS�)N+r6���r2���r5���r���r+��r���)r���r#���Zurlsplitr9���rsplitZ urlunsplit)r:��� pop_prefixr<���r2��r>���r?���r���revr&���r&���r'���_vcs_split_rev_from_urli��s����z$PackageIndex._vcs_split_rev_from_urlc�����������������C���sl���|�ddd�}|�j|dd\}}|�d||�td||f��|d�k rh|�d|�td ||f��|S�) Nr5���r6���r���Tr9��zDoing git clone from %s to %szgit clone --quiet %s %szChecking out %szgit -C %s checkout --quiet %sr9���r;��r���rR���r1��r���r:���rS���r:��r&���r&���r'���r&��{��s���� zPackageIndex._download_gitc�����������������C���sl���|�ddd�}|�j|dd\}}|�d||�td||f��|d�k rh|�d|�td ||f��|S�) Nr5���r6���r���Tr<��zDoing hg clone from %s to %szhg clone --quiet %s %szUpdating to %szhg --cwd %s up -C -r %s -qr=��r>��r&���r&���r'���r'����s���� zPackageIndex._download_hgc�����������������G���s���t�j|f|��d�S�rg���)r���r���r���r&���r&���r'���r�����s����zPackageIndex.debugc�����������������G���s���t�j|f|��d�S�rg���)r���r���r���r&���r&���r'���r�����s����zPackageIndex.infoc�����������������G���s���t�j|f|��d�S�rg���)r���r���r���r&���r&���r'���r�����s����zPackageIndex.warn)r���r���NT)F)F)F)N)N)FFFN)FF)N)F)(r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r3���r���r��r���r��r��r��r���r���r���r)��r*��r%�� staticmethodr;��r&��r'��r���r���r��� __classcell__r&���r&���r���r'���r���*��sX���������  3   +   #������ L )$ # z!&(#(\d+|x[\da-fA-F]+)|[\w.:-]+);?c�����������������C���s���|��d}t|S�)Nr���)rF���r���)rE���Zwhatr&���r&���r'��� decode_entity��s���� rA��c�����������������C���s ���t�t|�S�)a�� Decode HTML entities in the given text. >>> htmldecode( ... 'https://../package_name-0.1.2.tar.gz' ... '?tokena=A&tokenb=B">package_name-0.1.2.tar.gz') 'https://../package_name-0.1.2.tar.gz?tokena=A&tokenb=B">package_name-0.1.2.tar.gz' ) entity_subrA��)textr&���r&���r'���rv�����s���� rv���c��������������������s����fdd}|S�)Nc��������������������s����fdd}|S�)Nc��������������� ������s2���t��}t��z�|�|W�S�t�|�X�d�S�rg���)r��ZgetdefaulttimeoutZsetdefaulttimeout)rh���ri���Z old_timeout)rk���timeoutr&���r'���_socket_timeout��s ���� z@socket_timeout.<locals>._socket_timeout.<locals>._socket_timeoutr&���)rk���rE��rD��rj���r'���rE����s����z'socket_timeout.<locals>._socket_timeoutr&���)rD��rE��r&���rF��r'���socket_timeout��s���� rG��c�����������������C���s2���t�j|�}|�}t|}|�}|ddS�)aq�� A function compatible with Python 2.3-3.3 that will encode auth from a URL suitable for an HTTP header. >>> str(_encode_auth('username%3Apassword')) 'dXNlcm5hbWU6cGFzc3dvcmQ=' Long auth strings should not cause a newline to be inserted. >>> long_auth = 'username:' + 'password'*10 >>> chr(10) in str(_encode_auth(long_auth)) False  r���)r���r#���r8���encodebase64Z b64encoder���r$��)r4��Zauth_sZ auth_bytesZ encoded_bytesZencodedr&���r&���r'��� _encode_auth��s ����  rK��c�������������������@���s(���e�Zd�ZdZdd�Zdd�Zdd�ZdS�) Credentialz: A username/password pair. Use like a namedtuple. c�����������������C���s���||�_�||�_d�S�rg���usernamepassword)r���rN��rO��r&���r&���r'���r�����s����zCredential.__init__c�����������������c���s���|�j�V��|�jV��d�S�rg���rM��r���r&���r&���r'���__iter__��s����zCredential.__iter__c�����������������C���s ���dt�|��S�)Nz%(username)s:%(password)s)varsr���r&���r&���r'���__str__��s����zCredential.__str__N)r���r���r���r���r���rP��rR��r&���r&���r&���r'���rL����s���rL��c�������������������@���s0���e�Zd�Zdd�Zedd�Zdd�Zdd�Zd S�) PyPIConfigc�����������������C���sP���t�dddgd}tj|�|�tjtjdd}tj |rL|� |�dS�)z% Load from ~/.pypirc rN��rO�� repositoryr���~z.pypircN) dictfromkeysr���RawConfigParserr���rR���r>���r]��� expanduserr���r���)r���defaultsZrcr&���r&���r'���r�����s ���� zPyPIConfig.__init__c��������������������s&����fdd���D�}tt�j|S�)Nc��������������������s ���g�|�]}��|d��r|qS�)rT��)r���rs���)rV���sectionr���r&���r'���r����s���z2PyPIConfig.creds_by_repository.<locals>.<listcomp>)ZsectionsrV��r���_get_repo_cred)r���Zsections_with_repositoriesr&���r���r'���creds_by_repository��s���� zPyPIConfig.creds_by_repositoryc�����������������C���s6���|��|d�}|t|��|d�|��|d�fS�)NrT��rN��rO��)r���rs���rL��)r���r[��Zrepor&���r&���r'���r\����s ����zPyPIConfig._get_repo_credc�����������������C���s*���|�j��D�]\}}||r |��S�q dS�)z If the URL indicated appears to be a repository defined in this config, return the credential for that repository. N)r]��itemsr-���)r���r:���rT��credr&���r&���r'���find_credential ��s���� zPyPIConfig.find_credentialN)r���r���r���r���propertyr]��r\��r`��r&���r&���r&���r'���rS����s ���  rS��c�����������������C���s:��t�j|�}|\}}}}}}|dr0td|dkrFt|\} } nd} | s~t�|�} | r~t | } | j |�f} t j d | ��| rdt | �} || ||||f} t�j| }t�j|}|d| �n t�j|�}|dt�||}| r6t�j|j\}}}}}}||kr6|| kr6||||||f} t�j| |_|S�) z4Open a urllib2 request, handling HTTP authenticationr,��znonnumeric port: '')ZhttpZhttpsN*Authenticating as %s for %s (from .pypirc)zBasic Z Authorizationz User-Agent)rb��)r���r#���r7���r,���r���r��r/��rS��r`��rr���rN��r���r���rK��r0��r���ZRequestZ add_header user_agentr:���)r:���r���Zparsedr<���r2��r>���Zparamsr?���r���r4��Zaddressr_��r���r;���r���r���r��s2Zh2Zpath2Zparam2Zquery2Zfrag2r&���r&���r'���r����s8����          r��c�����������������C���s ���|��d\}}}�|r|nd|�fS�)zJsplituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.r+��N) rpartition)r5��r6��Zdelimr&���r&���r'���r/��F��s����r/��c�����������������C���s���|�S�rg���r&���)r:���r&���r&���r'��� fix_sf_urlP��s����rf��c�������������� ���C���s��t�j|�\}}}}}}t�j|}tj|r<t�j|�S�| drtj |rg�}t |D�]d} tj || } | dkrt | d} | �} W�5�Q�R�X��qntj | r| d7�} |dj| d�q`d} | j|�d |d} d \}}n d \}}} d d i}t| }t�j|�||||S�) z7Read a local path, with special support for directoriesr1���z index.htmlrz<a href="{name}">{name}</a>)r.���zB<html><head><title>{url}{files}rH)r:files)ZOK)rzPath not foundz Not foundrz text/html)rr#r7rr(rRr>isfilerr,rrr]rrrformatrStringIOrr)r:r<r=r>Zparamr?rrSrhrfilepathrZbodyrZstatusmessagerZ body_streamr&r&r'rTs.        r)N)N)N)N)r!)]rsysrRrUr rrJrrr- functoolsrZsetuptools.externrZsetuptools.extern.six.movesrrrrr"Z pkg_resourcesr r r r r rrrrrrrrrZ distutilsrZdistutils.errorsrZfnmatchrZsetuptools.py27compatrZsetuptools.py33compatrZsetuptools.wheelrtypeZ __metaclass__rrDIrtrrErr9rN__all__Z_SOCKET_TIMEOUTZ_tmplrk version_inforcr(rrArrCrTr rfrmror|r}rrrrBrArvrGrKrLrXrSrrrr/rfrr&r&r&r's  <           !  $   !  &/ site-packages/setuptools/__pycache__/unicode_utils.cpython-38.opt-1.pyc000064400000002177151464452710022172 0ustar00U Qab@s8ddlZddlZddlmZddZddZddZdS) N)sixcCsVt|tjrtd|Sz$|d}td|}|d}Wntk rPYnX|S)NZNFDutf-8) isinstancer text_type unicodedataZ normalizedecodeencode UnicodeError)pathr s   site-packages/setuptools/__pycache__/py27compat.cpython-38.pyc000064400000003325151464452710020366 0ustar00U Qab@sdZddlZddlZddlmZddZejr6ddZedkoFejZerPe ndd Z z,d d l m Z m Z mZmZd d l mZmZWnJek rddlZdd lm Z mZmZdddZ ddZddZYnXdS)z2 Compatibility Support for Python 2.7 and earlier N)sixcCs ||S)zH Given an HTTPMessage, return all headers matching a given key. )Zget_allmessagekeyr9/usr/lib/python3.8/site-packages/setuptools/py27compat.pyget_all_headers srcCs ||SN)Z getheadersrrrrrsZLinuxcCs|Sr r)xrrrr ) find_module PY_COMPILED PY_FROZEN PY_SOURCE)get_frozen_object get_module)rrrc Csj|d}|rf|d}t||\}}\}}}} |tjkrP|pFdg}|g}q |r td||fq | S)z7Just like 'imp.find_module()', but with package support.r__init__zCan't find %r in %s)splitpopimprZ PKG_DIRECTORY ImportError) modulepathspartspartfpathsuffixmodeZkindinforrrr's    rcCs t|Sr )rr)rrrrrr7srcCstj|f|tj|Sr )r load_modulesysmodules)rrr"rrrr:sr)N)__doc__r$platformZsetuptools.externrrZPY2systemZlinux_py2_asciistrZ rmtree_safe_imprrrrrrrrrrrrs&   site-packages/setuptools/__pycache__/py31compat.cpython-38.pyc000064400000002247151464452710020363 0ustar00U QabF@sPgZeZzddlmZWn2ek rJddlZddlZGdddZYnXdS))TemporaryDirectoryNc@s(eZdZdZddZddZddZdS) rz Very simple temporary directory context manager. Will try to delete afterward, but will also ignore OS and similar errors on deletion. cKsd|_tjf||_dSN)nametempfileZmkdtemp)selfkwargsr9/usr/lib/python3.8/site-packages/setuptools/py31compat.py__init__szTemporaryDirectory.__init__cCs|jSr)r)rrrr __enter__szTemporaryDirectory.__enter__cCs2zt|jdWntk r&YnXd|_dS)NT)shutilZrmtreerOSError)rexctypeZexcvalueZexctracerrr __exit__s zTemporaryDirectory.__exit__N)__name__ __module__ __qualname____doc__r r rrrrr r sr)__all__typeZ __metaclass__rr ImportErrorr rrrr ssite-packages/setuptools/__pycache__/glibc.cpython-38.opt-1.pyc000064400000002772151464452710020405 0ustar00U QabJ @sHddlmZddlZddlZddlZddZddZddZd d ZdS) )absolute_importNcCsRtd}z |j}Wntk r*YdSXtj|_|}t|tsN|d}|S)z9Returns glibc version string, or None if not using glibc.Nascii) ctypesZCDLLgnu_get_libc_versionAttributeErrorZc_char_pZrestype isinstancestrdecode)Zprocess_namespacer version_strr 4/usr/lib/python3.8/site-packages/setuptools/glibc.pyglibc_version_string s    r cCsHtd|}|s$td|tdSt|d|koFt|d|kS)Nz$(?P[0-9]+)\.(?P[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFmajorminor)rematchwarningswarnRuntimeWarningintgroup)r required_major minimum_minormr r r check_glibc_version$s rcCst}|dkrdSt|||S)NF)r r)rrr r r r have_compatible_glibc4srcCst}|dkrdSd|fSdS)zTry to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings in case the lookup fails. N)rZglibc)r )Z glibc_versionr r r libc_verLsr) Z __future__rrrrr rrrr r r r s site-packages/setuptools/__pycache__/extension.cpython-38.opt-1.pyc000064400000003651151464452710021336 0ustar00U Qab@s|ddlZddlZddlZddlZddlZddlmZddlm Z ddZ e Z e ej j ZGdddeZ Gd d d e ZdS) N)map) get_unpatchedcCs4d}zt|dgdjWdStk r.YnXdS)z0 Return True if Cython can be imported. zCython.Distutils.build_ext build_ext)fromlistTF) __import__r Exception)Z cython_implr 8/usr/lib/python3.8/site-packages/setuptools/extension.py _have_cython sr c@s eZdZdZddZddZdS) Extensionz7Extension that uses '.c' files in place of '.pyx' filescOs(|dd|_tj|||f||dS)Npy_limited_apiF)popr _Extension__init__)selfnamesourcesargskwr r r r#szExtension.__init__cCsNtr dS|jpd}|dkr$dnd}ttjd|}tt||j |_ dS)z Replace sources with .pyx extensions to sources with the target language extension. This mechanism allows language authors to supply pre-converted sources but to prefer the .pyx sources. Nzc++z.cppz.cz.pyx$) r Zlanguagelower functoolspartialresublistrr)rZlangZ target_extrr r r _convert_pyx_sources_to_lang)s  z&Extension._convert_pyx_sources_to_langN)__name__ __module__ __qualname____doc__rrr r r r r sr c@seZdZdZdS)Libraryz=Just like a regular Extension, but built as a library insteadN)rrr r!r r r r r"8sr")rrZdistutils.coreZ distutilsZdistutils.errorsZdistutils.extensionZsetuptools.extern.six.movesrZmonkeyrr Z have_pyrexZcorer rr"r r r r s   site-packages/setuptools/__pycache__/version.cpython-38.pyc000064400000000444151464452710020045 0ustar00U Qab@s6ddlZzedjZWnek r0dZYnXdS)NZ setuptoolsunknown)Z pkg_resourcesZget_distributionversion __version__ Exceptionrr6/usr/lib/python3.8/site-packages/setuptools/version.pyssite-packages/setuptools/__pycache__/archive_util.cpython-38.opt-1.pyc000064400000012010151464452710021765 0ustar00U Qab@sdZddlZddlZddlZddlZddlZddlZddlmZddl m Z ddddd d d gZ Gd d d eZ d dZ e dfddZe fdd Ze fddZe fddZeeefZdS)z/Utilities for extracting common archive formatsN)DistutilsError)ensure_directoryunpack_archiveunpack_zipfileunpack_tarfiledefault_filterUnrecognizedFormatextraction_driversunpack_directoryc@seZdZdZdS)rz#Couldn't recognize the archive typeN)__name__ __module__ __qualname____doc__rr;/usr/lib/python3.8/site-packages/setuptools/archive_util.pyrscCs|S)z@The default progress/filter callback; returns True for all filesr)srcdstrrrrsc CsN|ptD]4}z||||Wntk r4YqYqXdSqtd|dS)aUnpack `filename` to `extract_dir`, or raise ``UnrecognizedFormat`` `progress_filter` is a function taking two arguments: a source path internal to the archive ('/'-separated), and a filesystem path where it will be extracted. The callback must return the desired extract path (which may be the same as the one passed in), or else ``None`` to skip that file or directory. The callback can thus be used to report on the progress of the extraction, as well as to filter the items extracted or alter their extraction paths. `drivers`, if supplied, must be a non-empty sequence of functions with the same signature as this function (minus the `drivers` argument), that raise ``UnrecognizedFormat`` if they do not support extracting the designated archive type. The `drivers` are tried in sequence until one is found that does not raise an error, or until all are exhausted (in which case ``UnrecognizedFormat`` is raised). If you do not supply a sequence of drivers, the module's ``extraction_drivers`` constant will be used, which means that ``unpack_zipfile`` and ``unpack_tarfile`` will be tried, in that order. Nz!Not a recognized archive type: %s)r r)filename extract_dirprogress_filterZdriversZdriverrrrrs  c Cstj|std||d|fi}t|D]\}}}||\}}|D],} || dtj|| f|tj|| <qH|D]T} tj|| } ||| | } | sqzt| tj|| } t| | t | | qzq.dS)z"Unpack" a directory, using the same interface as for archives Raises ``UnrecognizedFormat`` if `filename` is not a directory z%s is not a directory/N) ospathisdirrwalkjoinrshutilZcopyfileZcopystat) rrrpathsbasedirsfilesrrdftargetrrrr ?s$   * c Cst|std|ft|}|D]}|j}|ds,d|dkrPq,tj j |f|d}|||}|sxq,| drt |n4t || |j}t|d}||W5QRX|jd?} | r,t|| q,W5QRXdS)zUnpack zip `filename` to `extract_dir` Raises ``UnrecognizedFormat`` if `filename` is not a zipfile (as determined by ``zipfile.is_zipfile()``). See ``unpack_archive()`` for an explanation of the `progress_filter` argument. z%s is not a zip filer..wbN)zipfileZ is_zipfilerZZipFileZinfolistr startswithsplitrrrendswithrreadopenwriteZ external_attrchmod) rrrzinfonamer$datar#Zunix_attributesrrrrZs(         c Csfzt|}Wn$tjk r2td|fYnXt|dd|_|D]}|j}|dsPd| dkrPt j j |f| d}|dk r| s|r|j}|rt|j}t ||}t|}||}q|dk rP|s|rP|||} | rP| t jr"| dd} z||| WqPtjk rJYqPXqPW5QRdSQRXdS) zUnpack tar/tar.gz/tar.bz2 `filename` to `extract_dir` Raises ``UnrecognizedFormat`` if `filename` is not a tarfile (as determined by ``tarfile.open()``). See ``unpack_archive()`` for an explanation of the `progress_filter` argument. z/%s is not a compressed or uncompressed tar filecWsdS)Nr)argsrrrz unpack_tarfile..rr%NT)tarfiler-ZTarErrorr contextlibclosingchownr2r)r*rrrZislnkZissymZlinkname posixpathdirnamenormpathZ _getmemberisfilerr+sepZ_extract_memberZ ExtractError) rrrZtarobjmemberr2Z prelim_dstZlinkpathrZ final_dstrrrrs:         )rr(r8rrr<r9Zdistutils.errorsrZ pkg_resourcesr__all__rrrr rrr rrrrs2   #  % .site-packages/setuptools/__pycache__/namespaces.cpython-38.opt-1.pyc000064400000007040151464452710021435 0ustar00U Qab @sRddlZddlmZddlZddlmZejjZGdddZ Gddde Z dS)N)log)mapc@sTeZdZdZddZddZddZdZd Zd d Z d d Z ddZ e ddZ dS) Installerz -nspkg.pthc Cs|}|sdStj|\}}||j7}|j|t d|t |j |}|j rdt |dSt|d}||W5QRXdS)Nz Installing %sZwt)_get_all_ns_packagesospathsplitext _get_target nspkg_extZoutputsappendrinfor_gen_nspkg_lineZdry_runlistopen writelines)selfZnspfilenameextlinesfr9/usr/lib/python3.8/site-packages/setuptools/namespaces.pyinstall_namespacess     zInstaller.install_namespacescCsHtj|\}}||j7}tj|s.dStd|t|dS)Nz Removing %s) rrrr r existsrr remove)rrrrrruninstall_namespaces!s    zInstaller.uninstall_namespacescCs|jSN)targetrrrrr )szInstaller._get_target) zimport sys, types, osz#has_mfs = sys.version_info > (3, 5)z$p = os.path.join(%(root)s, *%(pth)r)z4importlib = has_mfs and __import__('importlib.util')z-has_mfs and __import__('importlib.machinery')zm = has_mfs and sys.modules.setdefault(%(pkg)r, importlib.util.module_from_spec(importlib.machinery.PathFinder.find_spec(%(pkg)r, [os.path.dirname(p)])))zCm = m or sys.modules.setdefault(%(pkg)r, types.ModuleType(%(pkg)r))z7mp = (m or []) and m.__dict__.setdefault('__path__',[])z(p not in mp) and mp.append(p))z4m and setattr(sys.modules[%(parent)r], %(child)r, m)cCsdS)Nz$sys._getframe(1).f_locals['sitedir']rrrrr _get_rootCszInstaller._get_rootcCsVt|}t|d}|}|j}|d\}}}|rB||j7}d|tdS)N.; ) strtuplesplitr _nspkg_tmpl rpartition_nspkg_tmpl_multijoinlocals)rpkgZpthrootZ tmpl_linesparentsepZchildrrrr Fs zInstaller._gen_nspkg_linecCs |jjp g}ttt|j|S)z,Return sorted list of all package namespaces)Z distributionZnamespace_packagessortedflattenr _pkg_names)rZpkgsrrrrQs zInstaller._get_all_ns_packagesccs(|d}|r$d|V|q dS)z Given a namespace package, yield the components of that package. >>> names = Installer._pkg_names('a.b.c') >>> set(names) == set(['a', 'a.b', 'a.b.c']) True r N)r%r)pop)r+partsrrrr1Vs  zInstaller._pkg_namesN)__name__ __module__ __qualname__r rrr r&r(rr r staticmethodr1rrrrr s rc@seZdZddZddZdS)DevelopInstallercCstt|jSr)reprr#Zegg_pathrrrrrgszDevelopInstaller._get_rootcCs|jSr)Zegg_linkrrrrr jszDevelopInstaller._get_targetN)r4r5r6rr rrrrr8fsr8) rZ distutilsr itertoolsZsetuptools.extern.six.movesrchain from_iterabler0rr8rrrrs   [site-packages/setuptools/__pycache__/package_index.cpython-38.pyc000064400000100312151464452710021135 0ustar00U Qab@sdZddlZddlZddlZddlZddlZddlZddlZddlZddl Z ddl m Z ddl m Z ddlmZmZmZmZddlZddlmZmZmZmZmZmZmZmZmZmZmZm Z m!Z!ddlm"Z"ddl#m$Z$dd l%m&Z&dd l'm(Z(dd l)m*Z*dd l+m,Z,dd l-m.Z.e/Z0e1dZ2e1dej3Z4e1dZ5e1dej3j6Z7d8Z9ddddgZ:dZ;dZedZ?ddZ@ddZAddZBdGd dZCdHd!d"ZDdId#d$ZEdedfd%dZFdJd&d'ZGd(d)ZHe1d*ej3ZIeHd+d,ZJGd-d.d.ZKGd/d0d0eKZLGd1ddeZMe1d2jNZOd3d4ZPd5d6ZQdKd7d8ZRd9d:ZSGd;d<d<ZTGd=d>d>ejUZVejWjXfd?d@ZYdAdBZZeRe;eYZYdCdDZ[dEdFZ\dS)Lz#PyPI and direct package downloadingNwraps)six)urllib http_client configparsermap) CHECKOUT_DIST Distribution BINARY_DISTnormalize_path SOURCE_DIST Environmentfind_distributions safe_name safe_version to_filename Requirement DEVELOP_DISTEGG_DIST) ssl_support)log)DistutilsError) translate)get_all_headers)unescape)Wheelz^egg=([-A-Za-z0-9_.+!]+)$zhref\s*=\s*['"]?([^'"> ]+)z([^<]+)\n\s+\(md5\)z([-+.a-z0-9]{2,}):z.tar.gz .tar.bz2 .tar .zip .tgz PackageIndexdistros_for_urlparse_bdist_wininstinterpret_distro_namezrP)filenamerGr&r&r'distros_for_filenames  rTc cs||d}|s,tdd|ddDr,dStdt|dD]8}t||d|d|d||d|||dVq>dS)zGenerate alternative interpretations of a source distro name Note: if `location` is a filesystem filename, you should call ``pkg_resources.normalize_path()`` on it before passing it to this routine! rJcss|]}td|VqdS)z py\d\.\d$N)rerE).0pr&r&r' sz(interpret_distro_name..Nr6) py_versionrBrQ)r9anyrangerOr join)rKrPrGrZrBrQr;rWr&r&r'r s ccsft}|j}|dkr:tj|j|D]}|||Vq$n(|D]"}||}||kr>|||Vq>dS)zHList unique elements, preserving order. Remember all elements ever seen.N)setaddrZmoves filterfalse __contains__)iterablekeyseenZseen_addZelementkr&r&r'unique_everseens rfcstfdd}|S)zs Wrap a function returning an iterable such that the resulting iterable only ever yields unique items. cst||SN)rf)argskwargsfuncr&r'wrapperszunique_values..wrapperr)rkrlr&rjr' unique_valuessrmz(<([^>]*\srel\s*=\s*['"]?([^'">]+)[^>]*)>c cst|D]d}|\}}tttj|d}d|ksDd|kr t |D]}t j |t |dVqNq dD]@}||}|dkrtt ||}|rtt j |t |dVqtdS)zEFind rel="homepage" and rel="download" links in `page`, yielding URLs,Zhomepager3r6)z Home PagezDownload URLr2N)RELfinditergroupsr^rstrstripr+r9HREFrr#urljoin htmldecoderFfindsearch)r:pagerEtagZrelZrelsposr&r&r'find_external_linkss   r|c@s(eZdZdZddZddZddZdS) ContentCheckerzP A null content checker that defines the interface for checking content cCsdS)z3 Feed a block of data to the hash. Nr&selfblockr&r&r'feedszContentChecker.feedcCsdS)zC Check the hash. Return False if validation fails. Tr&rr&r&r'is_validszContentChecker.is_validcCsdS)zu Call reporter with information about the checker (hash name) substituted into the template. Nr&)rreportertemplater&r&r'reportszContentChecker.reportN)__name__ __module__ __qualname____doc__rrrr&r&r&r'r}sr}c@sBeZdZedZddZeddZddZ dd Z d d Z d S) HashCheckerzK(?Psha1|sha224|sha384|sha256|sha512|md5)=(?P[a-f0-9]+)cCs||_t||_||_dSrg) hash_namehashlibnewhashexpected)rrrr&r&r'__init__s zHashChecker.__init__cCs>tj|d}|stS|j|}|s0tS|f|S)z5Construct a (possibly null) ContentChecker from a URLr2)rr#r7r}patternrx groupdict)clsr:r@rEr&r&r'from_urls zHashChecker.from_urlcCs|j|dSrg)rupdater~r&r&r'rszHashChecker.feedcCs|j|jkSrg)rZ hexdigestrrr&r&r'r"szHashChecker.is_validcCs||j}||Srg)r)rrrmsgr&r&r'r%s zHashChecker.reportN) rrrrUcompilerr classmethodrrrrr&r&r&r'r s rcs<eZdZdZdJddZdKd d ZdLd d ZdMd dZddZddZ ddZ ddZ dNddZ ddZ dOfdd ZddZdd Zd!d"Zd#d$Zd%d&ZdPd'd(ZdQd)d*Zd+d,Zd-Zd.d/Zd0d1ZdRd2d3Zd4d5Zd6d7Zd8d9Zd:d;Zdd?Z!d@dAZ"dBdCZ#dDdEZ$dFdGZ%dHdIZ&Z'S)Trz;A distribution index that scans web pages for download URLshttps://pypi.org/simple/*NTcOstj|f|||dd|d |_i|_i|_i|_td t t |j |_ g|_|ortjor|prt}|rt||_n tjj|_dS)Nr1|)rrr, index_url scanned_urls fetched_urls package_pagesrUrr]rrrEallowsto_scanrZ is_availableZfind_ca_bundleZ opener_foropenerrrequesturlopen)rrZhostsZ ca_bundleZ verify_sslrhkwZuse_sslr&r&r'r-s zPackageIndex.__init__Fc Cs||jkr|sdSd|j|<t|s2||dStt|}|r\||sPdS|d||sn|rn||jkrtt|j |dS||sd|j|<dS| d|d|j|<d}| |||}|dkrdSd|j|j <d|j ddkr|dS|j }|}t|tsNt|tjjr0d }n|j d p@d }||d }|t|D](} tj|t| d } || q`| |j!rt"|d ddkr|#||}dS)zexistswarnisdirrealpathlistdirrr]rTrrrr_)rfnnestedr>itemrr&r&r'rus    zPackageIndex.process_filenamecCsbt|}|o|ddk}|s8|tj|drentryr&r&r'rXs   z.PackageIndex.scan_egg_links..)filterrRr>rr itertoolsstarmap scan_egg_link)rZ search_pathdirsZ egg_linksr&r&r'scan_egg_linkss zPackageIndex.scan_egg_linksc Csttj||}ttdttj|}W5QRXt |dkrDdS|\}}t tj||D](}tjj|f||_ t |_ ||q^dS)NrY)openrRr>r]rrrrrrsrOrrKr rBr_)rr>rZ raw_lineslinesZegg_pathZ setup_pathrHr&r&r'rs  zPackageIndex.scan_egg_linkc sfdd}t|D]:}z |tj|t|dWqtk rNYqXq||\}}|rt||D]H}t |\}} | dr| s|r|d||f7}n | |qlt dd|SdSd S) z#Process the contents of a PyPI pagecs|jrtttjj|tjdd}t|dkrd|dkrt |d}t |d}dj | i|<t|t|fSdS)Nr1rYr5r6rT)NN)r-rrrrr#r8rOr9rrr setdefaultr+r)rr;pkgverrr&r'scans   z(PackageIndex.process_index..scanr6.pyz #egg=%s-%scSsd|dddS)Nz%sr6rY)rF)mr&r&r'z,PackageIndex.process_index..rN)rtrprr#rurvrFr$r|rAr,need_version_infoscan_urlPYPI_MD5sub) rr:ryrrErrnew_urlr/fragr&rr'rs(      zPackageIndex.process_indexcCs|d|dS)NzPPage at %s links to .py file(s) without version info; an index scan is required.)scan_allrr:r&r&r'rszPackageIndex.need_version_infocGs:|j|jkr*|r |j|f||d||jdS)Nz6Scanning index of all packages (this may take a while))rrrrrrrrhr&r&r'rs zPackageIndex.scan_allcCsz||j|jd|j|js:||j|jd|j|jsR||t|j|jdD]}||qfdS)Nr1r&) rr unsafe_namerrrcrLnot_found_in_indexr)r requirementr:r&r&r' find_packagess zPackageIndex.find_packagescsR|||||jD]"}||kr0|S|d||qtt|||S)Nz%s does not match %s)prescanrrcrsuperrobtain)rrZ installerrH __class__r&r'rs zPackageIndex.obtaincCsL||jd||sH|t|td|jjtj |fdS)z- checker is a ContentChecker zValidating %%s checksum for %sz7%s validation failed for %s; possible download problem?N) rrrrrRunlinkrrr.r>rP)rcheckerrStfpr&r&r' check_hashs zPackageIndex.check_hashcCsN|D]D}|jdks0t|r0|ds0tt|r<||q|j|qdS)z;Add `urls` to the list that will be prescanned for searchesNfile:)rrr-rrrappend)rZurlsr:r&r&r'add_find_linkss  zPackageIndex.add_find_linkscCs"|jrtt|j|jd|_dS)z7Scan urls scheduled for prescanning (e.g. --find-links)N)rrrrrr&r&r'rszPackageIndex.prescancCs<||jr|jd}}n |jd}}|||j|dS)Nz#Couldn't retrieve index page for %rz3Couldn't find index page for %r (maybe misspelled?))rcrrrr)rrZmethrr&r&r'r#s  zPackageIndex.not_found_in_indexcCs~t|tsjt|}|rR||d||}t|\}}|drN||||}|Stj |rb|St |}t | ||ddS)aLocate and/or download `spec` to `tmpdir`, returning a local path `spec` may be a ``Requirement`` object, or a string containing a URL, an existing local filename, or a project/version requirement spec (i.e. the string form of a ``Requirement`` object). If it is the URL of a .py file with an unambiguous ``#egg=name-version`` tag (i.e., one that escapes ``-`` as ``_`` throughout), a trivial ``setup.py`` is automatically created alongside the downloaded file. If `spec` is a ``Requirement`` object or a string containing a project/version requirement spec, this method returns the location of a matching distribution (possibly after downloading it to `tmpdir`). If `spec` is a locally existing file or directory name, it is simply returned unchanged. If `spec` is a URL, it is downloaded to a subpath of `tmpdir`, and the local filename is returned. Various errors may be raised if a problem occurs during downloading. r6rrKN)rrr _download_urlrFrAr, gen_setuprRr>rr(rfetch_distribution)rr%tmpdirr<foundr/r@r&r&r'r3-s    zPackageIndex.downloadc sd|id}d fdd }|rH|||}|s^|dk r^|||}|dkrjdk rx||}|dkr|s|||}|dkrdrdpd|nd||j|jd SdS) a|Obtain a distribution suitable for fulfilling `requirement` `requirement` must be a ``pkg_resources.Requirement`` instance. If necessary, or if the `force_scan` flag is set, the requirement is searched for in the (online) package index as well as the locally installed packages. If a distribution matching `requirement` is found, the returned distribution's ``location`` is the value you would have gotten from calling the ``download()`` method with the matching distribution's URL or filename. If no matching distribution is found, ``None`` is returned. If the `source` flag is set, only source distributions and source checkout links will be considered. Unless the `develop_ok` flag is set, development and system eggs (i.e., those using the ``.egg-info`` format) will be ignored. zSearching for %sNcs|dkr }||jD]v}|jtkrFsF|krd|d|<q||ko\|jtkp\ }|r|j}||_tj |jr|SqdS)Nz&Skipping development or system egg: %sr6) rcrBrrr r3rKdownload_locationrRr>r)ZreqenvrHZtestZloc develop_okrZskippedsourcerr&r'rwgs&z-PackageIndex.fetch_distribution..findz:No local packages or working download links found for %s%sza source distribution of rzBest match: %s)rK)N)rrrrrZcloner) rrr force_scanrrZ local_indexrHrwr&rr'rOs2         zPackageIndex.fetch_distributioncCs"|||||}|dk r|jSdS)a3Obtain a file suitable for fulfilling `requirement` DEPRECATED; use the ``fetch_distribution()`` method now instead. For backward compatibility, this routine is identical but returns the ``location`` of the downloaded distribution instead of a distribution object. N)rrK)rrrrrrHr&r&r'fetchszPackageIndex.fetchc Cst|}|r*ddt||ddDp,g}t|dkrtj|}tj||krtj ||}ddl m }|||st |||}ttj |dd2} | d|dj|djtj|dfW5QRX|S|rtd ||fntd dS) NcSsg|]}|jr|qSr&)rM)rVdr&r&r' sz*PackageIndex.gen_setup..r6r)samefilezsetup.pywzIfrom setuptools import setup setup(name=%r, version=%r, py_modules=[%r]) zCan't unambiguously interpret project/version identifier %r; any dashes in the name or version should be escaped using underscores. %rzpCan't process plain .py files without an '#egg=name-version' suffix to enable automatic setup script generation.)rDrEr rFrOrRr>rPdirnamer]Zsetuptools.command.easy_installr shutilZcopy2rwriterLrMsplitextr) rrSr@rrErrPZdstr rr&r&r'rsB       zPackageIndex.gen_setupi c Cs|d|d}zt|}||}t|tjjrJt d||j |j f|}d}|j }d}d|krt |d} ttt| }||||||t|dV} ||} | r|| | | |d7}||||||qqq|||| W5QRX|WS|r|XdS) NzDownloading %szCan't download %s: %s %srr2zcontent-lengthzContent-Lengthwbr6)rrrrrrrrrrrr dl_blocksizermaxrint reporthookrrrr r) rr:rSfprrblocknumZbssizeZsizesrrr&r&r' _download_tos:        zPackageIndex._download_tocCsdSrgr&)rr:rSrZblksizerr&r&r'rszPackageIndex.reporthookc Cs|drt|Szt||jWSttjfk r}z.z%s %szDownload error for %s: %sz;%s returned a bad status line. The server might be down, %s)r- local_openopen_with_authrr$r InvalidURLr]rhrrrrrZURLErrorreasonZ BadStatusLinelineZ HTTPExceptionsocket)rr:Zwarningvrr&r&r'rs> "zPackageIndex.open_urlcCst|\}}|r0d|kr4|dddd}qnd}|drJ|dd}tj||}|dksj|d rv|||S|d ks|d r|||S|d r| ||S|d krt j t j |dS||d|||SdS)Nz...\_Z__downloaded__rIr*Zsvnzsvn+Zgitzgit+zhg+rrYT)rAreplacer,rRr>r]r- _download_svn _download_git _download_hgrr url2pathnamer#r7r_attempt_download)rr<r:rr.r@rSr&r&r'rs$        zPackageIndex._download_urlcCs||ddS)NT)rrr&r&r'r:szPackageIndex.scan_urlcCs6|||}d|ddkr.||||S|SdS)Nrrr)rrr+_download_html)rr:rSrr&r&r'r)=s zPackageIndex._attempt_downloadcCsnt|}|D]>}|r td|rF|t||||SqLq |t|td|dS)Nz ([^- ]+ - )?Revision \d+:zUnexpected HTML page found at ) r���rs���rU���rx���r���rR���r���r%��r���)r���r:���r���rS���r���r��r&���r&���r'���r*��D��s����   zPackageIndex._download_htmlc�����������������C���s��t�dt�|ddd�}d}|�drd|krtj|\}}}}}} |s|drd |d d��kr|d d��d d\}}t |\} } | rd | kr| d d\} } d | | f�}nd | �}| }|||||| f}tj |}|� d||�t d|||f��|S�)Nz"SVN download support is deprecatedr5���r6���r���r���zsvn:@z//r1���rY���:z --username=%s --password=%sz --username=z'Doing subversion checkout from %s to %szsvn checkout%s -q %s %s)warningsr��� UserWarningr9���r+���r-���r���r#���r7��� _splituser urlunparser���rR���system)r���r:���rS���Zcredsr<���netlocr>���rW���qr���authhostuserZpwr;���r&���r&���r'���r%��S��s&����   zPackageIndex._download_svnc�����������������C���sp���t�j|�\}}}}}|ddd�}|ddd�}d�}d|krR|dd\}}t�j||||df}�|�|fS�)N+r6���r2���r5���r���r+��r���)r���r#���Zurlsplitr9���rsplitZ urlunsplit)r:��� pop_prefixr<���r2��r>���r?���r���revr&���r&���r'���_vcs_split_rev_from_urli��s����z$PackageIndex._vcs_split_rev_from_urlc�����������������C���sl���|�ddd�}|�j|dd\}}|�d||�td||f��|d�k rh|�d|�td ||f��|S�) Nr5���r6���r���Tr9��zDoing git clone from %s to %szgit clone --quiet %s %szChecking out %szgit -C %s checkout --quiet %sr9���r;��r���rR���r1��r���r:���rS���r:��r&���r&���r'���r&��{��s���� zPackageIndex._download_gitc�����������������C���sl���|�ddd�}|�j|dd\}}|�d||�td||f��|d�k rh|�d|�td ||f��|S�) Nr5���r6���r���Tr<��zDoing hg clone from %s to %szhg clone --quiet %s %szUpdating to %szhg --cwd %s up -C -r %s -qr=��r>��r&���r&���r'���r'����s���� zPackageIndex._download_hgc�����������������G���s���t�j|f|��d�S�rg���)r���r���r���r&���r&���r'���r�����s����zPackageIndex.debugc�����������������G���s���t�j|f|��d�S�rg���)r���r���r���r&���r&���r'���r�����s����zPackageIndex.infoc�����������������G���s���t�j|f|��d�S�rg���)r���r���r���r&���r&���r'���r�����s����zPackageIndex.warn)r���r���NT)F)F)F)N)N)FFFN)FF)N)F)(r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r���r3���r���r��r���r��r��r��r���r���r���r)��r*��r%�� staticmethodr;��r&��r'��r���r���r��� __classcell__r&���r&���r���r'���r���*��sX���������  3   +   #������ L )$ # z!&(#(\d+|x[\da-fA-F]+)|[\w.:-]+);?c�����������������C���s���|��d}t|S�)Nr���)rF���r���)rE���Zwhatr&���r&���r'��� decode_entity��s���� rA��c�����������������C���s ���t�t|�S�)a�� Decode HTML entities in the given text. >>> htmldecode( ... 'https://../package_name-0.1.2.tar.gz' ... '?tokena=A&tokenb=B">package_name-0.1.2.tar.gz') 'https://../package_name-0.1.2.tar.gz?tokena=A&tokenb=B">package_name-0.1.2.tar.gz' ) entity_subrA��)textr&���r&���r'���rv�����s���� rv���c��������������������s����fdd}|S�)Nc��������������������s����fdd}|S�)Nc��������������� ������s2���t��}t��z�|�|W�S�t�|�X�d�S�rg���)r��ZgetdefaulttimeoutZsetdefaulttimeout)rh���ri���Z old_timeout)rk���timeoutr&���r'���_socket_timeout��s ���� z@socket_timeout.<locals>._socket_timeout.<locals>._socket_timeoutr&���)rk���rE��rD��rj���r'���rE����s����z'socket_timeout.<locals>._socket_timeoutr&���)rD��rE��r&���rF��r'���socket_timeout��s���� rG��c�����������������C���s2���t�j|�}|�}t|}|�}|ddS�)aq�� A function compatible with Python 2.3-3.3 that will encode auth from a URL suitable for an HTTP header. >>> str(_encode_auth('username%3Apassword')) 'dXNlcm5hbWU6cGFzc3dvcmQ=' Long auth strings should not cause a newline to be inserted. >>> long_auth = 'username:' + 'password'*10 >>> chr(10) in str(_encode_auth(long_auth)) False  r���)r���r#���r8���encodebase64Z b64encoder���r$��)r4��Zauth_sZ auth_bytesZ encoded_bytesZencodedr&���r&���r'��� _encode_auth��s ����  rK��c�������������������@���s(���e�Zd�ZdZdd�Zdd�Zdd�ZdS�) Credentialz: A username/password pair. Use like a namedtuple. c�����������������C���s���||�_�||�_d�S�rg���usernamepassword)r���rN��rO��r&���r&���r'���r�����s����zCredential.__init__c�����������������c���s���|�j�V��|�jV��d�S�rg���rM��r���r&���r&���r'���__iter__��s����zCredential.__iter__c�����������������C���s ���dt�|��S�)Nz%(username)s:%(password)s)varsr���r&���r&���r'���__str__��s����zCredential.__str__N)r���r���r���r���r���rP��rR��r&���r&���r&���r'���rL����s���rL��c�������������������@���s0���e�Zd�Zdd�Zedd�Zdd�Zdd�Zd S�) PyPIConfigc�����������������C���sP���t�dddgd}tj|�|�tjtjdd}tj |rL|� |�dS�)z% Load from ~/.pypirc rN��rO�� repositoryr���~z.pypircN) dictfromkeysr���RawConfigParserr���rR���r>���r]��� expanduserr���r���)r���defaultsZrcr&���r&���r'���r�����s ���� zPyPIConfig.__init__c��������������������s&����fdd���D�}tt�j|S�)Nc��������������������s ���g�|�]}��|d��r|qS�)rT��)r���rs���)rV���sectionr���r&���r'���r����s���z2PyPIConfig.creds_by_repository.<locals>.<listcomp>)ZsectionsrV��r���_get_repo_cred)r���Zsections_with_repositoriesr&���r���r'���creds_by_repository��s���� zPyPIConfig.creds_by_repositoryc�����������������C���s6���|��|d�}|t|��|d�|��|d�fS�)NrT��rN��rO��)r���rs���rL��)r���r[��Zrepor&���r&���r'���r\����s ����zPyPIConfig._get_repo_credc�����������������C���s*���|�j��D�]\}}||r |��S�q dS�)z If the URL indicated appears to be a repository defined in this config, return the credential for that repository. N)r]��itemsr-���)r���r:���rT��credr&���r&���r'���find_credential ��s���� zPyPIConfig.find_credentialN)r���r���r���r���propertyr]��r\��r`��r&���r&���r&���r'���rS����s ���  rS��c�����������������C���s:��t�j|�}|\}}}}}}|dr0td|dkrFt|\} } nd} | s~t�|�} | r~t | } | j |�f} t j d | ��| rdt | �} || ||||f} t�j| }t�j|}|d| �n t�j|�}|dt�||}| r6t�j|j\}}}}}}||kr6|| kr6||||||f} t�j| |_|S�) z4Open a urllib2 request, handling HTTP authenticationr,��znonnumeric port: '')ZhttpZhttpsN*Authenticating as %s for %s (from .pypirc)zBasic Z Authorizationz User-Agent)rb��)r���r#���r7���r,���r���r��r/��rS��r`��rr���rN��r���r���rK��r0��r���ZRequestZ add_header user_agentr:���)r:���r���Zparsedr<���r2��r>���Zparamsr?���r���r4��Zaddressr_��r���r;���r���r���r��s2Zh2Zpath2Zparam2Zquery2Zfrag2r&���r&���r'���r����s8����          r��c�����������������C���s ���|��d\}}}�|r|nd|�fS�)zJsplituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.r+��N) rpartition)r5��r6��Zdelimr&���r&���r'���r/��F��s����r/��c�����������������C���s���|�S�rg���r&���)r:���r&���r&���r'��� fix_sf_urlP��s����rf��c�������������� ���C���s��t�j|�\}}}}}}t�j|}tj|r<t�j|�S�| drtj |rg�}t |D�]d} tj || } | dkrt | d} | �} W�5�Q�R�X��qntj | r| d7�} |dj| d�q`d} | j|�d |d} d \}}n d \}}} d d i}t| }t�j|�||||S�) z7Read a local path, with special support for directoriesr1���z index.htmlrz<a href="{name}">{name}</a>)r.���zB<html><head><title>{url}{files}rH)r:files)ZOK)rzPath not foundz Not foundrz text/html)rr#r7rr(rRr>isfilerr,rrr]rrrformatrStringIOrr)r:r<r=r>Zparamr?rrSrhrfilepathrZbodyrZstatusmessagerZ body_streamr&r&r'rTs.        r)N)N)N)N)r!)]rsysrRrUr rrJrrr- functoolsrZsetuptools.externrZsetuptools.extern.six.movesrrrrr"Z pkg_resourcesr r r r r rrrrrrrrrZ distutilsrZdistutils.errorsrZfnmatchrZsetuptools.py27compatrZsetuptools.py33compatrZsetuptools.wheelrtypeZ __metaclass__rrDIrtrrErr9rN__all__Z_SOCKET_TIMEOUTZ_tmplrk version_inforcr(rrArrCrTr rfrmror|r}rrrrBrArvrGrKrLrXrSrrrr/rfrr&r&r&r's  <           !  $   !  &/ site-packages/setuptools/__pycache__/archive_util.cpython-38.pyc000064400000012010151464452710021026 0ustar00U Qab@sdZddlZddlZddlZddlZddlZddlZddlmZddl m Z ddddd d d gZ Gd d d eZ d dZ e dfddZe fdd Ze fddZe fddZeeefZdS)z/Utilities for extracting common archive formatsN)DistutilsError)ensure_directoryunpack_archiveunpack_zipfileunpack_tarfiledefault_filterUnrecognizedFormatextraction_driversunpack_directoryc@seZdZdZdS)rz#Couldn't recognize the archive typeN)__name__ __module__ __qualname____doc__rr;/usr/lib/python3.8/site-packages/setuptools/archive_util.pyrscCs|S)z@The default progress/filter callback; returns True for all filesr)srcdstrrrrsc CsN|ptD]4}z||||Wntk r4YqYqXdSqtd|dS)aUnpack `filename` to `extract_dir`, or raise ``UnrecognizedFormat`` `progress_filter` is a function taking two arguments: a source path internal to the archive ('/'-separated), and a filesystem path where it will be extracted. The callback must return the desired extract path (which may be the same as the one passed in), or else ``None`` to skip that file or directory. The callback can thus be used to report on the progress of the extraction, as well as to filter the items extracted or alter their extraction paths. `drivers`, if supplied, must be a non-empty sequence of functions with the same signature as this function (minus the `drivers` argument), that raise ``UnrecognizedFormat`` if they do not support extracting the designated archive type. The `drivers` are tried in sequence until one is found that does not raise an error, or until all are exhausted (in which case ``UnrecognizedFormat`` is raised). If you do not supply a sequence of drivers, the module's ``extraction_drivers`` constant will be used, which means that ``unpack_zipfile`` and ``unpack_tarfile`` will be tried, in that order. Nz!Not a recognized archive type: %s)r r)filename extract_dirprogress_filterZdriversZdriverrrrrs  c Cstj|std||d|fi}t|D]\}}}||\}}|D],} || dtj|| f|tj|| <qH|D]T} tj|| } ||| | } | sqzt| tj|| } t| | t | | qzq.dS)z"Unpack" a directory, using the same interface as for archives Raises ``UnrecognizedFormat`` if `filename` is not a directory z%s is not a directory/N) ospathisdirrwalkjoinrshutilZcopyfileZcopystat) rrrpathsbasedirsfilesrrdftargetrrrr ?s$   * c Cst|std|ft|}|D]}|j}|ds,d|dkrPq,tj j |f|d}|||}|sxq,| drt |n4t || |j}t|d}||W5QRX|jd?} | r,t|| q,W5QRXdS)zUnpack zip `filename` to `extract_dir` Raises ``UnrecognizedFormat`` if `filename` is not a zipfile (as determined by ``zipfile.is_zipfile()``). See ``unpack_archive()`` for an explanation of the `progress_filter` argument. z%s is not a zip filer..wbN)zipfileZ is_zipfilerZZipFileZinfolistr startswithsplitrrrendswithrreadopenwriteZ external_attrchmod) rrrzinfonamer$datar#Zunix_attributesrrrrZs(         c Csfzt|}Wn$tjk r2td|fYnXt|dd|_|D]}|j}|dsPd| dkrPt j j |f| d}|dk r| s|r|j}|rt|j}t ||}t|}||}q|dk rP|s|rP|||} | rP| t jr"| dd} z||| WqPtjk rJYqPXqPW5QRdSQRXdS) zUnpack tar/tar.gz/tar.bz2 `filename` to `extract_dir` Raises ``UnrecognizedFormat`` if `filename` is not a tarfile (as determined by ``tarfile.open()``). See ``unpack_archive()`` for an explanation of the `progress_filter` argument. z/%s is not a compressed or uncompressed tar filecWsdS)Nr)argsrrrz unpack_tarfile..rr%NT)tarfiler-ZTarErrorr contextlibclosingchownr2r)r*rrrZislnkZissymZlinkname posixpathdirnamenormpathZ _getmemberisfilerr+sepZ_extract_memberZ ExtractError) rrrZtarobjmemberr2Z prelim_dstZlinkpathrZ final_dstrrrrs:         )rr(r8rrr<r9Zdistutils.errorsrZ pkg_resourcesr__all__rrrr rrr rrrrs2   #  % .site-packages/setuptools/__pycache__/_deprecation_warning.cpython-38.pyc000064400000001004151464452710022532 0ustar00U Qab@sGdddeZdS)c@seZdZdZdS)SetuptoolsDeprecationWarningz Base class for warning deprecations in ``setuptools`` This class is not derived from ``DeprecationWarning``, and as such is visible by default. N)__name__ __module__ __qualname____doc__rrC/usr/lib/python3.8/site-packages/setuptools/_deprecation_warning.pyrsrN)Warningrrrrrsite-packages/setuptools/__pycache__/site-patch.cpython-38.pyc000064400000002702151464452710020420 0ustar00U Qab@sddZedkre[dS)c Cs ddl}ddl}|jd}|dks2|jdkr8|s8g}n ||j}t|di}|jt |d}|j t }|D]}||ksr|sqr||}|dk r| d}|dk r| dq.qrz ddl} | d|g\} } } Wntk rYqrYnX| dkrqrz| d| | | W5| Xq.qrtdtdd|jD} t|d d}d|_|D]}t|qX|j|7_t|d\}}d}g}|jD]b}t|\}}||kr|dkrt |}|| ks|dkr||n||||d 7}q||jdd<dS) N PYTHONPATHZwin32path_importer_cachesitez$Couldn't find the real 'site' modulecSsg|]}t|ddfqS))makepath).0itemr 9/usr/lib/python3.8/site-packages/setuptools/site-patch.py )sz__boot.. __egginsertr)sysosenvirongetplatformsplitpathsepgetattrpathlendirname__file__ find_module load_moduleimp ImportErrorclosedictr addsitedirrappendinsert)r rrZpicZstdpathZmydirrZimporterloaderrstreamrZdescr known_pathsZoldposdZndZ insert_atnew_pathpZnpr r r __boots`                 r(rN)r(__name__r r r r sGsite-packages/setuptools/__pycache__/monkey.cpython-38.pyc000064400000011044151464452710017660 0ustar00U Qab@sdZddlZddlZddlZddlZddlZddlmZddl Z ddl m Z ddl Z gZ ddZddZd d Zd d Zd dZddZddZddZdS)z Monkey patching of distutils. N) import_module)sixcCs"tdkr|f|jSt|S)am Returns the bases classes for cls sorted by the MRO. Works around an issue on Jython where inspect.getmro will not return all base classes if multiple classes share the same name. Instead, this function will return a tuple containing the class itself, and the contents of cls.__bases__. See https://github.com/pypa/setuptools/issues/1024. ZJython)platformZpython_implementation __bases__inspectZgetmro)clsr5/usr/lib/python3.8/site-packages/setuptools/monkey.py_get_mros  r cCs0t|tjrtnt|tjr tndd}||S)NcSsdS)Nr)itemrrr *zget_unpatched..) isinstancerZ class_typesget_unpatched_classtypes FunctionTypeget_unpatched_function)r lookuprrr get_unpatched&s rcCs:ddt|D}t|}|jds6d|}t||S)zProtect against re-patching the distutils if reloaded Also ensures that no other distutils extension monkeypatched the distutils first. css|]}|jds|VqdS) setuptoolsN) __module__ startswith).0rrrr 5s z&get_unpatched_class.. distutilsz(distutils has already been patched by %r)r nextrrAssertionError)rZexternal_basesbasemsgrrr r/s rcCstjtj_tjdk}|r"tjtj_tjdkp^dtjko@dknp^dtjkoZdkn}|rrd}|tjj _ t tj tjtj fD]}tj j|_qtjjtj_tjjtj_dtjkrtjjtjd_tdS)N)r) )r)rr$)rr zhttps://upload.pypi.org/legacy/zdistutils.command.build_ext)rZCommandrZcoresys version_infofindallZfilelistZconfigZ PyPIRCCommandZDEFAULT_REPOSITORY_patch_distribution_metadatadistcmdZ Distribution extensionZ Extensionmodules#patch_for_msvc_specialized_compiler)Zhas_issue_12885Zneeds_warehouseZ warehousemodulerrr patch_allAs*          r0cCs*dD] }ttj|}ttjj||qdS)zDPatch write_pkg_file and read_pkg_file for higher metadata standards)Zwrite_pkg_fileZ read_pkg_fileZget_metadata_versionN)getattrrr*setattrrZDistributionMetadata)attrZnew_valrrr r)hs r)cCs*t||}t|d|t|||dS)z Patch func_name in target_mod with replacement Important - original must be resolved by name to avoid patching an already patched function. unpatchedN)r1vars setdefaultr2)Z replacementZ target_mod func_nameoriginalrrr patch_funcos r9cCs t|dS)Nr4)r1) candidaterrr rsrcstdtdkrdSfdd}t|d}t|d}zt|dt|d Wntk rlYnXzt|d Wntk rYnXzt|d Wntk rYnXdS) z\ Patch functions in distutils to use standalone Microsoft Visual C++ compilers. zsetuptools.msvcZWindowsNcsLd|kr dnd}||d}t|}t|}t||sBt||||fS)zT Prepare the parameters for patch_func to patch indicated function. msvc9Zmsvc9_Zmsvc14__)lstripr1rhasattr ImportError)Zmod_namer7Z repl_prefixZ repl_namereplmodZmsvcrr patch_paramss  z9patch_for_msvc_specialized_compiler..patch_paramszdistutils.msvc9compilerzdistutils._msvccompilerZfind_vcvarsallZquery_vcvarsallZ _get_vc_envZgen_lib_options)rrsystem functoolspartialr9r?)rCr;Zmsvc14rrBr r.s&    r.)__doc__r&Zdistutils.filelistrrrrE importlibrrZsetuptools.externrr__all__r rrr0r)r9rr.rrrr s$   'site-packages/setuptools/__pycache__/dep_util.cpython-38.pyc000064400000001463151464452710020167 0ustar00U Qab@sddlmZddZdS)) newer_groupcCsht|t|krtdg}g}tt|D]2}t||||r,||||||q,||fS)zWalk both arguments in parallel, testing if each source group is newer than its corresponding target. Returns a pair of lists (sources_groups, targets) where sources is newer than target, according to the semantics of 'newer_group()'. z5'sources_group' and 'targets' must be the same length)len ValueErrorrangerappend)Zsources_groupsZtargetsZ n_sourcesZ n_targetsir7/usr/lib/python3.8/site-packages/setuptools/dep_util.pynewer_pairwise_groupsr N)Zdistutils.dep_utilrr rrrr s site-packages/setuptools/__pycache__/_imp.cpython-38.pyc000064400000003542151464452710017306 0ustar00U Qab@s\dZddlZddlZddlZddlmZdZdZdZ dZ dZ dd d Z dd d Z d dZdS)zX Re-implementation of find_module and get_frozen_object from the deprecated imp module. N)module_from_specc CsVtj||}|dkr"td||jsBt|drBtjd|j}d}d}t|jt }|j dkst|rt |jtj j rt}d}d}}n|j dks|rt |jtj jrt}d}d}}n|jr:|j }tj|d }|tj jkrd nd }|tj jkrt}n&|tj jkrt}n|tj jkr t}|tthkrFt||}n d}d}}|||||ffS) z7Just like 'imp.find_module()', but with package supportN Can't find %ssubmodule_search_locationsz __init__.pyfrozenzbuilt-inrrrb) importlibutil find_spec ImportError has_locationhasattrspec_from_loaderloader isinstancetypeorigin issubclass machineryFrozenImporter PY_FROZENBuiltinImporter C_BUILTINospathsplitextSOURCE_SUFFIXES PY_SOURCEBYTECODE_SUFFIXES PY_COMPILEDEXTENSION_SUFFIXES C_EXTENSIONopen) modulepathsspecZkindfileZstaticr!suffixmoder03/usr/lib/python3.8/site-packages/setuptools/_imp.py find_modulesJ      r2cCs*tj||}|std||j|SNr)rrrrrget_code)r*r+r,r0r0r1get_frozen_object>s r5cCs&tj||}|std|t|Sr3)rrrrr)r*r+infor,r0r0r1 get_moduleEs r7)N)N)__doc__r importlib.utilrZimportlib.machineryZ py34compatrr$r&r(rrr2r5r7r0r0r0r1s  * site-packages/setuptools/__pycache__/site-patch.cpython-38.opt-1.pyc000064400000002702151464452710021357 0ustar00U Qab@sddZedkre[dS)c Cs ddl}ddl}|jd}|dks2|jdkr8|s8g}n ||j}t|di}|jt |d}|j t }|D]}||ksr|sqr||}|dk r| d}|dk r| dq.qrz ddl} | d|g\} } } Wntk rYqrYnX| dkrqrz| d| | | W5| Xq.qrtdtdd|jD} t|d d}d|_|D]}t|qX|j|7_t|d\}}d}g}|jD]b}t|\}}||kr|dkrt |}|| ks|dkr||n||||d 7}q||jdd<dS) N PYTHONPATHZwin32path_importer_cachesitez$Couldn't find the real 'site' modulecSsg|]}t|ddfqS))makepath).0itemr 9/usr/lib/python3.8/site-packages/setuptools/site-patch.py )sz__boot.. __egginsertr)sysosenvirongetplatformsplitpathsepgetattrpathlendirname__file__ find_module load_moduleimp ImportErrorclosedictr addsitedirrappendinsert)r rrZpicZstdpathZmydirrZimporterloaderrstreamrZdescr known_pathsZoldposdZndZ insert_atnew_pathpZnpr r r __boots`                 r(rN)r(__name__r r r r sGsite-packages/setuptools/__pycache__/__init__.cpython-38.pyc000064400000017130151464452710020117 0ustar00U Qabs@sdZddlZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl m Z ddlmZddlmZmZdd lmZmZddlZdd lmZdd lmZmZdd lmZdd lm Z e!Z"ddddddddgZ#ere#$dej%j&Z&dZ'dZ(dgZ)GdddZ*Gddde*Z+e*j,Z-er,e+j,Z.ddZ/ddZ0ej1j0je0_e 2ej1j3Z4Gd dde4Z3d!d"Z5ej6fd#d$Z7e 8dS)%z@Extensions to the 'distutils' for large or complex distributionsN)DistutilsOptionError) convert_path fnmatchcase)SetuptoolsDeprecationWarning)PY3 string_types)filtermap) Extension) DistributionFeature)Require)monkeysetupr rCommandr rr find_packagesfind_namespace_packagesTz lib2to3.fixesc@sBeZdZdZedddZeddZed d Zed d Z d S) PackageFinderzI Generate a list of all Python packages found within a directory .*cCs&t|t||jd||j|S)a Return a list all Python packages found within directory 'where' 'where' is the root directory which will be searched for packages. It should be supplied as a "cross-platform" (i.e. URL-style) path; it will be converted to the appropriate local path syntax. 'exclude' is a sequence of package names to exclude; '*' can be used as a wildcard in the names, such that 'foo.*' will exclude all subpackages of 'foo' (but not 'foo' itself). 'include' is a sequence of package names to include. If it's specified, only the named packages will be included. If it's not specified, all found packages will be included. 'include' can contain shell style wildcard patterns just like 'exclude'. ez_setup *__pycache__)rr)list_find_packages_iterr _build_filter)clswhereexcludeincluderr7/usr/lib/python3.8/site-packages/setuptools/__init__.pyfind4s  zPackageFinder.findc cstj|ddD]\}}}|dd}g|dd<|D]d}tj||} tj| |} | tjjd} d|ks4|| sxq4|| r|| s| V||q4qdS)zy All the packages found in 'where' that pass the 'include' filter, but not the 'exclude' filter. T followlinksNr) oswalkpathjoinrelpathreplacesep_looks_like_packageappend) rr r!r"rootdirsfilesZall_dirsdir full_pathZrel_pathpackagerrr#rKs  z!PackageFinder._find_packages_itercCstjtj|dS)z%Does a directory look like a package?z __init__.py)r'r)isfiler*r)rrr#r.gsz!PackageFinder._looks_like_packagecs fddS)z Given a list of patterns, return a callable that will be true only if the input matches at least one of the patterns. cstfddDS)Nc3s|]}t|dVqdS))patNr).0r8namerr# rsz@PackageFinder._build_filter....)anyr:Zpatternsr:r#rz-PackageFinder._build_filter..rr>rr>r#rlszPackageFinder._build_filterN)rrr) __name__ __module__ __qualname____doc__ classmethodr$r staticmethodr.rrrrr#r/s   rc@seZdZeddZdS)PEP420PackageFindercCsdS)NTrr7rrr#r.vsz'PEP420PackageFinder._looks_like_packageN)rArBrCrFr.rrrr#rGusrGcCs@tjtdd|D}|jdd|jr<||jdS)Ncss"|]\}}|dkr||fVqdS))Zdependency_linkssetup_requiresNr)r9kvrrr#r<sz*_install_setup_requires..T)Zignore_option_errors) distutilscorer dictitemsZparse_config_filesrHZfetch_build_eggs)attrsdistrrr#_install_setup_requiress   rQcKst|tjjf|SN)rQrKrLr)rOrrr#rsc@s:eZdZejZdZddZd ddZddZd d d Z dS)rFcKst||t||dS)zj Construct the command for dist, updating vars(self) with any keyword parameters. N)_Command__init__varsupdate)selfrPkwrrr#rTs zCommand.__init__NcCsBt||}|dkr"t||||St|ts>td|||f|S)Nz'%s' must be a %s (got `%s`))getattrsetattr isinstancer r)rWoptionZwhatdefaultvalrrr#_ensure_stringlikes   zCommand._ensure_stringlikecCspt||}|dkrdSt|tr6t||td|n6t|trTtdd|D}nd}|sltd||fdS)zEnsure that 'option' is a list of strings. If 'option' is currently a string, we split it either on /,\s*/ or /\s+/, so "foo bar baz", "foo,bar,baz", and "foo, bar baz" all become ["foo", "bar", "baz"]. Nz,\s*|\s+css|]}t|tVqdSrR)r[r )r9rJrrr#r<sz-Command.ensure_string_list..Fz''%s' must be a list of strings (got %r)) rYr[r rZresplitrallr)rWr\r^okrrr#ensure_string_lists   zCommand.ensure_string_listrcKs t|||}t|||SrR)rSreinitialize_commandrUrV)rWZcommandZreinit_subcommandsrXcmdrrr#reszCommand.reinitialize_command)N)r) rArBrCrSrDZcommand_consumes_argumentsrTr_rdrerrrr#rs  cCs&ddtj|ddD}ttjj|S)z% Find all files under 'path' css,|]$\}}}|D]}tj||VqqdSrR)r'r)r*)r9baser1r2filerrr#r<sz#_find_all_simple..Tr%)r'r(r r)r6)r)resultsrrr#_find_all_simples rjcCs6t|}|tjkr.tjtjj|d}t||}t|S)z Find all files under 'dir' and return the list of full filenames. Unless dir is '.', return full filenames with dir prepended. )start) rjr'curdir functoolspartialr)r+r r)r3r2Zmake_relrrr#findalls   ro)9rDr'sysrmZdistutils.corerKZdistutils.filelistr`Zdistutils.errorsrZdistutils.utilrZfnmatchrZ_deprecation_warningrZsetuptools.extern.sixrr Zsetuptools.extern.six.movesr r Zsetuptools.versionZ setuptoolsZsetuptools.extensionr Zsetuptools.distr rZsetuptools.dependsrrtypeZ __metaclass____all__r/version __version__Zbootstrap_install_fromZrun_2to3_on_doctestsZlib2to3_fixer_packagesrrGr$rrrQrrLZ get_unpatchedrrSrjrlroZ patch_allrrrr#s\        F  2  site-packages/setuptools/__pycache__/py34compat.cpython-38.pyc000064400000000702151464452710020360 0ustar00U Qab@sXddlZz ddlZWnek r(YnXz ejjZWnek rRddZYnXdS)NcCs|j|jS)N)loader load_modulename)specr9/usr/lib/python3.8/site-packages/setuptools/py34compat.pymodule_from_spec sr) importlibimportlib.util ImportErrorutilrAttributeErrorrrrrs  site-packages/setuptools/__pycache__/py33compat.cpython-38.opt-1.pyc000064400000002600151464452710021315 0ustar00U Qab2@sddlZddlZddlZz ddlZWnek r<dZYnXddlmZddlmZe Z e ddZ GdddZ eede Zeed dZedkrejZdS) N)six) html_parserOpArgz opcode argc@seZdZddZddZdS)Bytecode_compatcCs ||_dS)N)code)selfrr9/usr/lib/python3.8/site-packages/setuptools/py33compat.py__init__szBytecode_compat.__init__ccstd|jj}t|jj}d}d}||kr||}|tjkr||d||dd|}|d7}|tjkrtjd}||d}q$n d }|d7}t ||Vq$d S) z>Yield '(op,arg)' pair for each operation in code object 'code'briN) arrayrco_codelendisZ HAVE_ARGUMENTZ EXTENDED_ARGrZ integer_typesr)rbyteseofZptrZ extended_argopargZ long_typerrr __iter__s       zBytecode_compat.__iter__N)__name__ __module__ __qualname__r rrrrr rsrBytecodeunescape)rr collectionsZhtml ImportErrorZsetuptools.externrZsetuptools.extern.six.movesrtypeZ __metaclass__ namedtuplerrgetattrrrZ HTMLParserrrrr s     "  site-packages/setuptools/__pycache__/config.cpython-38.pyc000064400000042725151464452710017635 0ustar00U Qab6P@sddlmZmZddlZddlZddlZddlZddlZddlm Z ddlm Z ddlm Z ddl m Z ddlmZmZddlmZmZdd lmZdd lmZmZeZdd d ZddZddZdddZGdddZGdddeZ GdddeZ!dS))absolute_importunicode_literalsN) defaultdict)partialwraps) import_module)DistutilsOptionErrorDistutilsFileError) LegacyVersionparse) SpecifierSet) string_typesPY3Fc Csddlm}m}tj|}tj|s4td|t}t tj |zJ|}|rb| ng}||krx| ||j ||dt||j|d}W5t |Xt|S)a,Read given configuration file and returns options from it as a dict. :param str|unicode filepath: Path to configuration file to get options from. :param bool find_others: Whether to search for other configuration files which could be on in various places. :param bool ignore_option_errors: Whether to silently ignore options, values of which could not be resolved (e.g. due to exceptions in directives such as file:, attr:, etc.). If False exceptions are propagated as expected. :rtype: dict r) Distribution _Distributionz%Configuration file %s does not exist.) filenames)ignore_option_errors)Zsetuptools.distrrospathabspathisfiler getcwdchdirdirnameZfind_config_filesappendZparse_config_filesparse_configurationcommand_optionsconfiguration_to_dict) filepathZ find_othersrrrZcurrent_directoryZdistrhandlersr!5/usr/lib/python3.8/site-packages/setuptools/config.pyread_configurations*     r#cCs.djft}tt||}t|||}|S)z Given a target object and option key, get that option from the target object, either through a get_{key} method or from an attribute directly. z get_{key})formatlocals functoolsrgetattr) target_objkeyZ getter_nameZ by_attributegetterr!r!r" _get_optionEs r+cCs<tt}|D]*}|jD]}t|j|}|||j|<qq |S)zReturns configuration data gathered by given handlers as a dict. :param list[ConfigHandler] handlers: Handlers list, usually from parse_configuration() :rtype: dict )rdict set_optionsr+r(section_prefix)r Z config_dictZhandlerZoptionvaluer!r!r"rQs   rcCs6t|||}|t|j|||j}|||fS)aPerforms additional parsing of configuration options for a distribution. Returns a list of used option handlers. :param Distribution distribution: :param dict command_options: :param bool ignore_option_errors: Whether to silently ignore options, values of which could not be resolved (e.g. due to exceptions in directives such as file:, attr:, etc.). If False exceptions are propagated as expected. :rtype: list )ConfigOptionsHandlerr ConfigMetadataHandlermetadata package_dir)Z distributionrroptionsmetar!r!r"rcsrc@seZdZdZdZiZd%ddZeddZdd Z e d&d d Z e d dZ e ddZ e ddZe ddZeddZeddZe d'ddZe ddZe d(ddZdd Zd!d"Zd#d$ZdS)) ConfigHandlerz1Handles metadata supplied in configuration files.NFcCs^i}|j}|D].\}}||s&q||dd}|||<q||_||_||_g|_dS)N.) r.items startswithreplacestriprr(sectionsr-)selfr(r4rr=r. section_namesection_optionsr!r!r"__init__s  zConfigHandler.__init__cCstd|jjdS).Metadata item name to parser function mapping.z!%s must provide .parsers propertyN)NotImplementedError __class____name__)r>r!r!r"parserss zConfigHandler.parsersc Cst}|j}|j||}t|||}||kr6t||r>dSd}|j|}|rz ||}Wn tk r~d}|jszYnX|rdSt|d|d}|dkrt |||n|||j |dS)NFTzset_%s) tupler(aliasesgetr'KeyErrorrF Exceptionrsetattrr-r) r>Z option_namer/unknownr(Z current_valueZ skip_optionparsersetterr!r!r" __setitem__s0   zConfigHandler.__setitem__,cCs8t|tr|Sd|kr |}n ||}dd|DS)zRepresents value as a list. Value is split either by separator (defaults to comma) or by lines. :param value: :param separator: List items separator character. :rtype: list  cSsg|]}|r|qSr!)r<).0chunkr!r!r" sz-ConfigHandler._parse_list..) isinstancelist splitlinessplit)clsr/ separatorr!r!r" _parse_lists   zConfigHandler._parse_listcCsPd}i}||D]8}||\}}}||kr:td||||<q|S)zPRepresents value as a dict. :param value: :rtype: dict =z(Unable to parse option value to dict: %s)r\ partitionr r<)rZr/r[resultliner)sepvalr!r!r" _parse_dictszConfigHandler._parse_dictcCs|}|dkS)zQRepresents value as boolean. :param value: :rtype: bool )1trueZyes)lower)rZr/r!r!r" _parse_boolszConfigHandler._parse_boolcsfdd}|S)zReturns a parser function to make sure field inputs are not files. Parses a value after getting the key so error messages are more informative. :param key: :rtype: callable cs d}||rtd|S)Nfile:zCOnly strings are accepted for the {0} field, files are not accepted)r: ValueErrorr$)r/Zexclude_directiver)r!r"rNs z3ConfigHandler._exclude_files_parser..parserr!)rZr)rNr!rjr"_exclude_files_parsers z#ConfigHandler._exclude_files_parsercs\d}t|ts|S||s |S|t|d}dd|dD}dfdd|DS)aORepresents value as a string, allowing including text from nearest files using `file:` directive. Directive is sandboxed and won't reach anything outside directory with setup.py. Examples: file: README.rst, CHANGELOG.md, src/file.txt :param str value: :rtype: str rhNcss|]}tj|VqdSN)rrrr<rSrr!r!r" %sz,ConfigHandler._parse_file..rQrRc3s.|]&}|stj|r|VqdS)TN) _assert_localrrr _read_filermrZr!r"rn&s   )rVrr:lenrYjoin)rZr/Zinclude_directivespecZ filepathsr!rqr" _parse_files  zConfigHandler._parse_filecCs|tstd|dS)Nz#`file:` directive can not access %s)r:rrr )rr!r!r"ro-szConfigHandler._assert_localc Cs.tj|dd}|W5QRSQRXdS)Nzutf-8)encoding)ioopenread)rfr!r!r"rp3szConfigHandler._read_filec Csd}||s|S||dd}|}d|}|p@d}t}|r|d|kr||d}|dd} t | dkrtj t| d}| d}q|}nd|krtj t|d}t j d|zt |} t| |}W5t j ddt _ X|S) zRepresents value as a module attribute. Examples: attr: package.attr attr: package.module.attr :param str value: :rtype: str zattr:r7r8rAr/N)r:r;r<rYpoprsrrrsplitrrrsysinsertrr') rZr/r3Zattr_directiveZ attrs_pathZ attr_nameZ module_name parent_pathZ custom_pathpartsmoduler!r!r" _parse_attr8s0        zConfigHandler._parse_attrcsfdd}|S)zReturns parser function to represents value as a list. Parses a value applying given methods one after another. :param parse_methods: :rtype: callable cs|}D] }||}q|Srlr!)r/parsedmethod parse_methodsr!r"r ns z1ConfigHandler._get_parser_compound..parser!)rZrr r!rr"_get_parser_compoundes z"ConfigHandler._get_parser_compoundcCs6i}|pdd}|D]\}\}}||||<q|S)zParses section options into a dictionary. Optionally applies a given parser to values. :param dict section_options: :param callable values_parser: :rtype: dict cSs|Srlr!)rbr!r!r"z6ConfigHandler._parse_section_to_dict..)r9)rZr@Z values_parserr/r)_rbr!r!r"_parse_section_to_dictxs  z$ConfigHandler._parse_section_to_dictc Cs<|D].\}\}}z |||<Wqtk r4YqXqdS)zQParses configuration file section. :param dict section_options: N)r9rJ)r>r@namerr/r!r!r" parse_sections  zConfigHandler.parse_sectioncCsb|jD]R\}}d}|r"d|}t|d|ddd}|dkrTtd|j|f||q dS)zTParses configuration file items from one or more related sections. r7z_%szparse_section%sr8__Nz0Unsupported distribution option section: [%s.%s])r=r9r'r;r r.)r>r?r@Zmethod_postfixZsection_parser_methodr!r!r"r s"zConfigHandler.parsecstfdd}|S)z this function will wrap around parameters that are deprecated :param msg: deprecation message :param warning_class: class of warning exception to be raised :param func: function to be wrapped around cst||Srl)warningswarn)argskwargsfuncmsg warning_classr!r"config_handlers z@ConfigHandler._deprecated_config_handler..config_handlerr)r>rrrrr!rr"_deprecated_config_handlersz(ConfigHandler._deprecated_config_handler)F)rQ)N)N)rE __module__ __qualname____doc__r.rHrApropertyrFrP classmethodr\rcrgrkru staticmethodrorprrrrr rr!r!r!r"r6~s<  &        ,   r6csHeZdZdZdddddZdZdfd d Zed d Zd dZ Z S)r1r2Zurl description classifiers platforms)Z home_pageZsummaryZ classifierplatformFNcstt||||||_dSrl)superr1rAr3)r>r(r4rr3rDr!r"rAszConfigMetadataHandler.__init__c CsL|j}|j}|j}|j}|||||dt|||||d|||j|d S)rBz[The requires parameter is deprecated, please use install_requires for runtime dependencies.license) rkeywordsZprovidesZrequiresZ obsoletesrrrZlong_descriptionversionZ project_urls)r\rurcrkrDeprecationWarningr_parse_version)r> parse_listZ parse_file parse_dictZexclude_files_parserr!r!r"rFs( zConfigMetadataHandler.parserscCs||}||krB|}tt|tr>d}t|jft|S|||j }t |r^|}t|t st |drd tt|}nd|}|S)zSParses `version` option value. :param value: :rtype: str zCVersion loaded from {value} does not comply with PEP 440: {version}__iter__r8z%s)rur<rVr r r r$r%rr3callablerhasattrrsmapstr)r>r/rZtmplr!r!r"rs    z$ConfigMetadataHandler._parse_version)FN) rErrr.rHZ strict_moderArrFr __classcell__r!r!rr"r1s r1c@s\eZdZdZeddZddZddZdd Zd d Z d d Z ddZ ddZ ddZ dS)r0r4cCsN|j}t|jdd}|j}|j}|||||||||||||||j|j|tdS)rB;r[)Zzip_safeZuse_2to3Zinclude_package_datar3Zuse_2to3_fixersZuse_2to3_exclude_fixersZconvert_2to3_doctestsZscriptsZeager_resourcesZdependency_linksZnamespace_packagesZinstall_requiresZsetup_requiresZ tests_requireZpackages entry_pointsZ py_modulesZpython_requires)r\rrgrc_parse_packagesrur )r>rZparse_list_semicolonZ parse_boolrr!r!r"rFs.zConfigOptionsHandler.parserscCszddg}|}||kr"||S||dk}|r>ts>td||jdi}|rdddlm}n ddlm }|f|S) zTParses `packages` option value. :param value: :rtype: list zfind:zfind_namespace:r|z8find_namespace: directive is unsupported on Python < 3.3z packages.findr)find_namespace_packages) find_packages) r<r\rr parse_section_packages__findr=rIZ setuptoolsrr)r>r/Zfind_directivesZ trimmed_valueZfindns find_kwargsrr!r!r"r1s     z$ConfigOptionsHandler._parse_packagescsT|||j}dddgtfdd|D}|d}|dk rP|d|d<|S)zParses `packages.find` configuration file section. To be used in conjunction with _parse_packages(). :param dict section_options: whereZincludeexcludecs$g|]\}}|kr|r||fqSr!r!rSkvZ valid_keysr!r"rUZszEConfigOptionsHandler.parse_section_packages__find..Nr)rr\r,r9rI)r>r@Z section_datarrr!rr"rMs   z1ConfigOptionsHandler.parse_section_packages__findcCs|||j}||d<dS)z`Parses `entry_points` configuration file section. :param dict section_options: rN)rr\r>r@rr!r!r"parse_section_entry_pointsbsz/ConfigOptionsHandler.parse_section_entry_pointscCs.|||j}|d}|r*||d<|d=|S)N*r7)rr\rI)r>r@rrootr!r!r"_parse_package_datajs  z(ConfigOptionsHandler._parse_package_datacCs|||d<dS)z`Parses `package_data` configuration file section. :param dict section_options: Z package_dataNrr>r@r!r!r"parse_section_package_datatsz/ConfigOptionsHandler.parse_section_package_datacCs|||d<dS)zhParses `exclude_package_data` configuration file section. :param dict section_options: Zexclude_package_dataNrrr!r!r""parse_section_exclude_package_data{sz7ConfigOptionsHandler.parse_section_exclude_package_datacCs"t|jdd}||||d<dS)zbParses `extras_require` configuration file section. :param dict section_options: rrZextras_requireN)rr\r)r>r@rr!r!r"parse_section_extras_requires z1ConfigOptionsHandler.parse_section_extras_requirecCs(|||j}dd|D|d<dS)z^Parses `data_files` configuration file section. :param dict section_options: cSsg|]\}}||fqSr!r!rr!r!r"rUszAConfigOptionsHandler.parse_section_data_files..Z data_filesN)rr\r9rr!r!r"parse_section_data_filessz-ConfigOptionsHandler.parse_section_data_filesN)rErrr.rrFrrrrrrrrr!r!r!r"r0s   r0)FF)F)"Z __future__rrrwrrrr& collectionsrrr importlibrZdistutils.errorsr r Z#setuptools.extern.packaging.versionr r Z&setuptools.extern.packaging.specifiersr Zsetuptools.extern.sixrrtypeZ __metaclass__r#r+rrr6r1r0r!r!r!r"s4      /  ?Usite-packages/setuptools/__pycache__/launch.cpython-38.opt-1.pyc000064400000001470151464452710020571 0ustar00U Qab@s.dZddlZddlZddZedkr*edS)z[ Launch the Python script on the command line after setuptools is bootstrapped via import. NcCsrttjd}t|ddd}tjddtjdd<ttdt}||}|dd}t ||d}t ||dS) zP Run the script in sys.argv[1] as if it had been invoked naturally. __main__N)__file____name____doc__openz\r\nz\nexec) __builtins__sysargvdictgetattrtokenizerreadreplacecompiler)Z script_name namespaceZopen_ZscriptZ norm_scriptcoder5/usr/lib/python3.8/site-packages/setuptools/launch.pyrun s     rr)rrr rrrrrrs site-packages/setuptools/__pycache__/build_meta.cpython-38.opt-1.pyc000064400000020415151464452710021424 0ustar00U Qab}%@s dZddlZddlZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z ddl mZdddd d d d gZGd d d eZGdddejjZddZddZddZddZGdddeZGdddeZeZejZejZejZejZej Z eZ!dS)a-A PEP 517 interface to setuptools Previously, when a user or a command line tool (let's call it a "frontend") needed to make a request of setuptools to take a certain action, for example, generating a list of installation requirements, the frontend would would call "setup.py egg_info" or "setup.py bdist_wheel" on the command line. PEP 517 defines a different method of interfacing with setuptools. Rather than calling "setup.py" directly, the frontend should: 1. Set the current directory to the directory with a setup.py file 2. Import this module into a safe python interpreter (one in which setuptools can potentially set global variables or crash hard). 3. Call one of the functions defined in PEP 517. What each function does is defined in PEP 517. However, here is a "casual" definition of the functions (this definition should not be relied on for bug reports or API stability): - `build_wheel`: build a wheel in the folder and return the basename - `get_requires_for_build_wheel`: get the `setup_requires` to build - `prepare_metadata_for_build_wheel`: get the `install_requires` - `build_sdist`: build an sdist in the folder and return the basename - `get_requires_for_build_sdist`: get the `setup_requires` to build Again, this is not a formal definition! Just a "taste" of the module. N)TemporaryDirectory)parse_requirements)makedirsget_requires_for_build_sdistget_requires_for_build_wheel prepare_metadata_for_build_wheel build_wheel build_sdist __legacy__SetupRequirementsErrorc@seZdZddZdS)r cCs ||_dSN) specifiers)selfr r9/usr/lib/python3.8/site-packages/setuptools/build_meta.py__init__4szSetupRequirementsError.__init__N)__name__ __module__ __qualname__rrrrrr 3sc@s&eZdZddZeejddZdS) DistributioncCstttt|}t|dSr )listmapstrrr )rr Zspecifier_listrrrfetch_build_eggs9szDistribution.fetch_build_eggsccs*tjj}|tj_z dVW5|tj_XdS)zw Replace distutils.dist.Distribution with this class for the duration of this context. N) distutilsZcorer)clsZorigrrrpatch>s  zDistribution.patchN)rrrr classmethod contextlibcontextmanagerrrrrrr8srcCs*tjddkr&t|ts&|tS|S)z Convert a filename to a string (on Python 2, explicitly a byte string, not Unicode) as distutils checks for the exact type str. r)sys version_info isinstancerencodegetfilesystemencoding)srrr_to_strNsr'csfddtDS)Ncs&g|]}tjtj|r|qSr)ospathisdirjoin).0nameZa_dirrr \sz1_get_immediate_subdirectories..r(listdirr.rr.r_get_immediate_subdirectories[sr2cs"fddt|D}|\}|S)Nc3s|]}|r|VqdSr endswithr,f extensionrr as z'_file_with_extension..r0)Z directoryr8Zmatchingfilerr7r_file_with_extension`s  r;cCs&tj|stdSttdt|S)Nz%from setuptools import setup; setup()open)r(r)existsioStringIOgetattrtokenizer< setup_scriptrrr_open_setup_scriptis  rDc@s`eZdZddZddZdddZdd d Zdd d Zdd dZddZ dddZ dddZ dS)_BuildMetaBackendcCs|pi}|dg|S)N--global-option) setdefaultrconfig_settingsrrr _fix_configss z_BuildMetaBackend._fix_configc Csz||}tjdddg|dt_z t|W5QRXWn,tk rt}z||j7}W5d}~XYnX|S)NZegg_inforF)rJr!argvrr run_setupr r )rrI requirementserrr_get_build_requiresxs  z%_BuildMetaBackend._get_build_requiressetup.pyc CsD|}d}t|}|dd}W5QRXtt||dtdS)N__main__z\r\nz\nexec)rDreadreplacerScompilelocals)rrC__file__rr6coderrrrMs  z_BuildMetaBackend.run_setupNcCs||}|j|dgdS)NZwheelrNrJrPrHrrrrs z._BuildMetaBackend.get_requires_for_build_wheelcCs||}|j|gdS)NrZr[rHrrrrs z._BuildMetaBackend.get_requires_for_build_sdistcCstjddddt|gt_||}ddt|D}t|dkrxtt|dkrxtj |t|d}q*qxq*||krt tj ||d|t j |dd|dS) NrKZ dist_infoz --egg-basecSsg|]}|dr|qS)z .dist-infor3r5rrrr/s zF_BuildMetaBackend.prepare_metadata_for_build_wheel..rT) ignore_errors) r!rLr'rMr(r1lenr2r)r+shutilZmoveZrmtree)rmetadata_directoryrIZdist_info_directoryZ dist_infosrrrrs,  z2_BuildMetaBackend.prepare_metadata_for_build_wheelc Cs||}tj|}t|ddt|dv}tjdd|d|g|dt_|t ||}tj ||}tj |rt |t tj |||W5QRX|S)NT)exist_ok)dirrKz --dist-dirrF)rJr(r)abspathrrr!rLrMr;r+r=removerename)rZ setup_commandZresult_extensionZresult_directoryrIZ tmp_dist_dirZresult_basenameZ result_pathrrr_build_with_temp_dirs         z&_BuildMetaBackend._build_with_temp_dircCs|dgd||S)NZ bdist_wheelz.whlre)rZwheel_directoryrIr_rrrrs z_BuildMetaBackend.build_wheelcCs|dddgd||S)NZsdistz --formatsZgztarz.tar.gzrf)rZsdist_directoryrIrrrr s  z_BuildMetaBackend.build_sdist)rQ)N)N)N)NN)N) rrrrJrPrMrrrrerr rrrrrEqs    rEcs"eZdZdZdfdd ZZS)_BuildMetaLegacyBackendaCCompatibility backend for setuptools This is a version of setuptools.build_meta that endeavors to maintain backwards compatibility with pre-PEP 517 modes of invocation. It exists as a temporary bridge between the old packaging mechanism and the new packaging mechanism, and will eventually be removed. rQc sbttj}tjtj|}|tjkr6tjd|ztt|j |dW5|tjdd<XdS)NrrB) rr!r)r(dirnamerbinsertsuperrgrM)rrCZsys_pathZ script_dir __class__rrrMs   z!_BuildMetaLegacyBackend.run_setup)rQ)rrr__doc__rM __classcell__rrrkrrgsrg)"rmr>r(r!rAr^rZ setuptoolsrZsetuptools.py31compatrZ pkg_resourcesrZpkg_resources.py31compatr__all__ BaseExceptionr Zdistrr'r2r;rDobjectrErgZ_BACKENDrrrrr r rrrrsD     hsite-packages/setuptools/__pycache__/_imp.cpython-38.opt-1.pyc000064400000003542151464452710020245 0ustar00U Qab@s\dZddlZddlZddlZddlmZdZdZdZ dZ dZ dd d Z dd d Z d dZdS)zX Re-implementation of find_module and get_frozen_object from the deprecated imp module. N)module_from_specc CsVtj||}|dkr"td||jsBt|drBtjd|j}d}d}t|jt }|j dkst|rt |jtj j rt}d}d}}n|j dks|rt |jtj jrt}d}d}}n|jr:|j }tj|d }|tj jkrd nd }|tj jkrt}n&|tj jkrt}n|tj jkr t}|tthkrFt||}n d}d}}|||||ffS) z7Just like 'imp.find_module()', but with package supportN Can't find %ssubmodule_search_locationsz __init__.pyfrozenzbuilt-inrrrb) importlibutil find_spec ImportError has_locationhasattrspec_from_loaderloader isinstancetypeorigin issubclass machineryFrozenImporter PY_FROZENBuiltinImporter C_BUILTINospathsplitextSOURCE_SUFFIXES PY_SOURCEBYTECODE_SUFFIXES PY_COMPILEDEXTENSION_SUFFIXES C_EXTENSIONopen) modulepathsspecZkindfileZstaticr!suffixmoder03/usr/lib/python3.8/site-packages/setuptools/_imp.py find_modulesJ      r2cCs*tj||}|std||j|SNr)rrrrrget_code)r*r+r,r0r0r1get_frozen_object>s r5cCs&tj||}|std|t|Sr3)rrrrr)r*r+infor,r0r0r1 get_moduleEs r7)N)N)__doc__r importlib.utilrZimportlib.machineryZ py34compatrr$r&r(rrr2r5r7r0r0r0r1s  * site-packages/setuptools/__pycache__/lib2to3_ex.cpython-38.opt-1.pyc000064400000004553151464452710021276 0ustar00U Qab@sXdZddlmZddlmZddlmZmZddl Z GdddeZ Gdd d eZdS) zy Customized Mixin2to3 support: - adds support for converting doctests This module raises an ImportError on Python 2. ) Mixin2to3)log)RefactoringToolget_fixers_from_packageNc@s$eZdZddZddZddZdS)DistutilsRefactoringToolcOstj|f|dSN)rerror)selfmsgargskwr 9/usr/lib/python3.8/site-packages/setuptools/lib2to3_ex.py log_errorsz"DistutilsRefactoringTool.log_errorcGstj|f|dSr)rinfor r r r r r log_messagesz$DistutilsRefactoringTool.log_messagecGstj|f|dSr)rdebugrr r r log_debugsz"DistutilsRefactoringTool.log_debugN)__name__ __module__ __qualname__rrrr r r rrsrc@s&eZdZd ddZddZddZdS) rFcCsr|jjdk rdS|sdStdd||||rbtjrnt |j }|j |dddn t ||dS)NTzFixing  )writeZ doctests_only) distributionZuse_2to3rrjoin_Mixin2to3__build_fixer_names_Mixin2to3__exclude_fixers setuptoolsZrun_2to3_on_doctestsr fixer_namesZrefactor _Mixin2to3run_2to3)r filesZdoctestsrr r rr!s  zMixin2to3.run_2to3cCsZ|jr dSg|_tjD]}|jt|q|jjdk rV|jjD]}|jt|q@dSr)rrZlib2to3_fixer_packagesextendrrZuse_2to3_fixers)r pr r rZ__build_fixer_names.s   zMixin2to3.__build_fixer_namescCsJt|dg}|jjdk r&||jj|D]}||jkr*|j|q*dS)NZexclude_fixers)getattrrZuse_2to3_exclude_fixersr$rremove)r Zexcluded_fixersZ fixer_namer r rZ__exclude_fixers8s    zMixin2to3.__exclude_fixersN)F)rrrr!rrr r r rrs  r) __doc__Zdistutils.utilrr Z distutilsrZlib2to3.refactorrrrrr r r rs    site-packages/setuptools/__pycache__/windows_support.cpython-38.opt-1.pyc000064400000001735151464452710022611 0ustar00U Qab@s(ddlZddlZddZeddZdS)NcCstdkrddS|S)NZWindowsc_sdS)N)argskwargsrr>/usr/lib/python3.8/site-packages/setuptools/windows_support.pyzwindows_only..)platformsystem)funcrrr windows_onlys r cCsLtdtjjj}tjjtjjf|_tjj |_ d}|||}|sHt dS)z Set the hidden attribute on a file or directory. From http://stackoverflow.com/questions/19622133/ `path` must be text. zctypes.wintypesN) __import__ctypesZwindllZkernel32ZSetFileAttributesWZwintypesZLPWSTRZDWORDZargtypesZBOOLZrestypeZWinError)pathZSetFileAttributesFILE_ATTRIBUTE_HIDDENretrrr hide_file s    r)rrr rrrrrssite-packages/setuptools/__pycache__/py27compat.cpython-38.opt-1.pyc000064400000003325151464452710021325 0ustar00U Qab@sdZddlZddlZddlmZddZejr6ddZedkoFejZerPe ndd Z z,d d l m Z m Z mZmZd d l mZmZWnJek rddlZdd lm Z mZmZdddZ ddZddZYnXdS)z2 Compatibility Support for Python 2.7 and earlier N)sixcCs ||S)zH Given an HTTPMessage, return all headers matching a given key. )Zget_allmessagekeyr9/usr/lib/python3.8/site-packages/setuptools/py27compat.pyget_all_headers srcCs ||SN)Z getheadersrrrrrsZLinuxcCs|Sr r)xrrrr ) find_module PY_COMPILED PY_FROZEN PY_SOURCE)get_frozen_object get_module)rrrc Csj|d}|rf|d}t||\}}\}}}} |tjkrP|pFdg}|g}q |r td||fq | S)z7Just like 'imp.find_module()', but with package support.r__init__zCan't find %r in %s)splitpopimprZ PKG_DIRECTORY ImportError) modulepathspartspartfpathsuffixmodeZkindinforrrr's    rcCs t|Sr )rr)rrrrrr7srcCstj|f|tj|Sr )r load_modulesysmodules)rrr"rrrr:sr)N)__doc__r$platformZsetuptools.externrrZPY2systemZlinux_py2_asciistrZ rmtree_safe_imprrrrrrrrrrrrs&   site-packages/setuptools/__pycache__/_deprecation_warning.cpython-38.opt-1.pyc000064400000001004151464452710023471 0ustar00U Qab@sGdddeZdS)c@seZdZdZdS)SetuptoolsDeprecationWarningz Base class for warning deprecations in ``setuptools`` This class is not derived from ``DeprecationWarning``, and as such is visible by default. N)__name__ __module__ __qualname____doc__rrC/usr/lib/python3.8/site-packages/setuptools/_deprecation_warning.pyrsrN)Warningrrrrrsite-packages/setuptools/__pycache__/lib2to3_ex.cpython-38.pyc000064400000004553151464452710020337 0ustar00U Qab@sXdZddlmZddlmZddlmZmZddl Z GdddeZ Gdd d eZdS) zy Customized Mixin2to3 support: - adds support for converting doctests This module raises an ImportError on Python 2. ) Mixin2to3)log)RefactoringToolget_fixers_from_packageNc@s$eZdZddZddZddZdS)DistutilsRefactoringToolcOstj|f|dSN)rerror)selfmsgargskwr 9/usr/lib/python3.8/site-packages/setuptools/lib2to3_ex.py log_errorsz"DistutilsRefactoringTool.log_errorcGstj|f|dSr)rinfor r r r r r log_messagesz$DistutilsRefactoringTool.log_messagecGstj|f|dSr)rdebugrr r r log_debugsz"DistutilsRefactoringTool.log_debugN)__name__ __module__ __qualname__rrrr r r rrsrc@s&eZdZd ddZddZddZdS) rFcCsr|jjdk rdS|sdStdd||||rbtjrnt |j }|j |dddn t ||dS)NTzFixing  )writeZ doctests_only) distributionZuse_2to3rrjoin_Mixin2to3__build_fixer_names_Mixin2to3__exclude_fixers setuptoolsZrun_2to3_on_doctestsr fixer_namesZrefactor _Mixin2to3run_2to3)r filesZdoctestsrr r rr!s  zMixin2to3.run_2to3cCsZ|jr dSg|_tjD]}|jt|q|jjdk rV|jjD]}|jt|q@dSr)rrZlib2to3_fixer_packagesextendrrZuse_2to3_fixers)r pr r rZ__build_fixer_names.s   zMixin2to3.__build_fixer_namescCsJt|dg}|jjdk r&||jj|D]}||jkr*|j|q*dS)NZexclude_fixers)getattrrZuse_2to3_exclude_fixersr$rremove)r Zexcluded_fixersZ fixer_namer r rZ__exclude_fixers8s    zMixin2to3.__exclude_fixersN)F)rrrr!rrr r r rrs  r) __doc__Zdistutils.utilrr Z distutilsrZlib2to3.refactorrrrrr r r rs    site-packages/setuptools/__pycache__/windows_support.cpython-38.pyc000064400000001735151464452710021652 0ustar00U Qab@s(ddlZddlZddZeddZdS)NcCstdkrddS|S)NZWindowsc_sdS)N)argskwargsrr>/usr/lib/python3.8/site-packages/setuptools/windows_support.pyzwindows_only..)platformsystem)funcrrr windows_onlys r cCsLtdtjjj}tjjtjjf|_tjj |_ d}|||}|sHt dS)z Set the hidden attribute on a file or directory. From http://stackoverflow.com/questions/19622133/ `path` must be text. zctypes.wintypesN) __import__ctypesZwindllZkernel32ZSetFileAttributesWZwintypesZLPWSTRZDWORDZargtypesZBOOLZrestypeZWinError)pathZSetFileAttributesFILE_ATTRIBUTE_HIDDENretrrr hide_file s    r)rrr rrrrrssite-packages/setuptools/__pycache__/ssl_support.cpython-38.opt-1.pyc000064400000015307151464452710021720 0ustar00U Qab-! @sddlZddlZddlZddlZddlZddlmZmZmZm Z ddl m Z m Z z ddl Z Wnek rtdZ YnXdddddgZd ZzejjZejZWnek reZZYnXe dk oeeefkZzdd l mZmZWnRek r:zdd lmZdd lmZWnek r4dZdZYnXYnXesRGd ddeZesjdddZddZGdddeZGdddeZd ddZ ddZ!e!ddZ"ddZ#ddZ$dS)!N)urllib http_clientmapfilter)ResolutionErrorExtractionErrorVerifyingHTTPSHandlerfind_ca_bundle is_available cert_paths opener_fora /etc/pki/tls/certs/ca-bundle.crt /etc/ssl/certs/ca-certificates.crt /usr/share/ssl/certs/ca-bundle.crt /usr/local/share/certs/ca-root.crt /etc/ssl/cert.pem /System/Library/OpenSSL/certs/cert.pem /usr/local/share/certs/ca-root-nss.crt /etc/ssl/ca-bundle.pem )CertificateErrormatch_hostname)r )rc@s eZdZdS)r N)__name__ __module__ __qualname__rr:/usr/lib/python3.8/site-packages/setuptools/ssl_support.pyr 5sr c Csg}|s dS|d}|d}|dd}|d}||krLtdt||s`||kS|dkrt|dn>|d s|d r|t|n|t| d d |D]}|t|qt d d |dtj } | |S)zqMatching according to RFC 6125, section 6.4.3 https://tools.ietf.org/html/rfc6125#section-6.4.3 F.rrN*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)splitcountr reprlowerappend startswithreescapereplacecompilejoin IGNORECASEmatch) ZdnhostnameZ max_wildcardsZpatspartsZleftmostZ remainderZ wildcardsZfragZpatrrr_dnsname_match;s,     r&cCs|s tdg}|dd}|D]*\}}|dkr t||r@dS||q |s|ddD]6}|D],\}}|dkrdt||rdS||qdq\t|dkrtd |d tt|fn*t|dkrtd ||d fntd dS)a=Verify that *cert* (in decoded format as returned by SSLSocket.getpeercert()) matches the *hostname*. RFC 2818 and RFC 6125 rules are followed, but IP addresses are not accepted for *hostname*. CertificateError is raised on failure. On success, the function returns nothing. zempty or no certificateZsubjectAltNamerZDNSNZsubjectZ commonNamerz&hostname %r doesn't match either of %sz, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found) ValueErrorgetr&rlenr r!rr)Zcertr$ZdnsnamesZsankeyvaluesubrrrros2         rc@s eZdZdZddZddZdS)rz=Simple verifying handler: no auth, subclasses, timeouts, etc.cCs||_t|dSN) ca_bundle HTTPSHandler__init__)selfr.rrrr0szVerifyingHTTPSHandler.__init__csfdd|S)Ncst|jf|Sr-)VerifyingHTTPSConnr.)hostkwr1rrz2VerifyingHTTPSHandler.https_open..)Zdo_open)r1Zreqrr5r https_opens z VerifyingHTTPSHandler.https_openN)rrr__doc__r0r8rrrrrsc@s eZdZdZddZddZdS)r2z@Simple verifying connection: no auth, subclasses, timeouts, etc.cKstj||f|||_dSr-)HTTPSConnectionr0r.)r1r3r.r4rrrr0szVerifyingHTTPSConn.__init__cCst|j|jft|dd}t|drHt|ddrH||_||j}n|j}tt drxt j |j d}|j ||d|_nt j |t j |j d|_zt|j|Wn.tk r|jtj|jYnXdS)NZsource_address_tunnel _tunnel_hostcreate_default_context)Zcafile)Zserver_hostname)Z cert_reqsZca_certs)socketZcreate_connectionr3Zportgetattrhasattrsockr;r<sslr=r.Z wrap_socketZ CERT_REQUIREDrZ getpeercertr ZshutdownZ SHUT_RDWRclose)r1rAZ actual_hostZctxrrrconnects.   zVerifyingHTTPSConn.connectN)rrrr9r0rDrrrrr2sr2cCstjt|ptjS)z@Get a urlopen() replacement that uses ca_bundle for verification)rrequestZ build_openerrr open)r.rrrr s cstfdd}|S)Ncstds||_jS)Nalways_returns)r@rG)argskwargsfuncrrwrappers  zonce..wrapper) functoolswraps)rKrLrrJroncesrOcsZz ddl}Wntk r"YdSXGfddd|j}|d|d|jS)Nrcs,eZdZfddZfddZZS)z"get_win_certfile..CertFilecst|t|jdSr-)superr0atexitregisterrCr5CertFile __class__rrr0sz+get_win_certfile..CertFile.__init__cs,zt|Wntk r&YnXdSr-)rPrCOSErrorr5rSrrrCsz(get_win_certfile..CertFile.close)rrrr0rC __classcell__rrT)rUrrTsrTZCAZROOT) wincertstore ImportErrorrTZaddstorename)rYZ _wincertsrrXrget_win_certfiles    r\cCs$ttjjt}tp"t|dp"tS)z*Return an existing CA bundle path, or NoneN)rospathisfiler r\next_certifi_where)Zextant_cert_pathsrrrr s c Cs.ztdWStttfk r(YnXdS)NZcertifi) __import__whererZrrrrrrrasra)r)N)%r]r>rQrrMZsetuptools.extern.six.movesrrrrZ pkg_resourcesrrrBrZ__all__striprr rEr/r:AttributeErrorobjectr r rZbackports.ssl_match_hostnamer'r&rr2r rOr\r rarrrrsZ      4) (    site-packages/setuptools/__pycache__/pep425tags.cpython-38.opt-1.pyc000064400000016043151464452710021217 0ustar00U Qabm*@sdZddlmZddlZddlmZddlZddlZddlZddl Z ddl Z ddl m Z ddl mZddlmZed Zd d Zd d ZddZddZddZd$ddZddZddZddZddZdd Zd%d"d#ZeZdS)&z2Generate and work with PEP 425 Compatibility Tags.)absolute_importN)log) OrderedDict)six)glibcz(.+)_(\d+)_(\d+)_(.+)c CsLz t|WStk rF}ztd|tWYdSd}~XYnXdS)Nz{}) sysconfigget_config_varIOErrorwarningswarnformatRuntimeWarning)varer9/usr/lib/python3.8/site-packages/setuptools/pep425tags.pyr s  r cCs:ttdrd}n&tjdr"d}ntjdkr2d}nd}|S)z'Return abbreviated implementation name.pypy_version_infoppjavaZjyZcliZipcp)hasattrsysplatform startswith)Zpyimplrrr get_abbr_impls   rcCs,td}|rtdkr(dttt}|S)zReturn implementation version.Zpy_version_nodotr)r rjoinmapstrget_impl_version_info)Zimpl_verrrr get_impl_ver+sr!cCs:tdkr"tjdtjjtjjfStjdtjdfSdS)zQReturn sys.version_info-like tuple for use in decrementing the minor version.rrrN)rr version_informajorminorrrrrr 3s  r cCsdttS)z; Returns the Tag for this specific implementation. z{}{})r rr!rrrr get_impl_tag>sr%TcCs.t|}|dkr&|r td||S||kS)zgUse a fallback method for determining SOABI flags if the needed config var is unset or unavailable.Nz>Config variable '%s' is unset, Python ABI tag may be incorrect)r rdebug)rZfallbackexpectedr valrrrget_flagEsr)cstd}t|sdkrttdrd}d}d}tddddkd rJd }td fd ddkd rhd }tdddddkotjdrtjrd}dt|||f}n@|r|drd| dd}n|r| dd dd}nd}|S)zXReturn the ABI tag based on SOABI (if available) or emulate SOABI (CPython 2, PyPy).ZSOABI>rr maxunicoderZPy_DEBUGcSs ttdS)NZgettotalrefcount)rrrrrr[zget_abi_tag..r)r dZ WITH_PYMALLOCcsdkS)Nrrrimplrrr+_r,mZPy_UNICODE_SIZEcSs tjdkS)Ni)rr*rrrrr+cr,)r'r uz %s%s%s%s%szcpython--r._N) r rrrr)rZPY2r!rsplitreplace)Zsoabir-r0r2abirr.r get_abi_tagQs@ r9cCs tjdkS)Ni)rmaxsizerrrr_is_running_32bitssr;cCstjdkr^t\}}}|d}|dkr6tr6d}n|dkrHtrHd}d|d|d |Stj dd  d d }|d krtrd }|S)z0Return our platform name 'win32', 'linux_x86_64'darwinr4x86_64i386ppc64ppczmacosx_{}_{}_{}rrr5r3 linux_x86_64 linux_i686) rrZmac_verr6r;r distutilsutil get_platformr7)releaser5machineZ split_verresultrrrrEws  rEc CsHtdkrdSzddl}t|jWSttfk r:YnXtddS)N>rArBFr)rE _manylinuxboolZmanylinux1_compatible ImportErrorAttributeErrorrZhave_compatible_glibc)rKrrris_manylinux1_compatibles  rOcsrg}fddtddddg|||r8||D]&}||kr<|||r<||q<|d|S)zReturn a list of supported arches (including group arches) for the given major, minor and machine architecture of a macOS machine. cs||dkr||fdkS|dkr(||fdkS|dkr<||fdkS|dkrP||fdkS|krx|D]}|||r`dSq`dS) Nr@) rJr?r>)rPr1r=TFr)r#r$archgarch_supports_archgroupsrrrTs      z)get_darwin_arches.._supports_arch)Zfat)r>r@)Zintel)r=r>)Zfat64)r=r?)Zfat32)r=r>r@Z universal)rappend)r#r$rGarchesrRrrSrget_darwin_archess$    rXFc Csg}|dkrTg}t}|dd}t|dddD] }|dtt||fq2|p\t}g} |pjt}|r~|g| dd<t} ddl } | D],} | d dr| | d dddq| tt| | d |sT|pt} | d rzt| }|rr|\}}}}d ||}g}ttt|dD]0}tt|||D]}||||fqRq>n| g}n*|dkrtr| d d | g}n| g}| D].}|D]"} |d||df|| fqq|ddD]F}|dkrq,| D]*}|D]} |d||f|| fqqq|D]"} |d|ddd | fq0|d||dfd df|d||ddfd dft|D]B\}}|d|fd df|dkr|d|dd dfq|S)acReturn a list of supported tags for each version specified in `versions`. :param versions: a list of string versions, of the form ["33", "32"], or None. The first version will be assumed to support our ABI. :param platform: specify the exact platform you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. :param abi: specify the exact abi you want valid tags for, or None. If None, use the local interpreter abi. Nrrz.abir4rIrZnoneZmacosxz {}_{}_%i_%sZlinuxZ manylinux1z%s%s>3031zpy%sany)r rangerVrrrrr9setimpZ get_suffixesraddr6extendsortedlistrE _osx_arch_patmatchrUr reversedintrXrOr7 enumerate)ZversionsZnoarchrr/r8Z supportedr"r#r$ZabisZabi3sr_suffixrQrenameZ actual_archZtplrWr0aversionirrr get_supportedsh         $ $   rn)TT)NFNNN) __doc__Z __future__rZdistutils.utilrCrrrerrr collectionsrZexternrrrcompilerdr rr!r r%r)r9r;rErOrXrnZimplementation_tagrrrrs8         "= `site-packages/setuptools/__pycache__/glob.cpython-38.opt-1.pyc000064400000007143151464452710020245 0ustar00U Qab@sdZddlZddlZddlZdddgZdddZdddZd d Zd d Zd dZ ddZ ddZ e dZ e dZddZddZddZdS)z Filename globbing utility. Mostly a copy of `glob` from Python 3.5. Changes include: * `yield from` and PEP3102 `*` removed. * Hidden files are not ignored. NglobiglobescapeFcCstt||dS)ayReturn a list of paths matching a pathname pattern. The pattern may contain simple shell-style wildcards a la fnmatch. However, unlike fnmatch, filenames starting with a dot are special cases that are not matched by '*' and '?' patterns. If recursive is true, the pattern '**' will match any files and zero or more directories and subdirectories. ) recursive)listr)pathnamerr3/usr/lib/python3.8/site-packages/setuptools/glob.pyrs cCs"t||}|rt|rt|}|S)aReturn an iterator which yields the paths matching a pathname pattern. The pattern may contain simple shell-style wildcards a la fnmatch. However, unlike fnmatch, filenames starting with a dot are special cases that are not matched by '*' and '?' patterns. If recursive is true, the pattern '**' will match any files and zero or more directories and subdirectories. )_iglob _isrecursivenext)rritsrrr rs  ccstj|\}}t|sF|r0tj|rB|Vntj|rB|VdS|s|rnt|rnt||D] }|Vq`nt||D] }|VqxdS||krt|rt ||}n|g}t|r|rt|rt}qt}nt }|D]$}|||D]}tj ||VqqdSN) ospathsplit has_magiclexistsisdirr glob2glob1r glob0join)rrdirnamebasenamexdirsZ glob_in_dirnamerrr r 0s4      r cCsV|s"t|trtjd}ntj}zt|}Wntk rHgYSXt||SNASCII) isinstancebytesrcurdirencodelistdirOSErrorfnmatchfilter)rpatternnamesrrr r[s  rcCs8|stj|r4|gSntjtj||r4|gSgSr)rrrrr)rrrrr rhs  rccs&|ddVt|D] }|VqdS)Nr) _rlistdir)rr)rrrr rxs rccs|s"t|trtjd}ntj}zt|}Wntjk rHYdSX|D]>}|V|rjtj||n|}t |D]}tj||VqvqNdSr) r!r"rr#r$r%errorrrr+)rr*rryrrr r+s  r+z([*?[])s([*?[])cCs(t|trt|}n t|}|dk Sr)r!r"magic_check_bytessearch magic_check)rmatchrrr rs   rcCst|tr|dkS|dkSdS)Ns**z**)r!r")r)rrr r s r cCs<tj|\}}t|tr(td|}n td|}||S)z#Escape all special characters. s[\1]z[\1])rr splitdriver!r"r.subr0)rZdriverrr rs   )F)F)__doc__rrer'__all__rrr rrrr+compiler0r.rr rrrrr s    +   site-packages/setuptools/__pycache__/build_meta.cpython-38.pyc000064400000020457151464452710020473 0ustar00U Qab}%@s dZddlZddlZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z ddl mZdddd d d d gZGd d d eZGdddejjZddZddZddZddZGdddeZGdddeZeZejZejZejZejZej Z eZ!dS)a-A PEP 517 interface to setuptools Previously, when a user or a command line tool (let's call it a "frontend") needed to make a request of setuptools to take a certain action, for example, generating a list of installation requirements, the frontend would would call "setup.py egg_info" or "setup.py bdist_wheel" on the command line. PEP 517 defines a different method of interfacing with setuptools. Rather than calling "setup.py" directly, the frontend should: 1. Set the current directory to the directory with a setup.py file 2. Import this module into a safe python interpreter (one in which setuptools can potentially set global variables or crash hard). 3. Call one of the functions defined in PEP 517. What each function does is defined in PEP 517. However, here is a "casual" definition of the functions (this definition should not be relied on for bug reports or API stability): - `build_wheel`: build a wheel in the folder and return the basename - `get_requires_for_build_wheel`: get the `setup_requires` to build - `prepare_metadata_for_build_wheel`: get the `install_requires` - `build_sdist`: build an sdist in the folder and return the basename - `get_requires_for_build_sdist`: get the `setup_requires` to build Again, this is not a formal definition! Just a "taste" of the module. N)TemporaryDirectory)parse_requirements)makedirsget_requires_for_build_sdistget_requires_for_build_wheel prepare_metadata_for_build_wheel build_wheel build_sdist __legacy__SetupRequirementsErrorc@seZdZddZdS)r cCs ||_dSN) specifiers)selfr r9/usr/lib/python3.8/site-packages/setuptools/build_meta.py__init__4szSetupRequirementsError.__init__N)__name__ __module__ __qualname__rrrrrr 3sc@s&eZdZddZeejddZdS) DistributioncCstttt|}t|dSr )listmapstrrr )rr Zspecifier_listrrrfetch_build_eggs9szDistribution.fetch_build_eggsccs*tjj}|tj_z dVW5|tj_XdS)zw Replace distutils.dist.Distribution with this class for the duration of this context. N) distutilsZcorer)clsZorigrrrpatch>s  zDistribution.patchN)rrrr classmethod contextlibcontextmanagerrrrrrr8srcCs*tjddkr&t|ts&|tS|S)z Convert a filename to a string (on Python 2, explicitly a byte string, not Unicode) as distutils checks for the exact type str. r)sys version_info isinstancerencodegetfilesystemencoding)srrr_to_strNsr'csfddtDS)Ncs&g|]}tjtj|r|qSr)ospathisdirjoin).0nameZa_dirrr \sz1_get_immediate_subdirectories..r(listdirr.rr.r_get_immediate_subdirectories[sr2cs"fddt|D}|\}|S)Nc3s|]}|r|VqdSr endswithr,f extensionrr as z'_file_with_extension..r0)Z directoryr8Zmatchingfilerr7r_file_with_extension`s  r;cCs&tj|stdSttdt|S)Nz%from setuptools import setup; setup()open)r(r)existsioStringIOgetattrtokenizer< setup_scriptrrr_open_setup_scriptis  rDc@s`eZdZddZddZdddZdd d Zdd d Zdd dZddZ dddZ dddZ dS)_BuildMetaBackendcCs|pi}|dg|S)N--global-option) setdefaultrconfig_settingsrrr _fix_configss z_BuildMetaBackend._fix_configc Csz||}tjdddg|dt_z t|W5QRXWn,tk rt}z||j7}W5d}~XYnX|S)NZegg_inforF)rJr!argvrr run_setupr r )rrI requirementserrr_get_build_requiresxs  z%_BuildMetaBackend._get_build_requiressetup.pyc CsD|}d}t|}|dd}W5QRXtt||dtdS)N__main__z\r\nz\nexec)rDreadreplacerScompilelocals)rrC__file__rr6coderrrrMs  z_BuildMetaBackend.run_setupNcCs||}|j|dgdS)NZwheelrNrJrPrHrrrrs z._BuildMetaBackend.get_requires_for_build_wheelcCs||}|j|gdS)NrZr[rHrrrrs z._BuildMetaBackend.get_requires_for_build_sdistcCstjddddt|gt_||}ddt|D}t|dkrttt|dkrttj |t|d}q*t|dkst qq*||krt tj ||d|t j |dd|dS) NrKZ dist_infoz --egg-basecSsg|]}|dr|qS)z .dist-infor3r5rrrr/s zF_BuildMetaBackend.prepare_metadata_for_build_wheel..rT) ignore_errors)r!rLr'rMr(r1lenr2r)r+AssertionErrorshutilZmoveZrmtree)rmetadata_directoryrIZdist_info_directoryZ dist_infosrrrrs.  z2_BuildMetaBackend.prepare_metadata_for_build_wheelc Cs||}tj|}t|ddt|dv}tjdd|d|g|dt_|t ||}tj ||}tj |rt |t tj |||W5QRX|S)NT)exist_ok)dirrKz --dist-dirrF)rJr(r)abspathrrr!rLrMr;r+r=removerename)rZ setup_commandZresult_extensionZresult_directoryrIZ tmp_dist_dirZresult_basenameZ result_pathrrr_build_with_temp_dirs         z&_BuildMetaBackend._build_with_temp_dircCs|dgd||S)NZ bdist_wheelz.whlrf)rZwheel_directoryrIr`rrrrs z_BuildMetaBackend.build_wheelcCs|dddgd||S)NZsdistz --formatsZgztarz.tar.gzrg)rZsdist_directoryrIrrrr s  z_BuildMetaBackend.build_sdist)rQ)N)N)N)NN)N) rrrrJrPrMrrrrfrr rrrrrEqs    rEcs"eZdZdZdfdd ZZS)_BuildMetaLegacyBackendaCCompatibility backend for setuptools This is a version of setuptools.build_meta that endeavors to maintain backwards compatibility with pre-PEP 517 modes of invocation. It exists as a temporary bridge between the old packaging mechanism and the new packaging mechanism, and will eventually be removed. rQc sbttj}tjtj|}|tjkr6tjd|ztt|j |dW5|tjdd<XdS)NrrB) rr!r)r(dirnamercinsertsuperrhrM)rrCZsys_pathZ script_dir __class__rrrMs   z!_BuildMetaLegacyBackend.run_setup)rQ)rrr__doc__rM __classcell__rrrlrrhsrh)"rnr>r(r!rAr_rZ setuptoolsrZsetuptools.py31compatrZ pkg_resourcesrZpkg_resources.py31compatr__all__ BaseExceptionr Zdistrr'r2r;rDobjectrErhZ_BACKENDrrrrr r rrrrsD     hsite-packages/setuptools/__pycache__/wheel.cpython-38.opt-1.pyc000064400000015564151464452710020434 0ustar00U Qab@sdZddlmZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl m Z ddl m Z ddlmZddl mZddlmZeZed ejjZd Zd d ZGd ddZdS)zWheels support.) get_platformN) parse_version)canonicalize_name)PY3) pep425tags)write_requirementsz^(?P.+?)-(?P\d.*?) ((-(?P\d.*?))?-(?P.+?)-(?P.+?)-(?P.+?) )\.whl$ztry: __import__('pkg_resources').declare_namespace(__name__) except ImportError: __path__ = __import__('pkgutil').extend_path(__path__, __name__) c Cst|D]\}}}tj||}|D].}tj||}tj|||}t||q&ttt|D]D\} } tj|| }tj||| }tj |sft|||| =qfq tj|ddD]\}}}t |qdS)zDMove everything under `src_dir` to `dst_dir`, and delete the former.T)topdownN) oswalkpathrelpathjoinrenamesreversedlist enumerateexistsrmdir) Zsrc_dirZdst_dirdirpathZdirnames filenamessubdirfsrcZdstndr4/usr/lib/python3.8/site-packages/setuptools/wheel.pyunpack%s   rc@sheZdZddZddZddZddZd d Zd d Zd dZ e ddZ e ddZ e ddZ dS)WheelcCsPttj|}|dkr$td|||_|D]\}}t|||q6dS)Nzinvalid wheel name: %r) WHEEL_NAMEr r basename ValueErrorfilename groupdictitemssetattr)selfr"matchkvrrr__init__=s  zWheel.__init__cCs&t|jd|jd|jdS)z>List tags (py_version, abi, platform) supported by this wheel..) itertoolsproductZ py_versionsplitZabiplatformr&rrrtagsEs    z Wheel.tagscs$ttfdd|DdS)z5Is the wheel is compatible with the current platform?c3s|]}|krdVqdS)TNr).0tZsupported_tagsrr Psz&Wheel.is_compatible..F)rZ get_supportednextr1r0rr4r is_compatibleMszWheel.is_compatiblecCs,tj|j|j|jdkrdntddS)Nany) project_nameversionr/z.egg) pkg_resources Distributionr9r:r/regg_namer0rrrr=RszWheel.egg_namecCsJ|D]4}t|}|drt|t|jr|SqtddS)Nz .dist-infoz.unsupported wheel format. .dist-info not found)Znamelist posixpathdirnameendswithr startswithr9r!)r&zfmemberr?rrr get_dist_infoXs    zWheel.get_dist_infoc Cs(t|j}|||W5QRXdS)z"Install wheel as an egg directory.N)zipfileZZipFiler"_install_as_egg)r&destination_eggdirrBrrrinstall_as_eggbszWheel.install_as_eggcCs\d|j|jf}||}d|}tj|d}|||||||||||dS)Nz%s-%sz%s.dataEGG-INFO) r9r:rDr r r _convert_metadata_move_data_entries_fix_namespace_packages)r&rGrBZ dist_basename dist_info dist_dataegg_inforrrrFgs  zWheel._install_as_eggc s&fdd}|d}t|d}td|ko>tdkn}|sTtd|t||tj|tj j |t |dd d t t tfd d jD}t|ttj|d tj|dtj t|dd} t| ddtj|ddS)Nc sTt|8}tr&|dn|}tj |W5QRSQRXdS)Nzutf-8) openr>r rreaddecodeemailparserZParserZparsestr)namefpvalue)rMrBrr get_metadatassz-Wheel._convert_metadata..get_metadataZWHEELz Wheel-Versionz1.0z2.0dev0z$unsupported wheel format version: %s)ZmetadatacSsd|_t|SN)Zmarkerstr)reqrrrraw_reqsz(Wheel._convert_metadata..raw_reqc s2i|]*}|tfddt|fDqS)c3s|]}|kr|VqdSrYr)r2r[)install_requiresrrr5sz5Wheel._convert_metadata...)sortedmaprequires)r2Zextra)distr]r\rr s  z+Wheel._convert_metadata..ZMETADATAzPKG-INFO)r]extras_require)ZattrsrOz requires.txt)rgetr!r mkdirZ extractallr r r;r<Z from_locationZ PathMetadatarr^r_r`Zextrasrename setuptoolsdictrZget_command_obj) rBrGrMrOrXZwheel_metadataZ wheel_versionZwheel_v1rcZ setup_distr)rarMr]r\rBrrJqsL       zWheel._convert_metadatacstj|tjd}tj|rtj|dd}t|t|D]D}|drpttj||qLttj||tj||qLt |t tjjfdddDD]}t ||qtjrt dS)z,Move data entries to their correct location.ZscriptsrIz.pycc3s|]}tj|VqdSrY)r r r )r2rrNrrr5sz+Wheel._move_data_entries..)dataZheadersZpurelibZplatlibN) r r r rrelistdirr@unlinkrfrfilterr)rGrNZdist_data_scriptsZegg_info_scriptsentryrrrirrKs.         zWheel._move_data_entriesc Cstj|d}tj|rt|}|}W5QRX|D]b}tjj|f|d}tj|d}tj|r>tj|s>t|d}|tW5QRXq>dS)Nznamespace_packages.txtr+z __init__.pyw) r r r rrPrQr.writeNAMESPACE_PACKAGE_INIT)rOrGZnamespace_packagesrVmodZmod_dirZmod_initrrrrLs   zWheel._fix_namespace_packagesN)__name__ __module__ __qualname__r*r1r7r=rDrHrF staticmethodrJrKrLrrrrr;s   9 r)__doc__Zdistutils.utilrrSr,r r>rerEr;rgrZ!setuptools.extern.packaging.utilsrZsetuptools.extern.sixrrZsetuptools.command.egg_infortypeZ __metaclass__compileVERBOSEr'rrqrrrrrrs,      site-packages/setuptools/__pycache__/__init__.cpython-38.opt-1.pyc000064400000017130151464452710021056 0ustar00U Qabs@sdZddlZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl m Z ddlmZddlmZmZdd lmZmZddlZdd lmZdd lmZmZdd lmZdd lm Z e!Z"ddddddddgZ#ere#$dej%j&Z&dZ'dZ(dgZ)GdddZ*Gddde*Z+e*j,Z-er,e+j,Z.ddZ/ddZ0ej1j0je0_e 2ej1j3Z4Gd dde4Z3d!d"Z5ej6fd#d$Z7e 8dS)%z@Extensions to the 'distutils' for large or complex distributionsN)DistutilsOptionError) convert_path fnmatchcase)SetuptoolsDeprecationWarning)PY3 string_types)filtermap) Extension) DistributionFeature)Require)monkeysetupr rCommandr rr find_packagesfind_namespace_packagesTz lib2to3.fixesc@sBeZdZdZedddZeddZed d Zed d Z d S) PackageFinderzI Generate a list of all Python packages found within a directory .*cCs&t|t||jd||j|S)a Return a list all Python packages found within directory 'where' 'where' is the root directory which will be searched for packages. It should be supplied as a "cross-platform" (i.e. URL-style) path; it will be converted to the appropriate local path syntax. 'exclude' is a sequence of package names to exclude; '*' can be used as a wildcard in the names, such that 'foo.*' will exclude all subpackages of 'foo' (but not 'foo' itself). 'include' is a sequence of package names to include. If it's specified, only the named packages will be included. If it's not specified, all found packages will be included. 'include' can contain shell style wildcard patterns just like 'exclude'. ez_setup *__pycache__)rr)list_find_packages_iterr _build_filter)clswhereexcludeincluderr7/usr/lib/python3.8/site-packages/setuptools/__init__.pyfind4s  zPackageFinder.findc cstj|ddD]\}}}|dd}g|dd<|D]d}tj||} tj| |} | tjjd} d|ks4|| sxq4|| r|| s| V||q4qdS)zy All the packages found in 'where' that pass the 'include' filter, but not the 'exclude' filter. T followlinksNr) oswalkpathjoinrelpathreplacesep_looks_like_packageappend) rr r!r"rootdirsfilesZall_dirsdir full_pathZrel_pathpackagerrr#rKs  z!PackageFinder._find_packages_itercCstjtj|dS)z%Does a directory look like a package?z __init__.py)r'r)isfiler*r)rrr#r.gsz!PackageFinder._looks_like_packagecs fddS)z Given a list of patterns, return a callable that will be true only if the input matches at least one of the patterns. cstfddDS)Nc3s|]}t|dVqdS))patNr).0r8namerr# rsz@PackageFinder._build_filter....)anyr:Zpatternsr:r#rz-PackageFinder._build_filter..rr>rr>r#rlszPackageFinder._build_filterN)rrr) __name__ __module__ __qualname____doc__ classmethodr$r staticmethodr.rrrrr#r/s   rc@seZdZeddZdS)PEP420PackageFindercCsdS)NTrr7rrr#r.vsz'PEP420PackageFinder._looks_like_packageN)rArBrCrFr.rrrr#rGusrGcCs@tjtdd|D}|jdd|jr<||jdS)Ncss"|]\}}|dkr||fVqdS))Zdependency_linkssetup_requiresNr)r9kvrrr#r<sz*_install_setup_requires..T)Zignore_option_errors) distutilscorer dictitemsZparse_config_filesrHZfetch_build_eggs)attrsdistrrr#_install_setup_requiress   rQcKst|tjjf|SN)rQrKrLr)rOrrr#rsc@s:eZdZejZdZddZd ddZddZd d d Z dS)rFcKst||t||dS)zj Construct the command for dist, updating vars(self) with any keyword parameters. N)_Command__init__varsupdate)selfrPkwrrr#rTs zCommand.__init__NcCsBt||}|dkr"t||||St|ts>td|||f|S)Nz'%s' must be a %s (got `%s`))getattrsetattr isinstancer r)rWoptionZwhatdefaultvalrrr#_ensure_stringlikes   zCommand._ensure_stringlikecCspt||}|dkrdSt|tr6t||td|n6t|trTtdd|D}nd}|sltd||fdS)zEnsure that 'option' is a list of strings. If 'option' is currently a string, we split it either on /,\s*/ or /\s+/, so "foo bar baz", "foo,bar,baz", and "foo, bar baz" all become ["foo", "bar", "baz"]. Nz,\s*|\s+css|]}t|tVqdSrR)r[r )r9rJrrr#r<sz-Command.ensure_string_list..Fz''%s' must be a list of strings (got %r)) rYr[r rZresplitrallr)rWr\r^okrrr#ensure_string_lists   zCommand.ensure_string_listrcKs t|||}t|||SrR)rSreinitialize_commandrUrV)rWZcommandZreinit_subcommandsrXcmdrrr#reszCommand.reinitialize_command)N)r) rArBrCrSrDZcommand_consumes_argumentsrTr_rdrerrrr#rs  cCs&ddtj|ddD}ttjj|S)z% Find all files under 'path' css,|]$\}}}|D]}tj||VqqdSrR)r'r)r*)r9baser1r2filerrr#r<sz#_find_all_simple..Tr%)r'r(r r)r6)r)resultsrrr#_find_all_simples rjcCs6t|}|tjkr.tjtjj|d}t||}t|S)z Find all files under 'dir' and return the list of full filenames. Unless dir is '.', return full filenames with dir prepended. )start) rjr'curdir functoolspartialr)r+r r)r3r2Zmake_relrrr#findalls   ro)9rDr'sysrmZdistutils.corerKZdistutils.filelistr`Zdistutils.errorsrZdistutils.utilrZfnmatchrZ_deprecation_warningrZsetuptools.extern.sixrr Zsetuptools.extern.six.movesr r Zsetuptools.versionZ setuptoolsZsetuptools.extensionr Zsetuptools.distr rZsetuptools.dependsrrtypeZ __metaclass____all__r/version __version__Zbootstrap_install_fromZrun_2to3_on_doctestsZlib2to3_fixer_packagesrrGr$rrrQrrLZ get_unpatchedrrSrjrlroZ patch_allrrrr#s\        F  2  site-packages/setuptools/__pycache__/msvc.cpython-38.pyc000064400000115317151464452710017336 0ustar00U Qab@sXdZddlZddlmZddlmZmZddlmZm Z m Z m Z ddl Z ddl Z ddlZddlZddlmZddlmZdd lmZe d krdd lmZdd lmZnGd ddZeZeejjfZzddlm Z Wnek rYnXddZ!d$ddZ"ddZ#ddZ$d%ddZ%GdddZ&GdddZ'Gd d!d!Z(Gd"d#d#Z)dS)&a Improved support for Microsoft Visual C++ compilers. Known supported compilers: -------------------------- Microsoft Visual C++ 9.0: Microsoft Visual C++ Compiler for Python 2.7 (x86, amd64) Microsoft Windows SDK 6.1 (x86, x64, ia64) Microsoft Windows SDK 7.0 (x86, x64, ia64) Microsoft Visual C++ 10.0: Microsoft Windows SDK 7.1 (x86, x64, ia64) Microsoft Visual C++ 14.X: Microsoft Visual C++ Build Tools 2015 (x86, x64, arm) Microsoft Visual Studio Build Tools 2017 (x86, x64, arm, arm64) Microsoft Visual Studio Build Tools 2019 (x86, x64, arm, arm64) This may also support compilers shipped with compatible Visual Studio versions. N)open)listdirpathsep)joinisfileisdirdirname) LegacyVersion) filterfalse) get_unpatchedWindows)winreg)environc@seZdZdZdZdZdZdS)rN)__name__ __module__ __qualname__ HKEY_USERSHKEY_CURRENT_USERHKEY_LOCAL_MACHINEHKEY_CLASSES_ROOTrr3/usr/lib/python3.8/site-packages/setuptools/msvc.pyr*sr)Regc Csd}|d|f}zt|d}WnJtk rjz|d|f}t|d}Wntk rdd}YnXYnX|rt|d}t|r|Stt|S)a Patched "distutils.msvc9compiler.find_vcvarsall" to use the standalone compiler build for Python (VCForPython / Microsoft Visual C++ Compiler for Python 2.7). Fall back to original behavior when the standalone compiler is not available. Redirect the path of "vcvarsall.bat". Parameters ---------- version: float Required Microsoft Visual C++ version. Return ------ str vcvarsall.bat path z-Software\%sMicrosoft\DevDiv\VCForPython\%0.1f installdirz Wow6432Node\Nz vcvarsall.bat)rZ get_valueKeyErrorrrr msvc9_find_vcvarsall)versionZvc_basekey productdir vcvarsallrrrrAs   rx86c Osztt}|||f||WStjjk r4Yntk rFYnXzt||WStjjk r}zt|||W5d}~XYnXdS)ao Patched "distutils.msvc9compiler.query_vcvarsall" for support extra Microsoft Visual C++ 9.0 and 10.0 compilers. Set environment without use of "vcvarsall.bat". Parameters ---------- ver: float Required Microsoft Visual C++ version. arch: str Target architecture. Return ------ dict environment N) r msvc9_query_vcvarsall distutilserrorsDistutilsPlatformError ValueErrorEnvironmentInfo return_env_augment_exception)verarchargskwargsZorigexcrrrr#ks r#c Csrztt|WStjjk r&YnXzt|ddWStjjk rl}zt|dW5d}~XYnXdS)a* Patched "distutils._msvccompiler._get_vc_env" for support extra Microsoft Visual C++ 14.X compilers. Set environment without use of "vcvarsall.bat". Parameters ---------- plat_spec: str Target architecture. Return ------ dict environment ,@) vc_min_verN)r msvc14_get_vc_envr$r%r&r(r)r*)Z plat_specr/rrrr2s r2cOsBdtjkr4ddl}t|jtdkr4|jjj||Stt ||S)z Patched "distutils._msvccompiler.gen_lib_options" for fix compatibility between "numpy.distutils" and "distutils._msvccompiler" (for Numpy < 1.11.2) znumpy.distutilsrNz1.11.2) sysmodulesZnumpyr __version__r$Z ccompilerZgen_lib_optionsr msvc14_gen_lib_options)r-r.Znprrrr6s  r6rcCs|jd}d|ks"d|krd}|jft}d}|dkrf|ddkr\|d 7}q|d 7}n.|d kr|d 7}||d 7}n|dkr|d7}|f|_dS)zl Add details to the exception message to help guide the user as to what action will resolve it. rr!zvisual cz0Microsoft Visual C++ {version:0.1f} is required.z-www.microsoft.com/download/details.aspx?id=%d"@Zia64z( Get it with "Microsoft Windows SDK 7.0"z% Get it from http://aka.ms/vcpython27$@z* Get it with "Microsoft Windows SDK 7.1": iW r0z[ Get it with "Build Tools for Visual Studio": https://visualstudio.microsoft.com/downloads/N)r-lowerformatlocalsfind)r/rr,messageZtmplZ msdownloadrrrr*s   r*c@sbeZdZdZeddZddZe ddZ dd Z d d Z dd dZ dddZdddZdS) PlatformInfoz Current and Target Architectures information. Parameters ---------- arch: str Target architecture. Zprocessor_architecturercCs|dd|_dS)Nx64amd64)r:replacer,)selfr,rrr__init__szPlatformInfo.__init__cCs|j|jdddS)zs Return Target CPU architecture. Return ------ str Target CPU _r N)r,r=rCrrr target_cpus zPlatformInfo.target_cpucCs |jdkS)z Return True if target CPU is x86 32 bits.. Return ------ bool CPU is x86 32 bits r"rGrFrrr target_is_x86s zPlatformInfo.target_is_x86cCs |jdkS)z Return True if current CPU is x86 32 bits.. Return ------ bool CPU is x86 32 bits r" current_cpurFrrrcurrent_is_x86s zPlatformInfo.current_is_x86FcCs.|jdkr|rdS|jdkr$|r$dSd|jS)uk Current platform specific subfolder. Parameters ---------- hidex86: bool return '' and not '†' if architecture is x86. x64: bool return 'd' and not 'md64' if architecture is amd64. Return ------ str subfolder: ' arget', or '' (see hidex86 parameter) r"rrA\x64\%srJrChidex86r@rrr current_dirszPlatformInfo.current_dircCs.|jdkr|rdS|jdkr$|r$dSd|jS)ar Target platform specific subfolder. Parameters ---------- hidex86: bool return '' and not '\x86' if architecture is x86. x64: bool return '\x64' and not '\amd64' if architecture is amd64. Return ------ str subfolder: '\current', or '' (see hidex86 parameter) r"rrArMrNrHrOrrr target_dir(szPlatformInfo.target_dircCs0|rdn|j}|j|krdS|dd|S)ap Cross platform specific subfolder. Parameters ---------- forcex86: bool Use 'x86' as current architecture even if current architecture is not x86. Return ------ str subfolder: '' if target architecture is current architecture, '\current_target' if not. r"r\z\%s_)rKrGrRrB)rCforcex86Zcurrentrrr cross_dir>szPlatformInfo.cross_dirN)FF)FF)F)rrr__doc__rgetr:rKrDpropertyrGrIrLrQrRrUrrrrr?s    r?c@seZdZdZejejejejfZ ddZ e ddZ e ddZ e dd Ze d d Ze d d Ze ddZe ddZe ddZe ddZdddZddZdS) RegistryInfoz Microsoft Visual Studio related registry information. Parameters ---------- platform_info: PlatformInfo "PlatformInfo" instance. cCs ||_dSN)pi)rCZ platform_inforrrrDcszRegistryInfo.__init__cCsdS)z Microsoft Visual Studio root registry key. Return ------ str Registry key Z VisualStudiorrFrrr visualstudiofs zRegistryInfo.visualstudiocCs t|jdS)z Microsoft Visual Studio SxS registry key. Return ------ str Registry key ZSxS)rr\rFrrrsxsrs zRegistryInfo.sxscCs t|jdS)z| Microsoft Visual C++ VC7 registry key. Return ------ str Registry key ZVC7rr]rFrrrvc~s zRegistryInfo.vccCs t|jdS)z Microsoft Visual Studio VS7 registry key. Return ------ str Registry key ZVS7r^rFrrrvss zRegistryInfo.vscCsdS)z Microsoft Visual C++ for Python registry key. Return ------ str Registry key zDevDiv\VCForPythonrrFrrr vc_for_pythons zRegistryInfo.vc_for_pythoncCsdS)zq Microsoft SDK registry key. Return ------ str Registry key zMicrosoft SDKsrrFrrr microsoft_sdks zRegistryInfo.microsoft_sdkcCs t|jdS)z Microsoft Windows/Platform SDK registry key. Return ------ str Registry key r rrbrFrrr windows_sdks zRegistryInfo.windows_sdkcCs t|jdS)z Microsoft .NET Framework SDK registry key. Return ------ str Registry key ZNETFXSDKrcrFrrr netfx_sdks zRegistryInfo.netfx_sdkcCsdS)z Microsoft Windows Kits Roots registry key. Return ------ str Registry key zWindows Kits\Installed RootsrrFrrrwindows_kits_rootss zRegistryInfo.windows_kits_rootsFcCs$|js|rdnd}td|d|S)a Return key in Microsoft software registry. Parameters ---------- key: str Registry key path where look. x86: str Force x86 software registry. Return ------ str Registry key rZ Wow6432NodeZSoftwareZ Microsoft)r[rLr)rCrr"Znode64rrr microsoftszRegistryInfo.microsoftc Cstj}tj}|j}|jD]}z||||d|}Wn`ttfk r|jsz||||dd|}Wqttfk rYYqYqXnYqYnXzt ||dWSttfk rYqXqdS)a Look for values in registry in Microsoft software registry. Parameters ---------- key: str Registry key path where look. name: str Value name to find. Return ------ str value rTN) rKEY_READOpenKeyrgHKEYSOSErrorIOErrorr[rLZ QueryValueEx)rCrnameZkey_readZopenkeymshkeybkeyrrrlookups"   zRegistryInfo.lookupN)F)rrrrVrrrrrrjrDrXr\r]r_r`rarbrdrerfrgrqrrrrrYUs6         rYc@s<eZdZdZeddZeddZedeZd7ddZ d d Z d d Z d dZ e ddZeddZeddZddZddZeddZeddZeddZedd Zed!d"Zed#d$Zed%d&Zed'd(Zed)d*Zed+d,Zed-d.Zed/d0Zed1d2Z d3d4Z!e d8d5d6Z"dS)9 SystemInfoz Microsoft Windows and Visual Studio related system information. Parameters ---------- registry_info: RegistryInfo "RegistryInfo" instance. vc_ver: float Required Microsoft Visual C++ version. WinDirr ProgramFileszProgramFiles(x86)NcCs2||_|jj|_||_|p$||_|_dSrZ)rir[find_programdata_vs_versknown_vs_paths_find_latest_available_vs_vervs_vervc_ver)rCZ registry_inforzrrrrDs    zSystemInfo.__init__cCs>|}|s|jstjdt|}||jt|dS)zm Find the latest VC version Return ------ float version z%No Microsoft Visual C++ version foundr8)find_reg_vs_versrwr$r%r&setupdatesorted)rCZ reg_vc_versZvc_versrrrrx%s   z(SystemInfo._find_latest_available_vs_verc Cs$|jj}|jj|jj|jjf}g}|jjD]}|D]}zt|||dtj}Wnt t fk rlYq2YnXt |\}}} t |D]D} z*t t|| d} | |kr|| Wqtk rYqXqt |D]B} z&t t|| } | |kr|| Wqtk rYqXqq2q*t|S)z Find Microsoft Visual Studio versions available in registry. Return ------ list of float Versions r)rurgr_rar`rjrrirhrkrlZ QueryInfoKeyrangefloatZ EnumValueappendr'ZEnumKeyr~) rCrnZvckeysZvs_versrorrpZsubkeysvaluesrEir+rrrr{8s2      zSystemInfo.find_reg_vs_versc Csi}d}z t|}Wnttfk r0|YSX|D]}z\t||d}t|ddd}t|}W5QRX|d}tt|d||||d<Wq6tttfk rYq6Yq6Xq6|S) z Find Visual studio 2017+ versions from information in "C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances". Return ------ dict float version as key, path as value. z9C:\ProgramData\Microsoft\VisualStudio\Packages\_Instancesz state.jsonZrtzutf-8)encodingZinstallationPath VC\Tools\MSVCZinstallationVersion) rrkrlrrjsonload_as_float_versionr) rCZ vs_versionsZ instances_dirZ hashed_namesrmZ state_pathZ state_filestateZvs_pathrrrrv[s*     z#SystemInfo.find_programdata_vs_verscCstd|dddS)z Return a string version as a simplified float version (major.minor) Parameters ---------- version: str Version. Return ------ float version .N)rrsplit)rrrrrszSystemInfo._as_float_versioncCs.t|jd|j}|j|jjd|jp,|S)zp Microsoft Visual Studio directory. Return ------ str path zMicrosoft Visual Studio %0.1f%0.1f)rProgramFilesx86ryrurqr`)rCdefaultrrr VSInstallDirs zSystemInfo.VSInstallDircCs,|p|}t|s(d}tj||S)zm Microsoft Visual C++ directory. Return ------ str path z(Microsoft Visual C++ directory not found) _guess_vc_guess_vc_legacyrr$r%r&)rCpathmsgrrr VCInstallDirs  zSystemInfo.VCInstallDirc Cs|jdkrdSz|j|j}Wntk r8|j}YnXt|d}z$t|d}|||_t||WStt t fk rYdSXdS)zl Locate Visual C++ for VS2017+. Return ------ str path r0rrr8N) ryrwrrrrrrzrkrl IndexError)rCZvs_dirZguess_vcrzrrrrs      zSystemInfo._guess_vccCsbt|jd|j}t|jjd|j}|j|d}|rBt|dn|}|j|jjd|jp`|S)z{ Locate Visual C++ for versions prior to 2017. Return ------ str path z Microsoft Visual Studio %0.1f\VCrrZVC)rrryrurarqr_)rCrZreg_pathZ python_vcZ default_vcrrrrs zSystemInfo._guess_vc_legacycCsJ|jdkrdS|jdkrdS|jdkr*dS|jdkr8dS|jd krFd Sd S) z Microsoft Windows SDK versions for specified MSVC++ version. Return ------ tuple of str versions r7)z7.0z6.1z6.0ar9)z7.1z7.0a&@)z8.0z8.0a(@)8.1z8.1ar0)z10.0rNryrFrrrWindowsSdkVersions     zSystemInfo.WindowsSdkVersioncCs|t|jdS)zt Microsoft Windows SDK last version. Return ------ str version lib)_use_last_dir_namer WindowsSdkDirrFrrrWindowsSdkLastVersions z SystemInfo.WindowsSdkLastVersioncCs d}|jD],}t|jjd|}|j|d}|r q8q |rDt|stt|jjd|j}|j|d}|rtt|d}|rt|s|jD]6}|d|d}d |}t|j |}t|r|}q|rt|s|jD]$}d |}t|j |}t|r|}q|st|j d }|S) zn Microsoft Windows SDK directory. Return ------ str path rzv%sinstallationfolderrrZWinSDKNrzMicrosoft SDKs\Windows Kits\%szMicrosoft SDKs\Windows\v%sZ PlatformSDK) rrrurdrqrrarzrfindrtr)rCsdkdirr+locrZ install_baseZintverdrrrr s6           zSystemInfo.WindowsSdkDirc Cs|jdkrd}d}n&d}|jdkr&dnd}|jjd|d}d ||d d f}g}|jd kr~|jD]}|t|jj||g7}qb|jD]}|t|jj d ||g7}q|D]}|j |d}|r|SqdS)zy Microsoft Windows SDK executable directory. Return ------ str path r#r(rTF)r@rPzWinSDK-NetFx%dTools%srS-r0zv%sArN) ryr[rQrBNetFxSdkVersionrrurerrdrq) rCZnetfxverr,rPZfxZregpathsr+rZexecpathrrrWindowsSDKExecutablePath7s"    z#SystemInfo.WindowsSDKExecutablePathcCs&t|jjd|j}|j|dp$dS)zl Microsoft Visual F# directory. Return ------ str path z%0.1f\Setup\F#r r)rrur\ryrq)rCrrrrFSharpInstallDirZs zSystemInfo.FSharpInstallDircCsF|jdkrdnd}|D]*}|j|jjd|}|r|p:dSqdS)zt Microsoft Universal CRT SDK directory. Return ------ str path r0)Z10Z81rz kitsroot%srN)ryrurqrf)rCZversr+rrrrUniversalCRTSdkDirgs  zSystemInfo.UniversalCRTSdkDircCs|t|jdS)z Microsoft Universal C Runtime SDK last version. Return ------ str version r)rrrrFrrrUniversalCRTSdkLastVersion{s z%SystemInfo.UniversalCRTSdkLastVersioncCs|jdkrdSdS)z Microsoft .NET Framework SDK versions. Return ------ tuple of str versions r0) z4.7.2z4.7.1z4.7z4.6.2z4.6.1z4.6z4.5.2z4.5.1z4.5rrrFrrrrszSystemInfo.NetFxSdkVersioncCs8d}|jD](}t|jj|}|j|d}|r q4q |S)zu Microsoft .NET Framework SDK directory. Return ------ str path rZkitsinstallationfolder)rrrurerq)rCrr+rrrr NetFxSdkDirs  zSystemInfo.NetFxSdkDircCs"t|jd}|j|jjdp |S)zw Microsoft .NET Framework 32bit directory. Return ------ str path zMicrosoft.NET\FrameworkZframeworkdir32rrsrurqr_rCZguess_fwrrrFrameworkDir32s zSystemInfo.FrameworkDir32cCs"t|jd}|j|jjdp |S)zw Microsoft .NET Framework 64bit directory. Return ------ str path zMicrosoft.NET\Framework64Zframeworkdir64rrrrrFrameworkDir64s zSystemInfo.FrameworkDir64cCs |dS)z Microsoft .NET Framework 32bit versions. Return ------ tuple of str versions _find_dot_net_versionsrFrrrFrameworkVersion32s zSystemInfo.FrameworkVersion32cCs |dS)z Microsoft .NET Framework 64bit versions. Return ------ tuple of str versions @rrFrrrFrameworkVersion64s zSystemInfo.FrameworkVersion64cCs|j|jjd|}t|d|}|p6||dp6d}|jdkrJ|dfS|jdkrt|dd d krld n|d fS|jd krdS|jdkrdSdS)z Find Microsoft .NET Framework versions. Parameters ---------- bits: int Platform number of bits: 32 or 64. Return ------ tuple of str versions zframeworkver%dzFrameworkDir%dvrrzv4.0r9NrZv4z v4.0.30319v3.5r7)r v2.0.50727g @)zv3.0r)rurqr_getattrrryr:)rCbitsZreg_verZ dot_net_dirr+rrrrs     z!SystemInfo._find_dot_net_versionscs*fddttD}t|dp(dS)a) Return name of the last dir in path or '' if no dir found. Parameters ---------- path: str Use dirs in this path prefix: str Use only dirs starting by this prefix Return ------ str name c3s*|]"}tt|r|r|VqdSrZ)rr startswith).0Zdir_namerprefixrr s z0SystemInfo._use_last_dir_name..Nr)reversedrnext)rrZ matching_dirsrrrrs  zSystemInfo._use_last_dir_name)N)r)#rrrrVrrWrsrtrrDrxr{rv staticmethodrrXrrrrrrrrrrrrrrrrrrrrrrrrr sZ    #*      * "         rrc@sbeZdZdZd?ddZeddZedd Zed d Zed d Z eddZ eddZ eddZ eddZ eddZeddZeddZddZeddZed d!Zed"d#Zed$d%Zed&d'Zed(d)Zed*d+Zed,d-Zed.d/Zed0d1Zed2d3Zed4d5Zed6d7Zd@d9d:Zd;d<Z e!dAd=d>Z"dS)Br(aY Return environment variables for specified Microsoft Visual C++ version and platform : Lib, Include, Path and libpath. This function is compatible with Microsoft Visual C++ 9.0 to 14.X. Script created by analysing Microsoft environment configuration files like "vcvars[...].bat", "SetEnv.Cmd", "vcbuildtools.bat", ... Parameters ---------- arch: str Target architecture. vc_ver: float Required Microsoft Visual C++ version. If not set, autodetect the last version. vc_min_ver: float Minimum Microsoft Visual C++ version. NrcCsBt||_t|j|_t|j||_|j|kr>d}tj |dS)Nz.No suitable Microsoft Visual C++ version found) r?r[rYrurrsirzr$r%r&)rCr,rzr1errrrrrD0s    zEnvironmentInfo.__init__cCs|jjS)zk Microsoft Visual Studio. Return ------ float version )rryrFrrrry9s zEnvironmentInfo.vs_vercCs|jjS)zp Microsoft Visual C++ version. Return ------ float version )rrzrFrrrrzEs zEnvironmentInfo.vc_vercsVddg}jdkrDjjddd}|dg7}|dg7}|d|g7}fd d |DS) zu Microsoft Visual Studio Tools. Return ------ list of str paths z Common7\IDEz Common7\Toolsr0TrPr@z1Common7\IDE\CommonExtensions\Microsoft\TestWindowzTeam Tools\Performance ToolszTeam Tools\Performance Tools%scsg|]}tjj|qSrrrrrrrFrr csz+EnvironmentInfo.VSTools..)ryr[rQ)rCpaths arch_subdirrrFrVSToolsQs    zEnvironmentInfo.VSToolscCst|jjdt|jjdgS)z Microsoft Visual C++ & Microsoft Foundation Class Includes. Return ------ list of str paths ZIncludezATLMFC\IncluderrrrFrrr VCIncludeses  zEnvironmentInfo.VCIncludescsbjdkrjjdd}njjdd}d|d|g}jdkrP|d|g7}fd d |DS) z Microsoft Visual C++ & Microsoft Foundation Class Libraries. Return ------ list of str paths .@Tr@rPLib%sz ATLMFC\Lib%sr0z Lib\store%scsg|]}tjj|qSrrrrFrrrsz/EnvironmentInfo.VCLibraries..)ryr[rR)rCrrrrFr VCLibrariesrs  zEnvironmentInfo.VCLibrariescCs|jdkrgSt|jjdgS)z Microsoft Visual C++ store references Libraries. Return ------ list of str paths r0zLib\store\references)ryrrrrFrrr VCStoreRefss zEnvironmentInfo.VCStoreRefscCs|j}t|jdg}|jdkr"dnd}|j|}|rL|t|jd|g7}|jdkr|d|jjdd}|t|j|g7}n|jdkr|jrd nd }|t|j||jjdd g7}|jj |jj kr|t|j||jjdd g7}n|t|jd g7}|S) zr Microsoft Visual C++ Tools. Return ------ list of str paths Z VCPackagesr9TFBin%sr0rrz bin\HostX86%sz bin\HostX64%srBin) rrrryr[rUrQrLrRrKrG)rCrtoolsrTrrZhost_dirrrrVCToolss0     zEnvironmentInfo.VCToolscCsh|jdkr.|jjddd}t|jjd|gS|jjdd}t|jjd}|j}t|d||fgSdS) zw Microsoft Windows SDK Libraries. Return ------ list of str paths r9Trrrrz%sum%sN)ryr[rRrrr _sdk_subdir)rCrrZlibverrrr OSLibrariess zEnvironmentInfo.OSLibrariescCsht|jjd}|jdkr&|t|dgS|jdkr8|j}nd}t|d|t|d|t|d|gSd S) zu Microsoft Windows SDK Include. Return ------ list of str paths includer9Zglr0rz%ssharedz%sumz%swinrtN)rrrryr)rCrsdkverrrr OSIncludess      zEnvironmentInfo.OSIncludescCst|jjd}g}|jdkr&||j7}|jdkr@|t|dg7}|jdkr||t|jjdt|ddt|d dt|d dt|jjd d d |jdddg7}|S)z} Microsoft Windows SDK Libraries Paths. Return ------ list of str paths Z Referencesr7rzCommonConfiguration\Neutralr0Z UnionMetadataz'Windows.Foundation.UniversalApiContractz1.0.0.0z%Windows.Foundation.FoundationContractz,Windows.Networking.Connectivity.WwanContractZ ExtensionSDKszMicrosoft.VCLibsrZCommonConfigurationZneutral)rrrryr)rCreflibpathrrr OSLibpaths.          zEnvironmentInfo.OSLibpathcCs t|S)zs Microsoft Windows SDK Tools. Return ------ list of str paths )list _sdk_toolsrFrrrSdkToolss zEnvironmentInfo.SdkToolsccs|jdkr,|jdkrdnd}t|jj|V|js\|jjdd}d|}t|jj|V|jdkr|jrvd }n|jjddd }d |}t|jj|VnB|jdkrt|jjd}|jjdd}|jj}t|d ||fV|jj r|jj Vd S)z Microsoft Windows SDK Tools paths generator. Return ------ generator of str paths rrrzBin\x86Trr)r9rrrzBin\NETFX 4.0 Tools%sz%s%sN) ryrrrr[rLrQrIrr)rCZbin_dirrrrrrrrs(     zEnvironmentInfo._sdk_toolscCs|jj}|rd|SdS)zu Microsoft Windows SDK version subdir. Return ------ str subdir %s\r)rrrCucrtverrrrr6s zEnvironmentInfo._sdk_subdircCs|jdkrgSt|jjdgS)zs Microsoft Windows SDK Setup. Return ------ list of str paths r7ZSetup)ryrrrrFrrrSdkSetupCs zEnvironmentInfo.SdkSetupcs|j}|j|jdkr0d}| o,| }n$|p>|}|jdkpR|jdk}g}|rt|fddjD7}|r|fddjD7}|S)zv Microsoft .NET Framework Tools. Return ------ list of str paths r9TrAcsg|]}tj|qSr)rrrr+rrrrhsz+EnvironmentInfo.FxTools..csg|]}tj|qSr)rrrrrrrks) r[rryrIrLrKrGrr)rCr[Z include32Z include64rrrrFxToolsRs"    zEnvironmentInfo.FxToolscCs8|jdks|jjsgS|jjdd}t|jjd|gS)z~ Microsoft .Net Framework SDK Libraries. Return ------ list of str paths r0Trzlib\um%s)ryrrr[rRr)rCrrrrNetFxSDKLibrariesos z!EnvironmentInfo.NetFxSDKLibrariescCs&|jdks|jjsgSt|jjdgS)z} Microsoft .Net Framework SDK Includes. Return ------ list of str paths r0z include\um)ryrrrrFrrrNetFxSDKIncludess z EnvironmentInfo.NetFxSDKIncludescCst|jjdgS)z Microsoft Visual Studio Team System Database. Return ------ list of str paths z VSTSDB\DeployrrFrrrVsTDbs zEnvironmentInfo.VsTDbcCsv|jdkrgS|jdkr0|jj}|jjdd}n |jj}d}d|j|f}t||g}|jdkrr|t||dg7}|S)zn Microsoft Build Engine. Return ------ list of str paths rrTrrzMSBuild\%0.1f\bin%sZRoslyn)ryrrr[rQrr)rC base_pathrrZbuildrrrMSBuilds    zEnvironmentInfo.MSBuildcCs|jdkrgSt|jjdgS)zt Microsoft HTML Help Workshop. Return ------ list of str paths rzHTML Help Workshop)ryrrrrFrrrHTMLHelpWorkshops z EnvironmentInfo.HTMLHelpWorkshopcCsD|jdkrgS|jjdd}t|jjd}|j}t|d||fgS)z Microsoft Universal C Runtime SDK Libraries. Return ------ list of str paths r0Trrz%sucrt%s)ryr[rRrrr _ucrt_subdir)rCrrrrrr UCRTLibrariess zEnvironmentInfo.UCRTLibrariescCs.|jdkrgSt|jjd}t|d|jgS)z Microsoft Universal C Runtime SDK Include. Return ------ list of str paths r0rz%sucrt)ryrrrr)rCrrrr UCRTIncludess zEnvironmentInfo.UCRTIncludescCs|jj}|rd|SdS)z Microsoft Universal C Runtime SDK version subdir. Return ------ str subdir rr)rrrrrrrs zEnvironmentInfo._ucrt_subdircCs(d|jkrdkrnngS|jjgS)zk Microsoft Visual F#. Return ------ list of str paths rr)ryrrrFrrrFSharps zEnvironmentInfo.FSharpc Csd|j}|jjddd}g}|jj}t|dd}t|rft |t |d}||t |dg7}|t |d g7}d |jd d t |j d f}t ||D]&\}}t ||||} t| r| Sqd S) z Microsoft Visual C++ runtime redistributable dll. Return ------ str path zvcruntime%d0.dllTrrSz\Toolsz\Redistr8ZonecoreZredistzMicrosoft.VC%d.CRT N)rzr[rRstriprrrrBrrrintry itertoolsproductr) rCZ vcruntimerprefixesZ tools_pathZ redist_pathZcrt_dirsrZcrt_dirrrrrVCRuntimeRedists  zEnvironmentInfo.VCRuntimeRedistTcCst|d|j|j|j|jg||d|j|j|j|j |j g||d|j|j|j |j g||d|j |j|j|j|j|j|j|j|jg |d}|jdkrt|jr|j|d<|S)z Return environment dict. Parameters ---------- exists: bool It True, only return existing paths. Return ------ dict environment rrrr)rrrrZpy_vcruntime_redist)dict _build_pathsrrrrrrrrrrrrrrrrrrrryrr)rCexistsenvrrrr)&sV   zEnvironmentInfo.return_envc Csptj|}t|dt}t||}|r A B C D _unique_everseen('ABBCcAD', str.lower) --> A B C D N)r|addr __contains__)iterablerseenZseen_addZelementkrrrrzs  z EnvironmentInfo._unique_everseen)Nr)T)N)#rrrrVrDrXryrzrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrrrr(sn        $    #             " 2"r()r")r)*rVriorosrrZos.pathrrrrr3platformrZdistutils.errorsr$Z#setuptools.extern.packaging.versionr Zsetuptools.extern.six.movesr Zmonkeyr systemrrr ImportErrorr%r&Z_msvc9_suppress_errorsZdistutils.msvc9compilerrrr#r2r6r*r?rYrrr(rrrrsJ       * &  $s5site-packages/setuptools/__pycache__/config.cpython-38.opt-1.pyc000064400000042725151464452710020574 0ustar00U Qab6P@sddlmZmZddlZddlZddlZddlZddlZddlm Z ddlm Z ddlm Z ddl m Z ddlmZmZddlmZmZdd lmZdd lmZmZeZdd d ZddZddZdddZGdddZGdddeZ GdddeZ!dS))absolute_importunicode_literalsN) defaultdict)partialwraps) import_module)DistutilsOptionErrorDistutilsFileError) LegacyVersionparse) SpecifierSet) string_typesPY3Fc Csddlm}m}tj|}tj|s4td|t}t tj |zJ|}|rb| ng}||krx| ||j ||dt||j|d}W5t |Xt|S)a,Read given configuration file and returns options from it as a dict. :param str|unicode filepath: Path to configuration file to get options from. :param bool find_others: Whether to search for other configuration files which could be on in various places. :param bool ignore_option_errors: Whether to silently ignore options, values of which could not be resolved (e.g. due to exceptions in directives such as file:, attr:, etc.). If False exceptions are propagated as expected. :rtype: dict r) Distribution _Distributionz%Configuration file %s does not exist.) filenames)ignore_option_errors)Zsetuptools.distrrospathabspathisfiler getcwdchdirdirnameZfind_config_filesappendZparse_config_filesparse_configurationcommand_optionsconfiguration_to_dict) filepathZ find_othersrrrZcurrent_directoryZdistrhandlersr!5/usr/lib/python3.8/site-packages/setuptools/config.pyread_configurations*     r#cCs.djft}tt||}t|||}|S)z Given a target object and option key, get that option from the target object, either through a get_{key} method or from an attribute directly. z get_{key})formatlocals functoolsrgetattr) target_objkeyZ getter_nameZ by_attributegetterr!r!r" _get_optionEs r+cCs<tt}|D]*}|jD]}t|j|}|||j|<qq |S)zReturns configuration data gathered by given handlers as a dict. :param list[ConfigHandler] handlers: Handlers list, usually from parse_configuration() :rtype: dict )rdict set_optionsr+r(section_prefix)r Z config_dictZhandlerZoptionvaluer!r!r"rQs   rcCs6t|||}|t|j|||j}|||fS)aPerforms additional parsing of configuration options for a distribution. Returns a list of used option handlers. :param Distribution distribution: :param dict command_options: :param bool ignore_option_errors: Whether to silently ignore options, values of which could not be resolved (e.g. due to exceptions in directives such as file:, attr:, etc.). If False exceptions are propagated as expected. :rtype: list )ConfigOptionsHandlerr ConfigMetadataHandlermetadata package_dir)Z distributionrroptionsmetar!r!r"rcsrc@seZdZdZdZiZd%ddZeddZdd Z e d&d d Z e d dZ e ddZ e ddZe ddZeddZeddZe d'ddZe ddZe d(ddZdd Zd!d"Zd#d$ZdS)) ConfigHandlerz1Handles metadata supplied in configuration files.NFcCs^i}|j}|D].\}}||s&q||dd}|||<q||_||_||_g|_dS)N.) r.items startswithreplacestriprr(sectionsr-)selfr(r4rr=r. section_namesection_optionsr!r!r"__init__s  zConfigHandler.__init__cCstd|jjdS).Metadata item name to parser function mapping.z!%s must provide .parsers propertyN)NotImplementedError __class____name__)r>r!r!r"parserss zConfigHandler.parsersc Cst}|j}|j||}t|||}||kr6t||r>dSd}|j|}|rz ||}Wn tk r~d}|jszYnX|rdSt|d|d}|dkrt |||n|||j |dS)NFTzset_%s) tupler(aliasesgetr'KeyErrorrF Exceptionrsetattrr-r) r>Z option_namer/unknownr(Z current_valueZ skip_optionparsersetterr!r!r" __setitem__s0   zConfigHandler.__setitem__,cCs8t|tr|Sd|kr |}n ||}dd|DS)zRepresents value as a list. Value is split either by separator (defaults to comma) or by lines. :param value: :param separator: List items separator character. :rtype: list  cSsg|]}|r|qSr!)r<).0chunkr!r!r" sz-ConfigHandler._parse_list..) isinstancelist splitlinessplit)clsr/ separatorr!r!r" _parse_lists   zConfigHandler._parse_listcCsPd}i}||D]8}||\}}}||kr:td||||<q|S)zPRepresents value as a dict. :param value: :rtype: dict =z(Unable to parse option value to dict: %s)r\ partitionr r<)rZr/r[resultliner)sepvalr!r!r" _parse_dictszConfigHandler._parse_dictcCs|}|dkS)zQRepresents value as boolean. :param value: :rtype: bool )1trueZyes)lower)rZr/r!r!r" _parse_boolszConfigHandler._parse_boolcsfdd}|S)zReturns a parser function to make sure field inputs are not files. Parses a value after getting the key so error messages are more informative. :param key: :rtype: callable cs d}||rtd|S)Nfile:zCOnly strings are accepted for the {0} field, files are not accepted)r: ValueErrorr$)r/Zexclude_directiver)r!r"rNs z3ConfigHandler._exclude_files_parser..parserr!)rZr)rNr!rjr"_exclude_files_parsers z#ConfigHandler._exclude_files_parsercs\d}t|ts|S||s |S|t|d}dd|dD}dfdd|DS)aORepresents value as a string, allowing including text from nearest files using `file:` directive. Directive is sandboxed and won't reach anything outside directory with setup.py. Examples: file: README.rst, CHANGELOG.md, src/file.txt :param str value: :rtype: str rhNcss|]}tj|VqdSN)rrrr<rSrr!r!r" %sz,ConfigHandler._parse_file..rQrRc3s.|]&}|stj|r|VqdS)TN) _assert_localrrr _read_filermrZr!r"rn&s   )rVrr:lenrYjoin)rZr/Zinclude_directivespecZ filepathsr!rqr" _parse_files  zConfigHandler._parse_filecCs|tstd|dS)Nz#`file:` directive can not access %s)r:rrr )rr!r!r"ro-szConfigHandler._assert_localc Cs.tj|dd}|W5QRSQRXdS)Nzutf-8)encoding)ioopenread)rfr!r!r"rp3szConfigHandler._read_filec Csd}||s|S||dd}|}d|}|p@d}t}|r|d|kr||d}|dd} t | dkrtj t| d}| d}q|}nd|krtj t|d}t j d|zt |} t| |}W5t j ddt _ X|S) zRepresents value as a module attribute. Examples: attr: package.attr attr: package.module.attr :param str value: :rtype: str zattr:r7r8rAr/N)r:r;r<rYpoprsrrrsplitrrrsysinsertrr') rZr/r3Zattr_directiveZ attrs_pathZ attr_nameZ module_name parent_pathZ custom_pathpartsmoduler!r!r" _parse_attr8s0        zConfigHandler._parse_attrcsfdd}|S)zReturns parser function to represents value as a list. Parses a value applying given methods one after another. :param parse_methods: :rtype: callable cs|}D] }||}q|Srlr!)r/parsedmethod parse_methodsr!r"r ns z1ConfigHandler._get_parser_compound..parser!)rZrr r!rr"_get_parser_compoundes z"ConfigHandler._get_parser_compoundcCs6i}|pdd}|D]\}\}}||||<q|S)zParses section options into a dictionary. Optionally applies a given parser to values. :param dict section_options: :param callable values_parser: :rtype: dict cSs|Srlr!)rbr!r!r"z6ConfigHandler._parse_section_to_dict..)r9)rZr@Z values_parserr/r)_rbr!r!r"_parse_section_to_dictxs  z$ConfigHandler._parse_section_to_dictc Cs<|D].\}\}}z |||<Wqtk r4YqXqdS)zQParses configuration file section. :param dict section_options: N)r9rJ)r>r@namerr/r!r!r" parse_sections  zConfigHandler.parse_sectioncCsb|jD]R\}}d}|r"d|}t|d|ddd}|dkrTtd|j|f||q dS)zTParses configuration file items from one or more related sections. r7z_%szparse_section%sr8__Nz0Unsupported distribution option section: [%s.%s])r=r9r'r;r r.)r>r?r@Zmethod_postfixZsection_parser_methodr!r!r"r s"zConfigHandler.parsecstfdd}|S)z this function will wrap around parameters that are deprecated :param msg: deprecation message :param warning_class: class of warning exception to be raised :param func: function to be wrapped around cst||Srl)warningswarn)argskwargsfuncmsg warning_classr!r"config_handlers z@ConfigHandler._deprecated_config_handler..config_handlerr)r>rrrrr!rr"_deprecated_config_handlersz(ConfigHandler._deprecated_config_handler)F)rQ)N)N)rE __module__ __qualname____doc__r.rHrApropertyrFrP classmethodr\rcrgrkru staticmethodrorprrrrr rr!r!r!r"r6~s<  &        ,   r6csHeZdZdZdddddZdZdfd d Zed d Zd dZ Z S)r1r2Zurl description classifiers platforms)Z home_pageZsummaryZ classifierplatformFNcstt||||||_dSrl)superr1rAr3)r>r(r4rr3rDr!r"rAszConfigMetadataHandler.__init__c CsL|j}|j}|j}|j}|||||dt|||||d|||j|d S)rBz[The requires parameter is deprecated, please use install_requires for runtime dependencies.license) rkeywordsZprovidesZrequiresZ obsoletesrrrZlong_descriptionversionZ project_urls)r\rurcrkrDeprecationWarningr_parse_version)r> parse_listZ parse_file parse_dictZexclude_files_parserr!r!r"rFs( zConfigMetadataHandler.parserscCs||}||krB|}tt|tr>d}t|jft|S|||j }t |r^|}t|t st |drd tt|}nd|}|S)zSParses `version` option value. :param value: :rtype: str zCVersion loaded from {value} does not comply with PEP 440: {version}__iter__r8z%s)rur<rVr r r r$r%rr3callablerhasattrrsmapstr)r>r/rZtmplr!r!r"rs    z$ConfigMetadataHandler._parse_version)FN) rErrr.rHZ strict_moderArrFr __classcell__r!r!rr"r1s r1c@s\eZdZdZeddZddZddZdd Zd d Z d d Z ddZ ddZ ddZ dS)r0r4cCsN|j}t|jdd}|j}|j}|||||||||||||||j|j|tdS)rB;r[)Zzip_safeZuse_2to3Zinclude_package_datar3Zuse_2to3_fixersZuse_2to3_exclude_fixersZconvert_2to3_doctestsZscriptsZeager_resourcesZdependency_linksZnamespace_packagesZinstall_requiresZsetup_requiresZ tests_requireZpackages entry_pointsZ py_modulesZpython_requires)r\rrgrc_parse_packagesrur )r>rZparse_list_semicolonZ parse_boolrr!r!r"rFs.zConfigOptionsHandler.parserscCszddg}|}||kr"||S||dk}|r>ts>td||jdi}|rdddlm}n ddlm }|f|S) zTParses `packages` option value. :param value: :rtype: list zfind:zfind_namespace:r|z8find_namespace: directive is unsupported on Python < 3.3z packages.findr)find_namespace_packages) find_packages) r<r\rr parse_section_packages__findr=rIZ setuptoolsrr)r>r/Zfind_directivesZ trimmed_valueZfindns find_kwargsrr!r!r"r1s     z$ConfigOptionsHandler._parse_packagescsT|||j}dddgtfdd|D}|d}|dk rP|d|d<|S)zParses `packages.find` configuration file section. To be used in conjunction with _parse_packages(). :param dict section_options: whereZincludeexcludecs$g|]\}}|kr|r||fqSr!r!rSkvZ valid_keysr!r"rUZszEConfigOptionsHandler.parse_section_packages__find..Nr)rr\r,r9rI)r>r@Z section_datarrr!rr"rMs   z1ConfigOptionsHandler.parse_section_packages__findcCs|||j}||d<dS)z`Parses `entry_points` configuration file section. :param dict section_options: rN)rr\r>r@rr!r!r"parse_section_entry_pointsbsz/ConfigOptionsHandler.parse_section_entry_pointscCs.|||j}|d}|r*||d<|d=|S)N*r7)rr\rI)r>r@rrootr!r!r"_parse_package_datajs  z(ConfigOptionsHandler._parse_package_datacCs|||d<dS)z`Parses `package_data` configuration file section. :param dict section_options: Z package_dataNrr>r@r!r!r"parse_section_package_datatsz/ConfigOptionsHandler.parse_section_package_datacCs|||d<dS)zhParses `exclude_package_data` configuration file section. :param dict section_options: Zexclude_package_dataNrrr!r!r""parse_section_exclude_package_data{sz7ConfigOptionsHandler.parse_section_exclude_package_datacCs"t|jdd}||||d<dS)zbParses `extras_require` configuration file section. :param dict section_options: rrZextras_requireN)rr\r)r>r@rr!r!r"parse_section_extras_requires z1ConfigOptionsHandler.parse_section_extras_requirecCs(|||j}dd|D|d<dS)z^Parses `data_files` configuration file section. :param dict section_options: cSsg|]\}}||fqSr!r!rr!r!r"rUszAConfigOptionsHandler.parse_section_data_files..Z data_filesN)rr\r9rr!r!r"parse_section_data_filessz-ConfigOptionsHandler.parse_section_data_filesN)rErrr.rrFrrrrrrrrr!r!r!r"r0s   r0)FF)F)"Z __future__rrrwrrrr& collectionsrrr importlibrZdistutils.errorsr r Z#setuptools.extern.packaging.versionr r Z&setuptools.extern.packaging.specifiersr Zsetuptools.extern.sixrrtypeZ __metaclass__r#r+rrr6r1r0r!r!r!r"s4      /  ?Usite-packages/setuptools/__pycache__/msvc.cpython-38.opt-1.pyc000064400000115317151464452710020275 0ustar00U Qab@sXdZddlZddlmZddlmZmZddlmZm Z m Z m Z ddl Z ddl Z ddlZddlZddlmZddlmZdd lmZe d krdd lmZdd lmZnGd ddZeZeejjfZzddlm Z Wnek rYnXddZ!d$ddZ"ddZ#ddZ$d%ddZ%GdddZ&GdddZ'Gd d!d!Z(Gd"d#d#Z)dS)&a Improved support for Microsoft Visual C++ compilers. Known supported compilers: -------------------------- Microsoft Visual C++ 9.0: Microsoft Visual C++ Compiler for Python 2.7 (x86, amd64) Microsoft Windows SDK 6.1 (x86, x64, ia64) Microsoft Windows SDK 7.0 (x86, x64, ia64) Microsoft Visual C++ 10.0: Microsoft Windows SDK 7.1 (x86, x64, ia64) Microsoft Visual C++ 14.X: Microsoft Visual C++ Build Tools 2015 (x86, x64, arm) Microsoft Visual Studio Build Tools 2017 (x86, x64, arm, arm64) Microsoft Visual Studio Build Tools 2019 (x86, x64, arm, arm64) This may also support compilers shipped with compatible Visual Studio versions. N)open)listdirpathsep)joinisfileisdirdirname) LegacyVersion) filterfalse) get_unpatchedWindows)winreg)environc@seZdZdZdZdZdZdS)rN)__name__ __module__ __qualname__ HKEY_USERSHKEY_CURRENT_USERHKEY_LOCAL_MACHINEHKEY_CLASSES_ROOTrr3/usr/lib/python3.8/site-packages/setuptools/msvc.pyr*sr)Regc Csd}|d|f}zt|d}WnJtk rjz|d|f}t|d}Wntk rdd}YnXYnX|rt|d}t|r|Stt|S)a Patched "distutils.msvc9compiler.find_vcvarsall" to use the standalone compiler build for Python (VCForPython / Microsoft Visual C++ Compiler for Python 2.7). Fall back to original behavior when the standalone compiler is not available. Redirect the path of "vcvarsall.bat". Parameters ---------- version: float Required Microsoft Visual C++ version. Return ------ str vcvarsall.bat path z-Software\%sMicrosoft\DevDiv\VCForPython\%0.1f installdirz Wow6432Node\Nz vcvarsall.bat)rZ get_valueKeyErrorrrr msvc9_find_vcvarsall)versionZvc_basekey productdir vcvarsallrrrrAs   rx86c Osztt}|||f||WStjjk r4Yntk rFYnXzt||WStjjk r}zt|||W5d}~XYnXdS)ao Patched "distutils.msvc9compiler.query_vcvarsall" for support extra Microsoft Visual C++ 9.0 and 10.0 compilers. Set environment without use of "vcvarsall.bat". Parameters ---------- ver: float Required Microsoft Visual C++ version. arch: str Target architecture. Return ------ dict environment N) r msvc9_query_vcvarsall distutilserrorsDistutilsPlatformError ValueErrorEnvironmentInfo return_env_augment_exception)verarchargskwargsZorigexcrrrr#ks r#c Csrztt|WStjjk r&YnXzt|ddWStjjk rl}zt|dW5d}~XYnXdS)a* Patched "distutils._msvccompiler._get_vc_env" for support extra Microsoft Visual C++ 14.X compilers. Set environment without use of "vcvarsall.bat". Parameters ---------- plat_spec: str Target architecture. Return ------ dict environment ,@) vc_min_verN)r msvc14_get_vc_envr$r%r&r(r)r*)Z plat_specr/rrrr2s r2cOsBdtjkr4ddl}t|jtdkr4|jjj||Stt ||S)z Patched "distutils._msvccompiler.gen_lib_options" for fix compatibility between "numpy.distutils" and "distutils._msvccompiler" (for Numpy < 1.11.2) znumpy.distutilsrNz1.11.2) sysmodulesZnumpyr __version__r$Z ccompilerZgen_lib_optionsr msvc14_gen_lib_options)r-r.Znprrrr6s  r6rcCs|jd}d|ks"d|krd}|jft}d}|dkrf|ddkr\|d 7}q|d 7}n.|d kr|d 7}||d 7}n|dkr|d7}|f|_dS)zl Add details to the exception message to help guide the user as to what action will resolve it. rr!zvisual cz0Microsoft Visual C++ {version:0.1f} is required.z-www.microsoft.com/download/details.aspx?id=%d"@Zia64z( Get it with "Microsoft Windows SDK 7.0"z% Get it from http://aka.ms/vcpython27$@z* Get it with "Microsoft Windows SDK 7.1": iW r0z[ Get it with "Build Tools for Visual Studio": https://visualstudio.microsoft.com/downloads/N)r-lowerformatlocalsfind)r/rr,messageZtmplZ msdownloadrrrr*s   r*c@sbeZdZdZeddZddZe ddZ dd Z d d Z dd dZ dddZdddZdS) PlatformInfoz Current and Target Architectures information. Parameters ---------- arch: str Target architecture. Zprocessor_architecturercCs|dd|_dS)Nx64amd64)r:replacer,)selfr,rrr__init__szPlatformInfo.__init__cCs|j|jdddS)zs Return Target CPU architecture. Return ------ str Target CPU _r N)r,r=rCrrr target_cpus zPlatformInfo.target_cpucCs |jdkS)z Return True if target CPU is x86 32 bits.. Return ------ bool CPU is x86 32 bits r"rGrFrrr target_is_x86s zPlatformInfo.target_is_x86cCs |jdkS)z Return True if current CPU is x86 32 bits.. Return ------ bool CPU is x86 32 bits r" current_cpurFrrrcurrent_is_x86s zPlatformInfo.current_is_x86FcCs.|jdkr|rdS|jdkr$|r$dSd|jS)uk Current platform specific subfolder. Parameters ---------- hidex86: bool return '' and not '†' if architecture is x86. x64: bool return 'd' and not 'md64' if architecture is amd64. Return ------ str subfolder: ' arget', or '' (see hidex86 parameter) r"rrA\x64\%srJrChidex86r@rrr current_dirszPlatformInfo.current_dircCs.|jdkr|rdS|jdkr$|r$dSd|jS)ar Target platform specific subfolder. Parameters ---------- hidex86: bool return '' and not '\x86' if architecture is x86. x64: bool return '\x64' and not '\amd64' if architecture is amd64. Return ------ str subfolder: '\current', or '' (see hidex86 parameter) r"rrArMrNrHrOrrr target_dir(szPlatformInfo.target_dircCs0|rdn|j}|j|krdS|dd|S)ap Cross platform specific subfolder. Parameters ---------- forcex86: bool Use 'x86' as current architecture even if current architecture is not x86. Return ------ str subfolder: '' if target architecture is current architecture, '\current_target' if not. r"r\z\%s_)rKrGrRrB)rCforcex86Zcurrentrrr cross_dir>szPlatformInfo.cross_dirN)FF)FF)F)rrr__doc__rgetr:rKrDpropertyrGrIrLrQrRrUrrrrr?s    r?c@seZdZdZejejejejfZ ddZ e ddZ e ddZ e dd Ze d d Ze d d Ze ddZe ddZe ddZe ddZdddZddZdS) RegistryInfoz Microsoft Visual Studio related registry information. Parameters ---------- platform_info: PlatformInfo "PlatformInfo" instance. cCs ||_dSN)pi)rCZ platform_inforrrrDcszRegistryInfo.__init__cCsdS)z Microsoft Visual Studio root registry key. Return ------ str Registry key Z VisualStudiorrFrrr visualstudiofs zRegistryInfo.visualstudiocCs t|jdS)z Microsoft Visual Studio SxS registry key. Return ------ str Registry key ZSxS)rr\rFrrrsxsrs zRegistryInfo.sxscCs t|jdS)z| Microsoft Visual C++ VC7 registry key. Return ------ str Registry key ZVC7rr]rFrrrvc~s zRegistryInfo.vccCs t|jdS)z Microsoft Visual Studio VS7 registry key. Return ------ str Registry key ZVS7r^rFrrrvss zRegistryInfo.vscCsdS)z Microsoft Visual C++ for Python registry key. Return ------ str Registry key zDevDiv\VCForPythonrrFrrr vc_for_pythons zRegistryInfo.vc_for_pythoncCsdS)zq Microsoft SDK registry key. Return ------ str Registry key zMicrosoft SDKsrrFrrr microsoft_sdks zRegistryInfo.microsoft_sdkcCs t|jdS)z Microsoft Windows/Platform SDK registry key. Return ------ str Registry key r rrbrFrrr windows_sdks zRegistryInfo.windows_sdkcCs t|jdS)z Microsoft .NET Framework SDK registry key. Return ------ str Registry key ZNETFXSDKrcrFrrr netfx_sdks zRegistryInfo.netfx_sdkcCsdS)z Microsoft Windows Kits Roots registry key. Return ------ str Registry key zWindows Kits\Installed RootsrrFrrrwindows_kits_rootss zRegistryInfo.windows_kits_rootsFcCs$|js|rdnd}td|d|S)a Return key in Microsoft software registry. Parameters ---------- key: str Registry key path where look. x86: str Force x86 software registry. Return ------ str Registry key rZ Wow6432NodeZSoftwareZ Microsoft)r[rLr)rCrr"Znode64rrr microsoftszRegistryInfo.microsoftc Cstj}tj}|j}|jD]}z||||d|}Wn`ttfk r|jsz||||dd|}Wqttfk rYYqYqXnYqYnXzt ||dWSttfk rYqXqdS)a Look for values in registry in Microsoft software registry. Parameters ---------- key: str Registry key path where look. name: str Value name to find. Return ------ str value rTN) rKEY_READOpenKeyrgHKEYSOSErrorIOErrorr[rLZ QueryValueEx)rCrnameZkey_readZopenkeymshkeybkeyrrrlookups"   zRegistryInfo.lookupN)F)rrrrVrrrrrrjrDrXr\r]r_r`rarbrdrerfrgrqrrrrrYUs6         rYc@s<eZdZdZeddZeddZedeZd7ddZ d d Z d d Z d dZ e ddZeddZeddZddZddZeddZeddZeddZedd Zed!d"Zed#d$Zed%d&Zed'd(Zed)d*Zed+d,Zed-d.Zed/d0Zed1d2Z d3d4Z!e d8d5d6Z"dS)9 SystemInfoz Microsoft Windows and Visual Studio related system information. Parameters ---------- registry_info: RegistryInfo "RegistryInfo" instance. vc_ver: float Required Microsoft Visual C++ version. WinDirr ProgramFileszProgramFiles(x86)NcCs2||_|jj|_||_|p$||_|_dSrZ)rir[find_programdata_vs_versknown_vs_paths_find_latest_available_vs_vervs_vervc_ver)rCZ registry_inforzrrrrDs    zSystemInfo.__init__cCs>|}|s|jstjdt|}||jt|dS)zm Find the latest VC version Return ------ float version z%No Microsoft Visual C++ version foundr8)find_reg_vs_versrwr$r%r&setupdatesorted)rCZ reg_vc_versZvc_versrrrrx%s   z(SystemInfo._find_latest_available_vs_verc Cs$|jj}|jj|jj|jjf}g}|jjD]}|D]}zt|||dtj}Wnt t fk rlYq2YnXt |\}}} t |D]D} z*t t|| d} | |kr|| Wqtk rYqXqt |D]B} z&t t|| } | |kr|| Wqtk rYqXqq2q*t|S)z Find Microsoft Visual Studio versions available in registry. Return ------ list of float Versions r)rurgr_rar`rjrrirhrkrlZ QueryInfoKeyrangefloatZ EnumValueappendr'ZEnumKeyr~) rCrnZvckeysZvs_versrorrpZsubkeysvaluesrEir+rrrr{8s2      zSystemInfo.find_reg_vs_versc Csi}d}z t|}Wnttfk r0|YSX|D]}z\t||d}t|ddd}t|}W5QRX|d}tt|d||||d<Wq6tttfk rYq6Yq6Xq6|S) z Find Visual studio 2017+ versions from information in "C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances". Return ------ dict float version as key, path as value. z9C:\ProgramData\Microsoft\VisualStudio\Packages\_Instancesz state.jsonZrtzutf-8)encodingZinstallationPath VC\Tools\MSVCZinstallationVersion) rrkrlrrjsonload_as_float_versionr) rCZ vs_versionsZ instances_dirZ hashed_namesrmZ state_pathZ state_filestateZvs_pathrrrrv[s*     z#SystemInfo.find_programdata_vs_verscCstd|dddS)z Return a string version as a simplified float version (major.minor) Parameters ---------- version: str Version. Return ------ float version .N)rrsplit)rrrrrszSystemInfo._as_float_versioncCs.t|jd|j}|j|jjd|jp,|S)zp Microsoft Visual Studio directory. Return ------ str path zMicrosoft Visual Studio %0.1f%0.1f)rProgramFilesx86ryrurqr`)rCdefaultrrr VSInstallDirs zSystemInfo.VSInstallDircCs,|p|}t|s(d}tj||S)zm Microsoft Visual C++ directory. Return ------ str path z(Microsoft Visual C++ directory not found) _guess_vc_guess_vc_legacyrr$r%r&)rCpathmsgrrr VCInstallDirs  zSystemInfo.VCInstallDirc Cs|jdkrdSz|j|j}Wntk r8|j}YnXt|d}z$t|d}|||_t||WStt t fk rYdSXdS)zl Locate Visual C++ for VS2017+. Return ------ str path r0rrr8N) ryrwrrrrrrzrkrl IndexError)rCZvs_dirZguess_vcrzrrrrs      zSystemInfo._guess_vccCsbt|jd|j}t|jjd|j}|j|d}|rBt|dn|}|j|jjd|jp`|S)z{ Locate Visual C++ for versions prior to 2017. Return ------ str path z Microsoft Visual Studio %0.1f\VCrrZVC)rrryrurarqr_)rCrZreg_pathZ python_vcZ default_vcrrrrs zSystemInfo._guess_vc_legacycCsJ|jdkrdS|jdkrdS|jdkr*dS|jdkr8dS|jd krFd Sd S) z Microsoft Windows SDK versions for specified MSVC++ version. Return ------ tuple of str versions r7)z7.0z6.1z6.0ar9)z7.1z7.0a&@)z8.0z8.0a(@)8.1z8.1ar0)z10.0rNryrFrrrWindowsSdkVersions     zSystemInfo.WindowsSdkVersioncCs|t|jdS)zt Microsoft Windows SDK last version. Return ------ str version lib)_use_last_dir_namer WindowsSdkDirrFrrrWindowsSdkLastVersions z SystemInfo.WindowsSdkLastVersioncCs d}|jD],}t|jjd|}|j|d}|r q8q |rDt|stt|jjd|j}|j|d}|rtt|d}|rt|s|jD]6}|d|d}d |}t|j |}t|r|}q|rt|s|jD]$}d |}t|j |}t|r|}q|st|j d }|S) zn Microsoft Windows SDK directory. Return ------ str path rzv%sinstallationfolderrrZWinSDKNrzMicrosoft SDKs\Windows Kits\%szMicrosoft SDKs\Windows\v%sZ PlatformSDK) rrrurdrqrrarzrfindrtr)rCsdkdirr+locrZ install_baseZintverdrrrr s6           zSystemInfo.WindowsSdkDirc Cs|jdkrd}d}n&d}|jdkr&dnd}|jjd|d}d ||d d f}g}|jd kr~|jD]}|t|jj||g7}qb|jD]}|t|jj d ||g7}q|D]}|j |d}|r|SqdS)zy Microsoft Windows SDK executable directory. Return ------ str path r#r(rTF)r@rPzWinSDK-NetFx%dTools%srS-r0zv%sArN) ryr[rQrBNetFxSdkVersionrrurerrdrq) rCZnetfxverr,rPZfxZregpathsr+rZexecpathrrrWindowsSDKExecutablePath7s"    z#SystemInfo.WindowsSDKExecutablePathcCs&t|jjd|j}|j|dp$dS)zl Microsoft Visual F# directory. Return ------ str path z%0.1f\Setup\F#r r)rrur\ryrq)rCrrrrFSharpInstallDirZs zSystemInfo.FSharpInstallDircCsF|jdkrdnd}|D]*}|j|jjd|}|r|p:dSqdS)zt Microsoft Universal CRT SDK directory. Return ------ str path r0)Z10Z81rz kitsroot%srN)ryrurqrf)rCZversr+rrrrUniversalCRTSdkDirgs  zSystemInfo.UniversalCRTSdkDircCs|t|jdS)z Microsoft Universal C Runtime SDK last version. Return ------ str version r)rrrrFrrrUniversalCRTSdkLastVersion{s z%SystemInfo.UniversalCRTSdkLastVersioncCs|jdkrdSdS)z Microsoft .NET Framework SDK versions. Return ------ tuple of str versions r0) z4.7.2z4.7.1z4.7z4.6.2z4.6.1z4.6z4.5.2z4.5.1z4.5rrrFrrrrszSystemInfo.NetFxSdkVersioncCs8d}|jD](}t|jj|}|j|d}|r q4q |S)zu Microsoft .NET Framework SDK directory. Return ------ str path rZkitsinstallationfolder)rrrurerq)rCrr+rrrr NetFxSdkDirs  zSystemInfo.NetFxSdkDircCs"t|jd}|j|jjdp |S)zw Microsoft .NET Framework 32bit directory. Return ------ str path zMicrosoft.NET\FrameworkZframeworkdir32rrsrurqr_rCZguess_fwrrrFrameworkDir32s zSystemInfo.FrameworkDir32cCs"t|jd}|j|jjdp |S)zw Microsoft .NET Framework 64bit directory. Return ------ str path zMicrosoft.NET\Framework64Zframeworkdir64rrrrrFrameworkDir64s zSystemInfo.FrameworkDir64cCs |dS)z Microsoft .NET Framework 32bit versions. Return ------ tuple of str versions _find_dot_net_versionsrFrrrFrameworkVersion32s zSystemInfo.FrameworkVersion32cCs |dS)z Microsoft .NET Framework 64bit versions. Return ------ tuple of str versions @rrFrrrFrameworkVersion64s zSystemInfo.FrameworkVersion64cCs|j|jjd|}t|d|}|p6||dp6d}|jdkrJ|dfS|jdkrt|dd d krld n|d fS|jd krdS|jdkrdSdS)z Find Microsoft .NET Framework versions. Parameters ---------- bits: int Platform number of bits: 32 or 64. Return ------ tuple of str versions zframeworkver%dzFrameworkDir%dvrrzv4.0r9NrZv4z v4.0.30319v3.5r7)r v2.0.50727g @)zv3.0r)rurqr_getattrrryr:)rCbitsZreg_verZ dot_net_dirr+rrrrs     z!SystemInfo._find_dot_net_versionscs*fddttD}t|dp(dS)a) Return name of the last dir in path or '' if no dir found. Parameters ---------- path: str Use dirs in this path prefix: str Use only dirs starting by this prefix Return ------ str name c3s*|]"}tt|r|r|VqdSrZ)rr startswith).0Zdir_namerprefixrr s z0SystemInfo._use_last_dir_name..Nr)reversedrnext)rrZ matching_dirsrrrrs  zSystemInfo._use_last_dir_name)N)r)#rrrrVrrWrsrtrrDrxr{rv staticmethodrrXrrrrrrrrrrrrrrrrrrrrrrrrr sZ    #*      * "         rrc@sbeZdZdZd?ddZeddZedd Zed d Zed d Z eddZ eddZ eddZ eddZ eddZeddZeddZddZeddZed d!Zed"d#Zed$d%Zed&d'Zed(d)Zed*d+Zed,d-Zed.d/Zed0d1Zed2d3Zed4d5Zed6d7Zd@d9d:Zd;d<Z e!dAd=d>Z"dS)Br(aY Return environment variables for specified Microsoft Visual C++ version and platform : Lib, Include, Path and libpath. This function is compatible with Microsoft Visual C++ 9.0 to 14.X. Script created by analysing Microsoft environment configuration files like "vcvars[...].bat", "SetEnv.Cmd", "vcbuildtools.bat", ... Parameters ---------- arch: str Target architecture. vc_ver: float Required Microsoft Visual C++ version. If not set, autodetect the last version. vc_min_ver: float Minimum Microsoft Visual C++ version. NrcCsBt||_t|j|_t|j||_|j|kr>d}tj |dS)Nz.No suitable Microsoft Visual C++ version found) r?r[rYrurrsirzr$r%r&)rCr,rzr1errrrrrD0s    zEnvironmentInfo.__init__cCs|jjS)zk Microsoft Visual Studio. Return ------ float version )rryrFrrrry9s zEnvironmentInfo.vs_vercCs|jjS)zp Microsoft Visual C++ version. Return ------ float version )rrzrFrrrrzEs zEnvironmentInfo.vc_vercsVddg}jdkrDjjddd}|dg7}|dg7}|d|g7}fd d |DS) zu Microsoft Visual Studio Tools. Return ------ list of str paths z Common7\IDEz Common7\Toolsr0TrPr@z1Common7\IDE\CommonExtensions\Microsoft\TestWindowzTeam Tools\Performance ToolszTeam Tools\Performance Tools%scsg|]}tjj|qSrrrrrrrFrr csz+EnvironmentInfo.VSTools..)ryr[rQ)rCpaths arch_subdirrrFrVSToolsQs    zEnvironmentInfo.VSToolscCst|jjdt|jjdgS)z Microsoft Visual C++ & Microsoft Foundation Class Includes. Return ------ list of str paths ZIncludezATLMFC\IncluderrrrFrrr VCIncludeses  zEnvironmentInfo.VCIncludescsbjdkrjjdd}njjdd}d|d|g}jdkrP|d|g7}fd d |DS) z Microsoft Visual C++ & Microsoft Foundation Class Libraries. Return ------ list of str paths .@Tr@rPLib%sz ATLMFC\Lib%sr0z Lib\store%scsg|]}tjj|qSrrrrFrrrsz/EnvironmentInfo.VCLibraries..)ryr[rR)rCrrrrFr VCLibrariesrs  zEnvironmentInfo.VCLibrariescCs|jdkrgSt|jjdgS)z Microsoft Visual C++ store references Libraries. Return ------ list of str paths r0zLib\store\references)ryrrrrFrrr VCStoreRefss zEnvironmentInfo.VCStoreRefscCs|j}t|jdg}|jdkr"dnd}|j|}|rL|t|jd|g7}|jdkr|d|jjdd}|t|j|g7}n|jdkr|jrd nd }|t|j||jjdd g7}|jj |jj kr|t|j||jjdd g7}n|t|jd g7}|S) zr Microsoft Visual C++ Tools. Return ------ list of str paths Z VCPackagesr9TFBin%sr0rrz bin\HostX86%sz bin\HostX64%srBin) rrrryr[rUrQrLrRrKrG)rCrtoolsrTrrZhost_dirrrrVCToolss0     zEnvironmentInfo.VCToolscCsh|jdkr.|jjddd}t|jjd|gS|jjdd}t|jjd}|j}t|d||fgSdS) zw Microsoft Windows SDK Libraries. Return ------ list of str paths r9Trrrrz%sum%sN)ryr[rRrrr _sdk_subdir)rCrrZlibverrrr OSLibrariess zEnvironmentInfo.OSLibrariescCsht|jjd}|jdkr&|t|dgS|jdkr8|j}nd}t|d|t|d|t|d|gSd S) zu Microsoft Windows SDK Include. Return ------ list of str paths includer9Zglr0rz%ssharedz%sumz%swinrtN)rrrryr)rCrsdkverrrr OSIncludess      zEnvironmentInfo.OSIncludescCst|jjd}g}|jdkr&||j7}|jdkr@|t|dg7}|jdkr||t|jjdt|ddt|d dt|d dt|jjd d d |jdddg7}|S)z} Microsoft Windows SDK Libraries Paths. Return ------ list of str paths Z Referencesr7rzCommonConfiguration\Neutralr0Z UnionMetadataz'Windows.Foundation.UniversalApiContractz1.0.0.0z%Windows.Foundation.FoundationContractz,Windows.Networking.Connectivity.WwanContractZ ExtensionSDKszMicrosoft.VCLibsrZCommonConfigurationZneutral)rrrryr)rCreflibpathrrr OSLibpaths.          zEnvironmentInfo.OSLibpathcCs t|S)zs Microsoft Windows SDK Tools. Return ------ list of str paths )list _sdk_toolsrFrrrSdkToolss zEnvironmentInfo.SdkToolsccs|jdkr,|jdkrdnd}t|jj|V|js\|jjdd}d|}t|jj|V|jdkr|jrvd }n|jjddd }d |}t|jj|VnB|jdkrt|jjd}|jjdd}|jj}t|d ||fV|jj r|jj Vd S)z Microsoft Windows SDK Tools paths generator. Return ------ generator of str paths rrrzBin\x86Trr)r9rrrzBin\NETFX 4.0 Tools%sz%s%sN) ryrrrr[rLrQrIrr)rCZbin_dirrrrrrrrs(     zEnvironmentInfo._sdk_toolscCs|jj}|rd|SdS)zu Microsoft Windows SDK version subdir. Return ------ str subdir %s\r)rrrCucrtverrrrr6s zEnvironmentInfo._sdk_subdircCs|jdkrgSt|jjdgS)zs Microsoft Windows SDK Setup. Return ------ list of str paths r7ZSetup)ryrrrrFrrrSdkSetupCs zEnvironmentInfo.SdkSetupcs|j}|j|jdkr0d}| o,| }n$|p>|}|jdkpR|jdk}g}|rt|fddjD7}|r|fddjD7}|S)zv Microsoft .NET Framework Tools. Return ------ list of str paths r9TrAcsg|]}tj|qSr)rrrr+rrrrhsz+EnvironmentInfo.FxTools..csg|]}tj|qSr)rrrrrrrks) r[rryrIrLrKrGrr)rCr[Z include32Z include64rrrrFxToolsRs"    zEnvironmentInfo.FxToolscCs8|jdks|jjsgS|jjdd}t|jjd|gS)z~ Microsoft .Net Framework SDK Libraries. Return ------ list of str paths r0Trzlib\um%s)ryrrr[rRr)rCrrrrNetFxSDKLibrariesos z!EnvironmentInfo.NetFxSDKLibrariescCs&|jdks|jjsgSt|jjdgS)z} Microsoft .Net Framework SDK Includes. Return ------ list of str paths r0z include\um)ryrrrrFrrrNetFxSDKIncludess z EnvironmentInfo.NetFxSDKIncludescCst|jjdgS)z Microsoft Visual Studio Team System Database. Return ------ list of str paths z VSTSDB\DeployrrFrrrVsTDbs zEnvironmentInfo.VsTDbcCsv|jdkrgS|jdkr0|jj}|jjdd}n |jj}d}d|j|f}t||g}|jdkrr|t||dg7}|S)zn Microsoft Build Engine. Return ------ list of str paths rrTrrzMSBuild\%0.1f\bin%sZRoslyn)ryrrr[rQrr)rC base_pathrrZbuildrrrMSBuilds    zEnvironmentInfo.MSBuildcCs|jdkrgSt|jjdgS)zt Microsoft HTML Help Workshop. Return ------ list of str paths rzHTML Help Workshop)ryrrrrFrrrHTMLHelpWorkshops z EnvironmentInfo.HTMLHelpWorkshopcCsD|jdkrgS|jjdd}t|jjd}|j}t|d||fgS)z Microsoft Universal C Runtime SDK Libraries. Return ------ list of str paths r0Trrz%sucrt%s)ryr[rRrrr _ucrt_subdir)rCrrrrrr UCRTLibrariess zEnvironmentInfo.UCRTLibrariescCs.|jdkrgSt|jjd}t|d|jgS)z Microsoft Universal C Runtime SDK Include. Return ------ list of str paths r0rz%sucrt)ryrrrr)rCrrrr UCRTIncludess zEnvironmentInfo.UCRTIncludescCs|jj}|rd|SdS)z Microsoft Universal C Runtime SDK version subdir. Return ------ str subdir rr)rrrrrrrs zEnvironmentInfo._ucrt_subdircCs(d|jkrdkrnngS|jjgS)zk Microsoft Visual F#. Return ------ list of str paths rr)ryrrrFrrrFSharps zEnvironmentInfo.FSharpc Csd|j}|jjddd}g}|jj}t|dd}t|rft |t |d}||t |dg7}|t |d g7}d |jd d t |j d f}t ||D]&\}}t ||||} t| r| Sqd S) z Microsoft Visual C++ runtime redistributable dll. Return ------ str path zvcruntime%d0.dllTrrSz\Toolsz\Redistr8ZonecoreZredistzMicrosoft.VC%d.CRT N)rzr[rRstriprrrrBrrrintry itertoolsproductr) rCZ vcruntimerprefixesZ tools_pathZ redist_pathZcrt_dirsrZcrt_dirrrrrVCRuntimeRedists  zEnvironmentInfo.VCRuntimeRedistTcCst|d|j|j|j|jg||d|j|j|j|j |j g||d|j|j|j |j g||d|j |j|j|j|j|j|j|j|jg |d}|jdkrt|jr|j|d<|S)z Return environment dict. Parameters ---------- exists: bool It True, only return existing paths. Return ------ dict environment rrrr)rrrrZpy_vcruntime_redist)dict _build_pathsrrrrrrrrrrrrrrrrrrrryrr)rCexistsenvrrrr)&sV   zEnvironmentInfo.return_envc Csptj|}t|dt}t||}|r A B C D _unique_everseen('ABBCcAD', str.lower) --> A B C D N)r|addr __contains__)iterablerseenZseen_addZelementkrrrrzs  z EnvironmentInfo._unique_everseen)Nr)T)N)#rrrrVrDrXryrzrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrrrr(sn        $    #             " 2"r()r")r)*rVriorosrrZos.pathrrrrr3platformrZdistutils.errorsr$Z#setuptools.extern.packaging.versionr Zsetuptools.extern.six.movesr Zmonkeyr systemrrr ImportErrorr%r&Z_msvc9_suppress_errorsZdistutils.msvc9compilerrrr#r2r6r*r?rYrrr(rrrrsJ       * &  $s5site-packages/setuptools/__pycache__/extension.cpython-38.pyc000064400000003651151464452710020377 0ustar00U Qab@s|ddlZddlZddlZddlZddlZddlmZddlm Z ddZ e Z e ej j ZGdddeZ Gd d d e ZdS) N)map) get_unpatchedcCs4d}zt|dgdjWdStk r.YnXdS)z0 Return True if Cython can be imported. zCython.Distutils.build_ext build_ext)fromlistTF) __import__r Exception)Z cython_implr 8/usr/lib/python3.8/site-packages/setuptools/extension.py _have_cython sr c@s eZdZdZddZddZdS) Extensionz7Extension that uses '.c' files in place of '.pyx' filescOs(|dd|_tj|||f||dS)Npy_limited_apiF)popr _Extension__init__)selfnamesourcesargskwr r r r#szExtension.__init__cCsNtr dS|jpd}|dkr$dnd}ttjd|}tt||j |_ dS)z Replace sources with .pyx extensions to sources with the target language extension. This mechanism allows language authors to supply pre-converted sources but to prefer the .pyx sources. Nzc++z.cppz.cz.pyx$) r Zlanguagelower functoolspartialresublistrr)rZlangZ target_extrr r r _convert_pyx_sources_to_lang)s  z&Extension._convert_pyx_sources_to_langN)__name__ __module__ __qualname____doc__rrr r r r r sr c@seZdZdZdS)Libraryz=Just like a regular Extension, but built as a library insteadN)rrr r!r r r r r"8sr")rrZdistutils.coreZ distutilsZdistutils.errorsZdistutils.extensionZsetuptools.extern.six.movesrZmonkeyrr Z have_pyrexZcorer rr"r r r r s   site-packages/setuptools/__pycache__/py33compat.cpython-38.pyc000064400000002600151464452710020356 0ustar00U Qab2@sddlZddlZddlZz ddlZWnek r<dZYnXddlmZddlmZe Z e ddZ GdddZ eede Zeed dZedkrejZdS) N)six) html_parserOpArgz opcode argc@seZdZddZddZdS)Bytecode_compatcCs ||_dS)N)code)selfrr9/usr/lib/python3.8/site-packages/setuptools/py33compat.py__init__szBytecode_compat.__init__ccstd|jj}t|jj}d}d}||kr||}|tjkr||d||dd|}|d7}|tjkrtjd}||d}q$n d }|d7}t ||Vq$d S) z>Yield '(op,arg)' pair for each operation in code object 'code'briN) arrayrco_codelendisZ HAVE_ARGUMENTZ EXTENDED_ARGrZ integer_typesr)rbyteseofZptrZ extended_argopargZ long_typerrr __iter__s       zBytecode_compat.__iter__N)__name__ __module__ __qualname__r rrrrr rsrBytecodeunescape)rr collectionsZhtml ImportErrorZsetuptools.externrZsetuptools.extern.six.movesrtypeZ __metaclass__ namedtuplerrgetattrrrZ HTMLParserrrrr s     "  site-packages/setuptools/__pycache__/glob.cpython-38.pyc000064400000007225151464452710017307 0ustar00U Qab@sdZddlZddlZddlZdddgZdddZdddZd d Zd d Zd dZ ddZ ddZ e dZ e dZddZddZddZdS)z Filename globbing utility. Mostly a copy of `glob` from Python 3.5. Changes include: * `yield from` and PEP3102 `*` removed. * Hidden files are not ignored. NglobiglobescapeFcCstt||dS)ayReturn a list of paths matching a pathname pattern. The pattern may contain simple shell-style wildcards a la fnmatch. However, unlike fnmatch, filenames starting with a dot are special cases that are not matched by '*' and '?' patterns. If recursive is true, the pattern '**' will match any files and zero or more directories and subdirectories. ) recursive)listr)pathnamerr3/usr/lib/python3.8/site-packages/setuptools/glob.pyrs cCs*t||}|r&t|r&t|}|r&t|S)aReturn an iterator which yields the paths matching a pathname pattern. The pattern may contain simple shell-style wildcards a la fnmatch. However, unlike fnmatch, filenames starting with a dot are special cases that are not matched by '*' and '?' patterns. If recursive is true, the pattern '**' will match any files and zero or more directories and subdirectories. )_iglob _isrecursivenextAssertionError)rritsrrr rs  ccstj|\}}t|sF|r0tj|rB|Vntj|rB|VdS|s|rnt|rnt||D] }|Vq`nt||D] }|VqxdS||krt|rt ||}n|g}t|r|rt|rt}qt}nt }|D]$}|||D]}tj ||VqqdSN) ospathsplit has_magiclexistsisdirr glob2glob1r glob0join)rrdirnamebasenamexdirsZ glob_in_dirnamerrr r 0s4      r cCsV|s"t|trtjd}ntj}zt|}Wntk rHgYSXt||SNASCII) isinstancebytesrcurdirencodelistdirOSErrorfnmatchfilter)rpatternnamesrrr r[s  rcCs8|stj|r4|gSntjtj||r4|gSgSr)rrrrr)rrrrr rhs  rccs2t|s t|ddVt|D] }|Vq"dS)Nr)r r _rlistdir)rr*rrrr rxs  rccs|s"t|trtjd}ntj}zt|}Wntjk rHYdSX|D]>}|V|rjtj||n|}t |D]}tj||VqvqNdSr ) r"r#rr$r%r&errorrrr,)rr+rryrrr r,s  r,z([*?[])s([*?[])cCs(t|trt|}n t|}|dk Sr)r"r#magic_check_bytessearch magic_check)rmatchrrr rs   rcCst|tr|dkS|dkSdS)Ns**z**)r"r#)r*rrr r s r cCs<tj|\}}t|tr(td|}n td|}||S)z#Escape all special characters. s[\1]z[\1])rr splitdriver"r#r/subr1)rZdriverrr rs   )F)F)__doc__rrer(__all__rrr rrrr,compiler1r/rr rrrrr s    +   site-packages/setuptools/__pycache__/py31compat.cpython-38.opt-1.pyc000064400000002247151464452710021322 0ustar00U QabF@sPgZeZzddlmZWn2ek rJddlZddlZGdddZYnXdS))TemporaryDirectoryNc@s(eZdZdZddZddZddZdS) rz Very simple temporary directory context manager. Will try to delete afterward, but will also ignore OS and similar errors on deletion. cKsd|_tjf||_dSN)nametempfileZmkdtemp)selfkwargsr9/usr/lib/python3.8/site-packages/setuptools/py31compat.py__init__szTemporaryDirectory.__init__cCs|jSr)r)rrrr __enter__szTemporaryDirectory.__enter__cCs2zt|jdWntk r&YnXd|_dS)NT)shutilZrmtreerOSError)rexctypeZexcvalueZexctracerrr __exit__s zTemporaryDirectory.__exit__N)__name__ __module__ __qualname____doc__r r rrrrr r sr)__all__typeZ __metaclass__rr ImportErrorr rrrr ssite-packages/setuptools/__pycache__/depends.cpython-38.opt-1.pyc000064400000012140151464452710020735 0ustar00U Qab@sddlZddlZddlZddlmZddlmZddlmZm Z m Z m Z ddl mZddd d gZ Gd ddZd d Zddd Zddd ZddZedS)N) StrictVersion)Bytecode) find_module PY_COMPILED PY_FROZEN PY_SOURCE) py27compatRequirerget_module_constantextract_constantc@sHeZdZdZdddZddZdd Zdd d Zdd dZdddZ dS)r z7A prerequisite to building or installing a distributionNcCsF|dkr|dk rt}|dk r0||}|dkr0d}|jt|`dS)N __version__)r__dict__updatelocalsself)rnamerequested_versionmoduleZhomepage attributeformatr6/usr/lib/python3.8/site-packages/setuptools/depends.py__init__szRequire.__init__cCs |jdk rd|j|jfS|jS)z0Return full package/distribution name, w/versionNz%s-%s)rr)rrrr full_name#s zRequire.full_namecCs*|jdkp(|jdkp(t|dko(||jkS)z%Is 'version' sufficiently up-to-date?Nunknown)rrstrr)rversionrrr version_ok)szRequire.version_okrcCs|jdkrFz$t|j|\}}}|r*||WStk rDYdSXt|j|j||}|dk r|||k r||jdk r|||S|S)aGet version number of installed module, 'None', or 'default' Search 'paths' for module. If not found, return 'None'. If found, return the extracted version attribute, or 'default' if no version attribute was specified, or the value cannot be determined without importing the module. The version is formatted according to the requirement's version format (if any), unless it is 'None' or the supplied 'default'. N)rrrclose ImportErrorr r)rpathsdefaultfpivrrr get_version.s  zRequire.get_versioncCs||dk S)z/Return true if dependency is present on 'paths'N)r()rr"rrr is_presentIszRequire.is_presentcCs ||}|dkrdS||S)z>Return true if dependency is present and up-to-date on 'paths'NF)r(r)rr"rrrr is_currentMs zRequire.is_current)r NN)Nr)N)N) __name__ __module__ __qualname____doc__rrrr(r)r*rrrrr s   cCs"tjdd}|s|St|S)Ncss dVdS)NrrrrremptyVszmaybe_close..empty) contextlibcontextmanagerclosing)r$r/rrr maybe_closeUs  r3c Cszt||\}}\}}}} Wntk r4YdSXt|z|tkr^|dt|} nV|tkrtt ||} n@|t krt ||d} n&t ||| } t | |dW5QRSW5QRXt| ||S)zFind 'module' by searching 'paths', and extract 'symbol' Return 'None' if 'module' does not exist on 'paths', or it does not define 'symbol'. If the module defines 'symbol' as a constant, return the constant. Otherwise, return 'default'.Nexec)rr!r3rreadmarshalloadrr get_frozen_objectrcompileZ get_modulegetattrr ) rsymbolr#r"r$pathsuffixmodeZkindinfocodeZimportedrrrr `s   "c Cs||jkrdSt|j|}d}d}d}|}t|D]H}|j} |j} | |krZ|j| }q6| |krz| |ksr| |krz|S|}q6dS)aExtract the constant value of 'symbol' from 'code' If the name 'symbol' is bound to a constant value by the Python code object 'code', return that value. If 'symbol' is bound to an expression, return 'default'. Otherwise, return 'None'. Return value is based on the first assignment to 'symbol'. 'symbol' must be a global, or at least a non-"fast" local in the code block. That is, only 'STORE_NAME' and 'STORE_GLOBAL' opcodes are checked, and 'symbol' must be present in 'code.co_names'. NZad)co_nameslistindexrZopcodearg co_consts) rBr=r#Zname_idxZ STORE_NAMEZ STORE_GLOBALZ LOAD_CONSTconstZ byte_codeoprIrrrr }s   cCs>tjdstjdkrdSd}|D]}t|=t|q"dS)z Patch the globals to remove the objects not available on some platforms. XXX it'd be better to test assertions about bytecode instead. javaZcliN)r r )sysplatform startswithglobals__all__remove)Z incompatiblerrrr_update_globalss rT)r4N)r4)rNr8r0Zdistutils.versionrZ py33compatrr rrrrr rRr r3r r rTrrrrs"   D  $site-packages/setuptools/__pycache__/dist.cpython-38.opt-1.pyc000064400000122205151464452710020262 0ustar00U Qab@sdgZddlZddlZddlZddlZddlZddlZddlZddl Zddl Zddl Zddl m Z ddlmZddlmZddlZddlmZddlmZddlmZmZmZdd l mZdd lmZdd lmZdd lm Z dd lm!Z!ddl"m#Z#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/ddl0Z0e1de1dddZ2ddZ3ddZ4ddZ5e6e7fZ8dd Z9d!d"Z:d#d$Z;d%d&Zd+d,Z?d-d.Z@d/d0ZAd1d2ZBd3d4ZCd5d6ZDe-ejEjFZGGd7ddeGZFGd8d9d9ZHGd:d;d;e'ZIdS)< DistributionN) strtobool)DEBUGtranslate_longopt) defaultdict)message_from_file)DistutilsOptionErrorDistutilsPlatformErrorDistutilsSetupError) rfc822_escape) StrictVersion)six) packaging) ordered_set)mapfilter filterfalse)SetuptoolsDeprecationWarning)Require)windows_support) get_unpatched)parse_configurationz&setuptools.extern.packaging.specifiersz#setuptools.extern.packaging.versioncCstdtt|S)NzDo not call this function)warningswarnDistDeprecationWarningr)clsr3/usr/lib/python3.8/site-packages/setuptools/dist.py_get_unpatched-s r cCst|dd}|dkr|js |jr*td}nd|jdk sT|jdk sTt|dddk sT|jr^td}n0|js||js||j s||j s||j rtd}ntd}||_ |S)Nmetadata_versionz2.1python_requires1.21.1z1.0) getattrlong_description_content_typeprovides_extrasr maintainermaintainer_email project_urlsprovidesrequires obsoletes classifiers download_urlr!)selfZmvrrrget_metadata_version2s*      r1cs t|fdd}fdd}td|_|d|_|d|_|d|_|d |_d |_|d |_d |_ |d |_ |d |_ dkr|d|_ nd |_ |d|_ |d|_dkr|dd|_|d|_|d|_|jtdkr |d|_|d|_|d|_nd |_d |_d |_d S)z-Reads the metadata values from a file object.cs|}|dkrdS|S)NZUNKNOWNr)namevaluemsgrr _read_fieldLsz"read_pkg_file.._read_fieldcs|d}|gkrdS|SN)Zget_all)r2valuesr4rr _read_listRs z!read_pkg_file.._read_listzmetadata-versionr2versionZsummaryauthorNz author-emailz home-pagelicensez download-url descriptionkeywords,platformZ classifierr$r,r+r-)rr r!r2r:r=r;r( author_emailr)Zurlr<r/Zlong_descriptionsplitr>Z platformsr.r,r+r-)r0filer6r9rr4r read_pkg_fileHs:                 rDc s}tjrfdd}n fdd}|dt||d|d|d|d|td kr|d  |d  n.d }|D]$\}}t |}|d k r|||q|d j r|dj jD]}|dd|qt} |d| d} | r:|d| |td krdD]} |d| qPndddddtdr|djjr|djjr jD]} |d| qd S)z5Write the PKG-INFO format data to a file object. csd||fdSNz%s: %s )writeZ _encode_fieldkeyr3rCr0rr write_fieldsz#write_pkg_file..write_fieldcsd||fdSrE)rFrG)rCrrrJszMetadata-VersionNameVersionZSummaryz Home-pager#Author Author-email))rMr;)rNrA)Z Maintainerr()zMaintainer-emailr)NZLicensez Download-URLz Project-URLz%s, %sZ Descriptionr?ZKeywordsZPlatformZ ClassifierZRequiresZProvidesZ Obsoletesr"zRequires-PythonzDescription-Content-TypezProvides-Extra)r1rPY2strZget_nameZ get_versionZget_descriptionZget_urlr Z get_contactZget_contact_emailr%Z get_licenser/r*itemsr Zget_long_descriptionjoinZ get_keywordsZ get_platformsZ _write_listZget_classifiersZ get_requiresZ get_providesZ get_obsoleteshasattrr"r&r') r0rCr:rJZoptional_fieldsZfieldattrZattr_valZ project_urlZ long_descr>r@extrarrIrwrite_pkg_file~sZ             rVc CsFztjd|}Wn,ttttfk r@td||fYnXdS)Nzx=z4%r must be importable 'module:attrs' string (got %r)) pkg_resources EntryPointparse TypeError ValueErrorAttributeErrorAssertionErrorr )distrTr3eprrrcheck_importablesr`c Cs6zWn,ttttfk r0td||fYnXdS)z"Verify that value is a string listz%%r must be a list of strings (got %r)N)rZr[r\r]r r^rTr3rrrassert_string_lists  rbcCsd|}t||||D]J}||s2tdd||d\}}}|r||krtjd||qdS)z(Verify that namespace packages are validz1Distribution contains no modules or packages for znamespace package %r.z^WARNING: %r is declared as a package namespace, but %r is not: please correct this in setup.pyN)rbhas_contents_forr rpartition distutilslogr)r^rTr3Z ns_packagesZnspparentsepZchildrrr check_nsps    rjc Cs@zttt|Wn"tttfk r:tdYnXdS)z+Verify that extras_require mapping is validz'extras_require' must be a dictionary whose values are strings or lists of strings containing valid project/version requirement specifiers.N) list itertoolsstarmap _check_extrarQrZr[r\r rarrr check_extrass rocCs<|d\}}}|r*t|r*td|tt|dS)N:zInvalid environment marker: ) partitionrWZinvalid_markerr rkparse_requirements)rUZreqsr2rimarkerrrrrns rncCs&t||kr"d}t|j||ddS)z)Verify that value is True, False, 0, or 1z0{attr!r} must be a boolean value (got {value!r}))rTr3N)boolr format)r^rTr3tmplrrr assert_bool s rwc Csjz(tt|t|ttfr&tdWn<ttfk rd}zd}t|j ||dW5d}~XYnXdS)z9Verify that install_requires is a valid requirements listzUnordered types are not allowedzm{attr!r} must be a string or list of strings containing valid project/version requirement specifiers; {error}rTerrorN) rkrWrr isinstancedictsetrZr[r rur^rTr3ryrvrrrcheck_requirementss r~c CsRztj|Wn<tjjk rL}zd}t|j||dW5d}~XYnXdS)z.Verify that value is a valid version specifierzF{attr!r} must be a string containing valid version specifiers; {error}rxN)rZ specifiersZ SpecifierSetZInvalidSpecifierr rur}rrrcheck_specifier s rc Cs@ztj|Wn*tk r:}z t|W5d}~XYnXdS)z)Verify that entry_points map is parseableN)rWrXZ parse_mapr[r )r^rTr3errrcheck_entry_points,srcCst|tjstddS)Nztest_suite must be a string)rzr string_typesr rarrrcheck_test_suite4s rcCs\t|tstd||D]6\}}t|tjsDtd||t|d||q dS)z@Verify that value is a dictionary of package names to glob listszT{!r} must be a dictionary mapping package names to lists of string wildcard patternsz,keys of {!r} dict must be strings (got {!r})zvalues of {!r} dictN)rzr{r rurQrrrb)r^rTr3kvrrrcheck_package_data9s  rcCs(|D]}td|stjd|qdS)Nz \w+(\.\w+)*z[WARNING: %r not a valid package name; please use only .-separated package names in setup.py)rematchrfrgr)r^rTr3Zpkgnamerrrcheck_packagesHs  rc@sReZdZdZdeejdZdZddZ dMddZ dd Z d d Z e d d ZddZddZdNddZe ddZdOddZdPddZddZddZdd Zd!d"Zd#d$Zd%d&Zd'd(Zd)d*Zd+d,Zd-d.Zd/d0Zd1d2Z d3d4Z!d5d6Z"d7d8Z#d9d:Z$d;d<Z%d=d>Z&d?d@Z'dAdBZ(dCdDZ)dEdFZ*dGdHZ+dIdJZ,dKdLZ-dS)QraDistribution with support for features, tests, and package data This is an enhanced version of 'distutils.dist.Distribution' that effectively adds the following new optional keyword arguments to 'setup()': 'install_requires' -- a string or sequence of strings specifying project versions that the distribution requires when installed, in the format used by 'pkg_resources.require()'. They will be installed automatically when the package is installed. If you wish to use packages that are not available in PyPI, or want to give your users an alternate download location, you can add a 'find_links' option to the '[easy_install]' section of your project's 'setup.cfg' file, and then setuptools will scan the listed web pages for links that satisfy the requirements. 'extras_require' -- a dictionary mapping names of optional "extras" to the additional requirement(s) that using those extras incurs. For example, this:: extras_require = dict(reST = ["docutils>=0.3", "reSTedit"]) indicates that the distribution can optionally provide an extra capability called "reST", but it can only be used if docutils and reSTedit are installed. If the user installs your package using EasyInstall and requests one of your extras, the corresponding additional requirements will be installed if needed. 'features' **deprecated** -- a dictionary mapping option names to 'setuptools.Feature' objects. Features are a portion of the distribution that can be included or excluded based on user options, inter-feature dependencies, and availability on the current system. Excluded features are omitted from all setup commands, including source and binary distributions, so you can create multiple distributions from the same source tree. Feature names should be valid Python identifiers, except that they may contain the '-' (minus) sign. Features can be included or excluded via the command line options '--with-X' and '--without-X', where 'X' is the name of the feature. Whether a feature is included by default, and whether you are allowed to control this from the command line, is determined by the Feature object. See the 'Feature' class for more information. 'test_suite' -- the name of a test suite to run for the 'test' command. If the user runs 'python setup.py test', the package will be installed, and the named test suite will be run. The format is the same as would be used on a 'unittest.py' command line. That is, it is the dotted name of an object to import and call to generate a test suite. 'package_data' -- a dictionary mapping package names to lists of filenames or globs to use to find data files contained in the named packages. If the dictionary has filenames or globs listed under '""' (the empty string), those names will be searched for in every package, in addition to any names for the specific package. Data files found using these names/globs will be installed along with the package, in the same location as the package. Note that globs are allowed to reference the contents of non-package subdirectories, as long as you use '/' as a path separator. (Globs are automatically converted to platform-specific paths at runtime.) In addition to these new keywords, this class also has several new methods for manipulating the distribution's contents. For example, the 'include()' and 'exclude()' methods can be thought of as in-place add and subtract commands that add or remove packages, modules, extensions, and so on from the distribution. They are used by the feature subsystem to configure the distribution for the included and excluded features. N)r&r*r'cCsl|rd|ksd|krdStt|d}tjj|}|dk rh|dshtt|d|_ ||_ dS)Nr2r:zPKG-INFO) rWZ safe_namerPlower working_setZby_keygetZ has_metadataZ safe_versionZ_version _patched_dist)r0attrsrHr^rrrpatch_missing_pkg_infosz#Distribution.patch_missing_pkg_infoc std}|si_|pi}d|ks,d|kr4tg_i_g_|dd_ ||dg_ |dg_ t dD]}t|jdqtfdd |DjD]L\}}jj|fD]}||kr||}qq|r|nd}tj||qtjjtjr4tjjj_jjdk rzHtjjj}t|} jj| krt d jj| f| j_Wn0tjj!t"fk rt d jjYnX#dS) N package_datafeaturesrequire_featuressrc_rootdependency_linkssetup_requiresdistutils.setup_keywordscs i|]\}}|jkr||qSr)_DISTUTILS_UNSUPPORTED_METADATA.0rrr0rr s z)Distribution.__init__..zNormalizing '%s' to '%s'zThe version specified (%r) is an invalid version, this may not work as expected with newer versions of setuptools, pip, and PyPI. Please see PEP 440 for more details.)$rSrFeaturewarn_deprecatedrrZ dist_filespoprrrrrWiter_entry_pointsvars setdefaultr2 _Distribution__init__rQrmetadata__dict__setattrrzr:numbersNumberrPrrLrrZInvalidVersionrZ_finalize_requires) r0rZhave_package_datar_optiondefaultsourcer3ZverZnormalized_versionrrrrs\    zDistribution.__init__cCsft|ddr|j|j_t|ddrR|jD]$}|dd}|r,|jj|q,|| dS)z Set `metadata.python_requires` and fix environment markers in `install_requires` and `extras_require`. r"Nextras_requirerpr) r%r"rrkeysrBr'add_convert_extras_requirements"_move_install_requirements_markers)r0rUrrrrs   zDistribution._finalize_requirescCsht|ddpi}tt|_|D]@\}}|j|t|D]"}||}|j|||q>q"dS)z Convert requirements in `extras_require` of the form `"extra": ["barbazquux; {marker}"]` to `"extra:{marker}": ["barbazquux"]`. rN) r%rrk_tmp_extras_requirerQrWrr _suffix_forappend)r0Z spec_ext_reqssectionrrsuffixrrrrs   z)Distribution._convert_extras_requirementscCs|jrdt|jSdS)ze For a requirement, return the 'extras_require' suffix for that requirement. rp)rsrPreqrrrr szDistribution._suffix_forcsdd}tddpd}tt|}t||}t||}ttt|_|D]}j dt|j  |qNt fddj D_dS) zv Move requirements in `install_requires` that are using environment markers `extras_require`. cSs|j Sr7rsrrrr is_simple_reqszFDistribution._move_install_requirements_markers..is_simple_reqinstall_requiresNrrpc3s,|]$\}}|ddtj|DfVqdS)cSsg|] }t|qSr)rPrrrrr )szMDistribution._move_install_requirements_markers...N)r _clean_reqrrrr (szBDistribution._move_install_requirements_markers..)r%rkrWrrrrrrPrrrsrr{rQr)r0rZspec_inst_reqsZ inst_reqsZ simple_reqsZ complex_reqsrrrrrs    z/Distribution._move_install_requirements_markerscCs d|_|S)zP Given a Requirement, remove environment markers and return it. Nr)r0rrrrr-szDistribution._clean_reqc Csddlm}tjr>tjtjkr>ddddddd d d d d ddg }ng}t|}|dkrZ|}t rh| d|}|D]}t j |dd4}t r| dj fttjr|jn|j|W5QRX|D]\}||}||} |D]>} | dkr| |kr|||| } | dd} || f| | <qq|qrd|jkr|jdD]\} \} } |j| } zF| r|t|| t|  n(| dkrt|| t| n t|| | Wn,tk r}z t|W5d}~XYnXqHdS)z Adapted from distutils.dist.Distribution.parse_config_files, this method provides the same functionality in subtly-improved ways. r) ConfigParserz install-basezinstall-platbasez install-libzinstall-platlibzinstall-purelibzinstall-headerszinstall-scriptsz install-dataprefixz exec-prefixhomeuserrootNz"Distribution.parse_config_files():utf-8)encodingz reading {filename}__name__-_global)verboseZdry_run)Z(setuptools.extern.six.moves.configparserrrPY3sysr base_prefix frozensetZfind_config_filesrannounceioopenrulocalsZ read_fileZreadfpZsectionsoptionsget_option_dict_try_strrreplacercommand_optionsrQ negative_optrrr[r )r0 filenamesrZignore_optionsparserfilenamereaderrrZopt_dictoptvalsrcaliasr5rrr_parse_config_files4s`           z Distribution._parse_config_filescCs.tjr |Sz |WStk r(YnX|S)ab On Python 2, much of distutils relies on string values being of type 'str' (bytes) and not unicode text. If the value can be safely encoded to bytes using the default encoding, prefer that. Why the default encoding? Because that value can be implicitly decoded back to text if needed. Ref #1653 )rrencodeUnicodeEncodeError)rrrrrrs  zDistribution._try_strc Cs^|}|dkr||}tr,|d||D]"\}\}}trZ|d|||fzdd|jD}Wntk rg}YnXz |j}Wntk ri}YnXz~t|t j } ||kr| rt |||t | nJ||kr| rt ||t |n,t ||rt |||ntd|||fWq4tk rV} z t| W5d} ~ XYq4Xq4dS)a Set the options for 'command_obj' from 'option_dict'. Basically this means copying elements of a dictionary ('option_dict') to attributes of an instance ('command'). 'command_obj' must be a Command instance. If 'option_dict' is not supplied, uses the standard option dictionary for this command (from 'self.command_options'). (Adopted from distutils.dist.Distribution._set_command_options) Nz# setting options for '%s' command:z %s = %s (from %s)cSsg|] }t|qSrr)rorrrrsz5Distribution._set_command_options..z1error in %s: command '%s' has no such option '%s')Zget_command_namerrrrQZboolean_optionsr\rrzrrrrrSr r[) r0Z command_objZ option_dictZ command_namerrr3Z bool_optsneg_optZ is_stringr5rrr_set_command_optionssF           z!Distribution._set_command_optionsFcCs(|j|dt||j|d|dS)zYParses configuration files from various levels and loads configuration. )r)ignore_option_errorsN)rrrr)r0rrrrrparse_config_filess  zDistribution.parse_config_filescCst|}|jr||S)z3Process features after parsing command line options)rparse_command_liner_finalize_features)r0resultrrrrs zDistribution.parse_command_linecCsd|ddS)z;Convert feature name to corresponding option attribute nameZwith_rrrr0r2rrr_feature_attrnameszDistribution._feature_attrnamecCs8tjjt||jdd}|D]}tjj|ddq|S)zResolve pre-setup requirementsT) installerZreplace_conflictingr)rWrresolverrfetch_build_eggr)r0r,Zresolved_distsr^rrrfetch_build_eggsszDistribution.fetch_build_eggscCst||jr|tdD]:}t||jd}|dk r"|j|j d| ||j|q"t|ddr~dd|j D|_ ng|_ dS)Nrrconvert_2to3_doctestscSsg|]}tj|qSr)ospathabspathrprrrrsz1Distribution.finalize_options..) rfinalize_optionsr_set_global_opts_from_featuresrWrr%r2requirerloadr)r0r_r3rrrrs   zDistribution.finalize_optionsc Csvtjtjd}tj|srt|t|tj|d}t|d$}| d| d| dW5QRX|S)Nz.eggsz README.txtwzcThis directory contains eggs that were downloaded by setuptools to build, test, and run plug-ins. zAThis directory caches those eggs to prevent repeated downloads. z/However, it is safe to delete this directory. ) rrrRcurdirexistsmkdirrZ hide_filerrF)r0Z egg_cache_dirZreadme_txt_filenamefrrrget_egg_cache_dirs      zDistribution.get_egg_cache_dirc Csddlm}|ddgi}|d}||dd|dD|jr|jdd}d|krx|dd |}d |f|d<|}||d g|d d dd d d d d d }| ||S)z Fetch an egg needed for buildingr) easy_installZ script_argsrcss"|]\}}|dkr||fVqdS)) find_links site_dirsZ index_urloptimizer Z allow_hostsNrrrrrrsz/Distribution.fetch_build_egg..NrrZsetupxTF) args install_dirZexclude_scriptsZ always_copyZbuild_directoryZeditableZupgradeZ multi_versionZ no_reportr) Zsetuptools.command.easy_installr __class__rclearupdaterQrrZensure_finalized)r0rrr^optsZlinksr cmdrrrrs8     zDistribution.fetch_build_eggc Csg}|j}|jD]\}}||d|||jr|j}d}d}|s\||}}d|dd||fd|dd||ff}| |d||d|<q||j |_ |_ ||_|_ dS)z;Add --with-X/--without-X options based on optional featuresNz (default)rzwith-zinclude zwithout-zexclude ) rcopyrrQ _set_featurevalidateoptionalr=include_by_defaultextendglobal_optionsZfeature_optionsZfeature_negopt) r0ZgoZnor2featuredescrZincdefZexcdefnewrrrrs$     z+Distribution._set_global_opts_from_featurescCs|jD]<\}}||}|s0|dkr |r ||||dq |jD](\}}||sR||||dqRdS)z9Add/remove features and resolve dependencies between themNrr)rrQfeature_is_includedr include_inr exclude_from)r0r2rZenabledrrrr0s    zDistribution._finalize_featurescCs\||jkr|j|Std|}|D]*}|j|jd||j|<}|St||S)z(Pluggable version of get_command_class()distutils.commandsrN)cmdclassrWrrrrrget_command_class)r0commandZepsr_r!rrrr"As   zDistribution.get_command_classcCs:tdD]$}|j|jkr |}||j|j<q t|SNr )rWrr2r!rrprint_commandsr0r_r!rrrr%Ns  zDistribution.print_commandscCs:tdD]$}|j|jkr |}||j|j<q t|Sr$)rWrr2r!rrget_command_listr&rrrr'Vs  zDistribution.get_command_listcCst||||dS)zSet feature's inclusion statusN)rr)r0r2Zstatusrrrr^szDistribution._set_featurecCst|||S)zAReturn 1 if feature is included, 0 if excluded, 'None' if unknown)r%rrrrrrbsz Distribution.feature_is_includedcCsF||dkr&|j|j}t|d|j||||ddS)z)Request inclusion of feature named 'name'rz2 is required, but was excluded or is not availablerN)rrr=r rr)r0r2rrrrinclude_featurefs zDistribution.include_featurecKs@|D]2\}}t|d|d}|r.||q|||qdS)aAdd items to distribution that are named in keyword arguments For example, 'dist.include(py_modules=["x"])' would add 'x' to the distribution's 'py_modules' attribute, if it was not already there. Currently, this method only supports inclusion for attributes that are lists or tuples. If you need to add support for adding to other attributes in this or a subclass, you can add an '_include_X' method, where 'X' is the name of the attribute. The method will be called with the value passed to 'include()'. So, 'dist.include(foo={"bar":"baz"})' will try to call 'dist._include_foo({"bar":"baz"})', which can then handle whatever special inclusion logic is needed. Z _include_N)rQr% _include_misc)r0rrrincluderrrr*qs  zDistribution.includecsfd|jr&fdd|jD|_|jrDfdd|jD|_|jrbfdd|jD|_dS)z9Remove packages, modules, and extensions in named packagerccs"g|]}|kr|s|qSr startswithrpackagepfxrrrs z0Distribution.exclude_package..cs"g|]}|kr|s|qSrr+rr-rrrs cs&g|]}|jkr|js|qSr)r2r,rr-rrrs N)packages py_modules ext_modules)r0r.rr-rexclude_packages   zDistribution.exclude_packagecCs2|d}|D]}||ks&||rdSqdS)z.rzsequencer r%r\r)r0r2r3oldrr9r _exclude_miscs    zDistribution._exclude_misccst|tstd||fzt||Wn tk rHtd|YnXdkr`t|||n:ttsxt|dn"fdd|D}t|||dS)zAHandle 'include()' for list/tuple attrs without a special handlerz%s: setting must be a list (%r)r5Nr6csg|]}|kr|qSrrr7r<rrrsz.Distribution._include_misc..r:)r0r2r3rrr>rr)s$    zDistribution._include_misccKs@|D]2\}}t|d|d}|r.||q|||qdS)aRemove items from distribution that are named in keyword arguments For example, 'dist.exclude(py_modules=["x"])' would remove 'x' from the distribution's 'py_modules' attribute. Excluding packages uses the 'exclude_package()' method, so all of the package's contained packages, modules, and extensions are also excluded. Currently, this method only supports exclusion from attributes that are lists or tuples. If you need to add support for excluding from other attributes in this or a subclass, you can add an '_exclude_X' method, where 'X' is the name of the attribute. The method will be called with the value passed to 'exclude()'. So, 'dist.exclude(foo={"bar":"baz"})' will try to call 'dist._exclude_foo({"bar":"baz"})', which can then handle whatever special exclusion logic is needed. Z _exclude_N)rQr%r=)r0rrrexcluderrrr?s  zDistribution.excludecCs,t|tstd|ftt|j|dS)Nz.packages: setting must be a list or tuple (%r))rzr;r rkrr3)r0r0rrr_exclude_packagess  zDistribution._exclude_packagesc Cs|jj|_|jj|_|d}|d}||krf||\}}||=ddl}||d|dd<|d}q&t|||}||} t | ddrd|f||d<|dk rgS|S)NraliasesTrZcommand_consumes_arguments command liner ) rrrrshlexrBr_parse_command_optsr"r%) r0rr r#rArrrCnargsZ cmd_classrrrrDs"       z Distribution._parse_command_optsc Csi}|jD]\}}|D]\}\}}|dkr4q|dd}|dkr||}|j}|t|di|D]\} } | |krv| }d}qqvtdn |dkrd}|| |i|<qq|S) ahReturn a '{cmd: {opt:val}}' map of all command-line options Option names are all long, but do not include the leading '--', and contain dashes rather than underscores. If the option doesn't take an argument (e.g. '--quiet'), the 'val' is 'None'. Note that options provided by config files are intentionally excluded. rBrrrrNzShouldn't be able to get herer) rrQrZget_command_objrrrr%r]r) r0drrrrrZcmdobjrnegposrrrget_cmdline_optionss(     z Distribution.get_cmdline_optionsccsv|jpdD] }|Vq |jpdD] }|Vq |jp4dD]:}t|trN|\}}n|j}|drj|dd}|Vq6dS)z@Yield all packages, modules, and extension names in distributionrmoduleNi)r0r1r2rztupler2endswith)r0ZpkgrJZextr2Z buildinforrrr40s    z$Distribution.iter_distribution_namesc Csddl}tjs|jr t||St|jtj s:t||S|jj dkrVt||S|jj }|jj }|j dkrtdpvd}|jj}t |jd||||_zt||WSt |j|||||_XdS)zIf there were any non-global "display-only" options (--help-commands or the metadata display options) on the command line, display the requested info and return true; else return false. rN)rutf8Zwin32 r)rrrOZ help_commandsrhandle_display_optionsrzstdoutr TextIOWrapperrrerrorsr@line_bufferingdetach)r0Z option_orderrrrRnewlinerSrrrrOBs6    z#Distribution.handle_display_options)N)N)N)NF).r __module__ __qualname____doc__r{rZ OrderedSetrrrrrr staticmethodrrrrrrrrrrrrrrrr"r%r'rrr(r*r3rdr=r)r?r@rDrIr4rOrrrrrTsXD ;  >  /     (c@sFeZdZdZeddZdddZd d Zd d Zd dZ ddZ dS)ra **deprecated** -- The `Feature` facility was never completely implemented or supported, `has reported issues `_ and will be removed in a future version. A subset of the distribution that can be excluded if unneeded/wanted Features are created using these keyword arguments: 'description' -- a short, human readable description of the feature, to be used in error messages, and option help messages. 'standard' -- if true, the feature is included by default if it is available on the current system. Otherwise, the feature is only included if requested via a command line '--with-X' option, or if another included feature requires it. The default setting is 'False'. 'available' -- if true, the feature is available for installation on the current system. The default setting is 'True'. 'optional' -- if true, the feature's inclusion can be controlled from the command line, using the '--with-X' or '--without-X' options. If false, the feature's inclusion status is determined automatically, based on 'availabile', 'standard', and whether any other feature requires it. The default setting is 'True'. 'require_features' -- a string or sequence of strings naming features that should also be included if this feature is included. Defaults to empty list. May also contain 'Require' objects that should be added/removed from the distribution. 'remove' -- a string or list of strings naming packages to be removed from the distribution if this feature is *not* included. If the feature *is* included, this argument is ignored. This argument exists to support removing features that "crosscut" a distribution, such as defining a 'tests' feature that removes all the 'tests' subpackages provided by other features. The default for this argument is an empty list. (Note: the named package(s) or modules must exist in the base distribution when the 'setup()' function is initially called.) other keywords -- any other keyword arguments are saved, and passed to the distribution's 'include()' and 'exclude()' methods when the feature is included or excluded, respectively. So, for example, you could pass 'packages=["a","b"]' to cause packages 'a' and 'b' to be added or removed from the distribution as appropriate. A feature must include at least one 'requires', 'remove', or other keyword argument. Otherwise, it can't affect the distribution in any way. Note also that you can subclass 'Feature' to create your own specialized feature types that modify the distribution in other ways when included or excluded. See the docstrings for the various methods here for more detail. Aside from the methods, the only feature attributes that distributions look at are 'description' and 'optional'. cCsd}tj|tdddS)NzrFeatures are deprecated and will be removed in a future version. See https://github.com/pypa/setuptools/issues/65.) stacklevel)rrrr4rrrrszFeature.warn_deprecatedFTrc Ks|||_||_||_||_t|ttfr4|f}dd|D|_dd|D}|r^||d<t|trn|f}||_ ||_ |s|s|st ddS)NcSsg|]}t|tr|qSrrzrPrrrrrs z$Feature.__init__..cSsg|]}t|ts|qSrr\rrrrrs rzgFeature %s: must define 'require_features', 'remove', or at least one of 'packages', 'py_modules', etc.) rr=standard availablerrzrPrrremoveextrasr ) r0r=r]r^rrr_r`Zerrrrrs*  zFeature.__init__cCs |jo |jS)z+Should this feature be included by default?)r^r]rrrrrszFeature.include_by_defaultcCs<|jst|jd|jf|j|jD]}||q(dS)aEnsure feature and its requirements are included in distribution You may override this in a subclass to perform additional operations on the distribution. Note that this method may be called more than once per feature, and so should be idempotent. z3 is required, but is not available on this platformN)r^r r=r*r`rr()r0r^rrrrrs  zFeature.include_incCs.|jf|j|jr*|jD]}||qdS)a2Ensure feature is excluded from distribution You may override this in a subclass to perform additional operations on the distribution. This method will be called at most once per feature, and only after all included features have been asked to include themselves. N)r?r`r_r3r0r^r8rrrrs  zFeature.exclude_fromcCs.|jD]"}||std|j||fqdS)aVerify that feature makes sense in context of distribution This method is called by the distribution just before it parses its command line. It checks to ensure that the 'remove' attribute, if any, contains only valid package/module names that are present in the base distribution when 'setup()' is called. You may override it in a subclass to perform any other required validation of the feature against a target distribution. zg%s wants to be able to remove %s, but the distribution doesn't contain any packages or modules under %sN)r_rdr r=rarrrrs   zFeature.validateN)FTTrr) rrVrWrXrYrrrrrrrrrrres8  rc@seZdZdZdS)rzrClass for warning about deprecations in dist in setuptools. Not ignored by default, unlike DeprecationWarning.N)rrVrWrXrrrrrsr)J__all__rrrrrrZ distutils.logrfZdistutils.coreZ distutils.cmdZdistutils.distZdistutils.utilrZdistutils.debugrZdistutils.fancy_getoptrrl collectionsrZemailrZdistutils.errorsr r r r Zdistutils.versionr Zsetuptools.externrrrZsetuptools.extern.six.movesrrrrrZsetuptools.dependsrZ setuptoolsrZsetuptools.monkeyrZsetuptools.configrrW __import__r r1rDrVrKrkr;r`rbrjrornrwr~rrrrrZcorerrrrrrrrsv               6L    site-packages/setuptools/__pycache__/wheel.cpython-38.pyc000064400000015616151464452710017473 0ustar00U Qab@sdZddlmZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl m Z ddl m Z ddlmZddl mZddlmZeZed ejjZd Zd d ZGd ddZdS)zWheels support.) get_platformN) parse_version)canonicalize_name)PY3) pep425tags)write_requirementsz^(?P.+?)-(?P\d.*?) ((-(?P\d.*?))?-(?P.+?)-(?P.+?)-(?P.+?) )\.whl$ztry: __import__('pkg_resources').declare_namespace(__name__) except ImportError: __path__ = __import__('pkgutil').extend_path(__path__, __name__) c Cst|D]\}}}tj||}|D].}tj||}tj|||}t||q&ttt|D]D\} } tj|| }tj||| }tj |sft|||| =qfq tj|ddD]\}}}|rt t |qdS)zDMove everything under `src_dir` to `dst_dir`, and delete the former.T)topdownN) oswalkpathrelpathjoinrenamesreversedlist enumerateexistsAssertionErrorrmdir) Zsrc_dirZdst_dirdirpathZdirnames filenamessubdirfsrcZdstndr4/usr/lib/python3.8/site-packages/setuptools/wheel.pyunpack%s   rc@sheZdZddZddZddZddZd d Zd d Zd dZ e ddZ e ddZ e ddZ dS)WheelcCsPttj|}|dkr$td|||_|D]\}}t|||q6dS)Nzinvalid wheel name: %r) WHEEL_NAMEr r basename ValueErrorfilename groupdictitemssetattr)selfr#matchkvrrr__init__=s  zWheel.__init__cCs&t|jd|jd|jdS)z>List tags (py_version, abi, platform) supported by this wheel..) itertoolsproductZ py_versionsplitZabiplatformr'rrrtagsEs    z Wheel.tagscs$ttfdd|DdS)z5Is the wheel is compatible with the current platform?c3s|]}|krdVqdS)TNr).0tZsupported_tagsrr Psz&Wheel.is_compatible..F)rZ get_supportednextr2r1rr5r is_compatibleMszWheel.is_compatiblecCs,tj|j|j|jdkrdntddS)Nany) project_nameversionr0z.egg) pkg_resources Distributionr:r;r0regg_namer1rrrr>RszWheel.egg_namecCsJ|D]4}t|}|drt|t|jr|SqtddS)Nz .dist-infoz.unsupported wheel format. .dist-info not found)Znamelist posixpathdirnameendswithr startswithr:r")r'zfmemberr@rrr get_dist_infoXs    zWheel.get_dist_infoc Cs(t|j}|||W5QRXdS)z"Install wheel as an egg directory.N)zipfileZZipFiler#_install_as_egg)r'destination_eggdirrCrrrinstall_as_eggbszWheel.install_as_eggcCs\d|j|jf}||}d|}tj|d}|||||||||||dS)Nz%s-%sz%s.dataEGG-INFO) r:r;rEr r r _convert_metadata_move_data_entries_fix_namespace_packages)r'rHrCZ dist_basename dist_info dist_dataegg_inforrrrGgs  zWheel._install_as_eggc s&fdd}|d}t|d}td|ko>tdkn}|sTtd|t||tj|tj j |t |dd d t t tfd d jD}t|ttj|d tj|dtj t|dd} t| ddtj|ddS)Nc sTt|8}tr&|dn|}tj |W5QRSQRXdS)Nzutf-8) openr?r rreaddecodeemailparserZParserZparsestr)namefpvalue)rNrCrr get_metadatassz-Wheel._convert_metadata..get_metadataZWHEELz Wheel-Versionz1.0z2.0dev0z$unsupported wheel format version: %s)ZmetadatacSsd|_t|SN)Zmarkerstr)reqrrrraw_reqsz(Wheel._convert_metadata..raw_reqc s2i|]*}|tfddt|fDqS)c3s|]}|kr|VqdSrZr)r3r\)install_requiresrrr6sz5Wheel._convert_metadata...)sortedmaprequires)r3Zextra)distr^r]rr s  z+Wheel._convert_metadata..ZMETADATAzPKG-INFO)r^extras_require)ZattrsrPz requires.txt)rgetr"r mkdirZ extractallr r r<r=Z from_locationZ PathMetadatarr_r`raZextrasrename setuptoolsdictrZget_command_obj) rCrHrNrPrYZwheel_metadataZ wheel_versionZwheel_v1rdZ setup_distr)rbrNr^r]rCrrKqsL       zWheel._convert_metadatacstj|tjd}tj|rtj|dd}t|t|D]D}|drpttj||qLttj||tj||qLt |t tjjfdddDD]}t ||qtjrt dS)z,Move data entries to their correct location.ZscriptsrJz.pycc3s|]}tj|VqdSrZ)r r r )r3rrOrrr6sz+Wheel._move_data_entries..)dataZheadersZpurelibZplatlibN) r r r rrflistdirrAunlinkrgrfilterr)rHrOZdist_data_scriptsZegg_info_scriptsentryrrrjrrLs.         zWheel._move_data_entriesc Cstj|d}tj|rt|}|}W5QRX|D]b}tjj|f|d}tj|d}tj|r>tj|s>t|d}|tW5QRXq>dS)Nznamespace_packages.txtr,z __init__.pyw) r r r rrQrRr/writeNAMESPACE_PACKAGE_INIT)rPrHZnamespace_packagesrWmodZmod_dirZmod_initrrrrMs   zWheel._fix_namespace_packagesN)__name__ __module__ __qualname__r+r2r8r>rErIrG staticmethodrKrLrMrrrrr;s   9 r)__doc__Zdistutils.utilrrTr-r r?rerFr<rhrZ!setuptools.extern.packaging.utilsrZsetuptools.extern.sixrrZsetuptools.command.egg_infortypeZ __metaclass__compileVERBOSEr(r rrrrrrrrs,      site-packages/setuptools/__pycache__/glibc.cpython-38.pyc000064400000002772151464452710017446 0ustar00U QabJ @sHddlmZddlZddlZddlZddZddZddZd d ZdS) )absolute_importNcCsRtd}z |j}Wntk r*YdSXtj|_|}t|tsN|d}|S)z9Returns glibc version string, or None if not using glibc.Nascii) ctypesZCDLLgnu_get_libc_versionAttributeErrorZc_char_pZrestype isinstancestrdecode)Zprocess_namespacer version_strr 4/usr/lib/python3.8/site-packages/setuptools/glibc.pyglibc_version_string s    r cCsHtd|}|s$td|tdSt|d|koFt|d|kS)Nz$(?P[0-9]+)\.(?P[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFmajorminor)rematchwarningswarnRuntimeWarningintgroup)r required_major minimum_minormr r r check_glibc_version$s rcCst}|dkrdSt|||S)NF)r r)rrr r r r have_compatible_glibc4srcCst}|dkrdSd|fSdS)zTry to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings in case the lookup fails. N)rZglibc)r )Z glibc_versionr r r libc_verLsr) Z __future__rrrrr rrrr r r r s site-packages/setuptools/__pycache__/pep425tags.cpython-38.pyc000064400000016043151464452710020260 0ustar00U Qabm*@sdZddlmZddlZddlmZddlZddlZddlZddl Z ddl Z ddl m Z ddl mZddlmZed Zd d Zd d ZddZddZddZd$ddZddZddZddZddZdd Zd%d"d#ZeZdS)&z2Generate and work with PEP 425 Compatibility Tags.)absolute_importN)log) OrderedDict)six)glibcz(.+)_(\d+)_(\d+)_(.+)c CsLz t|WStk rF}ztd|tWYdSd}~XYnXdS)Nz{}) sysconfigget_config_varIOErrorwarningswarnformatRuntimeWarning)varer9/usr/lib/python3.8/site-packages/setuptools/pep425tags.pyr s  r cCs:ttdrd}n&tjdr"d}ntjdkr2d}nd}|S)z'Return abbreviated implementation name.pypy_version_infoppjavaZjyZcliZipcp)hasattrsysplatform startswith)Zpyimplrrr get_abbr_impls   rcCs,td}|rtdkr(dttt}|S)zReturn implementation version.Zpy_version_nodotr)r rjoinmapstrget_impl_version_info)Zimpl_verrrr get_impl_ver+sr!cCs:tdkr"tjdtjjtjjfStjdtjdfSdS)zQReturn sys.version_info-like tuple for use in decrementing the minor version.rrrN)rr version_informajorminorrrrrr 3s  r cCsdttS)z; Returns the Tag for this specific implementation. z{}{})r rr!rrrr get_impl_tag>sr%TcCs.t|}|dkr&|r td||S||kS)zgUse a fallback method for determining SOABI flags if the needed config var is unset or unavailable.Nz>Config variable '%s' is unset, Python ABI tag may be incorrect)r rdebug)rZfallbackexpectedr valrrrget_flagEsr)cstd}t|sdkrttdrd}d}d}tddddkd rJd }td fd ddkd rhd }tdddddkotjdrtjrd}dt|||f}n@|r|drd| dd}n|r| dd dd}nd}|S)zXReturn the ABI tag based on SOABI (if available) or emulate SOABI (CPython 2, PyPy).ZSOABI>rr maxunicoderZPy_DEBUGcSs ttdS)NZgettotalrefcount)rrrrrr[zget_abi_tag..r)r dZ WITH_PYMALLOCcsdkS)Nrrrimplrrr+_r,mZPy_UNICODE_SIZEcSs tjdkS)Ni)rr*rrrrr+cr,)r'r uz %s%s%s%s%szcpython--r._N) r rrrr)rZPY2r!rsplitreplace)Zsoabir-r0r2abirr.r get_abi_tagQs@ r9cCs tjdkS)Ni)rmaxsizerrrr_is_running_32bitssr;cCstjdkr^t\}}}|d}|dkr6tr6d}n|dkrHtrHd}d|d|d |Stj dd  d d }|d krtrd }|S)z0Return our platform name 'win32', 'linux_x86_64'darwinr4x86_64i386ppc64ppczmacosx_{}_{}_{}rrr5r3 linux_x86_64 linux_i686) rrZmac_verr6r;r distutilsutil get_platformr7)releaser5machineZ split_verresultrrrrEws  rEc CsHtdkrdSzddl}t|jWSttfk r:YnXtddS)N>rArBFr)rE _manylinuxboolZmanylinux1_compatible ImportErrorAttributeErrorrZhave_compatible_glibc)rKrrris_manylinux1_compatibles  rOcsrg}fddtddddg|||r8||D]&}||kr<|||r<||q<|d|S)zReturn a list of supported arches (including group arches) for the given major, minor and machine architecture of a macOS machine. cs||dkr||fdkS|dkr(||fdkS|dkr<||fdkS|dkrP||fdkS|krx|D]}|||r`dSq`dS) Nr@) rJr?r>)rPr1r=TFr)r#r$archgarch_supports_archgroupsrrrTs      z)get_darwin_arches.._supports_arch)Zfat)r>r@)Zintel)r=r>)Zfat64)r=r?)Zfat32)r=r>r@Z universal)rappend)r#r$rGarchesrRrrSrget_darwin_archess$    rXFc Csg}|dkrTg}t}|dd}t|dddD] }|dtt||fq2|p\t}g} |pjt}|r~|g| dd<t} ddl } | D],} | d dr| | d dddq| tt| | d |sT|pt} | d rzt| }|rr|\}}}}d ||}g}ttt|dD]0}tt|||D]}||||fqRq>n| g}n*|dkrtr| d d | g}n| g}| D].}|D]"} |d||df|| fqq|ddD]F}|dkrq,| D]*}|D]} |d||f|| fqqq|D]"} |d|ddd | fq0|d||dfd df|d||ddfd dft|D]B\}}|d|fd df|dkr|d|dd dfq|S)acReturn a list of supported tags for each version specified in `versions`. :param versions: a list of string versions, of the form ["33", "32"], or None. The first version will be assumed to support our ABI. :param platform: specify the exact platform you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. :param abi: specify the exact abi you want valid tags for, or None. If None, use the local interpreter abi. Nrrz.abir4rIrZnoneZmacosxz {}_{}_%i_%sZlinuxZ manylinux1z%s%s>3031zpy%sany)r rangerVrrrrr9setimpZ get_suffixesraddr6extendsortedlistrE _osx_arch_patmatchrUr reversedintrXrOr7 enumerate)ZversionsZnoarchrr/r8Z supportedr"r#r$ZabisZabi3sr_suffixrQrenameZ actual_archZtplrWr0aversionirrr get_supportedsh         $ $   rn)TT)NFNNN) __doc__Z __future__rZdistutils.utilrCrrrerrr collectionsrZexternrrrcompilerdr rr!r r%r)r9r;rErOrXrnZimplementation_tagrrrrs8         "= `site-packages/setuptools/__pycache__/dep_util.cpython-38.opt-1.pyc000064400000001463151464452710021126 0ustar00U Qab@sddlmZddZdS)) newer_groupcCsht|t|krtdg}g}tt|D]2}t||||r,||||||q,||fS)zWalk both arguments in parallel, testing if each source group is newer than its corresponding target. Returns a pair of lists (sources_groups, targets) where sources is newer than target, according to the semantics of 'newer_group()'. z5'sources_group' and 'targets' must be the same length)len ValueErrorrangerappend)Zsources_groupsZtargetsZ n_sourcesZ n_targetsir7/usr/lib/python3.8/site-packages/setuptools/dep_util.pynewer_pairwise_groupsr N)Zdistutils.dep_utilrr rrrr s site-packages/setuptools/__pycache__/monkey.cpython-38.opt-1.pyc000064400000011044151464452710020617 0ustar00U Qab@sdZddlZddlZddlZddlZddlZddlmZddl Z ddl m Z ddl Z gZ ddZddZd d Zd d Zd dZddZddZddZdS)z Monkey patching of distutils. N) import_module)sixcCs"tdkr|f|jSt|S)am Returns the bases classes for cls sorted by the MRO. Works around an issue on Jython where inspect.getmro will not return all base classes if multiple classes share the same name. Instead, this function will return a tuple containing the class itself, and the contents of cls.__bases__. See https://github.com/pypa/setuptools/issues/1024. ZJython)platformZpython_implementation __bases__inspectZgetmro)clsr5/usr/lib/python3.8/site-packages/setuptools/monkey.py_get_mros  r cCs0t|tjrtnt|tjr tndd}||S)NcSsdS)Nr)itemrrr *zget_unpatched..) isinstancerZ class_typesget_unpatched_classtypes FunctionTypeget_unpatched_function)r lookuprrr get_unpatched&s rcCs:ddt|D}t|}|jds6d|}t||S)zProtect against re-patching the distutils if reloaded Also ensures that no other distutils extension monkeypatched the distutils first. css|]}|jds|VqdS) setuptoolsN) __module__ startswith).0rrrr 5s z&get_unpatched_class.. distutilsz(distutils has already been patched by %r)r nextrrAssertionError)rZexternal_basesbasemsgrrr r/s rcCstjtj_tjdk}|r"tjtj_tjdkp^dtjko@dknp^dtjkoZdkn}|rrd}|tjj _ t tj tjtj fD]}tj j|_qtjjtj_tjjtj_dtjkrtjjtjd_tdS)N)r) )r)rr$)rr zhttps://upload.pypi.org/legacy/zdistutils.command.build_ext)rZCommandrZcoresys version_infofindallZfilelistZconfigZ PyPIRCCommandZDEFAULT_REPOSITORY_patch_distribution_metadatadistcmdZ Distribution extensionZ Extensionmodules#patch_for_msvc_specialized_compiler)Zhas_issue_12885Zneeds_warehouseZ warehousemodulerrr patch_allAs*          r0cCs*dD] }ttj|}ttjj||qdS)zDPatch write_pkg_file and read_pkg_file for higher metadata standards)Zwrite_pkg_fileZ read_pkg_fileZget_metadata_versionN)getattrrr*setattrrZDistributionMetadata)attrZnew_valrrr r)hs r)cCs*t||}t|d|t|||dS)z Patch func_name in target_mod with replacement Important - original must be resolved by name to avoid patching an already patched function. unpatchedN)r1vars setdefaultr2)Z replacementZ target_mod func_nameoriginalrrr patch_funcos r9cCs t|dS)Nr4)r1) candidaterrr rsrcstdtdkrdSfdd}t|d}t|d}zt|dt|d Wntk rlYnXzt|d Wntk rYnXzt|d Wntk rYnXdS) z\ Patch functions in distutils to use standalone Microsoft Visual C++ compilers. zsetuptools.msvcZWindowsNcsLd|kr dnd}||d}t|}t|}t||sBt||||fS)zT Prepare the parameters for patch_func to patch indicated function. msvc9Zmsvc9_Zmsvc14__)lstripr1rhasattr ImportError)Zmod_namer7Z repl_prefixZ repl_namereplmodZmsvcrr patch_paramss  z9patch_for_msvc_specialized_compiler..patch_paramszdistutils.msvc9compilerzdistutils._msvccompilerZfind_vcvarsallZquery_vcvarsallZ _get_vc_envZgen_lib_options)rrsystem functoolspartialr9r?)rCr;Zmsvc14rrBr r.s&    r.)__doc__r&Zdistutils.filelistrrrrE importlibrrZsetuptools.externrr__all__r rrr0r)r9rr.rrrr s$   'site-packages/setuptools/lib2to3_ex.py000064400000003735151464452710014052 0ustar00""" Customized Mixin2to3 support: - adds support for converting doctests This module raises an ImportError on Python 2. """ from distutils.util import Mixin2to3 as _Mixin2to3 from distutils import log from lib2to3.refactor import RefactoringTool, get_fixers_from_package import setuptools class DistutilsRefactoringTool(RefactoringTool): def log_error(self, msg, *args, **kw): log.error(msg, *args) def log_message(self, msg, *args): log.info(msg, *args) def log_debug(self, msg, *args): log.debug(msg, *args) class Mixin2to3(_Mixin2to3): def run_2to3(self, files, doctests=False): # See of the distribution option has been set, otherwise check the # setuptools default. if self.distribution.use_2to3 is not True: return if not files: return log.info("Fixing " + " ".join(files)) self.__build_fixer_names() self.__exclude_fixers() if doctests: if setuptools.run_2to3_on_doctests: r = DistutilsRefactoringTool(self.fixer_names) r.refactor(files, write=True, doctests_only=True) else: _Mixin2to3.run_2to3(self, files) def __build_fixer_names(self): if self.fixer_names: return self.fixer_names = [] for p in setuptools.lib2to3_fixer_packages: self.fixer_names.extend(get_fixers_from_package(p)) if self.distribution.use_2to3_fixers is not None: for p in self.distribution.use_2to3_fixers: self.fixer_names.extend(get_fixers_from_package(p)) def __exclude_fixers(self): excluded_fixers = getattr(self, 'exclude_fixers', []) if self.distribution.use_2to3_exclude_fixers is not None: excluded_fixers.extend(self.distribution.use_2to3_exclude_fixers) for fixer_name in excluded_fixers: if fixer_name in self.fixer_names: self.fixer_names.remove(fixer_name) site-packages/setuptools/ssl_support.py000064400000020455151464452710014473 0ustar00import os import socket import atexit import re import functools from setuptools.extern.six.moves import urllib, http_client, map, filter from pkg_resources import ResolutionError, ExtractionError try: import ssl except ImportError: ssl = None __all__ = [ 'VerifyingHTTPSHandler', 'find_ca_bundle', 'is_available', 'cert_paths', 'opener_for' ] cert_paths = """ /etc/pki/tls/certs/ca-bundle.crt /etc/ssl/certs/ca-certificates.crt /usr/share/ssl/certs/ca-bundle.crt /usr/local/share/certs/ca-root.crt /etc/ssl/cert.pem /System/Library/OpenSSL/certs/cert.pem /usr/local/share/certs/ca-root-nss.crt /etc/ssl/ca-bundle.pem """.strip().split() try: HTTPSHandler = urllib.request.HTTPSHandler HTTPSConnection = http_client.HTTPSConnection except AttributeError: HTTPSHandler = HTTPSConnection = object is_available = ssl is not None and object not in (HTTPSHandler, HTTPSConnection) try: from ssl import CertificateError, match_hostname except ImportError: try: from backports.ssl_match_hostname import CertificateError from backports.ssl_match_hostname import match_hostname except ImportError: CertificateError = None match_hostname = None if not CertificateError: class CertificateError(ValueError): pass if not match_hostname: def _dnsname_match(dn, hostname, max_wildcards=1): """Matching according to RFC 6125, section 6.4.3 https://tools.ietf.org/html/rfc6125#section-6.4.3 """ pats = [] if not dn: return False # Ported from python3-syntax: # leftmost, *remainder = dn.split(r'.') parts = dn.split(r'.') leftmost = parts[0] remainder = parts[1:] wildcards = leftmost.count('*') if wildcards > max_wildcards: # Issue #17980: avoid denials of service by refusing more # than one wildcard per fragment. A survey of established # policy among SSL implementations showed it to be a # reasonable choice. raise CertificateError( "too many wildcards in certificate DNS name: " + repr(dn)) # speed up common case w/o wildcards if not wildcards: return dn.lower() == hostname.lower() # RFC 6125, section 6.4.3, subitem 1. # The client SHOULD NOT attempt to match a presented identifier in which # the wildcard character comprises a label other than the left-most label. if leftmost == '*': # When '*' is a fragment by itself, it matches a non-empty dotless # fragment. pats.append('[^.]+') elif leftmost.startswith('xn--') or hostname.startswith('xn--'): # RFC 6125, section 6.4.3, subitem 3. # The client SHOULD NOT attempt to match a presented identifier # where the wildcard character is embedded within an A-label or # U-label of an internationalized domain name. pats.append(re.escape(leftmost)) else: # Otherwise, '*' matches any dotless string, e.g. www* pats.append(re.escape(leftmost).replace(r'\*', '[^.]*')) # add the remaining fragments, ignore any wildcards for frag in remainder: pats.append(re.escape(frag)) pat = re.compile(r'\A' + r'\.'.join(pats) + r'\Z', re.IGNORECASE) return pat.match(hostname) def match_hostname(cert, hostname): """Verify that *cert* (in decoded format as returned by SSLSocket.getpeercert()) matches the *hostname*. RFC 2818 and RFC 6125 rules are followed, but IP addresses are not accepted for *hostname*. CertificateError is raised on failure. On success, the function returns nothing. """ if not cert: raise ValueError("empty or no certificate") dnsnames = [] san = cert.get('subjectAltName', ()) for key, value in san: if key == 'DNS': if _dnsname_match(value, hostname): return dnsnames.append(value) if not dnsnames: # The subject is only checked when there is no dNSName entry # in subjectAltName for sub in cert.get('subject', ()): for key, value in sub: # XXX according to RFC 2818, the most specific Common Name # must be used. if key == 'commonName': if _dnsname_match(value, hostname): return dnsnames.append(value) if len(dnsnames) > 1: raise CertificateError("hostname %r " "doesn't match either of %s" % (hostname, ', '.join(map(repr, dnsnames)))) elif len(dnsnames) == 1: raise CertificateError("hostname %r " "doesn't match %r" % (hostname, dnsnames[0])) else: raise CertificateError("no appropriate commonName or " "subjectAltName fields were found") class VerifyingHTTPSHandler(HTTPSHandler): """Simple verifying handler: no auth, subclasses, timeouts, etc.""" def __init__(self, ca_bundle): self.ca_bundle = ca_bundle HTTPSHandler.__init__(self) def https_open(self, req): return self.do_open( lambda host, **kw: VerifyingHTTPSConn(host, self.ca_bundle, **kw), req ) class VerifyingHTTPSConn(HTTPSConnection): """Simple verifying connection: no auth, subclasses, timeouts, etc.""" def __init__(self, host, ca_bundle, **kw): HTTPSConnection.__init__(self, host, **kw) self.ca_bundle = ca_bundle def connect(self): sock = socket.create_connection( (self.host, self.port), getattr(self, 'source_address', None) ) # Handle the socket if a (proxy) tunnel is present if hasattr(self, '_tunnel') and getattr(self, '_tunnel_host', None): self.sock = sock self._tunnel() # http://bugs.python.org/issue7776: Python>=3.4.1 and >=2.7.7 # change self.host to mean the proxy server host when tunneling is # being used. Adapt, since we are interested in the destination # host for the match_hostname() comparison. actual_host = self._tunnel_host else: actual_host = self.host if hasattr(ssl, 'create_default_context'): ctx = ssl.create_default_context(cafile=self.ca_bundle) self.sock = ctx.wrap_socket(sock, server_hostname=actual_host) else: # This is for python < 2.7.9 and < 3.4? self.sock = ssl.wrap_socket( sock, cert_reqs=ssl.CERT_REQUIRED, ca_certs=self.ca_bundle ) try: match_hostname(self.sock.getpeercert(), actual_host) except CertificateError: self.sock.shutdown(socket.SHUT_RDWR) self.sock.close() raise def opener_for(ca_bundle=None): """Get a urlopen() replacement that uses ca_bundle for verification""" return urllib.request.build_opener( VerifyingHTTPSHandler(ca_bundle or find_ca_bundle()) ).open # from jaraco.functools def once(func): @functools.wraps(func) def wrapper(*args, **kwargs): if not hasattr(func, 'always_returns'): func.always_returns = func(*args, **kwargs) return func.always_returns return wrapper @once def get_win_certfile(): try: import wincertstore except ImportError: return None class CertFile(wincertstore.CertFile): def __init__(self): super(CertFile, self).__init__() atexit.register(self.close) def close(self): try: super(CertFile, self).close() except OSError: pass _wincerts = CertFile() _wincerts.addstore('CA') _wincerts.addstore('ROOT') return _wincerts.name def find_ca_bundle(): """Return an existing CA bundle path, or None""" extant_cert_paths = filter(os.path.isfile, cert_paths) return ( get_win_certfile() or next(extant_cert_paths, None) or _certifi_where() ) def _certifi_where(): try: return __import__('certifi').where() except (ImportError, ResolutionError, ExtractionError): pass site-packages/setuptools/script.tmpl000064400000000212151464452710013713 0ustar00# EASY-INSTALL-SCRIPT: %(spec)r,%(script_name)r __requires__ = %(spec)r __import__('pkg_resources').run_script(%(spec)r, %(script_name)r) site-packages/setuptools/dist.py000064400000142205151464452710013037 0ustar00# -*- coding: utf-8 -*- __all__ = ['Distribution'] import io import sys import re import os import warnings import numbers import distutils.log import distutils.core import distutils.cmd import distutils.dist from distutils.util import strtobool from distutils.debug import DEBUG from distutils.fancy_getopt import translate_longopt import itertools from collections import defaultdict from email import message_from_file from distutils.errors import ( DistutilsOptionError, DistutilsPlatformError, DistutilsSetupError, ) from distutils.util import rfc822_escape from distutils.version import StrictVersion from setuptools.extern import six from setuptools.extern import packaging from setuptools.extern import ordered_set from setuptools.extern.six.moves import map, filter, filterfalse from . import SetuptoolsDeprecationWarning from setuptools.depends import Require from setuptools import windows_support from setuptools.monkey import get_unpatched from setuptools.config import parse_configuration import pkg_resources __import__('setuptools.extern.packaging.specifiers') __import__('setuptools.extern.packaging.version') def _get_unpatched(cls): warnings.warn("Do not call this function", DistDeprecationWarning) return get_unpatched(cls) def get_metadata_version(self): mv = getattr(self, 'metadata_version', None) if mv is None: if self.long_description_content_type or self.provides_extras: mv = StrictVersion('2.1') elif (self.maintainer is not None or self.maintainer_email is not None or getattr(self, 'python_requires', None) is not None or self.project_urls): mv = StrictVersion('1.2') elif (self.provides or self.requires or self.obsoletes or self.classifiers or self.download_url): mv = StrictVersion('1.1') else: mv = StrictVersion('1.0') self.metadata_version = mv return mv def read_pkg_file(self, file): """Reads the metadata values from a file object.""" msg = message_from_file(file) def _read_field(name): value = msg[name] if value == 'UNKNOWN': return None return value def _read_list(name): values = msg.get_all(name, None) if values == []: return None return values self.metadata_version = StrictVersion(msg['metadata-version']) self.name = _read_field('name') self.version = _read_field('version') self.description = _read_field('summary') # we are filling author only. self.author = _read_field('author') self.maintainer = None self.author_email = _read_field('author-email') self.maintainer_email = None self.url = _read_field('home-page') self.license = _read_field('license') if 'download-url' in msg: self.download_url = _read_field('download-url') else: self.download_url = None self.long_description = _read_field('description') self.description = _read_field('summary') if 'keywords' in msg: self.keywords = _read_field('keywords').split(',') self.platforms = _read_list('platform') self.classifiers = _read_list('classifier') # PEP 314 - these fields only exist in 1.1 if self.metadata_version == StrictVersion('1.1'): self.requires = _read_list('requires') self.provides = _read_list('provides') self.obsoletes = _read_list('obsoletes') else: self.requires = None self.provides = None self.obsoletes = None # Based on Python 3.5 version def write_pkg_file(self, file): """Write the PKG-INFO format data to a file object. """ version = self.get_metadata_version() if six.PY2: def write_field(key, value): file.write("%s: %s\n" % (key, self._encode_field(value))) else: def write_field(key, value): file.write("%s: %s\n" % (key, value)) write_field('Metadata-Version', str(version)) write_field('Name', self.get_name()) write_field('Version', self.get_version()) write_field('Summary', self.get_description()) write_field('Home-page', self.get_url()) if version < StrictVersion('1.2'): write_field('Author', self.get_contact()) write_field('Author-email', self.get_contact_email()) else: optional_fields = ( ('Author', 'author'), ('Author-email', 'author_email'), ('Maintainer', 'maintainer'), ('Maintainer-email', 'maintainer_email'), ) for field, attr in optional_fields: attr_val = getattr(self, attr) if attr_val is not None: write_field(field, attr_val) write_field('License', self.get_license()) if self.download_url: write_field('Download-URL', self.download_url) for project_url in self.project_urls.items(): write_field('Project-URL', '%s, %s' % project_url) long_desc = rfc822_escape(self.get_long_description()) write_field('Description', long_desc) keywords = ','.join(self.get_keywords()) if keywords: write_field('Keywords', keywords) if version >= StrictVersion('1.2'): for platform in self.get_platforms(): write_field('Platform', platform) else: self._write_list(file, 'Platform', self.get_platforms()) self._write_list(file, 'Classifier', self.get_classifiers()) # PEP 314 self._write_list(file, 'Requires', self.get_requires()) self._write_list(file, 'Provides', self.get_provides()) self._write_list(file, 'Obsoletes', self.get_obsoletes()) # Setuptools specific for PEP 345 if hasattr(self, 'python_requires'): write_field('Requires-Python', self.python_requires) # PEP 566 if self.long_description_content_type: write_field( 'Description-Content-Type', self.long_description_content_type ) if self.provides_extras: for extra in self.provides_extras: write_field('Provides-Extra', extra) sequence = tuple, list def check_importable(dist, attr, value): try: ep = pkg_resources.EntryPoint.parse('x=' + value) assert not ep.extras except (TypeError, ValueError, AttributeError, AssertionError): raise DistutilsSetupError( "%r must be importable 'module:attrs' string (got %r)" % (attr, value) ) def assert_string_list(dist, attr, value): """Verify that value is a string list""" try: # verify that value is a list or tuple to exclude unordered # or single-use iterables assert isinstance(value, (list, tuple)) # verify that elements of value are strings assert ''.join(value) != value except (TypeError, ValueError, AttributeError, AssertionError): raise DistutilsSetupError( "%r must be a list of strings (got %r)" % (attr, value) ) def check_nsp(dist, attr, value): """Verify that namespace packages are valid""" ns_packages = value assert_string_list(dist, attr, ns_packages) for nsp in ns_packages: if not dist.has_contents_for(nsp): raise DistutilsSetupError( "Distribution contains no modules or packages for " + "namespace package %r" % nsp ) parent, sep, child = nsp.rpartition('.') if parent and parent not in ns_packages: distutils.log.warn( "WARNING: %r is declared as a package namespace, but %r" " is not: please correct this in setup.py", nsp, parent ) def check_extras(dist, attr, value): """Verify that extras_require mapping is valid""" try: list(itertools.starmap(_check_extra, value.items())) except (TypeError, ValueError, AttributeError): raise DistutilsSetupError( "'extras_require' must be a dictionary whose values are " "strings or lists of strings containing valid project/version " "requirement specifiers." ) def _check_extra(extra, reqs): name, sep, marker = extra.partition(':') if marker and pkg_resources.invalid_marker(marker): raise DistutilsSetupError("Invalid environment marker: " + marker) list(pkg_resources.parse_requirements(reqs)) def assert_bool(dist, attr, value): """Verify that value is True, False, 0, or 1""" if bool(value) != value: tmpl = "{attr!r} must be a boolean value (got {value!r})" raise DistutilsSetupError(tmpl.format(attr=attr, value=value)) def check_requirements(dist, attr, value): """Verify that install_requires is a valid requirements list""" try: list(pkg_resources.parse_requirements(value)) if isinstance(value, (dict, set)): raise TypeError("Unordered types are not allowed") except (TypeError, ValueError) as error: tmpl = ( "{attr!r} must be a string or list of strings " "containing valid project/version requirement specifiers; {error}" ) raise DistutilsSetupError(tmpl.format(attr=attr, error=error)) def check_specifier(dist, attr, value): """Verify that value is a valid version specifier""" try: packaging.specifiers.SpecifierSet(value) except packaging.specifiers.InvalidSpecifier as error: tmpl = ( "{attr!r} must be a string " "containing valid version specifiers; {error}" ) raise DistutilsSetupError(tmpl.format(attr=attr, error=error)) def check_entry_points(dist, attr, value): """Verify that entry_points map is parseable""" try: pkg_resources.EntryPoint.parse_map(value) except ValueError as e: raise DistutilsSetupError(e) def check_test_suite(dist, attr, value): if not isinstance(value, six.string_types): raise DistutilsSetupError("test_suite must be a string") def check_package_data(dist, attr, value): """Verify that value is a dictionary of package names to glob lists""" if not isinstance(value, dict): raise DistutilsSetupError( "{!r} must be a dictionary mapping package names to lists of " "string wildcard patterns".format(attr)) for k, v in value.items(): if not isinstance(k, six.string_types): raise DistutilsSetupError( "keys of {!r} dict must be strings (got {!r})" .format(attr, k) ) assert_string_list(dist, 'values of {!r} dict'.format(attr), v) def check_packages(dist, attr, value): for pkgname in value: if not re.match(r'\w+(\.\w+)*', pkgname): distutils.log.warn( "WARNING: %r not a valid package name; please use only " ".-separated package names in setup.py", pkgname ) _Distribution = get_unpatched(distutils.core.Distribution) class Distribution(_Distribution): """Distribution with support for features, tests, and package data This is an enhanced version of 'distutils.dist.Distribution' that effectively adds the following new optional keyword arguments to 'setup()': 'install_requires' -- a string or sequence of strings specifying project versions that the distribution requires when installed, in the format used by 'pkg_resources.require()'. They will be installed automatically when the package is installed. If you wish to use packages that are not available in PyPI, or want to give your users an alternate download location, you can add a 'find_links' option to the '[easy_install]' section of your project's 'setup.cfg' file, and then setuptools will scan the listed web pages for links that satisfy the requirements. 'extras_require' -- a dictionary mapping names of optional "extras" to the additional requirement(s) that using those extras incurs. For example, this:: extras_require = dict(reST = ["docutils>=0.3", "reSTedit"]) indicates that the distribution can optionally provide an extra capability called "reST", but it can only be used if docutils and reSTedit are installed. If the user installs your package using EasyInstall and requests one of your extras, the corresponding additional requirements will be installed if needed. 'features' **deprecated** -- a dictionary mapping option names to 'setuptools.Feature' objects. Features are a portion of the distribution that can be included or excluded based on user options, inter-feature dependencies, and availability on the current system. Excluded features are omitted from all setup commands, including source and binary distributions, so you can create multiple distributions from the same source tree. Feature names should be valid Python identifiers, except that they may contain the '-' (minus) sign. Features can be included or excluded via the command line options '--with-X' and '--without-X', where 'X' is the name of the feature. Whether a feature is included by default, and whether you are allowed to control this from the command line, is determined by the Feature object. See the 'Feature' class for more information. 'test_suite' -- the name of a test suite to run for the 'test' command. If the user runs 'python setup.py test', the package will be installed, and the named test suite will be run. The format is the same as would be used on a 'unittest.py' command line. That is, it is the dotted name of an object to import and call to generate a test suite. 'package_data' -- a dictionary mapping package names to lists of filenames or globs to use to find data files contained in the named packages. If the dictionary has filenames or globs listed under '""' (the empty string), those names will be searched for in every package, in addition to any names for the specific package. Data files found using these names/globs will be installed along with the package, in the same location as the package. Note that globs are allowed to reference the contents of non-package subdirectories, as long as you use '/' as a path separator. (Globs are automatically converted to platform-specific paths at runtime.) In addition to these new keywords, this class also has several new methods for manipulating the distribution's contents. For example, the 'include()' and 'exclude()' methods can be thought of as in-place add and subtract commands that add or remove packages, modules, extensions, and so on from the distribution. They are used by the feature subsystem to configure the distribution for the included and excluded features. """ _DISTUTILS_UNSUPPORTED_METADATA = { 'long_description_content_type': None, 'project_urls': dict, 'provides_extras': ordered_set.OrderedSet, } _patched_dist = None def patch_missing_pkg_info(self, attrs): # Fake up a replacement for the data that would normally come from # PKG-INFO, but which might not yet be built if this is a fresh # checkout. # if not attrs or 'name' not in attrs or 'version' not in attrs: return key = pkg_resources.safe_name(str(attrs['name'])).lower() dist = pkg_resources.working_set.by_key.get(key) if dist is not None and not dist.has_metadata('PKG-INFO'): dist._version = pkg_resources.safe_version(str(attrs['version'])) self._patched_dist = dist def __init__(self, attrs=None): have_package_data = hasattr(self, "package_data") if not have_package_data: self.package_data = {} attrs = attrs or {} if 'features' in attrs or 'require_features' in attrs: Feature.warn_deprecated() self.require_features = [] self.features = {} self.dist_files = [] # Filter-out setuptools' specific options. self.src_root = attrs.pop("src_root", None) self.patch_missing_pkg_info(attrs) self.dependency_links = attrs.pop('dependency_links', []) self.setup_requires = attrs.pop('setup_requires', []) for ep in pkg_resources.iter_entry_points('distutils.setup_keywords'): vars(self).setdefault(ep.name, None) _Distribution.__init__(self, { k: v for k, v in attrs.items() if k not in self._DISTUTILS_UNSUPPORTED_METADATA }) # Fill-in missing metadata fields not supported by distutils. # Note some fields may have been set by other tools (e.g. pbr) # above; they are taken preferrentially to setup() arguments for option, default in self._DISTUTILS_UNSUPPORTED_METADATA.items(): for source in self.metadata.__dict__, attrs: if option in source: value = source[option] break else: value = default() if default else None setattr(self.metadata, option, value) if isinstance(self.metadata.version, numbers.Number): # Some people apparently take "version number" too literally :) self.metadata.version = str(self.metadata.version) if self.metadata.version is not None: try: ver = packaging.version.Version(self.metadata.version) normalized_version = str(ver) if self.metadata.version != normalized_version: warnings.warn( "Normalizing '%s' to '%s'" % ( self.metadata.version, normalized_version, ) ) self.metadata.version = normalized_version except (packaging.version.InvalidVersion, TypeError): warnings.warn( "The version specified (%r) is an invalid version, this " "may not work as expected with newer versions of " "setuptools, pip, and PyPI. Please see PEP 440 for more " "details." % self.metadata.version ) self._finalize_requires() def _finalize_requires(self): """ Set `metadata.python_requires` and fix environment markers in `install_requires` and `extras_require`. """ if getattr(self, 'python_requires', None): self.metadata.python_requires = self.python_requires if getattr(self, 'extras_require', None): for extra in self.extras_require.keys(): # Since this gets called multiple times at points where the # keys have become 'converted' extras, ensure that we are only # truly adding extras we haven't seen before here. extra = extra.split(':')[0] if extra: self.metadata.provides_extras.add(extra) self._convert_extras_requirements() self._move_install_requirements_markers() def _convert_extras_requirements(self): """ Convert requirements in `extras_require` of the form `"extra": ["barbazquux; {marker}"]` to `"extra:{marker}": ["barbazquux"]`. """ spec_ext_reqs = getattr(self, 'extras_require', None) or {} self._tmp_extras_require = defaultdict(list) for section, v in spec_ext_reqs.items(): # Do not strip empty sections. self._tmp_extras_require[section] for r in pkg_resources.parse_requirements(v): suffix = self._suffix_for(r) self._tmp_extras_require[section + suffix].append(r) @staticmethod def _suffix_for(req): """ For a requirement, return the 'extras_require' suffix for that requirement. """ return ':' + str(req.marker) if req.marker else '' def _move_install_requirements_markers(self): """ Move requirements in `install_requires` that are using environment markers `extras_require`. """ # divide the install_requires into two sets, simple ones still # handled by install_requires and more complex ones handled # by extras_require. def is_simple_req(req): return not req.marker spec_inst_reqs = getattr(self, 'install_requires', None) or () inst_reqs = list(pkg_resources.parse_requirements(spec_inst_reqs)) simple_reqs = filter(is_simple_req, inst_reqs) complex_reqs = filterfalse(is_simple_req, inst_reqs) self.install_requires = list(map(str, simple_reqs)) for r in complex_reqs: self._tmp_extras_require[':' + str(r.marker)].append(r) self.extras_require = dict( (k, [str(r) for r in map(self._clean_req, v)]) for k, v in self._tmp_extras_require.items() ) def _clean_req(self, req): """ Given a Requirement, remove environment markers and return it. """ req.marker = None return req def _parse_config_files(self, filenames=None): """ Adapted from distutils.dist.Distribution.parse_config_files, this method provides the same functionality in subtly-improved ways. """ from setuptools.extern.six.moves.configparser import ConfigParser # Ignore install directory options if we have a venv if six.PY3 and sys.prefix != sys.base_prefix: ignore_options = [ 'install-base', 'install-platbase', 'install-lib', 'install-platlib', 'install-purelib', 'install-headers', 'install-scripts', 'install-data', 'prefix', 'exec-prefix', 'home', 'user', 'root'] else: ignore_options = [] ignore_options = frozenset(ignore_options) if filenames is None: filenames = self.find_config_files() if DEBUG: self.announce("Distribution.parse_config_files():") parser = ConfigParser() for filename in filenames: with io.open(filename, encoding='utf-8') as reader: if DEBUG: self.announce(" reading {filename}".format(**locals())) (parser.read_file if six.PY3 else parser.readfp)(reader) for section in parser.sections(): options = parser.options(section) opt_dict = self.get_option_dict(section) for opt in options: if opt != '__name__' and opt not in ignore_options: val = self._try_str(parser.get(section, opt)) opt = opt.replace('-', '_') opt_dict[opt] = (filename, val) # Make the ConfigParser forget everything (so we retain # the original filenames that options come from) parser.__init__() # If there was a "global" section in the config file, use it # to set Distribution options. if 'global' in self.command_options: for (opt, (src, val)) in self.command_options['global'].items(): alias = self.negative_opt.get(opt) try: if alias: setattr(self, alias, not strtobool(val)) elif opt in ('verbose', 'dry_run'): # ugh! setattr(self, opt, strtobool(val)) else: setattr(self, opt, val) except ValueError as msg: raise DistutilsOptionError(msg) @staticmethod def _try_str(val): """ On Python 2, much of distutils relies on string values being of type 'str' (bytes) and not unicode text. If the value can be safely encoded to bytes using the default encoding, prefer that. Why the default encoding? Because that value can be implicitly decoded back to text if needed. Ref #1653 """ if six.PY3: return val try: return val.encode() except UnicodeEncodeError: pass return val def _set_command_options(self, command_obj, option_dict=None): """ Set the options for 'command_obj' from 'option_dict'. Basically this means copying elements of a dictionary ('option_dict') to attributes of an instance ('command'). 'command_obj' must be a Command instance. If 'option_dict' is not supplied, uses the standard option dictionary for this command (from 'self.command_options'). (Adopted from distutils.dist.Distribution._set_command_options) """ command_name = command_obj.get_command_name() if option_dict is None: option_dict = self.get_option_dict(command_name) if DEBUG: self.announce(" setting options for '%s' command:" % command_name) for (option, (source, value)) in option_dict.items(): if DEBUG: self.announce(" %s = %s (from %s)" % (option, value, source)) try: bool_opts = [translate_longopt(o) for o in command_obj.boolean_options] except AttributeError: bool_opts = [] try: neg_opt = command_obj.negative_opt except AttributeError: neg_opt = {} try: is_string = isinstance(value, six.string_types) if option in neg_opt and is_string: setattr(command_obj, neg_opt[option], not strtobool(value)) elif option in bool_opts and is_string: setattr(command_obj, option, strtobool(value)) elif hasattr(command_obj, option): setattr(command_obj, option, value) else: raise DistutilsOptionError( "error in %s: command '%s' has no such option '%s'" % (source, command_name, option)) except ValueError as msg: raise DistutilsOptionError(msg) def parse_config_files(self, filenames=None, ignore_option_errors=False): """Parses configuration files from various levels and loads configuration. """ self._parse_config_files(filenames=filenames) parse_configuration(self, self.command_options, ignore_option_errors=ignore_option_errors) self._finalize_requires() def parse_command_line(self): """Process features after parsing command line options""" result = _Distribution.parse_command_line(self) if self.features: self._finalize_features() return result def _feature_attrname(self, name): """Convert feature name to corresponding option attribute name""" return 'with_' + name.replace('-', '_') def fetch_build_eggs(self, requires): """Resolve pre-setup requirements""" resolved_dists = pkg_resources.working_set.resolve( pkg_resources.parse_requirements(requires), installer=self.fetch_build_egg, replace_conflicting=True, ) for dist in resolved_dists: pkg_resources.working_set.add(dist, replace=True) return resolved_dists def finalize_options(self): _Distribution.finalize_options(self) if self.features: self._set_global_opts_from_features() for ep in pkg_resources.iter_entry_points('distutils.setup_keywords'): value = getattr(self, ep.name, None) if value is not None: ep.require(installer=self.fetch_build_egg) ep.load()(self, ep.name, value) if getattr(self, 'convert_2to3_doctests', None): # XXX may convert to set here when we can rely on set being builtin self.convert_2to3_doctests = [ os.path.abspath(p) for p in self.convert_2to3_doctests ] else: self.convert_2to3_doctests = [] def get_egg_cache_dir(self): egg_cache_dir = os.path.join(os.curdir, '.eggs') if not os.path.exists(egg_cache_dir): os.mkdir(egg_cache_dir) windows_support.hide_file(egg_cache_dir) readme_txt_filename = os.path.join(egg_cache_dir, 'README.txt') with open(readme_txt_filename, 'w') as f: f.write('This directory contains eggs that were downloaded ' 'by setuptools to build, test, and run plug-ins.\n\n') f.write('This directory caches those eggs to prevent ' 'repeated downloads.\n\n') f.write('However, it is safe to delete this directory.\n\n') return egg_cache_dir def fetch_build_egg(self, req): """Fetch an egg needed for building""" from setuptools.command.easy_install import easy_install dist = self.__class__({'script_args': ['easy_install']}) opts = dist.get_option_dict('easy_install') opts.clear() opts.update( (k, v) for k, v in self.get_option_dict('easy_install').items() if k in ( # don't use any other settings 'find_links', 'site_dirs', 'index_url', 'optimize', 'site_dirs', 'allow_hosts', )) if self.dependency_links: links = self.dependency_links[:] if 'find_links' in opts: links = opts['find_links'][1] + links opts['find_links'] = ('setup', links) install_dir = self.get_egg_cache_dir() cmd = easy_install( dist, args=["x"], install_dir=install_dir, exclude_scripts=True, always_copy=False, build_directory=None, editable=False, upgrade=False, multi_version=True, no_report=True, user=False ) cmd.ensure_finalized() return cmd.easy_install(req) def _set_global_opts_from_features(self): """Add --with-X/--without-X options based on optional features""" go = [] no = self.negative_opt.copy() for name, feature in self.features.items(): self._set_feature(name, None) feature.validate(self) if feature.optional: descr = feature.description incdef = ' (default)' excdef = '' if not feature.include_by_default(): excdef, incdef = incdef, excdef new = ( ('with-' + name, None, 'include ' + descr + incdef), ('without-' + name, None, 'exclude ' + descr + excdef), ) go.extend(new) no['without-' + name] = 'with-' + name self.global_options = self.feature_options = go + self.global_options self.negative_opt = self.feature_negopt = no def _finalize_features(self): """Add/remove features and resolve dependencies between them""" # First, flag all the enabled items (and thus their dependencies) for name, feature in self.features.items(): enabled = self.feature_is_included(name) if enabled or (enabled is None and feature.include_by_default()): feature.include_in(self) self._set_feature(name, 1) # Then disable the rest, so that off-by-default features don't # get flagged as errors when they're required by an enabled feature for name, feature in self.features.items(): if not self.feature_is_included(name): feature.exclude_from(self) self._set_feature(name, 0) def get_command_class(self, command): """Pluggable version of get_command_class()""" if command in self.cmdclass: return self.cmdclass[command] eps = pkg_resources.iter_entry_points('distutils.commands', command) for ep in eps: ep.require(installer=self.fetch_build_egg) self.cmdclass[command] = cmdclass = ep.load() return cmdclass else: return _Distribution.get_command_class(self, command) def print_commands(self): for ep in pkg_resources.iter_entry_points('distutils.commands'): if ep.name not in self.cmdclass: # don't require extras as the commands won't be invoked cmdclass = ep.resolve() self.cmdclass[ep.name] = cmdclass return _Distribution.print_commands(self) def get_command_list(self): for ep in pkg_resources.iter_entry_points('distutils.commands'): if ep.name not in self.cmdclass: # don't require extras as the commands won't be invoked cmdclass = ep.resolve() self.cmdclass[ep.name] = cmdclass return _Distribution.get_command_list(self) def _set_feature(self, name, status): """Set feature's inclusion status""" setattr(self, self._feature_attrname(name), status) def feature_is_included(self, name): """Return 1 if feature is included, 0 if excluded, 'None' if unknown""" return getattr(self, self._feature_attrname(name)) def include_feature(self, name): """Request inclusion of feature named 'name'""" if self.feature_is_included(name) == 0: descr = self.features[name].description raise DistutilsOptionError( descr + " is required, but was excluded or is not available" ) self.features[name].include_in(self) self._set_feature(name, 1) def include(self, **attrs): """Add items to distribution that are named in keyword arguments For example, 'dist.include(py_modules=["x"])' would add 'x' to the distribution's 'py_modules' attribute, if it was not already there. Currently, this method only supports inclusion for attributes that are lists or tuples. If you need to add support for adding to other attributes in this or a subclass, you can add an '_include_X' method, where 'X' is the name of the attribute. The method will be called with the value passed to 'include()'. So, 'dist.include(foo={"bar":"baz"})' will try to call 'dist._include_foo({"bar":"baz"})', which can then handle whatever special inclusion logic is needed. """ for k, v in attrs.items(): include = getattr(self, '_include_' + k, None) if include: include(v) else: self._include_misc(k, v) def exclude_package(self, package): """Remove packages, modules, and extensions in named package""" pfx = package + '.' if self.packages: self.packages = [ p for p in self.packages if p != package and not p.startswith(pfx) ] if self.py_modules: self.py_modules = [ p for p in self.py_modules if p != package and not p.startswith(pfx) ] if self.ext_modules: self.ext_modules = [ p for p in self.ext_modules if p.name != package and not p.name.startswith(pfx) ] def has_contents_for(self, package): """Return true if 'exclude_package(package)' would do something""" pfx = package + '.' for p in self.iter_distribution_names(): if p == package or p.startswith(pfx): return True def _exclude_misc(self, name, value): """Handle 'exclude()' for list/tuple attrs without a special handler""" if not isinstance(value, sequence): raise DistutilsSetupError( "%s: setting must be a list or tuple (%r)" % (name, value) ) try: old = getattr(self, name) except AttributeError: raise DistutilsSetupError( "%s: No such distribution setting" % name ) if old is not None and not isinstance(old, sequence): raise DistutilsSetupError( name + ": this setting cannot be changed via include/exclude" ) elif old: setattr(self, name, [item for item in old if item not in value]) def _include_misc(self, name, value): """Handle 'include()' for list/tuple attrs without a special handler""" if not isinstance(value, sequence): raise DistutilsSetupError( "%s: setting must be a list (%r)" % (name, value) ) try: old = getattr(self, name) except AttributeError: raise DistutilsSetupError( "%s: No such distribution setting" % name ) if old is None: setattr(self, name, value) elif not isinstance(old, sequence): raise DistutilsSetupError( name + ": this setting cannot be changed via include/exclude" ) else: new = [item for item in value if item not in old] setattr(self, name, old + new) def exclude(self, **attrs): """Remove items from distribution that are named in keyword arguments For example, 'dist.exclude(py_modules=["x"])' would remove 'x' from the distribution's 'py_modules' attribute. Excluding packages uses the 'exclude_package()' method, so all of the package's contained packages, modules, and extensions are also excluded. Currently, this method only supports exclusion from attributes that are lists or tuples. If you need to add support for excluding from other attributes in this or a subclass, you can add an '_exclude_X' method, where 'X' is the name of the attribute. The method will be called with the value passed to 'exclude()'. So, 'dist.exclude(foo={"bar":"baz"})' will try to call 'dist._exclude_foo({"bar":"baz"})', which can then handle whatever special exclusion logic is needed. """ for k, v in attrs.items(): exclude = getattr(self, '_exclude_' + k, None) if exclude: exclude(v) else: self._exclude_misc(k, v) def _exclude_packages(self, packages): if not isinstance(packages, sequence): raise DistutilsSetupError( "packages: setting must be a list or tuple (%r)" % (packages,) ) list(map(self.exclude_package, packages)) def _parse_command_opts(self, parser, args): # Remove --with-X/--without-X options when processing command args self.global_options = self.__class__.global_options self.negative_opt = self.__class__.negative_opt # First, expand any aliases command = args[0] aliases = self.get_option_dict('aliases') while command in aliases: src, alias = aliases[command] del aliases[command] # ensure each alias can expand only once! import shlex args[:1] = shlex.split(alias, True) command = args[0] nargs = _Distribution._parse_command_opts(self, parser, args) # Handle commands that want to consume all remaining arguments cmd_class = self.get_command_class(command) if getattr(cmd_class, 'command_consumes_arguments', None): self.get_option_dict(command)['args'] = ("command line", nargs) if nargs is not None: return [] return nargs def get_cmdline_options(self): """Return a '{cmd: {opt:val}}' map of all command-line options Option names are all long, but do not include the leading '--', and contain dashes rather than underscores. If the option doesn't take an argument (e.g. '--quiet'), the 'val' is 'None'. Note that options provided by config files are intentionally excluded. """ d = {} for cmd, opts in self.command_options.items(): for opt, (src, val) in opts.items(): if src != "command line": continue opt = opt.replace('_', '-') if val == 0: cmdobj = self.get_command_obj(cmd) neg_opt = self.negative_opt.copy() neg_opt.update(getattr(cmdobj, 'negative_opt', {})) for neg, pos in neg_opt.items(): if pos == opt: opt = neg val = None break else: raise AssertionError("Shouldn't be able to get here") elif val == 1: val = None d.setdefault(cmd, {})[opt] = val return d def iter_distribution_names(self): """Yield all packages, modules, and extension names in distribution""" for pkg in self.packages or (): yield pkg for module in self.py_modules or (): yield module for ext in self.ext_modules or (): if isinstance(ext, tuple): name, buildinfo = ext else: name = ext.name if name.endswith('module'): name = name[:-6] yield name def handle_display_options(self, option_order): """If there were any non-global "display-only" options (--help-commands or the metadata display options) on the command line, display the requested info and return true; else return false. """ import sys if six.PY2 or self.help_commands: return _Distribution.handle_display_options(self, option_order) # Stdout may be StringIO (e.g. in tests) if not isinstance(sys.stdout, io.TextIOWrapper): return _Distribution.handle_display_options(self, option_order) # Don't wrap stdout if utf-8 is already the encoding. Provides # workaround for #334. if sys.stdout.encoding.lower() in ('utf-8', 'utf8'): return _Distribution.handle_display_options(self, option_order) # Print metadata in UTF-8 no matter the platform encoding = sys.stdout.encoding errors = sys.stdout.errors newline = sys.platform != 'win32' and '\n' or None line_buffering = sys.stdout.line_buffering sys.stdout = io.TextIOWrapper( sys.stdout.detach(), 'utf-8', errors, newline, line_buffering) try: return _Distribution.handle_display_options(self, option_order) finally: sys.stdout = io.TextIOWrapper( sys.stdout.detach(), encoding, errors, newline, line_buffering) class Feature: """ **deprecated** -- The `Feature` facility was never completely implemented or supported, `has reported issues `_ and will be removed in a future version. A subset of the distribution that can be excluded if unneeded/wanted Features are created using these keyword arguments: 'description' -- a short, human readable description of the feature, to be used in error messages, and option help messages. 'standard' -- if true, the feature is included by default if it is available on the current system. Otherwise, the feature is only included if requested via a command line '--with-X' option, or if another included feature requires it. The default setting is 'False'. 'available' -- if true, the feature is available for installation on the current system. The default setting is 'True'. 'optional' -- if true, the feature's inclusion can be controlled from the command line, using the '--with-X' or '--without-X' options. If false, the feature's inclusion status is determined automatically, based on 'availabile', 'standard', and whether any other feature requires it. The default setting is 'True'. 'require_features' -- a string or sequence of strings naming features that should also be included if this feature is included. Defaults to empty list. May also contain 'Require' objects that should be added/removed from the distribution. 'remove' -- a string or list of strings naming packages to be removed from the distribution if this feature is *not* included. If the feature *is* included, this argument is ignored. This argument exists to support removing features that "crosscut" a distribution, such as defining a 'tests' feature that removes all the 'tests' subpackages provided by other features. The default for this argument is an empty list. (Note: the named package(s) or modules must exist in the base distribution when the 'setup()' function is initially called.) other keywords -- any other keyword arguments are saved, and passed to the distribution's 'include()' and 'exclude()' methods when the feature is included or excluded, respectively. So, for example, you could pass 'packages=["a","b"]' to cause packages 'a' and 'b' to be added or removed from the distribution as appropriate. A feature must include at least one 'requires', 'remove', or other keyword argument. Otherwise, it can't affect the distribution in any way. Note also that you can subclass 'Feature' to create your own specialized feature types that modify the distribution in other ways when included or excluded. See the docstrings for the various methods here for more detail. Aside from the methods, the only feature attributes that distributions look at are 'description' and 'optional'. """ @staticmethod def warn_deprecated(): msg = ( "Features are deprecated and will be removed in a future " "version. See https://github.com/pypa/setuptools/issues/65." ) warnings.warn(msg, DistDeprecationWarning, stacklevel=3) def __init__( self, description, standard=False, available=True, optional=True, require_features=(), remove=(), **extras): self.warn_deprecated() self.description = description self.standard = standard self.available = available self.optional = optional if isinstance(require_features, (str, Require)): require_features = require_features, self.require_features = [ r for r in require_features if isinstance(r, str) ] er = [r for r in require_features if not isinstance(r, str)] if er: extras['require_features'] = er if isinstance(remove, str): remove = remove, self.remove = remove self.extras = extras if not remove and not require_features and not extras: raise DistutilsSetupError( "Feature %s: must define 'require_features', 'remove', or " "at least one of 'packages', 'py_modules', etc." ) def include_by_default(self): """Should this feature be included by default?""" return self.available and self.standard def include_in(self, dist): """Ensure feature and its requirements are included in distribution You may override this in a subclass to perform additional operations on the distribution. Note that this method may be called more than once per feature, and so should be idempotent. """ if not self.available: raise DistutilsPlatformError( self.description + " is required, " "but is not available on this platform" ) dist.include(**self.extras) for f in self.require_features: dist.include_feature(f) def exclude_from(self, dist): """Ensure feature is excluded from distribution You may override this in a subclass to perform additional operations on the distribution. This method will be called at most once per feature, and only after all included features have been asked to include themselves. """ dist.exclude(**self.extras) if self.remove: for item in self.remove: dist.exclude_package(item) def validate(self, dist): """Verify that feature makes sense in context of distribution This method is called by the distribution just before it parses its command line. It checks to ensure that the 'remove' attribute, if any, contains only valid package/module names that are present in the base distribution when 'setup()' is called. You may override it in a subclass to perform any other required validation of the feature against a target distribution. """ for item in self.remove: if not dist.has_contents_for(item): raise DistutilsSetupError( "%s wants to be able to remove %s, but the distribution" " doesn't contain any packages or modules under %s" % (self.description, item, item) ) class DistDeprecationWarning(SetuptoolsDeprecationWarning): """Class for warning about deprecations in dist in setuptools. Not ignored by default, unlike DeprecationWarning.""" site-packages/setuptools/extension.py000064400000003301151464452710014101 0ustar00import re import functools import distutils.core import distutils.errors import distutils.extension from setuptools.extern.six.moves import map from .monkey import get_unpatched def _have_cython(): """ Return True if Cython can be imported. """ cython_impl = 'Cython.Distutils.build_ext' try: # from (cython_impl) import build_ext __import__(cython_impl, fromlist=['build_ext']).build_ext return True except Exception: pass return False # for compatibility have_pyrex = _have_cython _Extension = get_unpatched(distutils.core.Extension) class Extension(_Extension): """Extension that uses '.c' files in place of '.pyx' files""" def __init__(self, name, sources, *args, **kw): # The *args is needed for compatibility as calls may use positional # arguments. py_limited_api may be set only via keyword. self.py_limited_api = kw.pop("py_limited_api", False) _Extension.__init__(self, name, sources, *args, **kw) def _convert_pyx_sources_to_lang(self): """ Replace sources with .pyx extensions to sources with the target language extension. This mechanism allows language authors to supply pre-converted sources but to prefer the .pyx sources. """ if _have_cython(): # the build has Cython, so allow it to compile the .pyx files return lang = self.language or '' target_ext = '.cpp' if lang.lower() == 'c++' else '.c' sub = functools.partial(re.sub, '.pyx$', target_ext) self.sources = list(map(sub, self.sources)) class Library(Extension): """Just like a regular Extension, but built as a library instead""" site-packages/setuptools/windows_support.py000064400000001312151464452710015353 0ustar00import platform import ctypes def windows_only(func): if platform.system() != 'Windows': return lambda *args, **kwargs: None return func @windows_only def hide_file(path): """ Set the hidden attribute on a file or directory. From http://stackoverflow.com/questions/19622133/ `path` must be text. """ __import__('ctypes.wintypes') SetFileAttributes = ctypes.windll.kernel32.SetFileAttributesW SetFileAttributes.argtypes = ctypes.wintypes.LPWSTR, ctypes.wintypes.DWORD SetFileAttributes.restype = ctypes.wintypes.BOOL FILE_ATTRIBUTE_HIDDEN = 0x02 ret = SetFileAttributes(path, FILE_ATTRIBUTE_HIDDEN) if not ret: raise ctypes.WinError() site-packages/setuptools/wheel.py000064400000017636151464452710013211 0ustar00"""Wheels support.""" from distutils.util import get_platform import email import itertools import os import posixpath import re import zipfile import pkg_resources import setuptools from pkg_resources import parse_version from setuptools.extern.packaging.utils import canonicalize_name from setuptools.extern.six import PY3 from setuptools import pep425tags from setuptools.command.egg_info import write_requirements __metaclass__ = type WHEEL_NAME = re.compile( r"""^(?P.+?)-(?P\d.*?) ((-(?P\d.*?))?-(?P.+?)-(?P.+?)-(?P.+?) )\.whl$""", re.VERBOSE).match NAMESPACE_PACKAGE_INIT = '''\ try: __import__('pkg_resources').declare_namespace(__name__) except ImportError: __path__ = __import__('pkgutil').extend_path(__path__, __name__) ''' def unpack(src_dir, dst_dir): '''Move everything under `src_dir` to `dst_dir`, and delete the former.''' for dirpath, dirnames, filenames in os.walk(src_dir): subdir = os.path.relpath(dirpath, src_dir) for f in filenames: src = os.path.join(dirpath, f) dst = os.path.join(dst_dir, subdir, f) os.renames(src, dst) for n, d in reversed(list(enumerate(dirnames))): src = os.path.join(dirpath, d) dst = os.path.join(dst_dir, subdir, d) if not os.path.exists(dst): # Directory does not exist in destination, # rename it and prune it from os.walk list. os.renames(src, dst) del dirnames[n] # Cleanup. for dirpath, dirnames, filenames in os.walk(src_dir, topdown=True): assert not filenames os.rmdir(dirpath) class Wheel: def __init__(self, filename): match = WHEEL_NAME(os.path.basename(filename)) if match is None: raise ValueError('invalid wheel name: %r' % filename) self.filename = filename for k, v in match.groupdict().items(): setattr(self, k, v) def tags(self): '''List tags (py_version, abi, platform) supported by this wheel.''' return itertools.product( self.py_version.split('.'), self.abi.split('.'), self.platform.split('.'), ) def is_compatible(self): '''Is the wheel is compatible with the current platform?''' supported_tags = pep425tags.get_supported() return next((True for t in self.tags() if t in supported_tags), False) def egg_name(self): return pkg_resources.Distribution( project_name=self.project_name, version=self.version, platform=(None if self.platform == 'any' else get_platform()), ).egg_name() + '.egg' def get_dist_info(self, zf): # find the correct name of the .dist-info dir in the wheel file for member in zf.namelist(): dirname = posixpath.dirname(member) if (dirname.endswith('.dist-info') and canonicalize_name(dirname).startswith( canonicalize_name(self.project_name))): return dirname raise ValueError("unsupported wheel format. .dist-info not found") def install_as_egg(self, destination_eggdir): '''Install wheel as an egg directory.''' with zipfile.ZipFile(self.filename) as zf: self._install_as_egg(destination_eggdir, zf) def _install_as_egg(self, destination_eggdir, zf): dist_basename = '%s-%s' % (self.project_name, self.version) dist_info = self.get_dist_info(zf) dist_data = '%s.data' % dist_basename egg_info = os.path.join(destination_eggdir, 'EGG-INFO') self._convert_metadata(zf, destination_eggdir, dist_info, egg_info) self._move_data_entries(destination_eggdir, dist_data) self._fix_namespace_packages(egg_info, destination_eggdir) @staticmethod def _convert_metadata(zf, destination_eggdir, dist_info, egg_info): def get_metadata(name): with zf.open(posixpath.join(dist_info, name)) as fp: value = fp.read().decode('utf-8') if PY3 else fp.read() return email.parser.Parser().parsestr(value) wheel_metadata = get_metadata('WHEEL') # Check wheel format version is supported. wheel_version = parse_version(wheel_metadata.get('Wheel-Version')) wheel_v1 = ( parse_version('1.0') <= wheel_version < parse_version('2.0dev0') ) if not wheel_v1: raise ValueError( 'unsupported wheel format version: %s' % wheel_version) # Extract to target directory. os.mkdir(destination_eggdir) zf.extractall(destination_eggdir) # Convert metadata. dist_info = os.path.join(destination_eggdir, dist_info) dist = pkg_resources.Distribution.from_location( destination_eggdir, dist_info, metadata=pkg_resources.PathMetadata(destination_eggdir, dist_info), ) # Note: Evaluate and strip markers now, # as it's difficult to convert back from the syntax: # foobar; "linux" in sys_platform and extra == 'test' def raw_req(req): req.marker = None return str(req) install_requires = list(sorted(map(raw_req, dist.requires()))) extras_require = { extra: sorted( req for req in map(raw_req, dist.requires((extra,))) if req not in install_requires ) for extra in dist.extras } os.rename(dist_info, egg_info) os.rename( os.path.join(egg_info, 'METADATA'), os.path.join(egg_info, 'PKG-INFO'), ) setup_dist = setuptools.Distribution( attrs=dict( install_requires=install_requires, extras_require=extras_require, ), ) write_requirements( setup_dist.get_command_obj('egg_info'), None, os.path.join(egg_info, 'requires.txt'), ) @staticmethod def _move_data_entries(destination_eggdir, dist_data): """Move data entries to their correct location.""" dist_data = os.path.join(destination_eggdir, dist_data) dist_data_scripts = os.path.join(dist_data, 'scripts') if os.path.exists(dist_data_scripts): egg_info_scripts = os.path.join( destination_eggdir, 'EGG-INFO', 'scripts') os.mkdir(egg_info_scripts) for entry in os.listdir(dist_data_scripts): # Remove bytecode, as it's not properly handled # during easy_install scripts install phase. if entry.endswith('.pyc'): os.unlink(os.path.join(dist_data_scripts, entry)) else: os.rename( os.path.join(dist_data_scripts, entry), os.path.join(egg_info_scripts, entry), ) os.rmdir(dist_data_scripts) for subdir in filter(os.path.exists, ( os.path.join(dist_data, d) for d in ('data', 'headers', 'purelib', 'platlib') )): unpack(subdir, destination_eggdir) if os.path.exists(dist_data): os.rmdir(dist_data) @staticmethod def _fix_namespace_packages(egg_info, destination_eggdir): namespace_packages = os.path.join( egg_info, 'namespace_packages.txt') if os.path.exists(namespace_packages): with open(namespace_packages) as fp: namespace_packages = fp.read().split() for mod in namespace_packages: mod_dir = os.path.join(destination_eggdir, *mod.split('.')) mod_init = os.path.join(mod_dir, '__init__.py') if os.path.exists(mod_dir) and not os.path.exists(mod_init): with open(mod_init, 'w') as fp: fp.write(NAMESPACE_PACKAGE_INIT) site-packages/setuptools/build_meta.py000064400000022575151464452710014210 0ustar00"""A PEP 517 interface to setuptools Previously, when a user or a command line tool (let's call it a "frontend") needed to make a request of setuptools to take a certain action, for example, generating a list of installation requirements, the frontend would would call "setup.py egg_info" or "setup.py bdist_wheel" on the command line. PEP 517 defines a different method of interfacing with setuptools. Rather than calling "setup.py" directly, the frontend should: 1. Set the current directory to the directory with a setup.py file 2. Import this module into a safe python interpreter (one in which setuptools can potentially set global variables or crash hard). 3. Call one of the functions defined in PEP 517. What each function does is defined in PEP 517. However, here is a "casual" definition of the functions (this definition should not be relied on for bug reports or API stability): - `build_wheel`: build a wheel in the folder and return the basename - `get_requires_for_build_wheel`: get the `setup_requires` to build - `prepare_metadata_for_build_wheel`: get the `install_requires` - `build_sdist`: build an sdist in the folder and return the basename - `get_requires_for_build_sdist`: get the `setup_requires` to build Again, this is not a formal definition! Just a "taste" of the module. """ import io import os import sys import tokenize import shutil import contextlib import setuptools import distutils from setuptools.py31compat import TemporaryDirectory from pkg_resources import parse_requirements from pkg_resources.py31compat import makedirs __all__ = ['get_requires_for_build_sdist', 'get_requires_for_build_wheel', 'prepare_metadata_for_build_wheel', 'build_wheel', 'build_sdist', '__legacy__', 'SetupRequirementsError'] class SetupRequirementsError(BaseException): def __init__(self, specifiers): self.specifiers = specifiers class Distribution(setuptools.dist.Distribution): def fetch_build_eggs(self, specifiers): specifier_list = list(map(str, parse_requirements(specifiers))) raise SetupRequirementsError(specifier_list) @classmethod @contextlib.contextmanager def patch(cls): """ Replace distutils.dist.Distribution with this class for the duration of this context. """ orig = distutils.core.Distribution distutils.core.Distribution = cls try: yield finally: distutils.core.Distribution = orig def _to_str(s): """ Convert a filename to a string (on Python 2, explicitly a byte string, not Unicode) as distutils checks for the exact type str. """ if sys.version_info[0] == 2 and not isinstance(s, str): # Assume it's Unicode, as that's what the PEP says # should be provided. return s.encode(sys.getfilesystemencoding()) return s def _get_immediate_subdirectories(a_dir): return [name for name in os.listdir(a_dir) if os.path.isdir(os.path.join(a_dir, name))] def _file_with_extension(directory, extension): matching = ( f for f in os.listdir(directory) if f.endswith(extension) ) file, = matching return file def _open_setup_script(setup_script): if not os.path.exists(setup_script): # Supply a default setup.py return io.StringIO(u"from setuptools import setup; setup()") return getattr(tokenize, 'open', open)(setup_script) class _BuildMetaBackend(object): def _fix_config(self, config_settings): config_settings = config_settings or {} config_settings.setdefault('--global-option', []) return config_settings def _get_build_requires(self, config_settings, requirements): config_settings = self._fix_config(config_settings) sys.argv = sys.argv[:1] + ['egg_info'] + \ config_settings["--global-option"] try: with Distribution.patch(): self.run_setup() except SetupRequirementsError as e: requirements += e.specifiers return requirements def run_setup(self, setup_script='setup.py'): # Note that we can reuse our build directory between calls # Correctness comes first, then optimization later __file__ = setup_script __name__ = '__main__' with _open_setup_script(__file__) as f: code = f.read().replace(r'\r\n', r'\n') exec(compile(code, __file__, 'exec'), locals()) def get_requires_for_build_wheel(self, config_settings=None): config_settings = self._fix_config(config_settings) return self._get_build_requires(config_settings, requirements=['wheel']) def get_requires_for_build_sdist(self, config_settings=None): config_settings = self._fix_config(config_settings) return self._get_build_requires(config_settings, requirements=[]) def prepare_metadata_for_build_wheel(self, metadata_directory, config_settings=None): sys.argv = sys.argv[:1] + ['dist_info', '--egg-base', _to_str(metadata_directory)] self.run_setup() dist_info_directory = metadata_directory while True: dist_infos = [f for f in os.listdir(dist_info_directory) if f.endswith('.dist-info')] if (len(dist_infos) == 0 and len(_get_immediate_subdirectories(dist_info_directory)) == 1): dist_info_directory = os.path.join( dist_info_directory, os.listdir(dist_info_directory)[0]) continue assert len(dist_infos) == 1 break # PEP 517 requires that the .dist-info directory be placed in the # metadata_directory. To comply, we MUST copy the directory to the root if dist_info_directory != metadata_directory: shutil.move( os.path.join(dist_info_directory, dist_infos[0]), metadata_directory) shutil.rmtree(dist_info_directory, ignore_errors=True) return dist_infos[0] def _build_with_temp_dir(self, setup_command, result_extension, result_directory, config_settings): config_settings = self._fix_config(config_settings) result_directory = os.path.abspath(result_directory) # Build in a temporary directory, then copy to the target. makedirs(result_directory, exist_ok=True) with TemporaryDirectory(dir=result_directory) as tmp_dist_dir: sys.argv = (sys.argv[:1] + setup_command + ['--dist-dir', tmp_dist_dir] + config_settings["--global-option"]) self.run_setup() result_basename = _file_with_extension(tmp_dist_dir, result_extension) result_path = os.path.join(result_directory, result_basename) if os.path.exists(result_path): # os.rename will fail overwriting on non-Unix. os.remove(result_path) os.rename(os.path.join(tmp_dist_dir, result_basename), result_path) return result_basename def build_wheel(self, wheel_directory, config_settings=None, metadata_directory=None): return self._build_with_temp_dir(['bdist_wheel'], '.whl', wheel_directory, config_settings) def build_sdist(self, sdist_directory, config_settings=None): return self._build_with_temp_dir(['sdist', '--formats', 'gztar'], '.tar.gz', sdist_directory, config_settings) class _BuildMetaLegacyBackend(_BuildMetaBackend): """Compatibility backend for setuptools This is a version of setuptools.build_meta that endeavors to maintain backwards compatibility with pre-PEP 517 modes of invocation. It exists as a temporary bridge between the old packaging mechanism and the new packaging mechanism, and will eventually be removed. """ def run_setup(self, setup_script='setup.py'): # In order to maintain compatibility with scripts assuming that # the setup.py script is in a directory on the PYTHONPATH, inject # '' into sys.path. (pypa/setuptools#1642) sys_path = list(sys.path) # Save the original path script_dir = os.path.dirname(os.path.abspath(setup_script)) if script_dir not in sys.path: sys.path.insert(0, script_dir) try: super(_BuildMetaLegacyBackend, self).run_setup(setup_script=setup_script) finally: # While PEP 517 frontends should be calling each hook in a fresh # subprocess according to the standard (and thus it should not be # strictly necessary to restore the old sys.path), we'll restore # the original path so that the path manipulation does not persist # within the hook after run_setup is called. sys.path[:] = sys_path # The primary backend _BACKEND = _BuildMetaBackend() get_requires_for_build_wheel = _BACKEND.get_requires_for_build_wheel get_requires_for_build_sdist = _BACKEND.get_requires_for_build_sdist prepare_metadata_for_build_wheel = _BACKEND.prepare_metadata_for_build_wheel build_wheel = _BACKEND.build_wheel build_sdist = _BACKEND.build_sdist # The legacy backend __legacy__ = _BuildMetaLegacyBackend() site-packages/setuptools/archive_util.py000064400000014700151464452710014550 0ustar00"""Utilities for extracting common archive formats""" import zipfile import tarfile import os import shutil import posixpath import contextlib from distutils.errors import DistutilsError from pkg_resources import ensure_directory __all__ = [ "unpack_archive", "unpack_zipfile", "unpack_tarfile", "default_filter", "UnrecognizedFormat", "extraction_drivers", "unpack_directory", ] class UnrecognizedFormat(DistutilsError): """Couldn't recognize the archive type""" def default_filter(src, dst): """The default progress/filter callback; returns True for all files""" return dst def unpack_archive(filename, extract_dir, progress_filter=default_filter, drivers=None): """Unpack `filename` to `extract_dir`, or raise ``UnrecognizedFormat`` `progress_filter` is a function taking two arguments: a source path internal to the archive ('/'-separated), and a filesystem path where it will be extracted. The callback must return the desired extract path (which may be the same as the one passed in), or else ``None`` to skip that file or directory. The callback can thus be used to report on the progress of the extraction, as well as to filter the items extracted or alter their extraction paths. `drivers`, if supplied, must be a non-empty sequence of functions with the same signature as this function (minus the `drivers` argument), that raise ``UnrecognizedFormat`` if they do not support extracting the designated archive type. The `drivers` are tried in sequence until one is found that does not raise an error, or until all are exhausted (in which case ``UnrecognizedFormat`` is raised). If you do not supply a sequence of drivers, the module's ``extraction_drivers`` constant will be used, which means that ``unpack_zipfile`` and ``unpack_tarfile`` will be tried, in that order. """ for driver in drivers or extraction_drivers: try: driver(filename, extract_dir, progress_filter) except UnrecognizedFormat: continue else: return else: raise UnrecognizedFormat( "Not a recognized archive type: %s" % filename ) def unpack_directory(filename, extract_dir, progress_filter=default_filter): """"Unpack" a directory, using the same interface as for archives Raises ``UnrecognizedFormat`` if `filename` is not a directory """ if not os.path.isdir(filename): raise UnrecognizedFormat("%s is not a directory" % filename) paths = { filename: ('', extract_dir), } for base, dirs, files in os.walk(filename): src, dst = paths[base] for d in dirs: paths[os.path.join(base, d)] = src + d + '/', os.path.join(dst, d) for f in files: target = os.path.join(dst, f) target = progress_filter(src + f, target) if not target: # skip non-files continue ensure_directory(target) f = os.path.join(base, f) shutil.copyfile(f, target) shutil.copystat(f, target) def unpack_zipfile(filename, extract_dir, progress_filter=default_filter): """Unpack zip `filename` to `extract_dir` Raises ``UnrecognizedFormat`` if `filename` is not a zipfile (as determined by ``zipfile.is_zipfile()``). See ``unpack_archive()`` for an explanation of the `progress_filter` argument. """ if not zipfile.is_zipfile(filename): raise UnrecognizedFormat("%s is not a zip file" % (filename,)) with zipfile.ZipFile(filename) as z: for info in z.infolist(): name = info.filename # don't extract absolute paths or ones with .. in them if name.startswith('/') or '..' in name.split('/'): continue target = os.path.join(extract_dir, *name.split('/')) target = progress_filter(name, target) if not target: continue if name.endswith('/'): # directory ensure_directory(target) else: # file ensure_directory(target) data = z.read(info.filename) with open(target, 'wb') as f: f.write(data) unix_attributes = info.external_attr >> 16 if unix_attributes: os.chmod(target, unix_attributes) def unpack_tarfile(filename, extract_dir, progress_filter=default_filter): """Unpack tar/tar.gz/tar.bz2 `filename` to `extract_dir` Raises ``UnrecognizedFormat`` if `filename` is not a tarfile (as determined by ``tarfile.open()``). See ``unpack_archive()`` for an explanation of the `progress_filter` argument. """ try: tarobj = tarfile.open(filename) except tarfile.TarError: raise UnrecognizedFormat( "%s is not a compressed or uncompressed tar file" % (filename,) ) with contextlib.closing(tarobj): # don't do any chowning! tarobj.chown = lambda *args: None for member in tarobj: name = member.name # don't extract absolute paths or ones with .. in them if not name.startswith('/') and '..' not in name.split('/'): prelim_dst = os.path.join(extract_dir, *name.split('/')) # resolve any links and to extract the link targets as normal # files while member is not None and (member.islnk() or member.issym()): linkpath = member.linkname if member.issym(): base = posixpath.dirname(member.name) linkpath = posixpath.join(base, linkpath) linkpath = posixpath.normpath(linkpath) member = tarobj._getmember(linkpath) if member is not None and (member.isfile() or member.isdir()): final_dst = progress_filter(name, prelim_dst) if final_dst: if final_dst.endswith(os.sep): final_dst = final_dst[:-1] try: # XXX Ugh tarobj._extract_member(member, final_dst) except tarfile.ExtractError: # chown/chmod/mkfifo/mknode/makedev failed pass return True extraction_drivers = unpack_directory, unpack_zipfile, unpack_tarfile site-packages/setuptools/py34compat.py000064400000000365151464452710014077 0ustar00import importlib try: import importlib.util except ImportError: pass try: module_from_spec = importlib.util.module_from_spec except AttributeError: def module_from_spec(spec): return spec.loader.load_module(spec.name) site-packages/setuptools/sandbox.py000064400000033704151464452710013535 0ustar00import os import sys import tempfile import operator import functools import itertools import re import contextlib import pickle import textwrap from setuptools.extern import six from setuptools.extern.six.moves import builtins, map import pkg_resources.py31compat if sys.platform.startswith('java'): import org.python.modules.posix.PosixModule as _os else: _os = sys.modules[os.name] try: _file = file except NameError: _file = None _open = open from distutils.errors import DistutilsError from pkg_resources import working_set __all__ = [ "AbstractSandbox", "DirectorySandbox", "SandboxViolation", "run_setup", ] def _execfile(filename, globals, locals=None): """ Python 3 implementation of execfile. """ mode = 'rb' with open(filename, mode) as stream: script = stream.read() if locals is None: locals = globals code = compile(script, filename, 'exec') exec(code, globals, locals) @contextlib.contextmanager def save_argv(repl=None): saved = sys.argv[:] if repl is not None: sys.argv[:] = repl try: yield saved finally: sys.argv[:] = saved @contextlib.contextmanager def save_path(): saved = sys.path[:] try: yield saved finally: sys.path[:] = saved @contextlib.contextmanager def override_temp(replacement): """ Monkey-patch tempfile.tempdir with replacement, ensuring it exists """ pkg_resources.py31compat.makedirs(replacement, exist_ok=True) saved = tempfile.tempdir tempfile.tempdir = replacement try: yield finally: tempfile.tempdir = saved @contextlib.contextmanager def pushd(target): saved = os.getcwd() os.chdir(target) try: yield saved finally: os.chdir(saved) class UnpickleableException(Exception): """ An exception representing another Exception that could not be pickled. """ @staticmethod def dump(type, exc): """ Always return a dumped (pickled) type and exc. If exc can't be pickled, wrap it in UnpickleableException first. """ try: return pickle.dumps(type), pickle.dumps(exc) except Exception: # get UnpickleableException inside the sandbox from setuptools.sandbox import UnpickleableException as cls return cls.dump(cls, cls(repr(exc))) class ExceptionSaver: """ A Context Manager that will save an exception, serialized, and restore it later. """ def __enter__(self): return self def __exit__(self, type, exc, tb): if not exc: return # dump the exception self._saved = UnpickleableException.dump(type, exc) self._tb = tb # suppress the exception return True def resume(self): "restore and re-raise any exception" if '_saved' not in vars(self): return type, exc = map(pickle.loads, self._saved) six.reraise(type, exc, self._tb) @contextlib.contextmanager def save_modules(): """ Context in which imported modules are saved. Translates exceptions internal to the context into the equivalent exception outside the context. """ saved = sys.modules.copy() with ExceptionSaver() as saved_exc: yield saved sys.modules.update(saved) # remove any modules imported since del_modules = ( mod_name for mod_name in sys.modules if mod_name not in saved # exclude any encodings modules. See #285 and not mod_name.startswith('encodings.') ) _clear_modules(del_modules) saved_exc.resume() def _clear_modules(module_names): for mod_name in list(module_names): del sys.modules[mod_name] @contextlib.contextmanager def save_pkg_resources_state(): saved = pkg_resources.__getstate__() try: yield saved finally: pkg_resources.__setstate__(saved) @contextlib.contextmanager def setup_context(setup_dir): temp_dir = os.path.join(setup_dir, 'temp') with save_pkg_resources_state(): with save_modules(): hide_setuptools() with save_path(): with save_argv(): with override_temp(temp_dir): with pushd(setup_dir): # ensure setuptools commands are available __import__('setuptools') yield def _needs_hiding(mod_name): """ >>> _needs_hiding('setuptools') True >>> _needs_hiding('pkg_resources') True >>> _needs_hiding('setuptools_plugin') False >>> _needs_hiding('setuptools.__init__') True >>> _needs_hiding('distutils') True >>> _needs_hiding('os') False >>> _needs_hiding('Cython') True """ pattern = re.compile(r'(setuptools|pkg_resources|distutils|Cython)(\.|$)') return bool(pattern.match(mod_name)) def hide_setuptools(): """ Remove references to setuptools' modules from sys.modules to allow the invocation to import the most appropriate setuptools. This technique is necessary to avoid issues such as #315 where setuptools upgrading itself would fail to find a function declared in the metadata. """ modules = filter(_needs_hiding, sys.modules) _clear_modules(modules) def run_setup(setup_script, args): """Run a distutils setup script, sandboxed in its directory""" setup_dir = os.path.abspath(os.path.dirname(setup_script)) with setup_context(setup_dir): try: sys.argv[:] = [setup_script] + list(args) sys.path.insert(0, setup_dir) # reset to include setup dir, w/clean callback list working_set.__init__() working_set.callbacks.append(lambda dist: dist.activate()) # __file__ should be a byte string on Python 2 (#712) dunder_file = ( setup_script if isinstance(setup_script, str) else setup_script.encode(sys.getfilesystemencoding()) ) with DirectorySandbox(setup_dir): ns = dict(__file__=dunder_file, __name__='__main__') _execfile(setup_script, ns) except SystemExit as v: if v.args and v.args[0]: raise # Normal exit, just return class AbstractSandbox: """Wrap 'os' module and 'open()' builtin for virtualizing setup scripts""" _active = False def __init__(self): self._attrs = [ name for name in dir(_os) if not name.startswith('_') and hasattr(self, name) ] def _copy(self, source): for name in self._attrs: setattr(os, name, getattr(source, name)) def __enter__(self): self._copy(self) if _file: builtins.file = self._file builtins.open = self._open self._active = True def __exit__(self, exc_type, exc_value, traceback): self._active = False if _file: builtins.file = _file builtins.open = _open self._copy(_os) def run(self, func): """Run 'func' under os sandboxing""" with self: return func() def _mk_dual_path_wrapper(name): original = getattr(_os, name) def wrap(self, src, dst, *args, **kw): if self._active: src, dst = self._remap_pair(name, src, dst, *args, **kw) return original(src, dst, *args, **kw) return wrap for name in ["rename", "link", "symlink"]: if hasattr(_os, name): locals()[name] = _mk_dual_path_wrapper(name) def _mk_single_path_wrapper(name, original=None): original = original or getattr(_os, name) def wrap(self, path, *args, **kw): if self._active: path = self._remap_input(name, path, *args, **kw) return original(path, *args, **kw) return wrap if _file: _file = _mk_single_path_wrapper('file', _file) _open = _mk_single_path_wrapper('open', _open) for name in [ "stat", "listdir", "chdir", "open", "chmod", "chown", "mkdir", "remove", "unlink", "rmdir", "utime", "lchown", "chroot", "lstat", "startfile", "mkfifo", "mknod", "pathconf", "access" ]: if hasattr(_os, name): locals()[name] = _mk_single_path_wrapper(name) def _mk_single_with_return(name): original = getattr(_os, name) def wrap(self, path, *args, **kw): if self._active: path = self._remap_input(name, path, *args, **kw) return self._remap_output(name, original(path, *args, **kw)) return original(path, *args, **kw) return wrap for name in ['readlink', 'tempnam']: if hasattr(_os, name): locals()[name] = _mk_single_with_return(name) def _mk_query(name): original = getattr(_os, name) def wrap(self, *args, **kw): retval = original(*args, **kw) if self._active: return self._remap_output(name, retval) return retval return wrap for name in ['getcwd', 'tmpnam']: if hasattr(_os, name): locals()[name] = _mk_query(name) def _validate_path(self, path): """Called to remap or validate any path, whether input or output""" return path def _remap_input(self, operation, path, *args, **kw): """Called for path inputs""" return self._validate_path(path) def _remap_output(self, operation, path): """Called for path outputs""" return self._validate_path(path) def _remap_pair(self, operation, src, dst, *args, **kw): """Called for path pairs like rename, link, and symlink operations""" return ( self._remap_input(operation + '-from', src, *args, **kw), self._remap_input(operation + '-to', dst, *args, **kw) ) if hasattr(os, 'devnull'): _EXCEPTIONS = [os.devnull,] else: _EXCEPTIONS = [] class DirectorySandbox(AbstractSandbox): """Restrict operations to a single subdirectory - pseudo-chroot""" write_ops = dict.fromkeys([ "open", "chmod", "chown", "mkdir", "remove", "unlink", "rmdir", "utime", "lchown", "chroot", "mkfifo", "mknod", "tempnam", ]) _exception_patterns = [ # Allow lib2to3 to attempt to save a pickled grammar object (#121) r'.*lib2to3.*\.pickle$', ] "exempt writing to paths that match the pattern" def __init__(self, sandbox, exceptions=_EXCEPTIONS): self._sandbox = os.path.normcase(os.path.realpath(sandbox)) self._prefix = os.path.join(self._sandbox, '') self._exceptions = [ os.path.normcase(os.path.realpath(path)) for path in exceptions ] AbstractSandbox.__init__(self) def _violation(self, operation, *args, **kw): from setuptools.sandbox import SandboxViolation raise SandboxViolation(operation, args, kw) if _file: def _file(self, path, mode='r', *args, **kw): if mode not in ('r', 'rt', 'rb', 'rU', 'U') and not self._ok(path): self._violation("file", path, mode, *args, **kw) return _file(path, mode, *args, **kw) def _open(self, path, mode='r', *args, **kw): if mode not in ('r', 'rt', 'rb', 'rU', 'U') and not self._ok(path): self._violation("open", path, mode, *args, **kw) return _open(path, mode, *args, **kw) def tmpnam(self): self._violation("tmpnam") def _ok(self, path): active = self._active try: self._active = False realpath = os.path.normcase(os.path.realpath(path)) return ( self._exempted(realpath) or realpath == self._sandbox or realpath.startswith(self._prefix) ) finally: self._active = active def _exempted(self, filepath): start_matches = ( filepath.startswith(exception) for exception in self._exceptions ) pattern_matches = ( re.match(pattern, filepath) for pattern in self._exception_patterns ) candidates = itertools.chain(start_matches, pattern_matches) return any(candidates) def _remap_input(self, operation, path, *args, **kw): """Called for path inputs""" if operation in self.write_ops and not self._ok(path): self._violation(operation, os.path.realpath(path), *args, **kw) return path def _remap_pair(self, operation, src, dst, *args, **kw): """Called for path pairs like rename, link, and symlink operations""" if not self._ok(src) or not self._ok(dst): self._violation(operation, src, dst, *args, **kw) return (src, dst) def open(self, file, flags, mode=0o777, *args, **kw): """Called for low-level os.open()""" if flags & WRITE_FLAGS and not self._ok(file): self._violation("os.open", file, flags, mode, *args, **kw) return _os.open(file, flags, mode, *args, **kw) WRITE_FLAGS = functools.reduce( operator.or_, [getattr(_os, a, 0) for a in "O_WRONLY O_RDWR O_APPEND O_CREAT O_TRUNC O_TEMPORARY".split()] ) class SandboxViolation(DistutilsError): """A setup script attempted to modify the filesystem outside the sandbox""" tmpl = textwrap.dedent(""" SandboxViolation: {cmd}{args!r} {kwargs} The package setup script has attempted to modify files on your system that are not within the EasyInstall build area, and has been aborted. This package cannot be safely installed by EasyInstall, and may not support alternate installation locations even if you run its setup script by hand. Please inform the package's author and the EasyInstall maintainers to find out if a fix or workaround is available. """).lstrip() def __str__(self): cmd, args, kwargs = self.args return self.tmpl.format(**locals()) site-packages/setuptools/_vendor/packaging/requirements.py000064400000010367151464452710020202 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import string import re from setuptools.extern.pyparsing import stringStart, stringEnd, originalTextFor, ParseException from setuptools.extern.pyparsing import ZeroOrMore, Word, Optional, Regex, Combine from setuptools.extern.pyparsing import Literal as L # noqa from setuptools.extern.six.moves.urllib import parse as urlparse from .markers import MARKER_EXPR, Marker from .specifiers import LegacySpecifier, Specifier, SpecifierSet class InvalidRequirement(ValueError): """ An invalid requirement was found, users should refer to PEP 508. """ ALPHANUM = Word(string.ascii_letters + string.digits) LBRACKET = L("[").suppress() RBRACKET = L("]").suppress() LPAREN = L("(").suppress() RPAREN = L(")").suppress() COMMA = L(",").suppress() SEMICOLON = L(";").suppress() AT = L("@").suppress() PUNCTUATION = Word("-_.") IDENTIFIER_END = ALPHANUM | (ZeroOrMore(PUNCTUATION) + ALPHANUM) IDENTIFIER = Combine(ALPHANUM + ZeroOrMore(IDENTIFIER_END)) NAME = IDENTIFIER("name") EXTRA = IDENTIFIER URI = Regex(r'[^ ]+')("url") URL = (AT + URI) EXTRAS_LIST = EXTRA + ZeroOrMore(COMMA + EXTRA) EXTRAS = (LBRACKET + Optional(EXTRAS_LIST) + RBRACKET)("extras") VERSION_PEP440 = Regex(Specifier._regex_str, re.VERBOSE | re.IGNORECASE) VERSION_LEGACY = Regex(LegacySpecifier._regex_str, re.VERBOSE | re.IGNORECASE) VERSION_ONE = VERSION_PEP440 ^ VERSION_LEGACY VERSION_MANY = Combine(VERSION_ONE + ZeroOrMore(COMMA + VERSION_ONE), joinString=",", adjacent=False)("_raw_spec") _VERSION_SPEC = Optional(((LPAREN + VERSION_MANY + RPAREN) | VERSION_MANY)) _VERSION_SPEC.setParseAction(lambda s, l, t: t._raw_spec or '') VERSION_SPEC = originalTextFor(_VERSION_SPEC)("specifier") VERSION_SPEC.setParseAction(lambda s, l, t: t[1]) MARKER_EXPR = originalTextFor(MARKER_EXPR())("marker") MARKER_EXPR.setParseAction( lambda s, l, t: Marker(s[t._original_start:t._original_end]) ) MARKER_SEPERATOR = SEMICOLON MARKER = MARKER_SEPERATOR + MARKER_EXPR VERSION_AND_MARKER = VERSION_SPEC + Optional(MARKER) URL_AND_MARKER = URL + Optional(MARKER) NAMED_REQUIREMENT = \ NAME + Optional(EXTRAS) + (URL_AND_MARKER | VERSION_AND_MARKER) REQUIREMENT = stringStart + NAMED_REQUIREMENT + stringEnd class Requirement(object): """Parse a requirement. Parse a given requirement string into its parts, such as name, specifier, URL, and extras. Raises InvalidRequirement on a badly-formed requirement string. """ # TODO: Can we test whether something is contained within a requirement? # If so how do we do that? Do we need to test against the _name_ of # the thing as well as the version? What about the markers? # TODO: Can we normalize the name and extra name? def __init__(self, requirement_string): try: req = REQUIREMENT.parseString(requirement_string) except ParseException as e: raise InvalidRequirement( "Invalid requirement, parse error at \"{0!r}\"".format( requirement_string[e.loc:e.loc + 8])) self.name = req.name if req.url: parsed_url = urlparse.urlparse(req.url) if not (parsed_url.scheme and parsed_url.netloc) or ( not parsed_url.scheme and not parsed_url.netloc): raise InvalidRequirement("Invalid URL given") self.url = req.url else: self.url = None self.extras = set(req.extras.asList() if req.extras else []) self.specifier = SpecifierSet(req.specifier) self.marker = req.marker if req.marker else None def __str__(self): parts = [self.name] if self.extras: parts.append("[{0}]".format(",".join(sorted(self.extras)))) if self.specifier: parts.append(str(self.specifier)) if self.url: parts.append("@ {0}".format(self.url)) if self.marker: parts.append("; {0}".format(self.marker)) return "".join(parts) def __repr__(self): return "".format(str(self)) site-packages/setuptools/_vendor/packaging/_structures.py000064400000002610151464452710020031 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function class Infinity(object): def __repr__(self): return "Infinity" def __hash__(self): return hash(repr(self)) def __lt__(self, other): return False def __le__(self, other): return False def __eq__(self, other): return isinstance(other, self.__class__) def __ne__(self, other): return not isinstance(other, self.__class__) def __gt__(self, other): return True def __ge__(self, other): return True def __neg__(self): return NegativeInfinity Infinity = Infinity() class NegativeInfinity(object): def __repr__(self): return "-Infinity" def __hash__(self): return hash(repr(self)) def __lt__(self, other): return True def __le__(self, other): return True def __eq__(self, other): return isinstance(other, self.__class__) def __ne__(self, other): return not isinstance(other, self.__class__) def __gt__(self, other): return False def __ge__(self, other): return False def __neg__(self): return Infinity NegativeInfinity = NegativeInfinity() site-packages/setuptools/_vendor/packaging/utils.py000064400000000645151464452710016615 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import re _canonicalize_regex = re.compile(r"[-_.]+") def canonicalize_name(name): # This is taken from PEP 503. return _canonicalize_regex.sub("-", name).lower() site-packages/setuptools/_vendor/packaging/version.py000064400000026444151464452710017147 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import collections import itertools import re from ._structures import Infinity __all__ = [ "parse", "Version", "LegacyVersion", "InvalidVersion", "VERSION_PATTERN" ] _Version = collections.namedtuple( "_Version", ["epoch", "release", "dev", "pre", "post", "local"], ) def parse(version): """ Parse the given version string and return either a :class:`Version` object or a :class:`LegacyVersion` object depending on if the given version is a valid PEP 440 version or a legacy version. """ try: return Version(version) except InvalidVersion: return LegacyVersion(version) class InvalidVersion(ValueError): """ An invalid version was found, users should refer to PEP 440. """ class _BaseVersion(object): def __hash__(self): return hash(self._key) def __lt__(self, other): return self._compare(other, lambda s, o: s < o) def __le__(self, other): return self._compare(other, lambda s, o: s <= o) def __eq__(self, other): return self._compare(other, lambda s, o: s == o) def __ge__(self, other): return self._compare(other, lambda s, o: s >= o) def __gt__(self, other): return self._compare(other, lambda s, o: s > o) def __ne__(self, other): return self._compare(other, lambda s, o: s != o) def _compare(self, other, method): if not isinstance(other, _BaseVersion): return NotImplemented return method(self._key, other._key) class LegacyVersion(_BaseVersion): def __init__(self, version): self._version = str(version) self._key = _legacy_cmpkey(self._version) def __str__(self): return self._version def __repr__(self): return "".format(repr(str(self))) @property def public(self): return self._version @property def base_version(self): return self._version @property def local(self): return None @property def is_prerelease(self): return False @property def is_postrelease(self): return False _legacy_version_component_re = re.compile( r"(\d+ | [a-z]+ | \.| -)", re.VERBOSE, ) _legacy_version_replacement_map = { "pre": "c", "preview": "c", "-": "final-", "rc": "c", "dev": "@", } def _parse_version_parts(s): for part in _legacy_version_component_re.split(s): part = _legacy_version_replacement_map.get(part, part) if not part or part == ".": continue if part[:1] in "0123456789": # pad for numeric comparison yield part.zfill(8) else: yield "*" + part # ensure that alpha/beta/candidate are before final yield "*final" def _legacy_cmpkey(version): # We hardcode an epoch of -1 here. A PEP 440 version can only have a epoch # greater than or equal to 0. This will effectively put the LegacyVersion, # which uses the defacto standard originally implemented by setuptools, # as before all PEP 440 versions. epoch = -1 # This scheme is taken from pkg_resources.parse_version setuptools prior to # it's adoption of the packaging library. parts = [] for part in _parse_version_parts(version.lower()): if part.startswith("*"): # remove "-" before a prerelease tag if part < "*final": while parts and parts[-1] == "*final-": parts.pop() # remove trailing zeros from each series of numeric parts while parts and parts[-1] == "00000000": parts.pop() parts.append(part) parts = tuple(parts) return epoch, parts # Deliberately not anchored to the start and end of the string, to make it # easier for 3rd party code to reuse VERSION_PATTERN = r""" v? (?: (?:(?P[0-9]+)!)? # epoch (?P[0-9]+(?:\.[0-9]+)*) # release segment (?P
                                          # pre-release
            [-_\.]?
            (?P(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P[0-9]+)?
        )?
        (?P                                         # post release
            (?:-(?P[0-9]+))
            |
            (?:
                [-_\.]?
                (?Ppost|rev|r)
                [-_\.]?
                (?P[0-9]+)?
            )
        )?
        (?P                                          # dev release
            [-_\.]?
            (?Pdev)
            [-_\.]?
            (?P[0-9]+)?
        )?
    )
    (?:\+(?P[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
"""


class Version(_BaseVersion):

    _regex = re.compile(
        r"^\s*" + VERSION_PATTERN + r"\s*$",
        re.VERBOSE | re.IGNORECASE,
    )

    def __init__(self, version):
        # Validate the version and parse it into pieces
        match = self._regex.search(version)
        if not match:
            raise InvalidVersion("Invalid version: '{0}'".format(version))

        # Store the parsed out pieces of the version
        self._version = _Version(
            epoch=int(match.group("epoch")) if match.group("epoch") else 0,
            release=tuple(int(i) for i in match.group("release").split(".")),
            pre=_parse_letter_version(
                match.group("pre_l"),
                match.group("pre_n"),
            ),
            post=_parse_letter_version(
                match.group("post_l"),
                match.group("post_n1") or match.group("post_n2"),
            ),
            dev=_parse_letter_version(
                match.group("dev_l"),
                match.group("dev_n"),
            ),
            local=_parse_local_version(match.group("local")),
        )

        # Generate a key which will be used for sorting
        self._key = _cmpkey(
            self._version.epoch,
            self._version.release,
            self._version.pre,
            self._version.post,
            self._version.dev,
            self._version.local,
        )

    def __repr__(self):
        return "".format(repr(str(self)))

    def __str__(self):
        parts = []

        # Epoch
        if self._version.epoch != 0:
            parts.append("{0}!".format(self._version.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self._version.release))

        # Pre-release
        if self._version.pre is not None:
            parts.append("".join(str(x) for x in self._version.pre))

        # Post-release
        if self._version.post is not None:
            parts.append(".post{0}".format(self._version.post[1]))

        # Development release
        if self._version.dev is not None:
            parts.append(".dev{0}".format(self._version.dev[1]))

        # Local version segment
        if self._version.local is not None:
            parts.append(
                "+{0}".format(".".join(str(x) for x in self._version.local))
            )

        return "".join(parts)

    @property
    def public(self):
        return str(self).split("+", 1)[0]

    @property
    def base_version(self):
        parts = []

        # Epoch
        if self._version.epoch != 0:
            parts.append("{0}!".format(self._version.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self._version.release))

        return "".join(parts)

    @property
    def local(self):
        version_string = str(self)
        if "+" in version_string:
            return version_string.split("+", 1)[1]

    @property
    def is_prerelease(self):
        return bool(self._version.dev or self._version.pre)

    @property
    def is_postrelease(self):
        return bool(self._version.post)


def _parse_letter_version(letter, number):
    if letter:
        # We consider there to be an implicit 0 in a pre-release if there is
        # not a numeral associated with it.
        if number is None:
            number = 0

        # We normalize any letters to their lower case form
        letter = letter.lower()

        # We consider some words to be alternate spellings of other words and
        # in those cases we want to normalize the spellings to our preferred
        # spelling.
        if letter == "alpha":
            letter = "a"
        elif letter == "beta":
            letter = "b"
        elif letter in ["c", "pre", "preview"]:
            letter = "rc"
        elif letter in ["rev", "r"]:
            letter = "post"

        return letter, int(number)
    if not letter and number:
        # We assume if we are given a number, but we are not given a letter
        # then this is using the implicit post release syntax (e.g. 1.0-1)
        letter = "post"

        return letter, int(number)


_local_version_seperators = re.compile(r"[\._-]")


def _parse_local_version(local):
    """
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    """
    if local is not None:
        return tuple(
            part.lower() if not part.isdigit() else int(part)
            for part in _local_version_seperators.split(local)
        )


def _cmpkey(epoch, release, pre, post, dev, local):
    # When we compare a release version, we want to compare it with all of the
    # trailing zeros removed. So we'll use a reverse the list, drop all the now
    # leading zeros until we come to something non zero, then take the rest
    # re-reverse it back into the correct order and make it a tuple and use
    # that for our sorting key.
    release = tuple(
        reversed(list(
            itertools.dropwhile(
                lambda x: x == 0,
                reversed(release),
            )
        ))
    )

    # We need to "trick" the sorting algorithm to put 1.0.dev0 before 1.0a0.
    # We'll do this by abusing the pre segment, but we _only_ want to do this
    # if there is not a pre or a post segment. If we have one of those then
    # the normal sorting rules will handle this case correctly.
    if pre is None and post is None and dev is not None:
        pre = -Infinity
    # Versions without a pre-release (except as noted above) should sort after
    # those with one.
    elif pre is None:
        pre = Infinity

    # Versions without a post segment should sort before those with one.
    if post is None:
        post = -Infinity

    # Versions without a development segment should sort after those with one.
    if dev is None:
        dev = Infinity

    if local is None:
        # Versions without a local segment should sort before those with one.
        local = -Infinity
    else:
        # Versions with a local segment need that segment parsed to implement
        # the sorting rules in PEP440.
        # - Alpha numeric segments sort before numeric segments
        # - Alpha numeric segments sort lexicographically
        # - Numeric segments sort numerically
        # - Shorter versions sort before longer versions when the prefixes
        #   match exactly
        local = tuple(
            (i, "") if isinstance(i, int) else (-Infinity, i)
            for i in local
        )

    return epoch, release, pre, post, dev, local
site-packages/setuptools/_vendor/packaging/__pycache__/version.cpython-38.opt-1.pyc000064400000024607151464452710024373 0ustar00U

Qab$-	@sddlmZmZmZddlZddlZddlZddlmZddddd	gZ	e
d
ddd
dddgZddZGddde
ZGdddeZGdddeZedejZddddddZddZddZdZGd ddeZd!d"Zed#Zd$d%Zd&d'ZdS)()absolute_importdivisionprint_functionN)InfinityparseVersion
LegacyVersionInvalidVersionVERSION_PATTERN_VersionepochreleasedevprepostlocalcCs,z
t|WStk
r&t|YSXdS)z
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rr
r	)versionrH/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/version.pyrs
c@seZdZdZdS)r
zF
    An invalid version was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrrrr
$sc@sLeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
dS)_BaseVersioncCs
t|jSN)hash_keyselfrrr__hash__,sz_BaseVersion.__hash__cCs||ddS)NcSs||kSrrsorrr0z%_BaseVersion.__lt__.._comparerotherrrr__lt__/sz_BaseVersion.__lt__cCs||ddS)NcSs||kSrrr!rrrr$3r%z%_BaseVersion.__le__..r&r(rrr__le__2sz_BaseVersion.__le__cCs||ddS)NcSs||kSrrr!rrrr$6r%z%_BaseVersion.__eq__..r&r(rrr__eq__5sz_BaseVersion.__eq__cCs||ddS)NcSs||kSrrr!rrrr$9r%z%_BaseVersion.__ge__..r&r(rrr__ge__8sz_BaseVersion.__ge__cCs||ddS)NcSs||kSrrr!rrrr$<r%z%_BaseVersion.__gt__..r&r(rrr__gt__;sz_BaseVersion.__gt__cCs||ddS)NcSs||kSrrr!rrrr$?r%z%_BaseVersion.__ne__..r&r(rrr__ne__>sz_BaseVersion.__ne__cCst|tstS||j|jSr)
isinstancerNotImplementedr)rr)methodrrrr'As
z_BaseVersion._compareN)rrrr r*r+r,r-r.r/r'rrrrr*src@s`eZdZddZddZddZeddZed	d
ZeddZ	ed
dZ
eddZdS)r	cCst||_t|j|_dSr)str_version_legacy_cmpkeyr)rrrrr__init__Js
zLegacyVersion.__init__cCs|jSrr4rrrr__str__NszLegacyVersion.__str__cCsdtt|S)Nzformatreprr3rrrr__repr__QszLegacyVersion.__repr__cCs|jSrr7rrrrpublicTszLegacyVersion.publiccCs|jSrr7rrrrbase_versionXszLegacyVersion.base_versioncCsdSrrrrrrr\szLegacyVersion.localcCsdSNFrrrrr
is_prerelease`szLegacyVersion.is_prereleasecCsdSr?rrrrris_postreleasedszLegacyVersion.is_postreleaseN)rrrr6r8r<propertyr=r>rr@rArrrrr	Hs



z(\d+ | [a-z]+ | \.| -)czfinal-@)rpreview-rcrccs\t|D]F}t||}|r
|dkr(q
|dddkrF|dVq
d|Vq
dVdS)N.r
0123456789**final)_legacy_version_component_resplit_legacy_version_replacement_mapgetzfill)r"partrrr_parse_version_partsrsrScCszd}g}t|D]T}|dr^|dkrD|rD|ddkrD|q*|r^|ddkr^|qD||qt|}||fS)NrKrLz*final-Z00000000)rSlower
startswithpopappendtuple)rr
partsrRrrrr5s


r5a
    v?
    (?:
        (?:(?P[0-9]+)!)?                           # epoch
        (?P[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P
                                          # pre-release
            [-_\.]?
            (?P(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P[0-9]+)?
        )?
        (?P                                         # post release
            (?:-(?P[0-9]+))
            |
            (?:
                [-_\.]?
                (?Ppost|rev|r)
                [-_\.]?
                (?P[0-9]+)?
            )
        )?
        (?P                                          # dev release
            [-_\.]?
            (?Pdev)
            [-_\.]?
            (?P[0-9]+)?
        )?
    )
    (?:\+(?P[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@s|eZdZededejejBZddZ	ddZ
ddZed	d
Z
eddZed
dZeddZeddZdS)rz^\s*z\s*$c
Cs|j|}|std|t|dr8t|dndtdd|ddDt	|d|d	t	|d
|dp|dt	|d
|dt
|dd|_t|jj
|jj|jj|jj|jj|jj|_dS)NzInvalid version: '{0}'r
rcss|]}t|VqdSr)int.0irrr	sz#Version.__init__..rrHZpre_lZpre_nZpost_lZpost_n1Zpost_n2Zdev_lZdev_nrr
rrrrr)_regexsearchr
r:rgroupr[rYrN_parse_letter_version_parse_local_versionr4_cmpkeyr
rrrrrr)rrmatchrrrr6s8zVersion.__init__cCsdtt|S)Nzr9rrrrr<szVersion.__repr__cCsg}|jjdkr$|d|jj|ddd|jjD|jjdk	rl|ddd|jjD|jjdk	r|d|jjd	|jjdk	r|d
|jjd	|jj	dk	r|dddd|jj	Dd|S)
Nr{0}!rHcss|]}t|VqdSrr3r]xrrrr_sz"Version.__str__..css|]}t|VqdSrrirjrrrr_sz.post{0}rz.dev{0}z+{0}css|]}t|VqdSrrirjrrrr_s)
r4r
rXr:joinrrrrrrrZrrrr8szVersion.__str__cCst|dddS)N+rrr3rNrrrrr=
szVersion.publiccCsLg}|jjdkr$|d|jj|ddd|jjDd|S)NrrhrHcss|]}t|VqdSrrirjrrrr_sz'Version.base_version..rl)r4r
rXr:rmrrnrrrr>s
zVersion.base_versioncCs$t|}d|kr |dddSdS)Nrorrp)rZversion_stringrrrrsz
Version.localcCst|jjp|jjSr)boolr4rrrrrrr@!szVersion.is_prereleasecCst|jjSr)rqr4rrrrrrA%szVersion.is_postreleaseN)rrrrecompilerVERBOSE
IGNORECASErar6r<r8rBr=r>rr@rArrrrrs"

#



cCsv|rZ|dkrd}|}|dkr&d}n(|dkr4d}n|dkrBd}n|dkrNd	}|t|fS|sr|rrd	}|t|fSdS)
NrZalphaaZbetab)rCrrErG)Zrevrr)rUr[)ZletterZnumberrrrrd*s rdz[\._-]cCs$|dk	r tddt|DSdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|s|nt|VqdSr)isdigitrUr[)r]rRrrrr_Qsz'_parse_local_version..)rY_local_version_seperatorsrN)rrrrreLsrecCsttttddt|}|dkr@|dkr@|dk	r@t}n|dkrLt}|dkrZt}|dkrft}|dkrvt}ntdd|D}||||||fS)NcSs|dkS)Nrr)rkrrrr$`r%z_cmpkey..css*|]"}t|tr|dfnt|fVqdS)rlN)r0r[rr\rrrr_sz_cmpkey..)rYreversedlist	itertools	dropwhilerr`rrrrfWs,
	rf)Z
__future__rrrcollectionsr}rrZ_structuresr__all__
namedtuplerr
ValueErrorr
objectrr	rsrtrMrOrSr5rrrdrzrerfrrrrsH! k
site-packages/setuptools/_vendor/packaging/__pycache__/version.cpython-38.pyc000064400000024607151464452710023434 0ustar00U

Qab$-	@sddlmZmZmZddlZddlZddlZddlmZddddd	gZ	e
d
ddd
dddgZddZGddde
ZGdddeZGdddeZedejZddddddZddZddZdZGd ddeZd!d"Zed#Zd$d%Zd&d'ZdS)()absolute_importdivisionprint_functionN)InfinityparseVersion
LegacyVersionInvalidVersionVERSION_PATTERN_VersionepochreleasedevprepostlocalcCs,z
t|WStk
r&t|YSXdS)z
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rr
r	)versionrH/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/version.pyrs
c@seZdZdZdS)r
zF
    An invalid version was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrrrr
$sc@sLeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
dS)_BaseVersioncCs
t|jSN)hash_keyselfrrr__hash__,sz_BaseVersion.__hash__cCs||ddS)NcSs||kSrrsorrr0z%_BaseVersion.__lt__.._comparerotherrrr__lt__/sz_BaseVersion.__lt__cCs||ddS)NcSs||kSrrr!rrrr$3r%z%_BaseVersion.__le__..r&r(rrr__le__2sz_BaseVersion.__le__cCs||ddS)NcSs||kSrrr!rrrr$6r%z%_BaseVersion.__eq__..r&r(rrr__eq__5sz_BaseVersion.__eq__cCs||ddS)NcSs||kSrrr!rrrr$9r%z%_BaseVersion.__ge__..r&r(rrr__ge__8sz_BaseVersion.__ge__cCs||ddS)NcSs||kSrrr!rrrr$<r%z%_BaseVersion.__gt__..r&r(rrr__gt__;sz_BaseVersion.__gt__cCs||ddS)NcSs||kSrrr!rrrr$?r%z%_BaseVersion.__ne__..r&r(rrr__ne__>sz_BaseVersion.__ne__cCst|tstS||j|jSr)
isinstancerNotImplementedr)rr)methodrrrr'As
z_BaseVersion._compareN)rrrr r*r+r,r-r.r/r'rrrrr*src@s`eZdZddZddZddZeddZed	d
ZeddZ	ed
dZ
eddZdS)r	cCst||_t|j|_dSr)str_version_legacy_cmpkeyr)rrrrr__init__Js
zLegacyVersion.__init__cCs|jSrr4rrrr__str__NszLegacyVersion.__str__cCsdtt|S)Nzformatreprr3rrrr__repr__QszLegacyVersion.__repr__cCs|jSrr7rrrrpublicTszLegacyVersion.publiccCs|jSrr7rrrrbase_versionXszLegacyVersion.base_versioncCsdSrrrrrrr\szLegacyVersion.localcCsdSNFrrrrr
is_prerelease`szLegacyVersion.is_prereleasecCsdSr?rrrrris_postreleasedszLegacyVersion.is_postreleaseN)rrrr6r8r<propertyr=r>rr@rArrrrr	Hs



z(\d+ | [a-z]+ | \.| -)czfinal-@)rpreview-rcrccs\t|D]F}t||}|r
|dkr(q
|dddkrF|dVq
d|Vq
dVdS)N.r
0123456789**final)_legacy_version_component_resplit_legacy_version_replacement_mapgetzfill)r"partrrr_parse_version_partsrsrScCszd}g}t|D]T}|dr^|dkrD|rD|ddkrD|q*|r^|ddkr^|qD||qt|}||fS)NrKrLz*final-Z00000000)rSlower
startswithpopappendtuple)rr
partsrRrrrr5s


r5a
    v?
    (?:
        (?:(?P[0-9]+)!)?                           # epoch
        (?P[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P
                                          # pre-release
            [-_\.]?
            (?P(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P[0-9]+)?
        )?
        (?P                                         # post release
            (?:-(?P[0-9]+))
            |
            (?:
                [-_\.]?
                (?Ppost|rev|r)
                [-_\.]?
                (?P[0-9]+)?
            )
        )?
        (?P                                          # dev release
            [-_\.]?
            (?Pdev)
            [-_\.]?
            (?P[0-9]+)?
        )?
    )
    (?:\+(?P[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@s|eZdZededejejBZddZ	ddZ
ddZed	d
Z
eddZed
dZeddZeddZdS)rz^\s*z\s*$c
Cs|j|}|std|t|dr8t|dndtdd|ddDt	|d|d	t	|d
|dp|dt	|d
|dt
|dd|_t|jj
|jj|jj|jj|jj|jj|_dS)NzInvalid version: '{0}'r
rcss|]}t|VqdSr)int.0irrr	sz#Version.__init__..rrHZpre_lZpre_nZpost_lZpost_n1Zpost_n2Zdev_lZdev_nrr
rrrrr)_regexsearchr
r:rgroupr[rYrN_parse_letter_version_parse_local_versionr4_cmpkeyr
rrrrrr)rrmatchrrrr6s8zVersion.__init__cCsdtt|S)Nzr9rrrrr<szVersion.__repr__cCsg}|jjdkr$|d|jj|ddd|jjD|jjdk	rl|ddd|jjD|jjdk	r|d|jjd	|jjdk	r|d
|jjd	|jj	dk	r|dddd|jj	Dd|S)
Nr{0}!rHcss|]}t|VqdSrr3r]xrrrr_sz"Version.__str__..css|]}t|VqdSrrirjrrrr_sz.post{0}rz.dev{0}z+{0}css|]}t|VqdSrrirjrrrr_s)
r4r
rXr:joinrrrrrrrZrrrr8szVersion.__str__cCst|dddS)N+rrr3rNrrrrr=
szVersion.publiccCsLg}|jjdkr$|d|jj|ddd|jjDd|S)NrrhrHcss|]}t|VqdSrrirjrrrr_sz'Version.base_version..rl)r4r
rXr:rmrrnrrrr>s
zVersion.base_versioncCs$t|}d|kr |dddSdS)Nrorrp)rZversion_stringrrrrsz
Version.localcCst|jjp|jjSr)boolr4rrrrrrr@!szVersion.is_prereleasecCst|jjSr)rqr4rrrrrrA%szVersion.is_postreleaseN)rrrrecompilerVERBOSE
IGNORECASErar6r<r8rBr=r>rr@rArrrrrs"

#



cCsv|rZ|dkrd}|}|dkr&d}n(|dkr4d}n|dkrBd}n|dkrNd	}|t|fS|sr|rrd	}|t|fSdS)
NrZalphaaZbetab)rCrrErG)Zrevrr)rUr[)ZletterZnumberrrrrd*s rdz[\._-]cCs$|dk	r tddt|DSdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|s|nt|VqdSr)isdigitrUr[)r]rRrrrr_Qsz'_parse_local_version..)rY_local_version_seperatorsrN)rrrrreLsrecCsttttddt|}|dkr@|dkr@|dk	r@t}n|dkrLt}|dkrZt}|dkrft}|dkrvt}ntdd|D}||||||fS)NcSs|dkS)Nrr)rkrrrr$`r%z_cmpkey..css*|]"}t|tr|dfnt|fVqdS)rlN)r0r[rr\rrrr_sz_cmpkey..)rYreversedlist	itertools	dropwhilerr`rrrrfWs,
	rf)Z
__future__rrrcollectionsr}rrZ_structuresr__all__
namedtuplerr
ValueErrorr
objectrr	rsrtrMrOrSr5rrrdrzrerfrrrrsH! k
site-packages/setuptools/_vendor/packaging/__pycache__/requirements.cpython-38.opt-1.pyc000064400000007435151464452710025431 0ustar00U

Qab@srddlmZmZmZddlZddlZddlmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZddlmZddlmZddlmZmZdd	lmZmZmZGd
ddeZeejejZ ed!Z"ed
!Z#ed!Z$ed!Z%ed!Z&ed!Z'ed!Z(edZ)e ee)e BZ*ee ee*Z+e+dZ,e+Z-eddZ.e(e.Z/e-ee&e-Z0e"e
e0e#dZ1eej2ej3ej4BZ5eej2ej3ej4BZ6e5e6AZ7ee7ee&e7ddddZ8e
e$e8e%e8BZ9e9:dde	e9dZ;e;:dde	edZe:d de'Ze/e
e=Z?e,e
e1e?e>BZ@ee@eZAGd!d"d"eBZCdS)#)absolute_importdivisionprint_functionN)stringStart	stringEndoriginalTextForParseException)
ZeroOrMoreWordOptionalRegexCombine)Literal)parse)MARKER_EXPRMarker)LegacySpecifier	SpecifierSpecifierSetc@seZdZdZdS)InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)__name__
__module____qualname____doc__rrM/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/requirements.pyrsr[](),;@z-_.namez[^ ]+urlextrasF)Z
joinStringZadjacent	_raw_speccCs
|jpdS)N)r'sltrrr6r-	specifiercCs|dS)Nrrr)rrrr-9r.markercCst||j|jS)N)rZ_original_startZ
_original_endr)rrrr-=r.c@s(eZdZdZddZddZddZdS)	RequirementzParse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    c
Cszt|}Wn@tk
rN}z"td||j|jdW5d}~XYnX|j|_|jrt|j}|j	r|j
r|j	s|j
std|j|_nd|_t|jr|j
ng|_t|j|_|jr|jnd|_dS)Nz+Invalid requirement, parse error at "{0!r}"zInvalid URL given)REQUIREMENTZparseStringrrformatZlocr$r%urlparseZschemeZnetlocsetr&ZasListrr/r0)selfZrequirement_stringZreqeZ
parsed_urlrrr__init__Xs,
zRequirement.__init__cCsz|jg}|jr*|ddt|j|jr@|t|j|jrX|d|j|j	rp|d|j	d|S)Nz[{0}]r!z@ {0}z; {0}r()
r$r&appendr4joinsortedr/strr%r0)r7partsrrr__str__mszRequirement.__str__cCsdt|S)Nz)r4r=)r7rrr__repr__~szRequirement.__repr__N)rrrrr9r?r@rrrrr1Ksr1)DZ
__future__rrrstringreZsetuptools.extern.pyparsingrrrrr	r
rrr
rLZ"setuptools.extern.six.moves.urllibrr5ZmarkersrrZ
specifiersrrr
ValueErrorrZ
ascii_lettersZdigitsZALPHANUMsuppressZLBRACKETZRBRACKETZLPARENZRPARENCOMMAZ	SEMICOLONATZPUNCTUATIONZIDENTIFIER_ENDZ
IDENTIFIERNAMEZEXTRAZURIZURLZEXTRAS_LISTZEXTRASZ
_regex_strVERBOSE
IGNORECASEZVERSION_PEP440ZVERSION_LEGACYZVERSION_ONEZVERSION_MANYZ
_VERSION_SPECZsetParseActionZVERSION_SPECZMARKER_SEPERATORZMARKERZVERSION_AND_MARKERZURL_AND_MARKERZNAMED_REQUIREMENTr3objectr1rrrrsfsite-packages/setuptools/_vendor/packaging/__pycache__/__init__.cpython-38.pyc000064400000001036151464452710023475 0ustar00U

Qab@sTddlmZmZmZddlmZmZmZmZm	Z	m
Z
mZmZdddddd	d
dgZ
dS)
)absolute_importdivisionprint_function)
__author__
__copyright__	__email____license____summary__	__title____uri____version__rr
rr
rrr	rN)Z
__future__rrr	__about__rrrr	r
rrr
__all__rrI/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/__init__.pys(site-packages/setuptools/_vendor/packaging/__pycache__/_compat.cpython-38.opt-1.pyc000064400000001722151464452710024321 0ustar00U

Qab\@sVddlmZmZmZddlZejddkZejddkZerDefZ	ne
fZ	ddZdS))absolute_importdivisionprint_functionNcs&Gfddd}t|ddiS)z/
    Create a base class with a metaclass.
    cseZdZfddZdS)z!with_metaclass..metaclasscs||S)N)clsnameZ
this_basesdbasesmetarH/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/_compat.py__new__sz)with_metaclass..metaclass.__new__N)__name__
__module____qualname__rrrrr	metaclasssrZtemporary_classr)typer)r
rrrrrwith_metaclasssr)Z
__future__rrrsysversion_infoZPY2ZPY3strZstring_typesZ
basestringrrrrrssite-packages/setuptools/_vendor/packaging/__pycache__/__about__.cpython-38.pyc000064400000001300151464452710023636 0ustar00U

Qab@sPddlmZmZmZdddddddd	gZd
ZdZdZd
ZdZ	dZ
dZde	ZdS))absolute_importdivisionprint_function	__title____summary____uri____version__
__author__	__email____license__
__copyright__Z	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz16.8z)Donald Stufft and individual contributorszdonald@stufft.ioz"BSD or Apache License, Version 2.0zCopyright 2014-2016 %sN)
Z
__future__rrr__all__rrrrr	r
rrrrJ/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/__about__.pys"site-packages/setuptools/_vendor/packaging/__pycache__/_compat.cpython-38.pyc000064400000001722151464452710023362 0ustar00U

Qab\@sVddlmZmZmZddlZejddkZejddkZerDefZ	ne
fZ	ddZdS))absolute_importdivisionprint_functionNcs&Gfddd}t|ddiS)z/
    Create a base class with a metaclass.
    cseZdZfddZdS)z!with_metaclass..metaclasscs||S)N)clsnameZ
this_basesdbasesmetarH/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/_compat.py__new__sz)with_metaclass..metaclass.__new__N)__name__
__module____qualname__rrrrr	metaclasssrZtemporary_classr)typer)r
rrrrrwith_metaclasssr)Z
__future__rrrsysversion_infoZPY2ZPY3strZstring_typesZ
basestringrrrrrssite-packages/setuptools/_vendor/packaging/__pycache__/__about__.cpython-38.opt-1.pyc000064400000001300151464452710024575 0ustar00U

Qab@sPddlmZmZmZdddddddd	gZd
ZdZdZd
ZdZ	dZ
dZde	ZdS))absolute_importdivisionprint_function	__title____summary____uri____version__
__author__	__email____license__
__copyright__Z	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz16.8z)Donald Stufft and individual contributorszdonald@stufft.ioz"BSD or Apache License, Version 2.0zCopyright 2014-2016 %sN)
Z
__future__rrr__all__rrrrr	r
rrrrJ/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/__about__.pys"site-packages/setuptools/_vendor/packaging/__pycache__/_structures.cpython-38.pyc000064400000005310151464452710024317 0ustar00U

Qab@sDddlmZmZmZGdddeZeZGdddeZeZdS))absolute_importdivisionprint_functionc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)InfinitycCsdS)NrselfrrL/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/_structures.py__repr__	szInfinity.__repr__cCstt|SNhashreprrrrr	__hash__szInfinity.__hash__cCsdSNFrrotherrrr	__lt__szInfinity.__lt__cCsdSrrrrrr	__le__szInfinity.__le__cCst||jSr
isinstance	__class__rrrr	__eq__szInfinity.__eq__cCst||jSrrrrrr	__ne__szInfinity.__ne__cCsdSNTrrrrr	__gt__szInfinity.__gt__cCsdSrrrrrr	__ge__szInfinity.__ge__cCstSr)NegativeInfinityrrrr	__neg__!szInfinity.__neg__N__name__
__module____qualname__r
rrrrrrrrrrrr	rsrc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)rcCsdS)Nz	-Infinityrrrrr	r
)szNegativeInfinity.__repr__cCstt|Srrrrrr	r,szNegativeInfinity.__hash__cCsdSrrrrrr	r/szNegativeInfinity.__lt__cCsdSrrrrrr	r2szNegativeInfinity.__le__cCst||jSrrrrrr	r5szNegativeInfinity.__eq__cCst||jSrrrrrr	r8szNegativeInfinity.__ne__cCsdSrrrrrr	r;szNegativeInfinity.__gt__cCsdSrrrrrr	r>szNegativeInfinity.__ge__cCstSr)rrrrr	rAszNegativeInfinity.__neg__Nrrrrr	r'srN)Z
__future__rrrobjectrrrrrr	ssite-packages/setuptools/_vendor/packaging/__pycache__/requirements.cpython-38.pyc000064400000007435151464452710024472 0ustar00U

Qab@srddlmZmZmZddlZddlZddlmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZddlmZddlmZddlmZmZdd	lmZmZmZGd
ddeZeejejZ ed!Z"ed
!Z#ed!Z$ed!Z%ed!Z&ed!Z'ed!Z(edZ)e ee)e BZ*ee ee*Z+e+dZ,e+Z-eddZ.e(e.Z/e-ee&e-Z0e"e
e0e#dZ1eej2ej3ej4BZ5eej2ej3ej4BZ6e5e6AZ7ee7ee&e7ddddZ8e
e$e8e%e8BZ9e9:dde	e9dZ;e;:dde	edZe:d de'Ze/e
e=Z?e,e
e1e?e>BZ@ee@eZAGd!d"d"eBZCdS)#)absolute_importdivisionprint_functionN)stringStart	stringEndoriginalTextForParseException)
ZeroOrMoreWordOptionalRegexCombine)Literal)parse)MARKER_EXPRMarker)LegacySpecifier	SpecifierSpecifierSetc@seZdZdZdS)InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)__name__
__module____qualname____doc__rrM/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/requirements.pyrsr[](),;@z-_.namez[^ ]+urlextrasF)Z
joinStringZadjacent	_raw_speccCs
|jpdS)N)r'sltrrr6r-	specifiercCs|dS)Nrrr)rrrr-9r.markercCst||j|jS)N)rZ_original_startZ
_original_endr)rrrr-=r.c@s(eZdZdZddZddZddZdS)	RequirementzParse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    c
Cszt|}Wn@tk
rN}z"td||j|jdW5d}~XYnX|j|_|jrt|j}|j	r|j
r|j	s|j
std|j|_nd|_t|jr|j
ng|_t|j|_|jr|jnd|_dS)Nz+Invalid requirement, parse error at "{0!r}"zInvalid URL given)REQUIREMENTZparseStringrrformatZlocr$r%urlparseZschemeZnetlocsetr&ZasListrr/r0)selfZrequirement_stringZreqeZ
parsed_urlrrr__init__Xs,
zRequirement.__init__cCsz|jg}|jr*|ddt|j|jr@|t|j|jrX|d|j|j	rp|d|j	d|S)Nz[{0}]r!z@ {0}z; {0}r()
r$r&appendr4joinsortedr/strr%r0)r7partsrrr__str__mszRequirement.__str__cCsdt|S)Nz)r4r=)r7rrr__repr__~szRequirement.__repr__N)rrrrr9r?r@rrrrr1Ksr1)DZ
__future__rrrstringreZsetuptools.extern.pyparsingrrrrr	r
rrr
rLZ"setuptools.extern.six.moves.urllibrr5ZmarkersrrZ
specifiersrrr
ValueErrorrZ
ascii_lettersZdigitsZALPHANUMsuppressZLBRACKETZRBRACKETZLPARENZRPARENCOMMAZ	SEMICOLONATZPUNCTUATIONZIDENTIFIER_ENDZ
IDENTIFIERNAMEZEXTRAZURIZURLZEXTRAS_LISTZEXTRASZ
_regex_strVERBOSE
IGNORECASEZVERSION_PEP440ZVERSION_LEGACYZVERSION_ONEZVERSION_MANYZ
_VERSION_SPECZsetParseActionZVERSION_SPECZMARKER_SEPERATORZMARKERZVERSION_AND_MARKERZURL_AND_MARKERZNAMED_REQUIREMENTr3objectr1rrrrsfsite-packages/setuptools/_vendor/packaging/__pycache__/_structures.cpython-38.opt-1.pyc000064400000005310151464452710025256 0ustar00U

Qab@sDddlmZmZmZGdddeZeZGdddeZeZdS))absolute_importdivisionprint_functionc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)InfinitycCsdS)NrselfrrL/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/_structures.py__repr__	szInfinity.__repr__cCstt|SNhashreprrrrr	__hash__szInfinity.__hash__cCsdSNFrrotherrrr	__lt__szInfinity.__lt__cCsdSrrrrrr	__le__szInfinity.__le__cCst||jSr
isinstance	__class__rrrr	__eq__szInfinity.__eq__cCst||jSrrrrrr	__ne__szInfinity.__ne__cCsdSNTrrrrr	__gt__szInfinity.__gt__cCsdSrrrrrr	__ge__szInfinity.__ge__cCstSr)NegativeInfinityrrrr	__neg__!szInfinity.__neg__N__name__
__module____qualname__r
rrrrrrrrrrrr	rsrc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)rcCsdS)Nz	-Infinityrrrrr	r
)szNegativeInfinity.__repr__cCstt|Srrrrrr	r,szNegativeInfinity.__hash__cCsdSrrrrrr	r/szNegativeInfinity.__lt__cCsdSrrrrrr	r2szNegativeInfinity.__le__cCst||jSrrrrrr	r5szNegativeInfinity.__eq__cCst||jSrrrrrr	r8szNegativeInfinity.__ne__cCsdSrrrrrr	r;szNegativeInfinity.__gt__cCsdSrrrrrr	r>szNegativeInfinity.__ge__cCstSr)rrrrr	rAszNegativeInfinity.__neg__Nrrrrr	r'srN)Z
__future__rrrobjectrrrrrr	ssite-packages/setuptools/_vendor/packaging/__pycache__/utils.cpython-38.pyc000064400000000717151464452710023103 0ustar00U

Qab@s2ddlmZmZmZddlZedZddZdS))absolute_importdivisionprint_functionNz[-_.]+cCstd|S)N-)_canonicalize_regexsublower)namer
F/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/utils.pycanonicalize_namesr)Z
__future__rrrrecompilerrr
r
r
rs
site-packages/setuptools/_vendor/packaging/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001036151464452710024434 0ustar00U

Qab@sTddlmZmZmZddlmZmZmZmZm	Z	m
Z
mZmZdddddd	d
dgZ
dS)
)absolute_importdivisionprint_function)
__author__
__copyright__	__email____license____summary__	__title____uri____version__rr
rr
rrr	rN)Z
__future__rrr	__about__rrrr	r
rrr
__all__rrI/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/__init__.pys(site-packages/setuptools/_vendor/packaging/__pycache__/markers.cpython-38.opt-1.pyc000064400000021154151464452710024344 0ustar00U

Qab/ 	@s@ddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
mZmZddlm
Z
mZmZmZddlmZddlmZddlmZmZd	d
ddd
gZGdd	d	eZGdd
d
eZGdddeZGdddeZGdddeZGdddeZ GdddeZ!ededBedBedBedBedBedBed Bed!Bed"Bed#Bed$Bed%Bed&Bed'Bed(Bed)Bed*BZ"d#d"ddddd+Z#e"$d,d-ed.ed/Bed0Bed1Bed2Bed3Bed4Bed5BZ%e%ed6Bed7BZ&e&$d8d-ed9ed:BZ'e'$d;d-ed<ed=BZ(e"e'BZ)ee)e&e)Z*e*$d>d-ed?+Z,ed@+Z-eZ.e*ee,e.e-BZ/e.e/e
e(e.>ee.eZ0dAdBZ1dSdDdEZ2dFd-dGd-ej3ej4ej5ej6ej7ej8dHZ9dIdJZ:eZ;dKdLZdQd
Z?GdRddeZ@dS)T)absolute_importdivisionprint_functionN)ParseExceptionParseResultsstringStart	stringEnd)
ZeroOrMoreGroupForwardQuotedString)Literal)string_types)	SpecifierInvalidSpecifier
InvalidMarkerUndefinedComparisonUndefinedEnvironmentNameMarkerdefault_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N__name__
__module____qualname____doc__rrH/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/markers.pyrsc@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrrrrrrsc@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrrrrrr%sc@s,eZdZddZddZddZddZd	S)
NodecCs
||_dSN)value)selfr rrr__init__.sz
Node.__init__cCs
t|jSr)strr r!rrr__str__1szNode.__str__cCsd|jjt|S)Nz<{0}({1!r})>)format	__class__rr#r$rrr__repr__4sz
Node.__repr__cCstdSr)NotImplementedErrorr$rrr	serialize7szNode.serializeN)rrrr"r%r(r*rrrrr,src@seZdZddZdS)VariablecCst|Srr#r$rrrr*=szVariable.serializeNrrrr*rrrrr+;sr+c@seZdZddZdS)ValuecCs
d|S)Nz"{0}")r&r$rrrr*CszValue.serializeNr-rrrrr.Asr.c@seZdZddZdS)OpcCst|Srr,r$rrrr*IszOp.serializeNr-rrrrr/Gsr/implementation_versionplatform_python_implementationimplementation_namepython_full_versionplatform_releaseplatform_versionplatform_machineplatform_systempython_versionsys_platformos_nameos.namesys.platformplatform.versionplatform.machineplatform.python_implementationpython_implementationZextra)r;r<r=r>r?r@cCstt|d|dSNr)r+ALIASESgetsltrrrirHz=====>=<=!=z~=><not inincCst|dSrA)r/rDrrrrHwrI'"cCst|dSrA)r.rDrrrrHzrIandorcCst|dSrA)tuplerDrrrrHrI()cCs t|trdd|DS|SdS)NcSsg|]}t|qSr)_coerce_parse_result).0irrr
sz(_coerce_parse_result..)
isinstancer)resultsrrrrYs
rYTcCst|tr4t|dkr4t|dttfr4t|dSt|trndd|D}|rZd|Sdd|dSn"t|trddd	|DS|SdS)
Nrrcss|]}t|ddVqdS)F)firstN)_format_markerrZmrrr	sz!_format_marker.. rWrXcSsg|]}|qSr)r*rarrrr\sz"_format_marker..)r]listlenrVr`join)markerr_innerrrrr`s


r`cCs||kSrrlhsrhsrrrrHrIcCs||kSrrrjrrrrHrI)rQrPrOrLrJrMrKrNcCslztd||g}Wntk
r.YnX||St|}|dkrbtd||||||S)Nz#Undefined {0!r} on {1!r} and {2!r}.)	rrgr*rcontains
_operatorsrCrr&)rkoprlspecZoperrrr_eval_ops
rrcCs&||t}|tkr"td||S)Nz/{0!r} does not exist in evaluation environment.)rC
_undefinedrr&)environmentnamer rrr_get_envsrvc	Csgg}|D]}t|tr.|dt||q
t|tr|\}}}t|tr`t||j}|j}n|j}t||j}|dt|||q
|dkr
|gq
t	dd|DS)NrUcss|]}t|VqdSr)all)rZitemrrrrcsz$_evaluate_markers..)
r]reappend_evaluate_markersrVr+rvr rrany)	ZmarkersrtgroupsrhrkrprlZ	lhs_valueZ	rhs_valuerrrr{s



r{cCs2d|}|j}|dkr.||dt|j7}|S)Nz{0.major}.{0.minor}.{0.micro}finalr)r&releaselevelr#serial)infoversionZkindrrrformat_full_versions

rcCslttdr ttjj}tjj}nd}d}||tjtt	t
ttttddtjdS)Nimplementation0rm)r2r0r:r6r4r7r5r3r1r8r9)
hasattrsysrrrruosplatformmachinereleasesystemr8r@)Ziverr2rrrrs"

c@s.eZdZddZddZddZd
dd	ZdS)rc
Cs`ztt||_WnFtk
rZ}z(d|||j|jd}t|W5d}~XYnXdS)Nz+Invalid marker: {0!r}, parse error at {1!r})rYMARKERZparseString_markersrr&Zlocr)r!rheZerr_strrrrr"szMarker.__init__cCs
t|jSr)r`rr$rrrr%szMarker.__str__cCsdt|S)Nz)r&r#r$rrrr(szMarker.__repr__NcCs$t}|dk	r||t|j|S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)rupdater{r)r!rtZcurrent_environmentrrrevaluate s	
zMarker.evaluate)N)rrrr"r%r(rrrrrrs)T)AZ
__future__rrroperatorrrrZsetuptools.extern.pyparsingrrrrr	r
rrr
LZ_compatrZ
specifiersrr__all__
ValueErrorrrrobjectrr+r.r/ZVARIABLErBZsetParseActionZVERSION_CMPZ	MARKER_OPZMARKER_VALUEZBOOLOPZ
MARKER_VARZMARKER_ITEMsuppressZLPARENZRPARENZMARKER_EXPRZMARKER_ATOMrrYr`ltleeqnegegtrorrrsrvr{rrrrrrrs	


site-packages/setuptools/_vendor/packaging/__pycache__/specifiers.cpython-38.opt-1.pyc000064400000046510151464452710025037 0ustar00U

Qabym@sddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZm
Z
mZGdddeZGdd	d	e
ejeZGd
ddeZGdd
d
eZddZGdddeZedZddZddZGdddeZdS))absolute_importdivisionprint_functionN)string_typeswith_metaclass)Version
LegacyVersionparsec@seZdZdZdS)InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrK/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/specifiers.pyrsrc@seZdZejddZejddZejddZejddZej	d	d
Z
e
jdd
Z
ejdd
dZejdddZ
dS)
BaseSpecifiercCsdS)z
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nrselfrrr__str__szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr__hash__szBaseSpecifier.__hash__cCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nrrotherrrr__eq__$szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrrrrr__ne__+szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrrprereleases2szBaseSpecifier.prereleasescCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrvaluerrrr9sNcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nrritemrrrrcontains@szBaseSpecifier.containscCsdS)z
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)riterablerrrrfilterFszBaseSpecifier.filter)N)N)rr
rabcabstractmethodrrrrabstractpropertyrsetterr r"rrrrrs 





rc@seZdZiZd ddZddZddZd	d
ZddZd
dZ	ddZ
ddZeddZ
eddZeddZejddZddZd!ddZd"ddZdS)#_IndividualSpecifierNcCsF|j|}|std||d|df|_||_dS)NzInvalid specifier: '{0}'operatorversion)_regexsearchrformatgroupstrip_spec_prereleases)rspecrmatchrrr__init__Rsz_IndividualSpecifier.__init__cCs0|jdk	rd|jnd}d|jjt||S)N, prereleases={0!r}r(z<{0}({1!r}{2})>)r1r-r	__class__rstrrZprerrr__repr___sz_IndividualSpecifier.__repr__cCsdj|jS)Nz{0}{1})r-r0rrrrrlsz_IndividualSpecifier.__str__cCs
t|jSN)hashr0rrrrrosz_IndividualSpecifier.__hash__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:
isinstancerr6rNotImplementedr0rrrrrrs
z_IndividualSpecifier.__eq__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:r<rrrrr}s
z_IndividualSpecifier.__ne__cCst|d|j|S)Nz_compare_{0})getattrr-
_operators)roprrr
_get_operatorsz"_IndividualSpecifier._get_operatorcCst|ttfst|}|Sr:)r=r	rr
rr*rrr_coerce_versionsz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nrr0rrrrr)sz_IndividualSpecifier.operatorcCs
|jdS)NrrErrrrr*sz_IndividualSpecifier.versioncCs|jSr:r1rrrrrsz _IndividualSpecifier.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:r rrrrr__contains__sz!_IndividualSpecifier.__contains__cCs:|dkr|j}||}|jr&|s&dS||j||jSNF)rrD
is_prereleaserBr)r*rrrrr s

z_IndividualSpecifier.containsccsd}g}d|dk	r|ndi}|D]B}||}|j|f|r |jrX|sX|jsX||q d}|Vq |s||r||D]
}|VqpdS)NFrT)rDr rKrappend)rr!rZyieldedfound_prereleaseskwr*parsed_versionrrrr"s"
z_IndividualSpecifier.filter)r(N)N)N)rr
rr@r4r9rrrrrBrDpropertyr)r*rr&rIr r"rrrrr'Ns(







r'c@sveZdZdZededejejBZdddddd	d
Z	ddZ
d
dZddZddZ
ddZddZddZdS)LegacySpecifiera
        (?P(==|!=|<=|>=|<|>))
        \s*
        (?P
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        ^\s*\s*$equal	not_equalless_than_equalgreater_than_equal	less_thangreater_than)==!=<=>=<>cCst|tstt|}|Sr:)r=r	r7rCrrrrDs
zLegacySpecifier._coerce_versioncCs|||kSr:rDrprospectiver2rrr_compare_equalszLegacySpecifier._compare_equalcCs|||kSr:r`rarrr_compare_not_equalsz"LegacySpecifier._compare_not_equalcCs|||kSr:r`rarrr_compare_less_than_equalsz(LegacySpecifier._compare_less_than_equalcCs|||kSr:r`rarrr_compare_greater_than_equalsz+LegacySpecifier._compare_greater_than_equalcCs|||kSr:r`rarrr_compare_less_thansz"LegacySpecifier._compare_less_thancCs|||kSr:r`rarrr_compare_greater_thansz%LegacySpecifier._compare_greater_thanN)rr
r
_regex_strrecompileVERBOSE
IGNORECASEr+r@rDrcrdrerfrgrhrrrrrQs(

	rQcstfdd}|S)Ncst|tsdS|||SrJ)r=rrafnrrwrappeds
z)_require_version_compare..wrapped)	functoolswraps)rorprrnr_require_version_compare
srsc	@seZdZdZededejejBZdddddd	d
ddZ	e
d
dZe
ddZe
ddZ
e
ddZe
ddZe
ddZe
ddZddZeddZejddZd S)!	Specifiera
        (?P(~=|==|!=|<=|>=|<|>|===))
        (?P
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?sz/Specifier._compare_compatible...*r]rZ)joinlist	itertools	takewhile_version_splitrB)rrbr2prefixrrr_compare_compatibleszSpecifier._compare_compatiblecCsp|drPt|j}t|dd}tt|}|dt|}t||\}}nt|}|jsht|j}||kS)Nr})endswithrZpublicrr7len_pad_versionlocalrarrrrcs


zSpecifier._compare_equalcCs|||Sr:)rcrarrrrdszSpecifier._compare_not_equalcCs|t|kSr:rrarrrresz"Specifier._compare_less_than_equalcCs|t|kSr:rrarrrrfsz%Specifier._compare_greater_than_equalcCs<t|}||ksdS|js8|jr8t|jt|jkr8dSdSNFT)rrKbase_versionrarrrrgszSpecifier._compare_less_thancCs^t|}||ksdS|js8|jr8t|jt|jkr8dS|jdk	rZt|jt|jkrZdSdSr)rZis_postreleaserrrarrrrhs
zSpecifier._compare_greater_thancCst|t|kSr:)r7lowerrarrr_compare_arbitraryszSpecifier._compare_arbitrarycCsR|jdk	r|jS|j\}}|dkrN|dkr@|dr@|dd}t|jrNdSdS)N)rZr]r\rurvrZr}rTF)r1r0rr
rK)rr)r*rrrrs


zSpecifier.prereleasescCs
||_dSr:rFrrrrrsN)rr
rrirjrkrlrmr+r@rsrrcrdrerfrgrhrrPrr&rrrrrtsD_



"





rtz^([0-9]+)((?:a|b|c|rc)[0-9]+)$cCs@g}|dD],}t|}|r0||q||q|S)Nrw)split
_prefix_regexr,extendgroupsrL)r*resultrr3rrrr's
rc
Csgg}}|ttdd||ttdd|||t|dd||t|dd|ddgtdt|dt|d|ddgtdt|dt|dttj|ttj|fS)NcSs|Sr:isdigitryrrrr{6z_pad_version..cSs|Sr:rryrrrr{7rrr0)rLrrrrinsertmaxchain)leftrightZ
left_splitZright_splitrrrr2s 
""rc@seZdZdddZddZddZd	d
ZddZd
dZddZ	ddZ
ddZeddZ
e
jddZ
ddZdddZd ddZdS)!SpecifierSetr(Nc	Csndd|dD}t}|D]:}z|t|Wqtk
rV|t|YqXqt||_||_dS)NcSsg|]}|r|qSr)r/.0srrr
Rsz)SpecifierSet.__init__..,)	rsetaddrtrrQ	frozenset_specsr1)rZ
specifiersrZparsed	specifierrrrr4Os
zSpecifierSet.__init__cCs*|jdk	rd|jnd}dt||S)Nr5r(z)r1r-rr7r8rrrr9ds
zSpecifierSet.__repr__cCsdtdd|jDS)Nrcss|]}t|VqdSr:)r7rrrr	nsz'SpecifierSet.__str__..)r~sortedrrrrrrmszSpecifierSet.__str__cCs
t|jSr:)r;rrrrrrpszSpecifierSet.__hash__cCst|trt|}nt|ts"tSt}t|j|jB|_|jdkrX|jdk	rX|j|_n<|jdk	rv|jdkrv|j|_n|j|jkr|j|_ntd|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)r=rrr>rrr1
ValueError)rrrrrr__and__ss 





zSpecifierSet.__and__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:r=rrr'r7r>rrrrrrs



zSpecifierSet.__eq__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:rrrrrrs



zSpecifierSet.__ne__cCs
t|jSr:)rrrrrr__len__szSpecifierSet.__len__cCs
t|jSr:)iterrrrrr__iter__szSpecifierSet.__iter__cCs.|jdk	r|jS|jsdStdd|jDS)Ncss|]}|jVqdSr:rrrrrrsz+SpecifierSet.prereleases..)r1ranyrrrrrs

zSpecifierSet.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:rGrHrrrrIszSpecifierSet.__contains__csLtttfstdkr$|js2jr2dStfdd|jDS)NFc3s|]}|jdVqdS)rNrGrrrrrrsz(SpecifierSet.contains..)r=r	rr
rrKallrrrrrr s
zSpecifierSet.containscCs|dkr|j}|jr6|jD]}|j|t|d}q|Sg}g}|D]P}t|ttfs^t|}n|}t|trnqB|jr|s|s|	|qB|	|qB|s|r|dkr|S|SdS)Nr)
rrr"boolr=r	rr
rKrL)rr!rr2ZfilteredrMrrOrrrr"s*



zSpecifierSet.filter)r(N)N)N)rr
rr4r9rrrrrrrrPrr&rIr r"rrrrrMs 
	




r)Z
__future__rrrr#rqrrjZ_compatrrr*rr	r
rrABCMetaobjectrr'rQrsrtrkrrrrrrrrs&9	4	
site-packages/setuptools/_vendor/packaging/__pycache__/utils.cpython-38.opt-1.pyc000064400000000717151464452710024042 0ustar00U

Qab@s2ddlmZmZmZddlZedZddZdS))absolute_importdivisionprint_functionNz[-_.]+cCstd|S)N-)_canonicalize_regexsublower)namer
F/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/utils.pycanonicalize_namesr)Z
__future__rrrrecompilerrr
r
r
rs
site-packages/setuptools/_vendor/packaging/__pycache__/specifiers.cpython-38.pyc000064400000046510151464452710024100 0ustar00U

Qabym@sddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZm
Z
mZGdddeZGdd	d	e
ejeZGd
ddeZGdd
d
eZddZGdddeZedZddZddZGdddeZdS))absolute_importdivisionprint_functionN)string_typeswith_metaclass)Version
LegacyVersionparsec@seZdZdZdS)InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrK/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/specifiers.pyrsrc@seZdZejddZejddZejddZejddZej	d	d
Z
e
jdd
Z
ejdd
dZejdddZ
dS)
BaseSpecifiercCsdS)z
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nrselfrrr__str__szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr__hash__szBaseSpecifier.__hash__cCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nrrotherrrr__eq__$szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrrrrr__ne__+szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrrprereleases2szBaseSpecifier.prereleasescCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrvaluerrrr9sNcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nrritemrrrrcontains@szBaseSpecifier.containscCsdS)z
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)riterablerrrrfilterFszBaseSpecifier.filter)N)N)rr
rabcabstractmethodrrrrabstractpropertyrsetterr r"rrrrrs 





rc@seZdZiZd ddZddZddZd	d
ZddZd
dZ	ddZ
ddZeddZ
eddZeddZejddZddZd!ddZd"ddZdS)#_IndividualSpecifierNcCsF|j|}|std||d|df|_||_dS)NzInvalid specifier: '{0}'operatorversion)_regexsearchrformatgroupstrip_spec_prereleases)rspecrmatchrrr__init__Rsz_IndividualSpecifier.__init__cCs0|jdk	rd|jnd}d|jjt||S)N, prereleases={0!r}r(z<{0}({1!r}{2})>)r1r-r	__class__rstrrZprerrr__repr___sz_IndividualSpecifier.__repr__cCsdj|jS)Nz{0}{1})r-r0rrrrrlsz_IndividualSpecifier.__str__cCs
t|jSN)hashr0rrrrrosz_IndividualSpecifier.__hash__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:
isinstancerr6rNotImplementedr0rrrrrrs
z_IndividualSpecifier.__eq__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:r<rrrrr}s
z_IndividualSpecifier.__ne__cCst|d|j|S)Nz_compare_{0})getattrr-
_operators)roprrr
_get_operatorsz"_IndividualSpecifier._get_operatorcCst|ttfst|}|Sr:)r=r	rr
rr*rrr_coerce_versionsz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nrr0rrrrr)sz_IndividualSpecifier.operatorcCs
|jdS)NrrErrrrr*sz_IndividualSpecifier.versioncCs|jSr:r1rrrrrsz _IndividualSpecifier.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:r rrrrr__contains__sz!_IndividualSpecifier.__contains__cCs:|dkr|j}||}|jr&|s&dS||j||jSNF)rrD
is_prereleaserBr)r*rrrrr s

z_IndividualSpecifier.containsccsd}g}d|dk	r|ndi}|D]B}||}|j|f|r |jrX|sX|jsX||q d}|Vq |s||r||D]
}|VqpdS)NFrT)rDr rKrappend)rr!rZyieldedfound_prereleaseskwr*parsed_versionrrrr"s"
z_IndividualSpecifier.filter)r(N)N)N)rr
rr@r4r9rrrrrBrDpropertyr)r*rr&rIr r"rrrrr'Ns(







r'c@sveZdZdZededejejBZdddddd	d
Z	ddZ
d
dZddZddZ
ddZddZddZdS)LegacySpecifiera
        (?P(==|!=|<=|>=|<|>))
        \s*
        (?P
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        ^\s*\s*$equal	not_equalless_than_equalgreater_than_equal	less_thangreater_than)==!=<=>=<>cCst|tstt|}|Sr:)r=r	r7rCrrrrDs
zLegacySpecifier._coerce_versioncCs|||kSr:rDrprospectiver2rrr_compare_equalszLegacySpecifier._compare_equalcCs|||kSr:r`rarrr_compare_not_equalsz"LegacySpecifier._compare_not_equalcCs|||kSr:r`rarrr_compare_less_than_equalsz(LegacySpecifier._compare_less_than_equalcCs|||kSr:r`rarrr_compare_greater_than_equalsz+LegacySpecifier._compare_greater_than_equalcCs|||kSr:r`rarrr_compare_less_thansz"LegacySpecifier._compare_less_thancCs|||kSr:r`rarrr_compare_greater_thansz%LegacySpecifier._compare_greater_thanN)rr
r
_regex_strrecompileVERBOSE
IGNORECASEr+r@rDrcrdrerfrgrhrrrrrQs(

	rQcstfdd}|S)Ncst|tsdS|||SrJ)r=rrafnrrwrappeds
z)_require_version_compare..wrapped)	functoolswraps)rorprrnr_require_version_compare
srsc	@seZdZdZededejejBZdddddd	d
ddZ	e
d
dZe
ddZe
ddZ
e
ddZe
ddZe
ddZe
ddZddZeddZejddZd S)!	Specifiera
        (?P(~=|==|!=|<=|>=|<|>|===))
        (?P
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?sz/Specifier._compare_compatible...*r]rZ)joinlist	itertools	takewhile_version_splitrB)rrbr2prefixrrr_compare_compatibleszSpecifier._compare_compatiblecCsp|drPt|j}t|dd}tt|}|dt|}t||\}}nt|}|jsht|j}||kS)Nr})endswithrZpublicrr7len_pad_versionlocalrarrrrcs


zSpecifier._compare_equalcCs|||Sr:)rcrarrrrdszSpecifier._compare_not_equalcCs|t|kSr:rrarrrresz"Specifier._compare_less_than_equalcCs|t|kSr:rrarrrrfsz%Specifier._compare_greater_than_equalcCs<t|}||ksdS|js8|jr8t|jt|jkr8dSdSNFT)rrKbase_versionrarrrrgszSpecifier._compare_less_thancCs^t|}||ksdS|js8|jr8t|jt|jkr8dS|jdk	rZt|jt|jkrZdSdSr)rZis_postreleaserrrarrrrhs
zSpecifier._compare_greater_thancCst|t|kSr:)r7lowerrarrr_compare_arbitraryszSpecifier._compare_arbitrarycCsR|jdk	r|jS|j\}}|dkrN|dkr@|dr@|dd}t|jrNdSdS)N)rZr]r\rurvrZr}rTF)r1r0rr
rK)rr)r*rrrrs


zSpecifier.prereleasescCs
||_dSr:rFrrrrrsN)rr
rrirjrkrlrmr+r@rsrrcrdrerfrgrhrrPrr&rrrrrtsD_



"





rtz^([0-9]+)((?:a|b|c|rc)[0-9]+)$cCs@g}|dD],}t|}|r0||q||q|S)Nrw)split
_prefix_regexr,extendgroupsrL)r*resultrr3rrrr's
rc
Csgg}}|ttdd||ttdd|||t|dd||t|dd|ddgtdt|dt|d|ddgtdt|dt|dttj|ttj|fS)NcSs|Sr:isdigitryrrrr{6z_pad_version..cSs|Sr:rryrrrr{7rrr0)rLrrrrinsertmaxchain)leftrightZ
left_splitZright_splitrrrr2s 
""rc@seZdZdddZddZddZd	d
ZddZd
dZddZ	ddZ
ddZeddZ
e
jddZ
ddZdddZd ddZdS)!SpecifierSetr(Nc	Csndd|dD}t}|D]:}z|t|Wqtk
rV|t|YqXqt||_||_dS)NcSsg|]}|r|qSr)r/.0srrr
Rsz)SpecifierSet.__init__..,)	rsetaddrtrrQ	frozenset_specsr1)rZ
specifiersrZparsed	specifierrrrr4Os
zSpecifierSet.__init__cCs*|jdk	rd|jnd}dt||S)Nr5r(z)r1r-rr7r8rrrr9ds
zSpecifierSet.__repr__cCsdtdd|jDS)Nrcss|]}t|VqdSr:)r7rrrr	nsz'SpecifierSet.__str__..)r~sortedrrrrrrmszSpecifierSet.__str__cCs
t|jSr:)r;rrrrrrpszSpecifierSet.__hash__cCst|trt|}nt|ts"tSt}t|j|jB|_|jdkrX|jdk	rX|j|_n<|jdk	rv|jdkrv|j|_n|j|jkr|j|_ntd|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)r=rrr>rrr1
ValueError)rrrrrr__and__ss 





zSpecifierSet.__and__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:r=rrr'r7r>rrrrrrs



zSpecifierSet.__eq__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:rrrrrrs



zSpecifierSet.__ne__cCs
t|jSr:)rrrrrr__len__szSpecifierSet.__len__cCs
t|jSr:)iterrrrrr__iter__szSpecifierSet.__iter__cCs.|jdk	r|jS|jsdStdd|jDS)Ncss|]}|jVqdSr:rrrrrrsz+SpecifierSet.prereleases..)r1ranyrrrrrs

zSpecifierSet.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:rGrHrrrrIszSpecifierSet.__contains__csLtttfstdkr$|js2jr2dStfdd|jDS)NFc3s|]}|jdVqdS)rNrGrrrrrrsz(SpecifierSet.contains..)r=r	rr
rrKallrrrrrr s
zSpecifierSet.containscCs|dkr|j}|jr6|jD]}|j|t|d}q|Sg}g}|D]P}t|ttfs^t|}n|}t|trnqB|jr|s|s|	|qB|	|qB|s|r|dkr|S|SdS)Nr)
rrr"boolr=r	rr
rKrL)rr!rr2ZfilteredrMrrOrrrr"s*



zSpecifierSet.filter)r(N)N)N)rr
rr4r9rrrrrrrrPrr&rIr r"rrrrrMs 
	




r)Z
__future__rrrr#rqrrjZ_compatrrr*rr	r
rrABCMetaobjectrr'rQrsrtrkrrrrrrrrs&9	4	
site-packages/setuptools/_vendor/packaging/__pycache__/markers.cpython-38.pyc000064400000021321151464452710023401 0ustar00U

Qab/ 	@s@ddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
mZmZddlm
Z
mZmZmZddlmZddlmZddlmZmZd	d
ddd
gZGdd	d	eZGdd
d
eZGdddeZGdddeZGdddeZGdddeZ GdddeZ!ededBedBedBedBedBedBed Bed!Bed"Bed#Bed$Bed%Bed&Bed'Bed(Bed)Bed*BZ"d#d"ddddd+Z#e"$d,d-ed.ed/Bed0Bed1Bed2Bed3Bed4Bed5BZ%e%ed6Bed7BZ&e&$d8d-ed9ed:BZ'e'$d;d-ed<ed=BZ(e"e'BZ)ee)e&e)Z*e*$d>d-ed?+Z,ed@+Z-eZ.e*ee,e.e-BZ/e.e/e
e(e.>ee.eZ0dAdBZ1dSdDdEZ2dFd-dGd-ej3ej4ej5ej6ej7ej8dHZ9dIdJZ:eZ;dKdLZdQd
Z?GdRddeZ@dS)T)absolute_importdivisionprint_functionN)ParseExceptionParseResultsstringStart	stringEnd)
ZeroOrMoreGroupForwardQuotedString)Literal)string_types)	SpecifierInvalidSpecifier
InvalidMarkerUndefinedComparisonUndefinedEnvironmentNameMarkerdefault_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N__name__
__module____qualname____doc__rrH/usr/lib/python3.8/site-packages/setuptools/_vendor/packaging/markers.pyrsc@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrrrrrrsc@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrrrrrr%sc@s,eZdZddZddZddZddZd	S)
NodecCs
||_dSN)value)selfr rrr__init__.sz
Node.__init__cCs
t|jSr)strr r!rrr__str__1szNode.__str__cCsd|jjt|S)Nz<{0}({1!r})>)format	__class__rr#r$rrr__repr__4sz
Node.__repr__cCstdSr)NotImplementedErrorr$rrr	serialize7szNode.serializeN)rrrr"r%r(r*rrrrr,src@seZdZddZdS)VariablecCst|Srr#r$rrrr*=szVariable.serializeNrrrr*rrrrr+;sr+c@seZdZddZdS)ValuecCs
d|S)Nz"{0}")r&r$rrrr*CszValue.serializeNr-rrrrr.Asr.c@seZdZddZdS)OpcCst|Srr,r$rrrr*IszOp.serializeNr-rrrrr/Gsr/implementation_versionplatform_python_implementationimplementation_namepython_full_versionplatform_releaseplatform_versionplatform_machineplatform_systempython_versionsys_platformos_nameos.namesys.platformplatform.versionplatform.machineplatform.python_implementationpython_implementationZextra)r;r<r=r>r?r@cCstt|d|dSNr)r+ALIASESgetsltrrrirHz=====>=<=!=z~=><not inincCst|dSrA)r/rDrrrrHwrI'"cCst|dSrA)r.rDrrrrHzrIandorcCst|dSrA)tuplerDrrrrHrI()cCs t|trdd|DS|SdS)NcSsg|]}t|qSr)_coerce_parse_result).0irrr
sz(_coerce_parse_result..)
isinstancer)resultsrrrrYs
rYTcCst|tttfstt|trHt|dkrHt|dttfrHt|dSt|trdd|D}|rnd|Sdd|dSn"t|trddd	|DS|SdS)
Nrrcss|]}t|ddVqdS)F)firstN)_format_markerrZmrrr	sz!_format_marker.. rWrXcSsg|]}|qSr)r*rarrrr\sz"_format_marker..)r]listrVrAssertionErrorlenr`join)markerr_innerrrrr`s


r`cCs||kSrrlhsrhsrrrrHrIcCs||kSrrrkrrrrHrI)rQrPrOrLrJrMrKrNcCslztd||g}Wntk
r.YnX||St|}|dkrbtd||||||S)Nz#Undefined {0!r} on {1!r} and {2!r}.)	rrhr*rcontains
_operatorsrCrr&)rloprmspecZoperrrr_eval_ops
rscCs&||t}|tkr"td||S)Nz/{0!r} does not exist in evaluation environment.)rC
_undefinedrr&)environmentnamer rrr_get_envsrwc	Csgg}|D]}t|tttfs"tt|trB|dt||q
t|tr|\}}}t|trtt||j	}|j	}n|j	}t||j	}|dt
|||q
|dkst|dkr
|gq
tdd|DS)N)rTrUrUcss|]}t|VqdSr)all)rZitemrrrrcsz$_evaluate_markers..)r]rerVrrfappend_evaluate_markersr+rwr rsany)	ZmarkersrugroupsrirlrqrmZ	lhs_valueZ	rhs_valuerrrr|s"



r|cCs2d|}|j}|dkr.||dt|j7}|S)Nz{0.major}.{0.minor}.{0.micro}finalr)r&releaselevelr#serial)infoversionZkindrrrformat_full_versions

rcCslttdr ttjj}tjj}nd}d}||tjtt	t
ttttddtjdS)Nimplementation0rn)r2r0r:r6r4r7r5r3r1r8r9)
hasattrsysrrrrvosplatformmachinereleasesystemr8r@)Ziverr2rrrrs"

c@s.eZdZddZddZddZd
dd	ZdS)rc
Cs`ztt||_WnFtk
rZ}z(d|||j|jd}t|W5d}~XYnXdS)Nz+Invalid marker: {0!r}, parse error at {1!r})rYMARKERZparseString_markersrr&Zlocr)r!rieZerr_strrrrr"szMarker.__init__cCs
t|jSr)r`rr$rrrr%szMarker.__str__cCsdt|S)Nz)r&r#r$rrrr(szMarker.__repr__NcCs$t}|dk	r||t|j|S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)rupdater|r)r!ruZcurrent_environmentrrrevaluate s	
zMarker.evaluate)N)rrrr"r%r(rrrrrrs)T)AZ
__future__rrroperatorrrrZsetuptools.extern.pyparsingrrrrr	r
rrr
LZ_compatrZ
specifiersrr__all__
ValueErrorrrrobjectrr+r.r/ZVARIABLErBZsetParseActionZVERSION_CMPZ	MARKER_OPZMARKER_VALUEZBOOLOPZ
MARKER_VARZMARKER_ITEMsuppressZLPARENZRPARENZMARKER_EXPRZMARKER_ATOMrrYr`ltleeqnegegtrprsrtrwr|rrrrrrrs	


site-packages/setuptools/_vendor/packaging/__about__.py000064400000001320151464452710017352 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

__all__ = [
    "__title__", "__summary__", "__uri__", "__version__", "__author__",
    "__email__", "__license__", "__copyright__",
]

__title__ = "packaging"
__summary__ = "Core utilities for Python packages"
__uri__ = "https://github.com/pypa/packaging"

__version__ = "16.8"

__author__ = "Donald Stufft and individual contributors"
__email__ = "donald@stufft.io"

__license__ = "BSD or Apache License, Version 2.0"
__copyright__ = "Copyright 2014-2016 %s" % __author__
site-packages/setuptools/_vendor/packaging/_compat.py000064400000001534151464452710017075 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import sys


PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3

# flake8: noqa

if PY3:
    string_types = str,
else:
    string_types = basestring,


def with_metaclass(meta, *bases):
    """
    Create a base class with a metaclass.
    """
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(meta):
        def __new__(cls, name, this_bases, d):
            return meta(name, bases, d)
    return type.__new__(metaclass, 'temporary_class', (), {})
site-packages/setuptools/_vendor/packaging/specifiers.py000064400000066571151464452710017623 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import abc
import functools
import itertools
import re

from ._compat import string_types, with_metaclass
from .version import Version, LegacyVersion, parse


class InvalidSpecifier(ValueError):
    """
    An invalid specifier was found, users should refer to PEP 440.
    """


class BaseSpecifier(with_metaclass(abc.ABCMeta, object)):

    @abc.abstractmethod
    def __str__(self):
        """
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        """

    @abc.abstractmethod
    def __hash__(self):
        """
        Returns a hash value for this Specifier like object.
        """

    @abc.abstractmethod
    def __eq__(self, other):
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        """

    @abc.abstractmethod
    def __ne__(self, other):
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        """

    @abc.abstractproperty
    def prereleases(self):
        """
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @prereleases.setter
    def prereleases(self, value):
        """
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @abc.abstractmethod
    def contains(self, item, prereleases=None):
        """
        Determines if the given item is contained within this specifier.
        """

    @abc.abstractmethod
    def filter(self, iterable, prereleases=None):
        """
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        """


class _IndividualSpecifier(BaseSpecifier):

    _operators = {}

    def __init__(self, spec="", prereleases=None):
        match = self._regex.search(spec)
        if not match:
            raise InvalidSpecifier("Invalid specifier: '{0}'".format(spec))

        self._spec = (
            match.group("operator").strip(),
            match.group("version").strip(),
        )

        # Store whether or not this Specifier should accept prereleases
        self._prereleases = prereleases

    def __repr__(self):
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "<{0}({1!r}{2})>".format(
            self.__class__.__name__,
            str(self),
            pre,
        )

    def __str__(self):
        return "{0}{1}".format(*self._spec)

    def __hash__(self):
        return hash(self._spec)

    def __eq__(self, other):
        if isinstance(other, string_types):
            try:
                other = self.__class__(other)
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec == other._spec

    def __ne__(self, other):
        if isinstance(other, string_types):
            try:
                other = self.__class__(other)
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec != other._spec

    def _get_operator(self, op):
        return getattr(self, "_compare_{0}".format(self._operators[op]))

    def _coerce_version(self, version):
        if not isinstance(version, (LegacyVersion, Version)):
            version = parse(version)
        return version

    @property
    def operator(self):
        return self._spec[0]

    @property
    def version(self):
        return self._spec[1]

    @property
    def prereleases(self):
        return self._prereleases

    @prereleases.setter
    def prereleases(self, value):
        self._prereleases = value

    def __contains__(self, item):
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # Determine if prereleases are to be allowed or not.
        if prereleases is None:
            prereleases = self.prereleases

        # Normalize item to a Version or LegacyVersion, this allows us to have
        # a shortcut for ``"2.0" in Specifier(">=2")
        item = self._coerce_version(item)

        # Determine if we should be supporting prereleases in this specifier
        # or not, if we do not support prereleases than we can short circuit
        # logic if this version is a prereleases.
        if item.is_prerelease and not prereleases:
            return False

        # Actually do the comparison to determine if this item is contained
        # within this Specifier or not.
        return self._get_operator(self.operator)(item, self.version)

    def filter(self, iterable, prereleases=None):
        yielded = False
        found_prereleases = []

        kw = {"prereleases": prereleases if prereleases is not None else True}

        # Attempt to iterate over all the values in the iterable and if any of
        # them match, yield them.
        for version in iterable:
            parsed_version = self._coerce_version(version)

            if self.contains(parsed_version, **kw):
                # If our version is a prerelease, and we were not set to allow
                # prereleases, then we'll store it for later incase nothing
                # else matches this specifier.
                if (parsed_version.is_prerelease and not
                        (prereleases or self.prereleases)):
                    found_prereleases.append(version)
                # Either this is not a prerelease, or we should have been
                # accepting prereleases from the begining.
                else:
                    yielded = True
                    yield version

        # Now that we've iterated over everything, determine if we've yielded
        # any values, and if we have not and we have any prereleases stored up
        # then we will go ahead and yield the prereleases.
        if not yielded and found_prereleases:
            for version in found_prereleases:
                yield version


class LegacySpecifier(_IndividualSpecifier):

    _regex_str = (
        r"""
        (?P(==|!=|<=|>=|<|>))
        \s*
        (?P
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        """
    )

    _regex = re.compile(
        r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
    }

    def _coerce_version(self, version):
        if not isinstance(version, LegacyVersion):
            version = LegacyVersion(str(version))
        return version

    def _compare_equal(self, prospective, spec):
        return prospective == self._coerce_version(spec)

    def _compare_not_equal(self, prospective, spec):
        return prospective != self._coerce_version(spec)

    def _compare_less_than_equal(self, prospective, spec):
        return prospective <= self._coerce_version(spec)

    def _compare_greater_than_equal(self, prospective, spec):
        return prospective >= self._coerce_version(spec)

    def _compare_less_than(self, prospective, spec):
        return prospective < self._coerce_version(spec)

    def _compare_greater_than(self, prospective, spec):
        return prospective > self._coerce_version(spec)


def _require_version_compare(fn):
    @functools.wraps(fn)
    def wrapped(self, prospective, spec):
        if not isinstance(prospective, Version):
            return False
        return fn(self, prospective, spec)
    return wrapped


class Specifier(_IndividualSpecifier):

    _regex_str = (
        r"""
        (?P(~=|==|!=|<=|>=|<|>|===))
        (?P
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
        "===": "arbitrary",
    }

    @_require_version_compare
    def _compare_compatible(self, prospective, spec):
        # Compatible releases have an equivalent combination of >= and ==. That
        # is that ~=2.2 is equivalent to >=2.2,==2.*. This allows us to
        # implement this in terms of the other specifiers instead of
        # implementing it ourselves. The only thing we need to do is construct
        # the other specifiers.

        # We want everything but the last item in the version, but we want to
        # ignore post and dev releases and we want to treat the pre-release as
        # it's own separate segment.
        prefix = ".".join(
            list(
                itertools.takewhile(
                    lambda x: (not x.startswith("post") and not
                               x.startswith("dev")),
                    _version_split(spec),
                )
            )[:-1]
        )

        # Add the prefix notation to the end of our string
        prefix += ".*"

        return (self._get_operator(">=")(prospective, spec) and
                self._get_operator("==")(prospective, prefix))

    @_require_version_compare
    def _compare_equal(self, prospective, spec):
        # We need special logic to handle prefix matching
        if spec.endswith(".*"):
            # In the case of prefix matching we want to ignore local segment.
            prospective = Version(prospective.public)
            # Split the spec out by dots, and pretend that there is an implicit
            # dot in between a release segment and a pre-release segment.
            spec = _version_split(spec[:-2])  # Remove the trailing .*

            # Split the prospective version out by dots, and pretend that there
            # is an implicit dot in between a release segment and a pre-release
            # segment.
            prospective = _version_split(str(prospective))

            # Shorten the prospective version to be the same length as the spec
            # so that we can determine if the specifier is a prefix of the
            # prospective version or not.
            prospective = prospective[:len(spec)]

            # Pad out our two sides with zeros so that they both equal the same
            # length.
            spec, prospective = _pad_version(spec, prospective)
        else:
            # Convert our spec string into a Version
            spec = Version(spec)

            # If the specifier does not have a local segment, then we want to
            # act as if the prospective version also does not have a local
            # segment.
            if not spec.local:
                prospective = Version(prospective.public)

        return prospective == spec

    @_require_version_compare
    def _compare_not_equal(self, prospective, spec):
        return not self._compare_equal(prospective, spec)

    @_require_version_compare
    def _compare_less_than_equal(self, prospective, spec):
        return prospective <= Version(spec)

    @_require_version_compare
    def _compare_greater_than_equal(self, prospective, spec):
        return prospective >= Version(spec)

    @_require_version_compare
    def _compare_less_than(self, prospective, spec):
        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec)

        # Check to see if the prospective version is less than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective < spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a pre-release version, that we do not accept pre-release
        # versions for the version mentioned in the specifier (e.g. <3.1 should
        # not match 3.1.dev0, but should match 3.0.dev0).
        if not spec.is_prerelease and prospective.is_prerelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # less than the spec version *and* it's not a pre-release of the same
        # version in the spec.
        return True

    @_require_version_compare
    def _compare_greater_than(self, prospective, spec):
        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec)

        # Check to see if the prospective version is greater than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective > spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a post-release version, that we do not accept
        # post-release versions for the version mentioned in the specifier
        # (e.g. >3.1 should not match 3.0.post0, but should match 3.2.post0).
        if not spec.is_postrelease and prospective.is_postrelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # Ensure that we do not allow a local version of the version mentioned
        # in the specifier, which is techincally greater than, to match.
        if prospective.local is not None:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # greater than the spec version *and* it's not a pre-release of the
        # same version in the spec.
        return True

    def _compare_arbitrary(self, prospective, spec):
        return str(prospective).lower() == str(spec).lower()

    @property
    def prereleases(self):
        # If there is an explicit prereleases set for this, then we'll just
        # blindly use that.
        if self._prereleases is not None:
            return self._prereleases

        # Look at all of our specifiers and determine if they are inclusive
        # operators, and if they are if they are including an explicit
        # prerelease.
        operator, version = self._spec
        if operator in ["==", ">=", "<=", "~=", "==="]:
            # The == specifier can include a trailing .*, if it does we
            # want to remove before parsing.
            if operator == "==" and version.endswith(".*"):
                version = version[:-2]

            # Parse the version, and if it is a pre-release than this
            # specifier allows pre-releases.
            if parse(version).is_prerelease:
                return True

        return False

    @prereleases.setter
    def prereleases(self, value):
        self._prereleases = value


_prefix_regex = re.compile(r"^([0-9]+)((?:a|b|c|rc)[0-9]+)$")


def _version_split(version):
    result = []
    for item in version.split("."):
        match = _prefix_regex.search(item)
        if match:
            result.extend(match.groups())
        else:
            result.append(item)
    return result


def _pad_version(left, right):
    left_split, right_split = [], []

    # Get the release segment of our versions
    left_split.append(list(itertools.takewhile(lambda x: x.isdigit(), left)))
    right_split.append(list(itertools.takewhile(lambda x: x.isdigit(), right)))

    # Get the rest of our versions
    left_split.append(left[len(left_split[0]):])
    right_split.append(right[len(right_split[0]):])

    # Insert our padding
    left_split.insert(
        1,
        ["0"] * max(0, len(right_split[0]) - len(left_split[0])),
    )
    right_split.insert(
        1,
        ["0"] * max(0, len(left_split[0]) - len(right_split[0])),
    )

    return (
        list(itertools.chain(*left_split)),
        list(itertools.chain(*right_split)),
    )


class SpecifierSet(BaseSpecifier):

    def __init__(self, specifiers="", prereleases=None):
        # Split on , to break each indidivual specifier into it's own item, and
        # strip each item to remove leading/trailing whitespace.
        specifiers = [s.strip() for s in specifiers.split(",") if s.strip()]

        # Parsed each individual specifier, attempting first to make it a
        # Specifier and falling back to a LegacySpecifier.
        parsed = set()
        for specifier in specifiers:
            try:
                parsed.add(Specifier(specifier))
            except InvalidSpecifier:
                parsed.add(LegacySpecifier(specifier))

        # Turn our parsed specifiers into a frozen set and save them for later.
        self._specs = frozenset(parsed)

        # Store our prereleases value so we can use it later to determine if
        # we accept prereleases or not.
        self._prereleases = prereleases

    def __repr__(self):
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "".format(str(self), pre)

    def __str__(self):
        return ",".join(sorted(str(s) for s in self._specs))

    def __hash__(self):
        return hash(self._specs)

    def __and__(self, other):
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        specifier = SpecifierSet()
        specifier._specs = frozenset(self._specs | other._specs)

        if self._prereleases is None and other._prereleases is not None:
            specifier._prereleases = other._prereleases
        elif self._prereleases is not None and other._prereleases is None:
            specifier._prereleases = self._prereleases
        elif self._prereleases == other._prereleases:
            specifier._prereleases = self._prereleases
        else:
            raise ValueError(
                "Cannot combine SpecifierSets with True and False prerelease "
                "overrides."
            )

        return specifier

    def __eq__(self, other):
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif isinstance(other, _IndividualSpecifier):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs == other._specs

    def __ne__(self, other):
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif isinstance(other, _IndividualSpecifier):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs != other._specs

    def __len__(self):
        return len(self._specs)

    def __iter__(self):
        return iter(self._specs)

    @property
    def prereleases(self):
        # If we have been given an explicit prerelease modifier, then we'll
        # pass that through here.
        if self._prereleases is not None:
            return self._prereleases

        # If we don't have any specifiers, and we don't have a forced value,
        # then we'll just return None since we don't know if this should have
        # pre-releases or not.
        if not self._specs:
            return None

        # Otherwise we'll see if any of the given specifiers accept
        # prereleases, if any of them do we'll return True, otherwise False.
        return any(s.prereleases for s in self._specs)

    @prereleases.setter
    def prereleases(self, value):
        self._prereleases = value

    def __contains__(self, item):
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # Ensure that our item is a Version or LegacyVersion instance.
        if not isinstance(item, (LegacyVersion, Version)):
            item = parse(item)

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # We can determine if we're going to allow pre-releases by looking to
        # see if any of the underlying items supports them. If none of them do
        # and this item is a pre-release then we do not allow it and we can
        # short circuit that here.
        # Note: This means that 1.0.dev1 would not be contained in something
        #       like >=1.0.devabc however it would be in >=1.0.debabc,>0.0.dev0
        if not prereleases and item.is_prerelease:
            return False

        # We simply dispatch to the underlying specs here to make sure that the
        # given version is contained within all of them.
        # Note: This use of all() here means that an empty set of specifiers
        #       will always return True, this is an explicit design decision.
        return all(
            s.contains(item, prereleases=prereleases)
            for s in self._specs
        )

    def filter(self, iterable, prereleases=None):
        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # If we have any specifiers, then we want to wrap our iterable in the
        # filter method for each one, this will act as a logical AND amongst
        # each specifier.
        if self._specs:
            for spec in self._specs:
                iterable = spec.filter(iterable, prereleases=bool(prereleases))
            return iterable
        # If we do not have any specifiers, then we need to have a rough filter
        # which will filter out any pre-releases, unless there are no final
        # releases, and which will filter out LegacyVersion in general.
        else:
            filtered = []
            found_prereleases = []

            for item in iterable:
                # Ensure that we some kind of Version class for this item.
                if not isinstance(item, (LegacyVersion, Version)):
                    parsed_version = parse(item)
                else:
                    parsed_version = item

                # Filter out any item which is parsed as a LegacyVersion
                if isinstance(parsed_version, LegacyVersion):
                    continue

                # Store any item which is a pre-release for later unless we've
                # already found a final version or we are accepting prereleases
                if parsed_version.is_prerelease and not prereleases:
                    if not filtered:
                        found_prereleases.append(item)
                else:
                    filtered.append(item)

            # If we've found no items except for pre-releases, then we'll go
            # ahead and use the pre-releases
            if not filtered and found_prereleases and prereleases is None:
                return found_prereleases

            return filtered
site-packages/setuptools/_vendor/packaging/markers.py000064400000020057151464452710017120 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import operator
import os
import platform
import sys

from setuptools.extern.pyparsing import ParseException, ParseResults, stringStart, stringEnd
from setuptools.extern.pyparsing import ZeroOrMore, Group, Forward, QuotedString
from setuptools.extern.pyparsing import Literal as L  # noqa

from ._compat import string_types
from .specifiers import Specifier, InvalidSpecifier


__all__ = [
    "InvalidMarker", "UndefinedComparison", "UndefinedEnvironmentName",
    "Marker", "default_environment",
]


class InvalidMarker(ValueError):
    """
    An invalid marker was found, users should refer to PEP 508.
    """


class UndefinedComparison(ValueError):
    """
    An invalid operation was attempted on a value that doesn't support it.
    """


class UndefinedEnvironmentName(ValueError):
    """
    A name was attempted to be used that does not exist inside of the
    environment.
    """


class Node(object):

    def __init__(self, value):
        self.value = value

    def __str__(self):
        return str(self.value)

    def __repr__(self):
        return "<{0}({1!r})>".format(self.__class__.__name__, str(self))

    def serialize(self):
        raise NotImplementedError


class Variable(Node):

    def serialize(self):
        return str(self)


class Value(Node):

    def serialize(self):
        return '"{0}"'.format(self)


class Op(Node):

    def serialize(self):
        return str(self)


VARIABLE = (
    L("implementation_version") |
    L("platform_python_implementation") |
    L("implementation_name") |
    L("python_full_version") |
    L("platform_release") |
    L("platform_version") |
    L("platform_machine") |
    L("platform_system") |
    L("python_version") |
    L("sys_platform") |
    L("os_name") |
    L("os.name") |  # PEP-345
    L("sys.platform") |  # PEP-345
    L("platform.version") |  # PEP-345
    L("platform.machine") |  # PEP-345
    L("platform.python_implementation") |  # PEP-345
    L("python_implementation") |  # undocumented setuptools legacy
    L("extra")
)
ALIASES = {
    'os.name': 'os_name',
    'sys.platform': 'sys_platform',
    'platform.version': 'platform_version',
    'platform.machine': 'platform_machine',
    'platform.python_implementation': 'platform_python_implementation',
    'python_implementation': 'platform_python_implementation'
}
VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0])))

VERSION_CMP = (
    L("===") |
    L("==") |
    L(">=") |
    L("<=") |
    L("!=") |
    L("~=") |
    L(">") |
    L("<")
)

MARKER_OP = VERSION_CMP | L("not in") | L("in")
MARKER_OP.setParseAction(lambda s, l, t: Op(t[0]))

MARKER_VALUE = QuotedString("'") | QuotedString('"')
MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0]))

BOOLOP = L("and") | L("or")

MARKER_VAR = VARIABLE | MARKER_VALUE

MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR)
MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0]))

LPAREN = L("(").suppress()
RPAREN = L(")").suppress()

MARKER_EXPR = Forward()
MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN)
MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR)

MARKER = stringStart + MARKER_EXPR + stringEnd


def _coerce_parse_result(results):
    if isinstance(results, ParseResults):
        return [_coerce_parse_result(i) for i in results]
    else:
        return results


def _format_marker(marker, first=True):
    assert isinstance(marker, (list, tuple, string_types))

    # Sometimes we have a structure like [[...]] which is a single item list
    # where the single item is itself it's own list. In that case we want skip
    # the rest of this function so that we don't get extraneous () on the
    # outside.
    if (isinstance(marker, list) and len(marker) == 1 and
            isinstance(marker[0], (list, tuple))):
        return _format_marker(marker[0])

    if isinstance(marker, list):
        inner = (_format_marker(m, first=False) for m in marker)
        if first:
            return " ".join(inner)
        else:
            return "(" + " ".join(inner) + ")"
    elif isinstance(marker, tuple):
        return " ".join([m.serialize() for m in marker])
    else:
        return marker


_operators = {
    "in": lambda lhs, rhs: lhs in rhs,
    "not in": lambda lhs, rhs: lhs not in rhs,
    "<": operator.lt,
    "<=": operator.le,
    "==": operator.eq,
    "!=": operator.ne,
    ">=": operator.ge,
    ">": operator.gt,
}


def _eval_op(lhs, op, rhs):
    try:
        spec = Specifier("".join([op.serialize(), rhs]))
    except InvalidSpecifier:
        pass
    else:
        return spec.contains(lhs)

    oper = _operators.get(op.serialize())
    if oper is None:
        raise UndefinedComparison(
            "Undefined {0!r} on {1!r} and {2!r}.".format(op, lhs, rhs)
        )

    return oper(lhs, rhs)


_undefined = object()


def _get_env(environment, name):
    value = environment.get(name, _undefined)

    if value is _undefined:
        raise UndefinedEnvironmentName(
            "{0!r} does not exist in evaluation environment.".format(name)
        )

    return value


def _evaluate_markers(markers, environment):
    groups = [[]]

    for marker in markers:
        assert isinstance(marker, (list, tuple, string_types))

        if isinstance(marker, list):
            groups[-1].append(_evaluate_markers(marker, environment))
        elif isinstance(marker, tuple):
            lhs, op, rhs = marker

            if isinstance(lhs, Variable):
                lhs_value = _get_env(environment, lhs.value)
                rhs_value = rhs.value
            else:
                lhs_value = lhs.value
                rhs_value = _get_env(environment, rhs.value)

            groups[-1].append(_eval_op(lhs_value, op, rhs_value))
        else:
            assert marker in ["and", "or"]
            if marker == "or":
                groups.append([])

    return any(all(item) for item in groups)


def format_full_version(info):
    version = '{0.major}.{0.minor}.{0.micro}'.format(info)
    kind = info.releaselevel
    if kind != 'final':
        version += kind[0] + str(info.serial)
    return version


def default_environment():
    if hasattr(sys, 'implementation'):
        iver = format_full_version(sys.implementation.version)
        implementation_name = sys.implementation.name
    else:
        iver = '0'
        implementation_name = ''

    return {
        "implementation_name": implementation_name,
        "implementation_version": iver,
        "os_name": os.name,
        "platform_machine": platform.machine(),
        "platform_release": platform.release(),
        "platform_system": platform.system(),
        "platform_version": platform.version(),
        "python_full_version": platform.python_version(),
        "platform_python_implementation": platform.python_implementation(),
        "python_version": platform.python_version()[:3],
        "sys_platform": sys.platform,
    }


class Marker(object):

    def __init__(self, marker):
        try:
            self._markers = _coerce_parse_result(MARKER.parseString(marker))
        except ParseException as e:
            err_str = "Invalid marker: {0!r}, parse error at {1!r}".format(
                marker, marker[e.loc:e.loc + 8])
            raise InvalidMarker(err_str)

    def __str__(self):
        return _format_marker(self._markers)

    def __repr__(self):
        return "".format(str(self))

    def evaluate(self, environment=None):
        """Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        """
        current_environment = default_environment()
        if environment is not None:
            current_environment.update(environment)

        return _evaluate_markers(self._markers, current_environment)
site-packages/setuptools/_vendor/packaging/__init__.py000064400000001001151464452710017177 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

from .__about__ import (
    __author__, __copyright__, __email__, __license__, __summary__, __title__,
    __uri__, __version__
)

__all__ = [
    "__title__", "__summary__", "__uri__", "__version__", "__author__",
    "__email__", "__license__", "__copyright__",
]
site-packages/setuptools/_vendor/__pycache__/pyparsing.cpython-38.pyc000064400000611637151464452710022044 0ustar00U

Qabwi@sdZdZdZdZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZzddlmZWn ek
rddlmZYnXzdd	lmZdd
lmZWn,ek
rdd	l
mZdd
l
mZYnXzddl
mZWnBek
rFzddlmZWnek
r@dZYnXYnXdd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtgiZee	jdduZeddukZ e rpe	j!Z"e#Z$e%Z&e#Z'e(e)e*e+e,ee-e.e/e0e1gZ2n`e	j3Z"e4Z5dvdwZ'gZ2ddl6Z6dx7D]8Z8ze29e:e6e8Wne;k
rYqYnXqeGd~dde?Z@ejAejBZCdZDeDdZEeCeDZFe%dZGdHddzejIDZJGdd#d#eKZLGdd%d%eLZMGdd'd'eLZNGdd)d)eNZOGdd,d,eKZPGddde?ZQGdd(d(e?ZReSeRdd?ZTddPZUddMZVddZWddZXddZYddWZZd/ddZ[Gdd*d*e?Z\Gdd2d2e\Z]Gddde]Z^Gddde]Z_Gddde]Z`e`Zae`e\_bGddde]ZcGddde`ZdGdd
d
ecZeGddrdre]ZfGdd5d5e]ZgGdd-d-e]ZhGdd+d+e]ZiGddde]ZjGdd4d4e]ZkGddde]ZlGdddelZmGdddelZnGdddelZoGdd0d0elZpGdd/d/elZqGdd7d7elZrGdd6d6elZsGdd&d&e\ZtGdddetZuGdd"d"etZvGdddetZwGdddetZxGdd$d$e\ZyGdddeyZzGdddeyZ{GdddeyZ|Gddde|Z}Gdd8d8e|Z~Gddde?ZeZGdd!d!eyZGdd.d.eyZGdddeyZGddÄdeZGdd3d3eyZGdddeZGdddeZGdddeZGdd1d1eZGdd d e?ZddhZd0ddFZd1ddBZddЄZddUZddTZddԄZd2ddYZddGZd3ddmZddnZddpZe^dIZendOZeodNZepdgZeqdfZegeGddd܍ddބZehd߃ddބZehdddބZeeBeBejdd{d܍BZeeedeZe`deddee}eeBddZddeZddSZddbZdd`ZddsZeddބZeddބZddZddQZddRZddkZe?e_d4ddqZe@Ze?e_e?e_ededfddoZeZeehdddZeehdddZeehddehddBdZeeadedZdddefddVZd5ddlZedZedZeegeCeFdd\ZZeed	7d
ZehddHeàġd
dZŐddaZeehdddZehddZehdɡdZehddZeehddeBdZeZehddZee}egeJdːdeegde`d˃eoϡdZeeeeBddd@ZGd dtdtZeӐd!kredd"Zedd#ZegeCeFd$Zee֐d%dՐd&eZeee׃d'Zؐd(eBZee֐d%dՐd&eZeeeڃd)ZeԐd*eِd'eeېd)Zeܠݐd+ejޠݐd,ejߠݐd,ejݐd-ddlZej᠝eejejݐd.dS(6a	
pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and executing simple grammars,
vs. the traditional lex/yacc approach, or the use of regular expressions.  With pyparsing, you
don't need to learn a new syntax for defining grammars or matching expressions - the parsing module
provides a library of classes that you use to construct the grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form 
C{", !"}), built up using L{Word}, L{Literal}, and L{And} elements 
(L{'+'} operator gives L{And} expressions, strings are auto-converted to
L{Literal} expressions)::

    from pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the self-explanatory
class names, and the use of '+', '|' and '^' operators.

The L{ParseResults} object returned from L{ParserElement.parseString} can be accessed as a nested list, a dictionary, or an
object with named attributes.

The pyparsing module handles some of the problems that are typically vexing when writing text parsers:
 - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello  ,  World  !", etc.)
 - quoted strings
 - embedded comments


Getting Started -
-----------------
Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:
 - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes
 - construct character word-group expressions using the L{Word} class
 - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes
 - use L{'+'}, L{'|'}, L{'^'}, and L{'&'} operators to combine simple expressions into more complex ones
 - associate names with your parsed results using L{ParserElement.setResultsName}
 - find some helpful expression short-cuts like L{delimitedList} and L{oneOf}
 - find more useful common expressions in the L{pyparsing_common} namespace class
z2.2.1z18 Sep 2018 00:49 UTCz*Paul McGuire N)ref)datetime)RLock)Iterable)MutableMapping)OrderedDictAndCaselessKeywordCaselessLiteral
CharsNotInCombineDictEachEmpty
FollowedByForward
GoToColumnGroupKeywordLineEnd	LineStartLiteral
MatchFirstNoMatchNotAny	OneOrMoreOnlyOnceOptionalOrParseBaseExceptionParseElementEnhanceParseExceptionParseExpressionParseFatalExceptionParseResultsParseSyntaxException
ParserElementQuotedStringRecursiveGrammarExceptionRegexSkipTo	StringEndStringStartSuppressTokenTokenConverterWhiteWordWordEnd	WordStart
ZeroOrMore	alphanumsalphas
alphas8bitanyCloseTag
anyOpenTag
cStyleCommentcolcommaSeparatedListcommonHTMLEntitycountedArraycppStyleCommentdblQuotedStringdblSlashComment
delimitedListdictOfdowncaseTokensemptyhexnumshtmlCommentjavaStyleCommentlinelineEnd	lineStartlinenomakeHTMLTagsmakeXMLTagsmatchOnlyAtColmatchPreviousExprmatchPreviousLiteral
nestedExprnullDebugActionnumsoneOfopAssocoperatorPrecedence
printablespunc8bitpythonStyleCommentquotedStringremoveQuotesreplaceHTMLEntityreplaceWith
restOfLinesglQuotedStringsrange	stringEndstringStarttraceParseAction
unicodeStringupcaseTokens
withAttribute
indentedBlockoriginalTextForungroup
infixNotationlocatedExpr	withClass
CloseMatchtokenMappyparsing_commoncCsft|tr|Sz
t|WStk
r`t|td}td}|dd|	|YSXdS)aDrop-in replacement for str(obj) that tries to be Unicode friendly. It first tries
           str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It
           then < returns the unicode object | encodes it with the default encoding | ... >.
        xmlcharrefreplacez&#\d+;cSs$dtt|dddddS)Nz\ur)hexinttry@/usr/lib/python3.8/site-packages/setuptools/_vendor/pyparsing.pyz_ustr..N)

isinstanceZunicodestrUnicodeEncodeErrorencodesysgetdefaultencodingr)setParseActiontransformString)objretZ
xmlcharrefryryrz_ustrs

rz6sum len sorted reversed list tuple set any all min maxccs|]
}|VqdSNry).0yryryrz	srcCs:d}dddD}t||D]\}}|||}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)&;Nry)rsryryrzrsz_xml_escape..zamp gt lt quot apos)splitzipreplace)dataZfrom_symbolsZ
to_symbolsZfrom_Zto_ryryrz_xml_escapes
rc@seZdZdS)
_ConstantsN)__name__
__module____qualname__ryryryrzrsr
0123456789ZABCDEFabcdef\ccs|]}|tjkr|VqdSr)stringZ
whitespacercryryrzrs
c@sPeZdZdZdddZeddZdd	Zd
dZdd
Z	dddZ
ddZdS)rz7base exception class for all parsing runtime exceptionsrNcCs>||_|dkr||_d|_n||_||_||_|||f|_dSNr)locmsgpstr
parserElementargs)selfrrrelemryryrz__init__szParseBaseException.__init__cCs||j|j|j|jS)z
        internal factory method to simplify creating one type of ParseException 
        from another - avoids having __init__ signature conflicts among subclasses
        )rrrr)clsperyryrz_from_exceptionsz"ParseBaseException._from_exceptioncCsN|dkrt|j|jS|dkr,t|j|jS|dkrBt|j|jSt|dS)zsupported attributes by name are:
            - lineno - returns the line number of the exception text
            - col - returns the column number of the exception text
            - line - returns the line containing the exception text
        rL)r;columnrIN)rLrrr;rIAttributeError)rZanameryryrz__getattr__szParseBaseException.__getattr__cCsd|j|j|j|jfS)Nz"%s (at char %d), (line:%d, col:%d))rrrLrrryryrz__str__szParseBaseException.__str__cCst|Srrrryryrz__repr__szParseBaseException.__repr__>!} ('-' operator) indicates that parsing is to stop 
       immediately because an unbacktrackable syntax error has been foundNrryryryrzr%sc@s eZdZdZddZddZdS)r(zZexception thrown by L{ParserElement.validate} if the grammar could be improperly recursivecCs
||_dSrZparseElementTracerparseElementListryryrzr4sz"RecursiveGrammarException.__init__cCs
d|jS)NzRecursiveGrammarException: %srrryryrzr7sz!RecursiveGrammarException.__str__N)rrrrrrryryryrzr(2sc@s,eZdZddZddZddZddZd	S)
_ParseResultsWithOffsetcCs||f|_dSrtup)rZp1Zp2ryryrzr;sz _ParseResultsWithOffset.__init__cCs
|j|Srrriryryrz__getitem__=sz#_ParseResultsWithOffset.__getitem__cCst|jdSNr)reprrrryryrzr?sz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSrrrryryrz	setOffsetAsz!_ParseResultsWithOffset.setOffsetN)rrrrrrrryryryrzr:src@seZdZdZd[ddZddddefddZdd	Zefd
dZdd
Z	ddZ
ddZddZeZ
ddZddZddZddZddZereZeZeZn$eZeZeZddZd d!Zd"d#Zd$d%Zd&d'Zd\d(d)Zd*d+Zd,d-Zd.d/Zd0d1Z d2d3Z!d4d5Z"d6d7Z#d8d9Z$d:d;Z%d} - see L{ParserElement.setResultsName})

    Example::
        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)
    prints::
        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    NTcCs"t||r|St|}d|_|SNT)r}object__new___ParseResults__doinit)rtoklistnameasListmodalZretobjryryrzrks


zParseResults.__new__c
Csb|jrvd|_d|_d|_i|_||_||_|dkr6g}||trP|dd|_n||trft||_n|g|_t	|_
|dk	r^|r^|sd|j|<||trt|}||_||t
dttfr|ddgfks^||tr|g}|r(||trt|d||<ntt|dd||<|||_n6z|d||<Wn$tttfk
r\|||<YnXdS)NFrr)r_ParseResults__name_ParseResults__parent_ParseResults__accumNames_ParseResults__asList_ParseResults__modallist_ParseResults__toklist_generatorTypedict_ParseResults__tokdictrvrr
basestringr$rcopyKeyError	TypeError
IndexError)rrrrrr}ryryrzrtsB



$
zParseResults.__init__cCsPt|ttfr|j|S||jkr4|j|ddStdd|j|DSdS)NrtrcSsg|]}|dqSrryrvryryrz
sz,ParseResults.__getitem__..)r}rvslicerrrr$rryryrzrs


zParseResults.__getitem__cCs||tr0|j|t|g|j|<|d}nD||ttfrN||j|<|}n&|j|tt|dg|j|<|}||trt||_	dSr)
rrgetrrvrrr$wkrefr)rkrr}subryryrz__setitem__s


"
zParseResults.__setitem__c
Cst|ttfrt|j}|j|=t|trH|dkr:||7}t||d}tt||}||j	
D]>\}}|D]0}t|D]"\}\}}	t||	|	|k||<qqxqln|j	|=dSNrr)
r}rvrlenrrrangeindicesreverseritems	enumerater)
rrZmylenZremovedroccurrencesjrvaluepositionryryrz__delitem__s

zParseResults.__delitem__cCs
||jkSr)r)rrryryrz__contains__szParseResults.__contains__cCs
t|jSr)rrrryryrz__len__r|zParseResults.__len__cCs
|jSrrrryryrz__bool__r|zParseResults.__bool__cCs
t|jSriterrrryryrz__iter__r|zParseResults.__iter__cCst|jdddSNrtrrryryrz__reversed__r|zParseResults.__reversed__cCs$t|jdr|jSt|jSdS)Niterkeys)hasattrrrrrryryrz	_iterkeyss
zParseResults._iterkeyscsfddDS)Nc3s|]}|VqdSrryrrrryrzrsz+ParseResults._itervalues..rrryrrz_itervaluesszParseResults._itervaluescsfddDS)Nc3s|]}||fVqdSrryrrryrzrsz*ParseResults._iteritems..rrryrrz
_iteritemsszParseResults._iteritemscCst|S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)rrrryryrzkeysszParseResults.keyscCst|S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r
itervaluesrryryrzvaluesszParseResults.valuescCst|S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r	iteritemsrryryrzrszParseResults.itemscCs
t|jS)zSince keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names.)boolrrryryrzhaskeysszParseResults.haskeyscOs|s
dg}|D]*\}}|dkr0|d|f}qtd|qt|dtsdt|dksd|d|kr~|d}||}||=|S|d}|SdS)a
        Removes and returns item at specified index (default=C{last}).
        Supports both C{list} and C{dict} semantics for C{pop()}. If passed no
        argument or an integer argument, it will use C{list} semantics
        and pop tokens from the list of parsed tokens. If passed a 
        non-integer argument (most likely a string), it will use C{dict}
        semantics and pop the corresponding value from any defined 
        results names. A second default return value argument is 
        supported, just as in C{dict.pop()}.

        Example::
            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())
        prints::
            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        rtdefaultrz-pop() got an unexpected keyword argument '%s'rN)rrr}rvr)rrkwargsrrindexrZdefaultvalueryryrzpops""

zParseResults.popcCs||kr||S|SdS)ai
        Returns named result matching the given key, or if there is no
        such name, then returns the given C{defaultValue} or C{None} if no
        C{defaultValue} is specified.

        Similar to C{dict.get()}.
        
        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        Nry)rkeydefaultValueryryrzr3szParseResults.getcCsR|j|||jD]4\}}t|D]"\}\}}t||||k||<q(qdS)a
        Inserts new element at location index in the list of parsed tokens.
        
        Similar to C{list.insert()}.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        N)rinsertrrrr)rr
ZinsStrrrrrrryryrzrIszParseResults.insertcCs|j|dS)a
        Add single element to end of ParseResults list of elements.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            
            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        N)rappend)ritemryryrzr]szParseResults.appendcCs$t|tr||7}n|j|dS)a
        Add sequence of elements to end of ParseResults list of elements.

        Example::
            patt = OneOrMore(Word(alphas))
            
            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        N)r}r$rextend)rZitemseqryryrzrks

zParseResults.extendcCs|jdd=|jdS)z7
        Clear all elements and results names.
        N)rrclearrryryrzr}szParseResults.clearcCsjz
||WStk
r YdSX||jkrb||jkrH|j|ddStdd|j|DSndSdS)NrrtrcSsg|]}|dqSrryrryryrzrsz,ParseResults.__getattr__..)rrrr$rrryryrzrs


zParseResults.__getattr__cCs|}||7}|Srr)rotherrryryrz__add__szParseResults.__add__cs|jrjt|jfdd|j}fdd|D}|D],\}}|||<t|dtr.c	s4g|],\}}|D]}|t|d|dfqqSrr)rrrvlistr)	addoffsetryrzrsz)ParseResults.__iadd__..r)
rrrrr}r$rrrupdate)rrZ
otheritemsZotherdictitemsrrry)rrrz__iadd__s


zParseResults.__iadd__cCs&t|tr|dkr|S||SdSr)r}rvrrrryryrz__radd__szParseResults.__radd__cCsdt|jt|jfS)Nz(%s, %s))rrrrryryrzrszParseResults.__repr__cCsdddd|jDdS)N[, css(|] }t|trt|nt|VqdSr)r}r$rrrrryryrzrsz'ParseResults.__str__..])rrrryryrzrszParseResults.__str__rcCsLg}|jD]<}|r |r ||t|tr8||7}q
|t|q
|Sr)rrr}r$
_asStringListr)rsepoutrryryrzr(s


zParseResults._asStringListcCsdd|jDS)a
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::
            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # ->  ['sldkj', 'lsdkj', 'sldkj']
            
            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # ->  ['sldkj', 'lsdkj', 'sldkj']
        cSs"g|]}t|tr|n|qSry)r}r$r)rresryryrzrsz'ParseResults.asList..rrryryrzrszParseResults.asListcs6tr|j}n|j}fddtfdd|DS)a
        Returns the named parse results as a nested dictionary.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # ->  (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})
            
            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # ->  {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        cs6t|tr.|r|Sfdd|DSn|SdS)Ncsg|]}|qSryryrtoItemryrzrsz7ParseResults.asDict..toItem..)r}r$r
asDict)rr,ryrzr-s

z#ParseResults.asDict..toItemc3s|]\}}||fVqdSrryrrrr,ryrzrsz&ParseResults.asDict..)PY_3rrr)rZitem_fnryr,rzr.s
	zParseResults.asDictcCs8t|j}|j|_|j|_|j|j|j|_|S)zA
        Returns a new copy of a C{ParseResults} object.
        )r$rrrrrr rrrryryrzrs
zParseResults.copyFcCsLd}g}tdd|jD}|d}|s8d}d}d}d}	|dk	rJ|}	n|jrV|j}	|	sf|rbdSd}	|||d|	d	g7}t|jD]\}
}t|tr|
|kr||||
|o|dk||g7}n||d|o|dk||g7}qd}|
|kr||
}|s|rqnd}t	t
|}
|||d|d	|
d
|d	g	7}q|||d
|	d	g7}d|S)z
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        
css(|] \}}|D]}|d|fVqqdSrNryrryryrzrsz%ParseResults.asXML..  rNZITEM<>.z
%s%s- %s: r4rcss|]}t|tVqdSr)r}r$)rvvryryrzrsz
%s%s[%d]:
%s%s%sr)
rrrr
sortedrr}r$dumpranyrr)rr9depthfullr*NLrrrrr?ryryrzrAgs,

4,zParseResults.dumpcOstj|f||dS)a
        Pretty-printer for parsed results as a list, using the C{pprint} module.
        Accepts additional positional or keyword args as defined for the 
        C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint})

        Example::
            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)
        prints::
            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        N)pprintrrrrryryrzrFszParseResults.pprintcCs.|j|j|jdk	r|p d|j|jffSr)rrrrrrrryryrz__getstate__szParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j||dk	rDt||_nd|_dSr)rrrrr rr)rstater=ZinAccumNamesryryrz__setstate__s
zParseResults.__setstate__cCs|j|j|j|jfSr)rrrrrryryrz__getnewargs__szParseResults.__getnewargs__cCstt|t|Sr)rrrrrryryrzrszParseResults.__dir__)NNTT)N)r)NFrT)rrT)4rrrrrr}rrrrrrr__nonzero__rrrrrr0rrrrrrr
rrrrrrrrr!r#rrr(rr.rr8r;r>rArFrHrJrKrryryryrzr$Dsh&
	'	
4

#
=%
-
cCsF|}d|krt|kr4nn||ddkr4dS||dd|S)aReturns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}} for more information
   on parsing strings containing C{}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   rrr2)rrfind)rstrgrryryrzr;s
cCs|dd|dS)aReturns current line number within a string, counting newlines as line separators.
   The first line is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}} for more information
   on parsing strings containing C{}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   r2rr)count)rrNryryrzrLs
cCsF|dd|}|d|}|dkr2||d|S||ddSdS)zfReturns the line of text containing loc within a string, counting newlines as line separators.
       r2rrN)rMfind)rrNZlastCRZnextCRryryrzrIs
cCs8tdt|dt|dt||t||fdS)NzMatch z at loc z(%d,%d))printrrLr;)instringrexprryryrz_defaultStartDebugActionsrTcCs$tdt|dt|dS)NzMatched z -> )rQrr~r)rRstartlocZendlocrStoksryryrz_defaultSuccessDebugActionsrWcCstdt|dS)NzException raised:)rQr)rRrrSexcryryrz_defaultExceptionDebugActionsrYcGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nry)rryryrzrSsrscstkrfddSdgdgtdddkrFddd}dd	d
ntj}tjd}|ddd
}|d|d|ffdd}d}ztdtdj}Wntk
rt}YnX||_|S)Ncs|Srryrlrx)funcryrzr{r|z_trim_arity..rFrs)rqcSs8tdkrdnd}tj||dd|}|ddgS)N)rqr]rrlimitrs)system_version	traceback
extract_stack)rar
frame_summaryryryrzrdsz"_trim_arity..extract_stackcSs$tj||d}|d}|ddgS)Nr`rtrs)rc
extract_tb)tbraZframesreryryrzrfsz_trim_arity..extract_tbr`rtrc	sz"|dd}dd<|WStk
rdr>n4z.td}|dddddksjW5~Xdkrdd7<YqYqXqdS)NrTrtrsr`r)rrexc_info)rrrgrfZ
foundArityr\ramaxargsZpa_call_line_synthryrzwrapper-s z_trim_arity..wrapperzr	__class__)r)r)	singleArgBuiltinsrbrcrdrfgetattrr	Exceptionr~)r\rkrdZ	LINE_DIFFZ	this_linerl	func_nameryrjrz_trim_aritys,

rrcseZdZdZdZdZeddZeddZddd	Z	d
dZ
dd
ZdddZdddZ
ddZddZddZddZddZddZddd Zd!d"Zdd#d$Zd%d&Zd'd(ZGd)d*d*eZed+k	rGd,d-d-eZnGd.d-d-eZiZeZd/d/gZ dd0d1Z!eZ"ed2d3Z#dZ$edd5d6Z%dd7d8Z&e'dfd9d:Z(d;d<Z)e'fd=d>Z*e'dfd?d@Z+dAdBZ,dCdDZ-dEdFZ.dGdHZ/dIdJZ0dKdLZ1dMdNZ2dOdPZ3dQdRZ4dSdTZ5dUdVZ6dWdXZ7dYdZZ8dd[d\Z9d]d^Z:d_d`Z;dadbZdgdhZ?ddidjZ@dkdlZAdmdnZBdodpZCdqdrZDgfdsdtZEddudvZFfdwdxZGdydzZHd{d|ZId}d~ZJddZKdddZLdddZMZNS)r&z)Abstract base level parser element class.z 
	
FcCs
|t_dS)a
        Overrides the default whitespace chars

        Example::
            # default whitespace chars are space,  and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']
            
            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        N)r&DEFAULT_WHITE_CHARScharsryryrzsetDefaultWhitespaceCharsTs
z'ParserElement.setDefaultWhitespaceCharscCs
|t_dS)a
        Set class to be used for inclusion of string literals into a parser.
        
        Example::
            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        N)r&_literalStringClass)rryryrzinlineLiteralsUsingcsz!ParserElement.inlineLiteralsUsingcCst|_d|_d|_d|_||_d|_tj|_	d|_
d|_d|_t|_
d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr)NNN)rparseAction
failActionstrReprresultsName
saveAsListskipWhitespacer&rs
whiteCharscopyDefaultWhiteCharsmayReturnEmptykeepTabsignoreExprsdebugstreamlined
mayIndexErrorerrmsgmodalResultsdebugActionsrecallPreparse
callDuringTry)rsavelistryryrzrxs(zParserElement.__init__cCs<t|}|jdd|_|jdd|_|jr8tj|_|S)a$
        Make a copy of this C{ParserElement}.  Useful for defining different parse actions
        for the same parsing pattern, using copies of the original parse element.
        
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
            
            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))
        prints::
            [5120, 100, 655360, 268435456]
        Equivalent form of C{expr.copy()} is just C{expr()}::
            integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
        N)rryrrr&rsr)rZcpyryryrzrs
zParserElement.copycCs*||_d|j|_t|dr&|j|j_|S)af
        Define name for this expression, makes debugging and exception messages clearer.
        
        Example::
            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        	Expected 	exception)rrrrrrryryrzsetNames


zParserElement.setNamecCs4|}|dr"|dd}d}||_||_|S)aP
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original C{ParserElement} object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        C{expr("name")} in place of C{expr.setResultsName("name")} - 
        see L{I{__call__}<__call__>}.

        Example::
            date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        *NrtT)rendswithr|r)rrlistAllMatchesZnewselfryryrzsetResultsNames
zParserElement.setResultsNameTcs@|r&|jdfdd	}|_||_nt|jdr<|jj|_|S)zMethod to invoke the Python pdb debugger when this element is
           about to be parsed. Set C{breakFlag} to True to enable, False to
           disable.
        Tcsddl}|||||Sr)pdbZ	set_trace)rRr	doActionscallPreParserZ_parseMethodryrzbreakersz'ParserElement.setBreak..breaker_originalParseMethod)TT)_parserr)rZ	breakFlagrryrrzsetBreaks
zParserElement.setBreakcOs&tttt||_|dd|_|S)a
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)},
        C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where:
         - s   = the original string being parsed (see note below)
         - loc = the location of the matching substring
         - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object
        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        Optional keyword arguments:
         - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See L{I{parseString}} for more information
        on parsing strings containing C{}s, and suggested methods to maintain a
        consistent view of the parsed string, the parse location, and line and column
        positions within the parsed string.
        
        Example::
            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        rF)rmaprrryrrrfnsrryryrzrs"zParserElement.setParseActioncOs4|jtttt|7_|jp,|dd|_|S)z
        Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}}.
        
        See examples in L{I{copy}}.
        rF)ryrrrrrrrryryrzaddParseActionszParserElement.addParseActioncs^|dd|ddrtnt|D] fdd}|j|q$|jpV|dd|_|S)aAdd a boolean predicate function to expression's list of parse actions. See 
        L{I{setParseAction}} for function call signatures. Unlike C{setParseAction}, 
        functions passed to C{addCondition} need to return boolean success/fail of the condition.

        Optional keyword arguments:
         - message = define a custom message to be used in the raised exception
         - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException
         
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        messagezfailed user-defined conditionfatalFcs$tt|||s ||dSr)r	rrrZexc_typefnrryrzpa&sz&ParserElement.addCondition..par)rr#r!ryrr)rrrrryrrzaddConditionszParserElement.addConditioncCs
||_|S)aDefine action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           C{fn(s,loc,expr,err)} where:
            - s = string being parsed
            - loc = location where expression match was attempted and failed
            - expr = the parse expression that failed
            - err = the exception thrown
           The function returns no value.  It may throw C{L{ParseFatalException}}
           if it is desired to stop parsing immediately.)rz)rrryryrz
setFailAction-s
zParserElement.setFailActionc	CsNd}|rJd}|jD]4}z|||\}}d}qWqtk
rDYqXqq|SNTF)rrr!)rrRrZ
exprsFoundeZdummyryryrz_skipIgnorables:s


zParserElement._skipIgnorablescCsH|jr|||}|jrD|j}t|}||krD|||krD|d7}q&|SNr)rrr~rr)rrRrZwtinstrlenryryrzpreParseGs
zParserElement.preParsecCs|gfSrryrrRrrryryrz	parseImplSszParserElement.parseImplcCs|SrryrrRr	tokenlistryryrz	postParseVszParserElement.postParsec
Cs|j}|s|jr|jdr,|jd||||rD|jrD|||}n|}|}zDz||||\}}Wn(tk
rt|t||j	|YnXWnXt
k
r}	z:|jdr|jd||||	|jr|||||	W5d}	~	XYnXn|r|jr|||}n|}|}|js&|t|krjz||||\}}Wn*tk
rft|t||j	|YnXn||||\}}||||}t
||j|j|jd}
|jr|s|jr|rTzN|jD]B}||||
}|dk	rt
||j|jot|t
tf|jd}
qWnFt
k
rP}	z&|jdr>|jd||||	W5d}	~	XYnXnJ|jD]B}||||
}|dk	rZt
||j|jot|t
tf|jd}
qZ|r|jdr|jd|||||
||
fS)Nrrs)rrr)rrzrrrrrr!rrrrrr$r|r}rryrr}r)rrRrrrZ	debuggingprelocZtokensStarttokenserrZ	retTokensrryryrz
_parseNoCacheZst








zParserElement._parseNoCachecCs@z|j||dddWStk
r:t|||j|YnXdS)NF)rr)rr#r!rrrRrryryrztryParseszParserElement.tryParsec	Cs4z|||Wnttfk
r*YdSXdSdS)NFT)rr!rrryryrzcanParseNexts
zParserElement.canParseNextc@seZdZddZdS)zParserElement._UnboundedCachecs~it|_fdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS)	Ncs|Srrrrcachenot_in_cacheryrzrsz3ParserElement._UnboundedCache.__init__..getcs||<dSrryrrrrryrzsetsz3ParserElement._UnboundedCache.__init__..setcsdSrrrrryrzrsz5ParserElement._UnboundedCache.__init__..clearcstSrrrrryrz	cache_lensz9ParserElement._UnboundedCache.__init__..cache_len)rrtypes
MethodTyperrrr)rrrrrryrrzrsz&ParserElement._UnboundedCache.__init__Nrrrrryryryrz_UnboundedCachesrNc@seZdZddZdS)ParserElement._FifoCachecst|_tfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS)	Ncs|Srrrrryrzrs.ParserElement._FifoCache.__init__..getcs>||<tkr:zdWqtk
r6YqXqdSNF)rpopitemrr)rsizeryrzrs.ParserElement._FifoCache.__init__..setcsdSrrrrryrzrs0ParserElement._FifoCache.__init__..clearcstSrrrrryrzrs4ParserElement._FifoCache.__init__..cache_len)	rr_OrderedDictrrrrrrrrrrrrry)rrrrzrs!ParserElement._FifoCache.__init__Nrryryryrz
_FifoCachesrc@seZdZddZdS)rcst|_itgfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_	dS)	Ncs|Srrrrryrzrsrcs4||<tkr&dq|dSr)rrpopleftrr)rkey_fiforryrzrsrcsdSrrr)rrryrzrsrcstSrrrrryrzrsr)
rrcollectionsdequerrrrrrrry)rrrrrzrsrNrryryryrzrsrcCsd\}}|||||f}tjtj}||}	|	|jkrtj|d7<z|||||}	Wn8tk
r}
z|||
j	|
j
W5d}
~
XYn.X|||	d|	df|	W5QRSn@tj|d7<t|	t
r|	|	d|	dfW5QRSW5QRXdS)Nrrr)r&packrat_cache_lock
packrat_cacherrpackrat_cache_statsrrrrmrrr}rp)rrRrrrZHITZMISSlookuprrrryryrz_parseCaches$


zParserElement._parseCachecCs(tjdgttjtjdd<dSr)r&rrrrryryryrz
resetCaches
zParserElement.resetCachecCs8tjs4dt_|dkr tt_nt|t_tjt_dS)aEnables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.
           
           Parameters:
            - cache_size_limit - (default=C{128}) - if an integer value is provided
              will limit the size of the packrat cache; if None is passed, then
              the cache size will be unbounded; if 0 is passed, the cache will
              be effectively disabled.
            
           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method C{ParserElement.enablePackrat()}.  If
           your program uses C{psyco} to "compile as you go", you must call
           C{enablePackrat} before calling C{psyco.full()}.  If you do not do this,
           Python will crash.  For best results, call C{enablePackrat()} immediately
           after importing pyparsing.
           
           Example::
               import pyparsing
               pyparsing.ParserElement.enablePackrat()
        TN)r&_packratEnabledrrrrr)Zcache_size_limitryryrz
enablePackrat%szParserElement.enablePackratc
Cst|js||jD]}|q|js8|}z<||d\}}|rr|||}t	t
}|||Wn0tk
r}ztjrn|W5d}~XYnX|SdS)aB
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set C{parseAll} to True (equivalent to ending
        the grammar with C{L{StringEnd()}}).

        Note: C{parseString} implicitly calls C{expandtabs()} on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the C{loc} argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:
         - calling C{parseWithTabs} on your grammar before calling C{parseString}
           (see L{I{parseWithTabs}})
         - define your parse action using the full C{(s,loc,toks)} signature, and
           reference the input string using the parse action's C{s} argument
         - explictly expand the tabs in your input string before calling
           C{parseString}
        
        Example::
            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        rN)
r&rr
streamlinerr
expandtabsrrrr+rverbose_stacktrace)rrRparseAllrrrZserXryryrzparseStringHs$

zParserElement.parseStringc
cs6|js||jD]}|q|js4t|}t|}d}|j}|j}t	
d}	z||kr|	|krz |||}
|||
dd\}}Wntk
r|
d}YqZX||kr|	d7}	||
|fV|r|||}
|
|kr|}q|d7}q|}qZ|
d}qZWn4tk
r0}zt	j
rn|W5d}~XYnXdS)a
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        C{maxMatches} argument, to clip scanning after 'n' matches are found.  If
        C{overlap} is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See L{I{parseString}} for more information on parsing
        strings with embedded tabs.

        Example::
            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens,start,end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])
        
        prints::
        
            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        rFrrN)rrrrrrrrrr&rr!rr)rrR
maxMatchesZoverlaprrrZ
preparseFnZparseFnmatchesrZnextLocrZnextlocrXryryrz
scanStringzsB




zParserElement.scanStringc
Csg}d}d|_z||D]Z\}}}|||||rpt|trR||7}nt|trf||7}n
|||}q|||ddd|D}dtt	t
|WStk
r}ztj
rƂn|W5d}~XYnXdS)af
        Extension to C{L{scanString}}, to modify matching text with modified tokens that may
        be returned from a parse action.  To use C{transformString}, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking C{transformString()} on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  C{transformString()} returns the resulting transformed string.
        
        Example::
            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())
            
            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))
        Prints::
            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        rTNcSsg|]}|r|qSryry)roryryrzrsz1ParserElement.transformString..r)rrrr}r$rrrrr_flattenrr&r)rrRr*ZlastErxrrrXryryrzrs(



zParserElement.transformStringc
CsRztdd|||DWStk
rL}ztjr8n|W5d}~XYnXdS)a
        Another extension to C{L{scanString}}, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        C{maxMatches} argument, to clip searching after 'n' matches are found.
        
        Example::
            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())
            
            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))
        prints::
            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        cSsg|]\}}}|qSryry)rrxrrryryrzrsz.ParserElement.searchString..N)r$rrr&r)rrRrrXryryrzsearchStringszParserElement.searchStringc	csTd}d}|j||dD]*\}}}|||V|r<|dV|}q||dVdS)a[
        Generator method to split a string using the given expression as a separator.
        May be called with optional C{maxsplit} argument, to limit the number of splits;
        and the optional C{includeSeparators} argument (default=C{False}), if the separating
        matching text should be included in the split results.
        
        Example::        
            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))
        prints::
            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        r)rN)r)	rrRmaxsplitZincludeSeparatorsZsplitsZlastrxrrryryrzrs

zParserElement.splitcCsFt|trt|}t|ts:tjdt|tdddSt||gS)a
        Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement
        converts them to L{Literal}s by default.
        
        Example::
            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))
        Prints::
            Hello, World! -> ['Hello', ',', 'World', '!']
        4Cannot combine element of type %s with ParserElementrs
stacklevelN)	r}rr&rwwarningswarnr
SyntaxWarningrr"ryryrzrs


zParserElement.__add__cCsBt|trt|}t|ts:tjdt|tdddS||S)z]
        Implementation of + operator when left operand is not a C{L{ParserElement}}
        rrsrNr}rr&rwrrrrr"ryryrzr#1s


zParserElement.__radd__cCsJt|trt|}t|ts:tjdt|tdddS|t	|S)zQ
        Implementation of - operator, returns C{L{And}} with error stop
        rrsrN)
r}rr&rwrrrrr
_ErrorStopr"ryryrz__sub__=s


zParserElement.__sub__cCsBt|trt|}t|ts:tjdt|tdddS||S)z]
        Implementation of - operator when left operand is not a C{L{ParserElement}}
        rrsrNrr"ryryrz__rsub__Is


zParserElement.__rsub__cst|tr|d}}nt|tr|ddd}|ddkrHd|df}t|dtr|ddkr|ddkrvtS|ddkrtS|dtSqt|dtrt|dtr|\}}||8}qtdt|dt|dntdt||dkrtd|dkrtd	||kr6dkrBnntd
|rfdd|r|dkrt|}ntg||}n|}n|dkr}ntg|}|S)
a
        Implementation of * operator, allows use of C{expr * 3} in place of
        C{expr + expr + expr}.  Expressions may also me multiplied by a 2-integer
        tuple, similar to C{{min,max}} multipliers in regular expressions.  Tuples
        may also include C{None} as in:
         - C{expr*(n,None)} or C{expr*(n,)} is equivalent
              to C{expr*n + L{ZeroOrMore}(expr)}
              (read as "at least n instances of C{expr}")
         - C{expr*(None,n)} is equivalent to C{expr*(0,n)}
              (read as "0 to n instances of C{expr}")
         - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)}
         - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)}

        Note that C{expr*(None,n)} does not raise an exception if
        more than n exprs exist in the input stream; that is,
        C{expr*(None,n)} does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        C{expr*(None,n) + ~expr}
        r)NNNrsrz7cannot multiply 'ParserElement' and ('%s','%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez+cannot multiply ParserElement by 0 or (0,0)cs(|dkrt|dStSdSr)rnmakeOptionalListrryrzrsz/ParserElement.__mul__..makeOptionalList)	r}rvtupler4rrr
ValueErrorr)rrZminElementsZoptElementsrryrrz__mul__UsD







zParserElement.__mul__cCs
||Sr)rr"ryryrz__rmul__szParserElement.__rmul__cCsFt|trt|}t|ts:tjdt|tdddSt||gS)zI
        Implementation of | operator - returns C{L{MatchFirst}}
        rrsrN)	r}rr&rwrrrrrr"ryryrz__or__s


zParserElement.__or__cCsBt|trt|}t|ts:tjdt|tdddS||BS)z]
        Implementation of | operator when left operand is not a C{L{ParserElement}}
        rrsrNrr"ryryrz__ror__s


zParserElement.__ror__cCsFt|trt|}t|ts:tjdt|tdddSt||gS)zA
        Implementation of ^ operator - returns C{L{Or}}
        rrsrN)	r}rr&rwrrrrrr"ryryrz__xor__s


zParserElement.__xor__cCsBt|trt|}t|ts:tjdt|tdddS||AS)z]
        Implementation of ^ operator when left operand is not a C{L{ParserElement}}
        rrsrNrr"ryryrz__rxor__s


zParserElement.__rxor__cCsFt|trt|}t|ts:tjdt|tdddSt||gS)zC
        Implementation of & operator - returns C{L{Each}}
        rrsrN)	r}rr&rwrrrrrr"ryryrz__and__s


zParserElement.__and__cCsBt|trt|}t|ts:tjdt|tdddS||@S)z]
        Implementation of & operator when left operand is not a C{L{ParserElement}}
        rrsrNrr"ryryrz__rand__s


zParserElement.__rand__cCst|S)zE
        Implementation of ~ operator - returns C{L{NotAny}}
        )rrryryrz
__invert__szParserElement.__invert__cCs|dk	r||S|SdS)a

        Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}.
        
        If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be
        passed as C{True}.
           
        If C{name} is omitted, same as calling C{L{copy}}.

        Example::
            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums+"-")("socsecno")             
        N)rrrryryrz__call__s
zParserElement.__call__cCst|S)z
        Suppresses the output of this C{ParserElement}; useful to keep punctuation from
        cluttering up returned output.
        )r-rryryrzsuppressszParserElement.suppresscCs
d|_|S)a
        Disables the skipping of whitespace before matching the characters in the
        C{ParserElement}'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        Fr~rryryrzleaveWhitespaceszParserElement.leaveWhitespacecCsd|_||_d|_|S)z8
        Overrides the default whitespace chars
        TF)r~rr)rruryryrzsetWhitespaceChars
sz ParserElement.setWhitespaceCharscCs
d|_|S)z
        Overrides default behavior to expand C{}s to spaces before parsing the input string.
        Must be called before C{parseString} when the input grammar contains elements that
        match C{} characters.
        T)rrryryrz
parseWithTabsszParserElement.parseWithTabscCsLt|trt|}t|tr4||jkrH|j|n|jt||S)a
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.
        
        Example::
            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']
            
            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        )r}rr-rrrr"ryryrzignores


zParserElement.ignorecCs"|pt|pt|ptf|_d|_|S)zT
        Enable display of debugging messages while doing pattern matching.
        T)rTrWrYrr)rZstartActionZ
successActionZexceptionActionryryrzsetDebugActions6szParserElement.setDebugActionscCs|r|tttnd|_|S)a
        Enable display of debugging messages while doing pattern matching.
        Set C{flag} to True to enable, False to disable.

        Example::
            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer
            
            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")
        
        prints::
            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using L{setDebugActions}. Prior to attempting
        to match the C{wd} expression, the debugging message C{"Match  at loc (,)"}
        is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"}
        message is shown. Also note the use of L{setName} to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}.
        F)rrTrWrYr)rflagryryrzsetDebug@s#zParserElement.setDebugcCs|jSr)rrryryrzriszParserElement.__str__cCst|SrrrryryrzrlszParserElement.__repr__cCsd|_d|_|Sr)rr{rryryrzroszParserElement.streamlinecCsdSrryrryryrzcheckRecursiontszParserElement.checkRecursioncCs|gdS)zj
        Check defined expressions for valid structure, check for infinite recursive definitions.
        N)r)r
validateTraceryryrzvalidatewszParserElement.validatecCsz|}Wn2tk
r>t|d}|}W5QRXYnXz|||WStk
r~}ztjrjn|W5d}~XYnXdS)z
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        rN)readropenrrr&r)rZfile_or_filenamerZ
file_contentsfrXryryrz	parseFile}szParserElement.parseFilecsHt|tr"||kp t|t|kSt|tr6||Stt||kSdSr)r}r&varsrrsuperr"rmryrz__eq__s



zParserElement.__eq__cCs
||kSrryr"ryryrz__ne__szParserElement.__ne__cCstt|Sr)hashidrryryrz__hash__szParserElement.__hash__cCs||kSrryr"ryryrz__req__szParserElement.__req__cCs
||kSrryr"ryryrz__rne__szParserElement.__rne__cCs4z|jt||dWdStk
r.YdSXdS)a
        Method for quick testing of a parser against a test string. Good for simple 
        inline microtests of sub expressions while building up larger parser.
           
        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
            
        Example::
            expr = Word(nums)
            assert expr.matches("100")
        rTFN)rrr)rZ
testStringrryryrzrs

zParserElement.matches#cCst|tr"tttj|}t|tr4t|}g}g}d}	|D]}
|dk	r^|	|
dsf|rr|
sr|
|
qD|
sxqDd||
g}g}z:|
dd}
|j
|
|d}|
|j|d|	o|}	Wntk
rr}
zt|
trdnd	}d|
kr*|
t|
j|
|
d
t|
j|
dd|n|
d
|
jd||
d
t|
|	o\|}	|
}W5d}
~
XYnDtk
r}z$|
dt||	o|}	|}W5d}~XYnX|r|r|
d	td||
|
|fqD|	|fS)a3
        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.
           
        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests           
         - comment - (default=C{'#'}) - expression for indicating embedded comments in the test 
              string; pass None to disable comment filtering
         - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default=C{True}) prints test output to stdout
         - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if C{failureTests} is True), and the results contain a list of lines of each 
        test's output
        
        Example::
            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")
        prints::
            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success
            
            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\n of strings that spans \n 3 lines")
        
        (Note that this is a raw string literal, you must include the leading 'r'.)
        TNFr2\nr%)rDz(FATAL)r r^zFAIL: zFAIL-EXCEPTION: )r}rrrr~rrstrip
splitlinesrrrrrrrArr#rIrr;rprQ)rZtestsrZcommentZfullDumpZprintResultsZfailureTestsZ
allResultsZcommentssuccessrxr*resultrrrXryryrzrunTestssNW




$


zParserElement.runTests)F)F)T)T)TT)TT)r)F)N)T)F)T)Tr&TTF)Orrrrrsrstaticmethodrvrxrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr_MAX_INTrrrrrr#rrrrrrrrrrrr	r
rr
rrrrrrrrrrrrr"r#r$rr.
__classcell__ryryrrzr&Os




&




G

"
2G+D
			

)

cs eZdZdZfddZZS)r.zT
    Abstract C{ParserElement} subclass, for defining atomic matching patterns.
    cstt|jdddSNFr)rr.rrrryrzr@	szToken.__init__rrrrrr1ryryrrzr.<	scs eZdZdZfddZZS)rz,
    An empty token, will always match.
    cs$tt|d|_d|_d|_dS)NrTF)rrrrrrrrryrzrH	szEmpty.__init__r4ryryrrzrD	scs*eZdZdZfddZdddZZS)rz(
    A token that will never match.
    cs*tt|d|_d|_d|_d|_dS)NrTFzUnmatchable token)rrrrrrrrrryrzrS	s
zNoMatch.__init__TcCst|||j|dSr)r!rrryryrzrZ	szNoMatch.parseImpl)Trrrrrrr1ryryrrzrO	scs*eZdZdZfddZdddZZS)ra
    Token to exactly match a specified string.
    
    Example::
        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"
    
    For case-insensitive matching, use L{CaselessLiteral}.
    
    For keyword matching (force word break before and after the matched string),
    use L{Keyword} or L{CaselessKeyword}.
    cstt|||_t||_z|d|_Wn*tk
rVtj	dt
ddt|_YnXdt
|j|_d|j|_d|_d|_dS)Nrz2null string passed to Literal; use Empty() insteadrsr"%s"rF)rrrmatchrmatchLenfirstMatchCharrrrrrrmrrrrrrmatchStringrryrzrl	s
zLiteral.__init__TcCsJ|||jkr6|jdks&||j|r6||j|jfSt|||j|dSr)r9r8
startswithr7r!rrryryrzr	szLiteral.parseImpl)Tr5ryryrrzr^	s
csLeZdZdZedZdfdd	Zddd	Zfd
dZe	dd
Z
ZS)ra\
    Token to exactly match a specified string as a keyword, that is, it must be
    immediately followed by a non-keyword character.  Compare with C{L{Literal}}:
     - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}.
     - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'}
    Accepts two optional constructor arguments in addition to the keyword string:
     - C{identChars} is a string of characters that would be valid identifier characters,
          defaulting to all alphanumerics + "_" and "$"
     - C{caseless} allows case-insensitive matching, default is C{False}.
       
    Example::
        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use L{CaselessKeyword}.
    _$NFcstt||dkrtj}||_t||_z|d|_Wn$tk
r^t	j
dtddYnXd|j|_d|j|_
d|_d|_||_|r||_|}t||_dS)Nrz2null string passed to Keyword; use Empty() insteadrsrr6rF)rrrDEFAULT_KEYWORD_CHARSr7rr8r9rrrrrrrrcaselessupper
caselessmatchr
identChars)rr;rBr?rryrzr	s*

zKeyword.__init__TcCs|jr|||||j|jkr|t||jksL|||j|jkr|dksj||d|jkr||j|jfSnv|||jkr|jdks||j|r|t||jks|||j|jkr|dks||d|jkr||j|jfSt	|||j
|dSr)r?r8r@rArrBr7r9r<r!rrryryrzr	s4zKeyword.parseImplcstt|}tj|_|Sr)rrrr>rB)rrrryrzr	szKeyword.copycCs
|t_dS)z,Overrides the default Keyword chars
        N)rr>rtryryrzsetDefaultKeywordChars	szKeyword.setDefaultKeywordChars)NF)T)rrrrr5r>rrrr/rCr1ryryrrzr	s
cs*eZdZdZfddZdddZZS)r
al
    Token to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::
        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']
        
    (Contrast with example for L{CaselessKeyword}.)
    cs6tt||||_d|j|_d|j|_dS)Nz'%s'r)rr
rr@returnStringrrr:rryrzr	szCaselessLiteral.__init__TcCs@||||j|jkr,||j|jfSt|||j|dSr)r8r@r7rDr!rrryryrzr	szCaselessLiteral.parseImpl)Tr5ryryrrzr
	s
cs,eZdZdZdfdd	Zd	ddZZS)
r	z
    Caseless version of L{Keyword}.

    Example::
        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']
        
    (Contrast with example for L{CaselessLiteral}.)
    Ncstt|j||dddS)NTr?)rr	r)rr;rBrryrzr	szCaselessKeyword.__init__TcCsj||||j|jkrV|t||jksF|||j|jkrV||j|jfSt|||j|dSr)r8r@rArrBr7r!rrryryrzr	szCaselessKeyword.parseImpl)N)Tr5ryryrrzr		scs,eZdZdZdfdd	Zd	ddZZS)
rnax
    A variation on L{Literal} which matches "close" matches, that is, 
    strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters:
     - C{match_string} - string to be matched
     - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match
    
    The results from a successful parse will contain the matched text from the input string and the following named results:
     - C{mismatches} - a list of the positions within the match_string where mismatches were found
     - C{original} - the original match_string used to compare against the input string
    
    If C{mismatches} is an empty list, then the match was an exact match.
    
    Example::
        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    rcsBtt|||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F)	rrnrrmatch_string
maxMismatchesrrr)rrFrGrryrzr

szCloseMatch.__init__TcCs|}t|}|t|j}||kr|j}d}g}	|j}
tt||||jD]2\}}|\}}
||
krN|	|t|	|
krNqqN|d}t|||g}|j|d<|	|d<||fSt|||j|dS)Nrroriginal
mismatches)	rrFrGrrrr$r!r)rrRrrstartrmaxlocrFZmatch_stringlocrIrGZs_msrcmatresultsryryrzr
s( 

zCloseMatch.parseImpl)r)Tr5ryryrrzrn	s	cs8eZdZdZd
fdd	Zdd	d
ZfddZZS)r1a	
    Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters,
    an optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction. An optional
    C{excludeChars} parameter can list characters that might be found in 
    the input C{bodyChars} string; useful to define a word of all printables
    except for one or two characters, for instance.
    
    L{srange} is useful for defining custom character set strings for defining 
    C{Word} expressions, using range notation from regular expression character sets.
    
    A common mistake is to use C{Word} to match a specific literal string, as in 
    C{Word("Address")}. Remember that C{Word} uses the string argument to define
    I{sets} of matchable characters. This expression would match "Add", "AAA",
    "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'.
    To match an exact literal string, use L{Literal} or L{Keyword}.

    pyparsing includes helper strings for building Words:
     - L{alphas}
     - L{nums}
     - L{alphanums}
     - L{hexnums}
     - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.)
     - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - L{printables} (any non-whitespace character)

    Example::
        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))
        
        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums+'-')
        
        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")
        
        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    NrrFcstt|rFdfdd|D}|rFdfdd|D}||_t||_|rl||_t||_n||_t||_|dk|_	|dkrt
d||_|dkr||_nt
|_|dkr||_||_t||_d|j|_d	|_||_d
|j|jkr|dkr|dkr|dkr|j|jkr8dt|j|_nHt|jdkrfdt|jt|jf|_nd
t|jt|jf|_|jrd|jd|_zt|j|_Wntk
rd|_YnXdS)Nrc3s|]}|kr|VqdSrryrexcludeCharsryrzr`
sz Word.__init__..c3s|]}|kr|VqdSrryrrOryrzrb
srrzZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedrFr(z[%s]+z%s[%s]*z	[%s][%s]*z\b)rr1rr
initCharsOrigr	initChars
bodyCharsOrig	bodyCharsmaxSpecifiedrminLenmaxLenr0rrrr	asKeyword_escapeRegexRangeCharsreStringrrescapecompilerp)rrRrTminmaxexactrXrPrrOrzr]
s\



0
z
Word.__init__Tc
Cs>|jr<|j||}|s(t|||j||}||fS|||jkrZt|||j||}|d7}t|}|j}||j	}t
||}||kr|||kr|d7}qd}	|||jkrd}	|jr||kr|||krd}	|j
r|dkr||d|ks||kr|||krd}	|	r.t|||j|||||fS)NrFTr)rr7r!rendgrouprRrrTrWr]rVrUrX)
rrRrrr-rJrZ	bodycharsrKZthrowExceptionryryrzr
s6


2zWord.parseImplcsvztt|WStk
r$YnX|jdkrpdd}|j|jkr`d||j||jf|_nd||j|_|jS)NcSs$t|dkr|dddS|SdS)N...rrryryrz
charsAsStr
sz Word.__str__..charsAsStrz	W:(%s,%s)zW:(%s))rr1rrpr{rQrS)rrerryrzr
s
zWord.__str__)NrrrFN)Trrrrrrrr1ryryrrzr1.
s.6
#csFeZdZdZeedZdfdd	ZdddZ	fd	d
Z
ZS)
r)a
    Token for matching strings that match a given regular expression.
    Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module.
    If the given regex contains named groups (defined using C{(?P...)}), these will be preserved as 
    named parse results.

    Example::
        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P\d{4})-(?P\d\d?)-(?P\d\d?)')
        # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
    z[A-Z]rcstt|t|tr|s,tjdtdd||_||_	zt
|j|j	|_
|j|_Wqt
jk
rtjd|tddYqXn2t|tjr||_
t||_|_||_	ntdt||_d|j|_d|_d|_d	S)
zThe parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags.z0null string passed to Regex; use Empty() insteadrsr$invalid pattern (%s) passed to RegexzCRegex may only be constructed with a string or a compiled RE objectrFTN)rr)rr}rrrrpatternflagsrr\rZ
sre_constantserrorcompiledREtyper~rrrrrr)rrhrirryrzr
s:



zRegex.__init__TcCs`|j||}|s"t|||j||}|}t|}|rX|D]}||||<qF||fSr)rr7r!rr`	groupdictr$ra)rrRrrr-drrryryrzr
szRegex.parseImplcsFztt|WStk
r$YnX|jdkr@dt|j|_|jS)NzRe:(%s))rr)rrpr{rrhrrryrzr
s
z
Regex.__str__)r)T)rrrrrrr\rlrrrr1ryryrrzr)
s
"

cs8eZdZdZdfdd	ZdddZfd	d
ZZS)
r'a
    Token for matching strings that are delimited by quoting characters.
    
    Defined with the following parameters:
        - quoteChar - string of one or more characters defining the quote delimiting string
        - escChar - character to escape quotes, typically backslash (default=C{None})
        - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None})
        - multiline - boolean indicating whether quotes can span multiple lines (default=C{False})
        - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True})
        - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True})

    Example::
        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))
    prints::
        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    NFTc
sNtt|}|s0tjdtddt|dkr>|}n"|}|s`tjdtddt|_t	|_
|d_|_t	|_
|_|_|_|_|rtjtjB_dtjtjd|dk	rt|pdf_n.rt)z|(?:%s)z|(?:%s.)z(.)z)*%srgrFT)%rr'rrrrrSyntaxError	quoteCharrquoteCharLenfirstQuoteCharroendQuoteCharLenescCharescQuoteunquoteResultsconvertWhitespaceEscapesr	MULTILINEDOTALLrir[rYrhrrescCharReplacePatternr\rZrjrkrrrrr)rrrrvrwZ	multilinerxroryrrrzr/s|





zQuotedString.__init__c	Cs|||jkr|j||pd}|s4t|||j||}|}|jr||j|j	}t
|trd|kr|jrddddd}|
D]\}}|||}q|jrt|jd|}|jr||j|j}||fS)N\	r2
)\tr'z\fz\rz\g<1>)rtrr7r!rr`rarxrsrur}rryrrrvrr|rwro)	rrRrrr-rZws_mapZwslitZwscharryryrzrps* 
zQuotedString.parseImplcsHztt|WStk
r$YnX|jdkrBd|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)rr'rrpr{rrrorrryrzrs
zQuotedString.__str__)NNFTNT)Trfryryrrzr'sA
#cs8eZdZdZdfdd	ZdddZfd	d
ZZS)
ra
    Token for matching words composed of characters I{not} in a given set (will
    include whitespace in matched characters if not listed in the provided exclusion set - see example).
    Defined with string containing all disallowed characters, and an optional
    minimum, maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction.

    Example::
        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))
    prints::
        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    rrcstt|d|_||_|dkr*td||_|dkr@||_nt|_|dkrZ||_||_t	||_
d|j
|_|jdk|_d|_
dS)NFrzfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr)rrrr~notCharsrrVrWr0rrrrr)rrr]r^r_rryrzrs 
zCharsNotIn.__init__TcCs|||jkrt|||j||}|d7}|j}t||jt|}||krb|||krb|d7}qD|||jkrt|||j|||||fSr)rr!rr]rWrrV)rrRrrrJZnotcharsmaxlenryryrzrs

zCharsNotIn.parseImplcsfztt|WStk
r$YnX|jdkr`t|jdkrTd|jdd|_nd|j|_|jS)Nrbz
!W:(%s...)z!W:(%s))rrrrpr{rrrrryrzrs
zCharsNotIn.__str__)rrr)Trfryryrrzrs
cs<eZdZdZddddddZdfdd	ZdddZZS)r0a
    Special matching class for matching whitespace.  Normally, whitespace is ignored
    by pyparsing grammars.  This class is included when some whitespace structures
    are significant.  Define with a string containing the whitespace characters to be
    matched; default is C{" \t\r\n"}.  Also takes optional C{min}, C{max}, and C{exact} arguments,
    as defined for the C{L{Word}} class.
    zzzzz)r(r~r2rr 	
rrcstt|_dfddjDdddjD_d_dj_	|_
|dkrt|_nt_|dkr|_|_
dS)Nrc3s|]}|jkr|VqdSr)
matchWhiterrryrzrs
z!White.__init__..css|]}tj|VqdSr)r0	whiteStrsrryryrzrsTrr)
rr0rrr
rrrrrrVrWr0)rZwsr]r^r_rrrzrs zWhite.__init__TcCs|||jkrt|||j||}|d7}||j}t|t|}||krb|||jkrb|d7}qB|||jkrt|||j|||||fSr)rr!rrWr]rrV)rrRrrrJrKryryrzr	s

zWhite.parseImpl)rrrr)T)rrrrrrrr1ryryrrzr0scseZdZfddZZS)_PositionTokencs(tt||jj|_d|_d|_dSr)rrrrmrrrrrrryrzrs
z_PositionToken.__init__rrrrr1ryryrrzrsrcs2eZdZdZfddZddZd	ddZZS)
rzb
    Token to advance to a specific column of input text; useful for tabular report scraping.
    cstt|||_dSr)rrrr;)rcolnorryrzr$szGoToColumn.__init__cCs\t|||jkrXt|}|jr*|||}||krX||rXt|||jkrX|d7}q*|Sr)r;rrrisspace)rrRrrryryrzr(s$
zGoToColumn.preParseTcCsDt||}||jkr"t||d|||j|}|||}||fS)NzText not in expected columnr;r!)rrRrrZthiscolZnewlocrryryrzr1s

zGoToColumn.parseImpl)T)rrrrrrrr1ryryrrzr s	cs*eZdZdZfddZdddZZS)ra
    Matches if current position is at the beginning of a line within the parse string
    
    Example::
    
        test = '''        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)
    
    Prints::
        ['AAA', ' this line']
        ['AAA', ' and this line']    

    cstt|d|_dS)NzExpected start of line)rrrrrrryrzrOszLineStart.__init__TcCs*t||dkr|gfSt|||j|dSr)r;r!rrryryrzrSszLineStart.parseImpl)Tr5ryryrrzr:scs*eZdZdZfddZdddZZS)rzU
    Matches if current position is at the end of a line within the parse string
    cs,tt||tjddd|_dS)Nr2rzExpected end of line)rrrr
r&rsrrrrryrzr\szLineEnd.__init__TcCsb|t|kr6||dkr$|ddfSt|||j|n(|t|krN|dgfSt|||j|dS)Nr2rrr!rrryryrzraszLineEnd.parseImpl)Tr5ryryrrzrXscs*eZdZdZfddZdddZZS)r,zM
    Matches if current position is at the beginning of the parse string
    cstt|d|_dS)NzExpected start of text)rr,rrrrryrzrpszStringStart.__init__TcCs0|dkr(|||dkr(t|||j||gfSr)rr!rrryryrzrtszStringStart.parseImpl)Tr5ryryrrzr,lscs*eZdZdZfddZdddZZS)r+zG
    Matches if current position is at the end of the parse string
    cstt|d|_dS)NzExpected end of text)rr+rrrrryrzrszStringEnd.__init__TcCs^|t|krt|||j|n<|t|kr6|dgfS|t|krJ|gfSt|||j|dSrrrryryrzrszStringEnd.parseImpl)Tr5ryryrrzr+{scs.eZdZdZeffdd	ZdddZZS)r3ap
    Matches if the current position is at the beginning of a Word, and
    is not preceded by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{} behavior of regular expressions,
    use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of
    the string being parsed, or at the beginning of a line.
    cs"tt|t||_d|_dS)NzNot at the start of a word)rr3rr	wordCharsrrrrryrzrs
zWordStart.__init__TcCs@|dkr8||d|jks(|||jkr8t|||j||gfSr)rr!rrryryrzrszWordStart.parseImpl)TrrrrrXrrr1ryryrrzr3scs.eZdZdZeffdd	ZdddZZS)r2aZ
    Matches if the current position is at the end of a Word, and
    is not followed by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{} behavior of regular expressions,
    use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of
    the string being parsed, or at the end of a line.
    cs(tt|t||_d|_d|_dS)NFzNot at the end of a word)rr2rrrr~rrrryrzrs
zWordEnd.__init__TcCsPt|}|dkrH||krH|||jks8||d|jkrHt|||j||gfSr)rrr!r)rrRrrrryryrzrszWordEnd.parseImpl)Trryryrrzr2scseZdZdZdfdd	ZddZddZd	d
ZfddZfd
dZ	fddZ
dfdd	ZgfddZfddZ
ZS)r"z^
    Abstract subclass of ParserElement, for combining and post-processing parsed tokens.
    Fcstt||t|tr"t|}t|tr.F)rr"rr}rrrr&rwexprsrallrrrrrrrryrzrs


zParseExpression.__init__cCs
|j|Sr)rrryryrzrszParseExpression.__getitem__cCs|j|d|_|Sr)rrr{r"ryryrzrszParseExpression.appendcCs0d|_dd|jD|_|jD]}|q|S)z~Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on
           all contained expressions.FcSsg|]}|qSryrrrryryrzrsz3ParseExpression.leaveWhitespace..)r~rr)rrryryrzrs


zParseExpression.leaveWhitespacecsrt|trB||jkrntt|||jD]}||jdq*n,tt|||jD]}||jdqX|Sr)r}r-rrr"rr)rrrrryrzrs



zParseExpression.ignorecsNztt|WStk
r$YnX|jdkrHd|jjt|jf|_|jSNz%s:(%s))	rr"rrpr{rmrrrrrryrzrs
zParseExpression.__str__cs*tt||jD]}|qt|jdkr|jd}t||jr|js|jdkr|j	s|jdd|jdg|_d|_
|j|jO_|j|jO_|jd}t||jr|js|jdkr|j	s|jdd|jdd|_d|_
|j|jO_|j|jO_dt
||_|S)Nrsrrrtr)rr"rrrr}rmryr|rr{rrrr)rrrrryrzrs<



zParseExpression.streamlinecstt|||}|Sr)rr"r)rrrrrryrzr
szParseExpression.setResultsNamecCs6|dd|g}|jD]}||q|gdSr)rrr)rrtmprryryrzr
s
zParseExpression.validatecs$tt|}dd|jD|_|S)NcSsg|]}|qSryrrryryrzr%
sz(ParseExpression.copy..)rr"rrr1rryrzr#
szParseExpression.copy)F)F)rrrrrrrrrrrrrrr1ryryrrzr"s	
"csTeZdZdZGdddeZdfdd	ZdddZd	d
ZddZ	d
dZ
ZS)ra

    Requires all given C{ParseExpression}s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'+'} operator.
    May also be constructed using the C{'-'} operator, which will suppress backtracking.

    Example::
        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"),name_expr("name"),integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    cseZdZfddZZS)zAnd._ErrorStopcs&ttj|j||d|_|dS)N-)rrrrrrrGrryrzr9
szAnd._ErrorStop.__init__rryryrrzr8
srTcsRtt|||tdd|jD|_||jdj|jdj|_d|_	dS)Ncss|]}|jVqdSrrrryryrzr@
szAnd.__init__..rT)
rrrrrrr
rr~rrrryrzr>
s
zAnd.__init__c	Cs|jdj|||dd\}}d}|jddD]}t|tjrDd}q.|rz||||\}}Wqtk
rtYqtk
r}zd|_t|W5d}~XYqt	k
rt|t
||j|YqXn||||\}}|s|r.||7}q.||fS)NrFrrT)
rrr}rrr%r
__traceback__rrrrr
)	rrRrr
resultlistZ	errorStoprZ
exprtokensrryryrzrE
s(
z
And.parseImplcCst|trt|}||Srr}rr&rwrr"ryryrzr!^
s

zAnd.__iadd__cCs6|dd|g}|jD]}|||jsq2qdSr)rrrrrsubRecCheckListrryryrzrc
s


zAnd.checkRecursioncCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nr{r(css|]}t|VqdSrrrryryrzro
szAnd.__str__..}rrr{rrrryryrzrj
s


 zAnd.__str__)T)T)rrrrrrrrr!rrr1ryryrrzr(
s
csDeZdZdZdfdd	ZdddZdd	Zd
dZdd
ZZ	S)ra
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the expression that matches the longest string will be used.
    May be constructed using the C{'^'} operator.

    Example::
        # construct Or using '^' operator
        
        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))
    prints::
        [['123'], ['3.1416'], ['789']]
    Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|]}|jVqdSrrrryryrzr
szOr.__init__..T)rrrrrBrrrryrzr
szOr.__init__TcCsRd}d}g}|jD]}z|||}Wnvtk
rb}	zd|	_|	j|krR|	}|	j}W5d}	~	XYqtk
rt||krt|t||j|}t|}YqX|||fq|r(|j	ddd|D]^\}
}z|
|||WStk
r$}	z d|	_|	j|kr|	}|	j}W5d}	~	XYqXq|dk	r@|j|_|nt||d|dS)NrtcSs
|dSrry)xryryrzr{
r|zOr.parseImpl..)r no defined alternatives to match)rrr!rrrrrrsortrr)rrRrr	maxExcLocmaxExceptionrrZloc2r_ryryrzr
s<


zOr.parseImplcCst|trt|}||Srrr"ryryrz__ixor__
s

zOr.__ixor__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz ^ css|]}t|VqdSrrrryryrzr
szOr.__str__..rrrryryrzr
s


 z
Or.__str__cCs,|dd|g}|jD]}||qdSrrrrryryrzr
s
zOr.checkRecursion)F)T)
rrrrrrrrrr1ryryrrzrt
s

&	csDeZdZdZdfdd	ZdddZdd	Zd
dZdd
ZZ	S)ra
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the first one listed is the one that will match.
    May be constructed using the C{'|'} operator.

    Example::
        # construct MatchFirst using '|' operator
        
        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|]}|jVqdSrrrryryrzr
sz&MatchFirst.__init__..T)rrrrrBrrrryrzr
szMatchFirst.__init__Tc	Csd}d}|jD]}z||||}|WStk
r`}z|j|krP|}|j}W5d}~XYqtk
rt||krt|t||j|}t|}YqXq|dk	r|j|_|nt||d|dS)Nrtr)rrr!rrrrr)	rrRrrrrrrrryryrzr
s$


zMatchFirst.parseImplcCst|trt|}||Srrr"ryryrz__ior__
s

zMatchFirst.__ior__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrr | css|]}t|VqdSrrrryryrzr
sz%MatchFirst.__str__..rrrryryrzr
s


 zMatchFirst.__str__cCs,|dd|g}|jD]}||qdSrrrryryrzrs
zMatchFirst.checkRecursion)F)T)
rrrrrrrrrr1ryryrrzr
s
	cs<eZdZdZdfdd	ZdddZddZd	d
ZZS)
ram
    Requires all given C{ParseExpression}s to be found, but in any order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'&'} operator.

    Example::
        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order 
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )
    prints::
        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    Tcs8tt|||tdd|jD|_d|_d|_dS)Ncss|]}|jVqdSrrrryryrzr?sz Each.__init__..T)rrrrrrr~initExprGroupsrrryrzr=sz
Each.__init__c	s|jrtdd|jD|_dd|jD}dd|jD}|||_dd|jD|_dd|jD|_dd|jD|_|j|j7_d	|_|}|jdd}|jddg}d
}	|	rj||j|j}
g}|
D]v}z|||}Wn t	k
r|
|YqX|
|jt||||kr@|
|q|kr܈
|qt|t|
krd	}	q|rddd|D}
t	||d
|
|fdd|jD7}g}|D]"}||||\}}|
|qt|tg}||fS)Ncss&|]}t|trt|j|fVqdSr)r}rr!rSrryryrzrEs
z!Each.parseImpl..cSsg|]}t|tr|jqSryr}rrSrryryrzrFs
z"Each.parseImpl..cSs g|]}|jrt|ts|qSry)rr}rrryryrzrGs
cSsg|]}t|tr|jqSry)r}r4rSrryryrzrIs
cSsg|]}t|tr|jqSry)r}rrSrryryrzrJs
cSs g|]}t|tttfs|qSry)r}rr4rrryryrzrKsFTr%css|]}t|VqdSrrrryryrzrfsz*Missing one or more required elements (%s)cs$g|]}t|tr|jkr|qSryrrZtmpOptryrzrjs

)rrrZopt1mapZ	optionalsZmultioptionalsZ
multirequiredZrequiredrr!rrr!removerrrsumr$)rrRrrZopt1Zopt2ZtmpLocZtmpReqdZ
matchOrderZkeepMatchingZtmpExprsZfailedrZmissingrrNZfinalResultsryrrzrCsP

zEach.parseImplcCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz & css|]}t|VqdSrrrryryrzryszEach.__str__..rrrryryrzrts


 zEach.__str__cCs,|dd|g}|jD]}||qdSrrrryryrzr}s
zEach.checkRecursion)T)T)	rrrrrrrrr1ryryrrzrs
5
1	csleZdZdZdfdd	ZdddZdd	Zfd
dZfdd
ZddZ	gfddZ
fddZZS)r za
    Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens.
    Fcstt||t|tr@ttjtr2t|}ntt	|}||_
d|_|dk	r|j|_|j
|_
||j|j|_|j|_|j|_|j|jdSr)rr rr}r
issubclassr&rwr.rrSr{rrr
rr~r}rrrrrSrrryrzrs
zParseElementEnhance.__init__TcCs2|jdk	r|jj|||ddStd||j|dS)NFrr)rSrr!rrryryrzrs
zParseElementEnhance.parseImplcCs*d|_|j|_|jdk	r&|j|Sr)r~rSrrrryryrzrs


z#ParseElementEnhance.leaveWhitespacecsrt|trB||jkrntt|||jdk	rn|j|jdn,tt|||jdk	rn|j|jd|Sr)r}r-rrr rrSr"rryrzrs



zParseElementEnhance.ignorecs&tt||jdk	r"|j|Sr)rr rrSrrryrzrs

zParseElementEnhance.streamlinecCsB||krt||g|dd|g}|jdk	r>|j|dSr)r(rSr)rrrryryrzrs

z"ParseElementEnhance.checkRecursioncCs6|dd|g}|jdk	r(|j||gdSrrSrrrrrryryrzrs
zParseElementEnhance.validatecsXztt|WStk
r$YnX|jdkrR|jdk	rRd|jjt|jf|_|jSr)	rr rrpr{rSrmrrrrryrzrszParseElementEnhance.__str__)F)T)
rrrrrrrrrrrrr1ryryrrzr s
cs*eZdZdZfddZdddZZS)ra
    Lookahead matching of the given parse expression.  C{FollowedBy}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression matches at the current
    position.  C{FollowedBy} always returns a null token list.

    Example::
        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()
    prints::
        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    cstt||d|_dSr)rrrrrrSrryrzrszFollowedBy.__init__TcCs|j|||gfSr)rSrrryryrzrszFollowedBy.parseImpl)Tr5ryryrrzrscs2eZdZdZfddZd	ddZddZZS)
ra
    Lookahead to disallow matching with the given parse expression.  C{NotAny}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression does I{not} match at the current
    position.  Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny}
    always returns a null token list.  May be constructed using the '~' operator.

    Example::
        
    cs0tt||d|_d|_dt|j|_dS)NFTzFound unwanted token, )rrrr~rrrSrrrryrzrszNotAny.__init__TcCs&|j||rt|||j||gfSr)rSrr!rrryryrzrszNotAny.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrz~{rrrr{rrSrryryrzrs


zNotAny.__str__)Trfryryrrzrs

cs(eZdZdfdd	ZdddZZS)	_MultipleMatchNcsFtt||d|_|}t|tr.t|}|dk	r<|nd|_dSr)	rrrr}r}rr&rw	not_ender)rrSstopOnZenderrryrzrs

z_MultipleMatch.__init__Tc	Cs|jj}|j}|jdk	}|r$|jj}|r2|||||||dd\}}zV|j}	|r`||||	rp|||}
n|}
|||
|\}}|s|rR||7}qRWnttfk
rYnX||fSNFr)	rSrrrrrr
r!r)rrRrrZself_expr_parseZself_skip_ignorablesZcheck_enderZ
try_not_enderrZhasIgnoreExprsrZ	tmptokensryryrzrs*



z_MultipleMatch.parseImpl)N)T)rrrrrr1ryryrrzr
src@seZdZdZddZdS)ra
    Repetition of one or more of the given expression.
    
    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]
        
        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    cCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrrz}...rrryryrzrJs


zOneOrMore.__str__N)rrrrrryryryrzr0scs8eZdZdZd
fdd	Zdfdd	Zdd	ZZS)r4aw
    Optional repetition of zero or more of the given expression.
    
    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example: similar to L{OneOrMore}
    Ncstt|j||dd|_dS)N)rT)rr4rr)rrSrrryrzr_szZeroOrMore.__init__Tc	s<ztt||||WSttfk
r6|gfYSXdSr)rr4rr!rrrryrzrcszZeroOrMore.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrr$]...rrryryrzris


zZeroOrMore.__str__)N)Trfryryrrzr4Ssc@s eZdZddZeZddZdS)
_NullTokencCsdSrryrryryrzrssz_NullToken.__bool__cCsdSrryrryryrzrvsz_NullToken.__str__N)rrrrrLrryryryrzrrsrcs6eZdZdZeffdd	Zd	ddZddZZS)
raa
    Optional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::
        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345
            
            # ZIP+4 form
            12101-0001
            
            # invalid ZIP
            98765-
            ''')
    prints::
        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    cs.tt|j|dd|jj|_||_d|_dS)NFr3T)rrrrSr}rr)rrSrrryrzrs
zOptional.__init__Tc	Cszz|jj|||dd\}}WnTttfk
rp|jtk	rh|jjr^t|jg}|j||jj<ql|jg}ng}YnX||fSr)rSrr!rr_optionalNotMatchedr|r$)rrRrrrryryrzrs


zOptional.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrr$r'rrryryrzrs


zOptional.__str__)T)	rrrrrrrrr1ryryrrzrzs"
cs,eZdZdZd	fdd	Zd
ddZZS)r*a	
    Token for skipping over all undefined text until the matched expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default=C{False}) if True, the target expression is also parsed 
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default=C{None}) used to define grammars (typically quoted strings and 
          comments) that might contain false matches to the target expression
     - failOn - (default=C{None}) define expressions that are not allowed to be 
          included in the skipped test; if found before the target expression is found, 
          the SkipTo is not a match

    Example::
        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP 
                      + string_data("sev") + SEP 
                      + string_data("desc") + SEP 
                      + integer("days_open"))
        
        for tkt in ticket_expr.searchString(report):
            print tkt.dump()
    prints::
        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    FNcs`tt||||_d|_d|_||_d|_t|t	rFt
||_n||_dt
|j|_dS)NTFzNo match found for )rr*r
ignoreExprrrincludeMatchrr}rr&rwfailOnrrSr)rrZincluderrrryrzrs
zSkipTo.__init__Tc	Cs&|}t|}|j}|jj}|jdk	r,|jjnd}|jdk	rB|jjnd}	|}
|
|kr|dk	rf|||
rfq|	dk	rz|	||
}
Wqntk
rYqYqnXqnz|||
dddWqtt	fk
r|
d7}
YqJXqqJt|||j
||
}|||}t|}|jr||||dd\}}
||
7}||fS)NF)rrrr)
rrSrrrrrrr!rrr$r)rrRrrrUrrSZ
expr_parseZself_failOn_canParseNextZself_ignoreExpr_tryParseZtmplocZskiptextZ
skipresultrMryryrzrs:
zSkipTo.parseImpl)FNN)Tr5ryryrrzr*s6
csbeZdZdZdfdd	ZddZddZd	d
ZddZgfd
dZ	ddZ
fddZZS)raK
    Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator.

    Note: take care when assigning to C{Forward} not to overlook precedence of operators.
    Specifically, '|' has a lower precedence than '<<', so that::
        fwdExpr << a | b | c
    will actually be evaluated as::
        (fwdExpr << a) | b | c
    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the C{Forward}::
        fwdExpr << (a | b | c)
    Converting to use the '<<=' operator instead will avoid this problem.

    See L{ParseResults.pprint} for an example of a recursive parser created using
    C{Forward}.
    Ncstt|j|dddSr2)rrrr"rryrzr@szForward.__init__cCsjt|trt|}||_d|_|jj|_|jj|_||jj	|jj
|_
|jj|_|j
|jj|Sr)r}rr&rwrSr{rrr
rr~r}rrr"ryryrz
__lshift__Cs





zForward.__lshift__cCs||>Srryr"ryryrz__ilshift__PszForward.__ilshift__cCs
d|_|SrrrryryrzrSszForward.leaveWhitespacecCs$|js d|_|jdk	r |j|Sr)rrSrrryryrzrWs


zForward.streamlinecCs>||kr0|dd|g}|jdk	r0|j||gdSrrrryryrzr^s

zForward.validatecCsVt|dr|jS|jjdSz|jdk	r4t|j}nd}W5|j|_X|jjd|S)Nrz: ...Nonez: )rrrmrZ_revertClass_ForwardNoRecurserSr)rZ	retStringryryrzres


zForward.__str__cs.|jdk	rtt|St}||K}|SdSr)rSrrrr1rryrzrvs

zForward.copy)N)
rrrrrrrrrrrrr1ryryrrzr-s
c@seZdZddZdS)rcCsdS)Nrcryrryryrzrsz_ForwardNoRecurse.__str__N)rrrrryryryrzr~srcs"eZdZdZdfdd	ZZS)r/zQ
    Abstract subclass of C{ParseExpression}, for converting parsed results.
    Fcstt||d|_dSr)rr/rr}rrryrzrszTokenConverter.__init__)Fr4ryryrrzr/scs6eZdZdZd
fdd	ZfddZdd	ZZS)ra
    Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the input string;
    this can be disabled by specifying C{'adjacent=False'} in the constructor.

    Example::
        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    rTcs8tt|||r|||_d|_||_d|_dSr)rrrradjacentr~
joinStringr)rrSrrrryrzrszCombine.__init__cs(|jrt||ntt|||Sr)rr&rrrr"rryrzrszCombine.ignorecCsP|}|dd=|td||jg|jd7}|jrH|rH|gS|SdS)Nr)r)rr$rr(rrr|r
)rrRrrZretToksryryrzrs
"zCombine.postParse)rT)rrrrrrrr1ryryrrzrs
cs(eZdZdZfddZddZZS)ra
    Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions.

    Example::
        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a,b,100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a,b,100"))  # -> ['fn', ['a', 'b', '100']]
    cstt||d|_dSr)rrrr}rrryrzrszGroup.__init__cCs|gSrryrryryrzrszGroup.postParserrrrrrr1ryryrrzrs
cs(eZdZdZfddZddZZS)r
aW
    Converter to return a repetitive expression as a list, but also as a dictionary.
    Each element can also be referenced using the first token in the expression as its key.
    Useful for tabular report scraping when the first column can be used as a item key.

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())
        
        # access named fields as dict entries, or output as dict
        print(result['shape'])        
        print(result.asDict())
    prints::
        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}
    See more examples at L{ParseResults} of accessing fields by results name.
    cstt||d|_dSr)rr
rr}rrryrzrsz
Dict.__init__cCst|D]\}}t|dkrq|d}t|tr@t|d}t|dkr\td|||<qt|dkrt|dtst|d|||<q|}|d=t|dkst|tr|	rt||||<qt|d|||<q|j
r|gS|SdS)Nrrrrs)rrr}rvrrrr$rr
r|)rrRrrrtokZikeyZ	dictvalueryryrzrs$
zDict.postParserryryrrzr
s#c@s eZdZdZddZddZdS)r-aV
    Converter for ignoring the results of a parsed expression.

    Example::
        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))
    prints::
        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']
    (See also L{delimitedList}.)
    cCsgSrryrryryrzrszSuppress.postParsecCs|Srryrryryrzr
"szSuppress.suppressN)rrrrrr
ryryryrzr-sc@s(eZdZdZddZddZddZdS)	rzI
    Wrapper for parse actions, to ensure they are only called once.
    cCst||_d|_dSr)rrcallablecalled)rZ
methodCallryryrzr*s
zOnlyOnce.__init__cCs.|js||||}d|_|St||ddS)NTr)rrr!)rrr[rxrNryryrzr	-s
zOnlyOnce.__call__cCs
d|_dSr)rrryryrzreset3szOnlyOnce.resetN)rrrrrr	rryryryrzr&scs:tfdd}zj|_Wntk
r4YnX|S)at
    Decorator for debugging parse actions. 
    
    When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".}
    When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised.

    Example::
        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))
    prints::
        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <>entering %s(line: '%s', %d, %r)
z<.z)rrrr)rrryrrzrd6s
,FcCs`t|dt|dt|d}|rBt|t|||S|tt|||SdS)a
    Helper to define a delimited list of expressions - the delimiter defaults to ','.
    By default, the list elements and delimiters can have intervening whitespace, and
    comments, but this can be overridden by passing C{combine=True} in the constructor.
    If C{combine} is set to C{True}, the matching tokens are returned as a single token
    string, with the delimiters included; otherwise, the matching tokens are returned
    as a list of tokens, with the delimiters suppressed.

    Example::
        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    z [r(rN)rrr4rr-)rSZdelimcombineZdlNameryryrzrBbs
$csjtfdd}|dkr0ttdd}n|}|d|j|dd|d	td
S)a:
    Helper to define a counted list of expressions.
    This helper defines a pattern of the form::
        integer expr expr expr...
    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed.
    
    If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value.

    Example::
        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    cs.|d}|r ttg|p&tt>gSr)rrrE)rr[rxrZ	arrayExprrSryrzcountFieldParseActions"z+countedArray..countFieldParseActionNcSst|dSr)rvrwryryrzr{r|zcountedArray..ZarrayLenTrz(len) rc)rr1rTrrrrr)rSZintExprrryrrzr>us
cCs6g}|D](}t|tr&|t|q||q|Sr)r}rrrr)Lrrryryrzrs
rcs6tfdd}|j|dddt|S)a*
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches a
    previous literal, will also match the leading C{"1:1"} in C{"1:10"}.
    If this is not desired, use C{matchPreviousExpr}.
    Do I{not} use with packrat parsing enabled.
    csP|rBt|dkr|d>qLt|}tdd|D>n
t>dS)Nrrcss|]}t|VqdSr)rrZttryryrzrszDmatchPreviousLiteral..copyTokenToRepeater..)rrrrr)rr[rxZtflatZrepryrzcopyTokenToRepeatersz1matchPreviousLiteral..copyTokenToRepeaterTr(prev) )rrrr)rSrryrrzrQs


csFt|}|Kfdd}|j|dddt|S)aS
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches by
    expressions, will I{not} match the leading C{"1:1"} in C{"1:10"};
    the expressions are evaluated first, and then compared, so
    C{"1"} is compared with C{"10"}.
    Do I{not} use with packrat parsing enabled.
    cs*t|fdd}j|dddS)Ncs$t|}|kr tddddS)Nrr)rrr!)rr[rxZtheseTokensZmatchTokensryrzmustMatchTheseTokensszLmatchPreviousExpr..copyTokenToRepeater..mustMatchTheseTokensTr)rrr)rr[rxrrrrzrsz.matchPreviousExpr..copyTokenToRepeaterTrr)rrrrr)rSZe2rryrrzrPscCs:dD]}||t|}q|dd}|dd}t|S)Nz\^-]r2r'r~r)r_bslashr)rrryryrzrYs
rYTc
s|rdd}dd}tndd}dd}tg}t|trF|}n$t|trZt|}ntjdt	dd|stt
Sd	}|t|d
kr||}t||d
dD]R\}}	||	|r|||d
=qxq|||	r|||d
=|
||	|	}qxq|d
7}qx|s|rzlt|td|krTtd
ddd|Dd|WStddd|Dd|WSWn&tk
rtjdt	ddYnXtfdd|Dd|S)a
    Helper to quickly define a set of alternative Literals, and makes sure to do
    longest-first testing when there is a conflict, regardless of the input order,
    but returns a C{L{MatchFirst}} for best performance.

    Parameters:
     - strs - a string of space-delimited literals, or a collection of string literals
     - caseless - (default=C{False}) - treat all literals as caseless
     - useRegex - (default=C{True}) - as an optimization, will generate a Regex
          object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or
          if creating a C{Regex} raises an exception)

    Example::
        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))
    prints::
        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    cSs||kSr)r@rbryryrzr{r|zoneOf..cSs||Sr)r@r<rryryrzr{r|cSs||kSrryrryryrzr{r|cSs
||Sr)r<rryryrzr{r|z6Invalid argument to oneOf, expected string or iterablersrrrNrz[%s]css|]}t|VqdSr)rYrZsymryryrzrszoneOf..r|css|]}t|VqdSr)rr[rryryrzrsz7Exception creating Regex for oneOf, building MatchFirstc3s|]}|VqdSrryrZparseElementClassryrzr$s)r
rr}rrrrrrrrrrrrr)rrpr)
Zstrsr?ZuseRegexZisequalZmasksZsymbolsrZcurrrryrrzrUsT






**cCsttt||S)a
    Helper to easily and clearly define a dictionary by specifying the respective patterns
    for the key and value.  Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens
    in the proper order.  The key pattern can include delimiting markers or punctuation,
    as long as they are suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the C{Dict} results can include named token
    fields.

    Example::
        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())
    prints::
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    )r
r4r)rrryryrzrC&s!cCs^tdd}|}d|_|d||d}|r@dd}ndd}|||j|_|S)	a
    Helper to return the original, untokenized text for a given expression.  Useful to
    restore the parsed fields of an HTML start tag into the raw tag text itself, or to
    revert separate tokens with intervening whitespace back to the original matching
    input text. By default, returns astring containing the original parsed text.  
       
    If the optional C{asString} argument is passed as C{False}, then the return value is a 
    C{L{ParseResults}} containing any results names that were originally matched, and a 
    single token containing the original matched text from the input string.  So if 
    the expression passed to C{L{originalTextFor}} contains expressions with defined
    results names, you must set C{asString} to C{False} if you want to preserve those
    results name values.

    Example::
        src = "this is test  bold text  normal text "
        for tag in ("b","i"):
            opener,closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])
    prints::
        [' bold text ']
        ['text']
    cSs|Srry)rrrxryryrzr{ar|z!originalTextFor..F_original_start
_original_endcSs||j|jSr)rrrZryryrzr{fr|cSs&||d|dg|dd<dS)Nrr)rrZryryrzextractTexthsz$originalTextFor..extractText)rrrrr)rSZasStringZ	locMarkerZendlocMarker	matchExprrryryrzriIs

cCst|ddS)zp
    Helper to undo pyparsing's default grouping of And expressions, even
    if all but one are non-empty.
    cSs|dSrryrwryryrzr{sr|zungroup..)r/r)rSryryrzrjnscCs4tdd}t|d|d|dS)a
    Helper to decorate a returned token with its starting and ending locations in the input string.
    This helper adds the following results names:
     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains C{} characters, you may want to call
    C{L{ParserElement.parseWithTabs}}

    Example::
        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)
    prints::
        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    cSs|SrryrZryryrzr{r|zlocatedExpr..Z
locn_startrZlocn_end)rrrrr)rSZlocatorryryrzrlusz\[]-*.$+^?()~ r_cCs|ddSrryrZryryrzr{r|r{z\\0?[xX][0-9a-fA-F]+cCstt|dddS)Nrz\0x)unichrrvlstriprZryryrzr{r|z	\\0[0-7]+cCstt|ddddS)Nrr)rrvrZryryrzr{r|z\]rr$r)negatebodyr'csFddz"dfddt|jDWStk
r@YdSXdS)a
    Helper to easily define string ranges for use in Word construction.  Borrows
    syntax from regexp '[]' string range definitions::
        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"
    The input string must be enclosed in []'s, and the returned string is the expanded
    character set joined into a single string.
    The values enclosed in the []'s may be:
     - a single character
     - an escaped character with a leading backslash (such as C{\-} or C{\]})
     - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) 
         (C{\0x##} is also supported for backwards compatibility) 
     - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character)
     - a range of any of the above, separated by a dash (C{'a-z'}, etc.)
     - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.)
    cSs<t|ts|Sdddtt|dt|ddDS)Nrcss|]}t|VqdSr)rrryryrzrsz+srange....rr)r}r$rrord)pryryrzr{r|zsrange..rc3s|]}|VqdSrry)rpartZ	_expandedryrzrszsrange..N)r_reBracketExprrrrprdryrrzras
"csfdd}|S)zt
    Helper method for defining parse actions that require matching at a specific
    column in the input text.
    cs"t||krt||ddS)Nzmatched token not at column %dr)rNZlocnrVrryrz	verifyColsz!matchOnlyAtCol..verifyColry)rrryrrzrOscsfddS)a
    Helper method for common parse actions that simply return a literal value.  Especially
    useful when used with C{L{transformString}()}.

    Example::
        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num
        
        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    csgSrryrZZreplStrryrzr{r|zreplaceWith..ryrryrrzr^scCs|dddS)a
    Helper parse action for removing quotation marks from parsed quoted strings.

    Example::
        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    rrrtryrZryryrzr\scsNfdd}ztdtdj}Wntk
rBt}YnX||_|S)aG
    Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional 
    args are passed, they are forwarded to the given function as additional arguments after
    the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the
    parsed data to an integer using base 16.

    Example (compare the last to example in L{ParserElement.transformString}::
        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')
        
        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')
    prints::
        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    csfdd|DS)Ncsg|]}|fqSryry)rZtoknrr\ryrzrsz(tokenMap..pa..ryrZrryrzrsztokenMap..parrm)rorrpr~)r\rrrqryrrzros 
cCst|Srrr@rwryryrzr{r|cCst|Srrlowerrwryryrzr{r|c	Cst|tr|}t||d}n|j}tttd}|rt	t
}td|dtt
t|td|tddgdd		d
dtd}nd
ddtD}t	t
t|B}td|dtt
t|	tttd|tddgdd		ddtd}ttd|d}|dd
|ddd|}|dd
|ddd|}||_||_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namerEz_-:r5tag=/FrrEcSs|ddkSNrrryrZryryrzr{r|z_makeTags..r6rcss|]}|dkr|VqdS)r6Nryrryryrzrsz_makeTags..cSs|ddkSrryrZryryrzr{r|r7rJ:r(z<%s>r`z)r}rrrr1r6r5r@rrr\r-r
r4rrrrrXr[rDr_Lrtitlerrr)tagStrZxmlZresnameZtagAttrNameZtagAttrValueZopenTagZprintablesLessRAbrackZcloseTagryryrz	_makeTagss>
..rcCs
t|dS)a 
    Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches
    tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values.

    Example::
        text = 'More info at the pyparsing wiki page'
        # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple
        a,a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end
        
        for link in link_expr.searchString(text):
            # attributes in the  tag (like "href" shown here) are also accessible as named results
            print(link.link_text, '->', link.href)
    prints::
        pyparsing -> http://pyparsing.wikispaces.com
    FrrryryrzrM(scCs
t|dS)z
    Helper to construct opening and closing tag expressions for XML, given a tag name. Matches
    tags only in the given upper/lower case.

    Example: similar to L{makeHTMLTags}
    TrrryryrzrN;scs8|r|ddn|ddDfdd}|S)a<
    Helper to create a validating parse action to be used with start tags created
    with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag
    with a required attribute value, to avoid false matches on common tags such as
    C{} or C{
}. Call C{withAttribute} with a series of attribute names and values. Specify the list of filter attributes names and values as: - keyword arguments, as in C{(align="right")}, or - as an explicit dict with C{**} operator, when an attribute name is also a Python reserved word, as in C{**{"class":"Customer", "align":"right"}} - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") ) For attribute names with a namespace prefix, you must use the second form. Attribute names are matched insensitive to upper/lower case. If just testing for C{class} (with or without a namespace), use C{L{withClass}}. To verify that the attribute exists, but without specifying a value, pass C{withAttribute.ANY_VALUE} as the value. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this has no type
''' div,div_end = makeHTMLTags("div") # only match div tag having a type attribute with value "grid" div_grid = div().setParseAction(withAttribute(type="grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) # construct a match with any div tag having a type attribute, regardless of the value div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 NcSsg|]\}}||fqSryryr/ryryrzrzsz!withAttribute..csZD]P\}}||kr$t||d||tjkr|||krt||d||||fqdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r!rg ANY_VALUE)rr[rZattrNameZ attrValueZattrsryrzr{s  zwithAttribute..pa)r)rZattrDictrryrrzrgDs 2 cCs|r d|nd}tf||iS)a Simplified version of C{L{withAttribute}} when matching on a div class - made difficult because C{class} is a reserved word in Python. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this <div> has no class
''' div,div_end = makeHTMLTags("div") div_grid = div().setParseAction(withClass("grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 z%s:classclass)rg)Z classname namespaceZ classattrryryrzrms (rpcCst}||||B}t|D]l\}}|ddd\}} } } | dkrPd|nd|} | dkr|dkstt|dkr|td|\} }t| }| tjkr^| d krt||t|t |}n| dkr|dk rt|||t|t ||}nt||t|t |}nD| dkrTt|| |||t|| |||}ntd n| tj krB| d krt |t st |}t|j |t||}n| dkr|dk rt|||t|t ||}nt||t|t |}nD| dkr8t|| |||t|| |||}ntd ntd | rvt | ttfrl|j| n || ||| |BK}|}q||K}|S) aD Helper method for constructing grammars of expressions made up of operators working in a precedence hierarchy. Operators may be unary or binary, left- or right-associative. Parse actions can also be attached to operator expressions. The generated parser will also recognize the use of parentheses to override operator precedences (see example below). Note: if you define a deep operator list, you may see performance issues when using infixNotation. See L{ParserElement.enablePackrat} for a mechanism to potentially improve your parser performance. Parameters: - baseExpr - expression representing the most basic element for the nested - opList - list of tuples, one for each operator precedence level in the expression grammar; each tuple is of the form (opExpr, numTerms, rightLeftAssoc, parseAction), where: - opExpr is the pyparsing expression for the operator; may also be a string, which will be converted to a Literal; if numTerms is 3, opExpr is a tuple of two expressions, for the two operators separating the 3 terms - numTerms is the number of terms for this operator (must be 1, 2, or 3) - rightLeftAssoc is the indicator whether the operator is right or left associative, using the pyparsing-defined constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}. - parseAction is the parse action to be associated with expressions matching this operator expression (the parse action tuple member may be omitted); if the parse action is passed a tuple or list of functions, this is equivalent to calling C{setParseAction(*fn)} (L{ParserElement.setParseAction}) - lpar - expression for matching left-parentheses (default=C{Suppress('(')}) - rpar - expression for matching right-parentheses (default=C{Suppress(')')}) Example:: # simple example of four-function arithmetic with ints and variable names integer = pyparsing_common.signed_integer varname = pyparsing_common.identifier arith_expr = infixNotation(integer | varname, [ ('-', 1, opAssoc.RIGHT), (oneOf('* /'), 2, opAssoc.LEFT), (oneOf('+ -'), 2, opAssoc.LEFT), ]) arith_expr.runTests(''' 5+3*6 (5+3)*6 -2--11 ''', fullDump=False) prints:: 5+3*6 [[5, '+', [3, '*', 6]]] (5+3)*6 [[[5, '+', 3], '*', 6]] -2--11 [[['-', 2], '-', ['-', 11]]] rNrbrqz%s termz %s%s termrsz@if numterms=3, opExpr must be a tuple or list of two expressionsrz6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)rrrrrrVLEFTrrrRIGHTr}rrSrrr)ZbaseExprZopListZlparZrparrZlastExprrZoperDefZopExprZarityZrightLeftAssocrZtermNameZopExpr1ZopExpr2ZthisExprrryryrzrksZ=   &       &    z4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*'z string enclosed in single quotesz*quotedString using single or double quotesuzunicode string literalcCs||krtd|dkr*t|tr"t|tr"t|dkrt|dkr|dk rtt|t||tjdd dd}n$t t||tj dd}nx|dk rtt|t |t |ttjdd dd}n4ttt |t |ttjdd d d}ntd t }|dk rd|tt|t||B|Bt|K}n$|tt|t||Bt|K}|d ||f|S) a~ Helper method for defining nested lists enclosed in opening and closing delimiters ("(" and ")" are the default). Parameters: - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression - content - expression for items within the nested lists (default=C{None}) - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString}) If an expression is not provided for the content argument, the nested expression will capture all whitespace-delimited content between delimiters as a list of separate values. Use the C{ignoreExpr} argument to define expressions that may contain opening or closing characters that should not be treated as opening or closing characters for nesting, such as quotedString or a comment expression. Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}. The default is L{quotedString}, but if no expressions are to be ignored, then pass C{None} for this argument. Example:: data_type = oneOf("void int short long char float double") decl_data_type = Combine(data_type + Optional(Word('*'))) ident = Word(alphas+'_', alphanums+'_') number = pyparsing_common.number arg = Group(decl_data_type + ident) LPAR,RPAR = map(Suppress, "()") code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment)) c_function = (decl_data_type("type") + ident("name") + LPAR + Optional(delimitedList(arg), [])("args") + RPAR + code_body("body")) c_function.ignore(cStyleComment) source_code = ''' int is_odd(int x) { return (x%2); } int dec_to_hex(char hchar) { if (hchar >= '0' && hchar <= '9') { return (ord(hchar)-ord('0')); } else { return (10+ord(hchar)-ord('A')); } } ''' for func in c_function.searchString(source_code): print("%(name)s (%(type)s) args: %(args)s" % func) prints:: is_odd (int) args: [['int', 'x']] dec_to_hex (int) args: [['char', 'hchar']] z.opening and closing strings cannot be the sameNrrcSs |dSrrrwryryrzr{gr|znestedExpr..cSs |dSrr rwryryrzr{jr|cSs |dSrr rwryryrzr{pr|cSs |dSrr rwryryrzr{tr|zOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)rr}rrr rr r&rsrrErrrrr-r4r)ZopenerZcloserZcontentrrryryrzrR%sH:    *$c sfdd}fdd}fdd}ttd}tt|d}t|d }t|d } |rtt||t|t|t|| } n$tt|t|t|t|} | t t| d S) a Helper method for defining space-delimited indentation blocks, such as those used to define block statements in Python source code. Parameters: - blockStatementExpr - expression defining syntax of statement that is repeated within the indented block - indentStack - list created by caller to manage indentation stack (multiple statementWithIndentedBlock expressions within a single grammar should share a common indentStack) - indent - boolean indicating whether block must be indented beyond the the current level; set to False for block of left-most statements (default=C{True}) A valid block must contain at least one C{blockStatement}. Example:: data = ''' def A(z): A1 B = 100 G = A2 A2 A3 B def BB(a,b,c): BB1 def BBA(): bba1 bba2 bba3 C D def spam(x,y): def eggs(z): pass ''' indentStack = [1] stmt = Forward() identifier = Word(alphas, alphanums) funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":") func_body = indentedBlock(stmt, indentStack) funcDef = Group( funcDecl + func_body ) rvalue = Forward() funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")") rvalue << (funcCall | identifier | Word(nums)) assignment = Group(identifier + "=" + rvalue) stmt << ( funcDef | assignment | identifier ) module_body = OneOrMore(stmt) parseTree = module_body.parseString(data) parseTree.pprint() prints:: [['def', 'A', ['(', 'z', ')'], ':', [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]], 'B', ['def', 'BB', ['(', 'a', 'b', 'c', ')'], ':', [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]], 'C', 'D', ['def', 'spam', ['(', 'x', 'y', ')'], ':', [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] csN|t|krdSt||}|dkrJ|dkr>t||dt||ddS)Nrtzillegal nestingznot a peer entry)rr;r#r!rr[rxZcurCol indentStackryrzcheckPeerIndents     z&indentedBlock..checkPeerIndentcs2t||}|dkr"|n t||ddS)Nrtznot a subentry)r;rr!rrryrzcheckSubIndents   z%indentedBlock..checkSubIndentcsN|t|krdSt||}r6|dkr6|dksBt||ddS)Nrtr_znot an unindent)rr;r!rrrryrz checkUnindents    z$indentedBlock..checkUnindentz INDENTrZUNINDENTzindented block) rrr r rrrrrrr) ZblockStatementExprrr9rrrrErZPEERZUNDENTZsmExprryrrzrhs(N   z#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z &(?Prz);zcommon HTML entitycCs t|jS)zRHelper parser action to replace common HTML entities with their special characters)_htmlEntityMaprZentityrwryryrzr]sz/\*(?:[^*]|\*(?!/))*z*/zC style commentzz HTML commentz.*z rest of linez//(?:\\\n|[^\n])*z // commentzC++ style commentz#.*zPython style commentrO commaItemrc@seZdZdZeeZeeZe e  d eZ e e d eedZed d eZe ede e dZed d eeeed eB d Zeeed  d eZed d eZeeBeBZed d eZe eded dZed dZ ed dZ!e!de!d dZ"ee!de!ddee!de!d dZ#e#$dd d e  d!Z%e&e"e%Be#B d" d"Z'ed# d$Z(e)d=d&d'Z*e)d>d)d*Z+ed+ d,Z,ed- d.Z-ed/ d0Z.e/e0BZ1e)d1d2Z2e&e3e4d3e5e e6d3d4ee7d5 d6Z8e9ee:;e8Bd7d8 d9Zd}, L{reals}, L{scientific notation}) - common L{programming identifiers} - network addresses (L{MAC}, L{IPv4}, L{IPv6}) - ISO8601 L{dates} and L{datetime} - L{UUID} - L{comma-separated list} Parse actions: - C{L{convertToInteger}} - C{L{convertToFloat}} - C{L{convertToDate}} - C{L{convertToDatetime}} - C{L{stripHTMLTags}} - C{L{upcaseTokens}} - C{L{downcaseTokens}} Example:: pyparsing_common.number.runTests(''' # any int or real number, returned as the appropriate type 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.fnumber.runTests(''' # any int or real number, returned as float 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.hex_integer.runTests(''' # hex numbers 100 FF ''') pyparsing_common.fraction.runTests(''' # fractions 1/2 -3/4 ''') pyparsing_common.mixed_integer.runTests(''' # mixed fractions 1 1/2 -3/4 1-3/4 ''') import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(''' # uuid 12345678-1234-5678-1234-567812345678 ''') prints:: # any int or real number, returned as the appropriate type 100 [100] -100 [-100] +100 [100] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # any int or real number, returned as float 100 [100.0] -100 [-100.0] +100 [100.0] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # hex numbers 100 [256] FF [255] # fractions 1/2 [0.5] -3/4 [-0.75] # mixed fractions 1 [1] 1/2 [0.5] -3/4 [-0.75] 1-3/4 [1.75] # uuid 12345678-1234-5678-1234-567812345678 [UUID('12345678-1234-5678-1234-567812345678')] integerz hex integerrz[+-]?\d+zsigned integerrfractioncCs|d|dS)Nrrtryrwryryrzr{r|zpyparsing_common.rz"fraction or mixed integer-fractionz [+-]?\d+\.\d*z real numberz+[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?fnumberr identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}z IPv4 addressz[0-9a-fA-F]{1,4} hex_integerrzfull IPv6 address)rrhz::zshort IPv6 addresscCstdd|DdkS)Ncss|]}tj|rdVqdSr3)rp _ipv6_partrrryryrzrs z,pyparsing_common...r)rrwryryrzr{r|z::ffff:zmixed IPv6 addressz IPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}z MAC address%Y-%m-%dcsfdd}|S)a Helper to create a parse action for converting parsed date string to Python datetime.date Params - - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"}) Example:: date_expr = pyparsing_common.iso8601_date.copy() date_expr.setParseAction(pyparsing_common.convertToDate()) print(date_expr.parseString("1999-12-31")) prints:: [datetime.date(1999, 12, 31)] c sNzt|dWStk rH}zt||t|W5d}~XYnXdSr)rstrptimedaterr!r~rr[rxZvefmtryrzcvt_fnsz.pyparsing_common.convertToDate..cvt_fnryr$r%ryr#rz convertToDates zpyparsing_common.convertToDate%Y-%m-%dT%H:%M:%S.%fcsfdd}|S)a Helper to create a parse action for converting parsed datetime string to Python datetime.datetime Params - - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"}) Example:: dt_expr = pyparsing_common.iso8601_datetime.copy() dt_expr.setParseAction(pyparsing_common.convertToDatetime()) print(dt_expr.parseString("1999-12-31T23:59:59.999")) prints:: [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)] c sJzt|dWStk rD}zt||t|W5d}~XYnXdSr)rr rr!r~r"r#ryrzr%sz2pyparsing_common.convertToDatetime..cvt_fnryr&ryr#rzconvertToDatetimes z"pyparsing_common.convertToDatetimez7(?P\d{4})(?:-(?P\d\d)(?:-(?P\d\d))?)?z ISO8601 datez(?P\d{4})-(?P\d\d)-(?P\d\d)[T ](?P\d\d):(?P\d\d)(:(?P\d\d(\.\d*)?)?)?(?PZ|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}UUIDcCstj|dS)a Parse action to remove HTML tags from web page HTML source Example:: # strip HTML links from normal text text = 'More info at the
pyparsing wiki page' td,td_end = makeHTMLTags("TD") table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page' r)rp_html_stripperr)rr[rryryrz stripHTMLTagss zpyparsing_common.stripHTMLTagsrrOrrrrzcomma separated listcCs t|Srrrwryryrzr{"r|cCs t|Srrrwryryrzr{%r|N)r)r()?rrrrrorvZconvertToIntegerfloatZconvertToFloatr1rTrrrrFrr)Zsigned_integerrrrr Z mixed_integerrrealZsci_realrnumberrr6r5rZ ipv4_addressrZ_full_ipv6_addressZ_short_ipv6_addressrZ_mixed_ipv6_addressr Z ipv6_addressZ mac_addressr/r'r)Z iso8601_dateZiso8601_datetimeuuidr9r8r+r,rrrrXr0 _commasepitemrBr[rZcomma_separated_listrfrDryryryrzrpsV"" 2     __main__Zselectfromr=r)rcolumnsrZtablesZcommandaK # '*' as column list and dotted table name select * from SYS.XYZZY # caseless match on "SELECT", and casts back to "select" SELECT * from XYZZY, ABC # list of column names, and mixed case SELECT keyword Select AA,BB,CC from Sys.dual # multiple tables Select A, B, C from Sys.dual, Table2 # invalid SELECT keyword - should fail Xelect A, B, C from Sys.dual # incomplete command - should fail Select # invalid column name - should fail Select ^^^ frox Sys.dual z] 100 -100 +100 3.14159 6.02e23 1e-12 z 100 FF z6 12345678-1234-5678-1234-567812345678 )rs)rF)N)FT)T)r)T)r __version__Z__versionTime__ __author__rweakrefrrrrrrrjrrFrcrr_threadr ImportErrorZ threadingZcollections.abcrrrrZ ordereddict__all__r version_inforbr0maxsizer0r~rchrrrrrr@reversedrrrBrr]r^rnZmaxintZxrangerZ __builtin__rZfnamerrorrrrrrZascii_uppercaseZascii_lowercaser6rTrFr5rrZ printablerXrprr!r#r%r(rr$registerr;rLrIrTrWrYrSrrr&r.rrrrrwrr r rnr1r)r'r r0rrrrr,r+r3r2r"rrrrr rrrrr4rrrr*rrr/r rr r-rrdrBr>rrQrPrYrUrCrirjrlrrErKrJrcrbrZ _escapedPuncZ_escapedHexCharZ_escapedOctCharZ _singleCharZ _charRangerrrarOr^r\rorfrDrrMrNrgrrmrVrr rkrWr@r`r[rerRrhr7rYr9r8rrrrr=r]r:rGr r_rAr?rHrZrr1r<rprZ selectTokenZ fromTokenZidentZ columnNameZcolumnNameListZ columnSpecZ tableNameZ tableNameListZ simpleSQLr.r/rrr0r*ryryryrzs4        8      @v &A= I G3pLOD|M &#@sQ,A,    I# %     0 ,   ? #p  Zr   (         "   site-packages/setuptools/_vendor/__pycache__/__init__.cpython-38.pyc000064400000000232151464452710021546 0ustar00U Qab@sdS)Nrrr?/usr/lib/python3.8/site-packages/setuptools/_vendor/__init__.pysite-packages/setuptools/_vendor/__pycache__/six.cpython-38.opt-1.pyc000064400000057553151464452710021573 0ustar00U QabuA@sRdZddlmZddlZddlZddlZddlZddlZdZdZ ej ddkZ ej ddkZ ej dddkZ e refZefZefZeZeZejZn~efZeefZeejfZeZeZejd red ZnHGd d d eZ ze!e Wne"k red ZYn Xed Z[ ddZ#ddZ$GdddeZ%Gddde%Z&Gdddej'Z(Gddde%Z)GdddeZ*e*e+Z,Gddde(Z-e)dddd e)d!d"d#d$d!e)d%d"d"d&d%e)d'd(d#d)d'e)d*d(d+e)d,d"d#d-d,e)d.d/d/d0d.e)d1d/d/d.d1e)d2d(d#d3d2e)d4d(e rd5nd6d7e)d8d(d9e)d:d;dd>d?e)d@d@d?e)dAdAd?e)d3d(d#d3d2e)dBd"d#dCdBe)dDd"d"dEdDe&d#d(e&dFdGe&dHdIe&dJdKdLe&dMdNdMe&dOdPdQe&dRdSdTe&dUdVdWe&dXdYdZe&d[d\d]e&d^d_d`e&dadbdce&dddedfe&dgdhdie&djdjdke&dldldke&dmdmdke&dndndoe&dpdqe&drdse&dtdue&dvdwdve&dxdye&dzd{d|e&d}d~de&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&de+dde&de+dde&de+de+de&ddde&ddde&dddg>Z.ejdkrRe.e&ddg7Z.e.D]2Z/e0e-e/j1e/e2e/e&rVe,3e/de/j1qV[/e.e-_.e-e+dZ4e,3e4dGddde(Z5e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d=dde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)dddgZ6e6D]Z/e0e5e/j1e/q[/e6e5_.e,3e5e+dddҡGddԄde(Z7e)ddde)ddde)dddgZ8e8D]Z/e0e7e/j1e/q[/e8e7_.e,3e7e+dddۡGdd݄de(Z9e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃g!Z:e:D]Z/e0e9e/j1e/q[/e:e9_.e,3e9e+dddGddde(Z;e)ddde)ddde)ddde)d ddgZe>D]Z/e0e=e/j1e/q[/e>e=_.e,3e=e+dddGdddej'Z?e,3e?e+ddddZ@ddZAe rHdZBdZCdZDdZEdZFdZGn$d ZBd!ZCd"ZDd#ZEd$ZFd%ZGzeHZIWn"eJk rd&d'ZIYnXeIZHzeKZKWn"eJk rd(d)ZKYnXe rd*d+ZLejMZNd,d-ZOeZPn>d.d+ZLd/d0ZNd1d-ZOGd2d3d3eZPeKZKe#eLd4eQeBZReQeCZSeQeDZTeQeEZUeQeFZVeQeGZWe rԐd5d6ZXd7d8ZYd9d:ZZd;d<Z[e\d=Z]e\d>Z^e\d?Z_nTd@d6ZXdAd8ZYdBd:ZZdCd<Z[e\dDZ]e\dEZ^e\dFZ_e#eXdGe#eYdHe#eZdIe#e[dJe rdKdLZ`dMdNZaebZcddldZdededOjfZg[dehdZiejjZkelZmddlnZnenjoZoenjpZpdPZqej dQdQk rdRZrdSZsn dTZrdUZsnjdVdLZ`dWdNZaecZcebZgdXdYZidZd[ZketejuevZmddloZoeojoZoZpd\ZqdRZrdSZse#e`d]e#ead^d_dPZwd`dTZxdadUZye reze4j{dbZ|d|dcddZ}nd}dedfZ|e|dgej dddhk re|din.ej dddhk re|djn dkdlZ~eze4j{dmdZedk rLdndoZej dddpk rreZdqdoZe#e}drej dddk rejejfdsdtZnejZdudvZdwdxZdydzZgZe+Zed{dk rge_ejrBeejD]4\ZZeej+dkrej1e+kreje=q>q[[eje,dS(~z6Utilities for writing code that runs on Python 2 and 3)absolute_importNz'Benjamin Peterson z1.10.0)rjavaic@seZdZddZdS)XcCsdS)Nlselfrr:/usr/lib/python3.8/site-packages/setuptools/_vendor/six.py__len__>sz X.__len__N)__name__ __module__ __qualname__r rrrr r<srlcCs ||_dS)z Add documentation to a function.N)__doc__)funcdocrrr _add_docKsrcCst|tj|S)z7Import module, returning the module after the last dot.) __import__sysmodulesnamerrr _import_modulePsrc@seZdZddZddZdS) _LazyDescrcCs ||_dSNrr rrrr __init__Xsz_LazyDescr.__init__cCsB|}t||j|zt|j|jWntk r<YnX|Sr)_resolvesetattrrdelattr __class__AttributeError)r objtpresultrrr __get__[sz_LazyDescr.__get__N)r rrrr&rrrr rVsrcs.eZdZdfdd ZddZddZZS) MovedModuleNcs2tt||tr(|dkr |}||_n||_dSr)superr'rPY3mod)r roldnewr!rr ris zMovedModule.__init__cCs t|jSr)rr*r rrr rrszMovedModule._resolvecCs"|}t||}t||||Sr)rgetattrr)r attr_modulevaluerrr __getattr__us  zMovedModule.__getattr__)N)r rrrrr2 __classcell__rrr-r r'gs r'cs(eZdZfddZddZgZZS) _LazyModulecstt|||jj|_dSr)r(r4rr!rrr-rr r~sz_LazyModule.__init__cCs ddg}|dd|jD7}|S)Nrr cSsg|] }|jqSrr).0r/rrr sz'_LazyModule.__dir__..)_moved_attributes)r Zattrsrrr __dir__sz_LazyModule.__dir__)r rrrr8r7r3rrr-r r4|s r4cs&eZdZdfdd ZddZZS)MovedAttributeNcsdtt||trH|dkr |}||_|dkr@|dkr<|}n|}||_n||_|dkrZ|}||_dSr)r(r9rr)r*r/)r rZold_modZnew_modZold_attrZnew_attrr-rr rszMovedAttribute.__init__cCst|j}t||jSr)rr*r.r/)r modulerrr rs zMovedAttribute._resolve)NN)r rrrrr3rrr-r r9sr9c@sVeZdZdZddZddZddZdd d Zd d Zd dZ ddZ ddZ e Z dS)_SixMetaPathImporterz A meta path importer to import six.moves and its submodules. This class implements a PEP302 finder and loader. It should be compatible with Python 2.5 and all existing versions of Python3 cCs||_i|_dSr)r known_modules)r Zsix_module_namerrr rsz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdSN.r<r)r r*Z fullnamesfullnamerrr _add_modulesz _SixMetaPathImporter._add_modulecCs|j|jd|Sr=r?r r@rrr _get_modulesz _SixMetaPathImporter._get_moduleNcCs||jkr|SdSr)r<)r r@pathrrr find_modules z _SixMetaPathImporter.find_modulecCs2z |j|WStk r,td|YnXdS)Nz!This loader does not know module )r<KeyError ImportErrorrBrrr Z __get_modules z!_SixMetaPathImporter.__get_modulecCsTz tj|WStk r YnX||}t|tr@|}n||_|tj|<|Sr)rrrF _SixMetaPathImporter__get_module isinstancer'r __loader__)r r@r*rrr load_modules     z _SixMetaPathImporter.load_modulecCst||dS)z Return true, if the named module is a package. We need this method to get correct spec objects with Python 3.4 (see PEP451) __path__)hasattrrHrBrrr is_packagesz_SixMetaPathImporter.is_packagecCs||dS)z;Return None Required, if is_package is implementedN)rHrBrrr get_codes z_SixMetaPathImporter.get_code)N) r rrrrrArCrErHrKrNrO get_sourcerrrr r;s  r;c@seZdZdZgZdS) _MovedItemszLazy loading of moved objectsN)r rrrrLrrrr rQsrQZ cStringIOioStringIOfilter itertoolsbuiltinsZifilter filterfalseZ ifilterfalseinputZ __builtin__Z raw_inputinternrmapimapgetcwdosZgetcwdugetcwdbrangeZxrangeZ reload_module importlibZimpreloadreduce functoolsZ shlex_quoteZpipesZshlexZquoteUserDict collectionsUserList UserStringzipZizip zip_longestZ izip_longestZ configparserZ ConfigParsercopyregZcopy_regZdbm_gnuZgdbmzdbm.gnuZ _dummy_threadZ dummy_threadZhttp_cookiejarZ cookielibzhttp.cookiejarZ http_cookiesZCookiez http.cookiesZ html_entitiesZhtmlentitydefsz html.entitiesZ html_parserZ HTMLParserz html.parserZ http_clientZhttplibz http.clientZemail_mime_multipartzemail.MIMEMultipartzemail.mime.multipartZemail_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipartZemail_mime_textzemail.MIMETextzemail.mime.textZemail_mime_basezemail.MIMEBasezemail.mime.baseZBaseHTTPServerz http.serverZ CGIHTTPServerZSimpleHTTPServerZcPicklepickleZqueueZQueuereprlibreprZ socketserverZ SocketServer_threadthreadZtkinterZTkinterZtkinter_dialogZDialogztkinter.dialogZtkinter_filedialogZ FileDialogztkinter.filedialogZtkinter_scrolledtextZ ScrolledTextztkinter.scrolledtextZtkinter_simpledialogZ SimpleDialogztkinter.simpledialogZ tkinter_tixZTixz tkinter.tixZ tkinter_ttkZttkz tkinter.ttkZtkinter_constantsZ Tkconstantsztkinter.constantsZ tkinter_dndZTkdndz tkinter.dndZtkinter_colorchooserZtkColorChooserztkinter.colorchooserZtkinter_commondialogZtkCommonDialogztkinter.commondialogZtkinter_tkfiledialogZ tkFileDialogZ tkinter_fontZtkFontz tkinter.fontZtkinter_messageboxZ tkMessageBoxztkinter.messageboxZtkinter_tksimpledialogZtkSimpleDialogZ urllib_parsez.moves.urllib_parsez urllib.parseZ urllib_errorz.moves.urllib_errorz urllib.errorZurllibz .moves.urllibZurllib_robotparser robotparserzurllib.robotparserZ xmlrpc_clientZ xmlrpclibz xmlrpc.clientZ xmlrpc_serverZSimpleXMLRPCServerz xmlrpc.serverZwin32winreg_winregzmoves.z.movesmovesc@seZdZdZdS)Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseNr rrrrrrr rt@srtZ ParseResultZurlparseZ SplitResultZparse_qsZ parse_qslZ urldefragZurljoinZurlsplitZ urlunparseZ urlunsplitZ quote_plusZunquoteZ unquote_plusZ urlencodeZ splitqueryZsplittagZ splituserZ uses_fragmentZ uses_netlocZ uses_paramsZ uses_queryZ uses_relativemoves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNrurrrr rwhsrwZURLErrorZurllib2Z HTTPErrorZContentTooShortErrorz.moves.urllib.errormoves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNrurrrr ry|sryZurlopenzurllib.requestZinstall_openerZ build_openerZ pathname2urlZ url2pathnameZ getproxiesZRequestZOpenerDirectorZHTTPDefaultErrorHandlerZHTTPRedirectHandlerZHTTPCookieProcessorZ ProxyHandlerZ BaseHandlerZHTTPPasswordMgrZHTTPPasswordMgrWithDefaultRealmZAbstractBasicAuthHandlerZHTTPBasicAuthHandlerZProxyBasicAuthHandlerZAbstractDigestAuthHandlerZHTTPDigestAuthHandlerZProxyDigestAuthHandlerZ HTTPHandlerZ HTTPSHandlerZ FileHandlerZ FTPHandlerZCacheFTPHandlerZUnknownHandlerZHTTPErrorProcessorZ urlretrieveZ urlcleanupZ URLopenerZFancyURLopenerZ proxy_bypassz.moves.urllib.requestmoves.urllib_requestzmoves.urllib.requestc@seZdZdZdS) Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNrurrrr r{sr{Zaddbasezurllib.responseZ addclosehookZaddinfoZ addinfourlz.moves.urllib.responsemoves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNrurrrr r}sr}ZRobotFileParserz.moves.urllib.robotparsermoves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZedZedZedZ edZ edZ ddZ d S) Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacervrxrzr|r~cCsdddddgS)Nparseerrorrequestresponserprr rrr r8szModule_six_moves_urllib.__dir__N) r rrrrL _importerrCrrrrrpr8rrrr rs     rz moves.urllibcCstt|j|dS)zAdd an item to six.moves.N)rrQr)Zmoverrr add_movesrc CsXztt|WnDtk rRz tj|=Wn"tk rLtd|fYnXYnXdS)zRemove item from six.moves.zno such move, %rN)r rQr"rs__dict__rFrrrr remove_moves r__func____self__ __closure____code__ __defaults__ __globals__im_funcZim_selfZ func_closureZ func_codeZ func_defaultsZ func_globalscCs|Sr)next)itrrr advance_iterator srcCstddt|jDS)Ncss|]}d|jkVqdS)__call__N)r)r5klassrrr szcallable..)anytype__mro__)r#rrr callablesrcCs|SrrZunboundrrr get_unbound_functionsrcCs|Srrrclsrrr create_unbound_methodsrcCs|jSr)rrrrr r"scCst|||jSr)types MethodTyper!)rr#rrr create_bound_method%srcCst|d|Sr)rrrrrr r(sc@seZdZddZdS)IteratorcCst||Sr)r__next__r rrr r-sz Iterator.nextN)r rrrrrrr r+srz3Get the function out of a possibly unbound functioncKst|jf|Sr)iterkeysdkwrrr iterkeys>srcKst|jf|Sr)rvaluesrrrr itervaluesAsrcKst|jf|Sr)ritemsrrrr iteritemsDsrcKst|jf|Sr)rZlistsrrrr iterlistsGsrrrrcKs |jf|Sr)rrrrr rPscKs |jf|Sr)rrrrr rSscKs |jf|Sr)rrrrr rVscKs |jf|Sr)rrrrr rYsviewkeys viewvalues viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs |dS)Nzlatin-1)encodesrrr bksrcCs|Srrrrrr unsrz>BassertCountEqualZassertRaisesRegexpZassertRegexpMatchesassertRaisesRegex assertRegexcCs|Srrrrrr rscCst|dddS)Nz\\z\\\\Zunicode_escape)unicodereplacerrrr rscCs t|dS)Nrord)Zbsrrr byte2intsrcCs t||Srr)Zbufirrr indexbytessrZassertItemsEqualz Byte literalz Text literalcOst|t||Sr)r._assertCountEqualr argskwargsrrr rscOst|t||Sr)r._assertRaisesRegexrrrr rscOst|t||Sr)r. _assertRegexrrrr rsexeccCs*|dkr|}|j|k r"|||dSr) __traceback__with_traceback)r$r1tbrrr reraises   rcCsB|dkr*td}|j}|dkr&|j}~n |dkr6|}tddS)zExecute code in a namespace.Nrzexec _code_ in _globs_, _locs_)r _getframe f_globalsf_localsr)Z_code_Z_globs_Z_locs_framerrr exec_s rz9def reraise(tp, value, tb=None): raise tp, value, tb )rrzrdef raise_from(value, from_value): if from_value is None: raise value raise value from from_value zCdef raise_from(value, from_value): raise value from from_value cCs|dSrr)r1Z from_valuerrr raise_fromsrprintc s.|dtjdkrdSfdd}d}|dd}|dk r`t|trNd}nt|ts`td|d d}|dk rt|trd}nt|tstd |rtd |s|D]}t|trd}qq|rtd }td }nd }d }|dkr|}|dkr|}t|D] \} }| r||||q||dS)z4The new-style print function for Python 2.4 and 2.5.fileNcsdt|tst|}ttrVt|trVjdk rVtdd}|dkrHd}|j|}|dS)Nerrorsstrict) rI basestringstrrrencodingr.rwrite)datarfprr rs   zprint_..writeFsepTzsep must be None or a stringendzend must be None or a stringz$invalid keyword arguments to print()  )poprstdoutrIrr TypeError enumerate) rrrZ want_unicoderrargnewlineZspacerrrr print_sL          r)rrcOs<|dtj}|dd}t|||r8|dk r8|dS)NrflushF)getrrr_printr)rrrrrrr r s    zReraise an exception.csfdd}|S)Ncst|}|_|Sr)rcwraps __wrapped__)fassignedupdatedwrappedrr wrapperszwraps..wrapperr)rrrrrrr rsrcs&Gfddd}t|ddiS)z%Create a base class with a metaclass.cseZdZfddZdS)z!with_metaclass..metaclasscs ||Srr)rrZ this_basesrbasesmetarr __new__'sz)with_metaclass..metaclass.__new__N)r rrrrrrr metaclass%srZtemporary_classr)rr)rrrrrr with_metaclass srcsfdd}|S)z6Class decorator for creating a class with a metaclass.csh|j}|d}|dk r@t|tr,|g}|D]}||q0|dd|dd|j|j|S)N __slots__r __weakref__)rcopyrrIrrr __bases__)rZ orig_varsslotsZ slots_varrrr r.s      zadd_metaclass..wrapperr)rrrrr add_metaclass,s rcCs2tr.d|jkrtd|j|j|_dd|_|S)a A decorator that defines __unicode__ and __str__ methods under Python 2. Under Python 3 it does nothing. To support Python 2 and 3 with a single code base, define a __str__ method returning text and apply this decorator to the class. __str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|dS)Nzutf-8) __unicode__rr rrr Jz-python_2_unicode_compatible..)PY2r ValueErrorr rr)rrrr python_2_unicode_compatible<s  r__spec__)N)NN)rZ __future__rrcrUoperatorrr __author__ __version__ version_inforr)ZPY34rZ string_typesintZ integer_typesrZ class_typesZ text_typebytesZ binary_typemaxsizeZMAXSIZErZlongZ ClassTyperplatform startswithobjectrlen OverflowErrorrrrr' ModuleTyper4r9r;r rrQr7r/rrrIrArsrtZ_urllib_parse_moved_attributesrwZ_urllib_error_moved_attributesryZ _urllib_request_moved_attributesr{Z!_urllib_response_moved_attributesr}Z$_urllib_robotparser_moved_attributesrrrZ _meth_funcZ _meth_selfZ _func_closureZ _func_codeZ_func_defaultsZ _func_globalsrr NameErrorrrrrrr attrgetterZget_method_functionZget_method_selfZget_function_closureZget_function_codeZget_function_defaultsZget_function_globalsrrrr methodcallerrrrrrchrZunichrstructStructpackZint2byte itemgetterrgetitemrrZ iterbytesrRrSBytesIOrrrpartialr[rrrrr.rVrrrrrWRAPPER_ASSIGNMENTSWRAPPER_UPDATESrrrrrL __package__globalsrrsubmodule_search_locations meta_pathrrZimporterappendrrrr s    >                                      D                                                               #                                                 5     site-packages/setuptools/_vendor/__pycache__/ordered_set.cpython-38.opt-1.pyc000064400000040034151464452710023251 0ustar00U Qab;@s|dZddlZddlmZzddlmZmZWn$ek rPddlmZmZYnXe dZ dZ ddZ Gdd d eeZ dS) z An OrderedSet is a custom MutableSet that remembers its order, so that every entry has an index that can be looked up. Based on a recipe originally posted to ActiveState Recipes by Raymond Hettiger, and released under the MIT license. N)deque) MutableSetSequencez3.1cCs"t|do t|t o t|t S)a  Are we being asked to look up a list of things, instead of a single thing? We check for the `__iter__` attribute so that this can cover types that don't have to be known by this module, such as NumPy arrays. Strings, however, should be considered as atomic values to look up, not iterables. The same goes for tuples, since they are immutable and therefore valid entries. We don't need to check for the Python 2 `unicode` type, because it doesn't have an `__iter__` attribute anyway. __iter__)hasattr isinstancestrtuple)objr B/usr/lib/python3.8/site-packages/setuptools/_vendor/ordered_set.py is_iterables    r c@seZdZdZd;ddZddZddZd d Zd d Zd dZ ddZ ddZ e Z ddZ ddZeZeZddZddZddZddZdd Zd!d"Zd#d$Zd%d&Zd'd(Zd)d*Zd+d,Zd-d.Zd/d0Zd1d2Zd3d4Zd5d6Z d7d8Z!d9d:Z"dS)< OrderedSetz An OrderedSet is a custom MutableSet that remembers its order, so that every entry has an index that can be looked up. Example: >>> OrderedSet([1, 1, 2, 3, 2]) OrderedSet([1, 2, 3]) NcCs g|_i|_|dk r||O}dSN)itemsmap)selfiterabler r r __init__4szOrderedSet.__init__cCs t|jS)z Returns the number of unique elements in the ordered set Example: >>> len(OrderedSet([])) 0 >>> len(OrderedSet([1, 2])) 2 )lenrrr r r __len__:s zOrderedSet.__len__cs|t|tr|tkrSt|r4fdd|DSt|dsHt|trlj|}t|trf|S|Sn t d|dS)aQ Get the item at a given index. If `index` is a slice, you will get back that slice of items, as a new OrderedSet. If `index` is a list or a similar iterable, you'll get a list of items corresponding to those indices. This is similar to NumPy's "fancy indexing". The result is not an OrderedSet because you may ask for duplicate indices, and the number of elements returned should be the number of elements asked for. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset[1] 2 csg|]}j|qSr )r).0irr r [sz*OrderedSet.__getitem__.. __index__z+Don't know how to index an OrderedSet by %rN) rslice SLICE_ALLcopyr rrlist __class__ TypeError)rindexresultr rr __getitem__Fs   zOrderedSet.__getitem__cCs ||S)z Return a shallow copy of this object. Example: >>> this = OrderedSet([1, 2, 3]) >>> other = this.copy() >>> this == other True >>> this is other False )r rr r r res zOrderedSet.copycCst|dkrdSt|SdS)Nrr)rrrr r r __getstate__ss zOrderedSet.__getstate__cCs"|dkr|gn ||dS)Nr)r)rstater r r __setstate__s zOrderedSet.__setstate__cCs ||jkS)z Test if the item is in this ordered set Example: >>> 1 in OrderedSet([1, 3, 2]) True >>> 5 in OrderedSet([1, 3, 2]) False )rrkeyr r r __contains__s zOrderedSet.__contains__cCs0||jkr&t|j|j|<|j||j|S)aE Add `key` as an item to this OrderedSet, then return its index. If `key` is already in the OrderedSet, return the index it already had. Example: >>> oset = OrderedSet() >>> oset.append(3) 0 >>> print(oset) OrderedSet([3]) )rrrappendr(r r r adds  zOrderedSet.addcCsFd}z|D]}||}q Wn$tk r@tdt|YnX|S)a< Update the set with the given iterable sequence, then return the index of the last element inserted. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.update([3, 1, 5, 1, 4]) 4 >>> print(oset) OrderedSet([1, 2, 3, 5, 4]) Nz(Argument needs to be an iterable, got %s)r,r! ValueErrortype)rZsequenceZ item_indexitemr r r updates  zOrderedSet.updatecs$t|rfdd|DSj|S)aH Get the index of a given entry, raising an IndexError if it's not present. `key` can be an iterable of entries that is not a string, in which case this returns a list of indices. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.index(2) 1 csg|]}|qSr )r")rZsubkeyrr r rsz$OrderedSet.index..)r rr(r rr r"s zOrderedSet.indexcCs,|jstd|jd}|jd=|j|=|S)z Remove and return the last element from the set. Raises KeyError if the set is empty. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.pop() 3 z Set is empty)rKeyErrorr)relemr r r pops  zOrderedSet.popcCsP||krL|j|}|j|=|j|=|jD]\}}||kr,|d|j|<q,dS)a Remove an element. Do not raise an exception if absent. The MutableSet mixin uses this to implement the .remove() method, which *does* raise an error when asked to remove a non-existent item. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.discard(2) >>> print(oset) OrderedSet([1, 3]) >>> oset.discard(2) >>> print(oset) OrderedSet([1, 3]) N)rr)rr)rkvr r r discards zOrderedSet.discardcCs|jdd=|jdS)z8 Remove all items from this OrderedSet. N)rrclearrr r r r9s zOrderedSet.clearcCs t|jS)zb Example: >>> list(iter(OrderedSet([1, 2, 3]))) [1, 2, 3] )iterrrr r r rszOrderedSet.__iter__cCs t|jS)zf Example: >>> list(reversed(OrderedSet([1, 2, 3]))) [3, 2, 1] )reversedrrr r r __reversed__ szOrderedSet.__reversed__cCs&|sd|jjfSd|jjt|fS)Nz%s()z%s(%r))r __name__rrr r r __repr__szOrderedSet.__repr__cCsRt|ttfrt|t|kSz t|}Wntk r@YdSXt||kSdS)a Returns true if the containers have the same items. If `other` is a Sequence, then order is checked, otherwise it is ignored. Example: >>> oset = OrderedSet([1, 3, 2]) >>> oset == [1, 3, 2] True >>> oset == [1, 2, 3] False >>> oset == [2, 3] False >>> oset == OrderedSet([3, 2, 1]) False FN)rrrrsetr!)rotherZ other_as_setr r r __eq__s zOrderedSet.__eq__cGs<t|tr|jnt}ttt|g|}tj|}||S)a Combines all unique items. Each items order is defined by its first appearance. Example: >>> oset = OrderedSet.union(OrderedSet([3, 1, 4, 1, 5]), [1, 3], [2, 0]) >>> print(oset) OrderedSet([3, 1, 4, 5, 2, 0]) >>> oset.union([8, 9]) OrderedSet([3, 1, 4, 5, 2, 0, 8, 9]) >>> oset | {10} OrderedSet([3, 1, 4, 5, 2, 0, 10]) )rrr rritchain from_iterable)rsetsclsZ containersrr r r union6s zOrderedSet.unioncCs ||Sr) intersectionrr@r r r __and__IszOrderedSet.__and__csHt|tr|jnt}|r>> oset = OrderedSet.intersection(OrderedSet([0, 1, 2, 3]), [1, 2, 3]) >>> print(oset) OrderedSet([1, 2, 3]) >>> oset.intersection([2, 4, 5], [1, 2, 3, 4]) OrderedSet([2]) >>> oset.intersection() OrderedSet([1, 2, 3]) c3s|]}|kr|VqdSrr rr/commonr r ^sz*OrderedSet.intersection..)rrr r?rHrrrErFrr rLr rHMs zOrderedSet.intersectioncs:|j}|r.tjtt|fdd|D}n|}||S)a Returns all elements that are in this set but not the others. Example: >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2])) OrderedSet([1, 3]) >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]), OrderedSet([3])) OrderedSet([1]) >>> OrderedSet([1, 2, 3]) - OrderedSet([2]) OrderedSet([1, 3]) >>> OrderedSet([1, 2, 3]).difference() OrderedSet([1, 2, 3]) c3s|]}|kr|VqdSrr rKr@r r rNtsz(OrderedSet.difference..)r r?rGrrOr rPr differencecs zOrderedSet.differencecs*t|tkrdStfdd|DS)a7 Report whether another set contains this set. Example: >>> OrderedSet([1, 2, 3]).issubset({1, 2}) False >>> OrderedSet([1, 2, 3]).issubset({1, 2, 3, 4}) True >>> OrderedSet([1, 2, 3]).issubset({1, 4, 3, 5}) False Fc3s|]}|kVqdSrr rKrPr r rNsz&OrderedSet.issubset..rallrIr rPr issubsetys zOrderedSet.issubsetcs*tt|krdStfdd|DS)a= Report whether this set contains another set. Example: >>> OrderedSet([1, 2]).issuperset([1, 2, 3]) False >>> OrderedSet([1, 2, 3, 4]).issuperset({1, 2, 3}) True >>> OrderedSet([1, 4, 3, 5]).issuperset({1, 2, 3}) False Fc3s|]}|kVqdSrr rKrr r rNsz(OrderedSet.issuperset..rRrIr rr issupersets zOrderedSet.issupersetcCs:t|tr|jnt}|||}|||}||S)a Return the symmetric difference of two OrderedSets as a new set. That is, the new set will contain all elements that are in exactly one of the sets. Their order will be preserved, with elements from `self` preceding elements from `other`. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.symmetric_difference(other) OrderedSet([4, 5, 9, 2]) )rrr rQrG)rr@rFZdiff1Zdiff2r r r symmetric_differenceszOrderedSet.symmetric_differencecCs||_ddt|D|_dS)zt Replace the 'items' list of this OrderedSet with a new one, updating self.map accordingly. cSsi|]\}}||qSr r )ridxr/r r r sz,OrderedSet._update_items..N)r enumerater)rrr r r _update_itemsszOrderedSet._update_itemscs:t|D]}t|Oq |fdd|jDdS)a Update this OrderedSet to remove items from one or more other sets. Example: >>> this = OrderedSet([1, 2, 3]) >>> this.difference_update(OrderedSet([2, 4])) >>> print(this) OrderedSet([1, 3]) >>> this = OrderedSet([1, 2, 3, 4, 5]) >>> this.difference_update(OrderedSet([2, 4]), OrderedSet([1, 4, 6])) >>> print(this) OrderedSet([3, 5]) csg|]}|kr|qSr r rKitems_to_remover r rsz0OrderedSet.difference_update..Nr?rZr)rrEr@r r[r difference_updateszOrderedSet.difference_updatecs&t|fdd|jDdS)a^ Update this OrderedSet to keep only items in another set, preserving their order in this set. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.intersection_update(other) >>> print(this) OrderedSet([1, 3, 7]) csg|]}|kr|qSr r rKrPr r rsz2OrderedSet.intersection_update..Nr]rIr rPr intersection_updates zOrderedSet.intersection_updatecs<fdd|D}t|fddjD|dS)a Update this OrderedSet to remove items from another set, then add items from the other set that were not present in this set. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.symmetric_difference_update(other) >>> print(this) OrderedSet([4, 5, 9, 2]) csg|]}|kr|qSr r rKrr r rsz:OrderedSet.symmetric_difference_update..csg|]}|kr|qSr r rKr[r r rsNr])rr@Z items_to_addr )r\rr symmetric_difference_updates z&OrderedSet.symmetric_difference_update)N)#r= __module__ __qualname____doc__rrr$rr%r'r*r,r+r0r"Zget_locZ get_indexerr4r8r9rr<r>rArGrJrHrQrTrUrVrZr^r_r`r r r r r*s@    r)rc itertoolsrB collectionsrZcollections.abcrr ImportErrorrr __version__r rr r r r s site-packages/setuptools/_vendor/__pycache__/pyparsing.cpython-38.opt-1.pyc000064400000611637151464452710023003 0ustar00U Qabwi@s dZdZdZdZddlZddlmZddlZddl Z ddl Z ddl Z ddl Z ddl Z ddlZddlZddlZddlmZzddlmZWn ek rddlmZYnXzdd lmZdd lmZWn,ek rdd l mZdd l mZYnXzdd l mZWnBek rFzdd lmZWnek r@dZYnXYnXd d ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtgiZee jdduZeddukZ e rpe j!Z"e#Z$e%Z&e#Z'e(e)e*e+e,ee-e.e/e0e1g Z2n`e j3Z"e4Z5dvdwZ'gZ2ddl6Z6dx7D]8Z8ze29e:e6e8Wne;k rYqYnXqeGd~dde?Z@ejAejBZCdZDeDdZEeCeDZFe%dZGdHddzejIDZJGdd#d#eKZLGdd%d%eLZMGdd'd'eLZNGdd)d)eNZOGdd,d,eKZPGddde?ZQGdd(d(e?ZReSeRdd?ZTddPZUddMZVddZWddZXddZYddWZZd/ddZ[Gdd*d*e?Z\Gdd2d2e\Z]Gddde]Z^Gddde]Z_Gddde]Z`e`Zae`e\_bGddde]ZcGddde`ZdGdd d ecZeGddrdre]ZfGdd5d5e]ZgGdd-d-e]ZhGdd+d+e]ZiGddde]ZjGdd4d4e]ZkGddde]ZlGdddelZmGdddelZnGdddelZoGdd0d0elZpGdd/d/elZqGdd7d7elZrGdd6d6elZsGdd&d&e\ZtGdd d etZuGdd"d"etZvGdddetZwGdddetZxGdd$d$e\ZyGdddeyZzGdddeyZ{GdddeyZ|Gddde|Z}Gdd8d8e|Z~Gddde?ZeZGdd!d!eyZGdd.d.eyZGdddeyZGddÄdeZGdd3d3eyZGdddeZGdddeZGdddeZGdd1d1eZGdd d e?ZddhZd0ddFZd1ddBZddЄZddUZddTZddԄZd2ddYZddGZd3ddmZddnZddpZe^dIZendOZeodNZepdgZeqdfZegeGddd܍ddބZehd߃ddބZehdddބZeeBeBejdd{d܍BZeeedeZe`deddee}eeBddZddeZddSZddbZdd`ZddsZeddބZeddބZddZddQZddRZddkZe?e_d4ddqZe@Ze?e_e?e_ededfddoZeZeehdddZeehdddZeehddehddBdZeeadedZdddefddVZd5ddlZedZedZeegeCeFdd\ZZeed 7d Zehd d Heàġd dZŐddaZeehdddZehddZehdɡdZehddZeehddeBdZeZehddZee}egeJdːdeegde`d˃eoϡdZeeeeBddd@ZGd dtdtZeӐd!k redd"Zedd#ZegeCeFd$Zee֐d%dՐd&eZeee׃d'Zؐd(eBZee֐d%dՐd&eZeeeڃd)ZeԐd*eِd'eeېd)Zeܠݐd+ejޠݐd,ejߠݐd,ejݐd-ddlZej᠝eejejݐd.dS(6a pyparsing module - Classes and methods to define and execute parsing grammars ============================================================================= The pyparsing module is an alternative approach to creating and executing simple grammars, vs. the traditional lex/yacc approach, or the use of regular expressions. With pyparsing, you don't need to learn a new syntax for defining grammars or matching expressions - the parsing module provides a library of classes that you use to construct the grammar directly in Python. Here is a program to parse "Hello, World!" (or any greeting of the form C{", !"}), built up using L{Word}, L{Literal}, and L{And} elements (L{'+'} operator gives L{And} expressions, strings are auto-converted to L{Literal} expressions):: from pyparsing import Word, alphas # define grammar of a greeting greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) The program outputs the following:: Hello, World! -> ['Hello', ',', 'World', '!'] The Python representation of the grammar is quite readable, owing to the self-explanatory class names, and the use of '+', '|' and '^' operators. The L{ParseResults} object returned from L{ParserElement.parseString} can be accessed as a nested list, a dictionary, or an object with named attributes. The pyparsing module handles some of the problems that are typically vexing when writing text parsers: - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello , World !", etc.) - quoted strings - embedded comments Getting Started - ----------------- Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing classes inherit from. Use the docstrings for examples of how to: - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes - construct character word-group expressions using the L{Word} class - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes - use L{'+'}, L{'|'}, L{'^'}, and L{'&'} operators to combine simple expressions into more complex ones - associate names with your parsed results using L{ParserElement.setResultsName} - find some helpful expression short-cuts like L{delimitedList} and L{oneOf} - find more useful common expressions in the L{pyparsing_common} namespace class z2.2.1z18 Sep 2018 00:49 UTCz*Paul McGuire N)ref)datetime)RLock)Iterable)MutableMapping) OrderedDictAndCaselessKeywordCaselessLiteral CharsNotInCombineDictEachEmpty FollowedByForward GoToColumnGroupKeywordLineEnd LineStartLiteral MatchFirstNoMatchNotAny OneOrMoreOnlyOnceOptionalOrParseBaseExceptionParseElementEnhanceParseExceptionParseExpressionParseFatalException ParseResultsParseSyntaxException ParserElement QuotedStringRecursiveGrammarExceptionRegexSkipTo StringEnd StringStartSuppressTokenTokenConverterWhiteWordWordEnd WordStart ZeroOrMore alphanumsalphas alphas8bit anyCloseTag anyOpenTag cStyleCommentcolcommaSeparatedListcommonHTMLEntity countedArraycppStyleCommentdblQuotedStringdblSlashComment delimitedListdictOfdowncaseTokensemptyhexnums htmlCommentjavaStyleCommentlinelineEnd lineStartlineno makeHTMLTags makeXMLTagsmatchOnlyAtColmatchPreviousExprmatchPreviousLiteral nestedExprnullDebugActionnumsoneOfopAssocoperatorPrecedence printablespunc8bitpythonStyleComment quotedString removeQuotesreplaceHTMLEntity replaceWith restOfLinesglQuotedStringsrange stringEnd stringStarttraceParseAction unicodeString upcaseTokens withAttribute indentedBlockoriginalTextForungroup infixNotation locatedExpr withClass CloseMatchtokenMappyparsing_commoncCsft|tr|Sz t|WStk r`t|td}td}|dd| |YSXdS)aDrop-in replacement for str(obj) that tries to be Unicode friendly. It first tries str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It then < returns the unicode object | encodes it with the default encoding | ... >. xmlcharrefreplacez&#\d+;cSs$dtt|dddddS)Nz\ur)hexinttry@/usr/lib/python3.8/site-packages/setuptools/_vendor/pyparsing.pyz_ustr..N) isinstanceZunicodestrUnicodeEncodeErrorencodesysgetdefaultencodingr)setParseActiontransformString)objretZ xmlcharrefryryrz_ustrs  rz6sum len sorted reversed list tuple set any all min maxccs|] }|VqdSNry).0yryryrz srcCs:d}dddD}t||D]\}}|||}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)&;Nry)rsryryrzrsz_xml_escape..zamp gt lt quot apos)splitzipreplace)dataZ from_symbolsZ to_symbolsZfrom_Zto_ryryrz _xml_escapes rc@s eZdZdS) _ConstantsN)__name__ __module__ __qualname__ryryryrzrsr 0123456789Z ABCDEFabcdef\ccs|]}|tjkr|VqdSr)stringZ whitespacercryryrzrs c@sPeZdZdZdddZeddZdd Zd d Zd d Z dddZ ddZ dS)rz7base exception class for all parsing runtime exceptionsrNcCs>||_|dkr||_d|_n ||_||_||_|||f|_dSNr)locmsgpstr parserElementargs)selfrrrelemryryrz__init__szParseBaseException.__init__cCs||j|j|j|jS)z internal factory method to simplify creating one type of ParseException from another - avoids having __init__ signature conflicts among subclasses )rrrr)clsperyryrz_from_exceptionsz"ParseBaseException._from_exceptioncCsN|dkrt|j|jS|dkr,t|j|jS|dkrBt|j|jSt|dS)zsupported attributes by name are: - lineno - returns the line number of the exception text - col - returns the column number of the exception text - line - returns the line containing the exception text rL)r;columnrIN)rLrrr;rIAttributeError)rZanameryryrz __getattr__szParseBaseException.__getattr__cCsd|j|j|j|jfS)Nz"%s (at char %d), (line:%d, col:%d))rrrLrrryryrz__str__szParseBaseException.__str__cCst|Srrrryryrz__repr__szParseBaseException.__repr__>!} ('-' operator) indicates that parsing is to stop immediately because an unbacktrackable syntax error has been foundNrryryryrzr%sc@s eZdZdZddZddZdS)r(zZexception thrown by L{ParserElement.validate} if the grammar could be improperly recursivecCs ||_dSrZparseElementTracerparseElementListryryrzr4sz"RecursiveGrammarException.__init__cCs d|jS)NzRecursiveGrammarException: %srrryryrzr7sz!RecursiveGrammarException.__str__N)rrrrrrryryryrzr(2sc@s,eZdZddZddZddZddZd S) _ParseResultsWithOffsetcCs||f|_dSrtup)rZp1Zp2ryryrzr;sz _ParseResultsWithOffset.__init__cCs |j|Srrriryryrz __getitem__=sz#_ParseResultsWithOffset.__getitem__cCst|jdSNr)reprrrryryrzr?sz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSrrrryryrz setOffsetAsz!_ParseResultsWithOffset.setOffsetN)rrrrrrrryryryrzr:src@seZdZdZd[ddZddddefddZdd Zefd d Zd d Z ddZ ddZ ddZ e Z ddZddZddZddZddZereZeZeZn$eZeZeZddZd d!Zd"d#Zd$d%Zd&d'Zd\d(d)Zd*d+Zd,d-Zd.d/Zd0d1Z d2d3Z!d4d5Z"d6d7Z#d8d9Z$d:d;Z%d} - see L{ParserElement.setResultsName}) Example:: integer = Word(nums) date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: # date_str = integer("year") + '/' + integer("month") + '/' + integer("day") # parseString returns a ParseResults object result = date_str.parseString("1999/12/31") def test(s, fn=repr): print("%s -> %s" % (s, fn(eval(s)))) test("list(result)") test("result[0]") test("result['month']") test("result.day") test("'month' in result") test("'minutes' in result") test("result.dump()", str) prints:: list(result) -> ['1999', '/', '12', '/', '31'] result[0] -> '1999' result['month'] -> '12' result.day -> '31' 'month' in result -> True 'minutes' in result -> False result.dump() -> ['1999', '/', '12', '/', '31'] - day: 31 - month: 12 - year: 1999 NTcCs"t||r|St|}d|_|SNT)r}object__new___ParseResults__doinit)rtoklistnameasListmodalZretobjryryrzrks   zParseResults.__new__c Csb|jrvd|_d|_d|_i|_||_||_|dkr6g}||trP|dd|_n||trft||_n|g|_t |_ |dk r^|r^|sd|j|<||t rt |}||_||t dttfr|ddgfks^||tr|g}|r(||trt|d||<ntt|dd||<|||_n6z|d||<Wn$tttfk r\|||<YnXdS)NFrr)r_ParseResults__name_ParseResults__parent_ParseResults__accumNames_ParseResults__asList_ParseResults__modallist_ParseResults__toklist_generatorTypedict_ParseResults__tokdictrvrr basestringr$rcopyKeyError TypeError IndexError)rrrrrr}ryryrzrtsB     $   zParseResults.__init__cCsPt|ttfr|j|S||jkr4|j|ddStdd|j|DSdS)NrtrcSsg|] }|dqSrryrvryryrz sz,ParseResults.__getitem__..)r}rvslicerrrr$rryryrzrs   zParseResults.__getitem__cCs||tr0|j|t|g|j|<|d}nD||ttfrN||j|<|}n&|j|tt|dg|j|<|}||trt||_ dSr) rrgetrrvrrr$wkrefr)rkrr}subryryrz __setitem__s   " zParseResults.__setitem__c Cst|ttfrt|j}|j|=t|trH|dkr:||7}t||d}tt||}||j D]>\}}|D]0}t |D]"\}\}} t || | |k||<qqxqln|j |=dSNrr) r}rvrlenrrrangeindicesreverseritems enumerater) rrZmylenZremovedr occurrencesjrvaluepositionryryrz __delitem__s  zParseResults.__delitem__cCs ||jkSr)r)rrryryrz __contains__szParseResults.__contains__cCs t|jSr)rrrryryrz__len__r|zParseResults.__len__cCs |j Srrrryryrz__bool__r|zParseResults.__bool__cCs t|jSriterrrryryrz__iter__r|zParseResults.__iter__cCst|jdddSNrtrrryryrz __reversed__r|zParseResults.__reversed__cCs$t|jdr|jSt|jSdS)Niterkeys)hasattrrrrrryryrz _iterkeyss  zParseResults._iterkeyscsfddDS)Nc3s|]}|VqdSrryrrrryrzrsz+ParseResults._itervalues..rrryrrz _itervaluesszParseResults._itervaluescsfddDS)Nc3s|]}||fVqdSrryrrryrzrsz*ParseResults._iteritems..rrryrrz _iteritemsszParseResults._iteritemscCs t|S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)rrrryryrzkeysszParseResults.keyscCs t|S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r itervaluesrryryrzvaluesszParseResults.valuescCs t|S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r iteritemsrryryrzrszParseResults.itemscCs t|jS)zSince keys() returns an iterator, this method is helpful in bypassing code that looks for the existence of any defined results names.)boolrrryryrzhaskeysszParseResults.haskeyscOs|s dg}|D]*\}}|dkr0|d|f}qtd|qt|dtsdt|dksd|d|kr~|d}||}||=|S|d}|SdS)a Removes and returns item at specified index (default=C{last}). Supports both C{list} and C{dict} semantics for C{pop()}. If passed no argument or an integer argument, it will use C{list} semantics and pop tokens from the list of parsed tokens. If passed a non-integer argument (most likely a string), it will use C{dict} semantics and pop the corresponding value from any defined results names. A second default return value argument is supported, just as in C{dict.pop()}. Example:: def remove_first(tokens): tokens.pop(0) print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321'] label = Word(alphas) patt = label("LABEL") + OneOrMore(Word(nums)) print(patt.parseString("AAB 123 321").dump()) # Use pop() in a parse action to remove named result (note that corresponding value is not # removed from list form of results) def remove_LABEL(tokens): tokens.pop("LABEL") return tokens patt.addParseAction(remove_LABEL) print(patt.parseString("AAB 123 321").dump()) prints:: ['AAB', '123', '321'] - LABEL: AAB ['AAB', '123', '321'] rtdefaultrz-pop() got an unexpected keyword argument '%s'rN)rrr}rvr)rrkwargsrrindexrZ defaultvalueryryrzpops""  zParseResults.popcCs||kr||S|SdS)ai Returns named result matching the given key, or if there is no such name, then returns the given C{defaultValue} or C{None} if no C{defaultValue} is specified. Similar to C{dict.get()}. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString("1999/12/31") print(result.get("year")) # -> '1999' print(result.get("hour", "not specified")) # -> 'not specified' print(result.get("hour")) # -> None Nry)rkey defaultValueryryrzr3szParseResults.getcCsR|j|||jD]4\}}t|D]"\}\}}t||||k||<q(qdS)a Inserts new element at location index in the list of parsed tokens. Similar to C{list.insert()}. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to insert the parse location in the front of the parsed results def insert_locn(locn, tokens): tokens.insert(0, locn) print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321'] N)rinsertrrrr)rr ZinsStrrrrrrryryrzrIszParseResults.insertcCs|j|dS)a Add single element to end of ParseResults list of elements. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to compute the sum of the parsed integers, and add it to the end def append_sum(tokens): tokens.append(sum(map(int, tokens))) print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444] N)rappend)ritemryryrzr]s zParseResults.appendcCs$t|tr||7}n |j|dS)a Add sequence of elements to end of ParseResults list of elements. Example:: patt = OneOrMore(Word(alphas)) # use a parse action to append the reverse of the matched strings, to make a palindrome def make_palindrome(tokens): tokens.extend(reversed([t[::-1] for t in tokens])) return ''.join(tokens) print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl' N)r}r$rextend)rZitemseqryryrzrks  zParseResults.extendcCs|jdd=|jdS)z7 Clear all elements and results names. N)rrclearrryryrzr}s zParseResults.clearcCsjz ||WStk r YdSX||jkrb||jkrH|j|ddStdd|j|DSndSdS)NrrtrcSsg|] }|dqSrryrryryrzrsz,ParseResults.__getattr__..)rrrr$rrryryrzrs   zParseResults.__getattr__cCs|}||7}|Srr)rotherrryryrz__add__szParseResults.__add__cs|jrjt|jfdd|j}fdd|D}|D],\}}|||<t|dtr.c s4g|],\}}|D]}|t|d|dfqqSrr)rrrvlistr) addoffsetryrzrsz)ParseResults.__iadd__..r) rrrrr}r$rrrupdate)rrZ otheritemsZotherdictitemsrrry)rrrz__iadd__s     zParseResults.__iadd__cCs&t|tr|dkr|S||SdSr)r}rvrrrryryrz__radd__szParseResults.__radd__cCsdt|jt|jfS)Nz(%s, %s))rrrrryryrzrszParseResults.__repr__cCsdddd|jDdS)N[, css(|] }t|trt|nt|VqdSr)r}r$rrrrryryrzrsz'ParseResults.__str__..])rrrryryrzrszParseResults.__str__rcCsLg}|jD]<}|r |r ||t|tr8||7}q |t|q |Sr)rrr}r$ _asStringListr)rsepoutrryryrzr(s   zParseResults._asStringListcCsdd|jDS)a Returns the parse results as a nested list of matching tokens, all converted to strings. Example:: patt = OneOrMore(Word(alphas)) result = patt.parseString("sldkj lsdkj sldkj") # even though the result prints in string-like form, it is actually a pyparsing ParseResults print(type(result), result) # -> ['sldkj', 'lsdkj', 'sldkj'] # Use asList() to create an actual list result_list = result.asList() print(type(result_list), result_list) # -> ['sldkj', 'lsdkj', 'sldkj'] cSs"g|]}t|tr|n|qSry)r}r$r)rresryryrzrsz'ParseResults.asList..rrryryrzrszParseResults.asListcs6tr |j}n|j}fddtfdd|DS)a Returns the named parse results as a nested dictionary. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString('12/31/1999') print(type(result), repr(result)) # -> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]}) result_dict = result.asDict() print(type(result_dict), repr(result_dict)) # -> {'day': '1999', 'year': '12', 'month': '31'} # even though a ParseResults supports dict-like access, sometime you just need to have a dict import json print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"} cs6t|tr.|r|Sfdd|DSn|SdS)Ncsg|] }|qSryryrtoItemryrzrsz7ParseResults.asDict..toItem..)r}r$r asDict)rr,ryrzr-s  z#ParseResults.asDict..toItemc3s|]\}}||fVqdSrryrrrr,ryrzrsz&ParseResults.asDict..)PY_3rrr)rZitem_fnryr,rzr.s  zParseResults.asDictcCs8t|j}|j|_|j|_|j|j|j|_|S)zA Returns a new copy of a C{ParseResults} object. )r$rrrrrr rrrryryrzrs   zParseResults.copyFc CsLd}g}tdd|jD}|d}|s8d}d}d}d} |dk rJ|} n |jrV|j} | sf|rbdSd} |||d| d g7}t|jD]\} } t| tr| |kr|| || |o|dk||g7}n|| d|o|dk||g7}qd} | |kr|| } | s|rqnd} t t | } |||d| d | d | d g 7}q|||d | d g7}d |S) z (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.  css(|] \}}|D]}|d|fVqqdSrNryrryryrzrsz%ParseResults.asXML.. rNZITEM<>.z %s%s- %s: r4rcss|]}t|tVqdSr)r}r$)rvvryryrzrsz %s%s[%d]: %s%s%sr) rrrr sortedrr}r$dumpranyrr) rr9depthfullr*NLrrrrr?ryryrzrAgs,    4,zParseResults.dumpcOstj|f||dS)a Pretty-printer for parsed results as a list, using the C{pprint} module. Accepts additional positional or keyword args as defined for the C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint}) Example:: ident = Word(alphas, alphanums) num = Word(nums) func = Forward() term = ident | num | Group('(' + func + ')') func <<= ident + Group(Optional(delimitedList(term))) result = func.parseString("fna a,b,(fnb c,d,200),100") result.pprint(width=40) prints:: ['fna', ['a', 'b', ['(', 'fnb', ['c', 'd', '200'], ')'], '100']] N)pprintrrrr ryryrzrFszParseResults.pprintcCs.|j|j|jdk r|p d|j|jffSr)rrrrrrrryryrz __getstate__szParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j||dk rDt||_nd|_dSr)rrrrr rr)rstater=Z inAccumNamesryryrz __setstate__s   zParseResults.__setstate__cCs|j|j|j|jfSr)rrrrrryryrz__getnewargs__szParseResults.__getnewargs__cCstt|t|Sr)rrrrrryryrzrszParseResults.__dir__)NNTT)N)r)NFrT)rrT)4rrrrrr}rrrrrrr __nonzero__rrrrrr0rrrrrrr rrrrrrrrr!r#rrr(rr.rr8r;r>rArFrHrJrKrryryryrzr$Dsh& ' 4  # =% - cCsF|}d|krt|kr4nn||ddkr4dS||dd|S)aReturns current column within a string, counting newlines as line separators. The first column is number 1. Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See L{I{ParserElement.parseString}} for more information on parsing strings containing C{}s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. rrr2)rrfind)rstrgrryryrzr;s cCs|dd|dS)aReturns current line number within a string, counting newlines as line separators. The first line is number 1. Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See L{I{ParserElement.parseString}} for more information on parsing strings containing C{}s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. r2rr)count)rrNryryrzrLs cCsF|dd|}|d|}|dkr2||d|S||ddSdS)zfReturns the line of text containing loc within a string, counting newlines as line separators. r2rrN)rMfind)rrNZlastCRZnextCRryryrzrIs  cCs8tdt|dt|dt||t||fdS)NzMatch z at loc z(%d,%d))printrrLr;)instringrexprryryrz_defaultStartDebugActionsrTcCs$tdt|dt|dS)NzMatched z -> )rQrr~r)rRstartlocZendlocrStoksryryrz_defaultSuccessDebugActionsrWcCstdt|dS)NzException raised:)rQr)rRrrSexcryryrz_defaultExceptionDebugActionsrYcGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nry)rryryrzrSsrscstkrfddSdgdgtdddkrFddd}dd d n tj}tjd }|dd d }|d|d|ffdd}d}ztdtdj}Wntk rt}YnX||_|S)Ncs|Srryrlrx)funcryrzr{r|z_trim_arity..rFrs)rqcSs8tdkr dnd}tj| |dd|}|ddgS)N)rqr]rrlimitrs)system_version traceback extract_stack)rar frame_summaryryryrzrdsz"_trim_arity..extract_stackcSs$tj||d}|d}|ddgS)Nr`rtrs)rc extract_tb)tbraZframesreryryrzrfsz_trim_arity..extract_tbr`rtrc sz"|dd}dd<|WStk rdr>n4z.td}|dddddksjW5~Xdkrdd7<YqYqXqdS)NrTrtrsr`r)rrexc_info)rrrgrfZ foundArityr\ramaxargsZpa_call_line_synthryrzwrapper-s   z_trim_arity..wrapperzr __class__)r)r) singleArgBuiltinsrbrcrdrfgetattrr Exceptionr~)r\rkrdZ LINE_DIFFZ this_linerl func_nameryrjrz _trim_aritys,    rrcseZdZdZdZdZeddZeddZddd Z d d Z d d Z dddZ dddZ ddZddZddZddZddZddZddd Zd!d"Zdd#d$Zd%d&Zd'd(ZGd)d*d*eZed+k rGd,d-d-eZnGd.d-d-eZiZeZd/d/gZ dd0d1Z!eZ"ed2d3Z#dZ$edd5d6Z%dd7d8Z&e'dfd9d:Z(d;d<Z)e'fd=d>Z*e'dfd?d@Z+dAdBZ,dCdDZ-dEdFZ.dGdHZ/dIdJZ0dKdLZ1dMdNZ2dOdPZ3dQdRZ4dSdTZ5dUdVZ6dWdXZ7dYdZZ8dd[d\Z9d]d^Z:d_d`Z;dadbZdgdhZ?ddidjZ@dkdlZAdmdnZBdodpZCdqdrZDgfdsdtZEddudvZFfdwdxZGdydzZHd{d|ZId}d~ZJddZKdddZLdddZMZNS)r&z)Abstract base level parser element class.z FcCs |t_dS)a Overrides the default whitespace chars Example:: # default whitespace chars are space, and newline OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def', 'ghi', 'jkl'] # change to just treat newline as significant ParserElement.setDefaultWhitespaceChars(" \t") OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def'] N)r&DEFAULT_WHITE_CHARScharsryryrzsetDefaultWhitespaceCharsTs z'ParserElement.setDefaultWhitespaceCharscCs |t_dS)a Set class to be used for inclusion of string literals into a parser. Example:: # default literal class used is Literal integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # change to Suppress ParserElement.inlineLiteralsUsing(Suppress) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '12', '31'] N)r&_literalStringClass)rryryrzinlineLiteralsUsingcsz!ParserElement.inlineLiteralsUsingcCst|_d|_d|_d|_||_d|_tj|_ d|_ d|_ d|_ t|_ d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr)NNN)r parseAction failActionstrRepr resultsName saveAsListskipWhitespacer&rs whiteCharscopyDefaultWhiteCharsmayReturnEmptykeepTabs ignoreExprsdebug streamlined mayIndexErrorerrmsg modalResults debugActionsre callPreparse callDuringTry)rsavelistryryrzrxs(zParserElement.__init__cCs<t|}|jdd|_|jdd|_|jr8tj|_|S)a$ Make a copy of this C{ParserElement}. Useful for defining different parse actions for the same parsing pattern, using copies of the original parse element. Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K") integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M") print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M")) prints:: [5120, 100, 655360, 268435456] Equivalent form of C{expr.copy()} is just C{expr()}:: integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M") N)rryrrr&rsr)rZcpyryryrzrs  zParserElement.copycCs*||_d|j|_t|dr&|j|j_|S)af Define name for this expression, makes debugging and exception messages clearer. Example:: Word(nums).parseString("ABC") # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1) Word(nums).setName("integer").parseString("ABC") # -> Exception: Expected integer (at char 0), (line:1, col:1) Expected exception)rrrrrrryryrzsetNames    zParserElement.setNamecCs4|}|dr"|dd}d}||_| |_|S)aP Define name for referencing matching tokens as a nested attribute of the returned parse results. NOTE: this returns a *copy* of the original C{ParserElement} object; this is so that the client can define a basic element, such as an integer, and reference it in multiple places with different names. You can also set results names using the abbreviated syntax, C{expr("name")} in place of C{expr.setResultsName("name")} - see L{I{__call__}<__call__>}. Example:: date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: date_str = integer("year") + '/' + integer("month") + '/' + integer("day") *NrtT)rendswithr|r)rrlistAllMatchesZnewselfryryrzsetResultsNames  zParserElement.setResultsNameTcs@|r&|jdfdd }|_||_nt|jdr<|jj|_|S)zMethod to invoke the Python pdb debugger when this element is about to be parsed. Set C{breakFlag} to True to enable, False to disable. Tcsddl}|||||Sr)pdbZ set_trace)rRr doActions callPreParserZ _parseMethodryrzbreakersz'ParserElement.setBreak..breaker_originalParseMethod)TT)_parserr)rZ breakFlagrryrrzsetBreaks  zParserElement.setBreakcOs&tttt||_|dd|_|S)a Define one or more actions to perform when successfully matching parse element definition. Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)}, C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where: - s = the original string being parsed (see note below) - loc = the location of the matching substring - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object If the functions in fns modify the tokens, they can return them as the return value from fn, and the modified list of tokens will replace the original. Otherwise, fn does not need to return any value. Optional keyword arguments: - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See L{I{parseString}} for more information on parsing strings containing C{}s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. Example:: integer = Word(nums) date_str = integer + '/' + integer + '/' + integer date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # use parse action to convert to ints at parse time integer = Word(nums).setParseAction(lambda toks: int(toks[0])) date_str = integer + '/' + integer + '/' + integer # note that integer fields are now ints, not strings date_str.parseString("1999/12/31") # -> [1999, '/', 12, '/', 31] rF)rmaprrryrrrfnsr ryryrzrs"zParserElement.setParseActioncOs4|jtttt|7_|jp,|dd|_|S)z Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}}. See examples in L{I{copy}}. rF)ryrrrrrrrryryrzaddParseActionszParserElement.addParseActioncs^|dd|ddrtnt|D] fdd}|j|q$|jpV|dd|_|S)aAdd a boolean predicate function to expression's list of parse actions. See L{I{setParseAction}} for function call signatures. Unlike C{setParseAction}, functions passed to C{addCondition} need to return boolean success/fail of the condition. Optional keyword arguments: - message = define a custom message to be used in the raised exception - fatal = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) year_int = integer.copy() year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later") date_str = year_int + '/' + integer + '/' + integer result = date_str.parseString("1999/12/31") # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1) messagezfailed user-defined conditionfatalFcs$tt|||s ||dSr)r rrrZexc_typefnrryrzpa&sz&ParserElement.addCondition..par)rr#r!ryrr)rrr rryrrz addConditions zParserElement.addConditioncCs ||_|S)a Define action to perform if parsing fails at this expression. Fail acton fn is a callable function that takes the arguments C{fn(s,loc,expr,err)} where: - s = string being parsed - loc = location where expression match was attempted and failed - expr = the parse expression that failed - err = the exception thrown The function returns no value. It may throw C{L{ParseFatalException}} if it is desired to stop parsing immediately.)rz)rrryryrz setFailAction-s zParserElement.setFailActionc CsNd}|rJd}|jD]4}z|||\}}d}qWqtk rDYqXqq|SNTF)rrr!)rrRrZ exprsFoundeZdummyryryrz_skipIgnorables:s   zParserElement._skipIgnorablescCsH|jr|||}|jrD|j}t|}||krD|||krD|d7}q&|SNr)rrr~rr)rrRrZwtinstrlenryryrzpreParseGs  zParserElement.preParsecCs|gfSrryrrRrrryryrz parseImplSszParserElement.parseImplcCs|SrryrrRr tokenlistryryrz postParseVszParserElement.postParsec Cs|j}|s|jr|jdr,|jd||||rD|jrD|||}n|}|}zDz||||\}}Wn(tk rt|t||j |YnXWnXt k r} z:|jdr|jd|||| |jr||||| W5d} ~ XYnXn|r|jr|||}n|}|}|j s&|t|krjz||||\}}Wn*tk rft|t||j |YnXn||||\}}| |||}t ||j|j|jd} |jr|s|jr|rTzN|jD]B} | ||| }|dk rt ||j|jot|t tf|jd} qWnFt k rP} z&|jdr>|jd|||| W5d} ~ XYnXnJ|jD]B} | ||| }|dk rZt ||j|jot|t tf|jd} qZ|r|jdr|jd||||| || fS)Nrrs)rrr)rrzrrrrrr!rrrrrr$r|r}rryrr}r) rrRrrrZ debuggingprelocZ tokensStarttokenserrZ retTokensrryryrz _parseNoCacheZst             zParserElement._parseNoCachecCs@z|j||dddWStk r:t|||j|YnXdS)NF)rr)rr#r!rrrRrryryrztryParseszParserElement.tryParsec Cs4z|||Wnttfk r*YdSXdSdS)NFT)rr!rrryryrz canParseNexts zParserElement.canParseNextc@seZdZddZdS)zParserElement._UnboundedCachecs~it|_fdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS) Ncs |Srrrrcache not_in_cacheryrzrsz3ParserElement._UnboundedCache.__init__..getcs ||<dSrryrrrrryrzsetsz3ParserElement._UnboundedCache.__init__..setcs dSrrrrryrzrsz5ParserElement._UnboundedCache.__init__..clearcstSrrrrryrz cache_lensz9ParserElement._UnboundedCache.__init__..cache_len)rrtypes MethodTyperrrr)rrrrrryrrzrs    z&ParserElement._UnboundedCache.__init__Nrrrrryryryrz_UnboundedCachesrNc@seZdZddZdS)ParserElement._FifoCachecst|_tfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS) Ncs |Srrrrryrzrs.ParserElement._FifoCache.__init__..getcs>||<tkr:zdWqtk r6YqXqdSNF)rpopitemrr)rsizeryrzrs  .ParserElement._FifoCache.__init__..setcs dSrrrrryrzrs0ParserElement._FifoCache.__init__..clearcstSrrrrryrzrs4ParserElement._FifoCache.__init__..cache_len) rr _OrderedDictrrrrrrrrrrrrry)rrrrzrs   !ParserElement._FifoCache.__init__Nrryryryrz _FifoCachesrc@seZdZddZdS)rcst|_itgfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_ dS) Ncs |Srrrrryrzrsrcs4||<tkr&dq|dSr)rrpopleftrr)rkey_fiforryrzrs rcsdSrrr)rrryrzrsrcstSrrrrryrzrsr) rr collectionsdequerrrrrrrry)rrrrrzrs   rNrryryryrzrsrc Csd\}}|||||f}tjtj}||} | |jkrtj|d7<z|||||} Wn8tk r} z||| j | j W5d} ~ XYn.X||| d| d f| W5QRSn@tj|d7<t | t r| | d| d fW5QRSW5QRXdS)Nrrr)r&packrat_cache_lock packrat_cacherrpackrat_cache_statsrrrrmrrr}rp) rrRrrrZHITZMISSlookuprrrryryrz _parseCaches$   zParserElement._parseCachecCs(tjdgttjtjdd<dSr)r&rrrrryryryrz resetCaches zParserElement.resetCachecCs8tjs4dt_|dkr tt_n t|t_tjt_dS)aEnables "packrat" parsing, which adds memoizing to the parsing logic. Repeated parse attempts at the same string location (which happens often in many complex grammars) can immediately return a cached value, instead of re-executing parsing/validating code. Memoizing is done of both valid results and parsing exceptions. Parameters: - cache_size_limit - (default=C{128}) - if an integer value is provided will limit the size of the packrat cache; if None is passed, then the cache size will be unbounded; if 0 is passed, the cache will be effectively disabled. This speedup may break existing programs that use parse actions that have side-effects. For this reason, packrat parsing is disabled when you first import pyparsing. To activate the packrat feature, your program must call the class method C{ParserElement.enablePackrat()}. If your program uses C{psyco} to "compile as you go", you must call C{enablePackrat} before calling C{psyco.full()}. If you do not do this, Python will crash. For best results, call C{enablePackrat()} immediately after importing pyparsing. Example:: import pyparsing pyparsing.ParserElement.enablePackrat() TN)r&_packratEnabledrrrrr)Zcache_size_limitryryrz enablePackrat%s   zParserElement.enablePackratc Cst|js||jD] }|q|js8|}z<||d\}}|rr|||}t t }|||Wn0t k r}ztj rn|W5d}~XYnX|SdS)aB Execute the parse expression with the given string. This is the main interface to the client code, once the complete expression has been built. If you want the grammar to require that the entire input string be successfully parsed, then set C{parseAll} to True (equivalent to ending the grammar with C{L{StringEnd()}}). Note: C{parseString} implicitly calls C{expandtabs()} on the input string, in order to report proper column numbers in parse actions. If the input string contains tabs and the grammar uses parse actions that use the C{loc} argument to index into the string being parsed, you can ensure you have a consistent view of the input string by: - calling C{parseWithTabs} on your grammar before calling C{parseString} (see L{I{parseWithTabs}}) - define your parse action using the full C{(s,loc,toks)} signature, and reference the input string using the parse action's C{s} argument - explictly expand the tabs in your input string before calling C{parseString} Example:: Word('a').parseString('aaaaabaaa') # -> ['aaaaa'] Word('a').parseString('aaaaabaaa', parseAll=True) # -> Exception: Expected end of text rN) r&rr streamlinerr expandtabsrrrr+rverbose_stacktrace)rrRparseAllrrrZserXryryrz parseStringHs$    zParserElement.parseStringc cs6|js||jD] }|q|js4t|}t|}d}|j}|j}t d} z||kr| |krz |||} ||| dd\} } Wnt k r| d}YqZX| |kr| d7} | | | fV|r|||} | |kr| }q|d7}q| }qZ| d}qZWn4t k r0}zt j rn|W5d}~XYnXdS)a Scan the input string for expression matches. Each match will return the matching tokens, start location, and end location. May be called with optional C{maxMatches} argument, to clip scanning after 'n' matches are found. If C{overlap} is specified, then overlapping matches will be reported. Note that the start and end locations are reported relative to the string being parsed. See L{I{parseString}} for more information on parsing strings with embedded tabs. Example:: source = "sldjf123lsdjjkf345sldkjf879lkjsfd987" print(source) for tokens,start,end in Word(alphas).scanString(source): print(' '*start + '^'*(end-start)) print(' '*start + tokens[0]) prints:: sldjf123lsdjjkf345sldkjf879lkjsfd987 ^^^^^ sldjf ^^^^^^^ lsdjjkf ^^^^^^ sldkjf ^^^^^^ lkjsfd rFrrN)rrrrrrrrrr&rr!rr)rrR maxMatchesZoverlaprrrZ preparseFnZparseFnmatchesrZnextLocrZnextlocrXryryrz scanStringzsB       zParserElement.scanStringc Csg}d}d|_z||D]Z\}}}|||||rpt|trR||7}nt|trf||7}n |||}q|||ddd|D}dtt t |WSt k r}zt j rƂn|W5d}~XYnXdS)af Extension to C{L{scanString}}, to modify matching text with modified tokens that may be returned from a parse action. To use C{transformString}, define a grammar and attach a parse action to it that modifies the returned token list. Invoking C{transformString()} on a target string will then scan for matches, and replace the matched text patterns according to the logic in the parse action. C{transformString()} returns the resulting transformed string. Example:: wd = Word(alphas) wd.setParseAction(lambda toks: toks[0].title()) print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york.")) Prints:: Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York. rTNcSsg|] }|r|qSryry)roryryrzrsz1ParserElement.transformString..r)rrrr}r$rrrrr_flattenrr&r)rrRr*ZlastErxrrrXryryrzrs(    zParserElement.transformStringc CsRztdd|||DWStk rL}ztjr8n|W5d}~XYnXdS)a Another extension to C{L{scanString}}, simplifying the access to the tokens found to match the given parse expression. May be called with optional C{maxMatches} argument, to clip searching after 'n' matches are found. Example:: # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters cap_word = Word(alphas.upper(), alphas.lower()) print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")) # the sum() builtin can be used to merge results into a single ParseResults object print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))) prints:: [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']] ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity'] cSsg|]\}}}|qSryry)rrxrrryryrzrsz.ParserElement.searchString..N)r$rrr&r)rrRrrXryryrz searchStrings zParserElement.searchStringc csTd}d}|j||dD]*\}}}|||V|r<|dV|}q||dVdS)a[ Generator method to split a string using the given expression as a separator. May be called with optional C{maxsplit} argument, to limit the number of splits; and the optional C{includeSeparators} argument (default=C{False}), if the separating matching text should be included in the split results. Example:: punc = oneOf(list(".,;:/-!?")) print(list(punc.split("This, this?, this sentence, is badly punctuated!"))) prints:: ['This', ' this', '', ' this sentence', ' is badly punctuated', ''] r)rN)r) rrRmaxsplitZincludeSeparatorsZsplitsZlastrxrrryryrzrs  zParserElement.splitcCsFt|trt|}t|ts:tjdt|tdddSt||gS)a Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement converts them to L{Literal}s by default. Example:: greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) Prints:: Hello, World! -> ['Hello', ',', 'World', '!'] 4Cannot combine element of type %s with ParserElementrs stacklevelN) r}rr&rwwarningswarnr SyntaxWarningrr"ryryrzrs   zParserElement.__add__cCsBt|trt|}t|ts:tjdt|tdddS||S)z] Implementation of + operator when left operand is not a C{L{ParserElement}} rrsrNr}rr&rwrrrrr"ryryrzr#1s   zParserElement.__radd__cCsJt|trt|}t|ts:tjdt|tdddS|t |S)zQ Implementation of - operator, returns C{L{And}} with error stop rrsrN) r}rr&rwrrrrr _ErrorStopr"ryryrz__sub__=s   zParserElement.__sub__cCsBt|trt|}t|ts:tjdt|tdddS||S)z] Implementation of - operator when left operand is not a C{L{ParserElement}} rrsrNrr"ryryrz__rsub__Is   zParserElement.__rsub__cst|tr|d}}nt|tr|ddd}|ddkrHd|df}t|dtr|ddkr|ddkrvtS|ddkrtS|dtSqt|dtrt|dtr|\}}||8}qtdt|dt|dntdt||dkr td|dkrtd ||kr6dkrBnntd |rfd d |r|dkrt|}ntg||}n|}n|dkr}ntg|}|S) a Implementation of * operator, allows use of C{expr * 3} in place of C{expr + expr + expr}. Expressions may also me multiplied by a 2-integer tuple, similar to C{{min,max}} multipliers in regular expressions. Tuples may also include C{None} as in: - C{expr*(n,None)} or C{expr*(n,)} is equivalent to C{expr*n + L{ZeroOrMore}(expr)} (read as "at least n instances of C{expr}") - C{expr*(None,n)} is equivalent to C{expr*(0,n)} (read as "0 to n instances of C{expr}") - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)} - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)} Note that C{expr*(None,n)} does not raise an exception if more than n exprs exist in the input stream; that is, C{expr*(None,n)} does not enforce a maximum number of expr occurrences. If this behavior is desired, then write C{expr*(None,n) + ~expr} r)NNNrsrz7cannot multiply 'ParserElement' and ('%s','%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez+cannot multiply ParserElement by 0 or (0,0)cs(|dkrt|dStSdSr)rnmakeOptionalListrryrzrsz/ParserElement.__mul__..makeOptionalList) r}rvtupler4rrr ValueErrorr)rrZ minElementsZ optElementsrryrrz__mul__UsD             zParserElement.__mul__cCs ||Sr)rr"ryryrz__rmul__szParserElement.__rmul__cCsFt|trt|}t|ts:tjdt|tdddSt||gS)zI Implementation of | operator - returns C{L{MatchFirst}} rrsrN) r}rr&rwrrrrrr"ryryrz__or__s   zParserElement.__or__cCsBt|trt|}t|ts:tjdt|tdddS||BS)z] Implementation of | operator when left operand is not a C{L{ParserElement}} rrsrNrr"ryryrz__ror__s   zParserElement.__ror__cCsFt|trt|}t|ts:tjdt|tdddSt||gS)zA Implementation of ^ operator - returns C{L{Or}} rrsrN) r}rr&rwrrrrrr"ryryrz__xor__s   zParserElement.__xor__cCsBt|trt|}t|ts:tjdt|tdddS||AS)z] Implementation of ^ operator when left operand is not a C{L{ParserElement}} rrsrNrr"ryryrz__rxor__s   zParserElement.__rxor__cCsFt|trt|}t|ts:tjdt|tdddSt||gS)zC Implementation of & operator - returns C{L{Each}} rrsrN) r}rr&rwrrrrrr"ryryrz__and__s   zParserElement.__and__cCsBt|trt|}t|ts:tjdt|tdddS||@S)z] Implementation of & operator when left operand is not a C{L{ParserElement}} rrsrNrr"ryryrz__rand__s   zParserElement.__rand__cCst|S)zE Implementation of ~ operator - returns C{L{NotAny}} )rrryryrz __invert__szParserElement.__invert__cCs|dk r||S|SdS)a  Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}. If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be passed as C{True}. If C{name} is omitted, same as calling C{L{copy}}. Example:: # these are equivalent userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno") userdata = Word(alphas)("name") + Word(nums+"-")("socsecno") N)rrrryryrz__call__s zParserElement.__call__cCst|S)z Suppresses the output of this C{ParserElement}; useful to keep punctuation from cluttering up returned output. )r-rryryrzsuppressszParserElement.suppresscCs d|_|S)a Disables the skipping of whitespace before matching the characters in the C{ParserElement}'s defined pattern. This is normally only used internally by the pyparsing module, but may be needed in some whitespace-sensitive grammars. Fr~rryryrzleaveWhitespaceszParserElement.leaveWhitespacecCsd|_||_d|_|S)z8 Overrides the default whitespace chars TF)r~rr)rruryryrzsetWhitespaceChars sz ParserElement.setWhitespaceCharscCs d|_|S)z Overrides default behavior to expand C{}s to spaces before parsing the input string. Must be called before C{parseString} when the input grammar contains elements that match C{} characters. T)rrryryrz parseWithTabsszParserElement.parseWithTabscCsLt|trt|}t|tr4||jkrH|j|n|jt||S)a Define expression to be ignored (e.g., comments) while doing pattern matching; may be called repeatedly, to define multiple comment or other ignorable patterns. Example:: patt = OneOrMore(Word(alphas)) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj'] patt.ignore(cStyleComment) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd'] )r}rr-rrrr"ryryrzignores   zParserElement.ignorecCs"|pt|p t|ptf|_d|_|S)zT Enable display of debugging messages while doing pattern matching. T)rTrWrYrr)rZ startActionZ successActionZexceptionActionryryrzsetDebugActions6s zParserElement.setDebugActionscCs|r|tttnd|_|S)a Enable display of debugging messages while doing pattern matching. Set C{flag} to True to enable, False to disable. Example:: wd = Word(alphas).setName("alphaword") integer = Word(nums).setName("numword") term = wd | integer # turn on debugging for wd wd.setDebug() OneOrMore(term).parseString("abc 123 xyz 890") prints:: Match alphaword at loc 0(1,1) Matched alphaword -> ['abc'] Match alphaword at loc 3(1,4) Exception raised:Expected alphaword (at char 4), (line:1, col:5) Match alphaword at loc 7(1,8) Matched alphaword -> ['xyz'] Match alphaword at loc 11(1,12) Exception raised:Expected alphaword (at char 12), (line:1, col:13) Match alphaword at loc 15(1,16) Exception raised:Expected alphaword (at char 15), (line:1, col:16) The output shown is that produced by the default debug actions - custom debug actions can be specified using L{setDebugActions}. Prior to attempting to match the C{wd} expression, the debugging message C{"Match at loc (,)"} is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"} message is shown. Also note the use of L{setName} to assign a human-readable name to the expression, which makes debugging and exception messages easier to understand - for instance, the default name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}. F)rrTrWrYr)rflagryryrzsetDebug@s#zParserElement.setDebugcCs|jSr)rrryryrzriszParserElement.__str__cCst|SrrrryryrzrlszParserElement.__repr__cCsd|_d|_|Sr)rr{rryryrzroszParserElement.streamlinecCsdSrryrryryrzcheckRecursiontszParserElement.checkRecursioncCs|gdS)zj Check defined expressions for valid structure, check for infinite recursive definitions. N)r)r validateTraceryryrzvalidatewszParserElement.validatec Csz |}Wn2tk r>t|d}|}W5QRXYnXz|||WStk r~}ztjrjn|W5d}~XYnXdS)z Execute the parse expression on the given file or filename. If a filename is specified (instead of a file object), the entire file is opened, read, and closed before parsing. rN)readropenrrr&r)rZfile_or_filenamerZ file_contentsfrXryryrz parseFile}s  zParserElement.parseFilecsHt|tr"||kp t|t|kSt|tr6||Stt||kSdSr)r}r&varsrrsuperr"rmryrz__eq__s    zParserElement.__eq__cCs ||k Srryr"ryryrz__ne__szParserElement.__ne__cCs tt|Sr)hashidrryryrz__hash__szParserElement.__hash__cCs||kSrryr"ryryrz__req__szParserElement.__req__cCs ||k Srryr"ryryrz__rne__szParserElement.__rne__cCs4z|jt||dWdStk r.YdSXdS)a Method for quick testing of a parser against a test string. Good for simple inline microtests of sub expressions while building up larger parser. Parameters: - testString - to test against this expression for a match - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests Example:: expr = Word(nums) assert expr.matches("100") rTFN)rrr)rZ testStringrryryrzrs zParserElement.matches#c Cst|tr"tttj|}t|tr4t|}g}g}d} |D]} |dk r^| | dsf|rr| sr| | qD| sxqDd || g} g}z:| dd} |j | |d} | | j|d| o| } Wntk rr} zt| trdnd }d| kr*| t| j| | d t| j| d d |n| d | jd || d t| | o\|} | } W5d} ~ XYnDtk r}z$| dt|| o|} |} W5d}~XYnX|r|r| d td | | | | fqD| |fS)a3 Execute the parse expression on a series of test strings, showing each test, the parsed results or where the parse failed. Quick and easy way to run a parse expression against a list of sample strings. Parameters: - tests - a list of separate test strings, or a multiline string of test strings - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests - comment - (default=C{'#'}) - expression for indicating embedded comments in the test string; pass None to disable comment filtering - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline; if False, only dump nested list - printResults - (default=C{True}) prints test output to stdout - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing Returns: a (success, results) tuple, where success indicates that all tests succeeded (or failed if C{failureTests} is True), and the results contain a list of lines of each test's output Example:: number_expr = pyparsing_common.number.copy() result = number_expr.runTests(''' # unsigned integer 100 # negative integer -100 # float with scientific notation 6.02e23 # integer with scientific notation 1e-12 ''') print("Success" if result[0] else "Failed!") result = number_expr.runTests(''' # stray character 100Z # missing leading digit before '.' -.100 # too many '.' 3.14.159 ''', failureTests=True) print("Success" if result[0] else "Failed!") prints:: # unsigned integer 100 [100] # negative integer -100 [-100] # float with scientific notation 6.02e23 [6.02e+23] # integer with scientific notation 1e-12 [1e-12] Success # stray character 100Z ^ FAIL: Expected end of text (at char 3), (line:1, col:4) # missing leading digit before '.' -.100 ^ FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1) # too many '.' 3.14.159 ^ FAIL: Expected end of text (at char 4), (line:1, col:5) Success Each test string must be on a single line. If you want to test a string that spans multiple lines, create a test like this:: expr.runTest(r"this is a test\n of strings that spans \n 3 lines") (Note that this is a raw string literal, you must include the leading 'r'.) TNFr2\nr%)rDz(FATAL)r r^zFAIL: zFAIL-EXCEPTION: )r}rrrr~rrstrip splitlinesrrrrrrrArr#rIrr;rprQ)rZtestsrZcommentZfullDumpZ printResultsZ failureTestsZ allResultsZcommentssuccessrxr*resultrrrXryryrzrunTestssNW      $   zParserElement.runTests)F)F)T)T)TT)TT)r)F)N)T)F)T)Tr&TTF)Orrrrrsr staticmethodrvrxrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr_MAX_INTrrrrrr#rrrrrrrrrrrr r r r rrrrrrrrrrrrr"r#r$rr. __classcell__ryryrrzr&Os     &     G   " 2G+    D           )    cs eZdZdZfddZZS)r.zT Abstract C{ParserElement} subclass, for defining atomic matching patterns. cstt|jdddSNFr)rr.rrrryrzr@ szToken.__init__rrrrrr1ryryrrzr.< scs eZdZdZfddZZS)rz, An empty token, will always match. cs$tt|d|_d|_d|_dS)NrTF)rrrrrrrrryrzrH szEmpty.__init__r4ryryrrzrD scs*eZdZdZfddZdddZZS)rz( A token that will never match. cs*tt|d|_d|_d|_d|_dS)NrTFzUnmatchable token)rrrrrrrrrryrzrS s zNoMatch.__init__TcCst|||j|dSr)r!rrryryrzrZ szNoMatch.parseImpl)Trrrrrrr1ryryrrzrO s cs*eZdZdZfddZdddZZS)ra Token to exactly match a specified string. Example:: Literal('blah').parseString('blah') # -> ['blah'] Literal('blah').parseString('blahfooblah') # -> ['blah'] Literal('blah').parseString('bla') # -> Exception: Expected "blah" For case-insensitive matching, use L{CaselessLiteral}. For keyword matching (force word break before and after the matched string), use L{Keyword} or L{CaselessKeyword}. cstt|||_t||_z|d|_Wn*tk rVtj dt ddt |_ YnXdt |j|_d|j|_d|_d|_dS)Nrz2null string passed to Literal; use Empty() insteadrsr"%s"rF)rrrmatchrmatchLenfirstMatchCharrrrrrrmrrrrrr matchStringrryrzrl s   zLiteral.__init__TcCsJ|||jkr6|jdks&||j|r6||j|jfSt|||j|dSr)r9r8 startswithr7r!rrryryrzr s zLiteral.parseImpl)Tr5ryryrrzr^ s csLeZdZdZedZdfdd Zddd Zfd d Ze d d Z Z S)ra\ Token to exactly match a specified string as a keyword, that is, it must be immediately followed by a non-keyword character. Compare with C{L{Literal}}: - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}. - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'} Accepts two optional constructor arguments in addition to the keyword string: - C{identChars} is a string of characters that would be valid identifier characters, defaulting to all alphanumerics + "_" and "$" - C{caseless} allows case-insensitive matching, default is C{False}. Example:: Keyword("start").parseString("start") # -> ['start'] Keyword("start").parseString("starting") # -> Exception For case-insensitive matching, use L{CaselessKeyword}. _$NFcstt||dkrtj}||_t||_z|d|_Wn$tk r^t j dt ddYnXd|j|_ d|j |_ d|_d|_||_|r||_|}t||_dS)Nrz2null string passed to Keyword; use Empty() insteadrsrr6rF)rrrDEFAULT_KEYWORD_CHARSr7rr8r9rrrrrrrrcaselessupper caselessmatchr identChars)rr;rBr?rryrzr s*     zKeyword.__init__TcCs|jr|||||j|jkr|t||jksL|||j|jkr|dksj||d|jkr||j|jfSnv|||jkr|jdks||j|r|t||jks|||j|jkr|dks||d|jkr||j|jfSt |||j |dSr) r?r8r@rArrBr7r9r<r!rrryryrzr s4 zKeyword.parseImplcstt|}tj|_|Sr)rrrr>rB)rrrryrzr sz Keyword.copycCs |t_dS)z,Overrides the default Keyword chars N)rr>rtryryrzsetDefaultKeywordChars szKeyword.setDefaultKeywordChars)NF)T) rrrrr5r>rrrr/rCr1ryryrrzr s  cs*eZdZdZfddZdddZZS)r al Token to match a specified string, ignoring case of letters. Note: the matched results will always be in the case of the given match string, NOT the case of the input text. Example:: OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD'] (Contrast with example for L{CaselessKeyword}.) cs6tt||||_d|j|_d|j|_dS)Nz'%s'r)rr rr@ returnStringrrr:rryrzr s zCaselessLiteral.__init__TcCs@||||j|jkr,||j|jfSt|||j|dSr)r8r@r7rDr!rrryryrzr szCaselessLiteral.parseImpl)Tr5ryryrrzr s cs,eZdZdZdfdd Zd ddZZS) r z Caseless version of L{Keyword}. Example:: OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD'] (Contrast with example for L{CaselessLiteral}.) Ncstt|j||dddS)NTr?)rr r)rr;rBrryrzr szCaselessKeyword.__init__TcCsj||||j|jkrV|t||jksF|||j|jkrV||j|jfSt|||j|dSr)r8r@rArrBr7r!rrryryrzr szCaselessKeyword.parseImpl)N)Tr5ryryrrzr scs,eZdZdZdfdd Zd ddZZS) rnax A variation on L{Literal} which matches "close" matches, that is, strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters: - C{match_string} - string to be matched - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match The results from a successful parse will contain the matched text from the input string and the following named results: - C{mismatches} - a list of the positions within the match_string where mismatches were found - C{original} - the original match_string used to compare against the input string If C{mismatches} is an empty list, then the match was an exact match. Example:: patt = CloseMatch("ATCATCGAATGGA") patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']}) patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1) # exact match patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']}) # close match allowing up to 2 mismatches patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2) patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']}) rcsBtt|||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F) rrnrr match_string maxMismatchesrrr)rrFrGrryrzr szCloseMatch.__init__TcCs|}t|}|t|j}||kr|j}d}g} |j} tt||||jD]2\}} | \} } | | krN| |t| | krNqqN|d}t|||g}|j|d<| |d<||fSt|||j|dS)Nrroriginal mismatches) rrFrGrrrr$r!r)rrRrrstartrmaxlocrFZmatch_stringlocrIrGZs_msrcmatresultsryryrzr s(    zCloseMatch.parseImpl)r)Tr5ryryrrzrn s cs8eZdZdZd fdd Zdd d Zfd d ZZS)r1a Token for matching words composed of allowed character sets. Defined with string containing all allowed initial characters, an optional string containing allowed body characters (if omitted, defaults to the initial character set), and an optional minimum, maximum, and/or exact length. The default value for C{min} is 1 (a minimum value < 1 is not valid); the default values for C{max} and C{exact} are 0, meaning no maximum or exact length restriction. An optional C{excludeChars} parameter can list characters that might be found in the input C{bodyChars} string; useful to define a word of all printables except for one or two characters, for instance. L{srange} is useful for defining custom character set strings for defining C{Word} expressions, using range notation from regular expression character sets. A common mistake is to use C{Word} to match a specific literal string, as in C{Word("Address")}. Remember that C{Word} uses the string argument to define I{sets} of matchable characters. This expression would match "Add", "AAA", "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an exact literal string, use L{Literal} or L{Keyword}. pyparsing includes helper strings for building Words: - L{alphas} - L{nums} - L{alphanums} - L{hexnums} - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.) - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.) - L{printables} (any non-whitespace character) Example:: # a word composed of digits integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9")) # a word with a leading capital, and zero or more lowercase capital_word = Word(alphas.upper(), alphas.lower()) # hostnames are alphanumeric, with leading alpha, and '-' hostname = Word(alphas, alphanums+'-') # roman numeral (not a strict parser, accepts invalid mix of characters) roman = Word("IVXLCDM") # any string of non-whitespace characters, except for ',' csv_value = Word(printables, excludeChars=",") NrrFcstt|rFdfdd|D}|rFdfdd|D}||_t||_|rl||_t||_n||_t||_|dk|_ |dkrt d||_ |dkr||_ nt |_ |dkr||_ ||_ t||_d|j|_d |_||_d |j|jkr|dkr|dkr|dkr|j|jkr8d t|j|_nHt|jdkrfd t|jt|jf|_nd t|jt|jf|_|jrd|jd|_zt|j|_Wntk rd|_YnXdS)Nrc3s|]}|kr|VqdSrryr excludeCharsryrzr` sz Word.__init__..c3s|]}|kr|VqdSrryrrOryrzrb srrzZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedrFr(z[%s]+z%s[%s]*z [%s][%s]*z\b)rr1rr initCharsOrigr initChars bodyCharsOrig bodyChars maxSpecifiedrminLenmaxLenr0rrrr asKeyword_escapeRegexRangeCharsreStringrrescapecompilerp)rrRrTminmaxexactrXrPrrOrzr] s\      0 z Word.__init__Tc Cs>|jr<|j||}|s(t|||j||}||fS|||jkrZt|||j||}|d7}t|}|j}||j }t ||}||kr|||kr|d7}qd} |||j krd} |j r||kr|||krd} |j r|dkr||d|ks||kr|||krd} | r.t|||j|||||fS)NrFTr)rr7r!rendgrouprRrrTrWr]rVrUrX) rrRrrr-rJrZ bodycharsrKZthrowExceptionryryrzr s6    2zWord.parseImplcsvztt|WStk r$YnX|jdkrpdd}|j|jkr`d||j||jf|_nd||j|_|jS)NcSs$t|dkr|dddS|SdS)N...rrryryrz charsAsStr s z Word.__str__..charsAsStrz W:(%s,%s)zW:(%s))rr1rrpr{rQrS)rrerryrzr s  z Word.__str__)NrrrFN)Trrrrrrrr1ryryrrzr1. s.6 #csFeZdZdZeedZd fdd Zd ddZ fd d Z Z S) r)a Token for matching strings that match a given regular expression. Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module. If the given regex contains named groups (defined using C{(?P...)}), these will be preserved as named parse results. Example:: realnum = Regex(r"[+-]?\d+\.\d*") date = Regex(r'(?P\d{4})-(?P\d\d?)-(?P\d\d?)') # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})") z[A-Z]rcstt|t|tr|s,tjdtdd||_||_ zt |j|j |_ |j|_ Wqt jk rtjd|tddYqXn2t|tjr||_ t||_|_ ||_ ntdt||_d|j|_d|_d|_d S) zThe parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags.z0null string passed to Regex; use Empty() insteadrsr$invalid pattern (%s) passed to RegexzCRegex may only be constructed with a string or a compiled RE objectrFTN)rr)rr}rrrrpatternflagsrr\rZ sre_constantserrorcompiledREtyper~rrrrrr)rrhrirryrzr s:       zRegex.__init__TcCs`|j||}|s"t|||j||}|}t|}|rX|D]}||||<qF||fSr)rr7r!rr` groupdictr$ra)rrRrrr-drrryryrzr s zRegex.parseImplcsFztt|WStk r$YnX|jdkr@dt|j|_|jS)NzRe:(%s))rr)rrpr{rrhrrryrzr s z Regex.__str__)r)T) rrrrrrr\rlrrrr1ryryrrzr) s  " cs8eZdZdZd fdd Zd ddZfd d ZZS) r'a Token for matching strings that are delimited by quoting characters. Defined with the following parameters: - quoteChar - string of one or more characters defining the quote delimiting string - escChar - character to escape quotes, typically backslash (default=C{None}) - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None}) - multiline - boolean indicating whether quotes can span multiple lines (default=C{False}) - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True}) - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar) - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True}) Example:: qs = QuotedString('"') print(qs.searchString('lsjdf "This is the quote" sldjf')) complex_qs = QuotedString('{{', endQuoteChar='}}') print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf')) sql_qs = QuotedString('"', escQuote='""') print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf')) prints:: [['This is the quote']] [['This is the "quote"']] [['This is the quote with "embedded" quotes']] NFTc sNtt|}|s0tjdtddt|dkr>|}n"|}|s`tjdtddt|_t |_ |d_ |_ t |_ |_|_|_|_|rtjtjB_dtjtj d|dk rt|pdf_n.rt)z|(?:%s)z|(?:%s.)z(.)z)*%srgrFT)%rr'rrrrr SyntaxError quoteCharr quoteCharLenfirstQuoteCharroendQuoteCharLenescCharescQuoteunquoteResultsconvertWhitespaceEscapesr MULTILINEDOTALLrir[rYrhrrescCharReplacePatternr\rZrjrkrrrrr)rrrrvrwZ multilinerxroryrrrzr/ s|           zQuotedString.__init__c Cs|||jkr|j||pd}|s4t|||j||}|}|jr||j|j }t |t rd|kr|j rddddd}| D]\}}|||}q|jrt|jd|}|jr||j|j}||fS)N\ r2  )\tr'z\fz\rz\g<1>)rtrr7r!rr`rarxrsrur}rryrrrvrr|rwro) rrRrrr-rZws_mapZwslitZwscharryryrzrp s*  zQuotedString.parseImplcsHztt|WStk r$YnX|jdkrBd|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)rr'rrpr{rrrorrryrzr s zQuotedString.__str__)NNFTNT)Trfryryrrzr' sA #cs8eZdZdZd fdd Zd ddZfd d ZZS) r a Token for matching words composed of characters I{not} in a given set (will include whitespace in matched characters if not listed in the provided exclusion set - see example). Defined with string containing all disallowed characters, and an optional minimum, maximum, and/or exact length. The default value for C{min} is 1 (a minimum value < 1 is not valid); the default values for C{max} and C{exact} are 0, meaning no maximum or exact length restriction. Example:: # define a comma-separated-value as anything that is not a ',' csv_value = CharsNotIn(',') print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213")) prints:: ['dkls', 'lsdkjf', 's12 34', '@!#', '213'] rrcstt|d|_||_|dkr*td||_|dkr@||_nt|_|dkrZ||_||_t ||_ d|j |_ |jdk|_ d|_ dS)NFrzfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr)rr rr~notCharsrrVrWr0rrrrr)rrr]r^r_rryrzr s    zCharsNotIn.__init__TcCs|||jkrt|||j||}|d7}|j}t||jt|}||krb|||krb|d7}qD|||jkrt|||j|||||fSr)rr!rr]rWrrV)rrRrrrJZnotcharsmaxlenryryrzr s  zCharsNotIn.parseImplcsfztt|WStk r$YnX|jdkr`t|jdkrTd|jdd|_n d|j|_|jS)Nrbz !W:(%s...)z!W:(%s))rr rrpr{rrrrryrzr s  zCharsNotIn.__str__)rrr)Trfryryrrzr s cs<eZdZdZddddddZdfd d ZdddZZS)r0a Special matching class for matching whitespace. Normally, whitespace is ignored by pyparsing grammars. This class is included when some whitespace structures are significant. Define with a string containing the whitespace characters to be matched; default is C{" \t\r\n"}. Also takes optional C{min}, C{max}, and C{exact} arguments, as defined for the C{L{Word}} class. zzzzz)r(r~r2rr rrcstt|_dfddjDdddjD_d_dj_ |_ |dkrt|_ nt _ |dkr|_ |_ dS)Nrc3s|]}|jkr|VqdSr) matchWhiterrryrzr s z!White.__init__..css|]}tj|VqdSr)r0 whiteStrsrryryrzr sTrr) rr0rrr rrrrrrVrWr0)rZwsr]r^r_rrrzr s  zWhite.__init__TcCs|||jkrt|||j||}|d7}||j}t|t|}||krb|||jkrb|d7}qB|||jkrt|||j|||||fSr)rr!rrWr]rrV)rrRrrrJrKryryrzr s  zWhite.parseImpl)rrrr)T)rrrrrrrr1ryryrrzr0 scseZdZfddZZS)_PositionTokencs(tt||jj|_d|_d|_dSr)rrrrmrrrrrrryrzr s z_PositionToken.__init__rrrrr1ryryrrzr srcs2eZdZdZfddZddZd ddZZS) rzb Token to advance to a specific column of input text; useful for tabular report scraping. cstt|||_dSr)rrrr;)rcolnorryrzr$ szGoToColumn.__init__cCs\t|||jkrXt|}|jr*|||}||krX||rXt|||jkrX|d7}q*|Sr)r;rrrisspace)rrRrrryryrzr( s $ zGoToColumn.preParseTcCsDt||}||jkr"t||d|||j|}|||}||fS)NzText not in expected columnr;r!)rrRrrZthiscolZnewlocrryryrzr1 s    zGoToColumn.parseImpl)T)rrrrrrrr1ryryrrzr s  cs*eZdZdZfddZdddZZS)ra Matches if current position is at the beginning of a line within the parse string Example:: test = ''' AAA this line AAA and this line AAA but not this one B AAA and definitely not this one ''' for t in (LineStart() + 'AAA' + restOfLine).searchString(test): print(t) Prints:: ['AAA', ' this line'] ['AAA', ' and this line'] cstt|d|_dS)NzExpected start of line)rrrrrrryrzrO szLineStart.__init__TcCs*t||dkr|gfSt|||j|dSr)r;r!rrryryrzrS szLineStart.parseImpl)Tr5ryryrrzr: s cs*eZdZdZfddZdddZZS)rzU Matches if current position is at the end of a line within the parse string cs,tt||tjddd|_dS)Nr2rzExpected end of line)rrrr r&rsrrrrryrzr\ szLineEnd.__init__TcCsb|t|kr6||dkr$|ddfSt|||j|n(|t|krN|dgfSt|||j|dS)Nr2rrr!rrryryrzra s     zLineEnd.parseImpl)Tr5ryryrrzrX s cs*eZdZdZfddZdddZZS)r,zM Matches if current position is at the beginning of the parse string cstt|d|_dS)NzExpected start of text)rr,rrrrryrzrp szStringStart.__init__TcCs0|dkr(|||dkr(t|||j||gfSr)rr!rrryryrzrt szStringStart.parseImpl)Tr5ryryrrzr,l s cs*eZdZdZfddZdddZZS)r+zG Matches if current position is at the end of the parse string cstt|d|_dS)NzExpected end of text)rr+rrrrryrzr szStringEnd.__init__TcCs^|t|krt|||j|n<|t|kr6|dgfS|t|krJ|gfSt|||j|dSrrrryryrzr s    zStringEnd.parseImpl)Tr5ryryrrzr+{ s cs.eZdZdZeffdd ZdddZZS)r3ap Matches if the current position is at the beginning of a Word, and is not preceded by any character in a given set of C{wordChars} (default=C{printables}). To emulate the C{} behavior of regular expressions, use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of the string being parsed, or at the beginning of a line. cs"tt|t||_d|_dS)NzNot at the start of a word)rr3rr wordCharsrrrrryrzr s zWordStart.__init__TcCs@|dkr8||d|jks(|||jkr8t|||j||gfSr)rr!rrryryrzr s  zWordStart.parseImpl)TrrrrrXrrr1ryryrrzr3 scs.eZdZdZeffdd ZdddZZS)r2aZ Matches if the current position is at the end of a Word, and is not followed by any character in a given set of C{wordChars} (default=C{printables}). To emulate the C{} behavior of regular expressions, use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of the string being parsed, or at the end of a line. cs(tt|t||_d|_d|_dS)NFzNot at the end of a word)rr2rrrr~rrrryrzr s zWordEnd.__init__TcCsPt|}|dkrH||krH|||jks8||d|jkrHt|||j||gfSr)rrr!r)rrRrrrryryrzr szWordEnd.parseImpl)Trryryrrzr2 scseZdZdZdfdd ZddZddZd d Zfd d Zfd dZ fddZ dfdd Z gfddZ fddZ ZS)r"z^ Abstract subclass of ParserElement, for combining and post-processing parsed tokens. Fcstt||t|tr"t|}t|tr.F)rr"rr}rrrr&rwexprsrallrrrrrrrryrzr s     zParseExpression.__init__cCs |j|Sr)rrryryrzr szParseExpression.__getitem__cCs|j|d|_|Sr)rrr{r"ryryrzr s zParseExpression.appendcCs0d|_dd|jD|_|jD] }|q|S)z~Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on all contained expressions.FcSsg|] }|qSryrrrryryrzr sz3ParseExpression.leaveWhitespace..)r~rr )rrryryrzr  s   zParseExpression.leaveWhitespacecsrt|trB||jkrntt|||jD]}||jdq*n,tt|||jD]}||jdqX|Sr)r}r-rrr"rr)rrrrryrzr s    zParseExpression.ignorecsNztt|WStk r$YnX|jdkrHd|jjt|jf|_|jSNz%s:(%s)) rr"rrpr{rmrrrrrryrzr s zParseExpression.__str__cs*tt||jD] }|qt|jdkr|jd}t||jr|js|jdkr|j s|jdd|jdg|_d|_ |j |j O_ |j |j O_ |jd}t||jr|js|jdkr|j s|jdd|jdd|_d|_ |j |j O_ |j |j O_ dt ||_|S)Nrsrrrtr)rr"rrrr}rmryr|rr{rrrr)rrrrryrzr s<     zParseExpression.streamlinecstt|||}|Sr)rr"r)rrrrrryrzr szParseExpression.setResultsNamecCs6|dd|g}|jD]}||q|gdSr)rrr)rrtmprryryrzr s  zParseExpression.validatecs$tt|}dd|jD|_|S)NcSsg|] }|qSryrrryryrzr% sz(ParseExpression.copy..)rr"rrr1rryrzr# szParseExpression.copy)F)F)rrrrrrrr rrrrrrr1ryryrrzr" s " csTeZdZdZGdddeZdfdd ZdddZd d Zd d Z d dZ Z S)ra  Requires all given C{ParseExpression}s to be found in the given order. Expressions may be separated by whitespace. May be constructed using the C{'+'} operator. May also be constructed using the C{'-'} operator, which will suppress backtracking. Example:: integer = Word(nums) name_expr = OneOrMore(Word(alphas)) expr = And([integer("id"),name_expr("name"),integer("age")]) # more easily written as: expr = integer("id") + name_expr("name") + integer("age") cseZdZfddZZS)zAnd._ErrorStopcs&ttj|j||d|_|dS)N-)rrrrrr rGrryrzr9 szAnd._ErrorStop.__init__rryryrrzr8 srTcsRtt|||tdd|jD|_||jdj|jdj|_d|_ dS)Ncss|] }|jVqdSrrrryryrzr@ szAnd.__init__..rT) rrrrrrr rr~rrrryrzr> s z And.__init__c Cs|jdj|||dd\}}d}|jddD]}t|tjrDd}q.|rz||||\}}Wqtk rtYqtk r}zd|_t|W5d}~XYqt k rt|t ||j |YqXn||||\}}|s| r.||7}q.||fS)NrFrrT) rrr}rrr%r __traceback__rrrrr ) rrRrr resultlistZ errorStoprZ exprtokensrryryrzrE s(   z And.parseImplcCst|trt|}||Srr}rr&rwrr"ryryrzr!^ s  z And.__iadd__cCs6|dd|g}|jD]}|||jsq2qdSr)rrrrrsubRecCheckListrryryrzrc s   zAnd.checkRecursioncCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nr{r(css|]}t|VqdSrrrryryrzro szAnd.__str__..}rrr{rrrryryrzrj s    z And.__str__)T)T) rrrrrrrrr!rrr1ryryrrzr( s csDeZdZdZdfdd ZdddZdd Zd d Zd d ZZ S)ra Requires that at least one C{ParseExpression} is found. If two expressions match, the expression that matches the longest string will be used. May be constructed using the C{'^'} operator. Example:: # construct Or using '^' operator number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) prints:: [['123'], ['3.1416'], ['789']] Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|] }|jVqdSrrrryryrzr szOr.__init__..T)rrrrrBrrrryrzr sz Or.__init__Tc CsRd}d}g}|jD]}z|||}Wnvtk rb} zd| _| j|krR| }| j}W5d} ~ XYqtk rt||krt|t||j|}t|}YqX|||fq|r(|j ddd|D]^\} }z| |||WStk r$} z d| _| j|kr| }| j}W5d} ~ XYqXq|dk r@|j|_ |nt||d|dS)NrtcSs |d Srry)xryryrzr{ r|zOr.parseImpl..)r no defined alternatives to match) rrr!rrrrrrsortrr) rrRrr maxExcLoc maxExceptionrrZloc2r_ryryrzr s<      z Or.parseImplcCst|trt|}||Srrr"ryryrz__ixor__ s  z Or.__ixor__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz ^ css|]}t|VqdSrrrryryrzr szOr.__str__..rrrryryrzr s    z Or.__str__cCs,|dd|g}|jD]}||qdSrrrrryryrzr s zOr.checkRecursion)F)T) rrrrrrrrrr1ryryrrzrt s   & csDeZdZdZdfdd ZdddZdd Zd d Zd d ZZ S)ra Requires that at least one C{ParseExpression} is found. If two expressions match, the first one listed is the one that will match. May be constructed using the C{'|'} operator. Example:: # construct MatchFirst using '|' operator # watch the order of expressions to match number = Word(nums) | Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) # Fail! -> [['123'], ['3'], ['1416'], ['789']] # put more selective expression first number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums) print(number.searchString("123 3.1416 789")) # Better -> [['123'], ['3.1416'], ['789']] Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|] }|jVqdSrrrryryrzr sz&MatchFirst.__init__..T)rrrrrBrrrryrzr szMatchFirst.__init__Tc Csd}d}|jD]}z||||}|WStk r`}z|j|krP|}|j}W5d}~XYqtk rt||krt|t||j|}t|}YqXq|dk r|j|_|nt||d|dS)Nrtr)rrr!rrrrr) rrRrrrrrrrryryrzr s$    zMatchFirst.parseImplcCst|trt|}||Srrr"ryryrz__ior__ s  zMatchFirst.__ior__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrr | css|]}t|VqdSrrrryryrzr sz%MatchFirst.__str__..rrrryryrzr s    zMatchFirst.__str__cCs,|dd|g}|jD]}||qdSrrrryryrzrs zMatchFirst.checkRecursion)F)T) rrrrrrrrrr1ryryrrzr s   cs<eZdZdZd fdd Zd ddZddZd d ZZS) ram Requires all given C{ParseExpression}s to be found, but in any order. Expressions may be separated by whitespace. May be constructed using the C{'&'} operator. Example:: color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN") shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON") integer = Word(nums) shape_attr = "shape:" + shape_type("shape") posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn") color_attr = "color:" + color("color") size_attr = "size:" + integer("size") # use Each (using operator '&') to accept attributes in any order # (shape and posn are required, color and size are optional) shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr) shape_spec.runTests(''' shape: SQUARE color: BLACK posn: 100, 120 shape: CIRCLE size: 50 color: BLUE posn: 50,80 color:GREEN size:20 shape:TRIANGLE posn:20,40 ''' ) prints:: shape: SQUARE color: BLACK posn: 100, 120 ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']] - color: BLACK - posn: ['100', ',', '120'] - x: 100 - y: 120 - shape: SQUARE shape: CIRCLE size: 50 color: BLUE posn: 50,80 ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']] - color: BLUE - posn: ['50', ',', '80'] - x: 50 - y: 80 - shape: CIRCLE - size: 50 color: GREEN size: 20 shape: TRIANGLE posn: 20,40 ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']] - color: GREEN - posn: ['20', ',', '40'] - x: 20 - y: 40 - shape: TRIANGLE - size: 20 Tcs8tt|||tdd|jD|_d|_d|_dS)Ncss|] }|jVqdSrrrryryrzr?sz Each.__init__..T)rrrrrrr~initExprGroupsrrryrzr=sz Each.__init__c s|jrtdd|jD|_dd|jD}dd|jD}|||_dd|jD|_dd|jD|_dd|jD|_|j|j7_d |_|}|jdd}|jddg}d } | rj||j|j} g} | D]v} z| ||}Wn t k r| | YqX| |j t | | | |kr@| | q| kr܈ | qt| t| krd } q|rd d d|D} t ||d | |fdd|jD7}g}|D]"} | |||\}}| |qt|tg}||fS)Ncss&|]}t|trt|j|fVqdSr)r}rr!rSrryryrzrEs z!Each.parseImpl..cSsg|]}t|tr|jqSryr}rrSrryryrzrFs z"Each.parseImpl..cSs g|]}|jrt|ts|qSry)rr}rrryryrzrGs cSsg|]}t|tr|jqSry)r}r4rSrryryrzrIs cSsg|]}t|tr|jqSry)r}rrSrryryrzrJs cSs g|]}t|tttfs|qSry)r}rr4rrryryrzrKsFTr%css|]}t|VqdSrrrryryrzrfsz*Missing one or more required elements (%s)cs$g|]}t|tr|jkr|qSryrrZtmpOptryrzrjs )rrrZopt1mapZ optionalsZmultioptionalsZ multirequiredZrequiredrr!rrr!removerrrsumr$)rrRrrZopt1Zopt2ZtmpLocZtmpReqdZ matchOrderZ keepMatchingZtmpExprsZfailedrZmissingrrNZ finalResultsryrrzrCsP    zEach.parseImplcCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz & css|]}t|VqdSrrrryryrzryszEach.__str__..rrrryryrzrts    z Each.__str__cCs,|dd|g}|jD]}||qdSrrrryryrzr}s zEach.checkRecursion)T)T) rrrrrrrrr1ryryrrzrs 5 1 csleZdZdZdfdd ZdddZdd Zfd d Zfd d ZddZ gfddZ fddZ Z S)r za Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens. Fcstt||t|tr@ttjtr2t|}ntt |}||_ d|_ |dk r|j |_ |j |_ ||j|j|_|j|_|j|_|j|jdSr)rr rr}r issubclassr&rwr.rrSr{rrr rr~r}rrrrrSrrryrzrs    zParseElementEnhance.__init__TcCs2|jdk r|jj|||ddStd||j|dS)NFrr)rSrr!rrryryrzrs zParseElementEnhance.parseImplcCs*d|_|j|_|jdk r&|j|Sr)r~rSrr rryryrzr s    z#ParseElementEnhance.leaveWhitespacecsrt|trB||jkrntt|||jdk rn|j|jdn,tt|||jdk rn|j|jd|Sr)r}r-rrr rrSr"rryrzrs    zParseElementEnhance.ignorecs&tt||jdk r"|j|Sr)rr rrSrrryrzrs  zParseElementEnhance.streamlinecCsB||krt||g|dd|g}|jdk r>|j|dSr)r(rSr)rrrryryrzrs  z"ParseElementEnhance.checkRecursioncCs6|dd|g}|jdk r(|j||gdSrrSrrrrrryryrzrs  zParseElementEnhance.validatecsXztt|WStk r$YnX|jdkrR|jdk rRd|jjt|jf|_|jSr) rr rrpr{rSrmrrrrryrzrszParseElementEnhance.__str__)F)T) rrrrrrr rrrrrr1ryryrrzr s   cs*eZdZdZfddZdddZZS)ra Lookahead matching of the given parse expression. C{FollowedBy} does I{not} advance the parsing position within the input string, it only verifies that the specified parse expression matches at the current position. C{FollowedBy} always returns a null token list. Example:: # use FollowedBy to match a label only if it is followed by a ':' data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint() prints:: [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']] cstt||d|_dSr)rrrrrrSrryrzrszFollowedBy.__init__TcCs|j|||gfSr)rSrrryryrzrszFollowedBy.parseImpl)Tr5ryryrrzrs cs2eZdZdZfddZd ddZddZZS) ra Lookahead to disallow matching with the given parse expression. C{NotAny} does I{not} advance the parsing position within the input string, it only verifies that the specified parse expression does I{not} match at the current position. Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny} always returns a null token list. May be constructed using the '~' operator. Example:: cs0tt||d|_d|_dt|j|_dS)NFTzFound unwanted token, )rrrr~rrrSrrrryrzrszNotAny.__init__TcCs&|j||rt|||j||gfSr)rSrr!rrryryrzrszNotAny.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrz~{rrrr{rrSrryryrzrs   zNotAny.__str__)Trfryryrrzrs  cs(eZdZdfdd ZdddZZS) _MultipleMatchNcsFtt||d|_|}t|tr.t|}|dk r<|nd|_dSr) rrrr}r}rr&rw not_ender)rrSstopOnZenderrryrzr s   z_MultipleMatch.__init__Tc Cs|jj}|j}|jdk }|r$|jj}|r2|||||||dd\}}zV|j } |r`|||| rp|||} n|} ||| |\}} | s| rR|| 7}qRWnttfk rYnX||fSNFr) rSrrrrrr r!r) rrRrrZself_expr_parseZself_skip_ignorablesZ check_enderZ try_not_enderrZhasIgnoreExprsrZ tmptokensryryrzrs*      z_MultipleMatch.parseImpl)N)T)rrrrrr1ryryrrzr src@seZdZdZddZdS)ra Repetition of one or more of the given expression. Parameters: - expr - expression that must match one or more times - stopOn - (default=C{None}) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: BLACK" OneOrMore(attr_expr).parseString(text).pprint() # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']] # use stopOn attribute for OneOrMore to avoid reading label string as part of the data attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']] # could also be written as (attr_expr * (1,)).parseString(text).pprint() cCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrrz}...rrryryrzrJs   zOneOrMore.__str__N)rrrrrryryryrzr0scs8eZdZdZd fdd Zd fdd Zdd ZZS) r4aw Optional repetition of zero or more of the given expression. Parameters: - expr - expression that must match zero or more times - stopOn - (default=C{None}) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example: similar to L{OneOrMore} Ncstt|j||dd|_dS)N)rT)rr4rr)rrSrrryrzr_szZeroOrMore.__init__Tc s<ztt||||WSttfk r6|gfYSXdSr)rr4rr!rrrryrzrcszZeroOrMore.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrr$]...rrryryrzris   zZeroOrMore.__str__)N)Trfryryrrzr4Ss c@s eZdZddZeZddZdS) _NullTokencCsdSrryrryryrzrssz_NullToken.__bool__cCsdSrryrryryrzrvsz_NullToken.__str__N)rrrrrLrryryryrzrrsrcs6eZdZdZeffdd Zd ddZddZZS) raa Optional matching of the given expression. Parameters: - expr - expression that must match zero or more times - default (optional) - value to be returned if the optional expression is not found. Example:: # US postal code can be a 5-digit zip, plus optional 4-digit qualifier zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4))) zip.runTests(''' # traditional ZIP code 12345 # ZIP+4 form 12101-0001 # invalid ZIP 98765- ''') prints:: # traditional ZIP code 12345 ['12345'] # ZIP+4 form 12101-0001 ['12101-0001'] # invalid ZIP 98765- ^ FAIL: Expected end of text (at char 5), (line:1, col:6) cs.tt|j|dd|jj|_||_d|_dS)NFr3T)rrrrSr}rr)rrSr rryrzrs zOptional.__init__Tc Cszz|jj|||dd\}}WnTttfk rp|jtk rh|jjr^t|jg}|j||jj<ql|jg}ng}YnX||fSr)rSrr!rr_optionalNotMatchedr|r$)rrRrrrryryrzrs    zOptional.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrr$r'rrryryrzrs   zOptional.__str__)T) rrrrrrrrr1ryryrrzrzs" cs,eZdZdZd fdd Zd ddZZS) r*a Token for skipping over all undefined text until the matched expression is found. Parameters: - expr - target expression marking the end of the data to be skipped - include - (default=C{False}) if True, the target expression is also parsed (the skipped text and target expression are returned as a 2-element list). - ignore - (default=C{None}) used to define grammars (typically quoted strings and comments) that might contain false matches to the target expression - failOn - (default=C{None}) define expressions that are not allowed to be included in the skipped test; if found before the target expression is found, the SkipTo is not a match Example:: report = ''' Outstanding Issues Report - 1 Jan 2000 # | Severity | Description | Days Open -----+----------+-------------------------------------------+----------- 101 | Critical | Intermittent system crash | 6 94 | Cosmetic | Spelling error on Login ('log|n') | 14 79 | Minor | System slow when running too many reports | 47 ''' integer = Word(nums) SEP = Suppress('|') # use SkipTo to simply match everything up until the next SEP # - ignore quoted strings, so that a '|' character inside a quoted string does not match # - parse action will call token.strip() for each matched token, i.e., the description body string_data = SkipTo(SEP, ignore=quotedString) string_data.setParseAction(tokenMap(str.strip)) ticket_expr = (integer("issue_num") + SEP + string_data("sev") + SEP + string_data("desc") + SEP + integer("days_open")) for tkt in ticket_expr.searchString(report): print tkt.dump() prints:: ['101', 'Critical', 'Intermittent system crash', '6'] - days_open: 6 - desc: Intermittent system crash - issue_num: 101 - sev: Critical ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14'] - days_open: 14 - desc: Spelling error on Login ('log|n') - issue_num: 94 - sev: Cosmetic ['79', 'Minor', 'System slow when running too many reports', '47'] - days_open: 47 - desc: System slow when running too many reports - issue_num: 79 - sev: Minor FNcs`tt||||_d|_d|_||_d|_t|t rFt ||_ n||_ dt |j|_dS)NTFzNo match found for )rr*r ignoreExprrr includeMatchrr}rr&rwfailOnrrSr)rrZincluderrrryrzrs zSkipTo.__init__Tc Cs&|}t|}|j}|jj}|jdk r,|jjnd}|jdk rB|jjnd} |} | |kr|dk rf||| rfq| dk rz| || } Wqntk rYqYqnXqnz||| dddWqtt fk r| d7} YqJXqqJt|||j || }|||} t | } |j r||||dd\}} | | 7} || fS)NF)rrrr) rrSrrrrrrr!rrr$r)rrRrrrUrrSZ expr_parseZself_failOn_canParseNextZself_ignoreExpr_tryParseZtmplocZskiptextZ skipresultrMryryrzrs:   zSkipTo.parseImpl)FNN)Tr5ryryrrzr*s6 csbeZdZdZdfdd ZddZddZd d Zd d Zgfd dZ ddZ fddZ Z S)raK Forward declaration of an expression to be defined later - used for recursive grammars, such as algebraic infix notation. When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator. Note: take care when assigning to C{Forward} not to overlook precedence of operators. Specifically, '|' has a lower precedence than '<<', so that:: fwdExpr << a | b | c will actually be evaluated as:: (fwdExpr << a) | b | c thereby leaving b and c out as parseable alternatives. It is recommended that you explicitly group the values inserted into the C{Forward}:: fwdExpr << (a | b | c) Converting to use the '<<=' operator instead will avoid this problem. See L{ParseResults.pprint} for an example of a recursive parser created using C{Forward}. Ncstt|j|dddSr2)rrrr"rryrzr@szForward.__init__cCsjt|trt|}||_d|_|jj|_|jj|_||jj |jj |_ |jj |_ |j |jj |Sr)r}rr&rwrSr{rrr rr~r}rrr"ryryrz __lshift__Cs      zForward.__lshift__cCs||>Srryr"ryryrz __ilshift__PszForward.__ilshift__cCs d|_|Srr rryryrzr SszForward.leaveWhitespacecCs$|js d|_|jdk r |j|Sr)rrSrrryryrzrWs   zForward.streamlinecCs>||kr0|dd|g}|jdk r0|j||gdSrrrryryrzr^s   zForward.validatecCsVt|dr|jS|jjdSz|jdk r4t|j}nd}W5|j|_X|jjd|S)Nrz: ...Nonez: )rrrmrZ _revertClass_ForwardNoRecurserSr)rZ retStringryryrzres     zForward.__str__cs.|jdk rtt|St}||K}|SdSr)rSrrrr1rryrzrvs  z Forward.copy)N) rrrrrrrr rrrrr1ryryrrzr-s  c@seZdZddZdS)rcCsdS)Nrcryrryryrzrsz_ForwardNoRecurse.__str__N)rrrrryryryrzr~srcs"eZdZdZdfdd ZZS)r/zQ Abstract subclass of C{ParseExpression}, for converting parsed results. Fcstt||d|_dSr)rr/rr}rrryrzrszTokenConverter.__init__)Fr4ryryrrzr/scs6eZdZdZd fdd ZfddZdd ZZS) r a Converter to concatenate all matching tokens to a single string. By default, the matching patterns must also be contiguous in the input string; this can be disabled by specifying C{'adjacent=False'} in the constructor. Example:: real = Word(nums) + '.' + Word(nums) print(real.parseString('3.1416')) # -> ['3', '.', '1416'] # will also erroneously match the following print(real.parseString('3. 1416')) # -> ['3', '.', '1416'] real = Combine(Word(nums) + '.' + Word(nums)) print(real.parseString('3.1416')) # -> ['3.1416'] # no match when there are internal spaces print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...) rTcs8tt|||r|||_d|_||_d|_dSr)rr rr adjacentr~ joinStringr)rrSrrrryrzrszCombine.__init__cs(|jrt||ntt|||Sr)rr&rrr r"rryrzrszCombine.ignorecCsP|}|dd=|td||jg|jd7}|jrH|rH|gS|SdS)Nr)r)rr$rr(rrr|r )rrRrrZretToksryryrzrs  "zCombine.postParse)rT)rrrrrrrr1ryryrrzr s cs(eZdZdZfddZddZZS)ra Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions. Example:: ident = Word(alphas) num = Word(nums) term = ident | num func = ident + Optional(delimitedList(term)) print(func.parseString("fn a,b,100")) # -> ['fn', 'a', 'b', '100'] func = ident + Group(Optional(delimitedList(term))) print(func.parseString("fn a,b,100")) # -> ['fn', ['a', 'b', '100']] cstt||d|_dSr)rrrr}rrryrzrszGroup.__init__cCs|gSrryrryryrzrszGroup.postParserrrrrrr1ryryrrzrs cs(eZdZdZfddZddZZS)r aW Converter to return a repetitive expression as a list, but also as a dictionary. Each element can also be referenced using the first token in the expression as its key. Useful for tabular report scraping when the first column can be used as a item key. Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) # print attributes as plain groups print(OneOrMore(attr_expr).parseString(text).dump()) # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names result = Dict(OneOrMore(Group(attr_expr))).parseString(text) print(result.dump()) # access named fields as dict entries, or output as dict print(result['shape']) print(result.asDict()) prints:: ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap'] [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'} See more examples at L{ParseResults} of accessing fields by results name. cstt||d|_dSr)rr rr}rrryrzrsz Dict.__init__cCst|D]\}}t|dkrq|d}t|tr@t|d}t|dkr\td|||<qt|dkrt|dtst|d|||<q|}|d=t|dkst|tr| rt||||<qt|d|||<q|j r|gS|SdS)Nrrrrs) rrr}rvrrrr$rr r|)rrRrrrtokZikeyZ dictvalueryryrzrs$   zDict.postParserryryrrzr s# c@s eZdZdZddZddZdS)r-aV Converter for ignoring the results of a parsed expression. Example:: source = "a, b, c,d" wd = Word(alphas) wd_list1 = wd + ZeroOrMore(',' + wd) print(wd_list1.parseString(source)) # often, delimiters that are useful during parsing are just in the # way afterward - use Suppress to keep them out of the parsed output wd_list2 = wd + ZeroOrMore(Suppress(',') + wd) print(wd_list2.parseString(source)) prints:: ['a', ',', 'b', ',', 'c', ',', 'd'] ['a', 'b', 'c', 'd'] (See also L{delimitedList}.) cCsgSrryrryryrzrszSuppress.postParsecCs|Srryrryryrzr "szSuppress.suppressN)rrrrrr ryryryrzr- sc@s(eZdZdZddZddZddZdS) rzI Wrapper for parse actions, to ensure they are only called once. cCst||_d|_dSr)rrcallablecalled)rZ methodCallryryrzr*s zOnlyOnce.__init__cCs.|js||||}d|_|St||ddS)NTr)rrr!)rrr[rxrNryryrzr -s zOnlyOnce.__call__cCs d|_dSr)rrryryrzreset3szOnlyOnce.resetN)rrrrrr rryryryrzr&scs:tfdd}z j|_Wntk r4YnX|S)at Decorator for debugging parse actions. When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".} When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised. Example:: wd = Word(alphas) @traceParseAction def remove_duplicate_chars(tokens): return ''.join(sorted(set(''.join(tokens)))) wds = OneOrMore(wd).setParseAction(remove_duplicate_chars) print(wds.parseString("slkdjs sld sldd sdlf sdljf")) prints:: >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {})) <>entering %s(line: '%s', %d, %r) z<.z)rrrr)rrryrrzrd6s  ,FcCs`t|dt|dt|d}|rBt|t|||S|tt|||SdS)a Helper to define a delimited list of expressions - the delimiter defaults to ','. By default, the list elements and delimiters can have intervening whitespace, and comments, but this can be overridden by passing C{combine=True} in the constructor. If C{combine} is set to C{True}, the matching tokens are returned as a single token string, with the delimiters included; otherwise, the matching tokens are returned as a list of tokens, with the delimiters suppressed. Example:: delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc'] delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE'] z [r(rN)rr r4rr-)rSZdelimcombineZdlNameryryrzrBbs $csjtfdd}|dkr0ttdd}n|}|d|j|dd|d td S) a: Helper to define a counted list of expressions. This helper defines a pattern of the form:: integer expr expr expr... where the leading integer tells how many expr expressions follow. The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed. If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value. Example:: countedArray(Word(alphas)).parseString('2 ab cd ef') # -> ['ab', 'cd'] # in this parser, the leading integer value is given in binary, # '10' indicating that 2 values are in the array binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2)) countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef') # -> ['ab', 'cd'] cs.|d}|r ttg|p&tt>gSr)rrrE)rr[rxrZ arrayExprrSryrzcountFieldParseActions"z+countedArray..countFieldParseActionNcSs t|dSr)rvrwryryrzr{r|zcountedArray..ZarrayLenTrz(len) rc)rr1rTrrrrr)rSZintExprrryrrzr>us cCs6g}|D](}t|tr&|t|q||q|Sr)r}rrrr)Lrrryryrzrs   rcs6tfdd}|j|dddt|S)a* Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousLiteral(first) matchExpr = first + ":" + second will match C{"1:1"}, but not C{"1:2"}. Because this matches a previous literal, will also match the leading C{"1:1"} in C{"1:10"}. If this is not desired, use C{matchPreviousExpr}. Do I{not} use with packrat parsing enabled. csP|rBt|dkr|d>qLt|}tdd|D>n t>dS)Nrrcss|]}t|VqdSr)rrZttryryrzrszDmatchPreviousLiteral..copyTokenToRepeater..)rrrrr)rr[rxZtflatZrepryrzcopyTokenToRepeaters   z1matchPreviousLiteral..copyTokenToRepeaterTr(prev) )rrrr)rSrryrrzrQs  csFt|}|Kfdd}|j|dddt|S)aS Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousExpr(first) matchExpr = first + ":" + second will match C{"1:1"}, but not C{"1:2"}. Because this matches by expressions, will I{not} match the leading C{"1:1"} in C{"1:10"}; the expressions are evaluated first, and then compared, so C{"1"} is compared with C{"10"}. Do I{not} use with packrat parsing enabled. cs*t|fdd}j|dddS)Ncs$t|}|kr tddddS)Nrr)rrr!)rr[rxZ theseTokensZ matchTokensryrzmustMatchTheseTokenss zLmatchPreviousExpr..copyTokenToRepeater..mustMatchTheseTokensTr)rrr)rr[rxrrrrzrs  z.matchPreviousExpr..copyTokenToRepeaterTrr)rrrrr)rSZe2rryrrzrPs cCs:dD]}||t|}q|dd}|dd}t|S)Nz\^-]r2r'r~r)r_bslashr)rrryryrzrYs   rYTc s|rdd}dd}tndd}dd}tg}t|trF|}n$t|trZt|}ntjdt dd|stt Sd }|t |d kr||}t ||d d D]R\}} || |r|||d =qxq||| r|||d =| || | }qxq|d 7}qx|s|rzlt |t d |krTtd d dd|Dd|WStddd|Dd|WSWn&tk rtjdt ddYnXtfdd|Dd|S)a Helper to quickly define a set of alternative Literals, and makes sure to do longest-first testing when there is a conflict, regardless of the input order, but returns a C{L{MatchFirst}} for best performance. Parameters: - strs - a string of space-delimited literals, or a collection of string literals - caseless - (default=C{False}) - treat all literals as caseless - useRegex - (default=C{True}) - as an optimization, will generate a Regex object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or if creating a C{Regex} raises an exception) Example:: comp_oper = oneOf("< = > <= >= !=") var = Word(alphas) number = Word(nums) term = var | number comparison_expr = term + comp_oper + term print(comparison_expr.searchString("B = 12 AA=23 B<=AA AA>12")) prints:: [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']] cSs||kSr)r@rbryryrzr{r|zoneOf..cSs||Sr)r@r<rryryrzr{r|cSs||kSrryrryryrzr{r|cSs ||Sr)r<rryryrzr{r|z6Invalid argument to oneOf, expected string or iterablersrrrNrz[%s]css|]}t|VqdSr)rYrZsymryryrzrszoneOf..r|css|]}t|VqdSr)rr[rryryrzrsz7Exception creating Regex for oneOf, building MatchFirstc3s|]}|VqdSrryrZparseElementClassryrzr$s)r rr}rrrrrrrrrrrrr)rrpr) Zstrsr?ZuseRegexZisequalZmasksZsymbolsrZcurrrryrrzrUsT         ** cCsttt||S)a Helper to easily and clearly define a dictionary by specifying the respective patterns for the key and value. Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens in the proper order. The key pattern can include delimiting markers or punctuation, as long as they are suppressed, thereby leaving the significant key text. The value pattern can include named results, so that the C{Dict} results can include named token fields. Example:: text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) print(OneOrMore(attr_expr).parseString(text).dump()) attr_label = label attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join) # similar to Dict, but simpler call format result = dictOf(attr_label, attr_value).parseString(text) print(result.dump()) print(result['shape']) print(result.shape) # object attribute access works too print(result.asDict()) prints:: [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE SQUARE {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'} )r r4r)rrryryrzrC&s!cCs^tdd}|}d|_|d||d}|r@dd}ndd}|||j|_|S) a Helper to return the original, untokenized text for a given expression. Useful to restore the parsed fields of an HTML start tag into the raw tag text itself, or to revert separate tokens with intervening whitespace back to the original matching input text. By default, returns astring containing the original parsed text. If the optional C{asString} argument is passed as C{False}, then the return value is a C{L{ParseResults}} containing any results names that were originally matched, and a single token containing the original matched text from the input string. So if the expression passed to C{L{originalTextFor}} contains expressions with defined results names, you must set C{asString} to C{False} if you want to preserve those results name values. Example:: src = "this is test bold text normal text " for tag in ("b","i"): opener,closer = makeHTMLTags(tag) patt = originalTextFor(opener + SkipTo(closer) + closer) print(patt.searchString(src)[0]) prints:: [' bold text '] ['text'] cSs|Srry)rrrxryryrzr{ar|z!originalTextFor..F_original_start _original_endcSs||j|jSr)rrrZryryrzr{fr|cSs&||d|dg|dd<dS)Nrr)rrZryryrz extractTexthsz$originalTextFor..extractText)rrrrr)rSZasStringZ locMarkerZ endlocMarker matchExprrryryrzriIs  cCst|ddS)zp Helper to undo pyparsing's default grouping of And expressions, even if all but one are non-empty. cSs|dSrryrwryryrzr{sr|zungroup..)r/r)rSryryrzrjnscCs4tdd}t|d|d|dS)a Helper to decorate a returned token with its starting and ending locations in the input string. This helper adds the following results names: - locn_start = location where matched expression begins - locn_end = location where matched expression ends - value = the actual parsed results Be careful if the input text contains C{} characters, you may want to call C{L{ParserElement.parseWithTabs}} Example:: wd = Word(alphas) for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"): print(match) prints:: [[0, 'ljsdf', 5]] [[8, 'lksdjjf', 15]] [[18, 'lkkjj', 23]] cSs|SrryrZryryrzr{r|zlocatedExpr..Z locn_startrZlocn_end)rrrrr )rSZlocatorryryrzrlusz\[]-*.$+^?()~ r_cCs |ddSrryrZryryrzr{r|r{z\\0?[xX][0-9a-fA-F]+cCstt|dddS)Nrz\0x)unichrrvlstriprZryryrzr{r|z \\0[0-7]+cCstt|ddddS)Nrr)rrvrZryryrzr{r|z\]rr$r)negatebodyr'csFddz"dfddt|jDWStk r@YdSXdS)a Helper to easily define string ranges for use in Word construction. Borrows syntax from regexp '[]' string range definitions:: srange("[0-9]") -> "0123456789" srange("[a-z]") -> "abcdefghijklmnopqrstuvwxyz" srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_" The input string must be enclosed in []'s, and the returned string is the expanded character set joined into a single string. The values enclosed in the []'s may be: - a single character - an escaped character with a leading backslash (such as C{\-} or C{\]}) - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) (C{\0x##} is also supported for backwards compatibility) - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character) - a range of any of the above, separated by a dash (C{'a-z'}, etc.) - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.) cSs<t|ts|Sdddtt|dt|ddDS)Nrcss|]}t|VqdSr)rrryryrzrsz+srange....rr)r}r$rrord)pryryrzr{r|zsrange..rc3s|]}|VqdSrry)rpartZ _expandedryrzrszsrange..N)r_reBracketExprrrrprdryrrzras "csfdd}|S)zt Helper method for defining parse actions that require matching at a specific column in the input text. cs"t||krt||ddS)Nzmatched token not at column %dr)rNZlocnrVrryrz verifyColsz!matchOnlyAtCol..verifyColry)rrryrrzrOs cs fddS)a Helper method for common parse actions that simply return a literal value. Especially useful when used with C{L{transformString}()}. Example:: num = Word(nums).setParseAction(lambda toks: int(toks[0])) na = oneOf("N/A NA").setParseAction(replaceWith(math.nan)) term = na | num OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234] csgSrryrZZreplStrryrzr{r|zreplaceWith..ryrryrrzr^s cCs|dddS)a Helper parse action for removing quotation marks from parsed quoted strings. Example:: # by default, quotation marks are included in parsed results quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"] # use removeQuotes to strip quotation marks from parsed results quotedString.setParseAction(removeQuotes) quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"] rrrtryrZryryrzr\s csNfdd}ztdtdj}Wntk rBt}YnX||_|S)aG Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional args are passed, they are forwarded to the given function as additional arguments after the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the parsed data to an integer using base 16. Example (compare the last to example in L{ParserElement.transformString}:: hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16)) hex_ints.runTests(''' 00 11 22 aa FF 0a 0d 1a ''') upperword = Word(alphas).setParseAction(tokenMap(str.upper)) OneOrMore(upperword).runTests(''' my kingdom for a horse ''') wd = Word(alphas).setParseAction(tokenMap(str.title)) OneOrMore(wd).setParseAction(' '.join).runTests(''' now is the winter of our discontent made glorious summer by this sun of york ''') prints:: 00 11 22 aa FF 0a 0d 1a [0, 17, 34, 170, 255, 10, 13, 26] my kingdom for a horse ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE'] now is the winter of our discontent made glorious summer by this sun of york ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York'] csfdd|DS)Ncsg|]}|fqSryry)rZtoknrr\ryrzrsz(tokenMap..pa..ryrZrryrzrsztokenMap..parrm)rorrpr~)r\rrrqryrrzros  cCs t|Srrr@rwryryrzr{r|cCs t|Srrlowerrwryryrzr{r|c Cst|tr|}t|| d}n|j}tttd}|rt t }t d|dt t t|t d|tddgdd  d d t d }nd ddtD}t t t|B}t d|dt t t| ttt d|tddgdd  dd t d }ttd|d }|dd |ddd|}|dd |ddd|}||_||_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namerEz_-:r5tag=/Fr rEcSs |ddkSNrrryrZryryrzr{r|z_makeTags..r6rcss|]}|dkr|VqdS)r6Nryrryryrzrsz_makeTags..cSs |ddkSrryrZryryrzr{r|r7rJ:r(z<%s>r`z)r}rrrr1r6r5r@rrr\r-r r4rrrrrXr[rDr _Lrtitlerrr)tagStrZxmlZresnameZ tagAttrNameZ tagAttrValueZopenTagZprintablesLessRAbrackZcloseTagryryrz _makeTags s> ..rcCs t|dS)a  Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values. Example:: text = 'More info at the pyparsing wiki page' # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple a,a_end = makeHTMLTags("A") link_expr = a + SkipTo(a_end)("link_text") + a_end for link in link_expr.searchString(text): # attributes in the tag (like "href" shown here) are also accessible as named results print(link.link_text, '->', link.href) prints:: pyparsing -> http://pyparsing.wikispaces.com FrrryryrzrM(scCs t|dS)z Helper to construct opening and closing tag expressions for XML, given a tag name. Matches tags only in the given upper/lower case. Example: similar to L{makeHTMLTags} TrrryryrzrN;scs8|r|ddn|ddDfdd}|S)a< Helper to create a validating parse action to be used with start tags created with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag with a required attribute value, to avoid false matches on common tags such as C{} or C{
}. Call C{withAttribute} with a series of attribute names and values. Specify the list of filter attributes names and values as: - keyword arguments, as in C{(align="right")}, or - as an explicit dict with C{**} operator, when an attribute name is also a Python reserved word, as in C{**{"class":"Customer", "align":"right"}} - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") ) For attribute names with a namespace prefix, you must use the second form. Attribute names are matched insensitive to upper/lower case. If just testing for C{class} (with or without a namespace), use C{L{withClass}}. To verify that the attribute exists, but without specifying a value, pass C{withAttribute.ANY_VALUE} as the value. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this has no type
''' div,div_end = makeHTMLTags("div") # only match div tag having a type attribute with value "grid" div_grid = div().setParseAction(withAttribute(type="grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) # construct a match with any div tag having a type attribute, regardless of the value div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 NcSsg|]\}}||fqSryryr/ryryrzrzsz!withAttribute..csZD]P\}}||kr$t||d||tjkr|||krt||d||||fqdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r!rg ANY_VALUE)rr[rZattrNameZ attrValueZattrsryrzr{s  zwithAttribute..pa)r)rZattrDictrryrrzrgDs 2 cCs|r d|nd}tf||iS)a Simplified version of C{L{withAttribute}} when matching on a div class - made difficult because C{class} is a reserved word in Python. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this <div> has no class
''' div,div_end = makeHTMLTags("div") div_grid = div().setParseAction(withClass("grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 z%s:classclass)rg)Z classname namespaceZ classattrryryrzrms (rpcCst}||||B}t|D]l\}}|ddd\}} } } | dkrPd|nd|} | dkr|dkstt|dkr|td|\} }t| }| tjkr^| d krt||t|t |}n| dkr|dk rt|||t|t ||}nt||t|t |}nD| dkrTt|| |||t|| |||}ntd n| tj krB| d krt |t st |}t|j |t||}n| dkr|dk rt|||t|t ||}nt||t|t |}nD| dkr8t|| |||t|| |||}ntd ntd | rvt | ttfrl|j| n || ||| |BK}|}q||K}|S) aD Helper method for constructing grammars of expressions made up of operators working in a precedence hierarchy. Operators may be unary or binary, left- or right-associative. Parse actions can also be attached to operator expressions. The generated parser will also recognize the use of parentheses to override operator precedences (see example below). Note: if you define a deep operator list, you may see performance issues when using infixNotation. See L{ParserElement.enablePackrat} for a mechanism to potentially improve your parser performance. Parameters: - baseExpr - expression representing the most basic element for the nested - opList - list of tuples, one for each operator precedence level in the expression grammar; each tuple is of the form (opExpr, numTerms, rightLeftAssoc, parseAction), where: - opExpr is the pyparsing expression for the operator; may also be a string, which will be converted to a Literal; if numTerms is 3, opExpr is a tuple of two expressions, for the two operators separating the 3 terms - numTerms is the number of terms for this operator (must be 1, 2, or 3) - rightLeftAssoc is the indicator whether the operator is right or left associative, using the pyparsing-defined constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}. - parseAction is the parse action to be associated with expressions matching this operator expression (the parse action tuple member may be omitted); if the parse action is passed a tuple or list of functions, this is equivalent to calling C{setParseAction(*fn)} (L{ParserElement.setParseAction}) - lpar - expression for matching left-parentheses (default=C{Suppress('(')}) - rpar - expression for matching right-parentheses (default=C{Suppress(')')}) Example:: # simple example of four-function arithmetic with ints and variable names integer = pyparsing_common.signed_integer varname = pyparsing_common.identifier arith_expr = infixNotation(integer | varname, [ ('-', 1, opAssoc.RIGHT), (oneOf('* /'), 2, opAssoc.LEFT), (oneOf('+ -'), 2, opAssoc.LEFT), ]) arith_expr.runTests(''' 5+3*6 (5+3)*6 -2--11 ''', fullDump=False) prints:: 5+3*6 [[5, '+', [3, '*', 6]]] (5+3)*6 [[[5, '+', 3], '*', 6]] -2--11 [[['-', 2], '-', ['-', 11]]] rNrbrqz%s termz %s%s termrsz@if numterms=3, opExpr must be a tuple or list of two expressionsrz6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)rrrrrrVLEFTrrrRIGHTr}rrSrrr)ZbaseExprZopListZlparZrparrZlastExprrZoperDefZopExprZarityZrightLeftAssocrZtermNameZopExpr1ZopExpr2ZthisExprrryryrzrksZ=   &       &    z4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*'z string enclosed in single quotesz*quotedString using single or double quotesuzunicode string literalcCs||krtd|dkr*t|tr"t|tr"t|dkrt|dkr|dk rtt|t||tjdd dd}n$t t||tj dd}nx|dk rtt|t |t |ttjdd dd}n4ttt |t |ttjdd d d}ntd t }|dk rd|tt|t||B|Bt|K}n$|tt|t||Bt|K}|d ||f|S) a~ Helper method for defining nested lists enclosed in opening and closing delimiters ("(" and ")" are the default). Parameters: - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression - content - expression for items within the nested lists (default=C{None}) - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString}) If an expression is not provided for the content argument, the nested expression will capture all whitespace-delimited content between delimiters as a list of separate values. Use the C{ignoreExpr} argument to define expressions that may contain opening or closing characters that should not be treated as opening or closing characters for nesting, such as quotedString or a comment expression. Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}. The default is L{quotedString}, but if no expressions are to be ignored, then pass C{None} for this argument. Example:: data_type = oneOf("void int short long char float double") decl_data_type = Combine(data_type + Optional(Word('*'))) ident = Word(alphas+'_', alphanums+'_') number = pyparsing_common.number arg = Group(decl_data_type + ident) LPAR,RPAR = map(Suppress, "()") code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment)) c_function = (decl_data_type("type") + ident("name") + LPAR + Optional(delimitedList(arg), [])("args") + RPAR + code_body("body")) c_function.ignore(cStyleComment) source_code = ''' int is_odd(int x) { return (x%2); } int dec_to_hex(char hchar) { if (hchar >= '0' && hchar <= '9') { return (ord(hchar)-ord('0')); } else { return (10+ord(hchar)-ord('A')); } } ''' for func in c_function.searchString(source_code): print("%(name)s (%(type)s) args: %(args)s" % func) prints:: is_odd (int) args: [['int', 'x']] dec_to_hex (int) args: [['char', 'hchar']] z.opening and closing strings cannot be the sameNrrcSs |dSrrrwryryrzr{gr|znestedExpr..cSs |dSrr rwryryrzr{jr|cSs |dSrr rwryryrzr{pr|cSs |dSrr rwryryrzr{tr|zOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)rr}rrr rr r&rsrrErrrrr-r4r)ZopenerZcloserZcontentrrryryrzrR%sH:    *$c sfdd}fdd}fdd}ttd}tt|d}t|d }t|d } |rtt||t|t|t|| } n$tt|t|t|t|} | t t| d S) a Helper method for defining space-delimited indentation blocks, such as those used to define block statements in Python source code. Parameters: - blockStatementExpr - expression defining syntax of statement that is repeated within the indented block - indentStack - list created by caller to manage indentation stack (multiple statementWithIndentedBlock expressions within a single grammar should share a common indentStack) - indent - boolean indicating whether block must be indented beyond the the current level; set to False for block of left-most statements (default=C{True}) A valid block must contain at least one C{blockStatement}. Example:: data = ''' def A(z): A1 B = 100 G = A2 A2 A3 B def BB(a,b,c): BB1 def BBA(): bba1 bba2 bba3 C D def spam(x,y): def eggs(z): pass ''' indentStack = [1] stmt = Forward() identifier = Word(alphas, alphanums) funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":") func_body = indentedBlock(stmt, indentStack) funcDef = Group( funcDecl + func_body ) rvalue = Forward() funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")") rvalue << (funcCall | identifier | Word(nums)) assignment = Group(identifier + "=" + rvalue) stmt << ( funcDef | assignment | identifier ) module_body = OneOrMore(stmt) parseTree = module_body.parseString(data) parseTree.pprint() prints:: [['def', 'A', ['(', 'z', ')'], ':', [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]], 'B', ['def', 'BB', ['(', 'a', 'b', 'c', ')'], ':', [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]], 'C', 'D', ['def', 'spam', ['(', 'x', 'y', ')'], ':', [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] csN|t|krdSt||}|dkrJ|dkr>t||dt||ddS)Nrtzillegal nestingznot a peer entry)rr;r#r!rr[rxZcurCol indentStackryrzcheckPeerIndents     z&indentedBlock..checkPeerIndentcs2t||}|dkr"|n t||ddS)Nrtznot a subentry)r;rr!rrryrzcheckSubIndents   z%indentedBlock..checkSubIndentcsN|t|krdSt||}r6|dkr6|dksBt||ddS)Nrtr_znot an unindent)rr;r!rrrryrz checkUnindents    z$indentedBlock..checkUnindentz INDENTrZUNINDENTzindented block) rrr r rrrrrrr) ZblockStatementExprrr9rrrrErZPEERZUNDENTZsmExprryrrzrhs(N   z#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z &(?Prz);zcommon HTML entitycCs t|jS)zRHelper parser action to replace common HTML entities with their special characters)_htmlEntityMaprZentityrwryryrzr]sz/\*(?:[^*]|\*(?!/))*z*/zC style commentzz HTML commentz.*z rest of linez//(?:\\\n|[^\n])*z // commentzC++ style commentz#.*zPython style commentrO commaItemrc@seZdZdZeeZeeZe e  d eZ e e d eedZed d eZe ede e dZed d eeeed eB d Zeeed  d eZed d eZeeBeBZed d eZe eded dZed dZ ed dZ!e!de!d dZ"ee!de!ddee!de!d dZ#e#$dd d e  d!Z%e&e"e%Be#B d" d"Z'ed# d$Z(e)d=d&d'Z*e)d>d)d*Z+ed+ d,Z,ed- d.Z-ed/ d0Z.e/e0BZ1e)d1d2Z2e&e3e4d3e5e e6d3d4ee7d5 d6Z8e9ee:;e8Bd7d8 d9Zd}, L{reals}, L{scientific notation}) - common L{programming identifiers} - network addresses (L{MAC}, L{IPv4}, L{IPv6}) - ISO8601 L{dates} and L{datetime} - L{UUID} - L{comma-separated list} Parse actions: - C{L{convertToInteger}} - C{L{convertToFloat}} - C{L{convertToDate}} - C{L{convertToDatetime}} - C{L{stripHTMLTags}} - C{L{upcaseTokens}} - C{L{downcaseTokens}} Example:: pyparsing_common.number.runTests(''' # any int or real number, returned as the appropriate type 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.fnumber.runTests(''' # any int or real number, returned as float 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.hex_integer.runTests(''' # hex numbers 100 FF ''') pyparsing_common.fraction.runTests(''' # fractions 1/2 -3/4 ''') pyparsing_common.mixed_integer.runTests(''' # mixed fractions 1 1/2 -3/4 1-3/4 ''') import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(''' # uuid 12345678-1234-5678-1234-567812345678 ''') prints:: # any int or real number, returned as the appropriate type 100 [100] -100 [-100] +100 [100] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # any int or real number, returned as float 100 [100.0] -100 [-100.0] +100 [100.0] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # hex numbers 100 [256] FF [255] # fractions 1/2 [0.5] -3/4 [-0.75] # mixed fractions 1 [1] 1/2 [0.5] -3/4 [-0.75] 1-3/4 [1.75] # uuid 12345678-1234-5678-1234-567812345678 [UUID('12345678-1234-5678-1234-567812345678')] integerz hex integerrz[+-]?\d+zsigned integerrfractioncCs|d|dS)Nrrtryrwryryrzr{r|zpyparsing_common.rz"fraction or mixed integer-fractionz [+-]?\d+\.\d*z real numberz+[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?fnumberr identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}z IPv4 addressz[0-9a-fA-F]{1,4} hex_integerrzfull IPv6 address)rrhz::zshort IPv6 addresscCstdd|DdkS)Ncss|]}tj|rdVqdSr3)rp _ipv6_partrrryryrzrs z,pyparsing_common...r)rrwryryrzr{r|z::ffff:zmixed IPv6 addressz IPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}z MAC address%Y-%m-%dcsfdd}|S)a Helper to create a parse action for converting parsed date string to Python datetime.date Params - - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"}) Example:: date_expr = pyparsing_common.iso8601_date.copy() date_expr.setParseAction(pyparsing_common.convertToDate()) print(date_expr.parseString("1999-12-31")) prints:: [datetime.date(1999, 12, 31)] c sNzt|dWStk rH}zt||t|W5d}~XYnXdSr)rstrptimedaterr!r~rr[rxZvefmtryrzcvt_fnsz.pyparsing_common.convertToDate..cvt_fnryr$r%ryr#rz convertToDates zpyparsing_common.convertToDate%Y-%m-%dT%H:%M:%S.%fcsfdd}|S)a Helper to create a parse action for converting parsed datetime string to Python datetime.datetime Params - - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"}) Example:: dt_expr = pyparsing_common.iso8601_datetime.copy() dt_expr.setParseAction(pyparsing_common.convertToDatetime()) print(dt_expr.parseString("1999-12-31T23:59:59.999")) prints:: [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)] c sJzt|dWStk rD}zt||t|W5d}~XYnXdSr)rr rr!r~r"r#ryrzr%sz2pyparsing_common.convertToDatetime..cvt_fnryr&ryr#rzconvertToDatetimes z"pyparsing_common.convertToDatetimez7(?P\d{4})(?:-(?P\d\d)(?:-(?P\d\d))?)?z ISO8601 datez(?P\d{4})-(?P\d\d)-(?P\d\d)[T ](?P\d\d):(?P\d\d)(:(?P\d\d(\.\d*)?)?)?(?PZ|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}UUIDcCstj|dS)a Parse action to remove HTML tags from web page HTML source Example:: # strip HTML links from normal text text = 'More info at the
pyparsing wiki page' td,td_end = makeHTMLTags("TD") table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page' r)rp_html_stripperr)rr[rryryrz stripHTMLTagss zpyparsing_common.stripHTMLTagsrrOrrrrzcomma separated listcCs t|Srrrwryryrzr{"r|cCs t|Srrrwryryrzr{%r|N)r)r()?rrrrrorvZconvertToIntegerfloatZconvertToFloatr1rTrrrrFrr)Zsigned_integerrrrr Z mixed_integerrrealZsci_realrnumberrr6r5rZ ipv4_addressrZ_full_ipv6_addressZ_short_ipv6_addressrZ_mixed_ipv6_addressr Z ipv6_addressZ mac_addressr/r'r)Z iso8601_dateZiso8601_datetimeuuidr9r8r+r,rrrrXr0 _commasepitemrBr[rZcomma_separated_listrfrDryryryrzrpsV"" 2     __main__Zselectfromr=r)rcolumnsrZtablesZcommandaK # '*' as column list and dotted table name select * from SYS.XYZZY # caseless match on "SELECT", and casts back to "select" SELECT * from XYZZY, ABC # list of column names, and mixed case SELECT keyword Select AA,BB,CC from Sys.dual # multiple tables Select A, B, C from Sys.dual, Table2 # invalid SELECT keyword - should fail Xelect A, B, C from Sys.dual # incomplete command - should fail Select # invalid column name - should fail Select ^^^ frox Sys.dual z] 100 -100 +100 3.14159 6.02e23 1e-12 z 100 FF z6 12345678-1234-5678-1234-567812345678 )rs)rF)N)FT)T)r)T)r __version__Z__versionTime__ __author__rweakrefrrrrrrrjrrFrcrr_threadr ImportErrorZ threadingZcollections.abcrrrrZ ordereddict__all__r version_inforbr0maxsizer0r~rchrrrrrr@reversedrrrBrr]r^rnZmaxintZxrangerZ __builtin__rZfnamerrorrrrrrZascii_uppercaseZascii_lowercaser6rTrFr5rrZ printablerXrprr!r#r%r(rr$registerr;rLrIrTrWrYrSrrr&r.rrrrrwrr r rnr1r)r'r r0rrrrr,r+r3r2r"rrrrr rrrrr4rrrr*rrr/r rr r-rrdrBr>rrQrPrYrUrCrirjrlrrErKrJrcrbrZ _escapedPuncZ_escapedHexCharZ_escapedOctCharZ _singleCharZ _charRangerrrarOr^r\rorfrDrrMrNrgrrmrVrr rkrWr@r`r[rerRrhr7rYr9r8rrrrr=r]r:rGr r_rAr?rHrZrr1r<rprZ selectTokenZ fromTokenZidentZ columnNameZcolumnNameListZ columnSpecZ tableNameZ tableNameListZ simpleSQLr.r/rrr0r*ryryryrzs4        8      @v &A= I G3pLOD|M &#@sQ,A,    I# %     0 ,   ? #p  Zr   (         "   site-packages/setuptools/_vendor/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000232151464452710022505 0ustar00U Qab@sdS)Nrrr?/usr/lib/python3.8/site-packages/setuptools/_vendor/__init__.pysite-packages/setuptools/_vendor/__pycache__/ordered_set.cpython-38.pyc000064400000040034151464452710022312 0ustar00U Qab;@s|dZddlZddlmZzddlmZmZWn$ek rPddlmZmZYnXe dZ dZ ddZ Gdd d eeZ dS) z An OrderedSet is a custom MutableSet that remembers its order, so that every entry has an index that can be looked up. Based on a recipe originally posted to ActiveState Recipes by Raymond Hettiger, and released under the MIT license. N)deque) MutableSetSequencez3.1cCs"t|do t|t o t|t S)a  Are we being asked to look up a list of things, instead of a single thing? We check for the `__iter__` attribute so that this can cover types that don't have to be known by this module, such as NumPy arrays. Strings, however, should be considered as atomic values to look up, not iterables. The same goes for tuples, since they are immutable and therefore valid entries. We don't need to check for the Python 2 `unicode` type, because it doesn't have an `__iter__` attribute anyway. __iter__)hasattr isinstancestrtuple)objr B/usr/lib/python3.8/site-packages/setuptools/_vendor/ordered_set.py is_iterables    r c@seZdZdZd;ddZddZddZd d Zd d Zd dZ ddZ ddZ e Z ddZ ddZeZeZddZddZddZddZdd Zd!d"Zd#d$Zd%d&Zd'd(Zd)d*Zd+d,Zd-d.Zd/d0Zd1d2Zd3d4Zd5d6Z d7d8Z!d9d:Z"dS)< OrderedSetz An OrderedSet is a custom MutableSet that remembers its order, so that every entry has an index that can be looked up. Example: >>> OrderedSet([1, 1, 2, 3, 2]) OrderedSet([1, 2, 3]) NcCs g|_i|_|dk r||O}dSN)itemsmap)selfiterabler r r __init__4szOrderedSet.__init__cCs t|jS)z Returns the number of unique elements in the ordered set Example: >>> len(OrderedSet([])) 0 >>> len(OrderedSet([1, 2])) 2 )lenrrr r r __len__:s zOrderedSet.__len__cs|t|tr|tkrSt|r4fdd|DSt|dsHt|trlj|}t|trf|S|Sn t d|dS)aQ Get the item at a given index. If `index` is a slice, you will get back that slice of items, as a new OrderedSet. If `index` is a list or a similar iterable, you'll get a list of items corresponding to those indices. This is similar to NumPy's "fancy indexing". The result is not an OrderedSet because you may ask for duplicate indices, and the number of elements returned should be the number of elements asked for. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset[1] 2 csg|]}j|qSr )r).0irr r [sz*OrderedSet.__getitem__.. __index__z+Don't know how to index an OrderedSet by %rN) rslice SLICE_ALLcopyr rrlist __class__ TypeError)rindexresultr rr __getitem__Fs   zOrderedSet.__getitem__cCs ||S)z Return a shallow copy of this object. Example: >>> this = OrderedSet([1, 2, 3]) >>> other = this.copy() >>> this == other True >>> this is other False )r rr r r res zOrderedSet.copycCst|dkrdSt|SdS)Nrr)rrrr r r __getstate__ss zOrderedSet.__getstate__cCs"|dkr|gn ||dS)Nr)r)rstater r r __setstate__s zOrderedSet.__setstate__cCs ||jkS)z Test if the item is in this ordered set Example: >>> 1 in OrderedSet([1, 3, 2]) True >>> 5 in OrderedSet([1, 3, 2]) False )rrkeyr r r __contains__s zOrderedSet.__contains__cCs0||jkr&t|j|j|<|j||j|S)aE Add `key` as an item to this OrderedSet, then return its index. If `key` is already in the OrderedSet, return the index it already had. Example: >>> oset = OrderedSet() >>> oset.append(3) 0 >>> print(oset) OrderedSet([3]) )rrrappendr(r r r adds  zOrderedSet.addcCsFd}z|D]}||}q Wn$tk r@tdt|YnX|S)a< Update the set with the given iterable sequence, then return the index of the last element inserted. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.update([3, 1, 5, 1, 4]) 4 >>> print(oset) OrderedSet([1, 2, 3, 5, 4]) Nz(Argument needs to be an iterable, got %s)r,r! ValueErrortype)rZsequenceZ item_indexitemr r r updates  zOrderedSet.updatecs$t|rfdd|DSj|S)aH Get the index of a given entry, raising an IndexError if it's not present. `key` can be an iterable of entries that is not a string, in which case this returns a list of indices. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.index(2) 1 csg|]}|qSr )r")rZsubkeyrr r rsz$OrderedSet.index..)r rr(r rr r"s zOrderedSet.indexcCs,|jstd|jd}|jd=|j|=|S)z Remove and return the last element from the set. Raises KeyError if the set is empty. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.pop() 3 z Set is empty)rKeyErrorr)relemr r r pops  zOrderedSet.popcCsP||krL|j|}|j|=|j|=|jD]\}}||kr,|d|j|<q,dS)a Remove an element. Do not raise an exception if absent. The MutableSet mixin uses this to implement the .remove() method, which *does* raise an error when asked to remove a non-existent item. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.discard(2) >>> print(oset) OrderedSet([1, 3]) >>> oset.discard(2) >>> print(oset) OrderedSet([1, 3]) N)rr)rr)rkvr r r discards zOrderedSet.discardcCs|jdd=|jdS)z8 Remove all items from this OrderedSet. N)rrclearrr r r r9s zOrderedSet.clearcCs t|jS)zb Example: >>> list(iter(OrderedSet([1, 2, 3]))) [1, 2, 3] )iterrrr r r rszOrderedSet.__iter__cCs t|jS)zf Example: >>> list(reversed(OrderedSet([1, 2, 3]))) [3, 2, 1] )reversedrrr r r __reversed__ szOrderedSet.__reversed__cCs&|sd|jjfSd|jjt|fS)Nz%s()z%s(%r))r __name__rrr r r __repr__szOrderedSet.__repr__cCsRt|ttfrt|t|kSz t|}Wntk r@YdSXt||kSdS)a Returns true if the containers have the same items. If `other` is a Sequence, then order is checked, otherwise it is ignored. Example: >>> oset = OrderedSet([1, 3, 2]) >>> oset == [1, 3, 2] True >>> oset == [1, 2, 3] False >>> oset == [2, 3] False >>> oset == OrderedSet([3, 2, 1]) False FN)rrrrsetr!)rotherZ other_as_setr r r __eq__s zOrderedSet.__eq__cGs<t|tr|jnt}ttt|g|}tj|}||S)a Combines all unique items. Each items order is defined by its first appearance. Example: >>> oset = OrderedSet.union(OrderedSet([3, 1, 4, 1, 5]), [1, 3], [2, 0]) >>> print(oset) OrderedSet([3, 1, 4, 5, 2, 0]) >>> oset.union([8, 9]) OrderedSet([3, 1, 4, 5, 2, 0, 8, 9]) >>> oset | {10} OrderedSet([3, 1, 4, 5, 2, 0, 10]) )rrr rritchain from_iterable)rsetsclsZ containersrr r r union6s zOrderedSet.unioncCs ||Sr) intersectionrr@r r r __and__IszOrderedSet.__and__csHt|tr|jnt}|r>> oset = OrderedSet.intersection(OrderedSet([0, 1, 2, 3]), [1, 2, 3]) >>> print(oset) OrderedSet([1, 2, 3]) >>> oset.intersection([2, 4, 5], [1, 2, 3, 4]) OrderedSet([2]) >>> oset.intersection() OrderedSet([1, 2, 3]) c3s|]}|kr|VqdSrr rr/commonr r ^sz*OrderedSet.intersection..)rrr r?rHrrrErFrr rLr rHMs zOrderedSet.intersectioncs:|j}|r.tjtt|fdd|D}n|}||S)a Returns all elements that are in this set but not the others. Example: >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2])) OrderedSet([1, 3]) >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]), OrderedSet([3])) OrderedSet([1]) >>> OrderedSet([1, 2, 3]) - OrderedSet([2]) OrderedSet([1, 3]) >>> OrderedSet([1, 2, 3]).difference() OrderedSet([1, 2, 3]) c3s|]}|kr|VqdSrr rKr@r r rNtsz(OrderedSet.difference..)r r?rGrrOr rPr differencecs zOrderedSet.differencecs*t|tkrdStfdd|DS)a7 Report whether another set contains this set. Example: >>> OrderedSet([1, 2, 3]).issubset({1, 2}) False >>> OrderedSet([1, 2, 3]).issubset({1, 2, 3, 4}) True >>> OrderedSet([1, 2, 3]).issubset({1, 4, 3, 5}) False Fc3s|]}|kVqdSrr rKrPr r rNsz&OrderedSet.issubset..rallrIr rPr issubsetys zOrderedSet.issubsetcs*tt|krdStfdd|DS)a= Report whether this set contains another set. Example: >>> OrderedSet([1, 2]).issuperset([1, 2, 3]) False >>> OrderedSet([1, 2, 3, 4]).issuperset({1, 2, 3}) True >>> OrderedSet([1, 4, 3, 5]).issuperset({1, 2, 3}) False Fc3s|]}|kVqdSrr rKrr r rNsz(OrderedSet.issuperset..rRrIr rr issupersets zOrderedSet.issupersetcCs:t|tr|jnt}|||}|||}||S)a Return the symmetric difference of two OrderedSets as a new set. That is, the new set will contain all elements that are in exactly one of the sets. Their order will be preserved, with elements from `self` preceding elements from `other`. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.symmetric_difference(other) OrderedSet([4, 5, 9, 2]) )rrr rQrG)rr@rFZdiff1Zdiff2r r r symmetric_differenceszOrderedSet.symmetric_differencecCs||_ddt|D|_dS)zt Replace the 'items' list of this OrderedSet with a new one, updating self.map accordingly. cSsi|]\}}||qSr r )ridxr/r r r sz,OrderedSet._update_items..N)r enumerater)rrr r r _update_itemsszOrderedSet._update_itemscs:t|D]}t|Oq |fdd|jDdS)a Update this OrderedSet to remove items from one or more other sets. Example: >>> this = OrderedSet([1, 2, 3]) >>> this.difference_update(OrderedSet([2, 4])) >>> print(this) OrderedSet([1, 3]) >>> this = OrderedSet([1, 2, 3, 4, 5]) >>> this.difference_update(OrderedSet([2, 4]), OrderedSet([1, 4, 6])) >>> print(this) OrderedSet([3, 5]) csg|]}|kr|qSr r rKitems_to_remover r rsz0OrderedSet.difference_update..Nr?rZr)rrEr@r r[r difference_updateszOrderedSet.difference_updatecs&t|fdd|jDdS)a^ Update this OrderedSet to keep only items in another set, preserving their order in this set. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.intersection_update(other) >>> print(this) OrderedSet([1, 3, 7]) csg|]}|kr|qSr r rKrPr r rsz2OrderedSet.intersection_update..Nr]rIr rPr intersection_updates zOrderedSet.intersection_updatecs<fdd|D}t|fddjD|dS)a Update this OrderedSet to remove items from another set, then add items from the other set that were not present in this set. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.symmetric_difference_update(other) >>> print(this) OrderedSet([4, 5, 9, 2]) csg|]}|kr|qSr r rKrr r rsz:OrderedSet.symmetric_difference_update..csg|]}|kr|qSr r rKr[r r rsNr])rr@Z items_to_addr )r\rr symmetric_difference_updates z&OrderedSet.symmetric_difference_update)N)#r= __module__ __qualname____doc__rrr$rr%r'r*r,r+r0r"Zget_locZ get_indexerr4r8r9rr<r>rArGrJrHrQrTrUrVrZr^r_r`r r r r r*s@    r)rc itertoolsrB collectionsrZcollections.abcrr ImportErrorrr __version__r rr r r r s site-packages/setuptools/_vendor/__pycache__/six.cpython-38.pyc000064400000057553151464452710020634 0ustar00U QabuA@sRdZddlmZddlZddlZddlZddlZddlZdZdZ ej ddkZ ej ddkZ ej dddkZ e refZefZefZeZeZejZn~efZeefZeejfZeZeZejd red ZnHGd d d eZ ze!e Wne"k red ZYn Xed Z[ ddZ#ddZ$GdddeZ%Gddde%Z&Gdddej'Z(Gddde%Z)GdddeZ*e*e+Z,Gddde(Z-e)dddd e)d!d"d#d$d!e)d%d"d"d&d%e)d'd(d#d)d'e)d*d(d+e)d,d"d#d-d,e)d.d/d/d0d.e)d1d/d/d.d1e)d2d(d#d3d2e)d4d(e rd5nd6d7e)d8d(d9e)d:d;dd>d?e)d@d@d?e)dAdAd?e)d3d(d#d3d2e)dBd"d#dCdBe)dDd"d"dEdDe&d#d(e&dFdGe&dHdIe&dJdKdLe&dMdNdMe&dOdPdQe&dRdSdTe&dUdVdWe&dXdYdZe&d[d\d]e&d^d_d`e&dadbdce&dddedfe&dgdhdie&djdjdke&dldldke&dmdmdke&dndndoe&dpdqe&drdse&dtdue&dvdwdve&dxdye&dzd{d|e&d}d~de&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&de+dde&de+dde&de+de+de&ddde&ddde&dddg>Z.ejdkrRe.e&ddg7Z.e.D]2Z/e0e-e/j1e/e2e/e&rVe,3e/de/j1qV[/e.e-_.e-e+dZ4e,3e4dGddde(Z5e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d=dde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)dddgZ6e6D]Z/e0e5e/j1e/q[/e6e5_.e,3e5e+dddҡGddԄde(Z7e)ddde)ddde)dddgZ8e8D]Z/e0e7e/j1e/q[/e8e7_.e,3e7e+dddۡGdd݄de(Z9e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃e)ddd߃g!Z:e:D]Z/e0e9e/j1e/q[/e:e9_.e,3e9e+dddGddde(Z;e)ddde)ddde)ddde)d ddgZe>D]Z/e0e=e/j1e/q[/e>e=_.e,3e=e+dddGdddej'Z?e,3e?e+ddddZ@ddZAe rHdZBdZCdZDdZEdZFdZGn$d ZBd!ZCd"ZDd#ZEd$ZFd%ZGzeHZIWn"eJk rd&d'ZIYnXeIZHzeKZKWn"eJk rd(d)ZKYnXe rd*d+ZLejMZNd,d-ZOeZPn>d.d+ZLd/d0ZNd1d-ZOGd2d3d3eZPeKZKe#eLd4eQeBZReQeCZSeQeDZTeQeEZUeQeFZVeQeGZWe rԐd5d6ZXd7d8ZYd9d:ZZd;d<Z[e\d=Z]e\d>Z^e\d?Z_nTd@d6ZXdAd8ZYdBd:ZZdCd<Z[e\dDZ]e\dEZ^e\dFZ_e#eXdGe#eYdHe#eZdIe#e[dJe rdKdLZ`dMdNZaebZcddldZdededOjfZg[dehdZiejjZkelZmddlnZnenjoZoenjpZpdPZqej dQdQk rdRZrdSZsn dTZrdUZsnjdVdLZ`dWdNZaecZcebZgdXdYZidZd[ZketejuevZmddloZoeojoZoZpd\ZqdRZrdSZse#e`d]e#ead^d_dPZwd`dTZxdadUZye reze4j{dbZ|d|dcddZ}nd}dedfZ|e|dgej dddhk re|din.ej dddhk re|djn dkdlZ~eze4j{dmdZedk rLdndoZej dddpk rreZdqdoZe#e}drej dddk rejejfdsdtZnejZdudvZdwdxZdydzZgZe+Zed{dk rge_ejrBeejD]4\ZZeej+dkrej1e+kreje=q>q[[eje,dS(~z6Utilities for writing code that runs on Python 2 and 3)absolute_importNz'Benjamin Peterson z1.10.0)rjavaic@seZdZddZdS)XcCsdS)Nlselfrr:/usr/lib/python3.8/site-packages/setuptools/_vendor/six.py__len__>sz X.__len__N)__name__ __module__ __qualname__r rrrr r<srlcCs ||_dS)z Add documentation to a function.N)__doc__)funcdocrrr _add_docKsrcCst|tj|S)z7Import module, returning the module after the last dot.) __import__sysmodulesnamerrr _import_modulePsrc@seZdZddZddZdS) _LazyDescrcCs ||_dSNrr rrrr __init__Xsz_LazyDescr.__init__cCsB|}t||j|zt|j|jWntk r<YnX|Sr)_resolvesetattrrdelattr __class__AttributeError)r objtpresultrrr __get__[sz_LazyDescr.__get__N)r rrrr&rrrr rVsrcs.eZdZdfdd ZddZddZZS) MovedModuleNcs2tt||tr(|dkr |}||_n||_dSr)superr'rPY3mod)r roldnewr!rr ris zMovedModule.__init__cCs t|jSr)rr*r rrr rrszMovedModule._resolvecCs"|}t||}t||||Sr)rgetattrr)r attr_modulevaluerrr __getattr__us  zMovedModule.__getattr__)N)r rrrrr2 __classcell__rrr-r r'gs r'cs(eZdZfddZddZgZZS) _LazyModulecstt|||jj|_dSr)r(r4rr!rrr-rr r~sz_LazyModule.__init__cCs ddg}|dd|jD7}|S)Nrr cSsg|] }|jqSrr).0r/rrr sz'_LazyModule.__dir__..)_moved_attributes)r Zattrsrrr __dir__sz_LazyModule.__dir__)r rrrr8r7r3rrr-r r4|s r4cs&eZdZdfdd ZddZZS)MovedAttributeNcsdtt||trH|dkr |}||_|dkr@|dkr<|}n|}||_n||_|dkrZ|}||_dSr)r(r9rr)r*r/)r rZold_modZnew_modZold_attrZnew_attrr-rr rszMovedAttribute.__init__cCst|j}t||jSr)rr*r.r/)r modulerrr rs zMovedAttribute._resolve)NN)r rrrrr3rrr-r r9sr9c@sVeZdZdZddZddZddZdd d Zd d Zd dZ ddZ ddZ e Z dS)_SixMetaPathImporterz A meta path importer to import six.moves and its submodules. This class implements a PEP302 finder and loader. It should be compatible with Python 2.5 and all existing versions of Python3 cCs||_i|_dSr)r known_modules)r Zsix_module_namerrr rsz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdSN.r<r)r r*Z fullnamesfullnamerrr _add_modulesz _SixMetaPathImporter._add_modulecCs|j|jd|Sr=r?r r@rrr _get_modulesz _SixMetaPathImporter._get_moduleNcCs||jkr|SdSr)r<)r r@pathrrr find_modules z _SixMetaPathImporter.find_modulecCs2z |j|WStk r,td|YnXdS)Nz!This loader does not know module )r<KeyError ImportErrorrBrrr Z __get_modules z!_SixMetaPathImporter.__get_modulecCsTz tj|WStk r YnX||}t|tr@|}n||_|tj|<|Sr)rrrF _SixMetaPathImporter__get_module isinstancer'r __loader__)r r@r*rrr load_modules     z _SixMetaPathImporter.load_modulecCst||dS)z Return true, if the named module is a package. We need this method to get correct spec objects with Python 3.4 (see PEP451) __path__)hasattrrHrBrrr is_packagesz_SixMetaPathImporter.is_packagecCs||dS)z;Return None Required, if is_package is implementedN)rHrBrrr get_codes z_SixMetaPathImporter.get_code)N) r rrrrrArCrErHrKrNrO get_sourcerrrr r;s  r;c@seZdZdZgZdS) _MovedItemszLazy loading of moved objectsN)r rrrrLrrrr rQsrQZ cStringIOioStringIOfilter itertoolsbuiltinsZifilter filterfalseZ ifilterfalseinputZ __builtin__Z raw_inputinternrmapimapgetcwdosZgetcwdugetcwdbrangeZxrangeZ reload_module importlibZimpreloadreduce functoolsZ shlex_quoteZpipesZshlexZquoteUserDict collectionsUserList UserStringzipZizip zip_longestZ izip_longestZ configparserZ ConfigParsercopyregZcopy_regZdbm_gnuZgdbmzdbm.gnuZ _dummy_threadZ dummy_threadZhttp_cookiejarZ cookielibzhttp.cookiejarZ http_cookiesZCookiez http.cookiesZ html_entitiesZhtmlentitydefsz html.entitiesZ html_parserZ HTMLParserz html.parserZ http_clientZhttplibz http.clientZemail_mime_multipartzemail.MIMEMultipartzemail.mime.multipartZemail_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipartZemail_mime_textzemail.MIMETextzemail.mime.textZemail_mime_basezemail.MIMEBasezemail.mime.baseZBaseHTTPServerz http.serverZ CGIHTTPServerZSimpleHTTPServerZcPicklepickleZqueueZQueuereprlibreprZ socketserverZ SocketServer_threadthreadZtkinterZTkinterZtkinter_dialogZDialogztkinter.dialogZtkinter_filedialogZ FileDialogztkinter.filedialogZtkinter_scrolledtextZ ScrolledTextztkinter.scrolledtextZtkinter_simpledialogZ SimpleDialogztkinter.simpledialogZ tkinter_tixZTixz tkinter.tixZ tkinter_ttkZttkz tkinter.ttkZtkinter_constantsZ Tkconstantsztkinter.constantsZ tkinter_dndZTkdndz tkinter.dndZtkinter_colorchooserZtkColorChooserztkinter.colorchooserZtkinter_commondialogZtkCommonDialogztkinter.commondialogZtkinter_tkfiledialogZ tkFileDialogZ tkinter_fontZtkFontz tkinter.fontZtkinter_messageboxZ tkMessageBoxztkinter.messageboxZtkinter_tksimpledialogZtkSimpleDialogZ urllib_parsez.moves.urllib_parsez urllib.parseZ urllib_errorz.moves.urllib_errorz urllib.errorZurllibz .moves.urllibZurllib_robotparser robotparserzurllib.robotparserZ xmlrpc_clientZ xmlrpclibz xmlrpc.clientZ xmlrpc_serverZSimpleXMLRPCServerz xmlrpc.serverZwin32winreg_winregzmoves.z.movesmovesc@seZdZdZdS)Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseNr rrrrrrr rt@srtZ ParseResultZurlparseZ SplitResultZparse_qsZ parse_qslZ urldefragZurljoinZurlsplitZ urlunparseZ urlunsplitZ quote_plusZunquoteZ unquote_plusZ urlencodeZ splitqueryZsplittagZ splituserZ uses_fragmentZ uses_netlocZ uses_paramsZ uses_queryZ uses_relativemoves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNrurrrr rwhsrwZURLErrorZurllib2Z HTTPErrorZContentTooShortErrorz.moves.urllib.errormoves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNrurrrr ry|sryZurlopenzurllib.requestZinstall_openerZ build_openerZ pathname2urlZ url2pathnameZ getproxiesZRequestZOpenerDirectorZHTTPDefaultErrorHandlerZHTTPRedirectHandlerZHTTPCookieProcessorZ ProxyHandlerZ BaseHandlerZHTTPPasswordMgrZHTTPPasswordMgrWithDefaultRealmZAbstractBasicAuthHandlerZHTTPBasicAuthHandlerZProxyBasicAuthHandlerZAbstractDigestAuthHandlerZHTTPDigestAuthHandlerZProxyDigestAuthHandlerZ HTTPHandlerZ HTTPSHandlerZ FileHandlerZ FTPHandlerZCacheFTPHandlerZUnknownHandlerZHTTPErrorProcessorZ urlretrieveZ urlcleanupZ URLopenerZFancyURLopenerZ proxy_bypassz.moves.urllib.requestmoves.urllib_requestzmoves.urllib.requestc@seZdZdZdS) Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNrurrrr r{sr{Zaddbasezurllib.responseZ addclosehookZaddinfoZ addinfourlz.moves.urllib.responsemoves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNrurrrr r}sr}ZRobotFileParserz.moves.urllib.robotparsermoves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZedZedZedZ edZ edZ ddZ d S) Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacervrxrzr|r~cCsdddddgS)Nparseerrorrequestresponserprr rrr r8szModule_six_moves_urllib.__dir__N) r rrrrL _importerrCrrrrrpr8rrrr rs     rz moves.urllibcCstt|j|dS)zAdd an item to six.moves.N)rrQr)Zmoverrr add_movesrc CsXztt|WnDtk rRz tj|=Wn"tk rLtd|fYnXYnXdS)zRemove item from six.moves.zno such move, %rN)r rQr"rs__dict__rFrrrr remove_moves r__func____self__ __closure____code__ __defaults__ __globals__im_funcZim_selfZ func_closureZ func_codeZ func_defaultsZ func_globalscCs|Sr)next)itrrr advance_iterator srcCstddt|jDS)Ncss|]}d|jkVqdS)__call__N)r)r5klassrrr szcallable..)anytype__mro__)r#rrr callablesrcCs|SrrZunboundrrr get_unbound_functionsrcCs|Srrrclsrrr create_unbound_methodsrcCs|jSr)rrrrr r"scCst|||jSr)types MethodTyper!)rr#rrr create_bound_method%srcCst|d|Sr)rrrrrr r(sc@seZdZddZdS)IteratorcCst||Sr)r__next__r rrr r-sz Iterator.nextN)r rrrrrrr r+srz3Get the function out of a possibly unbound functioncKst|jf|Sr)iterkeysdkwrrr iterkeys>srcKst|jf|Sr)rvaluesrrrr itervaluesAsrcKst|jf|Sr)ritemsrrrr iteritemsDsrcKst|jf|Sr)rZlistsrrrr iterlistsGsrrrrcKs |jf|Sr)rrrrr rPscKs |jf|Sr)rrrrr rSscKs |jf|Sr)rrrrr rVscKs |jf|Sr)rrrrr rYsviewkeys viewvalues viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs |dS)Nzlatin-1)encodesrrr bksrcCs|Srrrrrr unsrz>BassertCountEqualZassertRaisesRegexpZassertRegexpMatchesassertRaisesRegex assertRegexcCs|Srrrrrr rscCst|dddS)Nz\\z\\\\Zunicode_escape)unicodereplacerrrr rscCs t|dS)Nrord)Zbsrrr byte2intsrcCs t||Srr)Zbufirrr indexbytessrZassertItemsEqualz Byte literalz Text literalcOst|t||Sr)r._assertCountEqualr argskwargsrrr rscOst|t||Sr)r._assertRaisesRegexrrrr rscOst|t||Sr)r. _assertRegexrrrr rsexeccCs*|dkr|}|j|k r"|||dSr) __traceback__with_traceback)r$r1tbrrr reraises   rcCsB|dkr*td}|j}|dkr&|j}~n |dkr6|}tddS)zExecute code in a namespace.Nrzexec _code_ in _globs_, _locs_)r _getframe f_globalsf_localsr)Z_code_Z_globs_Z_locs_framerrr exec_s rz9def reraise(tp, value, tb=None): raise tp, value, tb )rrzrdef raise_from(value, from_value): if from_value is None: raise value raise value from from_value zCdef raise_from(value, from_value): raise value from from_value cCs|dSrr)r1Z from_valuerrr raise_fromsrprintc s.|dtjdkrdSfdd}d}|dd}|dk r`t|trNd}nt|ts`td|d d}|dk rt|trd}nt|tstd |rtd |s|D]}t|trd}qq|rtd }td }nd }d }|dkr|}|dkr|}t|D] \} }| r||||q||dS)z4The new-style print function for Python 2.4 and 2.5.fileNcsdt|tst|}ttrVt|trVjdk rVtdd}|dkrHd}|j|}|dS)Nerrorsstrict) rI basestringstrrrencodingr.rwrite)datarfprr rs   zprint_..writeFsepTzsep must be None or a stringendzend must be None or a stringz$invalid keyword arguments to print()  )poprstdoutrIrr TypeError enumerate) rrrZ want_unicoderrargnewlineZspacerrrr print_sL          r)rrcOs<|dtj}|dd}t|||r8|dk r8|dS)NrflushF)getrrr_printr)rrrrrrr r s    zReraise an exception.csfdd}|S)Ncst|}|_|Sr)rcwraps __wrapped__)fassignedupdatedwrappedrr wrapperszwraps..wrapperr)rrrrrrr rsrcs&Gfddd}t|ddiS)z%Create a base class with a metaclass.cseZdZfddZdS)z!with_metaclass..metaclasscs ||Srr)rrZ this_basesrbasesmetarr __new__'sz)with_metaclass..metaclass.__new__N)r rrrrrrr metaclass%srZtemporary_classr)rr)rrrrrr with_metaclass srcsfdd}|S)z6Class decorator for creating a class with a metaclass.csh|j}|d}|dk r@t|tr,|g}|D]}||q0|dd|dd|j|j|S)N __slots__r __weakref__)rcopyrrIrrr __bases__)rZ orig_varsslotsZ slots_varrrr r.s      zadd_metaclass..wrapperr)rrrrr add_metaclass,s rcCs2tr.d|jkrtd|j|j|_dd|_|S)a A decorator that defines __unicode__ and __str__ methods under Python 2. Under Python 3 it does nothing. To support Python 2 and 3 with a single code base, define a __str__ method returning text and apply this decorator to the class. __str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|dS)Nzutf-8) __unicode__rr rrr Jz-python_2_unicode_compatible..)PY2r ValueErrorr rr)rrrr python_2_unicode_compatible<s  r__spec__)N)NN)rZ __future__rrcrUoperatorrr __author__ __version__ version_inforr)ZPY34rZ string_typesintZ integer_typesrZ class_typesZ text_typebytesZ binary_typemaxsizeZMAXSIZErZlongZ ClassTyperplatform startswithobjectrlen OverflowErrorrrrr' ModuleTyper4r9r;r rrQr7r/rrrIrArsrtZ_urllib_parse_moved_attributesrwZ_urllib_error_moved_attributesryZ _urllib_request_moved_attributesr{Z!_urllib_response_moved_attributesr}Z$_urllib_robotparser_moved_attributesrrrZ _meth_funcZ _meth_selfZ _func_closureZ _func_codeZ_func_defaultsZ _func_globalsrr NameErrorrrrrrr attrgetterZget_method_functionZget_method_selfZget_function_closureZget_function_codeZget_function_defaultsZget_function_globalsrrrr methodcallerrrrrrchrZunichrstructStructpackZint2byte itemgetterrgetitemrrZ iterbytesrRrSBytesIOrrrpartialr[rrrrr.rVrrrrrWRAPPER_ASSIGNMENTSWRAPPER_UPDATESrrrrrL __package__globalsrrsubmodule_search_locations meta_pathrrZimporterappendrrrr s    >                                      D                                                               #                                                 5     site-packages/setuptools/_vendor/ordered_set.py000064400000035432151464452710016032 0ustar00""" An OrderedSet is a custom MutableSet that remembers its order, so that every entry has an index that can be looked up. Based on a recipe originally posted to ActiveState Recipes by Raymond Hettiger, and released under the MIT license. """ import itertools as it from collections import deque try: # Python 3 from collections.abc import MutableSet, Sequence except ImportError: # Python 2.7 from collections import MutableSet, Sequence SLICE_ALL = slice(None) __version__ = "3.1" def is_iterable(obj): """ Are we being asked to look up a list of things, instead of a single thing? We check for the `__iter__` attribute so that this can cover types that don't have to be known by this module, such as NumPy arrays. Strings, however, should be considered as atomic values to look up, not iterables. The same goes for tuples, since they are immutable and therefore valid entries. We don't need to check for the Python 2 `unicode` type, because it doesn't have an `__iter__` attribute anyway. """ return ( hasattr(obj, "__iter__") and not isinstance(obj, str) and not isinstance(obj, tuple) ) class OrderedSet(MutableSet, Sequence): """ An OrderedSet is a custom MutableSet that remembers its order, so that every entry has an index that can be looked up. Example: >>> OrderedSet([1, 1, 2, 3, 2]) OrderedSet([1, 2, 3]) """ def __init__(self, iterable=None): self.items = [] self.map = {} if iterable is not None: self |= iterable def __len__(self): """ Returns the number of unique elements in the ordered set Example: >>> len(OrderedSet([])) 0 >>> len(OrderedSet([1, 2])) 2 """ return len(self.items) def __getitem__(self, index): """ Get the item at a given index. If `index` is a slice, you will get back that slice of items, as a new OrderedSet. If `index` is a list or a similar iterable, you'll get a list of items corresponding to those indices. This is similar to NumPy's "fancy indexing". The result is not an OrderedSet because you may ask for duplicate indices, and the number of elements returned should be the number of elements asked for. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset[1] 2 """ if isinstance(index, slice) and index == SLICE_ALL: return self.copy() elif is_iterable(index): return [self.items[i] for i in index] elif hasattr(index, "__index__") or isinstance(index, slice): result = self.items[index] if isinstance(result, list): return self.__class__(result) else: return result else: raise TypeError("Don't know how to index an OrderedSet by %r" % index) def copy(self): """ Return a shallow copy of this object. Example: >>> this = OrderedSet([1, 2, 3]) >>> other = this.copy() >>> this == other True >>> this is other False """ return self.__class__(self) def __getstate__(self): if len(self) == 0: # The state can't be an empty list. # We need to return a truthy value, or else __setstate__ won't be run. # # This could have been done more gracefully by always putting the state # in a tuple, but this way is backwards- and forwards- compatible with # previous versions of OrderedSet. return (None,) else: return list(self) def __setstate__(self, state): if state == (None,): self.__init__([]) else: self.__init__(state) def __contains__(self, key): """ Test if the item is in this ordered set Example: >>> 1 in OrderedSet([1, 3, 2]) True >>> 5 in OrderedSet([1, 3, 2]) False """ return key in self.map def add(self, key): """ Add `key` as an item to this OrderedSet, then return its index. If `key` is already in the OrderedSet, return the index it already had. Example: >>> oset = OrderedSet() >>> oset.append(3) 0 >>> print(oset) OrderedSet([3]) """ if key not in self.map: self.map[key] = len(self.items) self.items.append(key) return self.map[key] append = add def update(self, sequence): """ Update the set with the given iterable sequence, then return the index of the last element inserted. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.update([3, 1, 5, 1, 4]) 4 >>> print(oset) OrderedSet([1, 2, 3, 5, 4]) """ item_index = None try: for item in sequence: item_index = self.add(item) except TypeError: raise ValueError( "Argument needs to be an iterable, got %s" % type(sequence) ) return item_index def index(self, key): """ Get the index of a given entry, raising an IndexError if it's not present. `key` can be an iterable of entries that is not a string, in which case this returns a list of indices. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.index(2) 1 """ if is_iterable(key): return [self.index(subkey) for subkey in key] return self.map[key] # Provide some compatibility with pd.Index get_loc = index get_indexer = index def pop(self): """ Remove and return the last element from the set. Raises KeyError if the set is empty. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.pop() 3 """ if not self.items: raise KeyError("Set is empty") elem = self.items[-1] del self.items[-1] del self.map[elem] return elem def discard(self, key): """ Remove an element. Do not raise an exception if absent. The MutableSet mixin uses this to implement the .remove() method, which *does* raise an error when asked to remove a non-existent item. Example: >>> oset = OrderedSet([1, 2, 3]) >>> oset.discard(2) >>> print(oset) OrderedSet([1, 3]) >>> oset.discard(2) >>> print(oset) OrderedSet([1, 3]) """ if key in self: i = self.map[key] del self.items[i] del self.map[key] for k, v in self.map.items(): if v >= i: self.map[k] = v - 1 def clear(self): """ Remove all items from this OrderedSet. """ del self.items[:] self.map.clear() def __iter__(self): """ Example: >>> list(iter(OrderedSet([1, 2, 3]))) [1, 2, 3] """ return iter(self.items) def __reversed__(self): """ Example: >>> list(reversed(OrderedSet([1, 2, 3]))) [3, 2, 1] """ return reversed(self.items) def __repr__(self): if not self: return "%s()" % (self.__class__.__name__,) return "%s(%r)" % (self.__class__.__name__, list(self)) def __eq__(self, other): """ Returns true if the containers have the same items. If `other` is a Sequence, then order is checked, otherwise it is ignored. Example: >>> oset = OrderedSet([1, 3, 2]) >>> oset == [1, 3, 2] True >>> oset == [1, 2, 3] False >>> oset == [2, 3] False >>> oset == OrderedSet([3, 2, 1]) False """ # In Python 2 deque is not a Sequence, so treat it as one for # consistent behavior with Python 3. if isinstance(other, (Sequence, deque)): # Check that this OrderedSet contains the same elements, in the # same order, as the other object. return list(self) == list(other) try: other_as_set = set(other) except TypeError: # If `other` can't be converted into a set, it's not equal. return False else: return set(self) == other_as_set def union(self, *sets): """ Combines all unique items. Each items order is defined by its first appearance. Example: >>> oset = OrderedSet.union(OrderedSet([3, 1, 4, 1, 5]), [1, 3], [2, 0]) >>> print(oset) OrderedSet([3, 1, 4, 5, 2, 0]) >>> oset.union([8, 9]) OrderedSet([3, 1, 4, 5, 2, 0, 8, 9]) >>> oset | {10} OrderedSet([3, 1, 4, 5, 2, 0, 10]) """ cls = self.__class__ if isinstance(self, OrderedSet) else OrderedSet containers = map(list, it.chain([self], sets)) items = it.chain.from_iterable(containers) return cls(items) def __and__(self, other): # the parent implementation of this is backwards return self.intersection(other) def intersection(self, *sets): """ Returns elements in common between all sets. Order is defined only by the first set. Example: >>> oset = OrderedSet.intersection(OrderedSet([0, 1, 2, 3]), [1, 2, 3]) >>> print(oset) OrderedSet([1, 2, 3]) >>> oset.intersection([2, 4, 5], [1, 2, 3, 4]) OrderedSet([2]) >>> oset.intersection() OrderedSet([1, 2, 3]) """ cls = self.__class__ if isinstance(self, OrderedSet) else OrderedSet if sets: common = set.intersection(*map(set, sets)) items = (item for item in self if item in common) else: items = self return cls(items) def difference(self, *sets): """ Returns all elements that are in this set but not the others. Example: >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2])) OrderedSet([1, 3]) >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]), OrderedSet([3])) OrderedSet([1]) >>> OrderedSet([1, 2, 3]) - OrderedSet([2]) OrderedSet([1, 3]) >>> OrderedSet([1, 2, 3]).difference() OrderedSet([1, 2, 3]) """ cls = self.__class__ if sets: other = set.union(*map(set, sets)) items = (item for item in self if item not in other) else: items = self return cls(items) def issubset(self, other): """ Report whether another set contains this set. Example: >>> OrderedSet([1, 2, 3]).issubset({1, 2}) False >>> OrderedSet([1, 2, 3]).issubset({1, 2, 3, 4}) True >>> OrderedSet([1, 2, 3]).issubset({1, 4, 3, 5}) False """ if len(self) > len(other): # Fast check for obvious cases return False return all(item in other for item in self) def issuperset(self, other): """ Report whether this set contains another set. Example: >>> OrderedSet([1, 2]).issuperset([1, 2, 3]) False >>> OrderedSet([1, 2, 3, 4]).issuperset({1, 2, 3}) True >>> OrderedSet([1, 4, 3, 5]).issuperset({1, 2, 3}) False """ if len(self) < len(other): # Fast check for obvious cases return False return all(item in self for item in other) def symmetric_difference(self, other): """ Return the symmetric difference of two OrderedSets as a new set. That is, the new set will contain all elements that are in exactly one of the sets. Their order will be preserved, with elements from `self` preceding elements from `other`. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.symmetric_difference(other) OrderedSet([4, 5, 9, 2]) """ cls = self.__class__ if isinstance(self, OrderedSet) else OrderedSet diff1 = cls(self).difference(other) diff2 = cls(other).difference(self) return diff1.union(diff2) def _update_items(self, items): """ Replace the 'items' list of this OrderedSet with a new one, updating self.map accordingly. """ self.items = items self.map = {item: idx for (idx, item) in enumerate(items)} def difference_update(self, *sets): """ Update this OrderedSet to remove items from one or more other sets. Example: >>> this = OrderedSet([1, 2, 3]) >>> this.difference_update(OrderedSet([2, 4])) >>> print(this) OrderedSet([1, 3]) >>> this = OrderedSet([1, 2, 3, 4, 5]) >>> this.difference_update(OrderedSet([2, 4]), OrderedSet([1, 4, 6])) >>> print(this) OrderedSet([3, 5]) """ items_to_remove = set() for other in sets: items_to_remove |= set(other) self._update_items([item for item in self.items if item not in items_to_remove]) def intersection_update(self, other): """ Update this OrderedSet to keep only items in another set, preserving their order in this set. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.intersection_update(other) >>> print(this) OrderedSet([1, 3, 7]) """ other = set(other) self._update_items([item for item in self.items if item in other]) def symmetric_difference_update(self, other): """ Update this OrderedSet to remove items from another set, then add items from the other set that were not present in this set. Example: >>> this = OrderedSet([1, 4, 3, 5, 7]) >>> other = OrderedSet([9, 7, 1, 3, 2]) >>> this.symmetric_difference_update(other) >>> print(this) OrderedSet([4, 5, 9, 2]) """ items_to_add = [item for item in other if item not in self] items_to_remove = set(other) self._update_items( [item for item in self.items if item not in items_to_remove] + items_to_add ) site-packages/setuptools/_vendor/six.py000064400000072622151464452710014340 0ustar00"""Utilities for writing code that runs on Python 2 and 3""" # Copyright (c) 2010-2015 Benjamin Peterson # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. from __future__ import absolute_import import functools import itertools import operator import sys import types __author__ = "Benjamin Peterson " __version__ = "1.10.0" # Useful for very coarse version differentiation. PY2 = sys.version_info[0] == 2 PY3 = sys.version_info[0] == 3 PY34 = sys.version_info[0:2] >= (3, 4) if PY3: string_types = str, integer_types = int, class_types = type, text_type = str binary_type = bytes MAXSIZE = sys.maxsize else: string_types = basestring, integer_types = (int, long) class_types = (type, types.ClassType) text_type = unicode binary_type = str if sys.platform.startswith("java"): # Jython always uses 32 bits. MAXSIZE = int((1 << 31) - 1) else: # It's possible to have sizeof(long) != sizeof(Py_ssize_t). class X(object): def __len__(self): return 1 << 31 try: len(X()) except OverflowError: # 32-bit MAXSIZE = int((1 << 31) - 1) else: # 64-bit MAXSIZE = int((1 << 63) - 1) del X def _add_doc(func, doc): """Add documentation to a function.""" func.__doc__ = doc def _import_module(name): """Import module, returning the module after the last dot.""" __import__(name) return sys.modules[name] class _LazyDescr(object): def __init__(self, name): self.name = name def __get__(self, obj, tp): result = self._resolve() setattr(obj, self.name, result) # Invokes __set__. try: # This is a bit ugly, but it avoids running this again by # removing this descriptor. delattr(obj.__class__, self.name) except AttributeError: pass return result class MovedModule(_LazyDescr): def __init__(self, name, old, new=None): super(MovedModule, self).__init__(name) if PY3: if new is None: new = name self.mod = new else: self.mod = old def _resolve(self): return _import_module(self.mod) def __getattr__(self, attr): _module = self._resolve() value = getattr(_module, attr) setattr(self, attr, value) return value class _LazyModule(types.ModuleType): def __init__(self, name): super(_LazyModule, self).__init__(name) self.__doc__ = self.__class__.__doc__ def __dir__(self): attrs = ["__doc__", "__name__"] attrs += [attr.name for attr in self._moved_attributes] return attrs # Subclasses should override this _moved_attributes = [] class MovedAttribute(_LazyDescr): def __init__(self, name, old_mod, new_mod, old_attr=None, new_attr=None): super(MovedAttribute, self).__init__(name) if PY3: if new_mod is None: new_mod = name self.mod = new_mod if new_attr is None: if old_attr is None: new_attr = name else: new_attr = old_attr self.attr = new_attr else: self.mod = old_mod if old_attr is None: old_attr = name self.attr = old_attr def _resolve(self): module = _import_module(self.mod) return getattr(module, self.attr) class _SixMetaPathImporter(object): """ A meta path importer to import six.moves and its submodules. This class implements a PEP302 finder and loader. It should be compatible with Python 2.5 and all existing versions of Python3 """ def __init__(self, six_module_name): self.name = six_module_name self.known_modules = {} def _add_module(self, mod, *fullnames): for fullname in fullnames: self.known_modules[self.name + "." + fullname] = mod def _get_module(self, fullname): return self.known_modules[self.name + "." + fullname] def find_module(self, fullname, path=None): if fullname in self.known_modules: return self return None def __get_module(self, fullname): try: return self.known_modules[fullname] except KeyError: raise ImportError("This loader does not know module " + fullname) def load_module(self, fullname): try: # in case of a reload return sys.modules[fullname] except KeyError: pass mod = self.__get_module(fullname) if isinstance(mod, MovedModule): mod = mod._resolve() else: mod.__loader__ = self sys.modules[fullname] = mod return mod def is_package(self, fullname): """ Return true, if the named module is a package. We need this method to get correct spec objects with Python 3.4 (see PEP451) """ return hasattr(self.__get_module(fullname), "__path__") def get_code(self, fullname): """Return None Required, if is_package is implemented""" self.__get_module(fullname) # eventually raises ImportError return None get_source = get_code # same as get_code _importer = _SixMetaPathImporter(__name__) class _MovedItems(_LazyModule): """Lazy loading of moved objects""" __path__ = [] # mark as package _moved_attributes = [ MovedAttribute("cStringIO", "cStringIO", "io", "StringIO"), MovedAttribute("filter", "itertools", "builtins", "ifilter", "filter"), MovedAttribute("filterfalse", "itertools", "itertools", "ifilterfalse", "filterfalse"), MovedAttribute("input", "__builtin__", "builtins", "raw_input", "input"), MovedAttribute("intern", "__builtin__", "sys"), MovedAttribute("map", "itertools", "builtins", "imap", "map"), MovedAttribute("getcwd", "os", "os", "getcwdu", "getcwd"), MovedAttribute("getcwdb", "os", "os", "getcwd", "getcwdb"), MovedAttribute("range", "__builtin__", "builtins", "xrange", "range"), MovedAttribute("reload_module", "__builtin__", "importlib" if PY34 else "imp", "reload"), MovedAttribute("reduce", "__builtin__", "functools"), MovedAttribute("shlex_quote", "pipes", "shlex", "quote"), MovedAttribute("StringIO", "StringIO", "io"), MovedAttribute("UserDict", "UserDict", "collections"), MovedAttribute("UserList", "UserList", "collections"), MovedAttribute("UserString", "UserString", "collections"), MovedAttribute("xrange", "__builtin__", "builtins", "xrange", "range"), MovedAttribute("zip", "itertools", "builtins", "izip", "zip"), MovedAttribute("zip_longest", "itertools", "itertools", "izip_longest", "zip_longest"), MovedModule("builtins", "__builtin__"), MovedModule("configparser", "ConfigParser"), MovedModule("copyreg", "copy_reg"), MovedModule("dbm_gnu", "gdbm", "dbm.gnu"), MovedModule("_dummy_thread", "dummy_thread", "_dummy_thread"), MovedModule("http_cookiejar", "cookielib", "http.cookiejar"), MovedModule("http_cookies", "Cookie", "http.cookies"), MovedModule("html_entities", "htmlentitydefs", "html.entities"), MovedModule("html_parser", "HTMLParser", "html.parser"), MovedModule("http_client", "httplib", "http.client"), MovedModule("email_mime_multipart", "email.MIMEMultipart", "email.mime.multipart"), MovedModule("email_mime_nonmultipart", "email.MIMENonMultipart", "email.mime.nonmultipart"), MovedModule("email_mime_text", "email.MIMEText", "email.mime.text"), MovedModule("email_mime_base", "email.MIMEBase", "email.mime.base"), MovedModule("BaseHTTPServer", "BaseHTTPServer", "http.server"), MovedModule("CGIHTTPServer", "CGIHTTPServer", "http.server"), MovedModule("SimpleHTTPServer", "SimpleHTTPServer", "http.server"), MovedModule("cPickle", "cPickle", "pickle"), MovedModule("queue", "Queue"), MovedModule("reprlib", "repr"), MovedModule("socketserver", "SocketServer"), MovedModule("_thread", "thread", "_thread"), MovedModule("tkinter", "Tkinter"), MovedModule("tkinter_dialog", "Dialog", "tkinter.dialog"), MovedModule("tkinter_filedialog", "FileDialog", "tkinter.filedialog"), MovedModule("tkinter_scrolledtext", "ScrolledText", "tkinter.scrolledtext"), MovedModule("tkinter_simpledialog", "SimpleDialog", "tkinter.simpledialog"), MovedModule("tkinter_tix", "Tix", "tkinter.tix"), MovedModule("tkinter_ttk", "ttk", "tkinter.ttk"), MovedModule("tkinter_constants", "Tkconstants", "tkinter.constants"), MovedModule("tkinter_dnd", "Tkdnd", "tkinter.dnd"), MovedModule("tkinter_colorchooser", "tkColorChooser", "tkinter.colorchooser"), MovedModule("tkinter_commondialog", "tkCommonDialog", "tkinter.commondialog"), MovedModule("tkinter_tkfiledialog", "tkFileDialog", "tkinter.filedialog"), MovedModule("tkinter_font", "tkFont", "tkinter.font"), MovedModule("tkinter_messagebox", "tkMessageBox", "tkinter.messagebox"), MovedModule("tkinter_tksimpledialog", "tkSimpleDialog", "tkinter.simpledialog"), MovedModule("urllib_parse", __name__ + ".moves.urllib_parse", "urllib.parse"), MovedModule("urllib_error", __name__ + ".moves.urllib_error", "urllib.error"), MovedModule("urllib", __name__ + ".moves.urllib", __name__ + ".moves.urllib"), MovedModule("urllib_robotparser", "robotparser", "urllib.robotparser"), MovedModule("xmlrpc_client", "xmlrpclib", "xmlrpc.client"), MovedModule("xmlrpc_server", "SimpleXMLRPCServer", "xmlrpc.server"), ] # Add windows specific modules. if sys.platform == "win32": _moved_attributes += [ MovedModule("winreg", "_winreg"), ] for attr in _moved_attributes: setattr(_MovedItems, attr.name, attr) if isinstance(attr, MovedModule): _importer._add_module(attr, "moves." + attr.name) del attr _MovedItems._moved_attributes = _moved_attributes moves = _MovedItems(__name__ + ".moves") _importer._add_module(moves, "moves") class Module_six_moves_urllib_parse(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_parse""" _urllib_parse_moved_attributes = [ MovedAttribute("ParseResult", "urlparse", "urllib.parse"), MovedAttribute("SplitResult", "urlparse", "urllib.parse"), MovedAttribute("parse_qs", "urlparse", "urllib.parse"), MovedAttribute("parse_qsl", "urlparse", "urllib.parse"), MovedAttribute("urldefrag", "urlparse", "urllib.parse"), MovedAttribute("urljoin", "urlparse", "urllib.parse"), MovedAttribute("urlparse", "urlparse", "urllib.parse"), MovedAttribute("urlsplit", "urlparse", "urllib.parse"), MovedAttribute("urlunparse", "urlparse", "urllib.parse"), MovedAttribute("urlunsplit", "urlparse", "urllib.parse"), MovedAttribute("quote", "urllib", "urllib.parse"), MovedAttribute("quote_plus", "urllib", "urllib.parse"), MovedAttribute("unquote", "urllib", "urllib.parse"), MovedAttribute("unquote_plus", "urllib", "urllib.parse"), MovedAttribute("urlencode", "urllib", "urllib.parse"), MovedAttribute("splitquery", "urllib", "urllib.parse"), MovedAttribute("splittag", "urllib", "urllib.parse"), MovedAttribute("splituser", "urllib", "urllib.parse"), MovedAttribute("uses_fragment", "urlparse", "urllib.parse"), MovedAttribute("uses_netloc", "urlparse", "urllib.parse"), MovedAttribute("uses_params", "urlparse", "urllib.parse"), MovedAttribute("uses_query", "urlparse", "urllib.parse"), MovedAttribute("uses_relative", "urlparse", "urllib.parse"), ] for attr in _urllib_parse_moved_attributes: setattr(Module_six_moves_urllib_parse, attr.name, attr) del attr Module_six_moves_urllib_parse._moved_attributes = _urllib_parse_moved_attributes _importer._add_module(Module_six_moves_urllib_parse(__name__ + ".moves.urllib_parse"), "moves.urllib_parse", "moves.urllib.parse") class Module_six_moves_urllib_error(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_error""" _urllib_error_moved_attributes = [ MovedAttribute("URLError", "urllib2", "urllib.error"), MovedAttribute("HTTPError", "urllib2", "urllib.error"), MovedAttribute("ContentTooShortError", "urllib", "urllib.error"), ] for attr in _urllib_error_moved_attributes: setattr(Module_six_moves_urllib_error, attr.name, attr) del attr Module_six_moves_urllib_error._moved_attributes = _urllib_error_moved_attributes _importer._add_module(Module_six_moves_urllib_error(__name__ + ".moves.urllib.error"), "moves.urllib_error", "moves.urllib.error") class Module_six_moves_urllib_request(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_request""" _urllib_request_moved_attributes = [ MovedAttribute("urlopen", "urllib2", "urllib.request"), MovedAttribute("install_opener", "urllib2", "urllib.request"), MovedAttribute("build_opener", "urllib2", "urllib.request"), MovedAttribute("pathname2url", "urllib", "urllib.request"), MovedAttribute("url2pathname", "urllib", "urllib.request"), MovedAttribute("getproxies", "urllib", "urllib.request"), MovedAttribute("Request", "urllib2", "urllib.request"), MovedAttribute("OpenerDirector", "urllib2", "urllib.request"), MovedAttribute("HTTPDefaultErrorHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPRedirectHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPCookieProcessor", "urllib2", "urllib.request"), MovedAttribute("ProxyHandler", "urllib2", "urllib.request"), MovedAttribute("BaseHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPPasswordMgr", "urllib2", "urllib.request"), MovedAttribute("HTTPPasswordMgrWithDefaultRealm", "urllib2", "urllib.request"), MovedAttribute("AbstractBasicAuthHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPBasicAuthHandler", "urllib2", "urllib.request"), MovedAttribute("ProxyBasicAuthHandler", "urllib2", "urllib.request"), MovedAttribute("AbstractDigestAuthHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPDigestAuthHandler", "urllib2", "urllib.request"), MovedAttribute("ProxyDigestAuthHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPSHandler", "urllib2", "urllib.request"), MovedAttribute("FileHandler", "urllib2", "urllib.request"), MovedAttribute("FTPHandler", "urllib2", "urllib.request"), MovedAttribute("CacheFTPHandler", "urllib2", "urllib.request"), MovedAttribute("UnknownHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPErrorProcessor", "urllib2", "urllib.request"), MovedAttribute("urlretrieve", "urllib", "urllib.request"), MovedAttribute("urlcleanup", "urllib", "urllib.request"), MovedAttribute("URLopener", "urllib", "urllib.request"), MovedAttribute("FancyURLopener", "urllib", "urllib.request"), MovedAttribute("proxy_bypass", "urllib", "urllib.request"), ] for attr in _urllib_request_moved_attributes: setattr(Module_six_moves_urllib_request, attr.name, attr) del attr Module_six_moves_urllib_request._moved_attributes = _urllib_request_moved_attributes _importer._add_module(Module_six_moves_urllib_request(__name__ + ".moves.urllib.request"), "moves.urllib_request", "moves.urllib.request") class Module_six_moves_urllib_response(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_response""" _urllib_response_moved_attributes = [ MovedAttribute("addbase", "urllib", "urllib.response"), MovedAttribute("addclosehook", "urllib", "urllib.response"), MovedAttribute("addinfo", "urllib", "urllib.response"), MovedAttribute("addinfourl", "urllib", "urllib.response"), ] for attr in _urllib_response_moved_attributes: setattr(Module_six_moves_urllib_response, attr.name, attr) del attr Module_six_moves_urllib_response._moved_attributes = _urllib_response_moved_attributes _importer._add_module(Module_six_moves_urllib_response(__name__ + ".moves.urllib.response"), "moves.urllib_response", "moves.urllib.response") class Module_six_moves_urllib_robotparser(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_robotparser""" _urllib_robotparser_moved_attributes = [ MovedAttribute("RobotFileParser", "robotparser", "urllib.robotparser"), ] for attr in _urllib_robotparser_moved_attributes: setattr(Module_six_moves_urllib_robotparser, attr.name, attr) del attr Module_six_moves_urllib_robotparser._moved_attributes = _urllib_robotparser_moved_attributes _importer._add_module(Module_six_moves_urllib_robotparser(__name__ + ".moves.urllib.robotparser"), "moves.urllib_robotparser", "moves.urllib.robotparser") class Module_six_moves_urllib(types.ModuleType): """Create a six.moves.urllib namespace that resembles the Python 3 namespace""" __path__ = [] # mark as package parse = _importer._get_module("moves.urllib_parse") error = _importer._get_module("moves.urllib_error") request = _importer._get_module("moves.urllib_request") response = _importer._get_module("moves.urllib_response") robotparser = _importer._get_module("moves.urllib_robotparser") def __dir__(self): return ['parse', 'error', 'request', 'response', 'robotparser'] _importer._add_module(Module_six_moves_urllib(__name__ + ".moves.urllib"), "moves.urllib") def add_move(move): """Add an item to six.moves.""" setattr(_MovedItems, move.name, move) def remove_move(name): """Remove item from six.moves.""" try: delattr(_MovedItems, name) except AttributeError: try: del moves.__dict__[name] except KeyError: raise AttributeError("no such move, %r" % (name,)) if PY3: _meth_func = "__func__" _meth_self = "__self__" _func_closure = "__closure__" _func_code = "__code__" _func_defaults = "__defaults__" _func_globals = "__globals__" else: _meth_func = "im_func" _meth_self = "im_self" _func_closure = "func_closure" _func_code = "func_code" _func_defaults = "func_defaults" _func_globals = "func_globals" try: advance_iterator = next except NameError: def advance_iterator(it): return it.next() next = advance_iterator try: callable = callable except NameError: def callable(obj): return any("__call__" in klass.__dict__ for klass in type(obj).__mro__) if PY3: def get_unbound_function(unbound): return unbound create_bound_method = types.MethodType def create_unbound_method(func, cls): return func Iterator = object else: def get_unbound_function(unbound): return unbound.im_func def create_bound_method(func, obj): return types.MethodType(func, obj, obj.__class__) def create_unbound_method(func, cls): return types.MethodType(func, None, cls) class Iterator(object): def next(self): return type(self).__next__(self) callable = callable _add_doc(get_unbound_function, """Get the function out of a possibly unbound function""") get_method_function = operator.attrgetter(_meth_func) get_method_self = operator.attrgetter(_meth_self) get_function_closure = operator.attrgetter(_func_closure) get_function_code = operator.attrgetter(_func_code) get_function_defaults = operator.attrgetter(_func_defaults) get_function_globals = operator.attrgetter(_func_globals) if PY3: def iterkeys(d, **kw): return iter(d.keys(**kw)) def itervalues(d, **kw): return iter(d.values(**kw)) def iteritems(d, **kw): return iter(d.items(**kw)) def iterlists(d, **kw): return iter(d.lists(**kw)) viewkeys = operator.methodcaller("keys") viewvalues = operator.methodcaller("values") viewitems = operator.methodcaller("items") else: def iterkeys(d, **kw): return d.iterkeys(**kw) def itervalues(d, **kw): return d.itervalues(**kw) def iteritems(d, **kw): return d.iteritems(**kw) def iterlists(d, **kw): return d.iterlists(**kw) viewkeys = operator.methodcaller("viewkeys") viewvalues = operator.methodcaller("viewvalues") viewitems = operator.methodcaller("viewitems") _add_doc(iterkeys, "Return an iterator over the keys of a dictionary.") _add_doc(itervalues, "Return an iterator over the values of a dictionary.") _add_doc(iteritems, "Return an iterator over the (key, value) pairs of a dictionary.") _add_doc(iterlists, "Return an iterator over the (key, [values]) pairs of a dictionary.") if PY3: def b(s): return s.encode("latin-1") def u(s): return s unichr = chr import struct int2byte = struct.Struct(">B").pack del struct byte2int = operator.itemgetter(0) indexbytes = operator.getitem iterbytes = iter import io StringIO = io.StringIO BytesIO = io.BytesIO _assertCountEqual = "assertCountEqual" if sys.version_info[1] <= 1: _assertRaisesRegex = "assertRaisesRegexp" _assertRegex = "assertRegexpMatches" else: _assertRaisesRegex = "assertRaisesRegex" _assertRegex = "assertRegex" else: def b(s): return s # Workaround for standalone backslash def u(s): return unicode(s.replace(r'\\', r'\\\\'), "unicode_escape") unichr = unichr int2byte = chr def byte2int(bs): return ord(bs[0]) def indexbytes(buf, i): return ord(buf[i]) iterbytes = functools.partial(itertools.imap, ord) import StringIO StringIO = BytesIO = StringIO.StringIO _assertCountEqual = "assertItemsEqual" _assertRaisesRegex = "assertRaisesRegexp" _assertRegex = "assertRegexpMatches" _add_doc(b, """Byte literal""") _add_doc(u, """Text literal""") def assertCountEqual(self, *args, **kwargs): return getattr(self, _assertCountEqual)(*args, **kwargs) def assertRaisesRegex(self, *args, **kwargs): return getattr(self, _assertRaisesRegex)(*args, **kwargs) def assertRegex(self, *args, **kwargs): return getattr(self, _assertRegex)(*args, **kwargs) if PY3: exec_ = getattr(moves.builtins, "exec") def reraise(tp, value, tb=None): if value is None: value = tp() if value.__traceback__ is not tb: raise value.with_traceback(tb) raise value else: def exec_(_code_, _globs_=None, _locs_=None): """Execute code in a namespace.""" if _globs_ is None: frame = sys._getframe(1) _globs_ = frame.f_globals if _locs_ is None: _locs_ = frame.f_locals del frame elif _locs_ is None: _locs_ = _globs_ exec("""exec _code_ in _globs_, _locs_""") exec_("""def reraise(tp, value, tb=None): raise tp, value, tb """) if sys.version_info[:2] == (3, 2): exec_("""def raise_from(value, from_value): if from_value is None: raise value raise value from from_value """) elif sys.version_info[:2] > (3, 2): exec_("""def raise_from(value, from_value): raise value from from_value """) else: def raise_from(value, from_value): raise value print_ = getattr(moves.builtins, "print", None) if print_ is None: def print_(*args, **kwargs): """The new-style print function for Python 2.4 and 2.5.""" fp = kwargs.pop("file", sys.stdout) if fp is None: return def write(data): if not isinstance(data, basestring): data = str(data) # If the file has an encoding, encode unicode with it. if (isinstance(fp, file) and isinstance(data, unicode) and fp.encoding is not None): errors = getattr(fp, "errors", None) if errors is None: errors = "strict" data = data.encode(fp.encoding, errors) fp.write(data) want_unicode = False sep = kwargs.pop("sep", None) if sep is not None: if isinstance(sep, unicode): want_unicode = True elif not isinstance(sep, str): raise TypeError("sep must be None or a string") end = kwargs.pop("end", None) if end is not None: if isinstance(end, unicode): want_unicode = True elif not isinstance(end, str): raise TypeError("end must be None or a string") if kwargs: raise TypeError("invalid keyword arguments to print()") if not want_unicode: for arg in args: if isinstance(arg, unicode): want_unicode = True break if want_unicode: newline = unicode("\n") space = unicode(" ") else: newline = "\n" space = " " if sep is None: sep = space if end is None: end = newline for i, arg in enumerate(args): if i: write(sep) write(arg) write(end) if sys.version_info[:2] < (3, 3): _print = print_ def print_(*args, **kwargs): fp = kwargs.get("file", sys.stdout) flush = kwargs.pop("flush", False) _print(*args, **kwargs) if flush and fp is not None: fp.flush() _add_doc(reraise, """Reraise an exception.""") if sys.version_info[0:2] < (3, 4): def wraps(wrapped, assigned=functools.WRAPPER_ASSIGNMENTS, updated=functools.WRAPPER_UPDATES): def wrapper(f): f = functools.wraps(wrapped, assigned, updated)(f) f.__wrapped__ = wrapped return f return wrapper else: wraps = functools.wraps def with_metaclass(meta, *bases): """Create a base class with a metaclass.""" # This requires a bit of explanation: the basic idea is to make a dummy # metaclass for one level of class instantiation that replaces itself with # the actual metaclass. class metaclass(meta): def __new__(cls, name, this_bases, d): return meta(name, bases, d) return type.__new__(metaclass, 'temporary_class', (), {}) def add_metaclass(metaclass): """Class decorator for creating a class with a metaclass.""" def wrapper(cls): orig_vars = cls.__dict__.copy() slots = orig_vars.get('__slots__') if slots is not None: if isinstance(slots, str): slots = [slots] for slots_var in slots: orig_vars.pop(slots_var) orig_vars.pop('__dict__', None) orig_vars.pop('__weakref__', None) return metaclass(cls.__name__, cls.__bases__, orig_vars) return wrapper def python_2_unicode_compatible(klass): """ A decorator that defines __unicode__ and __str__ methods under Python 2. Under Python 3 it does nothing. To support Python 2 and 3 with a single code base, define a __str__ method returning text and apply this decorator to the class. """ if PY2: if '__str__' not in klass.__dict__: raise ValueError("@python_2_unicode_compatible cannot be applied " "to %s because it doesn't define __str__()." % klass.__name__) klass.__unicode__ = klass.__str__ klass.__str__ = lambda self: self.__unicode__().encode('utf-8') return klass # Complete the moves implementation. # This code is at the end of this module to speed up module loading. # Turn this module into a package. __path__ = [] # required for PEP 302 and PEP 451 __package__ = __name__ # see PEP 366 @ReservedAssignment if globals().get("__spec__") is not None: __spec__.submodule_search_locations = [] # PEP 451 @UndefinedVariable # Remove other six meta path importers, since they cause problems. This can # happen if six is removed from sys.modules and then reloaded. (Setuptools does # this for some reason.) if sys.meta_path: for i, importer in enumerate(sys.meta_path): # Here's some real nastiness: Another "instance" of the six module might # be floating around. Therefore, we can't use isinstance() to check for # the six meta path importer, since the other six instance will have # inserted an importer with different class. if (type(importer).__name__ == "_SixMetaPathImporter" and importer.name == __name__): del sys.meta_path[i] break del i, importer # Finally, add the importer to the meta path import hook. sys.meta_path.append(_importer) site-packages/setuptools/_vendor/pyparsing.py000064400000705167151464452710015560 0ustar00# module pyparsing.py # # Copyright (c) 2003-2018 Paul T. McGuire # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be # included in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, # EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF # MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. # IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY # CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, # TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE # SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # __doc__ = \ """ pyparsing module - Classes and methods to define and execute parsing grammars ============================================================================= The pyparsing module is an alternative approach to creating and executing simple grammars, vs. the traditional lex/yacc approach, or the use of regular expressions. With pyparsing, you don't need to learn a new syntax for defining grammars or matching expressions - the parsing module provides a library of classes that you use to construct the grammar directly in Python. Here is a program to parse "Hello, World!" (or any greeting of the form C{", !"}), built up using L{Word}, L{Literal}, and L{And} elements (L{'+'} operator gives L{And} expressions, strings are auto-converted to L{Literal} expressions):: from pyparsing import Word, alphas # define grammar of a greeting greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) The program outputs the following:: Hello, World! -> ['Hello', ',', 'World', '!'] The Python representation of the grammar is quite readable, owing to the self-explanatory class names, and the use of '+', '|' and '^' operators. The L{ParseResults} object returned from L{ParserElement.parseString} can be accessed as a nested list, a dictionary, or an object with named attributes. The pyparsing module handles some of the problems that are typically vexing when writing text parsers: - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello , World !", etc.) - quoted strings - embedded comments Getting Started - ----------------- Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing classes inherit from. Use the docstrings for examples of how to: - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes - construct character word-group expressions using the L{Word} class - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes - use L{'+'}, L{'|'}, L{'^'}, and L{'&'} operators to combine simple expressions into more complex ones - associate names with your parsed results using L{ParserElement.setResultsName} - find some helpful expression short-cuts like L{delimitedList} and L{oneOf} - find more useful common expressions in the L{pyparsing_common} namespace class """ __version__ = "2.2.1" __versionTime__ = "18 Sep 2018 00:49 UTC" __author__ = "Paul McGuire " import string from weakref import ref as wkref import copy import sys import warnings import re import sre_constants import collections import pprint import traceback import types from datetime import datetime try: from _thread import RLock except ImportError: from threading import RLock try: # Python 3 from collections.abc import Iterable from collections.abc import MutableMapping except ImportError: # Python 2.7 from collections import Iterable from collections import MutableMapping try: from collections import OrderedDict as _OrderedDict except ImportError: try: from ordereddict import OrderedDict as _OrderedDict except ImportError: _OrderedDict = None #~ sys.stderr.write( "testing pyparsing module, version %s, %s\n" % (__version__,__versionTime__ ) ) __all__ = [ 'And', 'CaselessKeyword', 'CaselessLiteral', 'CharsNotIn', 'Combine', 'Dict', 'Each', 'Empty', 'FollowedBy', 'Forward', 'GoToColumn', 'Group', 'Keyword', 'LineEnd', 'LineStart', 'Literal', 'MatchFirst', 'NoMatch', 'NotAny', 'OneOrMore', 'OnlyOnce', 'Optional', 'Or', 'ParseBaseException', 'ParseElementEnhance', 'ParseException', 'ParseExpression', 'ParseFatalException', 'ParseResults', 'ParseSyntaxException', 'ParserElement', 'QuotedString', 'RecursiveGrammarException', 'Regex', 'SkipTo', 'StringEnd', 'StringStart', 'Suppress', 'Token', 'TokenConverter', 'White', 'Word', 'WordEnd', 'WordStart', 'ZeroOrMore', 'alphanums', 'alphas', 'alphas8bit', 'anyCloseTag', 'anyOpenTag', 'cStyleComment', 'col', 'commaSeparatedList', 'commonHTMLEntity', 'countedArray', 'cppStyleComment', 'dblQuotedString', 'dblSlashComment', 'delimitedList', 'dictOf', 'downcaseTokens', 'empty', 'hexnums', 'htmlComment', 'javaStyleComment', 'line', 'lineEnd', 'lineStart', 'lineno', 'makeHTMLTags', 'makeXMLTags', 'matchOnlyAtCol', 'matchPreviousExpr', 'matchPreviousLiteral', 'nestedExpr', 'nullDebugAction', 'nums', 'oneOf', 'opAssoc', 'operatorPrecedence', 'printables', 'punc8bit', 'pythonStyleComment', 'quotedString', 'removeQuotes', 'replaceHTMLEntity', 'replaceWith', 'restOfLine', 'sglQuotedString', 'srange', 'stringEnd', 'stringStart', 'traceParseAction', 'unicodeString', 'upcaseTokens', 'withAttribute', 'indentedBlock', 'originalTextFor', 'ungroup', 'infixNotation','locatedExpr', 'withClass', 'CloseMatch', 'tokenMap', 'pyparsing_common', ] system_version = tuple(sys.version_info)[:3] PY_3 = system_version[0] == 3 if PY_3: _MAX_INT = sys.maxsize basestring = str unichr = chr _ustr = str # build list of single arg builtins, that can be used as parse actions singleArgBuiltins = [sum, len, sorted, reversed, list, tuple, set, any, all, min, max] else: _MAX_INT = sys.maxint range = xrange def _ustr(obj): """Drop-in replacement for str(obj) that tries to be Unicode friendly. It first tries str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It then < returns the unicode object | encodes it with the default encoding | ... >. """ if isinstance(obj,unicode): return obj try: # If this works, then _ustr(obj) has the same behaviour as str(obj), so # it won't break any existing code. return str(obj) except UnicodeEncodeError: # Else encode it ret = unicode(obj).encode(sys.getdefaultencoding(), 'xmlcharrefreplace') xmlcharref = Regex(r'&#\d+;') xmlcharref.setParseAction(lambda t: '\\u' + hex(int(t[0][2:-1]))[2:]) return xmlcharref.transformString(ret) # build list of single arg builtins, tolerant of Python version, that can be used as parse actions singleArgBuiltins = [] import __builtin__ for fname in "sum len sorted reversed list tuple set any all min max".split(): try: singleArgBuiltins.append(getattr(__builtin__,fname)) except AttributeError: continue _generatorType = type((y for y in range(1))) def _xml_escape(data): """Escape &, <, >, ", ', etc. in a string of data.""" # ampersand must be replaced first from_symbols = '&><"\'' to_symbols = ('&'+s+';' for s in "amp gt lt quot apos".split()) for from_,to_ in zip(from_symbols, to_symbols): data = data.replace(from_, to_) return data class _Constants(object): pass alphas = string.ascii_uppercase + string.ascii_lowercase nums = "0123456789" hexnums = nums + "ABCDEFabcdef" alphanums = alphas + nums _bslash = chr(92) printables = "".join(c for c in string.printable if c not in string.whitespace) class ParseBaseException(Exception): """base exception class for all parsing runtime exceptions""" # Performance tuning: we construct a *lot* of these, so keep this # constructor as small and fast as possible def __init__( self, pstr, loc=0, msg=None, elem=None ): self.loc = loc if msg is None: self.msg = pstr self.pstr = "" else: self.msg = msg self.pstr = pstr self.parserElement = elem self.args = (pstr, loc, msg) @classmethod def _from_exception(cls, pe): """ internal factory method to simplify creating one type of ParseException from another - avoids having __init__ signature conflicts among subclasses """ return cls(pe.pstr, pe.loc, pe.msg, pe.parserElement) def __getattr__( self, aname ): """supported attributes by name are: - lineno - returns the line number of the exception text - col - returns the column number of the exception text - line - returns the line containing the exception text """ if( aname == "lineno" ): return lineno( self.loc, self.pstr ) elif( aname in ("col", "column") ): return col( self.loc, self.pstr ) elif( aname == "line" ): return line( self.loc, self.pstr ) else: raise AttributeError(aname) def __str__( self ): return "%s (at char %d), (line:%d, col:%d)" % \ ( self.msg, self.loc, self.lineno, self.column ) def __repr__( self ): return _ustr(self) def markInputline( self, markerString = ">!<" ): """Extracts the exception line from the input string, and marks the location of the exception with a special symbol. """ line_str = self.line line_column = self.column - 1 if markerString: line_str = "".join((line_str[:line_column], markerString, line_str[line_column:])) return line_str.strip() def __dir__(self): return "lineno col line".split() + dir(type(self)) class ParseException(ParseBaseException): """ Exception thrown when parse expressions don't match class; supported attributes by name are: - lineno - returns the line number of the exception text - col - returns the column number of the exception text - line - returns the line containing the exception text Example:: try: Word(nums).setName("integer").parseString("ABC") except ParseException as pe: print(pe) print("column: {}".format(pe.col)) prints:: Expected integer (at char 0), (line:1, col:1) column: 1 """ pass class ParseFatalException(ParseBaseException): """user-throwable exception thrown when inconsistent parse content is found; stops all parsing immediately""" pass class ParseSyntaxException(ParseFatalException): """just like L{ParseFatalException}, but thrown internally when an L{ErrorStop} ('-' operator) indicates that parsing is to stop immediately because an unbacktrackable syntax error has been found""" pass #~ class ReparseException(ParseBaseException): #~ """Experimental class - parse actions can raise this exception to cause #~ pyparsing to reparse the input string: #~ - with a modified input string, and/or #~ - with a modified start location #~ Set the values of the ReparseException in the constructor, and raise the #~ exception in a parse action to cause pyparsing to use the new string/location. #~ Setting the values as None causes no change to be made. #~ """ #~ def __init_( self, newstring, restartLoc ): #~ self.newParseText = newstring #~ self.reparseLoc = restartLoc class RecursiveGrammarException(Exception): """exception thrown by L{ParserElement.validate} if the grammar could be improperly recursive""" def __init__( self, parseElementList ): self.parseElementTrace = parseElementList def __str__( self ): return "RecursiveGrammarException: %s" % self.parseElementTrace class _ParseResultsWithOffset(object): def __init__(self,p1,p2): self.tup = (p1,p2) def __getitem__(self,i): return self.tup[i] def __repr__(self): return repr(self.tup[0]) def setOffset(self,i): self.tup = (self.tup[0],i) class ParseResults(object): """ Structured parse results, to provide multiple means of access to the parsed data: - as a list (C{len(results)}) - by list index (C{results[0], results[1]}, etc.) - by attribute (C{results.} - see L{ParserElement.setResultsName}) Example:: integer = Word(nums) date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: # date_str = integer("year") + '/' + integer("month") + '/' + integer("day") # parseString returns a ParseResults object result = date_str.parseString("1999/12/31") def test(s, fn=repr): print("%s -> %s" % (s, fn(eval(s)))) test("list(result)") test("result[0]") test("result['month']") test("result.day") test("'month' in result") test("'minutes' in result") test("result.dump()", str) prints:: list(result) -> ['1999', '/', '12', '/', '31'] result[0] -> '1999' result['month'] -> '12' result.day -> '31' 'month' in result -> True 'minutes' in result -> False result.dump() -> ['1999', '/', '12', '/', '31'] - day: 31 - month: 12 - year: 1999 """ def __new__(cls, toklist=None, name=None, asList=True, modal=True ): if isinstance(toklist, cls): return toklist retobj = object.__new__(cls) retobj.__doinit = True return retobj # Performance tuning: we construct a *lot* of these, so keep this # constructor as small and fast as possible def __init__( self, toklist=None, name=None, asList=True, modal=True, isinstance=isinstance ): if self.__doinit: self.__doinit = False self.__name = None self.__parent = None self.__accumNames = {} self.__asList = asList self.__modal = modal if toklist is None: toklist = [] if isinstance(toklist, list): self.__toklist = toklist[:] elif isinstance(toklist, _generatorType): self.__toklist = list(toklist) else: self.__toklist = [toklist] self.__tokdict = dict() if name is not None and name: if not modal: self.__accumNames[name] = 0 if isinstance(name,int): name = _ustr(name) # will always return a str, but use _ustr for consistency self.__name = name if not (isinstance(toklist, (type(None), basestring, list)) and toklist in (None,'',[])): if isinstance(toklist,basestring): toklist = [ toklist ] if asList: if isinstance(toklist,ParseResults): self[name] = _ParseResultsWithOffset(toklist.copy(),0) else: self[name] = _ParseResultsWithOffset(ParseResults(toklist[0]),0) self[name].__name = name else: try: self[name] = toklist[0] except (KeyError,TypeError,IndexError): self[name] = toklist def __getitem__( self, i ): if isinstance( i, (int,slice) ): return self.__toklist[i] else: if i not in self.__accumNames: return self.__tokdict[i][-1][0] else: return ParseResults([ v[0] for v in self.__tokdict[i] ]) def __setitem__( self, k, v, isinstance=isinstance ): if isinstance(v,_ParseResultsWithOffset): self.__tokdict[k] = self.__tokdict.get(k,list()) + [v] sub = v[0] elif isinstance(k,(int,slice)): self.__toklist[k] = v sub = v else: self.__tokdict[k] = self.__tokdict.get(k,list()) + [_ParseResultsWithOffset(v,0)] sub = v if isinstance(sub,ParseResults): sub.__parent = wkref(self) def __delitem__( self, i ): if isinstance(i,(int,slice)): mylen = len( self.__toklist ) del self.__toklist[i] # convert int to slice if isinstance(i, int): if i < 0: i += mylen i = slice(i, i+1) # get removed indices removed = list(range(*i.indices(mylen))) removed.reverse() # fixup indices in token dictionary for name,occurrences in self.__tokdict.items(): for j in removed: for k, (value, position) in enumerate(occurrences): occurrences[k] = _ParseResultsWithOffset(value, position - (position > j)) else: del self.__tokdict[i] def __contains__( self, k ): return k in self.__tokdict def __len__( self ): return len( self.__toklist ) def __bool__(self): return ( not not self.__toklist ) __nonzero__ = __bool__ def __iter__( self ): return iter( self.__toklist ) def __reversed__( self ): return iter( self.__toklist[::-1] ) def _iterkeys( self ): if hasattr(self.__tokdict, "iterkeys"): return self.__tokdict.iterkeys() else: return iter(self.__tokdict) def _itervalues( self ): return (self[k] for k in self._iterkeys()) def _iteritems( self ): return ((k, self[k]) for k in self._iterkeys()) if PY_3: keys = _iterkeys """Returns an iterator of all named result keys (Python 3.x only).""" values = _itervalues """Returns an iterator of all named result values (Python 3.x only).""" items = _iteritems """Returns an iterator of all named result key-value tuples (Python 3.x only).""" else: iterkeys = _iterkeys """Returns an iterator of all named result keys (Python 2.x only).""" itervalues = _itervalues """Returns an iterator of all named result values (Python 2.x only).""" iteritems = _iteritems """Returns an iterator of all named result key-value tuples (Python 2.x only).""" def keys( self ): """Returns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).""" return list(self.iterkeys()) def values( self ): """Returns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).""" return list(self.itervalues()) def items( self ): """Returns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).""" return list(self.iteritems()) def haskeys( self ): """Since keys() returns an iterator, this method is helpful in bypassing code that looks for the existence of any defined results names.""" return bool(self.__tokdict) def pop( self, *args, **kwargs): """ Removes and returns item at specified index (default=C{last}). Supports both C{list} and C{dict} semantics for C{pop()}. If passed no argument or an integer argument, it will use C{list} semantics and pop tokens from the list of parsed tokens. If passed a non-integer argument (most likely a string), it will use C{dict} semantics and pop the corresponding value from any defined results names. A second default return value argument is supported, just as in C{dict.pop()}. Example:: def remove_first(tokens): tokens.pop(0) print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321'] label = Word(alphas) patt = label("LABEL") + OneOrMore(Word(nums)) print(patt.parseString("AAB 123 321").dump()) # Use pop() in a parse action to remove named result (note that corresponding value is not # removed from list form of results) def remove_LABEL(tokens): tokens.pop("LABEL") return tokens patt.addParseAction(remove_LABEL) print(patt.parseString("AAB 123 321").dump()) prints:: ['AAB', '123', '321'] - LABEL: AAB ['AAB', '123', '321'] """ if not args: args = [-1] for k,v in kwargs.items(): if k == 'default': args = (args[0], v) else: raise TypeError("pop() got an unexpected keyword argument '%s'" % k) if (isinstance(args[0], int) or len(args) == 1 or args[0] in self): index = args[0] ret = self[index] del self[index] return ret else: defaultvalue = args[1] return defaultvalue def get(self, key, defaultValue=None): """ Returns named result matching the given key, or if there is no such name, then returns the given C{defaultValue} or C{None} if no C{defaultValue} is specified. Similar to C{dict.get()}. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString("1999/12/31") print(result.get("year")) # -> '1999' print(result.get("hour", "not specified")) # -> 'not specified' print(result.get("hour")) # -> None """ if key in self: return self[key] else: return defaultValue def insert( self, index, insStr ): """ Inserts new element at location index in the list of parsed tokens. Similar to C{list.insert()}. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to insert the parse location in the front of the parsed results def insert_locn(locn, tokens): tokens.insert(0, locn) print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321'] """ self.__toklist.insert(index, insStr) # fixup indices in token dictionary for name,occurrences in self.__tokdict.items(): for k, (value, position) in enumerate(occurrences): occurrences[k] = _ParseResultsWithOffset(value, position + (position > index)) def append( self, item ): """ Add single element to end of ParseResults list of elements. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to compute the sum of the parsed integers, and add it to the end def append_sum(tokens): tokens.append(sum(map(int, tokens))) print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444] """ self.__toklist.append(item) def extend( self, itemseq ): """ Add sequence of elements to end of ParseResults list of elements. Example:: patt = OneOrMore(Word(alphas)) # use a parse action to append the reverse of the matched strings, to make a palindrome def make_palindrome(tokens): tokens.extend(reversed([t[::-1] for t in tokens])) return ''.join(tokens) print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl' """ if isinstance(itemseq, ParseResults): self += itemseq else: self.__toklist.extend(itemseq) def clear( self ): """ Clear all elements and results names. """ del self.__toklist[:] self.__tokdict.clear() def __getattr__( self, name ): try: return self[name] except KeyError: return "" if name in self.__tokdict: if name not in self.__accumNames: return self.__tokdict[name][-1][0] else: return ParseResults([ v[0] for v in self.__tokdict[name] ]) else: return "" def __add__( self, other ): ret = self.copy() ret += other return ret def __iadd__( self, other ): if other.__tokdict: offset = len(self.__toklist) addoffset = lambda a: offset if a<0 else a+offset otheritems = other.__tokdict.items() otherdictitems = [(k, _ParseResultsWithOffset(v[0],addoffset(v[1])) ) for (k,vlist) in otheritems for v in vlist] for k,v in otherdictitems: self[k] = v if isinstance(v[0],ParseResults): v[0].__parent = wkref(self) self.__toklist += other.__toklist self.__accumNames.update( other.__accumNames ) return self def __radd__(self, other): if isinstance(other,int) and other == 0: # useful for merging many ParseResults using sum() builtin return self.copy() else: # this may raise a TypeError - so be it return other + self def __repr__( self ): return "(%s, %s)" % ( repr( self.__toklist ), repr( self.__tokdict ) ) def __str__( self ): return '[' + ', '.join(_ustr(i) if isinstance(i, ParseResults) else repr(i) for i in self.__toklist) + ']' def _asStringList( self, sep='' ): out = [] for item in self.__toklist: if out and sep: out.append(sep) if isinstance( item, ParseResults ): out += item._asStringList() else: out.append( _ustr(item) ) return out def asList( self ): """ Returns the parse results as a nested list of matching tokens, all converted to strings. Example:: patt = OneOrMore(Word(alphas)) result = patt.parseString("sldkj lsdkj sldkj") # even though the result prints in string-like form, it is actually a pyparsing ParseResults print(type(result), result) # -> ['sldkj', 'lsdkj', 'sldkj'] # Use asList() to create an actual list result_list = result.asList() print(type(result_list), result_list) # -> ['sldkj', 'lsdkj', 'sldkj'] """ return [res.asList() if isinstance(res,ParseResults) else res for res in self.__toklist] def asDict( self ): """ Returns the named parse results as a nested dictionary. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString('12/31/1999') print(type(result), repr(result)) # -> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]}) result_dict = result.asDict() print(type(result_dict), repr(result_dict)) # -> {'day': '1999', 'year': '12', 'month': '31'} # even though a ParseResults supports dict-like access, sometime you just need to have a dict import json print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"} """ if PY_3: item_fn = self.items else: item_fn = self.iteritems def toItem(obj): if isinstance(obj, ParseResults): if obj.haskeys(): return obj.asDict() else: return [toItem(v) for v in obj] else: return obj return dict((k,toItem(v)) for k,v in item_fn()) def copy( self ): """ Returns a new copy of a C{ParseResults} object. """ ret = ParseResults( self.__toklist ) ret.__tokdict = self.__tokdict.copy() ret.__parent = self.__parent ret.__accumNames.update( self.__accumNames ) ret.__name = self.__name return ret def asXML( self, doctag=None, namedItemsOnly=False, indent="", formatted=True ): """ (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names. """ nl = "\n" out = [] namedItems = dict((v[1],k) for (k,vlist) in self.__tokdict.items() for v in vlist) nextLevelIndent = indent + " " # collapse out indents if formatting is not desired if not formatted: indent = "" nextLevelIndent = "" nl = "" selfTag = None if doctag is not None: selfTag = doctag else: if self.__name: selfTag = self.__name if not selfTag: if namedItemsOnly: return "" else: selfTag = "ITEM" out += [ nl, indent, "<", selfTag, ">" ] for i,res in enumerate(self.__toklist): if isinstance(res,ParseResults): if i in namedItems: out += [ res.asXML(namedItems[i], namedItemsOnly and doctag is None, nextLevelIndent, formatted)] else: out += [ res.asXML(None, namedItemsOnly and doctag is None, nextLevelIndent, formatted)] else: # individual token, see if there is a name for it resTag = None if i in namedItems: resTag = namedItems[i] if not resTag: if namedItemsOnly: continue else: resTag = "ITEM" xmlBodyText = _xml_escape(_ustr(res)) out += [ nl, nextLevelIndent, "<", resTag, ">", xmlBodyText, "" ] out += [ nl, indent, "" ] return "".join(out) def __lookup(self,sub): for k,vlist in self.__tokdict.items(): for v,loc in vlist: if sub is v: return k return None def getName(self): r""" Returns the results name for this token expression. Useful when several different expressions might match at a particular location. Example:: integer = Word(nums) ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d") house_number_expr = Suppress('#') + Word(nums, alphanums) user_data = (Group(house_number_expr)("house_number") | Group(ssn_expr)("ssn") | Group(integer)("age")) user_info = OneOrMore(user_data) result = user_info.parseString("22 111-22-3333 #221B") for item in result: print(item.getName(), ':', item[0]) prints:: age : 22 ssn : 111-22-3333 house_number : 221B """ if self.__name: return self.__name elif self.__parent: par = self.__parent() if par: return par.__lookup(self) else: return None elif (len(self) == 1 and len(self.__tokdict) == 1 and next(iter(self.__tokdict.values()))[0][1] in (0,-1)): return next(iter(self.__tokdict.keys())) else: return None def dump(self, indent='', depth=0, full=True): """ Diagnostic method for listing out the contents of a C{ParseResults}. Accepts an optional C{indent} argument so that this string can be embedded in a nested display of other data. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString('12/31/1999') print(result.dump()) prints:: ['12', '/', '31', '/', '1999'] - day: 1999 - month: 31 - year: 12 """ out = [] NL = '\n' out.append( indent+_ustr(self.asList()) ) if full: if self.haskeys(): items = sorted((str(k), v) for k,v in self.items()) for k,v in items: if out: out.append(NL) out.append( "%s%s- %s: " % (indent,(' '*depth), k) ) if isinstance(v,ParseResults): if v: out.append( v.dump(indent,depth+1) ) else: out.append(_ustr(v)) else: out.append(repr(v)) elif any(isinstance(vv,ParseResults) for vv in self): v = self for i,vv in enumerate(v): if isinstance(vv,ParseResults): out.append("\n%s%s[%d]:\n%s%s%s" % (indent,(' '*(depth)),i,indent,(' '*(depth+1)),vv.dump(indent,depth+1) )) else: out.append("\n%s%s[%d]:\n%s%s%s" % (indent,(' '*(depth)),i,indent,(' '*(depth+1)),_ustr(vv))) return "".join(out) def pprint(self, *args, **kwargs): """ Pretty-printer for parsed results as a list, using the C{pprint} module. Accepts additional positional or keyword args as defined for the C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint}) Example:: ident = Word(alphas, alphanums) num = Word(nums) func = Forward() term = ident | num | Group('(' + func + ')') func <<= ident + Group(Optional(delimitedList(term))) result = func.parseString("fna a,b,(fnb c,d,200),100") result.pprint(width=40) prints:: ['fna', ['a', 'b', ['(', 'fnb', ['c', 'd', '200'], ')'], '100']] """ pprint.pprint(self.asList(), *args, **kwargs) # add support for pickle protocol def __getstate__(self): return ( self.__toklist, ( self.__tokdict.copy(), self.__parent is not None and self.__parent() or None, self.__accumNames, self.__name ) ) def __setstate__(self,state): self.__toklist = state[0] (self.__tokdict, par, inAccumNames, self.__name) = state[1] self.__accumNames = {} self.__accumNames.update(inAccumNames) if par is not None: self.__parent = wkref(par) else: self.__parent = None def __getnewargs__(self): return self.__toklist, self.__name, self.__asList, self.__modal def __dir__(self): return (dir(type(self)) + list(self.keys())) MutableMapping.register(ParseResults) def col (loc,strg): """Returns current column within a string, counting newlines as line separators. The first column is number 1. Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See L{I{ParserElement.parseString}} for more information on parsing strings containing C{}s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. """ s = strg return 1 if 0} for more information on parsing strings containing C{}s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. """ return strg.count("\n",0,loc) + 1 def line( loc, strg ): """Returns the line of text containing loc within a string, counting newlines as line separators. """ lastCR = strg.rfind("\n", 0, loc) nextCR = strg.find("\n", loc) if nextCR >= 0: return strg[lastCR+1:nextCR] else: return strg[lastCR+1:] def _defaultStartDebugAction( instring, loc, expr ): print (("Match " + _ustr(expr) + " at loc " + _ustr(loc) + "(%d,%d)" % ( lineno(loc,instring), col(loc,instring) ))) def _defaultSuccessDebugAction( instring, startloc, endloc, expr, toks ): print ("Matched " + _ustr(expr) + " -> " + str(toks.asList())) def _defaultExceptionDebugAction( instring, loc, expr, exc ): print ("Exception raised:" + _ustr(exc)) def nullDebugAction(*args): """'Do-nothing' debug action, to suppress debugging output during parsing.""" pass # Only works on Python 3.x - nonlocal is toxic to Python 2 installs #~ 'decorator to trim function calls to match the arity of the target' #~ def _trim_arity(func, maxargs=3): #~ if func in singleArgBuiltins: #~ return lambda s,l,t: func(t) #~ limit = 0 #~ foundArity = False #~ def wrapper(*args): #~ nonlocal limit,foundArity #~ while 1: #~ try: #~ ret = func(*args[limit:]) #~ foundArity = True #~ return ret #~ except TypeError: #~ if limit == maxargs or foundArity: #~ raise #~ limit += 1 #~ continue #~ return wrapper # this version is Python 2.x-3.x cross-compatible 'decorator to trim function calls to match the arity of the target' def _trim_arity(func, maxargs=2): if func in singleArgBuiltins: return lambda s,l,t: func(t) limit = [0] foundArity = [False] # traceback return data structure changed in Py3.5 - normalize back to plain tuples if system_version[:2] >= (3,5): def extract_stack(limit=0): # special handling for Python 3.5.0 - extra deep call stack by 1 offset = -3 if system_version == (3,5,0) else -2 frame_summary = traceback.extract_stack(limit=-offset+limit-1)[offset] return [frame_summary[:2]] def extract_tb(tb, limit=0): frames = traceback.extract_tb(tb, limit=limit) frame_summary = frames[-1] return [frame_summary[:2]] else: extract_stack = traceback.extract_stack extract_tb = traceback.extract_tb # synthesize what would be returned by traceback.extract_stack at the call to # user's parse action 'func', so that we don't incur call penalty at parse time LINE_DIFF = 6 # IF ANY CODE CHANGES, EVEN JUST COMMENTS OR BLANK LINES, BETWEEN THE NEXT LINE AND # THE CALL TO FUNC INSIDE WRAPPER, LINE_DIFF MUST BE MODIFIED!!!! this_line = extract_stack(limit=2)[-1] pa_call_line_synth = (this_line[0], this_line[1]+LINE_DIFF) def wrapper(*args): while 1: try: ret = func(*args[limit[0]:]) foundArity[0] = True return ret except TypeError: # re-raise TypeErrors if they did not come from our arity testing if foundArity[0]: raise else: try: tb = sys.exc_info()[-1] if not extract_tb(tb, limit=2)[-1][:2] == pa_call_line_synth: raise finally: del tb if limit[0] <= maxargs: limit[0] += 1 continue raise # copy func name to wrapper for sensible debug output func_name = "" try: func_name = getattr(func, '__name__', getattr(func, '__class__').__name__) except Exception: func_name = str(func) wrapper.__name__ = func_name return wrapper class ParserElement(object): """Abstract base level parser element class.""" DEFAULT_WHITE_CHARS = " \n\t\r" verbose_stacktrace = False @staticmethod def setDefaultWhitespaceChars( chars ): r""" Overrides the default whitespace chars Example:: # default whitespace chars are space, and newline OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def', 'ghi', 'jkl'] # change to just treat newline as significant ParserElement.setDefaultWhitespaceChars(" \t") OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def'] """ ParserElement.DEFAULT_WHITE_CHARS = chars @staticmethod def inlineLiteralsUsing(cls): """ Set class to be used for inclusion of string literals into a parser. Example:: # default literal class used is Literal integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # change to Suppress ParserElement.inlineLiteralsUsing(Suppress) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '12', '31'] """ ParserElement._literalStringClass = cls def __init__( self, savelist=False ): self.parseAction = list() self.failAction = None #~ self.name = "" # don't define self.name, let subclasses try/except upcall self.strRepr = None self.resultsName = None self.saveAsList = savelist self.skipWhitespace = True self.whiteChars = ParserElement.DEFAULT_WHITE_CHARS self.copyDefaultWhiteChars = True self.mayReturnEmpty = False # used when checking for left-recursion self.keepTabs = False self.ignoreExprs = list() self.debug = False self.streamlined = False self.mayIndexError = True # used to optimize exception handling for subclasses that don't advance parse index self.errmsg = "" self.modalResults = True # used to mark results names as modal (report only last) or cumulative (list all) self.debugActions = ( None, None, None ) #custom debug actions self.re = None self.callPreparse = True # used to avoid redundant calls to preParse self.callDuringTry = False def copy( self ): """ Make a copy of this C{ParserElement}. Useful for defining different parse actions for the same parsing pattern, using copies of the original parse element. Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K") integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M") print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M")) prints:: [5120, 100, 655360, 268435456] Equivalent form of C{expr.copy()} is just C{expr()}:: integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M") """ cpy = copy.copy( self ) cpy.parseAction = self.parseAction[:] cpy.ignoreExprs = self.ignoreExprs[:] if self.copyDefaultWhiteChars: cpy.whiteChars = ParserElement.DEFAULT_WHITE_CHARS return cpy def setName( self, name ): """ Define name for this expression, makes debugging and exception messages clearer. Example:: Word(nums).parseString("ABC") # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1) Word(nums).setName("integer").parseString("ABC") # -> Exception: Expected integer (at char 0), (line:1, col:1) """ self.name = name self.errmsg = "Expected " + self.name if hasattr(self,"exception"): self.exception.msg = self.errmsg return self def setResultsName( self, name, listAllMatches=False ): """ Define name for referencing matching tokens as a nested attribute of the returned parse results. NOTE: this returns a *copy* of the original C{ParserElement} object; this is so that the client can define a basic element, such as an integer, and reference it in multiple places with different names. You can also set results names using the abbreviated syntax, C{expr("name")} in place of C{expr.setResultsName("name")} - see L{I{__call__}<__call__>}. Example:: date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: date_str = integer("year") + '/' + integer("month") + '/' + integer("day") """ newself = self.copy() if name.endswith("*"): name = name[:-1] listAllMatches=True newself.resultsName = name newself.modalResults = not listAllMatches return newself def setBreak(self,breakFlag = True): """Method to invoke the Python pdb debugger when this element is about to be parsed. Set C{breakFlag} to True to enable, False to disable. """ if breakFlag: _parseMethod = self._parse def breaker(instring, loc, doActions=True, callPreParse=True): import pdb pdb.set_trace() return _parseMethod( instring, loc, doActions, callPreParse ) breaker._originalParseMethod = _parseMethod self._parse = breaker else: if hasattr(self._parse,"_originalParseMethod"): self._parse = self._parse._originalParseMethod return self def setParseAction( self, *fns, **kwargs ): """ Define one or more actions to perform when successfully matching parse element definition. Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)}, C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where: - s = the original string being parsed (see note below) - loc = the location of the matching substring - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object If the functions in fns modify the tokens, they can return them as the return value from fn, and the modified list of tokens will replace the original. Otherwise, fn does not need to return any value. Optional keyword arguments: - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See L{I{parseString}} for more information on parsing strings containing C{}s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. Example:: integer = Word(nums) date_str = integer + '/' + integer + '/' + integer date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # use parse action to convert to ints at parse time integer = Word(nums).setParseAction(lambda toks: int(toks[0])) date_str = integer + '/' + integer + '/' + integer # note that integer fields are now ints, not strings date_str.parseString("1999/12/31") # -> [1999, '/', 12, '/', 31] """ self.parseAction = list(map(_trim_arity, list(fns))) self.callDuringTry = kwargs.get("callDuringTry", False) return self def addParseAction( self, *fns, **kwargs ): """ Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}}. See examples in L{I{copy}}. """ self.parseAction += list(map(_trim_arity, list(fns))) self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False) return self def addCondition(self, *fns, **kwargs): """Add a boolean predicate function to expression's list of parse actions. See L{I{setParseAction}} for function call signatures. Unlike C{setParseAction}, functions passed to C{addCondition} need to return boolean success/fail of the condition. Optional keyword arguments: - message = define a custom message to be used in the raised exception - fatal = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) year_int = integer.copy() year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later") date_str = year_int + '/' + integer + '/' + integer result = date_str.parseString("1999/12/31") # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1) """ msg = kwargs.get("message", "failed user-defined condition") exc_type = ParseFatalException if kwargs.get("fatal", False) else ParseException for fn in fns: def pa(s,l,t): if not bool(_trim_arity(fn)(s,l,t)): raise exc_type(s,l,msg) self.parseAction.append(pa) self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False) return self def setFailAction( self, fn ): """Define action to perform if parsing fails at this expression. Fail acton fn is a callable function that takes the arguments C{fn(s,loc,expr,err)} where: - s = string being parsed - loc = location where expression match was attempted and failed - expr = the parse expression that failed - err = the exception thrown The function returns no value. It may throw C{L{ParseFatalException}} if it is desired to stop parsing immediately.""" self.failAction = fn return self def _skipIgnorables( self, instring, loc ): exprsFound = True while exprsFound: exprsFound = False for e in self.ignoreExprs: try: while 1: loc,dummy = e._parse( instring, loc ) exprsFound = True except ParseException: pass return loc def preParse( self, instring, loc ): if self.ignoreExprs: loc = self._skipIgnorables( instring, loc ) if self.skipWhitespace: wt = self.whiteChars instrlen = len(instring) while loc < instrlen and instring[loc] in wt: loc += 1 return loc def parseImpl( self, instring, loc, doActions=True ): return loc, [] def postParse( self, instring, loc, tokenlist ): return tokenlist #~ @profile def _parseNoCache( self, instring, loc, doActions=True, callPreParse=True ): debugging = ( self.debug ) #and doActions ) if debugging or self.failAction: #~ print ("Match",self,"at loc",loc,"(%d,%d)" % ( lineno(loc,instring), col(loc,instring) )) if (self.debugActions[0] ): self.debugActions[0]( instring, loc, self ) if callPreParse and self.callPreparse: preloc = self.preParse( instring, loc ) else: preloc = loc tokensStart = preloc try: try: loc,tokens = self.parseImpl( instring, preloc, doActions ) except IndexError: raise ParseException( instring, len(instring), self.errmsg, self ) except ParseBaseException as err: #~ print ("Exception raised:", err) if self.debugActions[2]: self.debugActions[2]( instring, tokensStart, self, err ) if self.failAction: self.failAction( instring, tokensStart, self, err ) raise else: if callPreParse and self.callPreparse: preloc = self.preParse( instring, loc ) else: preloc = loc tokensStart = preloc if self.mayIndexError or preloc >= len(instring): try: loc,tokens = self.parseImpl( instring, preloc, doActions ) except IndexError: raise ParseException( instring, len(instring), self.errmsg, self ) else: loc,tokens = self.parseImpl( instring, preloc, doActions ) tokens = self.postParse( instring, loc, tokens ) retTokens = ParseResults( tokens, self.resultsName, asList=self.saveAsList, modal=self.modalResults ) if self.parseAction and (doActions or self.callDuringTry): if debugging: try: for fn in self.parseAction: tokens = fn( instring, tokensStart, retTokens ) if tokens is not None: retTokens = ParseResults( tokens, self.resultsName, asList=self.saveAsList and isinstance(tokens,(ParseResults,list)), modal=self.modalResults ) except ParseBaseException as err: #~ print "Exception raised in user parse action:", err if (self.debugActions[2] ): self.debugActions[2]( instring, tokensStart, self, err ) raise else: for fn in self.parseAction: tokens = fn( instring, tokensStart, retTokens ) if tokens is not None: retTokens = ParseResults( tokens, self.resultsName, asList=self.saveAsList and isinstance(tokens,(ParseResults,list)), modal=self.modalResults ) if debugging: #~ print ("Matched",self,"->",retTokens.asList()) if (self.debugActions[1] ): self.debugActions[1]( instring, tokensStart, loc, self, retTokens ) return loc, retTokens def tryParse( self, instring, loc ): try: return self._parse( instring, loc, doActions=False )[0] except ParseFatalException: raise ParseException( instring, loc, self.errmsg, self) def canParseNext(self, instring, loc): try: self.tryParse(instring, loc) except (ParseException, IndexError): return False else: return True class _UnboundedCache(object): def __init__(self): cache = {} self.not_in_cache = not_in_cache = object() def get(self, key): return cache.get(key, not_in_cache) def set(self, key, value): cache[key] = value def clear(self): cache.clear() def cache_len(self): return len(cache) self.get = types.MethodType(get, self) self.set = types.MethodType(set, self) self.clear = types.MethodType(clear, self) self.__len__ = types.MethodType(cache_len, self) if _OrderedDict is not None: class _FifoCache(object): def __init__(self, size): self.not_in_cache = not_in_cache = object() cache = _OrderedDict() def get(self, key): return cache.get(key, not_in_cache) def set(self, key, value): cache[key] = value while len(cache) > size: try: cache.popitem(False) except KeyError: pass def clear(self): cache.clear() def cache_len(self): return len(cache) self.get = types.MethodType(get, self) self.set = types.MethodType(set, self) self.clear = types.MethodType(clear, self) self.__len__ = types.MethodType(cache_len, self) else: class _FifoCache(object): def __init__(self, size): self.not_in_cache = not_in_cache = object() cache = {} key_fifo = collections.deque([], size) def get(self, key): return cache.get(key, not_in_cache) def set(self, key, value): cache[key] = value while len(key_fifo) > size: cache.pop(key_fifo.popleft(), None) key_fifo.append(key) def clear(self): cache.clear() key_fifo.clear() def cache_len(self): return len(cache) self.get = types.MethodType(get, self) self.set = types.MethodType(set, self) self.clear = types.MethodType(clear, self) self.__len__ = types.MethodType(cache_len, self) # argument cache for optimizing repeated calls when backtracking through recursive expressions packrat_cache = {} # this is set later by enabledPackrat(); this is here so that resetCache() doesn't fail packrat_cache_lock = RLock() packrat_cache_stats = [0, 0] # this method gets repeatedly called during backtracking with the same arguments - # we can cache these arguments and save ourselves the trouble of re-parsing the contained expression def _parseCache( self, instring, loc, doActions=True, callPreParse=True ): HIT, MISS = 0, 1 lookup = (self, instring, loc, callPreParse, doActions) with ParserElement.packrat_cache_lock: cache = ParserElement.packrat_cache value = cache.get(lookup) if value is cache.not_in_cache: ParserElement.packrat_cache_stats[MISS] += 1 try: value = self._parseNoCache(instring, loc, doActions, callPreParse) except ParseBaseException as pe: # cache a copy of the exception, without the traceback cache.set(lookup, pe.__class__(*pe.args)) raise else: cache.set(lookup, (value[0], value[1].copy())) return value else: ParserElement.packrat_cache_stats[HIT] += 1 if isinstance(value, Exception): raise value return (value[0], value[1].copy()) _parse = _parseNoCache @staticmethod def resetCache(): ParserElement.packrat_cache.clear() ParserElement.packrat_cache_stats[:] = [0] * len(ParserElement.packrat_cache_stats) _packratEnabled = False @staticmethod def enablePackrat(cache_size_limit=128): """Enables "packrat" parsing, which adds memoizing to the parsing logic. Repeated parse attempts at the same string location (which happens often in many complex grammars) can immediately return a cached value, instead of re-executing parsing/validating code. Memoizing is done of both valid results and parsing exceptions. Parameters: - cache_size_limit - (default=C{128}) - if an integer value is provided will limit the size of the packrat cache; if None is passed, then the cache size will be unbounded; if 0 is passed, the cache will be effectively disabled. This speedup may break existing programs that use parse actions that have side-effects. For this reason, packrat parsing is disabled when you first import pyparsing. To activate the packrat feature, your program must call the class method C{ParserElement.enablePackrat()}. If your program uses C{psyco} to "compile as you go", you must call C{enablePackrat} before calling C{psyco.full()}. If you do not do this, Python will crash. For best results, call C{enablePackrat()} immediately after importing pyparsing. Example:: import pyparsing pyparsing.ParserElement.enablePackrat() """ if not ParserElement._packratEnabled: ParserElement._packratEnabled = True if cache_size_limit is None: ParserElement.packrat_cache = ParserElement._UnboundedCache() else: ParserElement.packrat_cache = ParserElement._FifoCache(cache_size_limit) ParserElement._parse = ParserElement._parseCache def parseString( self, instring, parseAll=False ): """ Execute the parse expression with the given string. This is the main interface to the client code, once the complete expression has been built. If you want the grammar to require that the entire input string be successfully parsed, then set C{parseAll} to True (equivalent to ending the grammar with C{L{StringEnd()}}). Note: C{parseString} implicitly calls C{expandtabs()} on the input string, in order to report proper column numbers in parse actions. If the input string contains tabs and the grammar uses parse actions that use the C{loc} argument to index into the string being parsed, you can ensure you have a consistent view of the input string by: - calling C{parseWithTabs} on your grammar before calling C{parseString} (see L{I{parseWithTabs}}) - define your parse action using the full C{(s,loc,toks)} signature, and reference the input string using the parse action's C{s} argument - explictly expand the tabs in your input string before calling C{parseString} Example:: Word('a').parseString('aaaaabaaa') # -> ['aaaaa'] Word('a').parseString('aaaaabaaa', parseAll=True) # -> Exception: Expected end of text """ ParserElement.resetCache() if not self.streamlined: self.streamline() #~ self.saveAsList = True for e in self.ignoreExprs: e.streamline() if not self.keepTabs: instring = instring.expandtabs() try: loc, tokens = self._parse( instring, 0 ) if parseAll: loc = self.preParse( instring, loc ) se = Empty() + StringEnd() se._parse( instring, loc ) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc else: return tokens def scanString( self, instring, maxMatches=_MAX_INT, overlap=False ): """ Scan the input string for expression matches. Each match will return the matching tokens, start location, and end location. May be called with optional C{maxMatches} argument, to clip scanning after 'n' matches are found. If C{overlap} is specified, then overlapping matches will be reported. Note that the start and end locations are reported relative to the string being parsed. See L{I{parseString}} for more information on parsing strings with embedded tabs. Example:: source = "sldjf123lsdjjkf345sldkjf879lkjsfd987" print(source) for tokens,start,end in Word(alphas).scanString(source): print(' '*start + '^'*(end-start)) print(' '*start + tokens[0]) prints:: sldjf123lsdjjkf345sldkjf879lkjsfd987 ^^^^^ sldjf ^^^^^^^ lsdjjkf ^^^^^^ sldkjf ^^^^^^ lkjsfd """ if not self.streamlined: self.streamline() for e in self.ignoreExprs: e.streamline() if not self.keepTabs: instring = _ustr(instring).expandtabs() instrlen = len(instring) loc = 0 preparseFn = self.preParse parseFn = self._parse ParserElement.resetCache() matches = 0 try: while loc <= instrlen and matches < maxMatches: try: preloc = preparseFn( instring, loc ) nextLoc,tokens = parseFn( instring, preloc, callPreParse=False ) except ParseException: loc = preloc+1 else: if nextLoc > loc: matches += 1 yield tokens, preloc, nextLoc if overlap: nextloc = preparseFn( instring, loc ) if nextloc > loc: loc = nextLoc else: loc += 1 else: loc = nextLoc else: loc = preloc+1 except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def transformString( self, instring ): """ Extension to C{L{scanString}}, to modify matching text with modified tokens that may be returned from a parse action. To use C{transformString}, define a grammar and attach a parse action to it that modifies the returned token list. Invoking C{transformString()} on a target string will then scan for matches, and replace the matched text patterns according to the logic in the parse action. C{transformString()} returns the resulting transformed string. Example:: wd = Word(alphas) wd.setParseAction(lambda toks: toks[0].title()) print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york.")) Prints:: Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York. """ out = [] lastE = 0 # force preservation of s, to minimize unwanted transformation of string, and to # keep string locs straight between transformString and scanString self.keepTabs = True try: for t,s,e in self.scanString( instring ): out.append( instring[lastE:s] ) if t: if isinstance(t,ParseResults): out += t.asList() elif isinstance(t,list): out += t else: out.append(t) lastE = e out.append(instring[lastE:]) out = [o for o in out if o] return "".join(map(_ustr,_flatten(out))) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def searchString( self, instring, maxMatches=_MAX_INT ): """ Another extension to C{L{scanString}}, simplifying the access to the tokens found to match the given parse expression. May be called with optional C{maxMatches} argument, to clip searching after 'n' matches are found. Example:: # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters cap_word = Word(alphas.upper(), alphas.lower()) print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")) # the sum() builtin can be used to merge results into a single ParseResults object print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))) prints:: [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']] ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity'] """ try: return ParseResults([ t for t,s,e in self.scanString( instring, maxMatches ) ]) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def split(self, instring, maxsplit=_MAX_INT, includeSeparators=False): """ Generator method to split a string using the given expression as a separator. May be called with optional C{maxsplit} argument, to limit the number of splits; and the optional C{includeSeparators} argument (default=C{False}), if the separating matching text should be included in the split results. Example:: punc = oneOf(list(".,;:/-!?")) print(list(punc.split("This, this?, this sentence, is badly punctuated!"))) prints:: ['This', ' this', '', ' this sentence', ' is badly punctuated', ''] """ splits = 0 last = 0 for t,s,e in self.scanString(instring, maxMatches=maxsplit): yield instring[last:s] if includeSeparators: yield t[0] last = e yield instring[last:] def __add__(self, other ): """ Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement converts them to L{Literal}s by default. Example:: greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) Prints:: Hello, World! -> ['Hello', ',', 'World', '!'] """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return And( [ self, other ] ) def __radd__(self, other ): """ Implementation of + operator when left operand is not a C{L{ParserElement}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other + self def __sub__(self, other): """ Implementation of - operator, returns C{L{And}} with error stop """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return self + And._ErrorStop() + other def __rsub__(self, other ): """ Implementation of - operator when left operand is not a C{L{ParserElement}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other - self def __mul__(self,other): """ Implementation of * operator, allows use of C{expr * 3} in place of C{expr + expr + expr}. Expressions may also me multiplied by a 2-integer tuple, similar to C{{min,max}} multipliers in regular expressions. Tuples may also include C{None} as in: - C{expr*(n,None)} or C{expr*(n,)} is equivalent to C{expr*n + L{ZeroOrMore}(expr)} (read as "at least n instances of C{expr}") - C{expr*(None,n)} is equivalent to C{expr*(0,n)} (read as "0 to n instances of C{expr}") - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)} - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)} Note that C{expr*(None,n)} does not raise an exception if more than n exprs exist in the input stream; that is, C{expr*(None,n)} does not enforce a maximum number of expr occurrences. If this behavior is desired, then write C{expr*(None,n) + ~expr} """ if isinstance(other,int): minElements, optElements = other,0 elif isinstance(other,tuple): other = (other + (None, None))[:2] if other[0] is None: other = (0, other[1]) if isinstance(other[0],int) and other[1] is None: if other[0] == 0: return ZeroOrMore(self) if other[0] == 1: return OneOrMore(self) else: return self*other[0] + ZeroOrMore(self) elif isinstance(other[0],int) and isinstance(other[1],int): minElements, optElements = other optElements -= minElements else: raise TypeError("cannot multiply 'ParserElement' and ('%s','%s') objects", type(other[0]),type(other[1])) else: raise TypeError("cannot multiply 'ParserElement' and '%s' objects", type(other)) if minElements < 0: raise ValueError("cannot multiply ParserElement by negative value") if optElements < 0: raise ValueError("second tuple value must be greater or equal to first tuple value") if minElements == optElements == 0: raise ValueError("cannot multiply ParserElement by 0 or (0,0)") if (optElements): def makeOptionalList(n): if n>1: return Optional(self + makeOptionalList(n-1)) else: return Optional(self) if minElements: if minElements == 1: ret = self + makeOptionalList(optElements) else: ret = And([self]*minElements) + makeOptionalList(optElements) else: ret = makeOptionalList(optElements) else: if minElements == 1: ret = self else: ret = And([self]*minElements) return ret def __rmul__(self, other): return self.__mul__(other) def __or__(self, other ): """ Implementation of | operator - returns C{L{MatchFirst}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return MatchFirst( [ self, other ] ) def __ror__(self, other ): """ Implementation of | operator when left operand is not a C{L{ParserElement}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other | self def __xor__(self, other ): """ Implementation of ^ operator - returns C{L{Or}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return Or( [ self, other ] ) def __rxor__(self, other ): """ Implementation of ^ operator when left operand is not a C{L{ParserElement}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other ^ self def __and__(self, other ): """ Implementation of & operator - returns C{L{Each}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return Each( [ self, other ] ) def __rand__(self, other ): """ Implementation of & operator when left operand is not a C{L{ParserElement}} """ if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) if not isinstance( other, ParserElement ): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other & self def __invert__( self ): """ Implementation of ~ operator - returns C{L{NotAny}} """ return NotAny( self ) def __call__(self, name=None): """ Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}. If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be passed as C{True}. If C{name} is omitted, same as calling C{L{copy}}. Example:: # these are equivalent userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno") userdata = Word(alphas)("name") + Word(nums+"-")("socsecno") """ if name is not None: return self.setResultsName(name) else: return self.copy() def suppress( self ): """ Suppresses the output of this C{ParserElement}; useful to keep punctuation from cluttering up returned output. """ return Suppress( self ) def leaveWhitespace( self ): """ Disables the skipping of whitespace before matching the characters in the C{ParserElement}'s defined pattern. This is normally only used internally by the pyparsing module, but may be needed in some whitespace-sensitive grammars. """ self.skipWhitespace = False return self def setWhitespaceChars( self, chars ): """ Overrides the default whitespace chars """ self.skipWhitespace = True self.whiteChars = chars self.copyDefaultWhiteChars = False return self def parseWithTabs( self ): """ Overrides default behavior to expand C{}s to spaces before parsing the input string. Must be called before C{parseString} when the input grammar contains elements that match C{} characters. """ self.keepTabs = True return self def ignore( self, other ): """ Define expression to be ignored (e.g., comments) while doing pattern matching; may be called repeatedly, to define multiple comment or other ignorable patterns. Example:: patt = OneOrMore(Word(alphas)) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj'] patt.ignore(cStyleComment) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd'] """ if isinstance(other, basestring): other = Suppress(other) if isinstance( other, Suppress ): if other not in self.ignoreExprs: self.ignoreExprs.append(other) else: self.ignoreExprs.append( Suppress( other.copy() ) ) return self def setDebugActions( self, startAction, successAction, exceptionAction ): """ Enable display of debugging messages while doing pattern matching. """ self.debugActions = (startAction or _defaultStartDebugAction, successAction or _defaultSuccessDebugAction, exceptionAction or _defaultExceptionDebugAction) self.debug = True return self def setDebug( self, flag=True ): """ Enable display of debugging messages while doing pattern matching. Set C{flag} to True to enable, False to disable. Example:: wd = Word(alphas).setName("alphaword") integer = Word(nums).setName("numword") term = wd | integer # turn on debugging for wd wd.setDebug() OneOrMore(term).parseString("abc 123 xyz 890") prints:: Match alphaword at loc 0(1,1) Matched alphaword -> ['abc'] Match alphaword at loc 3(1,4) Exception raised:Expected alphaword (at char 4), (line:1, col:5) Match alphaword at loc 7(1,8) Matched alphaword -> ['xyz'] Match alphaword at loc 11(1,12) Exception raised:Expected alphaword (at char 12), (line:1, col:13) Match alphaword at loc 15(1,16) Exception raised:Expected alphaword (at char 15), (line:1, col:16) The output shown is that produced by the default debug actions - custom debug actions can be specified using L{setDebugActions}. Prior to attempting to match the C{wd} expression, the debugging message C{"Match at loc (,)"} is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"} message is shown. Also note the use of L{setName} to assign a human-readable name to the expression, which makes debugging and exception messages easier to understand - for instance, the default name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}. """ if flag: self.setDebugActions( _defaultStartDebugAction, _defaultSuccessDebugAction, _defaultExceptionDebugAction ) else: self.debug = False return self def __str__( self ): return self.name def __repr__( self ): return _ustr(self) def streamline( self ): self.streamlined = True self.strRepr = None return self def checkRecursion( self, parseElementList ): pass def validate( self, validateTrace=[] ): """ Check defined expressions for valid structure, check for infinite recursive definitions. """ self.checkRecursion( [] ) def parseFile( self, file_or_filename, parseAll=False ): """ Execute the parse expression on the given file or filename. If a filename is specified (instead of a file object), the entire file is opened, read, and closed before parsing. """ try: file_contents = file_or_filename.read() except AttributeError: with open(file_or_filename, "r") as f: file_contents = f.read() try: return self.parseString(file_contents, parseAll) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def __eq__(self,other): if isinstance(other, ParserElement): return self is other or vars(self) == vars(other) elif isinstance(other, basestring): return self.matches(other) else: return super(ParserElement,self)==other def __ne__(self,other): return not (self == other) def __hash__(self): return hash(id(self)) def __req__(self,other): return self == other def __rne__(self,other): return not (self == other) def matches(self, testString, parseAll=True): """ Method for quick testing of a parser against a test string. Good for simple inline microtests of sub expressions while building up larger parser. Parameters: - testString - to test against this expression for a match - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests Example:: expr = Word(nums) assert expr.matches("100") """ try: self.parseString(_ustr(testString), parseAll=parseAll) return True except ParseBaseException: return False def runTests(self, tests, parseAll=True, comment='#', fullDump=True, printResults=True, failureTests=False): """ Execute the parse expression on a series of test strings, showing each test, the parsed results or where the parse failed. Quick and easy way to run a parse expression against a list of sample strings. Parameters: - tests - a list of separate test strings, or a multiline string of test strings - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests - comment - (default=C{'#'}) - expression for indicating embedded comments in the test string; pass None to disable comment filtering - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline; if False, only dump nested list - printResults - (default=C{True}) prints test output to stdout - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing Returns: a (success, results) tuple, where success indicates that all tests succeeded (or failed if C{failureTests} is True), and the results contain a list of lines of each test's output Example:: number_expr = pyparsing_common.number.copy() result = number_expr.runTests(''' # unsigned integer 100 # negative integer -100 # float with scientific notation 6.02e23 # integer with scientific notation 1e-12 ''') print("Success" if result[0] else "Failed!") result = number_expr.runTests(''' # stray character 100Z # missing leading digit before '.' -.100 # too many '.' 3.14.159 ''', failureTests=True) print("Success" if result[0] else "Failed!") prints:: # unsigned integer 100 [100] # negative integer -100 [-100] # float with scientific notation 6.02e23 [6.02e+23] # integer with scientific notation 1e-12 [1e-12] Success # stray character 100Z ^ FAIL: Expected end of text (at char 3), (line:1, col:4) # missing leading digit before '.' -.100 ^ FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1) # too many '.' 3.14.159 ^ FAIL: Expected end of text (at char 4), (line:1, col:5) Success Each test string must be on a single line. If you want to test a string that spans multiple lines, create a test like this:: expr.runTest(r"this is a test\\n of strings that spans \\n 3 lines") (Note that this is a raw string literal, you must include the leading 'r'.) """ if isinstance(tests, basestring): tests = list(map(str.strip, tests.rstrip().splitlines())) if isinstance(comment, basestring): comment = Literal(comment) allResults = [] comments = [] success = True for t in tests: if comment is not None and comment.matches(t, False) or comments and not t: comments.append(t) continue if not t: continue out = ['\n'.join(comments), t] comments = [] try: t = t.replace(r'\n','\n') result = self.parseString(t, parseAll=parseAll) out.append(result.dump(full=fullDump)) success = success and not failureTests except ParseBaseException as pe: fatal = "(FATAL)" if isinstance(pe, ParseFatalException) else "" if '\n' in t: out.append(line(pe.loc, t)) out.append(' '*(col(pe.loc,t)-1) + '^' + fatal) else: out.append(' '*pe.loc + '^' + fatal) out.append("FAIL: " + str(pe)) success = success and failureTests result = pe except Exception as exc: out.append("FAIL-EXCEPTION: " + str(exc)) success = success and failureTests result = exc if printResults: if fullDump: out.append('') print('\n'.join(out)) allResults.append((t, result)) return success, allResults class Token(ParserElement): """ Abstract C{ParserElement} subclass, for defining atomic matching patterns. """ def __init__( self ): super(Token,self).__init__( savelist=False ) class Empty(Token): """ An empty token, will always match. """ def __init__( self ): super(Empty,self).__init__() self.name = "Empty" self.mayReturnEmpty = True self.mayIndexError = False class NoMatch(Token): """ A token that will never match. """ def __init__( self ): super(NoMatch,self).__init__() self.name = "NoMatch" self.mayReturnEmpty = True self.mayIndexError = False self.errmsg = "Unmatchable token" def parseImpl( self, instring, loc, doActions=True ): raise ParseException(instring, loc, self.errmsg, self) class Literal(Token): """ Token to exactly match a specified string. Example:: Literal('blah').parseString('blah') # -> ['blah'] Literal('blah').parseString('blahfooblah') # -> ['blah'] Literal('blah').parseString('bla') # -> Exception: Expected "blah" For case-insensitive matching, use L{CaselessLiteral}. For keyword matching (force word break before and after the matched string), use L{Keyword} or L{CaselessKeyword}. """ def __init__( self, matchString ): super(Literal,self).__init__() self.match = matchString self.matchLen = len(matchString) try: self.firstMatchChar = matchString[0] except IndexError: warnings.warn("null string passed to Literal; use Empty() instead", SyntaxWarning, stacklevel=2) self.__class__ = Empty self.name = '"%s"' % _ustr(self.match) self.errmsg = "Expected " + self.name self.mayReturnEmpty = False self.mayIndexError = False # Performance tuning: this routine gets called a *lot* # if this is a single character match string and the first character matches, # short-circuit as quickly as possible, and avoid calling startswith #~ @profile def parseImpl( self, instring, loc, doActions=True ): if (instring[loc] == self.firstMatchChar and (self.matchLen==1 or instring.startswith(self.match,loc)) ): return loc+self.matchLen, self.match raise ParseException(instring, loc, self.errmsg, self) _L = Literal ParserElement._literalStringClass = Literal class Keyword(Token): """ Token to exactly match a specified string as a keyword, that is, it must be immediately followed by a non-keyword character. Compare with C{L{Literal}}: - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}. - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'} Accepts two optional constructor arguments in addition to the keyword string: - C{identChars} is a string of characters that would be valid identifier characters, defaulting to all alphanumerics + "_" and "$" - C{caseless} allows case-insensitive matching, default is C{False}. Example:: Keyword("start").parseString("start") # -> ['start'] Keyword("start").parseString("starting") # -> Exception For case-insensitive matching, use L{CaselessKeyword}. """ DEFAULT_KEYWORD_CHARS = alphanums+"_$" def __init__( self, matchString, identChars=None, caseless=False ): super(Keyword,self).__init__() if identChars is None: identChars = Keyword.DEFAULT_KEYWORD_CHARS self.match = matchString self.matchLen = len(matchString) try: self.firstMatchChar = matchString[0] except IndexError: warnings.warn("null string passed to Keyword; use Empty() instead", SyntaxWarning, stacklevel=2) self.name = '"%s"' % self.match self.errmsg = "Expected " + self.name self.mayReturnEmpty = False self.mayIndexError = False self.caseless = caseless if caseless: self.caselessmatch = matchString.upper() identChars = identChars.upper() self.identChars = set(identChars) def parseImpl( self, instring, loc, doActions=True ): if self.caseless: if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) and (loc == 0 or instring[loc-1].upper() not in self.identChars) ): return loc+self.matchLen, self.match else: if (instring[loc] == self.firstMatchChar and (self.matchLen==1 or instring.startswith(self.match,loc)) and (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen] not in self.identChars) and (loc == 0 or instring[loc-1] not in self.identChars) ): return loc+self.matchLen, self.match raise ParseException(instring, loc, self.errmsg, self) def copy(self): c = super(Keyword,self).copy() c.identChars = Keyword.DEFAULT_KEYWORD_CHARS return c @staticmethod def setDefaultKeywordChars( chars ): """Overrides the default Keyword chars """ Keyword.DEFAULT_KEYWORD_CHARS = chars class CaselessLiteral(Literal): """ Token to match a specified string, ignoring case of letters. Note: the matched results will always be in the case of the given match string, NOT the case of the input text. Example:: OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD'] (Contrast with example for L{CaselessKeyword}.) """ def __init__( self, matchString ): super(CaselessLiteral,self).__init__( matchString.upper() ) # Preserve the defining literal. self.returnString = matchString self.name = "'%s'" % self.returnString self.errmsg = "Expected " + self.name def parseImpl( self, instring, loc, doActions=True ): if instring[ loc:loc+self.matchLen ].upper() == self.match: return loc+self.matchLen, self.returnString raise ParseException(instring, loc, self.errmsg, self) class CaselessKeyword(Keyword): """ Caseless version of L{Keyword}. Example:: OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD'] (Contrast with example for L{CaselessLiteral}.) """ def __init__( self, matchString, identChars=None ): super(CaselessKeyword,self).__init__( matchString, identChars, caseless=True ) def parseImpl( self, instring, loc, doActions=True ): if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) ): return loc+self.matchLen, self.match raise ParseException(instring, loc, self.errmsg, self) class CloseMatch(Token): """ A variation on L{Literal} which matches "close" matches, that is, strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters: - C{match_string} - string to be matched - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match The results from a successful parse will contain the matched text from the input string and the following named results: - C{mismatches} - a list of the positions within the match_string where mismatches were found - C{original} - the original match_string used to compare against the input string If C{mismatches} is an empty list, then the match was an exact match. Example:: patt = CloseMatch("ATCATCGAATGGA") patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']}) patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1) # exact match patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']}) # close match allowing up to 2 mismatches patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2) patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']}) """ def __init__(self, match_string, maxMismatches=1): super(CloseMatch,self).__init__() self.name = match_string self.match_string = match_string self.maxMismatches = maxMismatches self.errmsg = "Expected %r (with up to %d mismatches)" % (self.match_string, self.maxMismatches) self.mayIndexError = False self.mayReturnEmpty = False def parseImpl( self, instring, loc, doActions=True ): start = loc instrlen = len(instring) maxloc = start + len(self.match_string) if maxloc <= instrlen: match_string = self.match_string match_stringloc = 0 mismatches = [] maxMismatches = self.maxMismatches for match_stringloc,s_m in enumerate(zip(instring[loc:maxloc], self.match_string)): src,mat = s_m if src != mat: mismatches.append(match_stringloc) if len(mismatches) > maxMismatches: break else: loc = match_stringloc + 1 results = ParseResults([instring[start:loc]]) results['original'] = self.match_string results['mismatches'] = mismatches return loc, results raise ParseException(instring, loc, self.errmsg, self) class Word(Token): """ Token for matching words composed of allowed character sets. Defined with string containing all allowed initial characters, an optional string containing allowed body characters (if omitted, defaults to the initial character set), and an optional minimum, maximum, and/or exact length. The default value for C{min} is 1 (a minimum value < 1 is not valid); the default values for C{max} and C{exact} are 0, meaning no maximum or exact length restriction. An optional C{excludeChars} parameter can list characters that might be found in the input C{bodyChars} string; useful to define a word of all printables except for one or two characters, for instance. L{srange} is useful for defining custom character set strings for defining C{Word} expressions, using range notation from regular expression character sets. A common mistake is to use C{Word} to match a specific literal string, as in C{Word("Address")}. Remember that C{Word} uses the string argument to define I{sets} of matchable characters. This expression would match "Add", "AAA", "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an exact literal string, use L{Literal} or L{Keyword}. pyparsing includes helper strings for building Words: - L{alphas} - L{nums} - L{alphanums} - L{hexnums} - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.) - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.) - L{printables} (any non-whitespace character) Example:: # a word composed of digits integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9")) # a word with a leading capital, and zero or more lowercase capital_word = Word(alphas.upper(), alphas.lower()) # hostnames are alphanumeric, with leading alpha, and '-' hostname = Word(alphas, alphanums+'-') # roman numeral (not a strict parser, accepts invalid mix of characters) roman = Word("IVXLCDM") # any string of non-whitespace characters, except for ',' csv_value = Word(printables, excludeChars=",") """ def __init__( self, initChars, bodyChars=None, min=1, max=0, exact=0, asKeyword=False, excludeChars=None ): super(Word,self).__init__() if excludeChars: initChars = ''.join(c for c in initChars if c not in excludeChars) if bodyChars: bodyChars = ''.join(c for c in bodyChars if c not in excludeChars) self.initCharsOrig = initChars self.initChars = set(initChars) if bodyChars : self.bodyCharsOrig = bodyChars self.bodyChars = set(bodyChars) else: self.bodyCharsOrig = initChars self.bodyChars = set(initChars) self.maxSpecified = max > 0 if min < 1: raise ValueError("cannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permitted") self.minLen = min if max > 0: self.maxLen = max else: self.maxLen = _MAX_INT if exact > 0: self.maxLen = exact self.minLen = exact self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayIndexError = False self.asKeyword = asKeyword if ' ' not in self.initCharsOrig+self.bodyCharsOrig and (min==1 and max==0 and exact==0): if self.bodyCharsOrig == self.initCharsOrig: self.reString = "[%s]+" % _escapeRegexRangeChars(self.initCharsOrig) elif len(self.initCharsOrig) == 1: self.reString = "%s[%s]*" % \ (re.escape(self.initCharsOrig), _escapeRegexRangeChars(self.bodyCharsOrig),) else: self.reString = "[%s][%s]*" % \ (_escapeRegexRangeChars(self.initCharsOrig), _escapeRegexRangeChars(self.bodyCharsOrig),) if self.asKeyword: self.reString = r"\b"+self.reString+r"\b" try: self.re = re.compile( self.reString ) except Exception: self.re = None def parseImpl( self, instring, loc, doActions=True ): if self.re: result = self.re.match(instring,loc) if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() return loc, result.group() if not(instring[ loc ] in self.initChars): raise ParseException(instring, loc, self.errmsg, self) start = loc loc += 1 instrlen = len(instring) bodychars = self.bodyChars maxloc = start + self.maxLen maxloc = min( maxloc, instrlen ) while loc < maxloc and instring[loc] in bodychars: loc += 1 throwException = False if loc - start < self.minLen: throwException = True if self.maxSpecified and loc < instrlen and instring[loc] in bodychars: throwException = True if self.asKeyword: if (start>0 and instring[start-1] in bodychars) or (loc4: return s[:4]+"..." else: return s if ( self.initCharsOrig != self.bodyCharsOrig ): self.strRepr = "W:(%s,%s)" % ( charsAsStr(self.initCharsOrig), charsAsStr(self.bodyCharsOrig) ) else: self.strRepr = "W:(%s)" % charsAsStr(self.initCharsOrig) return self.strRepr class Regex(Token): r""" Token for matching strings that match a given regular expression. Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module. If the given regex contains named groups (defined using C{(?P...)}), these will be preserved as named parse results. Example:: realnum = Regex(r"[+-]?\d+\.\d*") date = Regex(r'(?P\d{4})-(?P\d\d?)-(?P\d\d?)') # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})") """ compiledREtype = type(re.compile("[A-Z]")) def __init__( self, pattern, flags=0): """The parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags.""" super(Regex,self).__init__() if isinstance(pattern, basestring): if not pattern: warnings.warn("null string passed to Regex; use Empty() instead", SyntaxWarning, stacklevel=2) self.pattern = pattern self.flags = flags try: self.re = re.compile(self.pattern, self.flags) self.reString = self.pattern except sre_constants.error: warnings.warn("invalid pattern (%s) passed to Regex" % pattern, SyntaxWarning, stacklevel=2) raise elif isinstance(pattern, Regex.compiledREtype): self.re = pattern self.pattern = \ self.reString = str(pattern) self.flags = flags else: raise ValueError("Regex may only be constructed with a string or a compiled RE object") self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayIndexError = False self.mayReturnEmpty = True def parseImpl( self, instring, loc, doActions=True ): result = self.re.match(instring,loc) if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() d = result.groupdict() ret = ParseResults(result.group()) if d: for k in d: ret[k] = d[k] return loc,ret def __str__( self ): try: return super(Regex,self).__str__() except Exception: pass if self.strRepr is None: self.strRepr = "Re:(%s)" % repr(self.pattern) return self.strRepr class QuotedString(Token): r""" Token for matching strings that are delimited by quoting characters. Defined with the following parameters: - quoteChar - string of one or more characters defining the quote delimiting string - escChar - character to escape quotes, typically backslash (default=C{None}) - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None}) - multiline - boolean indicating whether quotes can span multiple lines (default=C{False}) - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True}) - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar) - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True}) Example:: qs = QuotedString('"') print(qs.searchString('lsjdf "This is the quote" sldjf')) complex_qs = QuotedString('{{', endQuoteChar='}}') print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf')) sql_qs = QuotedString('"', escQuote='""') print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf')) prints:: [['This is the quote']] [['This is the "quote"']] [['This is the quote with "embedded" quotes']] """ def __init__( self, quoteChar, escChar=None, escQuote=None, multiline=False, unquoteResults=True, endQuoteChar=None, convertWhitespaceEscapes=True): super(QuotedString,self).__init__() # remove white space from quote chars - wont work anyway quoteChar = quoteChar.strip() if not quoteChar: warnings.warn("quoteChar cannot be the empty string",SyntaxWarning,stacklevel=2) raise SyntaxError() if endQuoteChar is None: endQuoteChar = quoteChar else: endQuoteChar = endQuoteChar.strip() if not endQuoteChar: warnings.warn("endQuoteChar cannot be the empty string",SyntaxWarning,stacklevel=2) raise SyntaxError() self.quoteChar = quoteChar self.quoteCharLen = len(quoteChar) self.firstQuoteChar = quoteChar[0] self.endQuoteChar = endQuoteChar self.endQuoteCharLen = len(endQuoteChar) self.escChar = escChar self.escQuote = escQuote self.unquoteResults = unquoteResults self.convertWhitespaceEscapes = convertWhitespaceEscapes if multiline: self.flags = re.MULTILINE | re.DOTALL self.pattern = r'%s(?:[^%s%s]' % \ ( re.escape(self.quoteChar), _escapeRegexRangeChars(self.endQuoteChar[0]), (escChar is not None and _escapeRegexRangeChars(escChar) or '') ) else: self.flags = 0 self.pattern = r'%s(?:[^%s\n\r%s]' % \ ( re.escape(self.quoteChar), _escapeRegexRangeChars(self.endQuoteChar[0]), (escChar is not None and _escapeRegexRangeChars(escChar) or '') ) if len(self.endQuoteChar) > 1: self.pattern += ( '|(?:' + ')|(?:'.join("%s[^%s]" % (re.escape(self.endQuoteChar[:i]), _escapeRegexRangeChars(self.endQuoteChar[i])) for i in range(len(self.endQuoteChar)-1,0,-1)) + ')' ) if escQuote: self.pattern += (r'|(?:%s)' % re.escape(escQuote)) if escChar: self.pattern += (r'|(?:%s.)' % re.escape(escChar)) self.escCharReplacePattern = re.escape(self.escChar)+"(.)" self.pattern += (r')*%s' % re.escape(self.endQuoteChar)) try: self.re = re.compile(self.pattern, self.flags) self.reString = self.pattern except sre_constants.error: warnings.warn("invalid pattern (%s) passed to Regex" % self.pattern, SyntaxWarning, stacklevel=2) raise self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayIndexError = False self.mayReturnEmpty = True def parseImpl( self, instring, loc, doActions=True ): result = instring[loc] == self.firstQuoteChar and self.re.match(instring,loc) or None if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() ret = result.group() if self.unquoteResults: # strip off quotes ret = ret[self.quoteCharLen:-self.endQuoteCharLen] if isinstance(ret,basestring): # replace escaped whitespace if '\\' in ret and self.convertWhitespaceEscapes: ws_map = { r'\t' : '\t', r'\n' : '\n', r'\f' : '\f', r'\r' : '\r', } for wslit,wschar in ws_map.items(): ret = ret.replace(wslit, wschar) # replace escaped characters if self.escChar: ret = re.sub(self.escCharReplacePattern, r"\g<1>", ret) # replace escaped quotes if self.escQuote: ret = ret.replace(self.escQuote, self.endQuoteChar) return loc, ret def __str__( self ): try: return super(QuotedString,self).__str__() except Exception: pass if self.strRepr is None: self.strRepr = "quoted string, starting with %s ending with %s" % (self.quoteChar, self.endQuoteChar) return self.strRepr class CharsNotIn(Token): """ Token for matching words composed of characters I{not} in a given set (will include whitespace in matched characters if not listed in the provided exclusion set - see example). Defined with string containing all disallowed characters, and an optional minimum, maximum, and/or exact length. The default value for C{min} is 1 (a minimum value < 1 is not valid); the default values for C{max} and C{exact} are 0, meaning no maximum or exact length restriction. Example:: # define a comma-separated-value as anything that is not a ',' csv_value = CharsNotIn(',') print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213")) prints:: ['dkls', 'lsdkjf', 's12 34', '@!#', '213'] """ def __init__( self, notChars, min=1, max=0, exact=0 ): super(CharsNotIn,self).__init__() self.skipWhitespace = False self.notChars = notChars if min < 1: raise ValueError("cannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permitted") self.minLen = min if max > 0: self.maxLen = max else: self.maxLen = _MAX_INT if exact > 0: self.maxLen = exact self.minLen = exact self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayReturnEmpty = ( self.minLen == 0 ) self.mayIndexError = False def parseImpl( self, instring, loc, doActions=True ): if instring[loc] in self.notChars: raise ParseException(instring, loc, self.errmsg, self) start = loc loc += 1 notchars = self.notChars maxlen = min( start+self.maxLen, len(instring) ) while loc < maxlen and \ (instring[loc] not in notchars): loc += 1 if loc - start < self.minLen: raise ParseException(instring, loc, self.errmsg, self) return loc, instring[start:loc] def __str__( self ): try: return super(CharsNotIn, self).__str__() except Exception: pass if self.strRepr is None: if len(self.notChars) > 4: self.strRepr = "!W:(%s...)" % self.notChars[:4] else: self.strRepr = "!W:(%s)" % self.notChars return self.strRepr class White(Token): """ Special matching class for matching whitespace. Normally, whitespace is ignored by pyparsing grammars. This class is included when some whitespace structures are significant. Define with a string containing the whitespace characters to be matched; default is C{" \\t\\r\\n"}. Also takes optional C{min}, C{max}, and C{exact} arguments, as defined for the C{L{Word}} class. """ whiteStrs = { " " : "", "\t": "", "\n": "", "\r": "", "\f": "", } def __init__(self, ws=" \t\r\n", min=1, max=0, exact=0): super(White,self).__init__() self.matchWhite = ws self.setWhitespaceChars( "".join(c for c in self.whiteChars if c not in self.matchWhite) ) #~ self.leaveWhitespace() self.name = ("".join(White.whiteStrs[c] for c in self.matchWhite)) self.mayReturnEmpty = True self.errmsg = "Expected " + self.name self.minLen = min if max > 0: self.maxLen = max else: self.maxLen = _MAX_INT if exact > 0: self.maxLen = exact self.minLen = exact def parseImpl( self, instring, loc, doActions=True ): if not(instring[ loc ] in self.matchWhite): raise ParseException(instring, loc, self.errmsg, self) start = loc loc += 1 maxloc = start + self.maxLen maxloc = min( maxloc, len(instring) ) while loc < maxloc and instring[loc] in self.matchWhite: loc += 1 if loc - start < self.minLen: raise ParseException(instring, loc, self.errmsg, self) return loc, instring[start:loc] class _PositionToken(Token): def __init__( self ): super(_PositionToken,self).__init__() self.name=self.__class__.__name__ self.mayReturnEmpty = True self.mayIndexError = False class GoToColumn(_PositionToken): """ Token to advance to a specific column of input text; useful for tabular report scraping. """ def __init__( self, colno ): super(GoToColumn,self).__init__() self.col = colno def preParse( self, instring, loc ): if col(loc,instring) != self.col: instrlen = len(instring) if self.ignoreExprs: loc = self._skipIgnorables( instring, loc ) while loc < instrlen and instring[loc].isspace() and col( loc, instring ) != self.col : loc += 1 return loc def parseImpl( self, instring, loc, doActions=True ): thiscol = col( loc, instring ) if thiscol > self.col: raise ParseException( instring, loc, "Text not in expected column", self ) newloc = loc + self.col - thiscol ret = instring[ loc: newloc ] return newloc, ret class LineStart(_PositionToken): """ Matches if current position is at the beginning of a line within the parse string Example:: test = '''\ AAA this line AAA and this line AAA but not this one B AAA and definitely not this one ''' for t in (LineStart() + 'AAA' + restOfLine).searchString(test): print(t) Prints:: ['AAA', ' this line'] ['AAA', ' and this line'] """ def __init__( self ): super(LineStart,self).__init__() self.errmsg = "Expected start of line" def parseImpl( self, instring, loc, doActions=True ): if col(loc, instring) == 1: return loc, [] raise ParseException(instring, loc, self.errmsg, self) class LineEnd(_PositionToken): """ Matches if current position is at the end of a line within the parse string """ def __init__( self ): super(LineEnd,self).__init__() self.setWhitespaceChars( ParserElement.DEFAULT_WHITE_CHARS.replace("\n","") ) self.errmsg = "Expected end of line" def parseImpl( self, instring, loc, doActions=True ): if loc len(instring): return loc, [] else: raise ParseException(instring, loc, self.errmsg, self) class WordStart(_PositionToken): """ Matches if the current position is at the beginning of a Word, and is not preceded by any character in a given set of C{wordChars} (default=C{printables}). To emulate the C{\b} behavior of regular expressions, use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of the string being parsed, or at the beginning of a line. """ def __init__(self, wordChars = printables): super(WordStart,self).__init__() self.wordChars = set(wordChars) self.errmsg = "Not at the start of a word" def parseImpl(self, instring, loc, doActions=True ): if loc != 0: if (instring[loc-1] in self.wordChars or instring[loc] not in self.wordChars): raise ParseException(instring, loc, self.errmsg, self) return loc, [] class WordEnd(_PositionToken): """ Matches if the current position is at the end of a Word, and is not followed by any character in a given set of C{wordChars} (default=C{printables}). To emulate the C{\b} behavior of regular expressions, use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of the string being parsed, or at the end of a line. """ def __init__(self, wordChars = printables): super(WordEnd,self).__init__() self.wordChars = set(wordChars) self.skipWhitespace = False self.errmsg = "Not at the end of a word" def parseImpl(self, instring, loc, doActions=True ): instrlen = len(instring) if instrlen>0 and loc maxExcLoc: maxException = err maxExcLoc = err.loc except IndexError: if len(instring) > maxExcLoc: maxException = ParseException(instring,len(instring),e.errmsg,self) maxExcLoc = len(instring) else: # save match among all matches, to retry longest to shortest matches.append((loc2, e)) if matches: matches.sort(key=lambda x: -x[0]) for _,e in matches: try: return e._parse( instring, loc, doActions ) except ParseException as err: err.__traceback__ = None if err.loc > maxExcLoc: maxException = err maxExcLoc = err.loc if maxException is not None: maxException.msg = self.errmsg raise maxException else: raise ParseException(instring, loc, "no defined alternatives to match", self) def __ixor__(self, other ): if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) return self.append( other ) #Or( [ self, other ] ) def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "{" + " ^ ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr def checkRecursion( self, parseElementList ): subRecCheckList = parseElementList[:] + [ self ] for e in self.exprs: e.checkRecursion( subRecCheckList ) class MatchFirst(ParseExpression): """ Requires that at least one C{ParseExpression} is found. If two expressions match, the first one listed is the one that will match. May be constructed using the C{'|'} operator. Example:: # construct MatchFirst using '|' operator # watch the order of expressions to match number = Word(nums) | Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) # Fail! -> [['123'], ['3'], ['1416'], ['789']] # put more selective expression first number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums) print(number.searchString("123 3.1416 789")) # Better -> [['123'], ['3.1416'], ['789']] """ def __init__( self, exprs, savelist = False ): super(MatchFirst,self).__init__(exprs, savelist) if self.exprs: self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs) else: self.mayReturnEmpty = True def parseImpl( self, instring, loc, doActions=True ): maxExcLoc = -1 maxException = None for e in self.exprs: try: ret = e._parse( instring, loc, doActions ) return ret except ParseException as err: if err.loc > maxExcLoc: maxException = err maxExcLoc = err.loc except IndexError: if len(instring) > maxExcLoc: maxException = ParseException(instring,len(instring),e.errmsg,self) maxExcLoc = len(instring) # only got here if no expression matched, raise exception for match that made it the furthest else: if maxException is not None: maxException.msg = self.errmsg raise maxException else: raise ParseException(instring, loc, "no defined alternatives to match", self) def __ior__(self, other ): if isinstance( other, basestring ): other = ParserElement._literalStringClass( other ) return self.append( other ) #MatchFirst( [ self, other ] ) def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "{" + " | ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr def checkRecursion( self, parseElementList ): subRecCheckList = parseElementList[:] + [ self ] for e in self.exprs: e.checkRecursion( subRecCheckList ) class Each(ParseExpression): """ Requires all given C{ParseExpression}s to be found, but in any order. Expressions may be separated by whitespace. May be constructed using the C{'&'} operator. Example:: color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN") shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON") integer = Word(nums) shape_attr = "shape:" + shape_type("shape") posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn") color_attr = "color:" + color("color") size_attr = "size:" + integer("size") # use Each (using operator '&') to accept attributes in any order # (shape and posn are required, color and size are optional) shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr) shape_spec.runTests(''' shape: SQUARE color: BLACK posn: 100, 120 shape: CIRCLE size: 50 color: BLUE posn: 50,80 color:GREEN size:20 shape:TRIANGLE posn:20,40 ''' ) prints:: shape: SQUARE color: BLACK posn: 100, 120 ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']] - color: BLACK - posn: ['100', ',', '120'] - x: 100 - y: 120 - shape: SQUARE shape: CIRCLE size: 50 color: BLUE posn: 50,80 ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']] - color: BLUE - posn: ['50', ',', '80'] - x: 50 - y: 80 - shape: CIRCLE - size: 50 color: GREEN size: 20 shape: TRIANGLE posn: 20,40 ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']] - color: GREEN - posn: ['20', ',', '40'] - x: 20 - y: 40 - shape: TRIANGLE - size: 20 """ def __init__( self, exprs, savelist = True ): super(Each,self).__init__(exprs, savelist) self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs) self.skipWhitespace = True self.initExprGroups = True def parseImpl( self, instring, loc, doActions=True ): if self.initExprGroups: self.opt1map = dict((id(e.expr),e) for e in self.exprs if isinstance(e,Optional)) opt1 = [ e.expr for e in self.exprs if isinstance(e,Optional) ] opt2 = [ e for e in self.exprs if e.mayReturnEmpty and not isinstance(e,Optional)] self.optionals = opt1 + opt2 self.multioptionals = [ e.expr for e in self.exprs if isinstance(e,ZeroOrMore) ] self.multirequired = [ e.expr for e in self.exprs if isinstance(e,OneOrMore) ] self.required = [ e for e in self.exprs if not isinstance(e,(Optional,ZeroOrMore,OneOrMore)) ] self.required += self.multirequired self.initExprGroups = False tmpLoc = loc tmpReqd = self.required[:] tmpOpt = self.optionals[:] matchOrder = [] keepMatching = True while keepMatching: tmpExprs = tmpReqd + tmpOpt + self.multioptionals + self.multirequired failed = [] for e in tmpExprs: try: tmpLoc = e.tryParse( instring, tmpLoc ) except ParseException: failed.append(e) else: matchOrder.append(self.opt1map.get(id(e),e)) if e in tmpReqd: tmpReqd.remove(e) elif e in tmpOpt: tmpOpt.remove(e) if len(failed) == len(tmpExprs): keepMatching = False if tmpReqd: missing = ", ".join(_ustr(e) for e in tmpReqd) raise ParseException(instring,loc,"Missing one or more required elements (%s)" % missing ) # add any unmatched Optionals, in case they have default values defined matchOrder += [e for e in self.exprs if isinstance(e,Optional) and e.expr in tmpOpt] resultlist = [] for e in matchOrder: loc,results = e._parse(instring,loc,doActions) resultlist.append(results) finalResults = sum(resultlist, ParseResults([])) return loc, finalResults def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "{" + " & ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr def checkRecursion( self, parseElementList ): subRecCheckList = parseElementList[:] + [ self ] for e in self.exprs: e.checkRecursion( subRecCheckList ) class ParseElementEnhance(ParserElement): """ Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens. """ def __init__( self, expr, savelist=False ): super(ParseElementEnhance,self).__init__(savelist) if isinstance( expr, basestring ): if issubclass(ParserElement._literalStringClass, Token): expr = ParserElement._literalStringClass(expr) else: expr = ParserElement._literalStringClass(Literal(expr)) self.expr = expr self.strRepr = None if expr is not None: self.mayIndexError = expr.mayIndexError self.mayReturnEmpty = expr.mayReturnEmpty self.setWhitespaceChars( expr.whiteChars ) self.skipWhitespace = expr.skipWhitespace self.saveAsList = expr.saveAsList self.callPreparse = expr.callPreparse self.ignoreExprs.extend(expr.ignoreExprs) def parseImpl( self, instring, loc, doActions=True ): if self.expr is not None: return self.expr._parse( instring, loc, doActions, callPreParse=False ) else: raise ParseException("",loc,self.errmsg,self) def leaveWhitespace( self ): self.skipWhitespace = False self.expr = self.expr.copy() if self.expr is not None: self.expr.leaveWhitespace() return self def ignore( self, other ): if isinstance( other, Suppress ): if other not in self.ignoreExprs: super( ParseElementEnhance, self).ignore( other ) if self.expr is not None: self.expr.ignore( self.ignoreExprs[-1] ) else: super( ParseElementEnhance, self).ignore( other ) if self.expr is not None: self.expr.ignore( self.ignoreExprs[-1] ) return self def streamline( self ): super(ParseElementEnhance,self).streamline() if self.expr is not None: self.expr.streamline() return self def checkRecursion( self, parseElementList ): if self in parseElementList: raise RecursiveGrammarException( parseElementList+[self] ) subRecCheckList = parseElementList[:] + [ self ] if self.expr is not None: self.expr.checkRecursion( subRecCheckList ) def validate( self, validateTrace=[] ): tmp = validateTrace[:]+[self] if self.expr is not None: self.expr.validate(tmp) self.checkRecursion( [] ) def __str__( self ): try: return super(ParseElementEnhance,self).__str__() except Exception: pass if self.strRepr is None and self.expr is not None: self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.expr) ) return self.strRepr class FollowedBy(ParseElementEnhance): """ Lookahead matching of the given parse expression. C{FollowedBy} does I{not} advance the parsing position within the input string, it only verifies that the specified parse expression matches at the current position. C{FollowedBy} always returns a null token list. Example:: # use FollowedBy to match a label only if it is followed by a ':' data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint() prints:: [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']] """ def __init__( self, expr ): super(FollowedBy,self).__init__(expr) self.mayReturnEmpty = True def parseImpl( self, instring, loc, doActions=True ): self.expr.tryParse( instring, loc ) return loc, [] class NotAny(ParseElementEnhance): """ Lookahead to disallow matching with the given parse expression. C{NotAny} does I{not} advance the parsing position within the input string, it only verifies that the specified parse expression does I{not} match at the current position. Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny} always returns a null token list. May be constructed using the '~' operator. Example:: """ def __init__( self, expr ): super(NotAny,self).__init__(expr) #~ self.leaveWhitespace() self.skipWhitespace = False # do NOT use self.leaveWhitespace(), don't want to propagate to exprs self.mayReturnEmpty = True self.errmsg = "Found unwanted token, "+_ustr(self.expr) def parseImpl( self, instring, loc, doActions=True ): if self.expr.canParseNext(instring, loc): raise ParseException(instring, loc, self.errmsg, self) return loc, [] def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "~{" + _ustr(self.expr) + "}" return self.strRepr class _MultipleMatch(ParseElementEnhance): def __init__( self, expr, stopOn=None): super(_MultipleMatch, self).__init__(expr) self.saveAsList = True ender = stopOn if isinstance(ender, basestring): ender = ParserElement._literalStringClass(ender) self.not_ender = ~ender if ender is not None else None def parseImpl( self, instring, loc, doActions=True ): self_expr_parse = self.expr._parse self_skip_ignorables = self._skipIgnorables check_ender = self.not_ender is not None if check_ender: try_not_ender = self.not_ender.tryParse # must be at least one (but first see if we are the stopOn sentinel; # if so, fail) if check_ender: try_not_ender(instring, loc) loc, tokens = self_expr_parse( instring, loc, doActions, callPreParse=False ) try: hasIgnoreExprs = (not not self.ignoreExprs) while 1: if check_ender: try_not_ender(instring, loc) if hasIgnoreExprs: preloc = self_skip_ignorables( instring, loc ) else: preloc = loc loc, tmptokens = self_expr_parse( instring, preloc, doActions ) if tmptokens or tmptokens.haskeys(): tokens += tmptokens except (ParseException,IndexError): pass return loc, tokens class OneOrMore(_MultipleMatch): """ Repetition of one or more of the given expression. Parameters: - expr - expression that must match one or more times - stopOn - (default=C{None}) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: BLACK" OneOrMore(attr_expr).parseString(text).pprint() # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']] # use stopOn attribute for OneOrMore to avoid reading label string as part of the data attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']] # could also be written as (attr_expr * (1,)).parseString(text).pprint() """ def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "{" + _ustr(self.expr) + "}..." return self.strRepr class ZeroOrMore(_MultipleMatch): """ Optional repetition of zero or more of the given expression. Parameters: - expr - expression that must match zero or more times - stopOn - (default=C{None}) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example: similar to L{OneOrMore} """ def __init__( self, expr, stopOn=None): super(ZeroOrMore,self).__init__(expr, stopOn=stopOn) self.mayReturnEmpty = True def parseImpl( self, instring, loc, doActions=True ): try: return super(ZeroOrMore, self).parseImpl(instring, loc, doActions) except (ParseException,IndexError): return loc, [] def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "[" + _ustr(self.expr) + "]..." return self.strRepr class _NullToken(object): def __bool__(self): return False __nonzero__ = __bool__ def __str__(self): return "" _optionalNotMatched = _NullToken() class Optional(ParseElementEnhance): """ Optional matching of the given expression. Parameters: - expr - expression that must match zero or more times - default (optional) - value to be returned if the optional expression is not found. Example:: # US postal code can be a 5-digit zip, plus optional 4-digit qualifier zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4))) zip.runTests(''' # traditional ZIP code 12345 # ZIP+4 form 12101-0001 # invalid ZIP 98765- ''') prints:: # traditional ZIP code 12345 ['12345'] # ZIP+4 form 12101-0001 ['12101-0001'] # invalid ZIP 98765- ^ FAIL: Expected end of text (at char 5), (line:1, col:6) """ def __init__( self, expr, default=_optionalNotMatched ): super(Optional,self).__init__( expr, savelist=False ) self.saveAsList = self.expr.saveAsList self.defaultValue = default self.mayReturnEmpty = True def parseImpl( self, instring, loc, doActions=True ): try: loc, tokens = self.expr._parse( instring, loc, doActions, callPreParse=False ) except (ParseException,IndexError): if self.defaultValue is not _optionalNotMatched: if self.expr.resultsName: tokens = ParseResults([ self.defaultValue ]) tokens[self.expr.resultsName] = self.defaultValue else: tokens = [ self.defaultValue ] else: tokens = [] return loc, tokens def __str__( self ): if hasattr(self,"name"): return self.name if self.strRepr is None: self.strRepr = "[" + _ustr(self.expr) + "]" return self.strRepr class SkipTo(ParseElementEnhance): """ Token for skipping over all undefined text until the matched expression is found. Parameters: - expr - target expression marking the end of the data to be skipped - include - (default=C{False}) if True, the target expression is also parsed (the skipped text and target expression are returned as a 2-element list). - ignore - (default=C{None}) used to define grammars (typically quoted strings and comments) that might contain false matches to the target expression - failOn - (default=C{None}) define expressions that are not allowed to be included in the skipped test; if found before the target expression is found, the SkipTo is not a match Example:: report = ''' Outstanding Issues Report - 1 Jan 2000 # | Severity | Description | Days Open -----+----------+-------------------------------------------+----------- 101 | Critical | Intermittent system crash | 6 94 | Cosmetic | Spelling error on Login ('log|n') | 14 79 | Minor | System slow when running too many reports | 47 ''' integer = Word(nums) SEP = Suppress('|') # use SkipTo to simply match everything up until the next SEP # - ignore quoted strings, so that a '|' character inside a quoted string does not match # - parse action will call token.strip() for each matched token, i.e., the description body string_data = SkipTo(SEP, ignore=quotedString) string_data.setParseAction(tokenMap(str.strip)) ticket_expr = (integer("issue_num") + SEP + string_data("sev") + SEP + string_data("desc") + SEP + integer("days_open")) for tkt in ticket_expr.searchString(report): print tkt.dump() prints:: ['101', 'Critical', 'Intermittent system crash', '6'] - days_open: 6 - desc: Intermittent system crash - issue_num: 101 - sev: Critical ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14'] - days_open: 14 - desc: Spelling error on Login ('log|n') - issue_num: 94 - sev: Cosmetic ['79', 'Minor', 'System slow when running too many reports', '47'] - days_open: 47 - desc: System slow when running too many reports - issue_num: 79 - sev: Minor """ def __init__( self, other, include=False, ignore=None, failOn=None ): super( SkipTo, self ).__init__( other ) self.ignoreExpr = ignore self.mayReturnEmpty = True self.mayIndexError = False self.includeMatch = include self.asList = False if isinstance(failOn, basestring): self.failOn = ParserElement._literalStringClass(failOn) else: self.failOn = failOn self.errmsg = "No match found for "+_ustr(self.expr) def parseImpl( self, instring, loc, doActions=True ): startloc = loc instrlen = len(instring) expr = self.expr expr_parse = self.expr._parse self_failOn_canParseNext = self.failOn.canParseNext if self.failOn is not None else None self_ignoreExpr_tryParse = self.ignoreExpr.tryParse if self.ignoreExpr is not None else None tmploc = loc while tmploc <= instrlen: if self_failOn_canParseNext is not None: # break if failOn expression matches if self_failOn_canParseNext(instring, tmploc): break if self_ignoreExpr_tryParse is not None: # advance past ignore expressions while 1: try: tmploc = self_ignoreExpr_tryParse(instring, tmploc) except ParseBaseException: break try: expr_parse(instring, tmploc, doActions=False, callPreParse=False) except (ParseException, IndexError): # no match, advance loc in string tmploc += 1 else: # matched skipto expr, done break else: # ran off the end of the input string without matching skipto expr, fail raise ParseException(instring, loc, self.errmsg, self) # build up return values loc = tmploc skiptext = instring[startloc:loc] skipresult = ParseResults(skiptext) if self.includeMatch: loc, mat = expr_parse(instring,loc,doActions,callPreParse=False) skipresult += mat return loc, skipresult class Forward(ParseElementEnhance): """ Forward declaration of an expression to be defined later - used for recursive grammars, such as algebraic infix notation. When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator. Note: take care when assigning to C{Forward} not to overlook precedence of operators. Specifically, '|' has a lower precedence than '<<', so that:: fwdExpr << a | b | c will actually be evaluated as:: (fwdExpr << a) | b | c thereby leaving b and c out as parseable alternatives. It is recommended that you explicitly group the values inserted into the C{Forward}:: fwdExpr << (a | b | c) Converting to use the '<<=' operator instead will avoid this problem. See L{ParseResults.pprint} for an example of a recursive parser created using C{Forward}. """ def __init__( self, other=None ): super(Forward,self).__init__( other, savelist=False ) def __lshift__( self, other ): if isinstance( other, basestring ): other = ParserElement._literalStringClass(other) self.expr = other self.strRepr = None self.mayIndexError = self.expr.mayIndexError self.mayReturnEmpty = self.expr.mayReturnEmpty self.setWhitespaceChars( self.expr.whiteChars ) self.skipWhitespace = self.expr.skipWhitespace self.saveAsList = self.expr.saveAsList self.ignoreExprs.extend(self.expr.ignoreExprs) return self def __ilshift__(self, other): return self << other def leaveWhitespace( self ): self.skipWhitespace = False return self def streamline( self ): if not self.streamlined: self.streamlined = True if self.expr is not None: self.expr.streamline() return self def validate( self, validateTrace=[] ): if self not in validateTrace: tmp = validateTrace[:]+[self] if self.expr is not None: self.expr.validate(tmp) self.checkRecursion([]) def __str__( self ): if hasattr(self,"name"): return self.name return self.__class__.__name__ + ": ..." # stubbed out for now - creates awful memory and perf issues self._revertClass = self.__class__ self.__class__ = _ForwardNoRecurse try: if self.expr is not None: retString = _ustr(self.expr) else: retString = "None" finally: self.__class__ = self._revertClass return self.__class__.__name__ + ": " + retString def copy(self): if self.expr is not None: return super(Forward,self).copy() else: ret = Forward() ret <<= self return ret class _ForwardNoRecurse(Forward): def __str__( self ): return "..." class TokenConverter(ParseElementEnhance): """ Abstract subclass of C{ParseExpression}, for converting parsed results. """ def __init__( self, expr, savelist=False ): super(TokenConverter,self).__init__( expr )#, savelist ) self.saveAsList = False class Combine(TokenConverter): """ Converter to concatenate all matching tokens to a single string. By default, the matching patterns must also be contiguous in the input string; this can be disabled by specifying C{'adjacent=False'} in the constructor. Example:: real = Word(nums) + '.' + Word(nums) print(real.parseString('3.1416')) # -> ['3', '.', '1416'] # will also erroneously match the following print(real.parseString('3. 1416')) # -> ['3', '.', '1416'] real = Combine(Word(nums) + '.' + Word(nums)) print(real.parseString('3.1416')) # -> ['3.1416'] # no match when there are internal spaces print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...) """ def __init__( self, expr, joinString="", adjacent=True ): super(Combine,self).__init__( expr ) # suppress whitespace-stripping in contained parse expressions, but re-enable it on the Combine itself if adjacent: self.leaveWhitespace() self.adjacent = adjacent self.skipWhitespace = True self.joinString = joinString self.callPreparse = True def ignore( self, other ): if self.adjacent: ParserElement.ignore(self, other) else: super( Combine, self).ignore( other ) return self def postParse( self, instring, loc, tokenlist ): retToks = tokenlist.copy() del retToks[:] retToks += ParseResults([ "".join(tokenlist._asStringList(self.joinString)) ], modal=self.modalResults) if self.resultsName and retToks.haskeys(): return [ retToks ] else: return retToks class Group(TokenConverter): """ Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions. Example:: ident = Word(alphas) num = Word(nums) term = ident | num func = ident + Optional(delimitedList(term)) print(func.parseString("fn a,b,100")) # -> ['fn', 'a', 'b', '100'] func = ident + Group(Optional(delimitedList(term))) print(func.parseString("fn a,b,100")) # -> ['fn', ['a', 'b', '100']] """ def __init__( self, expr ): super(Group,self).__init__( expr ) self.saveAsList = True def postParse( self, instring, loc, tokenlist ): return [ tokenlist ] class Dict(TokenConverter): """ Converter to return a repetitive expression as a list, but also as a dictionary. Each element can also be referenced using the first token in the expression as its key. Useful for tabular report scraping when the first column can be used as a item key. Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) # print attributes as plain groups print(OneOrMore(attr_expr).parseString(text).dump()) # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names result = Dict(OneOrMore(Group(attr_expr))).parseString(text) print(result.dump()) # access named fields as dict entries, or output as dict print(result['shape']) print(result.asDict()) prints:: ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap'] [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'} See more examples at L{ParseResults} of accessing fields by results name. """ def __init__( self, expr ): super(Dict,self).__init__( expr ) self.saveAsList = True def postParse( self, instring, loc, tokenlist ): for i,tok in enumerate(tokenlist): if len(tok) == 0: continue ikey = tok[0] if isinstance(ikey,int): ikey = _ustr(tok[0]).strip() if len(tok)==1: tokenlist[ikey] = _ParseResultsWithOffset("",i) elif len(tok)==2 and not isinstance(tok[1],ParseResults): tokenlist[ikey] = _ParseResultsWithOffset(tok[1],i) else: dictvalue = tok.copy() #ParseResults(i) del dictvalue[0] if len(dictvalue)!= 1 or (isinstance(dictvalue,ParseResults) and dictvalue.haskeys()): tokenlist[ikey] = _ParseResultsWithOffset(dictvalue,i) else: tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0],i) if self.resultsName: return [ tokenlist ] else: return tokenlist class Suppress(TokenConverter): """ Converter for ignoring the results of a parsed expression. Example:: source = "a, b, c,d" wd = Word(alphas) wd_list1 = wd + ZeroOrMore(',' + wd) print(wd_list1.parseString(source)) # often, delimiters that are useful during parsing are just in the # way afterward - use Suppress to keep them out of the parsed output wd_list2 = wd + ZeroOrMore(Suppress(',') + wd) print(wd_list2.parseString(source)) prints:: ['a', ',', 'b', ',', 'c', ',', 'd'] ['a', 'b', 'c', 'd'] (See also L{delimitedList}.) """ def postParse( self, instring, loc, tokenlist ): return [] def suppress( self ): return self class OnlyOnce(object): """ Wrapper for parse actions, to ensure they are only called once. """ def __init__(self, methodCall): self.callable = _trim_arity(methodCall) self.called = False def __call__(self,s,l,t): if not self.called: results = self.callable(s,l,t) self.called = True return results raise ParseException(s,l,"") def reset(self): self.called = False def traceParseAction(f): """ Decorator for debugging parse actions. When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".} When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised. Example:: wd = Word(alphas) @traceParseAction def remove_duplicate_chars(tokens): return ''.join(sorted(set(''.join(tokens)))) wds = OneOrMore(wd).setParseAction(remove_duplicate_chars) print(wds.parseString("slkdjs sld sldd sdlf sdljf")) prints:: >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {})) <3: thisFunc = paArgs[0].__class__.__name__ + '.' + thisFunc sys.stderr.write( ">>entering %s(line: '%s', %d, %r)\n" % (thisFunc,line(l,s),l,t) ) try: ret = f(*paArgs) except Exception as exc: sys.stderr.write( "< ['aa', 'bb', 'cc'] delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE'] """ dlName = _ustr(expr)+" ["+_ustr(delim)+" "+_ustr(expr)+"]..." if combine: return Combine( expr + ZeroOrMore( delim + expr ) ).setName(dlName) else: return ( expr + ZeroOrMore( Suppress( delim ) + expr ) ).setName(dlName) def countedArray( expr, intExpr=None ): """ Helper to define a counted list of expressions. This helper defines a pattern of the form:: integer expr expr expr... where the leading integer tells how many expr expressions follow. The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed. If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value. Example:: countedArray(Word(alphas)).parseString('2 ab cd ef') # -> ['ab', 'cd'] # in this parser, the leading integer value is given in binary, # '10' indicating that 2 values are in the array binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2)) countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef') # -> ['ab', 'cd'] """ arrayExpr = Forward() def countFieldParseAction(s,l,t): n = t[0] arrayExpr << (n and Group(And([expr]*n)) or Group(empty)) return [] if intExpr is None: intExpr = Word(nums).setParseAction(lambda t:int(t[0])) else: intExpr = intExpr.copy() intExpr.setName("arrayLen") intExpr.addParseAction(countFieldParseAction, callDuringTry=True) return ( intExpr + arrayExpr ).setName('(len) ' + _ustr(expr) + '...') def _flatten(L): ret = [] for i in L: if isinstance(i,list): ret.extend(_flatten(i)) else: ret.append(i) return ret def matchPreviousLiteral(expr): """ Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousLiteral(first) matchExpr = first + ":" + second will match C{"1:1"}, but not C{"1:2"}. Because this matches a previous literal, will also match the leading C{"1:1"} in C{"1:10"}. If this is not desired, use C{matchPreviousExpr}. Do I{not} use with packrat parsing enabled. """ rep = Forward() def copyTokenToRepeater(s,l,t): if t: if len(t) == 1: rep << t[0] else: # flatten t tokens tflat = _flatten(t.asList()) rep << And(Literal(tt) for tt in tflat) else: rep << Empty() expr.addParseAction(copyTokenToRepeater, callDuringTry=True) rep.setName('(prev) ' + _ustr(expr)) return rep def matchPreviousExpr(expr): """ Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousExpr(first) matchExpr = first + ":" + second will match C{"1:1"}, but not C{"1:2"}. Because this matches by expressions, will I{not} match the leading C{"1:1"} in C{"1:10"}; the expressions are evaluated first, and then compared, so C{"1"} is compared with C{"10"}. Do I{not} use with packrat parsing enabled. """ rep = Forward() e2 = expr.copy() rep <<= e2 def copyTokenToRepeater(s,l,t): matchTokens = _flatten(t.asList()) def mustMatchTheseTokens(s,l,t): theseTokens = _flatten(t.asList()) if theseTokens != matchTokens: raise ParseException("",0,"") rep.setParseAction( mustMatchTheseTokens, callDuringTry=True ) expr.addParseAction(copyTokenToRepeater, callDuringTry=True) rep.setName('(prev) ' + _ustr(expr)) return rep def _escapeRegexRangeChars(s): #~ escape these chars: ^-] for c in r"\^-]": s = s.replace(c,_bslash+c) s = s.replace("\n",r"\n") s = s.replace("\t",r"\t") return _ustr(s) def oneOf( strs, caseless=False, useRegex=True ): """ Helper to quickly define a set of alternative Literals, and makes sure to do longest-first testing when there is a conflict, regardless of the input order, but returns a C{L{MatchFirst}} for best performance. Parameters: - strs - a string of space-delimited literals, or a collection of string literals - caseless - (default=C{False}) - treat all literals as caseless - useRegex - (default=C{True}) - as an optimization, will generate a Regex object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or if creating a C{Regex} raises an exception) Example:: comp_oper = oneOf("< = > <= >= !=") var = Word(alphas) number = Word(nums) term = var | number comparison_expr = term + comp_oper + term print(comparison_expr.searchString("B = 12 AA=23 B<=AA AA>12")) prints:: [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']] """ if caseless: isequal = ( lambda a,b: a.upper() == b.upper() ) masks = ( lambda a,b: b.upper().startswith(a.upper()) ) parseElementClass = CaselessLiteral else: isequal = ( lambda a,b: a == b ) masks = ( lambda a,b: b.startswith(a) ) parseElementClass = Literal symbols = [] if isinstance(strs,basestring): symbols = strs.split() elif isinstance(strs, Iterable): symbols = list(strs) else: warnings.warn("Invalid argument to oneOf, expected string or iterable", SyntaxWarning, stacklevel=2) if not symbols: return NoMatch() i = 0 while i < len(symbols)-1: cur = symbols[i] for j,other in enumerate(symbols[i+1:]): if ( isequal(other, cur) ): del symbols[i+j+1] break elif ( masks(cur, other) ): del symbols[i+j+1] symbols.insert(i,other) cur = other break else: i += 1 if not caseless and useRegex: #~ print (strs,"->", "|".join( [ _escapeRegexChars(sym) for sym in symbols] )) try: if len(symbols)==len("".join(symbols)): return Regex( "[%s]" % "".join(_escapeRegexRangeChars(sym) for sym in symbols) ).setName(' | '.join(symbols)) else: return Regex( "|".join(re.escape(sym) for sym in symbols) ).setName(' | '.join(symbols)) except Exception: warnings.warn("Exception creating Regex for oneOf, building MatchFirst", SyntaxWarning, stacklevel=2) # last resort, just use MatchFirst return MatchFirst(parseElementClass(sym) for sym in symbols).setName(' | '.join(symbols)) def dictOf( key, value ): """ Helper to easily and clearly define a dictionary by specifying the respective patterns for the key and value. Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens in the proper order. The key pattern can include delimiting markers or punctuation, as long as they are suppressed, thereby leaving the significant key text. The value pattern can include named results, so that the C{Dict} results can include named token fields. Example:: text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) print(OneOrMore(attr_expr).parseString(text).dump()) attr_label = label attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join) # similar to Dict, but simpler call format result = dictOf(attr_label, attr_value).parseString(text) print(result.dump()) print(result['shape']) print(result.shape) # object attribute access works too print(result.asDict()) prints:: [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE SQUARE {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'} """ return Dict( ZeroOrMore( Group ( key + value ) ) ) def originalTextFor(expr, asString=True): """ Helper to return the original, untokenized text for a given expression. Useful to restore the parsed fields of an HTML start tag into the raw tag text itself, or to revert separate tokens with intervening whitespace back to the original matching input text. By default, returns astring containing the original parsed text. If the optional C{asString} argument is passed as C{False}, then the return value is a C{L{ParseResults}} containing any results names that were originally matched, and a single token containing the original matched text from the input string. So if the expression passed to C{L{originalTextFor}} contains expressions with defined results names, you must set C{asString} to C{False} if you want to preserve those results name values. Example:: src = "this is test bold text normal text " for tag in ("b","i"): opener,closer = makeHTMLTags(tag) patt = originalTextFor(opener + SkipTo(closer) + closer) print(patt.searchString(src)[0]) prints:: [' bold text '] ['text'] """ locMarker = Empty().setParseAction(lambda s,loc,t: loc) endlocMarker = locMarker.copy() endlocMarker.callPreparse = False matchExpr = locMarker("_original_start") + expr + endlocMarker("_original_end") if asString: extractText = lambda s,l,t: s[t._original_start:t._original_end] else: def extractText(s,l,t): t[:] = [s[t.pop('_original_start'):t.pop('_original_end')]] matchExpr.setParseAction(extractText) matchExpr.ignoreExprs = expr.ignoreExprs return matchExpr def ungroup(expr): """ Helper to undo pyparsing's default grouping of And expressions, even if all but one are non-empty. """ return TokenConverter(expr).setParseAction(lambda t:t[0]) def locatedExpr(expr): """ Helper to decorate a returned token with its starting and ending locations in the input string. This helper adds the following results names: - locn_start = location where matched expression begins - locn_end = location where matched expression ends - value = the actual parsed results Be careful if the input text contains C{} characters, you may want to call C{L{ParserElement.parseWithTabs}} Example:: wd = Word(alphas) for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"): print(match) prints:: [[0, 'ljsdf', 5]] [[8, 'lksdjjf', 15]] [[18, 'lkkjj', 23]] """ locator = Empty().setParseAction(lambda s,l,t: l) return Group(locator("locn_start") + expr("value") + locator.copy().leaveWhitespace()("locn_end")) # convenience constants for positional expressions empty = Empty().setName("empty") lineStart = LineStart().setName("lineStart") lineEnd = LineEnd().setName("lineEnd") stringStart = StringStart().setName("stringStart") stringEnd = StringEnd().setName("stringEnd") _escapedPunc = Word( _bslash, r"\[]-*.$+^?()~ ", exact=2 ).setParseAction(lambda s,l,t:t[0][1]) _escapedHexChar = Regex(r"\\0?[xX][0-9a-fA-F]+").setParseAction(lambda s,l,t:unichr(int(t[0].lstrip(r'\0x'),16))) _escapedOctChar = Regex(r"\\0[0-7]+").setParseAction(lambda s,l,t:unichr(int(t[0][1:],8))) _singleChar = _escapedPunc | _escapedHexChar | _escapedOctChar | CharsNotIn(r'\]', exact=1) _charRange = Group(_singleChar + Suppress("-") + _singleChar) _reBracketExpr = Literal("[") + Optional("^").setResultsName("negate") + Group( OneOrMore( _charRange | _singleChar ) ).setResultsName("body") + "]" def srange(s): r""" Helper to easily define string ranges for use in Word construction. Borrows syntax from regexp '[]' string range definitions:: srange("[0-9]") -> "0123456789" srange("[a-z]") -> "abcdefghijklmnopqrstuvwxyz" srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_" The input string must be enclosed in []'s, and the returned string is the expanded character set joined into a single string. The values enclosed in the []'s may be: - a single character - an escaped character with a leading backslash (such as C{\-} or C{\]}) - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) (C{\0x##} is also supported for backwards compatibility) - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character) - a range of any of the above, separated by a dash (C{'a-z'}, etc.) - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.) """ _expanded = lambda p: p if not isinstance(p,ParseResults) else ''.join(unichr(c) for c in range(ord(p[0]),ord(p[1])+1)) try: return "".join(_expanded(part) for part in _reBracketExpr.parseString(s).body) except Exception: return "" def matchOnlyAtCol(n): """ Helper method for defining parse actions that require matching at a specific column in the input text. """ def verifyCol(strg,locn,toks): if col(locn,strg) != n: raise ParseException(strg,locn,"matched token not at column %d" % n) return verifyCol def replaceWith(replStr): """ Helper method for common parse actions that simply return a literal value. Especially useful when used with C{L{transformString}()}. Example:: num = Word(nums).setParseAction(lambda toks: int(toks[0])) na = oneOf("N/A NA").setParseAction(replaceWith(math.nan)) term = na | num OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234] """ return lambda s,l,t: [replStr] def removeQuotes(s,l,t): """ Helper parse action for removing quotation marks from parsed quoted strings. Example:: # by default, quotation marks are included in parsed results quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"] # use removeQuotes to strip quotation marks from parsed results quotedString.setParseAction(removeQuotes) quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"] """ return t[0][1:-1] def tokenMap(func, *args): """ Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional args are passed, they are forwarded to the given function as additional arguments after the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the parsed data to an integer using base 16. Example (compare the last to example in L{ParserElement.transformString}:: hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16)) hex_ints.runTests(''' 00 11 22 aa FF 0a 0d 1a ''') upperword = Word(alphas).setParseAction(tokenMap(str.upper)) OneOrMore(upperword).runTests(''' my kingdom for a horse ''') wd = Word(alphas).setParseAction(tokenMap(str.title)) OneOrMore(wd).setParseAction(' '.join).runTests(''' now is the winter of our discontent made glorious summer by this sun of york ''') prints:: 00 11 22 aa FF 0a 0d 1a [0, 17, 34, 170, 255, 10, 13, 26] my kingdom for a horse ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE'] now is the winter of our discontent made glorious summer by this sun of york ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York'] """ def pa(s,l,t): return [func(tokn, *args) for tokn in t] try: func_name = getattr(func, '__name__', getattr(func, '__class__').__name__) except Exception: func_name = str(func) pa.__name__ = func_name return pa upcaseTokens = tokenMap(lambda t: _ustr(t).upper()) """(Deprecated) Helper parse action to convert tokens to upper case. Deprecated in favor of L{pyparsing_common.upcaseTokens}""" downcaseTokens = tokenMap(lambda t: _ustr(t).lower()) """(Deprecated) Helper parse action to convert tokens to lower case. Deprecated in favor of L{pyparsing_common.downcaseTokens}""" def _makeTags(tagStr, xml): """Internal helper to construct opening and closing tag expressions, given a tag name""" if isinstance(tagStr,basestring): resname = tagStr tagStr = Keyword(tagStr, caseless=not xml) else: resname = tagStr.name tagAttrName = Word(alphas,alphanums+"_-:") if (xml): tagAttrValue = dblQuotedString.copy().setParseAction( removeQuotes ) openTag = Suppress("<") + tagStr("tag") + \ Dict(ZeroOrMore(Group( tagAttrName + Suppress("=") + tagAttrValue ))) + \ Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">") else: printablesLessRAbrack = "".join(c for c in printables if c not in ">") tagAttrValue = quotedString.copy().setParseAction( removeQuotes ) | Word(printablesLessRAbrack) openTag = Suppress("<") + tagStr("tag") + \ Dict(ZeroOrMore(Group( tagAttrName.setParseAction(downcaseTokens) + \ Optional( Suppress("=") + tagAttrValue ) ))) + \ Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">") closeTag = Combine(_L("") openTag = openTag.setResultsName("start"+"".join(resname.replace(":"," ").title().split())).setName("<%s>" % resname) closeTag = closeTag.setResultsName("end"+"".join(resname.replace(":"," ").title().split())).setName("" % resname) openTag.tag = resname closeTag.tag = resname return openTag, closeTag def makeHTMLTags(tagStr): """ Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values. Example:: text = 'More info at the pyparsing wiki page' # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple a,a_end = makeHTMLTags("A") link_expr = a + SkipTo(a_end)("link_text") + a_end for link in link_expr.searchString(text): # attributes in the tag (like "href" shown here) are also accessible as named results print(link.link_text, '->', link.href) prints:: pyparsing -> http://pyparsing.wikispaces.com """ return _makeTags( tagStr, False ) def makeXMLTags(tagStr): """ Helper to construct opening and closing tag expressions for XML, given a tag name. Matches tags only in the given upper/lower case. Example: similar to L{makeHTMLTags} """ return _makeTags( tagStr, True ) def withAttribute(*args,**attrDict): """ Helper to create a validating parse action to be used with start tags created with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag with a required attribute value, to avoid false matches on common tags such as C{} or C{
}. Call C{withAttribute} with a series of attribute names and values. Specify the list of filter attributes names and values as: - keyword arguments, as in C{(align="right")}, or - as an explicit dict with C{**} operator, when an attribute name is also a Python reserved word, as in C{**{"class":"Customer", "align":"right"}} - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") ) For attribute names with a namespace prefix, you must use the second form. Attribute names are matched insensitive to upper/lower case. If just testing for C{class} (with or without a namespace), use C{L{withClass}}. To verify that the attribute exists, but without specifying a value, pass C{withAttribute.ANY_VALUE} as the value. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this has no type
''' div,div_end = makeHTMLTags("div") # only match div tag having a type attribute with value "grid" div_grid = div().setParseAction(withAttribute(type="grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) # construct a match with any div tag having a type attribute, regardless of the value div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 """ if args: attrs = args[:] else: attrs = attrDict.items() attrs = [(k,v) for k,v in attrs] def pa(s,l,tokens): for attrName,attrValue in attrs: if attrName not in tokens: raise ParseException(s,l,"no matching attribute " + attrName) if attrValue != withAttribute.ANY_VALUE and tokens[attrName] != attrValue: raise ParseException(s,l,"attribute '%s' has value '%s', must be '%s'" % (attrName, tokens[attrName], attrValue)) return pa withAttribute.ANY_VALUE = object() def withClass(classname, namespace=''): """ Simplified version of C{L{withAttribute}} when matching on a div class - made difficult because C{class} is a reserved word in Python. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this <div> has no class
''' div,div_end = makeHTMLTags("div") div_grid = div().setParseAction(withClass("grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 """ classattr = "%s:class" % namespace if namespace else "class" return withAttribute(**{classattr : classname}) opAssoc = _Constants() opAssoc.LEFT = object() opAssoc.RIGHT = object() def infixNotation( baseExpr, opList, lpar=Suppress('('), rpar=Suppress(')') ): """ Helper method for constructing grammars of expressions made up of operators working in a precedence hierarchy. Operators may be unary or binary, left- or right-associative. Parse actions can also be attached to operator expressions. The generated parser will also recognize the use of parentheses to override operator precedences (see example below). Note: if you define a deep operator list, you may see performance issues when using infixNotation. See L{ParserElement.enablePackrat} for a mechanism to potentially improve your parser performance. Parameters: - baseExpr - expression representing the most basic element for the nested - opList - list of tuples, one for each operator precedence level in the expression grammar; each tuple is of the form (opExpr, numTerms, rightLeftAssoc, parseAction), where: - opExpr is the pyparsing expression for the operator; may also be a string, which will be converted to a Literal; if numTerms is 3, opExpr is a tuple of two expressions, for the two operators separating the 3 terms - numTerms is the number of terms for this operator (must be 1, 2, or 3) - rightLeftAssoc is the indicator whether the operator is right or left associative, using the pyparsing-defined constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}. - parseAction is the parse action to be associated with expressions matching this operator expression (the parse action tuple member may be omitted); if the parse action is passed a tuple or list of functions, this is equivalent to calling C{setParseAction(*fn)} (L{ParserElement.setParseAction}) - lpar - expression for matching left-parentheses (default=C{Suppress('(')}) - rpar - expression for matching right-parentheses (default=C{Suppress(')')}) Example:: # simple example of four-function arithmetic with ints and variable names integer = pyparsing_common.signed_integer varname = pyparsing_common.identifier arith_expr = infixNotation(integer | varname, [ ('-', 1, opAssoc.RIGHT), (oneOf('* /'), 2, opAssoc.LEFT), (oneOf('+ -'), 2, opAssoc.LEFT), ]) arith_expr.runTests(''' 5+3*6 (5+3)*6 -2--11 ''', fullDump=False) prints:: 5+3*6 [[5, '+', [3, '*', 6]]] (5+3)*6 [[[5, '+', 3], '*', 6]] -2--11 [[['-', 2], '-', ['-', 11]]] """ ret = Forward() lastExpr = baseExpr | ( lpar + ret + rpar ) for i,operDef in enumerate(opList): opExpr,arity,rightLeftAssoc,pa = (operDef + (None,))[:4] termName = "%s term" % opExpr if arity < 3 else "%s%s term" % opExpr if arity == 3: if opExpr is None or len(opExpr) != 2: raise ValueError("if numterms=3, opExpr must be a tuple or list of two expressions") opExpr1, opExpr2 = opExpr thisExpr = Forward().setName(termName) if rightLeftAssoc == opAssoc.LEFT: if arity == 1: matchExpr = FollowedBy(lastExpr + opExpr) + Group( lastExpr + OneOrMore( opExpr ) ) elif arity == 2: if opExpr is not None: matchExpr = FollowedBy(lastExpr + opExpr + lastExpr) + Group( lastExpr + OneOrMore( opExpr + lastExpr ) ) else: matchExpr = FollowedBy(lastExpr+lastExpr) + Group( lastExpr + OneOrMore(lastExpr) ) elif arity == 3: matchExpr = FollowedBy(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr) + \ Group( lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr ) else: raise ValueError("operator must be unary (1), binary (2), or ternary (3)") elif rightLeftAssoc == opAssoc.RIGHT: if arity == 1: # try to avoid LR with this extra test if not isinstance(opExpr, Optional): opExpr = Optional(opExpr) matchExpr = FollowedBy(opExpr.expr + thisExpr) + Group( opExpr + thisExpr ) elif arity == 2: if opExpr is not None: matchExpr = FollowedBy(lastExpr + opExpr + thisExpr) + Group( lastExpr + OneOrMore( opExpr + thisExpr ) ) else: matchExpr = FollowedBy(lastExpr + thisExpr) + Group( lastExpr + OneOrMore( thisExpr ) ) elif arity == 3: matchExpr = FollowedBy(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr) + \ Group( lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr ) else: raise ValueError("operator must be unary (1), binary (2), or ternary (3)") else: raise ValueError("operator must indicate right or left associativity") if pa: if isinstance(pa, (tuple, list)): matchExpr.setParseAction(*pa) else: matchExpr.setParseAction(pa) thisExpr <<= ( matchExpr.setName(termName) | lastExpr ) lastExpr = thisExpr ret <<= lastExpr return ret operatorPrecedence = infixNotation """(Deprecated) Former name of C{L{infixNotation}}, will be dropped in a future release.""" dblQuotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"').setName("string enclosed in double quotes") sglQuotedString = Combine(Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("string enclosed in single quotes") quotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"'| Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("quotedString using single or double quotes") unicodeString = Combine(_L('u') + quotedString.copy()).setName("unicode string literal") def nestedExpr(opener="(", closer=")", content=None, ignoreExpr=quotedString.copy()): """ Helper method for defining nested lists enclosed in opening and closing delimiters ("(" and ")" are the default). Parameters: - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression - content - expression for items within the nested lists (default=C{None}) - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString}) If an expression is not provided for the content argument, the nested expression will capture all whitespace-delimited content between delimiters as a list of separate values. Use the C{ignoreExpr} argument to define expressions that may contain opening or closing characters that should not be treated as opening or closing characters for nesting, such as quotedString or a comment expression. Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}. The default is L{quotedString}, but if no expressions are to be ignored, then pass C{None} for this argument. Example:: data_type = oneOf("void int short long char float double") decl_data_type = Combine(data_type + Optional(Word('*'))) ident = Word(alphas+'_', alphanums+'_') number = pyparsing_common.number arg = Group(decl_data_type + ident) LPAR,RPAR = map(Suppress, "()") code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment)) c_function = (decl_data_type("type") + ident("name") + LPAR + Optional(delimitedList(arg), [])("args") + RPAR + code_body("body")) c_function.ignore(cStyleComment) source_code = ''' int is_odd(int x) { return (x%2); } int dec_to_hex(char hchar) { if (hchar >= '0' && hchar <= '9') { return (ord(hchar)-ord('0')); } else { return (10+ord(hchar)-ord('A')); } } ''' for func in c_function.searchString(source_code): print("%(name)s (%(type)s) args: %(args)s" % func) prints:: is_odd (int) args: [['int', 'x']] dec_to_hex (int) args: [['char', 'hchar']] """ if opener == closer: raise ValueError("opening and closing strings cannot be the same") if content is None: if isinstance(opener,basestring) and isinstance(closer,basestring): if len(opener) == 1 and len(closer)==1: if ignoreExpr is not None: content = (Combine(OneOrMore(~ignoreExpr + CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS,exact=1)) ).setParseAction(lambda t:t[0].strip())) else: content = (empty.copy()+CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS ).setParseAction(lambda t:t[0].strip())) else: if ignoreExpr is not None: content = (Combine(OneOrMore(~ignoreExpr + ~Literal(opener) + ~Literal(closer) + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1)) ).setParseAction(lambda t:t[0].strip())) else: content = (Combine(OneOrMore(~Literal(opener) + ~Literal(closer) + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1)) ).setParseAction(lambda t:t[0].strip())) else: raise ValueError("opening and closing arguments must be strings if no content expression is given") ret = Forward() if ignoreExpr is not None: ret <<= Group( Suppress(opener) + ZeroOrMore( ignoreExpr | ret | content ) + Suppress(closer) ) else: ret <<= Group( Suppress(opener) + ZeroOrMore( ret | content ) + Suppress(closer) ) ret.setName('nested %s%s expression' % (opener,closer)) return ret def indentedBlock(blockStatementExpr, indentStack, indent=True): """ Helper method for defining space-delimited indentation blocks, such as those used to define block statements in Python source code. Parameters: - blockStatementExpr - expression defining syntax of statement that is repeated within the indented block - indentStack - list created by caller to manage indentation stack (multiple statementWithIndentedBlock expressions within a single grammar should share a common indentStack) - indent - boolean indicating whether block must be indented beyond the the current level; set to False for block of left-most statements (default=C{True}) A valid block must contain at least one C{blockStatement}. Example:: data = ''' def A(z): A1 B = 100 G = A2 A2 A3 B def BB(a,b,c): BB1 def BBA(): bba1 bba2 bba3 C D def spam(x,y): def eggs(z): pass ''' indentStack = [1] stmt = Forward() identifier = Word(alphas, alphanums) funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":") func_body = indentedBlock(stmt, indentStack) funcDef = Group( funcDecl + func_body ) rvalue = Forward() funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")") rvalue << (funcCall | identifier | Word(nums)) assignment = Group(identifier + "=" + rvalue) stmt << ( funcDef | assignment | identifier ) module_body = OneOrMore(stmt) parseTree = module_body.parseString(data) parseTree.pprint() prints:: [['def', 'A', ['(', 'z', ')'], ':', [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]], 'B', ['def', 'BB', ['(', 'a', 'b', 'c', ')'], ':', [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]], 'C', 'D', ['def', 'spam', ['(', 'x', 'y', ')'], ':', [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] """ def checkPeerIndent(s,l,t): if l >= len(s): return curCol = col(l,s) if curCol != indentStack[-1]: if curCol > indentStack[-1]: raise ParseFatalException(s,l,"illegal nesting") raise ParseException(s,l,"not a peer entry") def checkSubIndent(s,l,t): curCol = col(l,s) if curCol > indentStack[-1]: indentStack.append( curCol ) else: raise ParseException(s,l,"not a subentry") def checkUnindent(s,l,t): if l >= len(s): return curCol = col(l,s) if not(indentStack and curCol < indentStack[-1] and curCol <= indentStack[-2]): raise ParseException(s,l,"not an unindent") indentStack.pop() NL = OneOrMore(LineEnd().setWhitespaceChars("\t ").suppress()) INDENT = (Empty() + Empty().setParseAction(checkSubIndent)).setName('INDENT') PEER = Empty().setParseAction(checkPeerIndent).setName('') UNDENT = Empty().setParseAction(checkUnindent).setName('UNINDENT') if indent: smExpr = Group( Optional(NL) + #~ FollowedBy(blockStatementExpr) + INDENT + (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) + UNDENT) else: smExpr = Group( Optional(NL) + (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) ) blockStatementExpr.ignore(_bslash + LineEnd()) return smExpr.setName('indented block') alphas8bit = srange(r"[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]") punc8bit = srange(r"[\0xa1-\0xbf\0xd7\0xf7]") anyOpenTag,anyCloseTag = makeHTMLTags(Word(alphas,alphanums+"_:").setName('any tag')) _htmlEntityMap = dict(zip("gt lt amp nbsp quot apos".split(),'><& "\'')) commonHTMLEntity = Regex('&(?P' + '|'.join(_htmlEntityMap.keys()) +");").setName("common HTML entity") def replaceHTMLEntity(t): """Helper parser action to replace common HTML entities with their special characters""" return _htmlEntityMap.get(t.entity) # it's easy to get these comment structures wrong - they're very common, so may as well make them available cStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/').setName("C style comment") "Comment of the form C{/* ... */}" htmlComment = Regex(r"").setName("HTML comment") "Comment of the form C{}" restOfLine = Regex(r".*").leaveWhitespace().setName("rest of line") dblSlashComment = Regex(r"//(?:\\\n|[^\n])*").setName("// comment") "Comment of the form C{// ... (to end of line)}" cppStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/'| dblSlashComment).setName("C++ style comment") "Comment of either form C{L{cStyleComment}} or C{L{dblSlashComment}}" javaStyleComment = cppStyleComment "Same as C{L{cppStyleComment}}" pythonStyleComment = Regex(r"#.*").setName("Python style comment") "Comment of the form C{# ... (to end of line)}" _commasepitem = Combine(OneOrMore(Word(printables, excludeChars=',') + Optional( Word(" \t") + ~Literal(",") + ~LineEnd() ) ) ).streamline().setName("commaItem") commaSeparatedList = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("commaSeparatedList") """(Deprecated) Predefined expression of 1 or more printable words or quoted strings, separated by commas. This expression is deprecated in favor of L{pyparsing_common.comma_separated_list}.""" # some other useful expressions - using lower-case class name since we are really using this as a namespace class pyparsing_common: """ Here are some common low-level expressions that may be useful in jump-starting parser development: - numeric forms (L{integers}, L{reals}, L{scientific notation}) - common L{programming identifiers} - network addresses (L{MAC}, L{IPv4}, L{IPv6}) - ISO8601 L{dates} and L{datetime} - L{UUID} - L{comma-separated list} Parse actions: - C{L{convertToInteger}} - C{L{convertToFloat}} - C{L{convertToDate}} - C{L{convertToDatetime}} - C{L{stripHTMLTags}} - C{L{upcaseTokens}} - C{L{downcaseTokens}} Example:: pyparsing_common.number.runTests(''' # any int or real number, returned as the appropriate type 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.fnumber.runTests(''' # any int or real number, returned as float 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.hex_integer.runTests(''' # hex numbers 100 FF ''') pyparsing_common.fraction.runTests(''' # fractions 1/2 -3/4 ''') pyparsing_common.mixed_integer.runTests(''' # mixed fractions 1 1/2 -3/4 1-3/4 ''') import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(''' # uuid 12345678-1234-5678-1234-567812345678 ''') prints:: # any int or real number, returned as the appropriate type 100 [100] -100 [-100] +100 [100] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # any int or real number, returned as float 100 [100.0] -100 [-100.0] +100 [100.0] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # hex numbers 100 [256] FF [255] # fractions 1/2 [0.5] -3/4 [-0.75] # mixed fractions 1 [1] 1/2 [0.5] -3/4 [-0.75] 1-3/4 [1.75] # uuid 12345678-1234-5678-1234-567812345678 [UUID('12345678-1234-5678-1234-567812345678')] """ convertToInteger = tokenMap(int) """ Parse action for converting parsed integers to Python int """ convertToFloat = tokenMap(float) """ Parse action for converting parsed numbers to Python float """ integer = Word(nums).setName("integer").setParseAction(convertToInteger) """expression that parses an unsigned integer, returns an int""" hex_integer = Word(hexnums).setName("hex integer").setParseAction(tokenMap(int,16)) """expression that parses a hexadecimal integer, returns an int""" signed_integer = Regex(r'[+-]?\d+').setName("signed integer").setParseAction(convertToInteger) """expression that parses an integer with optional leading sign, returns an int""" fraction = (signed_integer().setParseAction(convertToFloat) + '/' + signed_integer().setParseAction(convertToFloat)).setName("fraction") """fractional expression of an integer divided by an integer, returns a float""" fraction.addParseAction(lambda t: t[0]/t[-1]) mixed_integer = (fraction | signed_integer + Optional(Optional('-').suppress() + fraction)).setName("fraction or mixed integer-fraction") """mixed integer of the form 'integer - fraction', with optional leading integer, returns float""" mixed_integer.addParseAction(sum) real = Regex(r'[+-]?\d+\.\d*').setName("real number").setParseAction(convertToFloat) """expression that parses a floating point number and returns a float""" sci_real = Regex(r'[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)').setName("real number with scientific notation").setParseAction(convertToFloat) """expression that parses a floating point number with optional scientific notation and returns a float""" # streamlining this expression makes the docs nicer-looking number = (sci_real | real | signed_integer).streamline() """any numeric expression, returns the corresponding Python type""" fnumber = Regex(r'[+-]?\d+\.?\d*([eE][+-]?\d+)?').setName("fnumber").setParseAction(convertToFloat) """any int or real number, returned as float""" identifier = Word(alphas+'_', alphanums+'_').setName("identifier") """typical code identifier (leading alpha or '_', followed by 0 or more alphas, nums, or '_')""" ipv4_address = Regex(r'(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}').setName("IPv4 address") "IPv4 address (C{0.0.0.0 - 255.255.255.255})" _ipv6_part = Regex(r'[0-9a-fA-F]{1,4}').setName("hex_integer") _full_ipv6_address = (_ipv6_part + (':' + _ipv6_part)*7).setName("full IPv6 address") _short_ipv6_address = (Optional(_ipv6_part + (':' + _ipv6_part)*(0,6)) + "::" + Optional(_ipv6_part + (':' + _ipv6_part)*(0,6))).setName("short IPv6 address") _short_ipv6_address.addCondition(lambda t: sum(1 for tt in t if pyparsing_common._ipv6_part.matches(tt)) < 8) _mixed_ipv6_address = ("::ffff:" + ipv4_address).setName("mixed IPv6 address") ipv6_address = Combine((_full_ipv6_address | _mixed_ipv6_address | _short_ipv6_address).setName("IPv6 address")).setName("IPv6 address") "IPv6 address (long, short, or mixed form)" mac_address = Regex(r'[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}').setName("MAC address") "MAC address xx:xx:xx:xx:xx (may also have '-' or '.' delimiters)" @staticmethod def convertToDate(fmt="%Y-%m-%d"): """ Helper to create a parse action for converting parsed date string to Python datetime.date Params - - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"}) Example:: date_expr = pyparsing_common.iso8601_date.copy() date_expr.setParseAction(pyparsing_common.convertToDate()) print(date_expr.parseString("1999-12-31")) prints:: [datetime.date(1999, 12, 31)] """ def cvt_fn(s,l,t): try: return datetime.strptime(t[0], fmt).date() except ValueError as ve: raise ParseException(s, l, str(ve)) return cvt_fn @staticmethod def convertToDatetime(fmt="%Y-%m-%dT%H:%M:%S.%f"): """ Helper to create a parse action for converting parsed datetime string to Python datetime.datetime Params - - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"}) Example:: dt_expr = pyparsing_common.iso8601_datetime.copy() dt_expr.setParseAction(pyparsing_common.convertToDatetime()) print(dt_expr.parseString("1999-12-31T23:59:59.999")) prints:: [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)] """ def cvt_fn(s,l,t): try: return datetime.strptime(t[0], fmt) except ValueError as ve: raise ParseException(s, l, str(ve)) return cvt_fn iso8601_date = Regex(r'(?P\d{4})(?:-(?P\d\d)(?:-(?P\d\d))?)?').setName("ISO8601 date") "ISO8601 date (C{yyyy-mm-dd})" iso8601_datetime = Regex(r'(?P\d{4})-(?P\d\d)-(?P\d\d)[T ](?P\d\d):(?P\d\d)(:(?P\d\d(\.\d*)?)?)?(?PZ|[+-]\d\d:?\d\d)?').setName("ISO8601 datetime") "ISO8601 datetime (C{yyyy-mm-ddThh:mm:ss.s(Z|+-00:00)}) - trailing seconds, milliseconds, and timezone optional; accepts separating C{'T'} or C{' '}" uuid = Regex(r'[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}').setName("UUID") "UUID (C{xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx})" _html_stripper = anyOpenTag.suppress() | anyCloseTag.suppress() @staticmethod def stripHTMLTags(s, l, tokens): """ Parse action to remove HTML tags from web page HTML source Example:: # strip HTML links from normal text text = 'More info at the
pyparsing wiki page' td,td_end = makeHTMLTags("TD") table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page' """ return pyparsing_common._html_stripper.transformString(tokens[0]) _commasepitem = Combine(OneOrMore(~Literal(",") + ~LineEnd() + Word(printables, excludeChars=',') + Optional( White(" \t") ) ) ).streamline().setName("commaItem") comma_separated_list = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("comma separated list") """Predefined expression of 1 or more printable words or quoted strings, separated by commas.""" upcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).upper())) """Parse action to convert tokens to upper case.""" downcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).lower())) """Parse action to convert tokens to lower case.""" if __name__ == "__main__": selectToken = CaselessLiteral("select") fromToken = CaselessLiteral("from") ident = Word(alphas, alphanums + "_$") columnName = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens) columnNameList = Group(delimitedList(columnName)).setName("columns") columnSpec = ('*' | columnNameList) tableName = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens) tableNameList = Group(delimitedList(tableName)).setName("tables") simpleSQL = selectToken("command") + columnSpec("columns") + fromToken + tableNameList("tables") # demo runTests method, including embedded comments in test string simpleSQL.runTests(""" # '*' as column list and dotted table name select * from SYS.XYZZY # caseless match on "SELECT", and casts back to "select" SELECT * from XYZZY, ABC # list of column names, and mixed case SELECT keyword Select AA,BB,CC from Sys.dual # multiple tables Select A, B, C from Sys.dual, Table2 # invalid SELECT keyword - should fail Xelect A, B, C from Sys.dual # incomplete command - should fail Select # invalid column name - should fail Select ^^^ frox Sys.dual """) pyparsing_common.number.runTests(""" 100 -100 +100 3.14159 6.02e23 1e-12 """) # any int or real number, returned as float pyparsing_common.fnumber.runTests(""" 100 -100 +100 3.14159 6.02e23 1e-12 """) pyparsing_common.hex_integer.runTests(""" 100 FF """) import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(""" 12345678-1234-5678-1234-567812345678 """) site-packages/setuptools/_vendor/__init__.py000064400000000000151464452710015251 0ustar00site-packages/setuptools/script (dev).tmpl000064400000000332151464452710014576 0ustar00# EASY-INSTALL-DEV-SCRIPT: %(spec)r,%(script_name)r __requires__ = %(spec)r __import__('pkg_resources').require(%(spec)r) __file__ = %(dev_path)r with open(__file__) as f: exec(compile(f.read(), __file__, 'exec')) site-packages/setuptools/namespaces.py000064400000006177151464452710014222 0ustar00import os from distutils import log import itertools from setuptools.extern.six.moves import map flatten = itertools.chain.from_iterable class Installer: nspkg_ext = '-nspkg.pth' def install_namespaces(self): nsp = self._get_all_ns_packages() if not nsp: return filename, ext = os.path.splitext(self._get_target()) filename += self.nspkg_ext self.outputs.append(filename) log.info("Installing %s", filename) lines = map(self._gen_nspkg_line, nsp) if self.dry_run: # always generate the lines, even in dry run list(lines) return with open(filename, 'wt') as f: f.writelines(lines) def uninstall_namespaces(self): filename, ext = os.path.splitext(self._get_target()) filename += self.nspkg_ext if not os.path.exists(filename): return log.info("Removing %s", filename) os.remove(filename) def _get_target(self): return self.target _nspkg_tmpl = ( "import sys, types, os", "has_mfs = sys.version_info > (3, 5)", "p = os.path.join(%(root)s, *%(pth)r)", "importlib = has_mfs and __import__('importlib.util')", "has_mfs and __import__('importlib.machinery')", "m = has_mfs and " "sys.modules.setdefault(%(pkg)r, " "importlib.util.module_from_spec(" "importlib.machinery.PathFinder.find_spec(%(pkg)r, " "[os.path.dirname(p)])))", "m = m or " "sys.modules.setdefault(%(pkg)r, types.ModuleType(%(pkg)r))", "mp = (m or []) and m.__dict__.setdefault('__path__',[])", "(p not in mp) and mp.append(p)", ) "lines for the namespace installer" _nspkg_tmpl_multi = ( 'm and setattr(sys.modules[%(parent)r], %(child)r, m)', ) "additional line(s) when a parent package is indicated" def _get_root(self): return "sys._getframe(1).f_locals['sitedir']" def _gen_nspkg_line(self, pkg): # ensure pkg is not a unicode string under Python 2.7 pkg = str(pkg) pth = tuple(pkg.split('.')) root = self._get_root() tmpl_lines = self._nspkg_tmpl parent, sep, child = pkg.rpartition('.') if parent: tmpl_lines += self._nspkg_tmpl_multi return ';'.join(tmpl_lines) % locals() + '\n' def _get_all_ns_packages(self): """Return sorted list of all package namespaces""" pkgs = self.distribution.namespace_packages or [] return sorted(flatten(map(self._pkg_names, pkgs))) @staticmethod def _pkg_names(pkg): """ Given a namespace package, yield the components of that package. >>> names = Installer._pkg_names('a.b.c') >>> set(names) == set(['a', 'a.b', 'a.b.c']) True """ parts = pkg.split('.') while parts: yield '.'.join(parts) parts.pop() class DevelopInstaller(Installer): def _get_root(self): return repr(str(self.egg_path)) def _get_target(self): return self.egg_link site-packages/setuptools/depends.py000064400000012615151464452710013517 0ustar00import sys import marshal import contextlib from distutils.version import StrictVersion from .py33compat import Bytecode from .py27compat import find_module, PY_COMPILED, PY_FROZEN, PY_SOURCE from . import py27compat __all__ = [ 'Require', 'find_module', 'get_module_constant', 'extract_constant' ] class Require: """A prerequisite to building or installing a distribution""" def __init__( self, name, requested_version, module, homepage='', attribute=None, format=None): if format is None and requested_version is not None: format = StrictVersion if format is not None: requested_version = format(requested_version) if attribute is None: attribute = '__version__' self.__dict__.update(locals()) del self.self def full_name(self): """Return full package/distribution name, w/version""" if self.requested_version is not None: return '%s-%s' % (self.name, self.requested_version) return self.name def version_ok(self, version): """Is 'version' sufficiently up-to-date?""" return self.attribute is None or self.format is None or \ str(version) != "unknown" and version >= self.requested_version def get_version(self, paths=None, default="unknown"): """Get version number of installed module, 'None', or 'default' Search 'paths' for module. If not found, return 'None'. If found, return the extracted version attribute, or 'default' if no version attribute was specified, or the value cannot be determined without importing the module. The version is formatted according to the requirement's version format (if any), unless it is 'None' or the supplied 'default'. """ if self.attribute is None: try: f, p, i = find_module(self.module, paths) if f: f.close() return default except ImportError: return None v = get_module_constant(self.module, self.attribute, default, paths) if v is not None and v is not default and self.format is not None: return self.format(v) return v def is_present(self, paths=None): """Return true if dependency is present on 'paths'""" return self.get_version(paths) is not None def is_current(self, paths=None): """Return true if dependency is present and up-to-date on 'paths'""" version = self.get_version(paths) if version is None: return False return self.version_ok(version) def maybe_close(f): @contextlib.contextmanager def empty(): yield return if not f: return empty() return contextlib.closing(f) def get_module_constant(module, symbol, default=-1, paths=None): """Find 'module' by searching 'paths', and extract 'symbol' Return 'None' if 'module' does not exist on 'paths', or it does not define 'symbol'. If the module defines 'symbol' as a constant, return the constant. Otherwise, return 'default'.""" try: f, path, (suffix, mode, kind) = info = find_module(module, paths) except ImportError: # Module doesn't exist return None with maybe_close(f): if kind == PY_COMPILED: f.read(8) # skip magic & date code = marshal.load(f) elif kind == PY_FROZEN: code = py27compat.get_frozen_object(module, paths) elif kind == PY_SOURCE: code = compile(f.read(), path, 'exec') else: # Not something we can parse; we'll have to import it. :( imported = py27compat.get_module(module, paths, info) return getattr(imported, symbol, None) return extract_constant(code, symbol, default) def extract_constant(code, symbol, default=-1): """Extract the constant value of 'symbol' from 'code' If the name 'symbol' is bound to a constant value by the Python code object 'code', return that value. If 'symbol' is bound to an expression, return 'default'. Otherwise, return 'None'. Return value is based on the first assignment to 'symbol'. 'symbol' must be a global, or at least a non-"fast" local in the code block. That is, only 'STORE_NAME' and 'STORE_GLOBAL' opcodes are checked, and 'symbol' must be present in 'code.co_names'. """ if symbol not in code.co_names: # name's not there, can't possibly be an assignment return None name_idx = list(code.co_names).index(symbol) STORE_NAME = 90 STORE_GLOBAL = 97 LOAD_CONST = 100 const = default for byte_code in Bytecode(code): op = byte_code.opcode arg = byte_code.arg if op == LOAD_CONST: const = code.co_consts[arg] elif arg == name_idx and (op == STORE_NAME or op == STORE_GLOBAL): return const else: const = default def _update_globals(): """ Patch the globals to remove the objects not available on some platforms. XXX it'd be better to test assertions about bytecode instead. """ if not sys.platform.startswith('java') and sys.platform != 'cli': return incompatible = 'extract_constant', 'get_module_constant' for name in incompatible: del globals()[name] __all__.remove(name) _update_globals() site-packages/setuptools/__init__.py000064400000016163151464452710013636 0ustar00"""Extensions to the 'distutils' for large or complex distributions""" import os import sys import functools import distutils.core import distutils.filelist import re from distutils.errors import DistutilsOptionError from distutils.util import convert_path from fnmatch import fnmatchcase from ._deprecation_warning import SetuptoolsDeprecationWarning from setuptools.extern.six import PY3, string_types from setuptools.extern.six.moves import filter, map import setuptools.version from setuptools.extension import Extension from setuptools.dist import Distribution, Feature from setuptools.depends import Require from . import monkey __metaclass__ = type __all__ = [ 'setup', 'Distribution', 'Feature', 'Command', 'Extension', 'Require', 'SetuptoolsDeprecationWarning', 'find_packages' ] if PY3: __all__.append('find_namespace_packages') __version__ = setuptools.version.__version__ bootstrap_install_from = None # If we run 2to3 on .py files, should we also convert docstrings? # Default: yes; assume that we can detect doctests reliably run_2to3_on_doctests = True # Standard package names for fixer packages lib2to3_fixer_packages = ['lib2to3.fixes'] class PackageFinder: """ Generate a list of all Python packages found within a directory """ @classmethod def find(cls, where='.', exclude=(), include=('*',)): """Return a list all Python packages found within directory 'where' 'where' is the root directory which will be searched for packages. It should be supplied as a "cross-platform" (i.e. URL-style) path; it will be converted to the appropriate local path syntax. 'exclude' is a sequence of package names to exclude; '*' can be used as a wildcard in the names, such that 'foo.*' will exclude all subpackages of 'foo' (but not 'foo' itself). 'include' is a sequence of package names to include. If it's specified, only the named packages will be included. If it's not specified, all found packages will be included. 'include' can contain shell style wildcard patterns just like 'exclude'. """ return list(cls._find_packages_iter( convert_path(where), cls._build_filter('ez_setup', '*__pycache__', *exclude), cls._build_filter(*include))) @classmethod def _find_packages_iter(cls, where, exclude, include): """ All the packages found in 'where' that pass the 'include' filter, but not the 'exclude' filter. """ for root, dirs, files in os.walk(where, followlinks=True): # Copy dirs to iterate over it, then empty dirs. all_dirs = dirs[:] dirs[:] = [] for dir in all_dirs: full_path = os.path.join(root, dir) rel_path = os.path.relpath(full_path, where) package = rel_path.replace(os.path.sep, '.') # Skip directory trees that are not valid packages if ('.' in dir or not cls._looks_like_package(full_path)): continue # Should this package be included? if include(package) and not exclude(package): yield package # Keep searching subdirectories, as there may be more packages # down there, even if the parent was excluded. dirs.append(dir) @staticmethod def _looks_like_package(path): """Does a directory look like a package?""" return os.path.isfile(os.path.join(path, '__init__.py')) @staticmethod def _build_filter(*patterns): """ Given a list of patterns, return a callable that will be true only if the input matches at least one of the patterns. """ return lambda name: any(fnmatchcase(name, pat=pat) for pat in patterns) class PEP420PackageFinder(PackageFinder): @staticmethod def _looks_like_package(path): return True find_packages = PackageFinder.find if PY3: find_namespace_packages = PEP420PackageFinder.find def _install_setup_requires(attrs): # Note: do not use `setuptools.Distribution` directly, as # our PEP 517 backend patch `distutils.core.Distribution`. dist = distutils.core.Distribution(dict( (k, v) for k, v in attrs.items() if k in ('dependency_links', 'setup_requires') )) # Honor setup.cfg's options. dist.parse_config_files(ignore_option_errors=True) if dist.setup_requires: dist.fetch_build_eggs(dist.setup_requires) def setup(**attrs): # Make sure we have any requirements needed to interpret 'attrs'. _install_setup_requires(attrs) return distutils.core.setup(**attrs) setup.__doc__ = distutils.core.setup.__doc__ _Command = monkey.get_unpatched(distutils.core.Command) class Command(_Command): __doc__ = _Command.__doc__ command_consumes_arguments = False def __init__(self, dist, **kw): """ Construct the command for dist, updating vars(self) with any keyword parameters. """ _Command.__init__(self, dist) vars(self).update(kw) def _ensure_stringlike(self, option, what, default=None): val = getattr(self, option) if val is None: setattr(self, option, default) return default elif not isinstance(val, string_types): raise DistutilsOptionError("'%s' must be a %s (got `%s`)" % (option, what, val)) return val def ensure_string_list(self, option): r"""Ensure that 'option' is a list of strings. If 'option' is currently a string, we split it either on /,\s*/ or /\s+/, so "foo bar baz", "foo,bar,baz", and "foo, bar baz" all become ["foo", "bar", "baz"]. """ val = getattr(self, option) if val is None: return elif isinstance(val, string_types): setattr(self, option, re.split(r',\s*|\s+', val)) else: if isinstance(val, list): ok = all(isinstance(v, string_types) for v in val) else: ok = False if not ok: raise DistutilsOptionError( "'%s' must be a list of strings (got %r)" % (option, val)) def reinitialize_command(self, command, reinit_subcommands=0, **kw): cmd = _Command.reinitialize_command(self, command, reinit_subcommands) vars(cmd).update(kw) return cmd def _find_all_simple(path): """ Find all files under 'path' """ results = ( os.path.join(base, file) for base, dirs, files in os.walk(path, followlinks=True) for file in files ) return filter(os.path.isfile, results) def findall(dir=os.curdir): """ Find all files under 'dir' and return the list of full filenames. Unless dir is '.', return full filenames with dir prepended. """ files = _find_all_simple(dir) if dir == os.curdir: make_rel = functools.partial(os.path.relpath, start=dir) files = map(make_rel, files) return list(files) # Apply monkey patches monkey.patch_all() site-packages/setuptools/extern/__pycache__/__init__.cpython-38.pyc000064400000004564151464452710021433 0ustar00U Qab @s.ddlZGdddZdZeeeddS)Nc@s@eZdZdZdddZeddZddd Zd d Zd d Z dS)VendorImporterz A PEP 302 meta path importer for finding optionally-vendored or otherwise naturally-installed packages from root_name. NcCs&||_t||_|p|dd|_dS)NZexternZ_vendor) root_namesetvendored_namesreplace vendor_pkg)selfrrrrr>/usr/lib/python3.8/site-packages/setuptools/extern/__init__.py__init__ s zVendorImporter.__init__ccs|jdVdVdS)zL Search first the vendor package then as a natural package. .N)rr rrr search_paths zVendorImporter.search_pathcCs8||jd\}}}|rdStt|j|js4dS|S)z Return self when fullname starts with root_name and the target module is one vendored through this importer. r N) partitionranymap startswithr)r fullnamepathrootbasetargetrrr find_modules zVendorImporter.find_modulec Cs||jd\}}}|jD]Z}z@||}t|tj|}|tj|<tjdkrXtj|=|WStk rtYqXqtdjft dS)zK Iterate over the search path to locate and load fullname. r )zThe '{target}' package is required; normally this is bundled with this package so if you get this warning, consult the packager of your distribution.N) rrr __import__sysmodules version_info ImportErrorformatlocals)r rrrrprefixZextantmodrrr load_module#s"     zVendorImporter.load_modulecCs|tjkrtj|dS)zR Install this importer into sys.meta_path if not already present. N)r meta_pathappendrrrr install@s zVendorImporter.install)rN)N) __name__ __module__ __qualname____doc__r propertyrrr$r'rrrr rs   r)ZsixZ packagingZ pyparsingZ ordered_setzsetuptools._vendor)rrnamesr(r'rrrr sDsite-packages/setuptools/extern/__pycache__/__init__.cpython-38.opt-1.pyc000064400000004564151464452710022372 0ustar00U Qab @s.ddlZGdddZdZeeeddS)Nc@s@eZdZdZdddZeddZddd Zd d Zd d Z dS)VendorImporterz A PEP 302 meta path importer for finding optionally-vendored or otherwise naturally-installed packages from root_name. NcCs&||_t||_|p|dd|_dS)NZexternZ_vendor) root_namesetvendored_namesreplace vendor_pkg)selfrrrrr>/usr/lib/python3.8/site-packages/setuptools/extern/__init__.py__init__ s zVendorImporter.__init__ccs|jdVdVdS)zL Search first the vendor package then as a natural package. .N)rr rrr search_paths zVendorImporter.search_pathcCs8||jd\}}}|rdStt|j|js4dS|S)z Return self when fullname starts with root_name and the target module is one vendored through this importer. r N) partitionranymap startswithr)r fullnamepathrootbasetargetrrr find_modules zVendorImporter.find_modulec Cs||jd\}}}|jD]Z}z@||}t|tj|}|tj|<tjdkrXtj|=|WStk rtYqXqtdjft dS)zK Iterate over the search path to locate and load fullname. r )zThe '{target}' package is required; normally this is bundled with this package so if you get this warning, consult the packager of your distribution.N) rrr __import__sysmodules version_info ImportErrorformatlocals)r rrrrprefixZextantmodrrr load_module#s"     zVendorImporter.load_modulecCs|tjkrtj|dS)zR Install this importer into sys.meta_path if not already present. N)r meta_pathappendrrrr install@s zVendorImporter.install)rN)N) __name__ __module__ __qualname____doc__r propertyrrr$r'rrrr rs   r)ZsixZ packagingZ pyparsingZ ordered_setzsetuptools._vendor)rrnamesr(r'rrrr sDsite-packages/setuptools/extern/__init__.py000064400000004722151464452710015141 0ustar00import sys class VendorImporter: """ A PEP 302 meta path importer for finding optionally-vendored or otherwise naturally-installed packages from root_name. """ def __init__(self, root_name, vendored_names=(), vendor_pkg=None): self.root_name = root_name self.vendored_names = set(vendored_names) self.vendor_pkg = vendor_pkg or root_name.replace('extern', '_vendor') @property def search_path(self): """ Search first the vendor package then as a natural package. """ yield self.vendor_pkg + '.' yield '' def find_module(self, fullname, path=None): """ Return self when fullname starts with root_name and the target module is one vendored through this importer. """ root, base, target = fullname.partition(self.root_name + '.') if root: return if not any(map(target.startswith, self.vendored_names)): return return self def load_module(self, fullname): """ Iterate over the search path to locate and load fullname. """ root, base, target = fullname.partition(self.root_name + '.') for prefix in self.search_path: try: extant = prefix + target __import__(extant) mod = sys.modules[extant] sys.modules[fullname] = mod # mysterious hack: # Remove the reference to the extant package/module # on later Python versions to cause relative imports # in the vendor package to resolve the same modules # as those going through this importer. if sys.version_info >= (3, ): del sys.modules[extant] return mod except ImportError: pass else: raise ImportError( "The '{target}' package is required; " "normally this is bundled with this package so if you get " "this warning, consult the packager of your " "distribution.".format(**locals()) ) def install(self): """ Install this importer into sys.meta_path if not already present. """ if self not in sys.meta_path: sys.meta_path.append(self) names = 'six', 'packaging', 'pyparsing', 'ordered_set', VendorImporter(__name__, names, 'setuptools._vendor').install() site-packages/setuptools/launch.py000064400000001423151464452710013342 0ustar00""" Launch the Python script on the command line after setuptools is bootstrapped via import. """ # Note that setuptools gets imported implicitly by the # invocation of this script using python -m setuptools.launch import tokenize import sys def run(): """ Run the script in sys.argv[1] as if it had been invoked naturally. """ __builtins__ script_name = sys.argv[1] namespace = dict( __file__=script_name, __name__='__main__', __doc__=None, ) sys.argv[:] = sys.argv[1:] open_ = getattr(tokenize, 'open', open) script = open_(script_name).read() norm_script = script.replace('\\r\\n', '\\n') code = compile(norm_script, script_name, 'exec') exec(code, namespace) if __name__ == '__main__': run() site-packages/pip/_internal/pep425tags.py000064400000037105151464452710014316 0ustar00"""Generate and work with PEP 425 Compatibility Tags.""" from __future__ import absolute_import import distutils.util import logging import platform import re import sys import sysconfig import warnings from collections import OrderedDict import pip._internal.utils.glibc from pip._internal.utils.compat import get_extension_suffixes from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ( Tuple, Callable, List, Optional, Union, Dict, Set ) Pep425Tag = Tuple[str, str, str] logger = logging.getLogger(__name__) _osx_arch_pat = re.compile(r'(.+)_(\d+)_(\d+)_(.+)') def get_config_var(var): # type: (str) -> Optional[str] try: return sysconfig.get_config_var(var) except IOError as e: # Issue #1074 warnings.warn("{}".format(e), RuntimeWarning) return None def get_abbr_impl(): # type: () -> str """Return abbreviated implementation name.""" if hasattr(sys, 'pypy_version_info'): pyimpl = 'pp' elif sys.platform.startswith('java'): pyimpl = 'jy' elif sys.platform == 'cli': pyimpl = 'ip' else: pyimpl = 'cp' return pyimpl def version_info_to_nodot(version_info): # type: (Tuple[int, ...]) -> str # Only use up to the first two numbers. return ''.join(map(str, version_info[:2])) def get_impl_ver(): # type: () -> str """Return implementation version.""" impl_ver = get_config_var("py_version_nodot") if not impl_ver or get_abbr_impl() == 'pp': impl_ver = ''.join(map(str, get_impl_version_info())) return impl_ver def get_impl_version_info(): # type: () -> Tuple[int, ...] """Return sys.version_info-like tuple for use in decrementing the minor version.""" if get_abbr_impl() == 'pp': # as per https://github.com/pypa/pip/issues/2882 # attrs exist only on pypy return (sys.version_info[0], sys.pypy_version_info.major, # type: ignore sys.pypy_version_info.minor) # type: ignore else: return sys.version_info[0], sys.version_info[1] def get_impl_tag(): # type: () -> str """ Returns the Tag for this specific implementation. """ return "{}{}".format(get_abbr_impl(), get_impl_ver()) def get_flag(var, fallback, expected=True, warn=True): # type: (str, Callable[..., bool], Union[bool, int], bool) -> bool """Use a fallback method for determining SOABI flags if the needed config var is unset or unavailable.""" val = get_config_var(var) if val is None: if warn: logger.debug("Config variable '%s' is unset, Python ABI tag may " "be incorrect", var) return fallback() return val == expected def get_abi_tag(): # type: () -> Optional[str] """Return the ABI tag based on SOABI (if available) or emulate SOABI (CPython 2, PyPy).""" soabi = get_config_var('SOABI') impl = get_abbr_impl() abi = None # type: Optional[str] if not soabi and impl in {'cp', 'pp'} and hasattr(sys, 'maxunicode'): d = '' m = '' u = '' is_cpython = (impl == 'cp') if get_flag( 'Py_DEBUG', lambda: hasattr(sys, 'gettotalrefcount'), warn=is_cpython): d = 'd' if sys.version_info < (3, 8) and get_flag( 'WITH_PYMALLOC', lambda: is_cpython, warn=is_cpython): m = 'm' if sys.version_info < (3, 3) and get_flag( 'Py_UNICODE_SIZE', lambda: sys.maxunicode == 0x10ffff, expected=4, warn=is_cpython): u = 'u' abi = '%s%s%s%s%s' % (impl, get_impl_ver(), d, m, u) elif soabi and soabi.startswith('cpython-'): abi = 'cp' + soabi.split('-')[1] elif soabi: abi = soabi.replace('.', '_').replace('-', '_') return abi def _is_running_32bit(): # type: () -> bool return sys.maxsize == 2147483647 def get_platform(): # type: () -> str """Return our platform name 'win32', 'linux_x86_64'""" if sys.platform == 'darwin': # distutils.util.get_platform() returns the release based on the value # of MACOSX_DEPLOYMENT_TARGET on which Python was built, which may # be significantly older than the user's current machine. release, _, machine = platform.mac_ver() split_ver = release.split('.') if machine == "x86_64" and _is_running_32bit(): machine = "i386" elif machine == "ppc64" and _is_running_32bit(): machine = "ppc" return 'macosx_{}_{}_{}'.format(split_ver[0], split_ver[1], machine) # XXX remove distutils dependency result = distutils.util.get_platform().replace('.', '_').replace('-', '_') if result == "linux_x86_64" and _is_running_32bit(): # 32 bit Python program (running on a 64 bit Linux): pip should only # install and run 32 bit compiled extensions in that case. result = "linux_i686" return result def is_linux_armhf(): # type: () -> bool if get_platform() != "linux_armv7l": return False # hard-float ABI can be detected from the ELF header of the running # process try: with open(sys.executable, 'rb') as f: elf_header_raw = f.read(40) # read 40 first bytes of ELF header except (IOError, OSError, TypeError): return False if elf_header_raw is None or len(elf_header_raw) < 40: return False if isinstance(elf_header_raw, str): elf_header = [ord(c) for c in elf_header_raw] else: elf_header = [b for b in elf_header_raw] result = elf_header[0:4] == [0x7f, 0x45, 0x4c, 0x46] # ELF magic number result &= elf_header[4:5] == [1] # 32-bit ELF result &= elf_header[5:6] == [1] # little-endian result &= elf_header[18:20] == [0x28, 0] # ARM machine result &= elf_header[39:40] == [5] # ARM EABIv5 result &= (elf_header[37:38][0] & 4) == 4 # EF_ARM_ABI_FLOAT_HARD return result def is_manylinux1_compatible(): # type: () -> bool # Only Linux, and only x86-64 / i686 if get_platform() not in {"linux_x86_64", "linux_i686"}: return False # Check for presence of _manylinux module try: import _manylinux return bool(_manylinux.manylinux1_compatible) except (ImportError, AttributeError): # Fall through to heuristic check below pass # Check glibc version. CentOS 5 uses glibc 2.5. return pip._internal.utils.glibc.have_compatible_glibc(2, 5) def is_manylinux2010_compatible(): # type: () -> bool # Only Linux, and only x86-64 / i686 if get_platform() not in {"linux_x86_64", "linux_i686"}: return False # Check for presence of _manylinux module try: import _manylinux return bool(_manylinux.manylinux2010_compatible) except (ImportError, AttributeError): # Fall through to heuristic check below pass # Check glibc version. CentOS 6 uses glibc 2.12. return pip._internal.utils.glibc.have_compatible_glibc(2, 12) def is_manylinux2014_compatible(): # type: () -> bool # Only Linux, and only supported architectures platform = get_platform() if platform not in {"linux_x86_64", "linux_i686", "linux_aarch64", "linux_armv7l", "linux_ppc64", "linux_ppc64le", "linux_s390x"}: return False # check for hard-float ABI in case we're running linux_armv7l not to # install hard-float ABI wheel in a soft-float ABI environment if platform == "linux_armv7l" and not is_linux_armhf(): return False # Check for presence of _manylinux module try: import _manylinux return bool(_manylinux.manylinux2014_compatible) except (ImportError, AttributeError): # Fall through to heuristic check below pass # Check glibc version. CentOS 7 uses glibc 2.17. return pip._internal.utils.glibc.have_compatible_glibc(2, 17) def get_darwin_arches(major, minor, machine): # type: (int, int, str) -> List[str] """Return a list of supported arches (including group arches) for the given major, minor and machine architecture of an macOS machine. """ arches = [] def _supports_arch(major, minor, arch): # type: (int, int, str) -> bool # Looking at the application support for macOS versions in the chart # provided by https://en.wikipedia.org/wiki/OS_X#Versions it appears # our timeline looks roughly like: # # 10.0 - Introduces ppc support. # 10.4 - Introduces ppc64, i386, and x86_64 support, however the ppc64 # and x86_64 support is CLI only, and cannot be used for GUI # applications. # 10.5 - Extends ppc64 and x86_64 support to cover GUI applications. # 10.6 - Drops support for ppc64 # 10.7 - Drops support for ppc # # Given that we do not know if we're installing a CLI or a GUI # application, we must be conservative and assume it might be a GUI # application and behave as if ppc64 and x86_64 support did not occur # until 10.5. # # Note: The above information is taken from the "Application support" # column in the chart not the "Processor support" since I believe # that we care about what instruction sets an application can use # not which processors the OS supports. if arch == 'ppc': return (major, minor) <= (10, 5) if arch == 'ppc64': return (major, minor) == (10, 5) if arch == 'i386': return (major, minor) >= (10, 4) if arch == 'x86_64': return (major, minor) >= (10, 5) if arch in groups: for garch in groups[arch]: if _supports_arch(major, minor, garch): return True return False groups = OrderedDict([ ("fat", ("i386", "ppc")), ("intel", ("x86_64", "i386")), ("fat64", ("x86_64", "ppc64")), ("fat32", ("x86_64", "i386", "ppc")), ]) # type: Dict[str, Tuple[str, ...]] if _supports_arch(major, minor, machine): arches.append(machine) for garch in groups: if machine in groups[garch] and _supports_arch(major, minor, garch): arches.append(garch) arches.append('universal') return arches def get_all_minor_versions_as_strings(version_info): # type: (Tuple[int, ...]) -> List[str] versions = [] major = version_info[:-1] # Support all previous minor Python versions. for minor in range(version_info[-1], -1, -1): versions.append(''.join(map(str, major + (minor,)))) return versions def get_supported( versions=None, # type: Optional[List[str]] noarch=False, # type: bool platform=None, # type: Optional[str] impl=None, # type: Optional[str] abi=None # type: Optional[str] ): # type: (...) -> List[Pep425Tag] """Return a list of supported tags for each version specified in `versions`. :param versions: a list of string versions, of the form ["33", "32"], or None. The first version will be assumed to support our ABI. :param platform: specify the exact platform you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. :param abi: specify the exact abi you want valid tags for, or None. If None, use the local interpreter abi. """ supported = [] # Versions must be given with respect to the preference if versions is None: version_info = get_impl_version_info() versions = get_all_minor_versions_as_strings(version_info) impl = impl or get_abbr_impl() abis = [] # type: List[str] abi = abi or get_abi_tag() if abi: abis[0:0] = [abi] abi3s = set() # type: Set[str] for suffix in get_extension_suffixes(): if suffix.startswith('.abi'): abi3s.add(suffix.split('.', 2)[1]) abis.extend(sorted(list(abi3s))) abis.append('none') if not noarch: arch = platform or get_platform() arch_prefix, arch_sep, arch_suffix = arch.partition('_') if arch.startswith('macosx'): # support macosx-10.6-intel on macosx-10.9-x86_64 match = _osx_arch_pat.match(arch) if match: name, major, minor, actual_arch = match.groups() tpl = '{}_{}_%i_%s'.format(name, major) arches = [] for m in reversed(range(int(minor) + 1)): for a in get_darwin_arches(int(major), m, actual_arch): arches.append(tpl % (m, a)) else: # arch pattern didn't match (?!) arches = [arch] elif arch_prefix == 'manylinux2014': arches = [arch] # manylinux1/manylinux2010 wheels run on most manylinux2014 systems # with the exception of wheels depending on ncurses. PEP 599 states # manylinux1/manylinux2010 wheels should be considered # manylinux2014 wheels: # https://www.python.org/dev/peps/pep-0599/#backwards-compatibility-with-manylinux2010-wheels if arch_suffix in {'i686', 'x86_64'}: arches.append('manylinux2010' + arch_sep + arch_suffix) arches.append('manylinux1' + arch_sep + arch_suffix) elif arch_prefix == 'manylinux2010': # manylinux1 wheels run on most manylinux2010 systems with the # exception of wheels depending on ncurses. PEP 571 states # manylinux1 wheels should be considered manylinux2010 wheels: # https://www.python.org/dev/peps/pep-0571/#backwards-compatibility-with-manylinux1-wheels arches = [arch, 'manylinux1' + arch_sep + arch_suffix] elif platform is None: arches = [] if is_manylinux2014_compatible(): arches.append('manylinux2014' + arch_sep + arch_suffix) if is_manylinux2010_compatible(): arches.append('manylinux2010' + arch_sep + arch_suffix) if is_manylinux1_compatible(): arches.append('manylinux1' + arch_sep + arch_suffix) arches.append(arch) else: arches = [arch] # Current version, current API (built specifically for our Python): for abi in abis: for arch in arches: supported.append(('%s%s' % (impl, versions[0]), abi, arch)) # abi3 modules compatible with older version of Python for version in versions[1:]: # abi3 was introduced in Python 3.2 if version in {'31', '30'}: break for abi in abi3s: # empty set if not Python 3 for arch in arches: supported.append(("%s%s" % (impl, version), abi, arch)) # Has binaries, does not use the Python API: for arch in arches: supported.append(('py%s' % (versions[0][0]), 'none', arch)) # No abi / arch, but requires our implementation: supported.append(('%s%s' % (impl, versions[0]), 'none', 'any')) # Tagged specifically as being cross-version compatible # (with just the major version specified) supported.append(('%s%s' % (impl, versions[0][0]), 'none', 'any')) # No abi / arch, generic Python for i, version in enumerate(versions): supported.append(('py%s' % (version,), 'none', 'any')) if i == 0: supported.append(('py%s' % (version[0]), 'none', 'any')) return supported implementation_tag = get_impl_tag() site-packages/pip/_internal/cache.py000064400000020263151464452710013460 0ustar00"""Cache Management """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False import errno import hashlib import logging import os from pip._vendor.packaging.utils import canonicalize_name from pip._internal.models.link import Link from pip._internal.utils.compat import expanduser from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import path_to_url from pip._internal.wheel import InvalidWheelFilename, Wheel if MYPY_CHECK_RUNNING: from typing import Optional, Set, List, Any from pip._internal.index import FormatControl from pip._internal.pep425tags import Pep425Tag logger = logging.getLogger(__name__) class Cache(object): """An abstract class - provides cache directories for data from links :param cache_dir: The root of the cache. :param format_control: An object of FormatControl class to limit binaries being read from the cache. :param allowed_formats: which formats of files the cache should store. ('binary' and 'source' are the only allowed values) """ def __init__(self, cache_dir, format_control, allowed_formats): # type: (str, FormatControl, Set[str]) -> None super(Cache, self).__init__() self.cache_dir = expanduser(cache_dir) if cache_dir else None self.format_control = format_control self.allowed_formats = allowed_formats _valid_formats = {"source", "binary"} assert self.allowed_formats.union(_valid_formats) == _valid_formats def _get_cache_path_parts(self, link): # type: (Link) -> List[str] """Get parts of part that must be os.path.joined with cache_dir """ # We want to generate an url to use as our cache key, we don't want to # just re-use the URL because it might have other items in the fragment # and we don't care about those. key_parts = [link.url_without_fragment] if link.hash_name is not None and link.hash is not None: key_parts.append("=".join([link.hash_name, link.hash])) key_url = "#".join(key_parts) # Encode our key url with sha224, we'll use this because it has similar # security properties to sha256, but with a shorter total output (and # thus less secure). However the differences don't make a lot of # difference for our use case here. hashed = hashlib.sha224(key_url.encode()).hexdigest() # We want to nest the directories some to prevent having a ton of top # level directories where we might run out of sub directories on some # FS. parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]] return parts def _get_candidates(self, link, package_name): # type: (Link, Optional[str]) -> List[Any] can_not_cache = ( not self.cache_dir or not package_name or not link ) if can_not_cache: return [] canonical_name = canonicalize_name(package_name) formats = self.format_control.get_allowed_formats( canonical_name ) if not self.allowed_formats.intersection(formats): return [] root = self.get_path_for_link(link) try: return os.listdir(root) except OSError as err: if err.errno in {errno.ENOENT, errno.ENOTDIR}: return [] raise def get_path_for_link(self, link): # type: (Link) -> str """Return a directory to store cached items in for link. """ raise NotImplementedError() def get( self, link, # type: Link package_name, # type: Optional[str] supported_tags, # type: List[Pep425Tag] ): # type: (...) -> Link """Returns a link to a cached item if it exists, otherwise returns the passed link. """ raise NotImplementedError() def _link_for_candidate(self, link, candidate): # type: (Link, str) -> Link root = self.get_path_for_link(link) path = os.path.join(root, candidate) return Link(path_to_url(path)) def cleanup(self): # type: () -> None pass class SimpleWheelCache(Cache): """A cache of wheels for future installs. """ def __init__(self, cache_dir, format_control): # type: (str, FormatControl) -> None super(SimpleWheelCache, self).__init__( cache_dir, format_control, {"binary"} ) def get_path_for_link(self, link): # type: (Link) -> str """Return a directory to store cached wheels for link Because there are M wheels for any one sdist, we provide a directory to cache them in, and then consult that directory when looking up cache hits. We only insert things into the cache if they have plausible version numbers, so that we don't contaminate the cache with things that were not unique. E.g. ./package might have dozens of installs done for it and build a version of 0.0...and if we built and cached a wheel, we'd end up using the same wheel even if the source has been edited. :param link: The link of the sdist for which this will cache wheels. """ parts = self._get_cache_path_parts(link) # Store wheels within the root cache_dir return os.path.join(self.cache_dir, "wheels", *parts) def get( self, link, # type: Link package_name, # type: Optional[str] supported_tags, # type: List[Pep425Tag] ): # type: (...) -> Link candidates = [] for wheel_name in self._get_candidates(link, package_name): try: wheel = Wheel(wheel_name) except InvalidWheelFilename: continue if not wheel.supported(supported_tags): # Built for a different python/arch/etc continue candidates.append( (wheel.support_index_min(supported_tags), wheel_name) ) if not candidates: return link return self._link_for_candidate(link, min(candidates)[1]) class EphemWheelCache(SimpleWheelCache): """A SimpleWheelCache that creates it's own temporary cache directory """ def __init__(self, format_control): # type: (FormatControl) -> None self._temp_dir = TempDirectory(kind="ephem-wheel-cache") super(EphemWheelCache, self).__init__( self._temp_dir.path, format_control ) def cleanup(self): # type: () -> None self._temp_dir.cleanup() class WheelCache(Cache): """Wraps EphemWheelCache and SimpleWheelCache into a single Cache This Cache allows for gracefully degradation, using the ephem wheel cache when a certain link is not found in the simple wheel cache first. """ def __init__(self, cache_dir, format_control): # type: (str, FormatControl) -> None super(WheelCache, self).__init__( cache_dir, format_control, {'binary'} ) self._wheel_cache = SimpleWheelCache(cache_dir, format_control) self._ephem_cache = EphemWheelCache(format_control) def get_path_for_link(self, link): # type: (Link) -> str return self._wheel_cache.get_path_for_link(link) def get_ephem_path_for_link(self, link): # type: (Link) -> str return self._ephem_cache.get_path_for_link(link) def get( self, link, # type: Link package_name, # type: Optional[str] supported_tags, # type: List[Pep425Tag] ): # type: (...) -> Link retval = self._wheel_cache.get( link=link, package_name=package_name, supported_tags=supported_tags, ) if retval is not link: return retval return self._ephem_cache.get( link=link, package_name=package_name, supported_tags=supported_tags, ) def cleanup(self): # type: () -> None self._wheel_cache.cleanup() self._ephem_cache.cleanup() site-packages/pip/_internal/distributions/installed.py000064400000001036151464452710017273 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from pip._internal.distributions.base import AbstractDistribution class InstalledDistribution(AbstractDistribution): """Represents an installed package. This does not need any preparation as the required information has already been computed. """ def get_pkg_resources_distribution(self): return self.req.satisfied_by def prepare_distribution_metadata(self, finder, build_isolation): pass site-packages/pip/_internal/distributions/base.py000064400000002124151464452710016225 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import abc from pip._vendor.six import add_metaclass @add_metaclass(abc.ABCMeta) class AbstractDistribution(object): """A base class for handling installable artifacts. The requirements for anything installable are as follows: - we must be able to determine the requirement name (or we can't correctly handle the non-upgrade case). - for packages with setup requirements, we must also be able to determine their requirements without installing additional packages (for the same reason as run-time dependencies) - we must be able to create a Distribution object exposing the above metadata. """ def __init__(self, req): super(AbstractDistribution, self).__init__() self.req = req @abc.abstractmethod def get_pkg_resources_distribution(self): raise NotImplementedError() @abc.abstractmethod def prepare_distribution_metadata(self, finder, build_isolation): raise NotImplementedError() site-packages/pip/_internal/distributions/__pycache__/base.cpython-38.pyc000064400000003124151464452710022514 0ustar00U .eT@s2ddlZddlmZeejGdddeZdS)N) add_metaclasscs<eZdZdZfddZejddZejddZZ S)AbstractDistributiona A base class for handling installable artifacts. The requirements for anything installable are as follows: - we must be able to determine the requirement name (or we can't correctly handle the non-upgrade case). - for packages with setup requirements, we must also be able to determine their requirements without installing additional packages (for the same reason as run-time dependencies) - we must be able to create a Distribution object exposing the above metadata. cstt|||_dSN)superr__init__req)selfr __class__D/usr/lib/python3.8/site-packages/pip/_internal/distributions/base.pyrszAbstractDistribution.__init__cCs tdSrNotImplementedError)rr r r get_pkg_resources_distributionsz3AbstractDistribution.get_pkg_resources_distributioncCs tdSrr )rfinderZbuild_isolationr r r prepare_distribution_metadata"sz2AbstractDistribution.prepare_distribution_metadata) __name__ __module__ __qualname____doc__rabcabstractmethodrr __classcell__r r r r r s   r)rZpip._vendor.sixrABCMetaobjectrr r r r s site-packages/pip/_internal/distributions/__pycache__/installed.cpython-38.opt-1.pyc000064400000001700151464452710024516 0ustar00U .e@s ddlmZGdddeZdS))AbstractDistributionc@s eZdZdZddZddZdS)InstalledDistributionzRepresents an installed package. This does not need any preparation as the required information has already been computed. cCs|jjSN)ZreqZ satisfied_by)selfrI/usr/lib/python3.8/site-packages/pip/_internal/distributions/installed.pyget_pkg_resources_distributionsz4InstalledDistribution.get_pkg_resources_distributioncCsdSrr)rfinderZbuild_isolationrrrprepare_distribution_metadatasz3InstalledDistribution.prepare_distribution_metadataN)__name__ __module__ __qualname____doc__rr rrrrrsrN)Z pip._internal.distributions.baserrrrrrs site-packages/pip/_internal/distributions/__pycache__/__init__.cpython-38.pyc000064400000001464151464452710023346 0ustar00U .e@sLddlmZddlmZddlmZer@ddlmZddlm Z ddZ dS) )SourceDistribution)WheelDistribution)MYPY_CHECK_RUNNING)AbstractDistribution)InstallRequirementcCs$|jrt|S|jrt|St|S)zs     site-packages/pip/_internal/distributions/__pycache__/base.cpython-38.opt-1.pyc000064400000003124151464452710023453 0ustar00U .eT@s2ddlZddlmZeejGdddeZdS)N) add_metaclasscs<eZdZdZfddZejddZejddZZ S)AbstractDistributiona A base class for handling installable artifacts. The requirements for anything installable are as follows: - we must be able to determine the requirement name (or we can't correctly handle the non-upgrade case). - for packages with setup requirements, we must also be able to determine their requirements without installing additional packages (for the same reason as run-time dependencies) - we must be able to create a Distribution object exposing the above metadata. cstt|||_dSN)superr__init__req)selfr __class__D/usr/lib/python3.8/site-packages/pip/_internal/distributions/base.pyrszAbstractDistribution.__init__cCs tdSrNotImplementedError)rr r r get_pkg_resources_distributionsz3AbstractDistribution.get_pkg_resources_distributioncCs tdSrr )rfinderZbuild_isolationr r r prepare_distribution_metadata"sz2AbstractDistribution.prepare_distribution_metadata) __name__ __module__ __qualname____doc__rabcabstractmethodrr __classcell__r r r r r s   r)rZpip._vendor.sixrABCMetaobjectrr r r r s site-packages/pip/_internal/distributions/__pycache__/wheel.cpython-38.opt-1.pyc000064400000002000151464452710023635 0ustar00U .eh@s,ddlmZddlmZGdddeZdS)) pkg_resources)AbstractDistributionc@s eZdZdZddZddZdS)WheelDistributionzqRepresents a wheel distribution. This does not need any preparation as wheels can be directly unpacked. cCstt|jjdS)Nr)listrfind_distributionsZreqZ source_dir)selfrE/usr/lib/python3.8/site-packages/pip/_internal/distributions/wheel.pyget_pkg_resources_distributions z0WheelDistribution.get_pkg_resources_distributioncCsdS)Nr)rfinderZbuild_isolationrrr prepare_distribution_metadatasz/WheelDistribution.prepare_distribution_metadataN)__name__ __module__ __qualname____doc__r r rrrr r srN)Z pip._vendorrZ pip._internal.distributions.baserrrrrr s  site-packages/pip/_internal/distributions/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001464151464452710024305 0ustar00U .e@sLddlmZddlmZddlmZer@ddlmZddlm Z ddZ dS) )SourceDistribution)WheelDistribution)MYPY_CHECK_RUNNING)AbstractDistribution)InstallRequirementcCs$|jrt|S|jrt|St|S)zs     site-packages/pip/_internal/distributions/__pycache__/installed.cpython-38.pyc000064400000001700151464452710023557 0ustar00U .e@s ddlmZGdddeZdS))AbstractDistributionc@s eZdZdZddZddZdS)InstalledDistributionzRepresents an installed package. This does not need any preparation as the required information has already been computed. cCs|jjSN)ZreqZ satisfied_by)selfrI/usr/lib/python3.8/site-packages/pip/_internal/distributions/installed.pyget_pkg_resources_distributionsz4InstalledDistribution.get_pkg_resources_distributioncCsdSrr)rfinderZbuild_isolationrrrprepare_distribution_metadatasz3InstalledDistribution.prepare_distribution_metadataN)__name__ __module__ __qualname____doc__rr rrrrrsrN)Z pip._internal.distributions.baserrrrrrs site-packages/pip/_internal/distributions/__pycache__/wheel.cpython-38.pyc000064400000002000151464452710022676 0ustar00U .eh@s,ddlmZddlmZGdddeZdS)) pkg_resources)AbstractDistributionc@s eZdZdZddZddZdS)WheelDistributionzqRepresents a wheel distribution. This does not need any preparation as wheels can be directly unpacked. cCstt|jjdS)Nr)listrfind_distributionsZreqZ source_dir)selfrE/usr/lib/python3.8/site-packages/pip/_internal/distributions/wheel.pyget_pkg_resources_distributions z0WheelDistribution.get_pkg_resources_distributioncCsdS)Nr)rfinderZbuild_isolationrrr prepare_distribution_metadatasz/WheelDistribution.prepare_distribution_metadataN)__name__ __module__ __qualname____doc__r r rrrr r srN)Z pip._vendorrZ pip._internal.distributions.baserrrrrr s  site-packages/pip/_internal/distributions/wheel.py000064400000001150151464452710016415 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from pip._vendor import pkg_resources from pip._internal.distributions.base import AbstractDistribution class WheelDistribution(AbstractDistribution): """Represents a wheel distribution. This does not need any preparation as wheels can be directly unpacked. """ def get_pkg_resources_distribution(self): return list(pkg_resources.find_distributions( self.req.source_dir))[0] def prepare_distribution_metadata(self, finder, build_isolation): pass site-packages/pip/_internal/distributions/source/__pycache__/legacy.cpython-38.opt-1.pyc000064400000006617151464452710025317 0ustar00U .ee@sVddlZddlmZddlmZddlmZddlmZe e Z GdddeZ dS)N)BuildEnvironment)AbstractDistribution)InstallationError)runner_with_spinner_messagec@s(eZdZdZddZddZddZdS) SourceDistributionaRepresents a source distribution. The preparation step for these needs metadata for the packages to be generated, either using PEP 517 or using the legacy `setup.py egg_info`. NOTE from @pradyunsg (14 June 2019) I expect SourceDistribution class will need to be split into `legacy_source` (setup.py based) and `source` (PEP 517 based) when we start bringing logic for preparation out of InstallRequirement into this class. cCs |jSN)reqZget_dist)selfr M/usr/lib/python3.8/site-packages/pip/_internal/distributions/source/legacy.pyget_pkg_resources_distributionsz1SourceDistribution.get_pkg_resources_distributioncCs<|j|jjo|}|r$|||j|jdSr)rZload_pyproject_tomlZ use_pep517_setup_isolationZprepare_metadataZassert_source_matches_version)r finderZbuild_isolationZshould_isolater r r prepare_distribution_metadatas     z0SourceDistribution.prepare_distribution_metadatac sfdd}tj_jj|jjddjjjj\}rT|d|rtdjtdd t t t |jj4t d }jj}|||}W5QRXW5QRXjj|\}r|d jj||d d dS) Ncs6d}|jj|dddtDd}t|dS)NzZSome build dependencies for {requirement} conflict with {conflicting_with}: {description}.z, css|]\}}d||fVqdS)z%s is incompatible with %sNr ).0Z installedZwantedr r r 3szPSourceDistribution._setup_isolation.._raise_conflicts..)Z requirementconflicting_with description)formatrjoinsortedr)rZconflicting_reqs format_stringZ error_messageZ conflictingr r r _raise_conflicts+s z=SourceDistribution._setup_isolation.._raise_conflictsZoverlayzInstalling build dependenciesz"PEP 517/518 supported requirementsz4Missing build requirements in pyproject.toml for %s.z`The project does not specify a build backend, and pip cannot fall back to setuptools without %s.z and z#Getting requirements to build wheelzthe backend dependenciesZnormalzInstalling backend dependencies)rrZ build_envZinstall_requirementsZpyproject_requiresZcheck_requirementsZrequirements_to_checkloggerZwarningrmapreprrrZpep517_backendZsubprocess_runnerZget_requires_for_build_wheel)r rrZmissingZrunnerZbackendZreqsr rr r *sP    z#SourceDistribution._setup_isolationN)__name__ __module__ __qualname____doc__r rr r r r r rs  r) ZloggingZpip._internal.build_envrZ pip._internal.distributions.baserZpip._internal.exceptionsrZpip._internal.utils.subprocessrZ getLoggerrrrr r r r s      site-packages/pip/_internal/distributions/source/__pycache__/__init__.cpython-38.pyc000064400000000252151464452710024640 0ustar00U .e@sdS)NrrrO/usr/lib/python3.8/site-packages/pip/_internal/distributions/source/__init__.pysite-packages/pip/_internal/distributions/source/__pycache__/legacy.cpython-38.pyc000064400000006617151464452710024360 0ustar00U .ee@sVddlZddlmZddlmZddlmZddlmZe e Z GdddeZ dS)N)BuildEnvironment)AbstractDistribution)InstallationError)runner_with_spinner_messagec@s(eZdZdZddZddZddZdS) SourceDistributionaRepresents a source distribution. The preparation step for these needs metadata for the packages to be generated, either using PEP 517 or using the legacy `setup.py egg_info`. NOTE from @pradyunsg (14 June 2019) I expect SourceDistribution class will need to be split into `legacy_source` (setup.py based) and `source` (PEP 517 based) when we start bringing logic for preparation out of InstallRequirement into this class. cCs |jSN)reqZget_dist)selfr M/usr/lib/python3.8/site-packages/pip/_internal/distributions/source/legacy.pyget_pkg_resources_distributionsz1SourceDistribution.get_pkg_resources_distributioncCs<|j|jjo|}|r$|||j|jdSr)rZload_pyproject_tomlZ use_pep517_setup_isolationZprepare_metadataZassert_source_matches_version)r finderZbuild_isolationZshould_isolater r r prepare_distribution_metadatas     z0SourceDistribution.prepare_distribution_metadatac sfdd}tj_jj|jjddjjjj\}rT|d|rtdjtdd t t t |jj4t d }jj}|||}W5QRXW5QRXjj|\}r|d jj||d d dS) Ncs6d}|jj|dddtDd}t|dS)NzZSome build dependencies for {requirement} conflict with {conflicting_with}: {description}.z, css|]\}}d||fVqdS)z%s is incompatible with %sNr ).0Z installedZwantedr r r 3szPSourceDistribution._setup_isolation.._raise_conflicts..)Z requirementconflicting_with description)formatrjoinsortedr)rZconflicting_reqs format_stringZ error_messageZ conflictingr r r _raise_conflicts+s z=SourceDistribution._setup_isolation.._raise_conflictsZoverlayzInstalling build dependenciesz"PEP 517/518 supported requirementsz4Missing build requirements in pyproject.toml for %s.z`The project does not specify a build backend, and pip cannot fall back to setuptools without %s.z and z#Getting requirements to build wheelzthe backend dependenciesZnormalzInstalling backend dependencies)rrZ build_envZinstall_requirementsZpyproject_requiresZcheck_requirementsZrequirements_to_checkloggerZwarningrmapreprrrZpep517_backendZsubprocess_runnerZget_requires_for_build_wheel)r rrZmissingZrunnerZbackendZreqsr rr r *sP    z#SourceDistribution._setup_isolationN)__name__ __module__ __qualname____doc__r rr r r r r rs  r) ZloggingZpip._internal.build_envrZ pip._internal.distributions.baserZpip._internal.exceptionsrZpip._internal.utils.subprocessrZ getLoggerrrrr r r r s      site-packages/pip/_internal/distributions/source/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000252151464452710025577 0ustar00U .e@sdS)NrrrO/usr/lib/python3.8/site-packages/pip/_internal/distributions/source/__init__.pysite-packages/pip/_internal/distributions/source/legacy.py000064400000007545151464452710020073 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import logging from pip._internal.build_env import BuildEnvironment from pip._internal.distributions.base import AbstractDistribution from pip._internal.exceptions import InstallationError from pip._internal.utils.subprocess import runner_with_spinner_message logger = logging.getLogger(__name__) class SourceDistribution(AbstractDistribution): """Represents a source distribution. The preparation step for these needs metadata for the packages to be generated, either using PEP 517 or using the legacy `setup.py egg_info`. NOTE from @pradyunsg (14 June 2019) I expect SourceDistribution class will need to be split into `legacy_source` (setup.py based) and `source` (PEP 517 based) when we start bringing logic for preparation out of InstallRequirement into this class. """ def get_pkg_resources_distribution(self): return self.req.get_dist() def prepare_distribution_metadata(self, finder, build_isolation): # Prepare for building. We need to: # 1. Load pyproject.toml (if it exists) # 2. Set up the build environment self.req.load_pyproject_toml() should_isolate = self.req.use_pep517 and build_isolation if should_isolate: self._setup_isolation(finder) self.req.prepare_metadata() self.req.assert_source_matches_version() def _setup_isolation(self, finder): def _raise_conflicts(conflicting_with, conflicting_reqs): format_string = ( "Some build dependencies for {requirement} " "conflict with {conflicting_with}: {description}." ) error_message = format_string.format( requirement=self.req, conflicting_with=conflicting_with, description=', '.join( '%s is incompatible with %s' % (installed, wanted) for installed, wanted in sorted(conflicting) ) ) raise InstallationError(error_message) # Isolate in a BuildEnvironment and install the build-time # requirements. self.req.build_env = BuildEnvironment() self.req.build_env.install_requirements( finder, self.req.pyproject_requires, 'overlay', "Installing build dependencies" ) conflicting, missing = self.req.build_env.check_requirements( self.req.requirements_to_check ) if conflicting: _raise_conflicts("PEP 517/518 supported requirements", conflicting) if missing: logger.warning( "Missing build requirements in pyproject.toml for %s.", self.req, ) logger.warning( "The project does not specify a build backend, and " "pip cannot fall back to setuptools without %s.", " and ".join(map(repr, sorted(missing))) ) # Install any extra build dependencies that the backend requests. # This must be done in a second pass, as the pyproject.toml # dependencies must be installed before we can call the backend. with self.req.build_env: runner = runner_with_spinner_message( "Getting requirements to build wheel" ) backend = self.req.pep517_backend with backend.subprocess_runner(runner): reqs = backend.get_requires_for_build_wheel() conflicting, missing = self.req.build_env.check_requirements(reqs) if conflicting: _raise_conflicts("the backend dependencies", conflicting) self.req.build_env.install_requirements( finder, missing, 'normal', "Installing backend dependencies" ) site-packages/pip/_internal/distributions/source/__init__.py000064400000000000151464452710020341 0ustar00site-packages/pip/_internal/distributions/__init__.py000064400000001707151464452710017060 0ustar00from pip._internal.distributions.source.legacy import SourceDistribution from pip._internal.distributions.wheel import WheelDistribution from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from pip._internal.distributions.base import AbstractDistribution from pip._internal.req.req_install import InstallRequirement def make_distribution_for_install_requirement(install_req): # type: (InstallRequirement) -> AbstractDistribution """Returns a Distribution for the given InstallRequirement """ # Editable requirements will always be source distributions. They use the # legacy logic until we create a modern standard for them. if install_req.editable: return SourceDistribution(install_req) # If it's a wheel, it's a WheelDistribution if install_req.is_wheel: return WheelDistribution(install_req) # Otherwise, a SourceDistribution return SourceDistribution(install_req) site-packages/pip/_internal/pyproject.py000064400000014532151464452710014436 0ustar00from __future__ import absolute_import import io import os import sys from pip._vendor import pytoml, six from pip._internal.exceptions import InstallationError from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Any, Tuple, Optional, List def _is_list_of_str(obj): # type: (Any) -> bool return ( isinstance(obj, list) and all(isinstance(item, six.string_types) for item in obj) ) def make_pyproject_path(unpacked_source_directory): # type: (str) -> str path = os.path.join(unpacked_source_directory, 'pyproject.toml') # Python2 __file__ should not be unicode if six.PY2 and isinstance(path, six.text_type): path = path.encode(sys.getfilesystemencoding()) return path def load_pyproject_toml( use_pep517, # type: Optional[bool] pyproject_toml, # type: str setup_py, # type: str req_name # type: str ): # type: (...) -> Optional[Tuple[List[str], str, List[str]]] """Load the pyproject.toml file. Parameters: use_pep517 - Has the user requested PEP 517 processing? None means the user hasn't explicitly specified. pyproject_toml - Location of the project's pyproject.toml file setup_py - Location of the project's setup.py file req_name - The name of the requirement we're processing (for error reporting) Returns: None if we should use the legacy code path, otherwise a tuple ( requirements from pyproject.toml, name of PEP 517 backend, requirements we should check are installed after setting up the build environment ) """ has_pyproject = os.path.isfile(pyproject_toml) has_setup = os.path.isfile(setup_py) if has_pyproject: with io.open(pyproject_toml, encoding="utf-8") as f: pp_toml = pytoml.load(f) build_system = pp_toml.get("build-system") else: build_system = None # The following cases must use PEP 517 # We check for use_pep517 being non-None and falsey because that means # the user explicitly requested --no-use-pep517. The value 0 as # opposed to False can occur when the value is provided via an # environment variable or config file option (due to the quirk of # strtobool() returning an integer in pip's configuration code). if has_pyproject and not has_setup: if use_pep517 is not None and not use_pep517: raise InstallationError( "Disabling PEP 517 processing is invalid: " "project does not have a setup.py" ) use_pep517 = True elif build_system and "build-backend" in build_system: if use_pep517 is not None and not use_pep517: raise InstallationError( "Disabling PEP 517 processing is invalid: " "project specifies a build backend of {} " "in pyproject.toml".format( build_system["build-backend"] ) ) use_pep517 = True # If we haven't worked out whether to use PEP 517 yet, # and the user hasn't explicitly stated a preference, # we do so if the project has a pyproject.toml file. elif use_pep517 is None: use_pep517 = has_pyproject # At this point, we know whether we're going to use PEP 517. assert use_pep517 is not None # If we're using the legacy code path, there is nothing further # for us to do here. if not use_pep517: return None if build_system is None: # Either the user has a pyproject.toml with no build-system # section, or the user has no pyproject.toml, but has opted in # explicitly via --use-pep517. # In the absence of any explicit backend specification, we # assume the setuptools backend that most closely emulates the # traditional direct setup.py execution, and require wheel and # a version of setuptools that supports that backend. build_system = { "requires": ["setuptools>=40.8.0", "wheel"], "build-backend": "setuptools.build_meta:__legacy__", } # If we're using PEP 517, we have build system information (either # from pyproject.toml, or defaulted by the code above). # Note that at this point, we do not know if the user has actually # specified a backend, though. assert build_system is not None # Ensure that the build-system section in pyproject.toml conforms # to PEP 518. error_template = ( "{package} has a pyproject.toml file that does not comply " "with PEP 518: {reason}" ) # Specifying the build-system table but not the requires key is invalid if "requires" not in build_system: raise InstallationError( error_template.format(package=req_name, reason=( "it has a 'build-system' table but not " "'build-system.requires' which is mandatory in the table" )) ) # Error out if requires is not a list of strings requires = build_system["requires"] if not _is_list_of_str(requires): raise InstallationError(error_template.format( package=req_name, reason="'build-system.requires' is not a list of strings.", )) backend = build_system.get("build-backend") check = [] # type: List[str] if backend is None: # If the user didn't specify a backend, we assume they want to use # the setuptools backend. But we can't be sure they have included # a version of setuptools which supplies the backend, or wheel # (which is needed by the backend) in their requirements. So we # make a note to check that those requirements are present once # we have set up the environment. # This is quite a lot of work to check for a very specific case. But # the problem is, that case is potentially quite common - projects that # adopted PEP 518 early for the ability to specify requirements to # execute setup.py, but never considered needing to mention the build # tools themselves. The original PEP 518 code had a similar check (but # implemented in a different way). backend = "setuptools.build_meta:__legacy__" check = ["setuptools>=40.8.0", "wheel"] return (requires, backend, check) site-packages/pip/_internal/operations/freeze.py000064400000023142151464452710016057 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import collections import logging import os import re from pip._vendor import six from pip._vendor.packaging.utils import canonicalize_name from pip._vendor.pkg_resources import RequirementParseError from pip._internal.exceptions import BadCommand, InstallationError from pip._internal.req.constructors import ( install_req_from_editable, install_req_from_line, ) from pip._internal.req.req_file import COMMENT_RE from pip._internal.utils.misc import ( dist_is_editable, get_installed_distributions, ) from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ( Iterator, Optional, List, Container, Set, Dict, Tuple, Iterable, Union ) from pip._internal.cache import WheelCache from pip._vendor.pkg_resources import ( Distribution, Requirement ) RequirementInfo = Tuple[Optional[Union[str, Requirement]], bool, List[str]] logger = logging.getLogger(__name__) def freeze( requirement=None, # type: Optional[List[str]] find_links=None, # type: Optional[List[str]] local_only=None, # type: Optional[bool] user_only=None, # type: Optional[bool] paths=None, # type: Optional[List[str]] skip_regex=None, # type: Optional[str] isolated=False, # type: bool wheel_cache=None, # type: Optional[WheelCache] exclude_editable=False, # type: bool skip=() # type: Container[str] ): # type: (...) -> Iterator[str] find_links = find_links or [] skip_match = None if skip_regex: skip_match = re.compile(skip_regex).search for link in find_links: yield '-f %s' % link installations = {} # type: Dict[str, FrozenRequirement] for dist in get_installed_distributions(local_only=local_only, skip=(), user_only=user_only, paths=paths): try: req = FrozenRequirement.from_dist(dist) except RequirementParseError as exc: # We include dist rather than dist.project_name because the # dist string includes more information, like the version and # location. We also include the exception message to aid # troubleshooting. logger.warning( 'Could not generate requirement for distribution %r: %s', dist, exc ) continue if exclude_editable and req.editable: continue installations[req.name] = req if requirement: # the options that don't get turned into an InstallRequirement # should only be emitted once, even if the same option is in multiple # requirements files, so we need to keep track of what has been emitted # so that we don't emit it again if it's seen again emitted_options = set() # type: Set[str] # keep track of which files a requirement is in so that we can # give an accurate warning if a requirement appears multiple times. req_files = collections.defaultdict(list) # type: Dict[str, List[str]] for req_file_path in requirement: with open(req_file_path) as req_file: for line in req_file: if (not line.strip() or line.strip().startswith('#') or (skip_match and skip_match(line)) or line.startswith(( '-r', '--requirement', '-Z', '--always-unzip', '-f', '--find-links', '-i', '--index-url', '--pre', '--trusted-host', '--process-dependency-links', '--extra-index-url'))): line = line.rstrip() if line not in emitted_options: emitted_options.add(line) yield line continue if line.startswith('-e') or line.startswith('--editable'): if line.startswith('-e'): line = line[2:].strip() else: line = line[len('--editable'):].strip().lstrip('=') line_req = install_req_from_editable( line, isolated=isolated, wheel_cache=wheel_cache, ) else: line_req = install_req_from_line( COMMENT_RE.sub('', line).strip(), isolated=isolated, wheel_cache=wheel_cache, ) if not line_req.name: logger.info( "Skipping line in requirement file [%s] because " "it's not clear what it would install: %s", req_file_path, line.strip(), ) logger.info( " (add #egg=PackageName to the URL to avoid" " this warning)" ) elif line_req.name not in installations: # either it's not installed, or it is installed # but has been processed already if not req_files[line_req.name]: logger.warning( "Requirement file [%s] contains %s, but " "package %r is not installed", req_file_path, COMMENT_RE.sub('', line).strip(), line_req.name ) else: req_files[line_req.name].append(req_file_path) else: yield str(installations[line_req.name]).rstrip() del installations[line_req.name] req_files[line_req.name].append(req_file_path) # Warn about requirements that were included multiple times (in a # single requirements file or in different requirements files). for name, files in six.iteritems(req_files): if len(files) > 1: logger.warning("Requirement %s included multiple times [%s]", name, ', '.join(sorted(set(files)))) yield( '## The following requirements were added by ' 'pip freeze:' ) for installation in sorted( installations.values(), key=lambda x: x.name.lower()): if canonicalize_name(installation.name) not in skip: yield str(installation).rstrip() def get_requirement_info(dist): # type: (Distribution) -> RequirementInfo """ Compute and return values (req, editable, comments) for use in FrozenRequirement.from_dist(). """ if not dist_is_editable(dist): return (None, False, []) location = os.path.normcase(os.path.abspath(dist.location)) from pip._internal.vcs import vcs, RemoteNotFoundError vcs_backend = vcs.get_backend_for_dir(location) if vcs_backend is None: req = dist.as_requirement() logger.debug( 'No VCS found for editable requirement "%s" in: %r', req, location, ) comments = [ '# Editable install with no version control ({})'.format(req) ] return (location, True, comments) try: req = vcs_backend.get_src_requirement(location, dist.project_name) except RemoteNotFoundError: req = dist.as_requirement() comments = [ '# Editable {} install with no remote ({})'.format( type(vcs_backend).__name__, req, ) ] return (location, True, comments) except BadCommand: logger.warning( 'cannot determine version of editable source in %s ' '(%s command not found in path)', location, vcs_backend.name, ) return (None, True, []) except InstallationError as exc: logger.warning( "Error when trying to get requirement for VCS system %s, " "falling back to uneditable format", exc ) else: if req is not None: return (req, True, []) logger.warning( 'Could not determine repository location of %s', location ) comments = ['## !! Could not determine repository location'] return (None, False, comments) class FrozenRequirement(object): def __init__(self, name, req, editable, comments=()): # type: (str, Union[str, Requirement], bool, Iterable[str]) -> None self.name = name self.req = req self.editable = editable self.comments = comments @classmethod def from_dist(cls, dist): # type: (Distribution) -> FrozenRequirement req, editable, comments = get_requirement_info(dist) if req is None: req = dist.as_requirement() return cls(dist.project_name, req, editable, comments=comments) def __str__(self): req = self.req if self.editable: req = '-e %s' % req return '\n'.join(list(self.comments) + [str(req)]) + '\n' site-packages/pip/_internal/operations/generate_metadata.py000064400000011133151464452710020226 0ustar00"""Metadata generation logic for source distributions. """ import logging import os from pip._internal.exceptions import InstallationError from pip._internal.utils.misc import ensure_dir from pip._internal.utils.setuptools_build import make_setuptools_shim_args from pip._internal.utils.subprocess import call_subprocess from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.vcs import vcs if MYPY_CHECK_RUNNING: from typing import Callable, List from pip._internal.req.req_install import InstallRequirement logger = logging.getLogger(__name__) def get_metadata_generator(install_req): # type: (InstallRequirement) -> Callable[[InstallRequirement], str] """Return a callable metadata generator for this InstallRequirement. A metadata generator takes an InstallRequirement (install_req) as an input, generates metadata via the appropriate process for that install_req and returns the generated metadata directory. """ if not install_req.use_pep517: return _generate_metadata_legacy return _generate_metadata def _find_egg_info(source_directory, is_editable): # type: (str, bool) -> str """Find an .egg-info in `source_directory`, based on `is_editable`. """ def looks_like_virtual_env(path): # type: (str) -> bool return ( os.path.lexists(os.path.join(path, 'bin', 'python')) or os.path.exists(os.path.join(path, 'Scripts', 'Python.exe')) ) def locate_editable_egg_info(base): # type: (str) -> List[str] candidates = [] # type: List[str] for root, dirs, files in os.walk(base): for dir_ in vcs.dirnames: if dir_ in dirs: dirs.remove(dir_) # Iterate over a copy of ``dirs``, since mutating # a list while iterating over it can cause trouble. # (See https://github.com/pypa/pip/pull/462.) for dir_ in list(dirs): if looks_like_virtual_env(os.path.join(root, dir_)): dirs.remove(dir_) # Also don't search through tests elif dir_ == 'test' or dir_ == 'tests': dirs.remove(dir_) candidates.extend(os.path.join(root, dir_) for dir_ in dirs) return [f for f in candidates if f.endswith('.egg-info')] def depth_of_directory(dir_): # type: (str) -> int return ( dir_.count(os.path.sep) + (os.path.altsep and dir_.count(os.path.altsep) or 0) ) base = source_directory if is_editable: filenames = locate_editable_egg_info(base) else: base = os.path.join(base, 'pip-egg-info') filenames = os.listdir(base) if not filenames: raise InstallationError( "Files/directories not found in %s" % base ) # If we have more than one match, we pick the toplevel one. This # can easily be the case if there is a dist folder which contains # an extracted tarball for testing purposes. if len(filenames) > 1: filenames.sort(key=depth_of_directory) return os.path.join(base, filenames[0]) def _generate_metadata_legacy(install_req): # type: (InstallRequirement) -> str req_details_str = install_req.name or "from {}".format(install_req.link) logger.debug( 'Running setup.py (path:%s) egg_info for package %s', install_req.setup_py_path, req_details_str, ) # Compose arguments for subprocess call base_cmd = make_setuptools_shim_args(install_req.setup_py_path) if install_req.isolated: base_cmd += ["--no-user-cfg"] # For non-editable installs, don't put the .egg-info files at the root, # to avoid confusion due to the source code being considered an installed # egg. egg_base_option = [] # type: List[str] if not install_req.editable: egg_info_dir = os.path.join( install_req.unpacked_source_directory, 'pip-egg-info', ) egg_base_option = ['--egg-base', egg_info_dir] # setuptools complains if the target directory does not exist. ensure_dir(egg_info_dir) with install_req.build_env: call_subprocess( base_cmd + ["egg_info"] + egg_base_option, cwd=install_req.unpacked_source_directory, command_desc='python setup.py egg_info', ) # Return the .egg-info directory. return _find_egg_info( install_req.unpacked_source_directory, install_req.editable, ) def _generate_metadata(install_req): # type: (InstallRequirement) -> str return install_req.prepare_pep517_metadata() site-packages/pip/_internal/operations/check.py000064400000012352151464452710015655 0ustar00"""Validation of dependencies of packages """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False import logging from collections import namedtuple from pip._vendor.packaging.utils import canonicalize_name from pip._vendor.pkg_resources import RequirementParseError from pip._internal.distributions import ( make_distribution_for_install_requirement, ) from pip._internal.utils.misc import get_installed_distributions from pip._internal.utils.typing import MYPY_CHECK_RUNNING logger = logging.getLogger(__name__) if MYPY_CHECK_RUNNING: from pip._internal.req.req_install import InstallRequirement from typing import ( Any, Callable, Dict, Optional, Set, Tuple, List ) # Shorthands PackageSet = Dict[str, 'PackageDetails'] Missing = Tuple[str, Any] Conflicting = Tuple[str, str, Any] MissingDict = Dict[str, List[Missing]] ConflictingDict = Dict[str, List[Conflicting]] CheckResult = Tuple[MissingDict, ConflictingDict] PackageDetails = namedtuple('PackageDetails', ['version', 'requires']) def create_package_set_from_installed(**kwargs): # type: (**Any) -> Tuple[PackageSet, bool] """Converts a list of distributions into a PackageSet. """ # Default to using all packages installed on the system if kwargs == {}: kwargs = {"local_only": False, "skip": ()} package_set = {} problems = False for dist in get_installed_distributions(**kwargs): name = canonicalize_name(dist.project_name) try: package_set[name] = PackageDetails(dist.version, dist.requires()) except RequirementParseError as e: # Don't crash on broken metadata logging.warning("Error parsing requirements for %s: %s", name, e) problems = True return package_set, problems def check_package_set(package_set, should_ignore=None): # type: (PackageSet, Optional[Callable[[str], bool]]) -> CheckResult """Check if a package set is consistent If should_ignore is passed, it should be a callable that takes a package name and returns a boolean. """ if should_ignore is None: def should_ignore(name): return False missing = {} conflicting = {} for package_name in package_set: # Info about dependencies of package_name missing_deps = set() # type: Set[Missing] conflicting_deps = set() # type: Set[Conflicting] if should_ignore(package_name): continue for req in package_set[package_name].requires: name = canonicalize_name(req.project_name) # type: str # Check if it's missing if name not in package_set: missed = True if req.marker is not None: missed = req.marker.evaluate() if missed: missing_deps.add((name, req)) continue # Check if there's a conflict version = package_set[name].version # type: str if not req.specifier.contains(version, prereleases=True): conflicting_deps.add((name, version, req)) if missing_deps: missing[package_name] = sorted(missing_deps, key=str) if conflicting_deps: conflicting[package_name] = sorted(conflicting_deps, key=str) return missing, conflicting def check_install_conflicts(to_install): # type: (List[InstallRequirement]) -> Tuple[PackageSet, CheckResult] """For checking if the dependency graph would be consistent after \ installing given requirements """ # Start from the current state package_set, _ = create_package_set_from_installed() # Install packages would_be_installed = _simulate_installation_of(to_install, package_set) # Only warn about directly-dependent packages; create a whitelist of them whitelist = _create_whitelist(would_be_installed, package_set) return ( package_set, check_package_set( package_set, should_ignore=lambda name: name not in whitelist ) ) def _simulate_installation_of(to_install, package_set): # type: (List[InstallRequirement], PackageSet) -> Set[str] """Computes the version of packages after installing to_install. """ # Keep track of packages that were installed installed = set() # Modify it as installing requirement_set would (assuming no errors) for inst_req in to_install: abstract_dist = make_distribution_for_install_requirement(inst_req) dist = abstract_dist.get_pkg_resources_distribution() name = canonicalize_name(dist.key) package_set[name] = PackageDetails(dist.version, dist.requires()) installed.add(name) return installed def _create_whitelist(would_be_installed, package_set): # type: (Set[str], PackageSet) -> Set[str] packages_affected = set(would_be_installed) for package_name in package_set: if package_name in packages_affected: continue for req in package_set[package_name].requires: if canonicalize_name(req.name) in packages_affected: packages_affected.add(package_name) break return packages_affected site-packages/pip/_internal/operations/prepare.py000064400000026017151464452710016241 0ustar00"""Prepares a distribution for installation """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False import logging import os from pip._vendor import requests from pip._internal.distributions import ( make_distribution_for_install_requirement, ) from pip._internal.distributions.installed import InstalledDistribution from pip._internal.download import unpack_url from pip._internal.exceptions import ( DirectoryUrlHashUnsupported, HashUnpinned, InstallationError, PreviousBuildDirError, VcsHashUnsupported, ) from pip._internal.utils.compat import expanduser from pip._internal.utils.hashes import MissingHashes from pip._internal.utils.logging import indent_log from pip._internal.utils.marker_files import write_delete_marker_file from pip._internal.utils.misc import display_path, normalize_path from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional from pip._internal.distributions import AbstractDistribution from pip._internal.index import PackageFinder from pip._internal.network.session import PipSession from pip._internal.req.req_install import InstallRequirement from pip._internal.req.req_tracker import RequirementTracker logger = logging.getLogger(__name__) def _get_prepared_distribution(req, req_tracker, finder, build_isolation): """Prepare a distribution for installation. """ abstract_dist = make_distribution_for_install_requirement(req) with req_tracker.track(req): abstract_dist.prepare_distribution_metadata(finder, build_isolation) return abstract_dist class RequirementPreparer(object): """Prepares a Requirement """ def __init__( self, build_dir, # type: str download_dir, # type: Optional[str] src_dir, # type: str wheel_download_dir, # type: Optional[str] progress_bar, # type: str build_isolation, # type: bool req_tracker # type: RequirementTracker ): # type: (...) -> None super(RequirementPreparer, self).__init__() self.src_dir = src_dir self.build_dir = build_dir self.req_tracker = req_tracker # Where still-packed archives should be written to. If None, they are # not saved, and are deleted immediately after unpacking. if download_dir: download_dir = expanduser(download_dir) self.download_dir = download_dir # Where still-packed .whl files should be written to. If None, they are # written to the download_dir parameter. Separate to download_dir to # permit only keeping wheel archives for pip wheel. if wheel_download_dir: wheel_download_dir = normalize_path(wheel_download_dir) self.wheel_download_dir = wheel_download_dir # NOTE # download_dir and wheel_download_dir overlap semantically and may # be combined if we're willing to have non-wheel archives present in # the wheelhouse output by 'pip wheel'. self.progress_bar = progress_bar # Is build isolation allowed? self.build_isolation = build_isolation @property def _download_should_save(self): # type: () -> bool if not self.download_dir: return False if os.path.exists(self.download_dir): return True logger.critical('Could not find download directory') raise InstallationError( "Could not find or access download directory '%s'" % display_path(self.download_dir)) def prepare_linked_requirement( self, req, # type: InstallRequirement session, # type: PipSession finder, # type: PackageFinder require_hashes, # type: bool ): # type: (...) -> AbstractDistribution """Prepare a requirement that would be obtained from req.link """ assert req.link link = req.link # TODO: Breakup into smaller functions if link.scheme == 'file': path = link.file_path logger.info('Processing %s', display_path(path)) else: logger.info('Collecting %s', req.req or req) with indent_log(): # @@ if filesystem packages are not marked # editable in a req, a non deterministic error # occurs when the script attempts to unpack the # build directory req.ensure_has_source_dir(self.build_dir) # If a checkout exists, it's unwise to keep going. version # inconsistencies are logged later, but do not fail the # installation. # FIXME: this won't upgrade when there's an existing # package unpacked in `req.source_dir` if os.path.exists(os.path.join(req.source_dir, 'setup.py')): raise PreviousBuildDirError( "pip can't proceed with requirements '%s' due to a" " pre-existing build directory (%s). This is " "likely due to a previous installation that failed" ". pip is being responsible and not assuming it " "can delete this. Please delete it and try again." % (req, req.source_dir) ) # Now that we have the real link, we can tell what kind of # requirements we have and raise some more informative errors # than otherwise. (For example, we can raise VcsHashUnsupported # for a VCS URL rather than HashMissing.) if require_hashes: # We could check these first 2 conditions inside # unpack_url and save repetition of conditions, but then # we would report less-useful error messages for # unhashable requirements, complaining that there's no # hash provided. if link.is_vcs: raise VcsHashUnsupported() elif link.is_existing_dir(): raise DirectoryUrlHashUnsupported() if not req.original_link and not req.is_pinned: # Unpinned packages are asking for trouble when a new # version is uploaded. This isn't a security check, but # it saves users a surprising hash mismatch in the # future. # # file:/// URLs aren't pinnable, so don't complain # about them not being pinned. raise HashUnpinned() hashes = req.hashes(trust_internet=not require_hashes) if require_hashes and not hashes: # Known-good hashes are missing for this requirement, so # shim it with a facade object that will provoke hash # computation and then raise a HashMissing exception # showing the user what the hash should be. hashes = MissingHashes() download_dir = self.download_dir if link.is_wheel and self.wheel_download_dir: # when doing 'pip wheel` we download wheels to a # dedicated dir. download_dir = self.wheel_download_dir try: unpack_url( link, req.source_dir, download_dir, session=session, hashes=hashes, progress_bar=self.progress_bar ) except requests.HTTPError as exc: logger.critical( 'Could not install requirement %s because of error %s', req, exc, ) raise InstallationError( 'Could not install requirement %s because of HTTP ' 'error %s for URL %s' % (req, exc, link) ) if link.is_wheel: if download_dir: # When downloading, we only unpack wheels to get # metadata. autodelete_unpacked = True else: # When installing a wheel, we use the unpacked # wheel. autodelete_unpacked = False else: # We always delete unpacked sdists after pip runs. autodelete_unpacked = True if autodelete_unpacked: write_delete_marker_file(req.source_dir) abstract_dist = _get_prepared_distribution( req, self.req_tracker, finder, self.build_isolation, ) if self._download_should_save: # Make a .zip of the source_dir we already created. if link.is_vcs: req.archive(self.download_dir) return abstract_dist def prepare_editable_requirement( self, req, # type: InstallRequirement require_hashes, # type: bool use_user_site, # type: bool finder # type: PackageFinder ): # type: (...) -> AbstractDistribution """Prepare an editable requirement """ assert req.editable, "cannot prepare a non-editable req as editable" logger.info('Obtaining %s', req) with indent_log(): if require_hashes: raise InstallationError( 'The editable requirement %s cannot be installed when ' 'requiring hashes, because there is no single file to ' 'hash.' % req ) req.ensure_has_source_dir(self.src_dir) req.update_editable(not self._download_should_save) abstract_dist = _get_prepared_distribution( req, self.req_tracker, finder, self.build_isolation, ) if self._download_should_save: req.archive(self.download_dir) req.check_if_exists(use_user_site) return abstract_dist def prepare_installed_requirement( self, req, # type: InstallRequirement require_hashes, # type: bool skip_reason # type: str ): # type: (...) -> AbstractDistribution """Prepare an already-installed requirement """ assert req.satisfied_by, "req should have been satisfied but isn't" assert skip_reason is not None, ( "did not get skip reason skipped but req.satisfied_by " "is set to %r" % (req.satisfied_by,) ) logger.info( 'Requirement %s: %s (%s)', skip_reason, req, req.satisfied_by.version ) with indent_log(): if require_hashes: logger.debug( 'Since it is already installed, we are trusting this ' 'package without checking its hash. To ensure a ' 'completely repeatable environment, install into an ' 'empty virtualenv.' ) abstract_dist = InstalledDistribution(req) return abstract_dist site-packages/pip/_internal/operations/__pycache__/freeze.cpython-38.opt-1.pyc000064400000013142151464452710023303 0ustar00U .eb& @s4ddlmZddlZddlZddlZddlZddlmZddlm Z ddl m Z ddl m Z mZddlmZmZddlmZdd lmZmZdd lmZerdd lmZmZmZmZmZmZm Z m!Z!m"Z"dd l#m$Z$dd l m%Z%m&Z&e ee"e'e&fe(ee'fZ)e*e+Z,dddZ-ddZ.Gddde/Z0dS))absolute_importN)six)canonicalize_name)RequirementParseError) BadCommandInstallationError)install_req_from_editableinstall_req_from_line) COMMENT_RE)dist_is_editableget_installed_distributions)MYPY_CHECK_RUNNING) IteratorOptionalList ContainerSetDictTupleIterableUnion) WheelCache) Distribution RequirementFc  cs|pg}d} |rt|j} |D]} d| Vq i} t|d||dD]b} zt| }Wn8tk r}ztd| |WYqDW5d}~XYnX|r|j rqD|| |j <qD|rt }t t}|D]}t|}|D]r}|r|ds| r| |s|dr6|}||kr|||Vq|dsN|dr|drl|d d}n|tddd }t|||d }nttd |||d }|j std ||tdq|j | kr ||j std|td ||j n||j |qt| |j V| |j =||j |qW5QRXqt|D]4\}}t|dkrdtd|dt t |qddVt | !dddD]$}t"|j | krt|VqdS)Nz-f %sr) local_onlyskip user_onlypathsz6Could not generate requirement for distribution %r: %s#) z-rz --requirementz-Zz--always-unzipz-fz --find-linksz-iz --index-urlz--prez--trusted-hostz--process-dependency-linksz--extra-index-urlz-ez --editable=)isolated wheel_cachezWSkipping line in requirement file [%s] because it's not clear what it would install: %sz9 (add #egg=PackageName to the URL to avoid this warning)zBRequirement file [%s] contains %s, but package %r is not installedz+Requirement %s included multiple times [%s]z, z7## The following requirements were added by pip freeze:cSs |jSN)namelower)xrrC/usr/lib/python3.8/site-packages/pip/_internal/operations/freeze.pyzfreeze..)key)#recompilesearchr FrozenRequirement from_distrloggerwarningeditabler'set collections defaultdictlistopenstrip startswithrstripaddlenlstriprr r subinfoappendstrrZ iteritemsjoinsortedvaluesr)Z requirementZ find_linksrrrZ skip_regexr"r#Zexclude_editablerZ skip_matchlinkZ installationsdistreqexcZemitted_optionsZ req_filesZ req_file_pathZreq_filelineZline_reqr'filesZ installationrrr*freeze+s                  rNc CsNt|sddgfStjtj|j}ddlm}m}| |}|dkrv| }t d||d |g}|d|fSz|||j}Wn|k r| }d t|j|g}|d|fYStk rt d ||jddgfYStk r}zt d |W5d}~XYnX|dk r2|dgfSt d |d g}dd|fS) zk Compute and return values (req, editable, comments) for use in FrozenRequirement.from_dist(). NFr)vcsRemoteNotFoundErrorz1No VCS found for editable requirement "%s" in: %rz/# Editable install with no version control ({})Tz)# Editable {} install with no remote ({})zPcannot determine version of editable source in %s (%s command not found in path)zYError when trying to get requirement for VCS system %s, falling back to uneditable formatz-Could not determine repository location of %sz-## !! Could not determine repository location)r ospathnormcaseabspathlocationZpip._internal.vcsrOrPZget_backend_for_diras_requirementr3debugformatZget_src_requirement project_nametype__name__rr4r'r)rIrUrOrPZ vcs_backendrJcommentsrKrrr*get_requirement_infosZ     r]c@s*eZdZd ddZeddZddZdS) r1rcCs||_||_||_||_dSr&)r'rJr5r\)selfr'rJr5r\rrr*__init__szFrozenRequirement.__init__cCs0t|\}}}|dkr|}||j|||dS)N)r\)r]rVrY)clsrIrJr5r\rrr*r2szFrozenRequirement.from_distcCs2|j}|jrd|}dt|jt|gdS)Nz-e %s )rJr5rEr9r\rD)r^rJrrr*__str__szFrozenRequirement.__str__N)r)r[ __module__ __qualname__r_ classmethodr2rbrrrr*r1s  r1) NNNNNNFNFr)1Z __future__rr7ZloggingrQr.Z pip._vendorrZpip._vendor.packaging.utilsrZpip._vendor.pkg_resourcesrZpip._internal.exceptionsrrZpip._internal.req.constructorsrr Zpip._internal.req.req_filer Zpip._internal.utils.miscr r Zpip._internal.utils.typingr typingrrrrrrrrrZpip._internal.cacherrrrDboolZRequirementInfoZ getLoggerr[r3rNr]objectr1rrrr*s@      ,   >site-packages/pip/_internal/operations/__pycache__/check.cpython-38.pyc000064400000007110151464452710022137 0ustar00U .e@s dZddlZddlmZddlmZddlmZddlm Z ddl m Z ddl m Z eeZe rdd lmZdd lmZmZmZmZmZmZmZeed fZeeefZeeeefZeeeefZeeeefZ eee fZ!ed d d gZ"ddZ#dddZ$ddZ%ddZ&ddZ'dS)z'Validation of dependencies of packages N) namedtuple)canonicalize_name)RequirementParseError))make_distribution_for_install_requirement)get_installed_distributions)MYPY_CHECK_RUNNING)InstallRequirement)AnyCallableDictOptionalSetTupleListPackageDetailsversionrequiresc Ks|ikrddd}i}d}tf|D]\}t|j}zt|j|||<Wq$tk r~}ztd||d}W5d}~XYq$Xq$||fS)z8Converts a list of distributions into a PackageSet. F)Z local_onlyskipz%Error parsing requirements for %s: %sTN) rr project_namerrrrloggingZwarning)kwargs package_setZproblemsdistnameerrB/usr/lib/python3.8/site-packages/pip/_internal/operations/check.py!create_package_set_from_installed(s  rc Cs|dkrdd}i}i}|D]}t}t}||r6q||jD]l}t|j}||krd} |jdk rn|j} | r@|||fq@||j} |jj | dds@||| |fq@|rt |t d||<|rt |t d||<q||fS)zCheck if a package set is consistent If should_ignore is passed, it should be a callable that takes a package name and returns a boolean. NcSsdS)NFrrrrr should_ignoreEsz(check_package_set..should_ignoreT)Z prereleases)key) setrrrZmarkerZevaluateaddrZ specifiercontainssortedstr) rrZmissingZ conflicting package_nameZ missing_depsZconflicting_depsreqrZmissedrrrrcheck_package_set=s4    r(cs6t\}}t||}t|||t|fdddfS)zeFor checking if the dependency graph would be consistent after installing given requirements cs|kSNrrZ whitelistrr|z)check_install_conflicts..)r)r_simulate_installation_of_create_whitelistr() to_installr_would_be_installedrr*rcheck_install_conflictsls    r2cCsLt}|D]<}t|}|}t|j}t|j|||<||q |S)zBComputes the version of packages after installing to_install. ) r!rZget_pkg_resources_distributionrr rrrr")r/rZ installedZinst_reqZ abstract_distrrrrrr-s  r-cCsLt|}|D]:}||krq ||jD] }t|j|kr$||q q$q |Sr))r!rrrr")r1rZpackages_affectedr&r'rrrr.s r.)N)(__doc__r collectionsrZpip._vendor.packaging.utilsrZpip._vendor.pkg_resourcesrZpip._internal.distributionsrZpip._internal.utils.miscrZpip._internal.utils.typingrZ getLogger__name__ZloggerZpip._internal.req.req_installrtypingr r r r r rrr%Z PackageSetZMissingZ ConflictingZ MissingDictZConflictingDictZ CheckResultrrr(r2r-r.rrrrs.        $    /site-packages/pip/_internal/operations/__pycache__/__init__.cpython-38.pyc000064400000000240151464452710022616 0ustar00U .e@sdS)NrrrE/usr/lib/python3.8/site-packages/pip/_internal/operations/__init__.pysite-packages/pip/_internal/operations/__pycache__/check.cpython-38.opt-1.pyc000064400000007110151464452710023076 0ustar00U .e@s dZddlZddlmZddlmZddlmZddlm Z ddl m Z ddl m Z eeZe rdd lmZdd lmZmZmZmZmZmZmZeed fZeeefZeeeefZeeeefZeeeefZ eee fZ!ed d d gZ"ddZ#dddZ$ddZ%ddZ&ddZ'dS)z'Validation of dependencies of packages N) namedtuple)canonicalize_name)RequirementParseError))make_distribution_for_install_requirement)get_installed_distributions)MYPY_CHECK_RUNNING)InstallRequirement)AnyCallableDictOptionalSetTupleListPackageDetailsversionrequiresc Ks|ikrddd}i}d}tf|D]\}t|j}zt|j|||<Wq$tk r~}ztd||d}W5d}~XYq$Xq$||fS)z8Converts a list of distributions into a PackageSet. F)Z local_onlyskipz%Error parsing requirements for %s: %sTN) rr project_namerrrrloggingZwarning)kwargs package_setZproblemsdistnameerrB/usr/lib/python3.8/site-packages/pip/_internal/operations/check.py!create_package_set_from_installed(s  rc Cs|dkrdd}i}i}|D]}t}t}||r6q||jD]l}t|j}||krd} |jdk rn|j} | r@|||fq@||j} |jj | dds@||| |fq@|rt |t d||<|rt |t d||<q||fS)zCheck if a package set is consistent If should_ignore is passed, it should be a callable that takes a package name and returns a boolean. NcSsdS)NFrrrrr should_ignoreEsz(check_package_set..should_ignoreT)Z prereleases)key) setrrrZmarkerZevaluateaddrZ specifiercontainssortedstr) rrZmissingZ conflicting package_nameZ missing_depsZconflicting_depsreqrZmissedrrrrcheck_package_set=s4    r(cs6t\}}t||}t|||t|fdddfS)zeFor checking if the dependency graph would be consistent after installing given requirements cs|kSNrrZ whitelistrr|z)check_install_conflicts..)r)r_simulate_installation_of_create_whitelistr() to_installr_would_be_installedrr*rcheck_install_conflictsls    r2cCsLt}|D]<}t|}|}t|j}t|j|||<||q |S)zBComputes the version of packages after installing to_install. ) r!rZget_pkg_resources_distributionrr rrrr")r/rZ installedZinst_reqZ abstract_distrrrrrr-s  r-cCsLt|}|D]:}||krq ||jD] }t|j|kr$||q q$q |Sr))r!rrrr")r1rZpackages_affectedr&r'rrrr.s r.)N)(__doc__r collectionsrZpip._vendor.packaging.utilsrZpip._vendor.pkg_resourcesrZpip._internal.distributionsrZpip._internal.utils.miscrZpip._internal.utils.typingrZ getLogger__name__ZloggerZpip._internal.req.req_installrtypingr r r r r rrr%Z PackageSetZMissingZ ConflictingZ MissingDictZConflictingDictZ CheckResultrrr(r2r-r.rrrrs.        $    /site-packages/pip/_internal/operations/__pycache__/prepare.cpython-38.pyc000064400000013410151464452710022520 0ustar00U .e,@sdZddlZddlZddlmZddlmZddlmZddl m Z ddl m Z m Z mZmZmZddlmZdd lmZdd lmZdd lmZdd lmZmZdd lmZerddlmZddlm Z ddl!m"Z"ddl#m$Z$ddl%m&Z&ddl'm(Z(e)e*Z+ddZ,Gddde-Z.dS)z)Prepares a distribution for installation N)requests))make_distribution_for_install_requirement)InstalledDistribution) unpack_url)DirectoryUrlHashUnsupported HashUnpinnedInstallationErrorPreviousBuildDirErrorVcsHashUnsupported) expanduser) MissingHashes) indent_log)write_delete_marker_file) display_pathnormalize_path)MYPY_CHECK_RUNNING)Optional)AbstractDistribution) PackageFinder) PipSession)InstallRequirement)RequirementTrackerc Cs.t|}|||||W5QRX|S)z-Prepare a distribution for installation. )rZtrackZprepare_distribution_metadata)req req_trackerfinderbuild_isolation abstract_distrD/usr/lib/python3.8/site-packages/pip/_internal/operations/prepare.py_get_prepared_distribution,s rcsDeZdZdZfddZeddZddZdd Zd d Z Z S) RequirementPreparerzPrepares a Requirement csTtt|||_||_||_|r,t|}||_|r>t|}||_ ||_ ||_ dS)N) superr __init__src_dir build_dirrr download_dirrwheel_download_dir progress_barr)selfr$r%r#r&r'rr __class__rrr"9s zRequirementPreparer.__init__cCs<|js dStj|jrdStdtdt|jdS)NFTz!Could not find download directoryz0Could not find or access download directory '%s')r%ospathexistsloggercriticalrr)r(rrr_download_should_saveas z)RequirementPreparer._download_should_savec Cs|js t|j}|jdkr2|j}tdt|ntd|jp@|tb| |j t j t j |jdrtd||jf|r|jrtn|rt|js|jst|j| d}|r|st}|j}|jr|jr|j}zt||j||||jdWnFtj k rL} z$t!d|| t"d || |fW5d } ~ XYnX|jrh|rbd } nd } nd } | r|t#|jt$||j%||j&} |j'r|jr|(|jW5QRX| S) zCPrepare a requirement that would be obtained from req.link filez Processing %sz Collecting %szsetup.pyzpip can't proceed with requirements '%s' due to a pre-existing build directory (%s). This is likely due to a previous installation that failed. pip is being responsible and not assuming it can delete this. Please delete it and try again.)Ztrust_internet)sessionhashesr'z4Could not install requirement %s because of error %szDCould not install requirement %s because of HTTP error %s for URL %sNTF))linkAssertionErrorZschemeZ file_pathr.inforrr ensure_has_source_dirr$r+r,r-joinZ source_dirr Zis_vcsr Zis_existing_dirrZ original_linkZ is_pinnedrr3r r%Zis_wheelr&rr'rZ HTTPErrorr/rrrrrr0archive) r(rr2rrequire_hashesr4r,r3r%excZautodelete_unpackedrrrrprepare_linked_requirementos         z.RequirementPreparer.prepare_linked_requirementc Cs|jstdtd|t^|r2td|||j||j t ||j ||j }|j rp| |j||W5QRX|S)z(Prepare an editable requirement z-cannot prepare a non-editable req as editablez Obtaining %szoThe editable requirement %s cannot be installed when requiring hashes, because there is no single file to hash.)Zeditabler5r.r6r rr7r#Zupdate_editabler0rrrr9r%Zcheck_if_exists)r(rr:Z use_user_siterrrrrprepare_editable_requirements*    z0RequirementPreparer.prepare_editable_requirementc Csf|jstd|dk s&td|jftd|||jjt|rPtdt|}W5QRX|S)z1Prepare an already-installed requirement z(req should have been satisfied but isn'tNzAdid not get skip reason skipped but req.satisfied_by is set to %rzRequirement %s: %s (%s)zSince it is already installed, we are trusting this package without checking its hash. To ensure a completely repeatable environment, install into an empty virtualenv.)Z satisfied_byr5r.r6versionr debugr)r(rr:Z skip_reasonrrrrprepare_installed_requirement s&  z1RequirementPreparer.prepare_installed_requirement) __name__ __module__ __qualname____doc__r"propertyr0r<r=r@ __classcell__rrr)rr 5s ( z"r )/rDZloggingr+Z pip._vendorrZpip._internal.distributionsrZ%pip._internal.distributions.installedrZpip._internal.downloadrZpip._internal.exceptionsrrrr r Zpip._internal.utils.compatr Zpip._internal.utils.hashesr Zpip._internal.utils.loggingr Z pip._internal.utils.marker_filesrZpip._internal.utils.miscrrZpip._internal.utils.typingrtypingrrZpip._internal.indexrZpip._internal.network.sessionrZpip._internal.req.req_installrZpip._internal.req.req_trackerrZ getLoggerrAr.robjectr rrrrs.                 site-packages/pip/_internal/operations/__pycache__/freeze.cpython-38.pyc000064400000013142151464452710022344 0ustar00U .eb& @s4ddlmZddlZddlZddlZddlZddlmZddlm Z ddl m Z ddl m Z mZddlmZmZddlmZdd lmZmZdd lmZerdd lmZmZmZmZmZmZm Z m!Z!m"Z"dd l#m$Z$dd l m%Z%m&Z&e ee"e'e&fe(ee'fZ)e*e+Z,dddZ-ddZ.Gddde/Z0dS))absolute_importN)six)canonicalize_name)RequirementParseError) BadCommandInstallationError)install_req_from_editableinstall_req_from_line) COMMENT_RE)dist_is_editableget_installed_distributions)MYPY_CHECK_RUNNING) IteratorOptionalList ContainerSetDictTupleIterableUnion) WheelCache) Distribution RequirementFc  cs|pg}d} |rt|j} |D]} d| Vq i} t|d||dD]b} zt| }Wn8tk r}ztd| |WYqDW5d}~XYnX|r|j rqD|| |j <qD|rt }t t}|D]}t|}|D]r}|r|ds| r| |s|dr6|}||kr|||Vq|dsN|dr|drl|d d}n|tddd }t|||d }nttd |||d }|j std ||tdq|j | kr ||j std|td ||j n||j |qt| |j V| |j =||j |qW5QRXqt|D]4\}}t|dkrdtd|dt t |qddVt | !dddD]$}t"|j | krt|VqdS)Nz-f %sr) local_onlyskip user_onlypathsz6Could not generate requirement for distribution %r: %s#) z-rz --requirementz-Zz--always-unzipz-fz --find-linksz-iz --index-urlz--prez--trusted-hostz--process-dependency-linksz--extra-index-urlz-ez --editable=)isolated wheel_cachezWSkipping line in requirement file [%s] because it's not clear what it would install: %sz9 (add #egg=PackageName to the URL to avoid this warning)zBRequirement file [%s] contains %s, but package %r is not installedz+Requirement %s included multiple times [%s]z, z7## The following requirements were added by pip freeze:cSs |jSN)namelower)xrrC/usr/lib/python3.8/site-packages/pip/_internal/operations/freeze.pyzfreeze..)key)#recompilesearchr FrozenRequirement from_distrloggerwarningeditabler'set collections defaultdictlistopenstrip startswithrstripaddlenlstriprr r subinfoappendstrrZ iteritemsjoinsortedvaluesr)Z requirementZ find_linksrrrZ skip_regexr"r#Zexclude_editablerZ skip_matchlinkZ installationsdistreqexcZemitted_optionsZ req_filesZ req_file_pathZreq_filelineZline_reqr'filesZ installationrrr*freeze+s                  rNc CsNt|sddgfStjtj|j}ddlm}m}| |}|dkrv| }t d||d |g}|d|fSz|||j}Wn|k r| }d t|j|g}|d|fYStk rt d ||jddgfYStk r}zt d |W5d}~XYnX|dk r2|dgfSt d |d g}dd|fS) zk Compute and return values (req, editable, comments) for use in FrozenRequirement.from_dist(). NFr)vcsRemoteNotFoundErrorz1No VCS found for editable requirement "%s" in: %rz/# Editable install with no version control ({})Tz)# Editable {} install with no remote ({})zPcannot determine version of editable source in %s (%s command not found in path)zYError when trying to get requirement for VCS system %s, falling back to uneditable formatz-Could not determine repository location of %sz-## !! Could not determine repository location)r ospathnormcaseabspathlocationZpip._internal.vcsrOrPZget_backend_for_diras_requirementr3debugformatZget_src_requirement project_nametype__name__rr4r'r)rIrUrOrPZ vcs_backendrJcommentsrKrrr*get_requirement_infosZ     r]c@s*eZdZd ddZeddZddZdS) r1rcCs||_||_||_||_dSr&)r'rJr5r\)selfr'rJr5r\rrr*__init__szFrozenRequirement.__init__cCs0t|\}}}|dkr|}||j|||dS)N)r\)r]rVrY)clsrIrJr5r\rrr*r2szFrozenRequirement.from_distcCs2|j}|jrd|}dt|jt|gdS)Nz-e %s )rJr5rEr9r\rD)r^rJrrr*__str__szFrozenRequirement.__str__N)r)r[ __module__ __qualname__r_ classmethodr2rbrrrr*r1s  r1) NNNNNNFNFr)1Z __future__rr7ZloggingrQr.Z pip._vendorrZpip._vendor.packaging.utilsrZpip._vendor.pkg_resourcesrZpip._internal.exceptionsrrZpip._internal.req.constructorsrr Zpip._internal.req.req_filer Zpip._internal.utils.miscr r Zpip._internal.utils.typingr typingrrrrrrrrrZpip._internal.cacherrrrDboolZRequirementInfoZ getLoggerr[r3rNr]objectr1rrrr*s@      ,   >site-packages/pip/_internal/operations/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000240151464452710023555 0ustar00U .e@sdS)NrrrE/usr/lib/python3.8/site-packages/pip/_internal/operations/__init__.pysite-packages/pip/_internal/operations/__pycache__/prepare.cpython-38.opt-1.pyc000064400000012771151464452710023470 0ustar00U .e,@sdZddlZddlZddlmZddlmZddlmZddl m Z ddl m Z m Z mZmZmZddlmZdd lmZdd lmZdd lmZdd lmZmZdd lmZerddlmZddlm Z ddl!m"Z"ddl#m$Z$ddl%m&Z&ddl'm(Z(e)e*Z+ddZ,Gddde-Z.dS)z)Prepares a distribution for installation N)requests))make_distribution_for_install_requirement)InstalledDistribution) unpack_url)DirectoryUrlHashUnsupported HashUnpinnedInstallationErrorPreviousBuildDirErrorVcsHashUnsupported) expanduser) MissingHashes) indent_log)write_delete_marker_file) display_pathnormalize_path)MYPY_CHECK_RUNNING)Optional)AbstractDistribution) PackageFinder) PipSession)InstallRequirement)RequirementTrackerc Cs.t|}|||||W5QRX|S)z-Prepare a distribution for installation. )rZtrackZprepare_distribution_metadata)req req_trackerfinderbuild_isolation abstract_distrD/usr/lib/python3.8/site-packages/pip/_internal/operations/prepare.py_get_prepared_distribution,s rcsDeZdZdZfddZeddZddZdd Zd d Z Z S) RequirementPreparerzPrepares a Requirement csTtt|||_||_||_|r,t|}||_|r>t|}||_ ||_ ||_ dS)N) superr __init__src_dir build_dirrr download_dirrwheel_download_dir progress_barr)selfr$r%r#r&r'rr __class__rrr"9s zRequirementPreparer.__init__cCs<|js dStj|jrdStdtdt|jdS)NFTz!Could not find download directoryz0Could not find or access download directory '%s')r%ospathexistsloggercriticalrr)r(rrr_download_should_saveas z)RequirementPreparer._download_should_savec Cs|j}|jdkr(|j}tdt|ntd|jp6|tb||j t j t j |jdrztd||jf|r|jrtn|rt|js|jst|j| d}|r|st}|j}|jr|jr|j}zt||j||||jdWnFtjk rB} z$t d|| t!d || |fW5d } ~ XYnX|jr^|rXd } nd } nd } | rrt"|jt#||j$||j%} |j&r|jr|'|jW5QRX| S) zCPrepare a requirement that would be obtained from req.link filez Processing %sz Collecting %szsetup.pyzpip can't proceed with requirements '%s' due to a pre-existing build directory (%s). This is likely due to a previous installation that failed. pip is being responsible and not assuming it can delete this. Please delete it and try again.)Ztrust_internet)sessionhashesr'z4Could not install requirement %s because of error %szDCould not install requirement %s because of HTTP error %s for URL %sNTF)(linkZschemeZ file_pathr.inforrr ensure_has_source_dirr$r+r,r-joinZ source_dirr Zis_vcsr Zis_existing_dirrZ original_linkZ is_pinnedrr3r r%Zis_wheelr&rr'rZ HTTPErrorr/rrrrrr0archive) r(rr2rrequire_hashesr4r,r3r%excZautodelete_unpackedrrrrprepare_linked_requirementos~         z.RequirementPreparer.prepare_linked_requirementc Csztd|t^|r$td|||j||j t||j ||j }|jrb| |j | |W5QRX|S)z(Prepare an editable requirement z Obtaining %szoThe editable requirement %s cannot be installed when requiring hashes, because there is no single file to hash.)r.r5r rr6r#Zupdate_editabler0rrrr8r%Zcheck_if_exists)r(rr9Z use_user_siterrrrrprepare_editable_requirements(   z0RequirementPreparer.prepare_editable_requirementc Cs@td|||jjt|r*tdt|}W5QRX|S)z1Prepare an already-installed requirement zRequirement %s: %s (%s)zSince it is already installed, we are trusting this package without checking its hash. To ensure a completely repeatable environment, install into an empty virtualenv.)r.r5Z satisfied_byversionr debugr)r(rr9Z skip_reasonrrrrprepare_installed_requirement sz1RequirementPreparer.prepare_installed_requirement) __name__ __module__ __qualname____doc__r"propertyr0r;r<r? __classcell__rrr)rr 5s ( z"r )/rCZloggingr+Z pip._vendorrZpip._internal.distributionsrZ%pip._internal.distributions.installedrZpip._internal.downloadrZpip._internal.exceptionsrrrr r Zpip._internal.utils.compatr Zpip._internal.utils.hashesr Zpip._internal.utils.loggingr Z pip._internal.utils.marker_filesrZpip._internal.utils.miscrrZpip._internal.utils.typingrtypingrrZpip._internal.indexrZpip._internal.network.sessionrZpip._internal.req.req_installrZpip._internal.req.req_trackerrZ getLoggerr@r.robjectr rrrrs.                 site-packages/pip/_internal/operations/__pycache__/generate_metadata.cpython-38.pyc000064400000007463151464452710024527 0ustar00U .e[@sdZddlZddlZddlmZddlmZddlmZddl m Z ddl m Z ddl mZe r|dd lmZmZdd lmZeeZd d Zd dZddZddZdS)z4Metadata generation logic for source distributions. N)InstallationError) ensure_dir)make_setuptools_shim_args)call_subprocess)MYPY_CHECK_RUNNING)vcs)CallableList)InstallRequirementcCs|js tStS)aReturn a callable metadata generator for this InstallRequirement. A metadata generator takes an InstallRequirement (install_req) as an input, generates metadata via the appropriate process for that install_req and returns the generated metadata directory. )Z use_pep517_generate_metadata_legacy_generate_metadata install_reqrN/usr/lib/python3.8/site-packages/pip/_internal/operations/generate_metadata.pyget_metadata_generatorsrcsddfdd}dd}|}|r.||}ntj|d}t|}|sVtd|t|d krn|j|d tj||d S) zEFind an .egg-info in `source_directory`, based on `is_editable`. cSs0tjtj|ddp.tjtj|ddS)NbinpythonZScriptsz Python.exe)ospathlexistsjoinexists)rrrrlooks_like_virtual_env(sz._find_egg_info..looks_like_virtual_envcsg}t|D]\}}tjD]}||kr||qt|D]<}tj|r`||q>|dksp|dkr>||q>|fdd|Dqdd|DS)NZtestZtestsc3s|]}tj|VqdSN)rrr).0dir_rootrr ?szC_find_egg_info..locate_editable_egg_info..cSsg|]}|dr|qS)z .egg-info)endswith)rfrrr @s zD_find_egg_info..locate_editable_egg_info..) rwalkrZdirnamesremovelistrrextend)baseZ candidatesdirsfilesrrrrlocate_editable_egg_info/s     z0_find_egg_info..locate_editable_egg_infocSs(|tjjtjjr"|tjjp$dS)Nr)countrrsepaltsep)rrrrdepth_of_directoryBs z*_find_egg_info..depth_of_directory pip-egg-infoz!Files/directories not found in %s)keyr)rrrlistdirrlensort)Zsource_directoryZ is_editabler+r/r' filenamesrr*r_find_egg_info#s     r7c Cs|jpd|j}td|j|t|j}|jr<|dg7}g}|jsft j |j d}d|g}t ||j t|dg||j ddW5QRXt|j |jS) Nzfrom {}z2Running setup.py (path:%s) egg_info for package %sz --no-user-cfgr0z --egg-baseZegg_infozpython setup.py egg_info)cwdZ command_desc)nameformatlinkloggerdebugZ setup_py_pathrisolatedZeditablerrrZunpacked_source_directoryrZ build_envrr7)rZreq_details_strZbase_cmdZegg_base_optionZ egg_info_dirrrrr ^s6   r cCs|Sr)Zprepare_pep517_metadatar rrrr sr )__doc__ZloggingrZpip._internal.exceptionsrZpip._internal.utils.miscrZ$pip._internal.utils.setuptools_buildrZpip._internal.utils.subprocessrZpip._internal.utils.typingrZpip._internal.vcsrtypingrr Zpip._internal.req.req_installr Z getLogger__name__r<rr7r r rrrrs         ;(site-packages/pip/_internal/operations/__pycache__/generate_metadata.cpython-38.opt-1.pyc000064400000007463151464452710025466 0ustar00U .e[@sdZddlZddlZddlmZddlmZddlmZddl m Z ddl m Z ddl mZe r|dd lmZmZdd lmZeeZd d Zd dZddZddZdS)z4Metadata generation logic for source distributions. N)InstallationError) ensure_dir)make_setuptools_shim_args)call_subprocess)MYPY_CHECK_RUNNING)vcs)CallableList)InstallRequirementcCs|js tStS)aReturn a callable metadata generator for this InstallRequirement. A metadata generator takes an InstallRequirement (install_req) as an input, generates metadata via the appropriate process for that install_req and returns the generated metadata directory. )Z use_pep517_generate_metadata_legacy_generate_metadata install_reqrN/usr/lib/python3.8/site-packages/pip/_internal/operations/generate_metadata.pyget_metadata_generatorsrcsddfdd}dd}|}|r.||}ntj|d}t|}|sVtd|t|d krn|j|d tj||d S) zEFind an .egg-info in `source_directory`, based on `is_editable`. cSs0tjtj|ddp.tjtj|ddS)NbinpythonZScriptsz Python.exe)ospathlexistsjoinexists)rrrrlooks_like_virtual_env(sz._find_egg_info..looks_like_virtual_envcsg}t|D]\}}tjD]}||kr||qt|D]<}tj|r`||q>|dksp|dkr>||q>|fdd|Dqdd|DS)NZtestZtestsc3s|]}tj|VqdSN)rrr).0dir_rootrr ?szC_find_egg_info..locate_editable_egg_info..cSsg|]}|dr|qS)z .egg-info)endswith)rfrrr @s zD_find_egg_info..locate_editable_egg_info..) rwalkrZdirnamesremovelistrrextend)baseZ candidatesdirsfilesrrrrlocate_editable_egg_info/s     z0_find_egg_info..locate_editable_egg_infocSs(|tjjtjjr"|tjjp$dS)Nr)countrrsepaltsep)rrrrdepth_of_directoryBs z*_find_egg_info..depth_of_directory pip-egg-infoz!Files/directories not found in %s)keyr)rrrlistdirrlensort)Zsource_directoryZ is_editabler+r/r' filenamesrr*r_find_egg_info#s     r7c Cs|jpd|j}td|j|t|j}|jr<|dg7}g}|jsft j |j d}d|g}t ||j t|dg||j ddW5QRXt|j |jS) Nzfrom {}z2Running setup.py (path:%s) egg_info for package %sz --no-user-cfgr0z --egg-baseZegg_infozpython setup.py egg_info)cwdZ command_desc)nameformatlinkloggerdebugZ setup_py_pathrisolatedZeditablerrrZunpacked_source_directoryrZ build_envrr7)rZreq_details_strZbase_cmdZegg_base_optionZ egg_info_dirrrrr ^s6   r cCs|Sr)Zprepare_pep517_metadatar rrrr sr )__doc__ZloggingrZpip._internal.exceptionsrZpip._internal.utils.miscrZ$pip._internal.utils.setuptools_buildrZpip._internal.utils.subprocessrZpip._internal.utils.typingrZpip._internal.vcsrtypingrr Zpip._internal.req.req_installr Z getLogger__name__r<rr7r r rrrrs         ;(site-packages/pip/_internal/operations/__init__.py000064400000000000151464452710016322 0ustar00site-packages/pip/_internal/configuration.py000064400000033612151464452710015266 0ustar00"""Configuration management setup Some terminology: - name As written in config files. - value Value associated with a name - key Name combined with it's section (section.name) - variant A single word describing where the configuration key-value pair came from """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False import locale import logging import os import sys from pip._vendor.six.moves import configparser from pip._internal.exceptions import ( ConfigurationError, ConfigurationFileCouldNotBeLoaded, ) from pip._internal.utils import appdirs from pip._internal.utils.compat import WINDOWS, expanduser from pip._internal.utils.misc import ensure_dir, enum from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ( Any, Dict, Iterable, List, NewType, Optional, Tuple ) RawConfigParser = configparser.RawConfigParser # Shorthand Kind = NewType("Kind", str) logger = logging.getLogger(__name__) # NOTE: Maybe use the optionx attribute to normalize keynames. def _normalize_name(name): # type: (str) -> str """Make a name consistent regardless of source (environment or file) """ name = name.lower().replace('_', '-') if name.startswith('--'): name = name[2:] # only prefer long opts return name def _disassemble_key(name): # type: (str) -> List[str] if "." not in name: error_message = ( "Key does not contain dot separated section and key. " "Perhaps you wanted to use 'global.{}' instead?" ).format(name) raise ConfigurationError(error_message) return name.split(".", 1) # The kinds of configurations there are. kinds = enum( USER="user", # User Specific GLOBAL="global", # System Wide SITE="site", # [Virtual] Environment Specific ENV="env", # from PIP_CONFIG_FILE ENV_VAR="env-var", # from Environment Variables ) CONFIG_BASENAME = 'pip.ini' if WINDOWS else 'pip.conf' def get_configuration_files(): global_config_files = [ os.path.join(path, CONFIG_BASENAME) for path in appdirs.site_config_dirs('pip') ] site_config_file = os.path.join(sys.prefix, CONFIG_BASENAME) legacy_config_file = os.path.join( expanduser('~'), 'pip' if WINDOWS else '.pip', CONFIG_BASENAME, ) new_config_file = os.path.join( appdirs.user_config_dir("pip"), CONFIG_BASENAME ) return { kinds.GLOBAL: global_config_files, kinds.SITE: [site_config_file], kinds.USER: [legacy_config_file, new_config_file], } class Configuration(object): """Handles management of configuration. Provides an interface to accessing and managing configuration files. This class converts provides an API that takes "section.key-name" style keys and stores the value associated with it as "key-name" under the section "section". This allows for a clean interface wherein the both the section and the key-name are preserved in an easy to manage form in the configuration files and the data stored is also nice. """ def __init__(self, isolated, load_only=None): # type: (bool, Kind) -> None super(Configuration, self).__init__() _valid_load_only = [kinds.USER, kinds.GLOBAL, kinds.SITE, None] if load_only not in _valid_load_only: raise ConfigurationError( "Got invalid value for load_only - should be one of {}".format( ", ".join(map(repr, _valid_load_only[:-1])) ) ) self.isolated = isolated # type: bool self.load_only = load_only # type: Optional[Kind] # The order here determines the override order. self._override_order = [ kinds.GLOBAL, kinds.USER, kinds.SITE, kinds.ENV, kinds.ENV_VAR ] self._ignore_env_names = ["version", "help"] # Because we keep track of where we got the data from self._parsers = { variant: [] for variant in self._override_order } # type: Dict[Kind, List[Tuple[str, RawConfigParser]]] self._config = { variant: {} for variant in self._override_order } # type: Dict[Kind, Dict[str, Any]] self._modified_parsers = [] # type: List[Tuple[str, RawConfigParser]] def load(self): # type: () -> None """Loads configuration from configuration files and environment """ self._load_config_files() if not self.isolated: self._load_environment_vars() def get_file_to_edit(self): # type: () -> Optional[str] """Returns the file with highest priority in configuration """ assert self.load_only is not None, \ "Need to be specified a file to be editing" try: return self._get_parser_to_modify()[0] except IndexError: return None def items(self): # type: () -> Iterable[Tuple[str, Any]] """Returns key-value pairs like dict.items() representing the loaded configuration """ return self._dictionary.items() def get_value(self, key): # type: (str) -> Any """Get a value from the configuration. """ try: return self._dictionary[key] except KeyError: raise ConfigurationError("No such key - {}".format(key)) def set_value(self, key, value): # type: (str, Any) -> None """Modify a value in the configuration. """ self._ensure_have_load_only() fname, parser = self._get_parser_to_modify() if parser is not None: section, name = _disassemble_key(key) # Modify the parser and the configuration if not parser.has_section(section): parser.add_section(section) parser.set(section, name, value) self._config[self.load_only][key] = value self._mark_as_modified(fname, parser) def unset_value(self, key): # type: (str) -> None """Unset a value in the configuration. """ self._ensure_have_load_only() if key not in self._config[self.load_only]: raise ConfigurationError("No such key - {}".format(key)) fname, parser = self._get_parser_to_modify() if parser is not None: section, name = _disassemble_key(key) # Remove the key in the parser modified_something = False if parser.has_section(section): # Returns whether the option was removed or not modified_something = parser.remove_option(section, name) if modified_something: # name removed from parser, section may now be empty section_iter = iter(parser.items(section)) try: val = next(section_iter) except StopIteration: val = None if val is None: parser.remove_section(section) self._mark_as_modified(fname, parser) else: raise ConfigurationError( "Fatal Internal error [id=1]. Please report as a bug." ) del self._config[self.load_only][key] def save(self): # type: () -> None """Save the current in-memory state. """ self._ensure_have_load_only() for fname, parser in self._modified_parsers: logger.info("Writing to %s", fname) # Ensure directory exists. ensure_dir(os.path.dirname(fname)) with open(fname, "w") as f: parser.write(f) # # Private routines # def _ensure_have_load_only(self): # type: () -> None if self.load_only is None: raise ConfigurationError("Needed a specific file to be modifying.") logger.debug("Will be working with %s variant only", self.load_only) @property def _dictionary(self): # type: () -> Dict[str, Any] """A dictionary representing the loaded configuration. """ # NOTE: Dictionaries are not populated if not loaded. So, conditionals # are not needed here. retval = {} for variant in self._override_order: retval.update(self._config[variant]) return retval def _load_config_files(self): # type: () -> None """Loads configuration from configuration files """ config_files = dict(self._iter_config_files()) if config_files[kinds.ENV][0:1] == [os.devnull]: logger.debug( "Skipping loading configuration files due to " "environment's PIP_CONFIG_FILE being os.devnull" ) return for variant, files in config_files.items(): for fname in files: # If there's specific variant set in `load_only`, load only # that variant, not the others. if self.load_only is not None and variant != self.load_only: logger.debug( "Skipping file '%s' (variant: %s)", fname, variant ) continue parser = self._load_file(variant, fname) # Keeping track of the parsers used self._parsers[variant].append((fname, parser)) def _load_file(self, variant, fname): # type: (Kind, str) -> RawConfigParser logger.debug("For variant '%s', will try loading '%s'", variant, fname) parser = self._construct_parser(fname) for section in parser.sections(): items = parser.items(section) self._config[variant].update(self._normalized_keys(section, items)) return parser def _construct_parser(self, fname): # type: (str) -> RawConfigParser parser = configparser.RawConfigParser() # If there is no such file, don't bother reading it but create the # parser anyway, to hold the data. # Doing this is useful when modifying and saving files, where we don't # need to construct a parser. if os.path.exists(fname): try: parser.read(fname) except UnicodeDecodeError: # See https://github.com/pypa/pip/issues/4963 raise ConfigurationFileCouldNotBeLoaded( reason="contains invalid {} characters".format( locale.getpreferredencoding(False) ), fname=fname, ) except configparser.Error as error: # See https://github.com/pypa/pip/issues/4893 raise ConfigurationFileCouldNotBeLoaded(error=error) return parser def _load_environment_vars(self): # type: () -> None """Loads configuration from environment variables """ self._config[kinds.ENV_VAR].update( self._normalized_keys(":env:", self._get_environ_vars()) ) def _normalized_keys(self, section, items): # type: (str, Iterable[Tuple[str, Any]]) -> Dict[str, Any] """Normalizes items to construct a dictionary with normalized keys. This routine is where the names become keys and are made the same regardless of source - configuration files or environment. """ normalized = {} for name, val in items: key = section + "." + _normalize_name(name) normalized[key] = val return normalized def _get_environ_vars(self): # type: () -> Iterable[Tuple[str, str]] """Returns a generator with all environmental vars with prefix PIP_""" for key, val in os.environ.items(): should_be_yielded = ( key.startswith("PIP_") and key[4:].lower() not in self._ignore_env_names ) if should_be_yielded: yield key[4:].lower(), val # XXX: This is patched in the tests. def _iter_config_files(self): # type: () -> Iterable[Tuple[Kind, List[str]]] """Yields variant and configuration files associated with it. This should be treated like items of a dictionary. """ # SMELL: Move the conditions out of this function # environment variables have the lowest priority config_file = os.environ.get('PIP_CONFIG_FILE', None) if config_file is not None: yield kinds.ENV, [config_file] else: yield kinds.ENV, [] config_files = get_configuration_files() # at the base we have any global configuration yield kinds.GLOBAL, config_files[kinds.GLOBAL] # per-user configuration next should_load_user_config = not self.isolated and not ( config_file and os.path.exists(config_file) ) if should_load_user_config: # The legacy config file is overridden by the new config file yield kinds.USER, config_files[kinds.USER] # finally virtualenv configuration first trumping others yield kinds.SITE, config_files[kinds.SITE] def _get_parser_to_modify(self): # type: () -> Tuple[str, RawConfigParser] # Determine which parser to modify parsers = self._parsers[self.load_only] if not parsers: # This should not happen if everything works correctly. raise ConfigurationError( "Fatal Internal error [id=2]. Please report as a bug." ) # Use the highest priority parser. return parsers[-1] # XXX: This is patched in the tests. def _mark_as_modified(self, fname, parser): # type: (str, RawConfigParser) -> None file_parser_tuple = (fname, parser) if file_parser_tuple not in self._modified_parsers: self._modified_parsers.append(file_parser_tuple) site-packages/pip/_internal/main.py000064400000002517151464452710013343 0ustar00"""Primary application entrypoint. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import locale import logging import os import sys from pip._internal.cli.autocompletion import autocomplete from pip._internal.cli.main_parser import parse_command from pip._internal.commands import create_command from pip._internal.exceptions import PipError from pip._internal.utils import deprecation logger = logging.getLogger(__name__) def main(args=None): if args is None: args = sys.argv[1:] # Configure our deprecation warnings to be sent through loggers deprecation.install_warning_logger() autocomplete() try: cmd_name, cmd_args = parse_command(args) except PipError as exc: sys.stderr.write("ERROR: %s" % exc) sys.stderr.write(os.linesep) sys.exit(1) # Needed for locale.getpreferredencoding(False) to work # in pip._internal.utils.encoding.auto_decode try: locale.setlocale(locale.LC_ALL, '') except locale.Error as e: # setlocale can apparently crash if locale are uninitialized logger.debug("Ignoring error %s when setting locale", e) command = create_command(cmd_name, isolated=("--isolated" in cmd_args)) return command.main(cmd_args) site-packages/pip/_internal/cli/parser.py000064400000022417151464452710014463 0ustar00"""Base option parser setup""" # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import optparse import sys import textwrap from distutils.util import strtobool from pip._vendor.six import string_types from pip._internal.cli.status_codes import UNKNOWN_ERROR from pip._internal.configuration import Configuration, ConfigurationError from pip._internal.utils.compat import get_terminal_size logger = logging.getLogger(__name__) class PrettyHelpFormatter(optparse.IndentedHelpFormatter): """A prettier/less verbose help formatter for optparse.""" def __init__(self, *args, **kwargs): # help position must be aligned with __init__.parseopts.description kwargs['max_help_position'] = 30 kwargs['indent_increment'] = 1 kwargs['width'] = get_terminal_size()[0] - 2 optparse.IndentedHelpFormatter.__init__(self, *args, **kwargs) def format_option_strings(self, option): return self._format_option_strings(option, ' <%s>', ', ') def _format_option_strings(self, option, mvarfmt=' <%s>', optsep=', '): """ Return a comma-separated list of option strings and metavars. :param option: tuple of (short opt, long opt), e.g: ('-f', '--format') :param mvarfmt: metavar format string - evaluated as mvarfmt % metavar :param optsep: separator """ opts = [] if option._short_opts: opts.append(option._short_opts[0]) if option._long_opts: opts.append(option._long_opts[0]) if len(opts) > 1: opts.insert(1, optsep) if option.takes_value(): metavar = option.metavar or option.dest.lower() opts.append(mvarfmt % metavar.lower()) return ''.join(opts) def format_heading(self, heading): if heading == 'Options': return '' return heading + ':\n' def format_usage(self, usage): """ Ensure there is only one newline between usage and the first heading if there is no description. """ msg = '\nUsage: %s\n' % self.indent_lines(textwrap.dedent(usage), " ") return msg def format_description(self, description): # leave full control over description to us if description: if hasattr(self.parser, 'main'): label = 'Commands' else: label = 'Description' # some doc strings have initial newlines, some don't description = description.lstrip('\n') # some doc strings have final newlines and spaces, some don't description = description.rstrip() # dedent, then reindent description = self.indent_lines(textwrap.dedent(description), " ") description = '%s:\n%s\n' % (label, description) return description else: return '' def format_epilog(self, epilog): # leave full control over epilog to us if epilog: return epilog else: return '' def indent_lines(self, text, indent): new_lines = [indent + line for line in text.split('\n')] return "\n".join(new_lines) class UpdatingDefaultsHelpFormatter(PrettyHelpFormatter): """Custom help formatter for use in ConfigOptionParser. This is updates the defaults before expanding them, allowing them to show up correctly in the help listing. """ def expand_default(self, option): if self.parser is not None: self.parser._update_defaults(self.parser.defaults) return optparse.IndentedHelpFormatter.expand_default(self, option) class CustomOptionParser(optparse.OptionParser): def insert_option_group(self, idx, *args, **kwargs): """Insert an OptionGroup at a given position.""" group = self.add_option_group(*args, **kwargs) self.option_groups.pop() self.option_groups.insert(idx, group) return group @property def option_list_all(self): """Get a list of all options, including those in option groups.""" res = self.option_list[:] for i in self.option_groups: res.extend(i.option_list) return res class ConfigOptionParser(CustomOptionParser): """Custom option parser which updates its defaults by checking the configuration files and environmental variables""" def __init__(self, *args, **kwargs): self.name = kwargs.pop('name') isolated = kwargs.pop("isolated", False) self.config = Configuration(isolated) assert self.name optparse.OptionParser.__init__(self, *args, **kwargs) def check_default(self, option, key, val): try: return option.check_value(key, val) except optparse.OptionValueError as exc: print("An error occurred during configuration: %s" % exc) sys.exit(3) def _get_ordered_configuration_items(self): # Configuration gives keys in an unordered manner. Order them. override_order = ["global", self.name, ":env:"] # Pool the options into different groups section_items = {name: [] for name in override_order} for section_key, val in self.config.items(): # ignore empty values if not val: logger.debug( "Ignoring configuration key '%s' as it's value is empty.", section_key ) continue section, key = section_key.split(".", 1) if section in override_order: section_items[section].append((key, val)) # Yield each group in their override order for section in override_order: for key, val in section_items[section]: yield key, val def _update_defaults(self, defaults): """Updates the given defaults with values from the config files and the environ. Does a little special handling for certain types of options (lists).""" # Accumulate complex default state. self.values = optparse.Values(self.defaults) late_eval = set() # Then set the options with those values for key, val in self._get_ordered_configuration_items(): # '--' because configuration supports only long names option = self.get_option('--' + key) # Ignore options not present in this parser. E.g. non-globals put # in [global] by users that want them to apply to all applicable # commands. if option is None: continue if option.action in ('store_true', 'store_false', 'count'): try: val = strtobool(val) except ValueError: error_msg = invalid_config_error_message( option.action, key, val ) self.error(error_msg) elif option.action == 'append': val = val.split() val = [self.check_default(option, key, v) for v in val] elif option.action == 'callback': late_eval.add(option.dest) opt_str = option.get_opt_string() val = option.convert_value(opt_str, val) # From take_action args = option.callback_args or () kwargs = option.callback_kwargs or {} option.callback(option, opt_str, val, self, *args, **kwargs) else: val = self.check_default(option, key, val) defaults[option.dest] = val for key in late_eval: defaults[key] = getattr(self.values, key) self.values = None return defaults def get_default_values(self): """Overriding to make updating the defaults after instantiation of the option parser possible, _update_defaults() does the dirty work.""" if not self.process_default_values: # Old, pre-Optik 1.5 behaviour. return optparse.Values(self.defaults) # Load the configuration, or error out in case of an error try: self.config.load() except ConfigurationError as err: self.exit(UNKNOWN_ERROR, str(err)) defaults = self._update_defaults(self.defaults.copy()) # ours for option in self._get_all_options(): default = defaults.get(option.dest) if isinstance(default, string_types): opt_str = option.get_opt_string() defaults[option.dest] = option.check_value(opt_str, default) return optparse.Values(defaults) def error(self, msg): self.print_usage(sys.stderr) self.exit(UNKNOWN_ERROR, "%s\n" % msg) def invalid_config_error_message(action, key, val): """Returns a better error message when invalid configuration option is provided.""" if action in ('store_true', 'store_false'): return ("{0} is not a valid value for {1} option, " "please specify a boolean value like yes/no, " "true/false or 1/0 instead.").format(val, key) return ("{0} is not a valid value for {1} option, " "please specify a numerical value like 1/0 " "instead.").format(val, key) site-packages/pip/_internal/cli/autocompletion.py000064400000014031151464452710016222 0ustar00"""Logic that powers autocompletion installed by ``pip completion``. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import optparse import os import sys from pip._internal.cli.main_parser import create_main_parser from pip._internal.commands import commands_dict, create_command from pip._internal.utils.misc import get_installed_distributions def autocomplete(): """Entry Point for completion of main and subcommand options. """ # Don't complete if user hasn't sourced bash_completion file. if 'PIP_AUTO_COMPLETE' not in os.environ: return cwords = os.environ['COMP_WORDS'].split()[1:] cword = int(os.environ['COMP_CWORD']) try: current = cwords[cword - 1] except IndexError: current = '' subcommands = list(commands_dict) options = [] # subcommand try: subcommand_name = [w for w in cwords if w in subcommands][0] except IndexError: subcommand_name = None parser = create_main_parser() # subcommand options if subcommand_name: # special case: 'help' subcommand has no options if subcommand_name == 'help': sys.exit(1) # special case: list locally installed dists for show and uninstall should_list_installed = ( subcommand_name in ['show', 'uninstall'] and not current.startswith('-') ) if should_list_installed: installed = [] lc = current.lower() for dist in get_installed_distributions(local_only=True): if dist.key.startswith(lc) and dist.key not in cwords[1:]: installed.append(dist.key) # if there are no dists installed, fall back to option completion if installed: for dist in installed: print(dist) sys.exit(1) subcommand = create_command(subcommand_name) for opt in subcommand.parser.option_list_all: if opt.help != optparse.SUPPRESS_HELP: for opt_str in opt._long_opts + opt._short_opts: options.append((opt_str, opt.nargs)) # filter out previously specified options from available options prev_opts = [x.split('=')[0] for x in cwords[1:cword - 1]] options = [(x, v) for (x, v) in options if x not in prev_opts] # filter options by current input options = [(k, v) for k, v in options if k.startswith(current)] # get completion type given cwords and available subcommand options completion_type = get_path_completion_type( cwords, cword, subcommand.parser.option_list_all, ) # get completion files and directories if ``completion_type`` is # ````, ```` or ```` if completion_type: options = auto_complete_paths(current, completion_type) options = ((opt, 0) for opt in options) for option in options: opt_label = option[0] # append '=' to options which require args if option[1] and option[0][:2] == "--": opt_label += '=' print(opt_label) else: # show main parser options only when necessary opts = [i.option_list for i in parser.option_groups] opts.append(parser.option_list) opts = (o for it in opts for o in it) if current.startswith('-'): for opt in opts: if opt.help != optparse.SUPPRESS_HELP: subcommands += opt._long_opts + opt._short_opts else: # get completion type given cwords and all available options completion_type = get_path_completion_type(cwords, cword, opts) if completion_type: subcommands = auto_complete_paths(current, completion_type) print(' '.join([x for x in subcommands if x.startswith(current)])) sys.exit(1) def get_path_completion_type(cwords, cword, opts): """Get the type of path completion (``file``, ``dir``, ``path`` or None) :param cwords: same as the environmental variable ``COMP_WORDS`` :param cword: same as the environmental variable ``COMP_CWORD`` :param opts: The available options to check :return: path completion type (``file``, ``dir``, ``path`` or None) """ if cword < 2 or not cwords[cword - 2].startswith('-'): return for opt in opts: if opt.help == optparse.SUPPRESS_HELP: continue for o in str(opt).split('/'): if cwords[cword - 2].split('=')[0] == o: if not opt.metavar or any( x in ('path', 'file', 'dir') for x in opt.metavar.split('/')): return opt.metavar def auto_complete_paths(current, completion_type): """If ``completion_type`` is ``file`` or ``path``, list all regular files and directories starting with ``current``; otherwise only list directories starting with ``current``. :param current: The word to be completed :param completion_type: path completion type(`file`, `path` or `dir`)i :return: A generator of regular files and/or directories """ directory, filename = os.path.split(current) current_path = os.path.abspath(directory) # Don't complete paths if they can't be accessed if not os.access(current_path, os.R_OK): return filename = os.path.normcase(filename) # list all files that start with ``filename`` file_list = (x for x in os.listdir(current_path) if os.path.normcase(x).startswith(filename)) for f in file_list: opt = os.path.join(current_path, f) comp_file = os.path.normcase(os.path.join(directory, f)) # complete regular files when there is not ```` after option # complete directories when there is ````, ```` or # ````after option if completion_type != 'dir' and os.path.isfile(opt): yield comp_file elif os.path.isdir(opt): yield os.path.join(comp_file, '') site-packages/pip/_internal/cli/main_parser.py000064400000005403151464452710015463 0ustar00"""A single place for constructing and exposing the main parser """ import os import sys from pip._internal.cli import cmdoptions from pip._internal.cli.parser import ( ConfigOptionParser, UpdatingDefaultsHelpFormatter, ) from pip._internal.commands import commands_dict, get_similar_commands from pip._internal.exceptions import CommandError from pip._internal.utils.misc import get_pip_version, get_prog from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Tuple, List __all__ = ["create_main_parser", "parse_command"] def create_main_parser(): # type: () -> ConfigOptionParser """Creates and returns the main parser for pip's CLI """ parser_kw = { 'usage': '\n%prog [options]', 'add_help_option': False, 'formatter': UpdatingDefaultsHelpFormatter(), 'name': 'global', 'prog': get_prog(), } parser = ConfigOptionParser(**parser_kw) parser.disable_interspersed_args() parser.version = get_pip_version() # add the general options gen_opts = cmdoptions.make_option_group(cmdoptions.general_group, parser) parser.add_option_group(gen_opts) # so the help formatter knows parser.main = True # type: ignore # create command listing for description description = [''] + [ '%-27s %s' % (name, command_info.summary) for name, command_info in commands_dict.items() ] parser.description = '\n'.join(description) return parser def parse_command(args): # type: (List[str]) -> Tuple[str, List[str]] parser = create_main_parser() # Note: parser calls disable_interspersed_args(), so the result of this # call is to split the initial args into the general options before the # subcommand and everything else. # For example: # args: ['--timeout=5', 'install', '--user', 'INITools'] # general_options: ['--timeout==5'] # args_else: ['install', '--user', 'INITools'] general_options, args_else = parser.parse_args(args) # --version if general_options.version: sys.stdout.write(parser.version) # type: ignore sys.stdout.write(os.linesep) sys.exit() # pip || pip help -> print_help() if not args_else or (args_else[0] == 'help' and len(args_else) == 1): parser.print_help() sys.exit() # the subcommand name cmd_name = args_else[0] if cmd_name not in commands_dict: guess = get_similar_commands(cmd_name) msg = ['unknown command "%s"' % cmd_name] if guess: msg.append('maybe you meant "%s"' % guess) raise CommandError(' - '.join(msg)) # all the args without the subcommand cmd_args = args[:] cmd_args.remove(cmd_name) return cmd_name, cmd_args site-packages/pip/_internal/cli/cmdoptions.py000064400000064327151464452710015354 0ustar00""" shared options and groups The principle here is to define options once, but *not* instantiate them globally. One reason being that options with action='append' can carry state between parses. pip parses general options twice internally, and shouldn't pass on state. To be consistent, all options will follow this design. """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import textwrap import warnings from distutils.util import strtobool from functools import partial from optparse import SUPPRESS_HELP, Option, OptionGroup from textwrap import dedent from pip._internal.exceptions import CommandError from pip._internal.locations import USER_CACHE_DIR, get_src_prefix from pip._internal.models.format_control import FormatControl from pip._internal.models.index import PyPI from pip._internal.models.target_python import TargetPython from pip._internal.utils.hashes import STRONG_HASHES from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.ui import BAR_TYPES if MYPY_CHECK_RUNNING: from typing import Any, Callable, Dict, Optional, Tuple from optparse import OptionParser, Values from pip._internal.cli.parser import ConfigOptionParser logger = logging.getLogger(__name__) def raise_option_error(parser, option, msg): """ Raise an option parsing error using parser.error(). Args: parser: an OptionParser instance. option: an Option instance. msg: the error text. """ msg = '{} error: {}'.format(option, msg) msg = textwrap.fill(' '.join(msg.split())) parser.error(msg) def make_option_group(group, parser): # type: (Dict[str, Any], ConfigOptionParser) -> OptionGroup """ Return an OptionGroup object group -- assumed to be dict with 'name' and 'options' keys parser -- an optparse Parser """ option_group = OptionGroup(parser, group['name']) for option in group['options']: option_group.add_option(option()) return option_group def check_install_build_global(options, check_options=None): # type: (Values, Optional[Values]) -> None """Disable wheels if per-setup.py call options are set. :param options: The OptionParser options to update. :param check_options: The options to check, if not supplied defaults to options. """ if check_options is None: check_options = options def getname(n): return getattr(check_options, n, None) names = ["build_options", "global_options", "install_options"] if any(map(getname, names)): control = options.format_control control.disallow_binaries() warnings.warn( 'Disabling all use of wheels due to the use of --build-options ' '/ --global-options / --install-options.', stacklevel=2, ) def check_dist_restriction(options, check_target=False): # type: (Values, bool) -> None """Function for determining if custom platform options are allowed. :param options: The OptionParser options. :param check_target: Whether or not to check if --target is being used. """ dist_restriction_set = any([ options.python_version, options.platform, options.abi, options.implementation, ]) binary_only = FormatControl(set(), {':all:'}) sdist_dependencies_allowed = ( options.format_control != binary_only and not options.ignore_dependencies ) # Installations or downloads using dist restrictions must not combine # source distributions and dist-specific wheels, as they are not # guaranteed to be locally compatible. if dist_restriction_set and sdist_dependencies_allowed: raise CommandError( "When restricting platform and interpreter constraints using " "--python-version, --platform, --abi, or --implementation, " "either --no-deps must be set, or --only-binary=:all: must be " "set and --no-binary must not be set (or must be set to " ":none:)." ) if check_target: if dist_restriction_set and not options.target_dir: raise CommandError( "Can not use any platform or abi specific options unless " "installing via '--target'" ) ########### # options # ########### help_ = partial( Option, '-h', '--help', dest='help', action='help', help='Show help.', ) # type: Callable[..., Option] isolated_mode = partial( Option, "--isolated", dest="isolated_mode", action="store_true", default=False, help=( "Run pip in an isolated mode, ignoring environment variables and user " "configuration." ), ) # type: Callable[..., Option] require_virtualenv = partial( Option, # Run only if inside a virtualenv, bail if not. '--require-virtualenv', '--require-venv', dest='require_venv', action='store_true', default=False, help=SUPPRESS_HELP ) # type: Callable[..., Option] verbose = partial( Option, '-v', '--verbose', dest='verbose', action='count', default=0, help='Give more output. Option is additive, and can be used up to 3 times.' ) # type: Callable[..., Option] no_color = partial( Option, '--no-color', dest='no_color', action='store_true', default=False, help="Suppress colored output", ) # type: Callable[..., Option] version = partial( Option, '-V', '--version', dest='version', action='store_true', help='Show version and exit.', ) # type: Callable[..., Option] quiet = partial( Option, '-q', '--quiet', dest='quiet', action='count', default=0, help=( 'Give less output. Option is additive, and can be used up to 3' ' times (corresponding to WARNING, ERROR, and CRITICAL logging' ' levels).' ), ) # type: Callable[..., Option] progress_bar = partial( Option, '--progress-bar', dest='progress_bar', type='choice', choices=list(BAR_TYPES.keys()), default='on', help=( 'Specify type of progress to be displayed [' + '|'.join(BAR_TYPES.keys()) + '] (default: %default)' ), ) # type: Callable[..., Option] log = partial( Option, "--log", "--log-file", "--local-log", dest="log", metavar="path", help="Path to a verbose appending log." ) # type: Callable[..., Option] no_input = partial( Option, # Don't ask for input '--no-input', dest='no_input', action='store_true', default=False, help=SUPPRESS_HELP ) # type: Callable[..., Option] proxy = partial( Option, '--proxy', dest='proxy', type='str', default='', help="Specify a proxy in the form [user:passwd@]proxy.server:port." ) # type: Callable[..., Option] retries = partial( Option, '--retries', dest='retries', type='int', default=5, help="Maximum number of retries each connection should attempt " "(default %default times).", ) # type: Callable[..., Option] timeout = partial( Option, '--timeout', '--default-timeout', metavar='sec', dest='timeout', type='float', default=15, help='Set the socket timeout (default %default seconds).', ) # type: Callable[..., Option] skip_requirements_regex = partial( Option, # A regex to be used to skip requirements '--skip-requirements-regex', dest='skip_requirements_regex', type='str', default='', help=SUPPRESS_HELP, ) # type: Callable[..., Option] def exists_action(): # type: () -> Option return Option( # Option when path already exist '--exists-action', dest='exists_action', type='choice', choices=['s', 'i', 'w', 'b', 'a'], default=[], action='append', metavar='action', help="Default action when a path already exists: " "(s)witch, (i)gnore, (w)ipe, (b)ackup, (a)bort.", ) cert = partial( Option, '--cert', dest='cert', type='str', metavar='path', help="Path to alternate CA bundle.", ) # type: Callable[..., Option] client_cert = partial( Option, '--client-cert', dest='client_cert', type='str', default=None, metavar='path', help="Path to SSL client certificate, a single file containing the " "private key and the certificate in PEM format.", ) # type: Callable[..., Option] index_url = partial( Option, '-i', '--index-url', '--pypi-url', dest='index_url', metavar='URL', default=PyPI.simple_url, help="Base URL of the Python Package Index (default %default). " "This should point to a repository compliant with PEP 503 " "(the simple repository API) or a local directory laid out " "in the same format.", ) # type: Callable[..., Option] def extra_index_url(): return Option( '--extra-index-url', dest='extra_index_urls', metavar='URL', action='append', default=[], help="Extra URLs of package indexes to use in addition to " "--index-url. Should follow the same rules as " "--index-url.", ) no_index = partial( Option, '--no-index', dest='no_index', action='store_true', default=False, help='Ignore package index (only looking at --find-links URLs instead).', ) # type: Callable[..., Option] def find_links(): # type: () -> Option return Option( '-f', '--find-links', dest='find_links', action='append', default=[], metavar='url', help="If a url or path to an html file, then parse for links to " "archives. If a local path or file:// url that's a directory, " "then look for archives in the directory listing.", ) def trusted_host(): # type: () -> Option return Option( "--trusted-host", dest="trusted_hosts", action="append", metavar="HOSTNAME", default=[], help="Mark this host or host:port pair as trusted, even though it " "does not have valid or any HTTPS.", ) def constraints(): # type: () -> Option return Option( '-c', '--constraint', dest='constraints', action='append', default=[], metavar='file', help='Constrain versions using the given constraints file. ' 'This option can be used multiple times.' ) def requirements(): # type: () -> Option return Option( '-r', '--requirement', dest='requirements', action='append', default=[], metavar='file', help='Install from the given requirements file. ' 'This option can be used multiple times.' ) def editable(): # type: () -> Option return Option( '-e', '--editable', dest='editables', action='append', default=[], metavar='path/url', help=('Install a project in editable mode (i.e. setuptools ' '"develop mode") from a local project path or a VCS url.'), ) src = partial( Option, '--src', '--source', '--source-dir', '--source-directory', dest='src_dir', metavar='dir', default=get_src_prefix(), help='Directory to check out editable projects into. ' 'The default in a virtualenv is "/src". ' 'The default for global installs is "/src".' ) # type: Callable[..., Option] def _get_format_control(values, option): # type: (Values, Option) -> Any """Get a format_control object.""" return getattr(values, option.dest) def _handle_no_binary(option, opt_str, value, parser): # type: (Option, str, str, OptionParser) -> None existing = _get_format_control(parser.values, option) FormatControl.handle_mutual_excludes( value, existing.no_binary, existing.only_binary, ) def _handle_only_binary(option, opt_str, value, parser): # type: (Option, str, str, OptionParser) -> None existing = _get_format_control(parser.values, option) FormatControl.handle_mutual_excludes( value, existing.only_binary, existing.no_binary, ) def no_binary(): # type: () -> Option format_control = FormatControl(set(), set()) return Option( "--no-binary", dest="format_control", action="callback", callback=_handle_no_binary, type="str", default=format_control, help="Do not use binary packages. Can be supplied multiple times, and " "each time adds to the existing value. Accepts either :all: to " "disable all binary packages, :none: to empty the set, or one or " "more package names with commas between them (no colons). Note " "that some packages are tricky to compile and may fail to " "install when this option is used on them.", ) def only_binary(): # type: () -> Option format_control = FormatControl(set(), set()) return Option( "--only-binary", dest="format_control", action="callback", callback=_handle_only_binary, type="str", default=format_control, help="Do not use source packages. Can be supplied multiple times, and " "each time adds to the existing value. Accepts either :all: to " "disable all source packages, :none: to empty the set, or one or " "more package names with commas between them. Packages without " "binary distributions will fail to install when this option is " "used on them.", ) platform = partial( Option, '--platform', dest='platform', metavar='platform', default=None, help=("Only use wheels compatible with . " "Defaults to the platform of the running system."), ) # type: Callable[..., Option] # This was made a separate function for unit-testing purposes. def _convert_python_version(value): # type: (str) -> Tuple[Tuple[int, ...], Optional[str]] """ Convert a version string like "3", "37", or "3.7.3" into a tuple of ints. :return: A 2-tuple (version_info, error_msg), where `error_msg` is non-None if and only if there was a parsing error. """ if not value: # The empty string is the same as not providing a value. return (None, None) parts = value.split('.') if len(parts) > 3: return ((), 'at most three version parts are allowed') if len(parts) == 1: # Then we are in the case of "3" or "37". value = parts[0] if len(value) > 1: parts = [value[0], value[1:]] try: version_info = tuple(int(part) for part in parts) except ValueError: return ((), 'each version part must be an integer') return (version_info, None) def _handle_python_version(option, opt_str, value, parser): # type: (Option, str, str, OptionParser) -> None """ Handle a provided --python-version value. """ version_info, error_msg = _convert_python_version(value) if error_msg is not None: msg = ( 'invalid --python-version value: {!r}: {}'.format( value, error_msg, ) ) raise_option_error(parser, option=option, msg=msg) parser.values.python_version = version_info python_version = partial( Option, '--python-version', dest='python_version', metavar='python_version', action='callback', callback=_handle_python_version, type='str', default=None, help=dedent("""\ The Python interpreter version to use for wheel and "Requires-Python" compatibility checks. Defaults to a version derived from the running interpreter. The version can be specified using up to three dot-separated integers (e.g. "3" for 3.0.0, "3.7" for 3.7.0, or "3.7.3"). A major-minor version can also be given as a string without dots (e.g. "37" for 3.7.0). """), ) # type: Callable[..., Option] implementation = partial( Option, '--implementation', dest='implementation', metavar='implementation', default=None, help=("Only use wheels compatible with Python " "implementation , e.g. 'pp', 'jy', 'cp', " " or 'ip'. If not specified, then the current " "interpreter implementation is used. Use 'py' to force " "implementation-agnostic wheels."), ) # type: Callable[..., Option] abi = partial( Option, '--abi', dest='abi', metavar='abi', default=None, help=("Only use wheels compatible with Python " "abi , e.g. 'pypy_41'. If not specified, then the " "current interpreter abi tag is used. Generally " "you will need to specify --implementation, " "--platform, and --python-version when using " "this option."), ) # type: Callable[..., Option] def add_target_python_options(cmd_opts): # type: (OptionGroup) -> None cmd_opts.add_option(platform()) cmd_opts.add_option(python_version()) cmd_opts.add_option(implementation()) cmd_opts.add_option(abi()) def make_target_python(options): # type: (Values) -> TargetPython target_python = TargetPython( platform=options.platform, py_version_info=options.python_version, abi=options.abi, implementation=options.implementation, ) return target_python def prefer_binary(): # type: () -> Option return Option( "--prefer-binary", dest="prefer_binary", action="store_true", default=False, help="Prefer older binary packages over newer source packages." ) cache_dir = partial( Option, "--cache-dir", dest="cache_dir", default=USER_CACHE_DIR, metavar="dir", help="Store the cache data in ." ) # type: Callable[..., Option] def _handle_no_cache_dir(option, opt, value, parser): # type: (Option, str, str, OptionParser) -> None """ Process a value provided for the --no-cache-dir option. This is an optparse.Option callback for the --no-cache-dir option. """ # The value argument will be None if --no-cache-dir is passed via the # command-line, since the option doesn't accept arguments. However, # the value can be non-None if the option is triggered e.g. by an # environment variable, like PIP_NO_CACHE_DIR=true. if value is not None: # Then parse the string value to get argument error-checking. try: strtobool(value) except ValueError as exc: raise_option_error(parser, option=option, msg=str(exc)) # Originally, setting PIP_NO_CACHE_DIR to a value that strtobool() # converted to 0 (like "false" or "no") caused cache_dir to be disabled # rather than enabled (logic would say the latter). Thus, we disable # the cache directory not just on values that parse to True, but (for # backwards compatibility reasons) also on values that parse to False. # In other words, always set it to False if the option is provided in # some (valid) form. parser.values.cache_dir = False no_cache = partial( Option, "--no-cache-dir", dest="cache_dir", action="callback", callback=_handle_no_cache_dir, help="Disable the cache.", ) # type: Callable[..., Option] no_deps = partial( Option, '--no-deps', '--no-dependencies', dest='ignore_dependencies', action='store_true', default=False, help="Don't install package dependencies.", ) # type: Callable[..., Option] build_dir = partial( Option, '-b', '--build', '--build-dir', '--build-directory', dest='build_dir', metavar='dir', help='Directory to unpack packages into and build in. Note that ' 'an initial build still takes place in a temporary directory. ' 'The location of temporary directories can be controlled by setting ' 'the TMPDIR environment variable (TEMP on Windows) appropriately. ' 'When passed, build directories are not cleaned in case of failures.' ) # type: Callable[..., Option] ignore_requires_python = partial( Option, '--ignore-requires-python', dest='ignore_requires_python', action='store_true', help='Ignore the Requires-Python information.' ) # type: Callable[..., Option] no_build_isolation = partial( Option, '--no-build-isolation', dest='build_isolation', action='store_false', default=True, help='Disable isolation when building a modern source distribution. ' 'Build dependencies specified by PEP 518 must be already installed ' 'if this option is used.' ) # type: Callable[..., Option] def _handle_no_use_pep517(option, opt, value, parser): # type: (Option, str, str, OptionParser) -> None """ Process a value provided for the --no-use-pep517 option. This is an optparse.Option callback for the no_use_pep517 option. """ # Since --no-use-pep517 doesn't accept arguments, the value argument # will be None if --no-use-pep517 is passed via the command-line. # However, the value can be non-None if the option is triggered e.g. # by an environment variable, for example "PIP_NO_USE_PEP517=true". if value is not None: msg = """A value was passed for --no-use-pep517, probably using either the PIP_NO_USE_PEP517 environment variable or the "no-use-pep517" config file option. Use an appropriate value of the PIP_USE_PEP517 environment variable or the "use-pep517" config file option instead. """ raise_option_error(parser, option=option, msg=msg) # Otherwise, --no-use-pep517 was passed via the command-line. parser.values.use_pep517 = False use_pep517 = partial( Option, '--use-pep517', dest='use_pep517', action='store_true', default=None, help='Use PEP 517 for building source distributions ' '(use --no-use-pep517 to force legacy behaviour).' ) # type: Any no_use_pep517 = partial( Option, '--no-use-pep517', dest='use_pep517', action='callback', callback=_handle_no_use_pep517, default=None, help=SUPPRESS_HELP ) # type: Any install_options = partial( Option, '--install-option', dest='install_options', action='append', metavar='options', help="Extra arguments to be supplied to the setup.py install " "command (use like --install-option=\"--install-scripts=/usr/local/" "bin\"). Use multiple --install-option options to pass multiple " "options to setup.py install. If you are using an option with a " "directory path, be sure to use absolute path.", ) # type: Callable[..., Option] global_options = partial( Option, '--global-option', dest='global_options', action='append', metavar='options', help="Extra global options to be supplied to the setup.py " "call before the install command.", ) # type: Callable[..., Option] no_clean = partial( Option, '--no-clean', action='store_true', default=False, help="Don't clean up build directories." ) # type: Callable[..., Option] pre = partial( Option, '--pre', action='store_true', default=False, help="Include pre-release and development versions. By default, " "pip only finds stable versions.", ) # type: Callable[..., Option] disable_pip_version_check = partial( Option, "--disable-pip-version-check", dest="disable_pip_version_check", action="store_true", default=False, help="Don't periodically check PyPI to determine whether a new version " "of pip is available for download. Implied with --no-index.", ) # type: Callable[..., Option] # Deprecated, Remove later always_unzip = partial( Option, '-Z', '--always-unzip', dest='always_unzip', action='store_true', help=SUPPRESS_HELP, ) # type: Callable[..., Option] def _handle_merge_hash(option, opt_str, value, parser): # type: (Option, str, str, OptionParser) -> None """Given a value spelled "algo:digest", append the digest to a list pointed to in a dict by the algo name.""" if not parser.values.hashes: parser.values.hashes = {} try: algo, digest = value.split(':', 1) except ValueError: parser.error('Arguments to %s must be a hash name ' 'followed by a value, like --hash=sha256:abcde...' % opt_str) if algo not in STRONG_HASHES: parser.error('Allowed hash algorithms for %s are %s.' % (opt_str, ', '.join(STRONG_HASHES))) parser.values.hashes.setdefault(algo, []).append(digest) hash = partial( Option, '--hash', # Hash values eventually end up in InstallRequirement.hashes due to # __dict__ copying in process_line(). dest='hashes', action='callback', callback=_handle_merge_hash, type='string', help="Verify that the package's archive matches this " 'hash before installing. Example: --hash=sha256:abcdef...', ) # type: Callable[..., Option] require_hashes = partial( Option, '--require-hashes', dest='require_hashes', action='store_true', default=False, help='Require a hash to check each requirement against, for ' 'repeatable installs. This option is implied when any package in a ' 'requirements file has a --hash option.', ) # type: Callable[..., Option] list_path = partial( Option, '--path', dest='path', action='append', help='Restrict to the specified installation path for listing ' 'packages (can be used multiple times).' ) # type: Callable[..., Option] def check_list_path_option(options): # type: (Values) -> None if options.path and (options.user or options.local): raise CommandError( "Cannot combine '--path' with '--user' or '--local'" ) ########## # groups # ########## general_group = { 'name': 'General Options', 'options': [ help_, isolated_mode, require_virtualenv, verbose, version, quiet, log, no_input, proxy, retries, timeout, skip_requirements_regex, exists_action, trusted_host, cert, client_cert, cache_dir, no_cache, disable_pip_version_check, no_color, ] } # type: Dict[str, Any] index_group = { 'name': 'Package Index Options', 'options': [ index_url, extra_index_url, no_index, find_links, ] } # type: Dict[str, Any] site-packages/pip/_internal/cli/base_command.py000064400000014550151464452710015576 0ustar00"""Base Command class, and related routines""" from __future__ import absolute_import, print_function import logging import logging.config import optparse import os import platform import sys import traceback from pip._internal.cli import cmdoptions from pip._internal.cli.command_context import CommandContextMixIn from pip._internal.cli.parser import ( ConfigOptionParser, UpdatingDefaultsHelpFormatter, ) from pip._internal.cli.status_codes import ( ERROR, PREVIOUS_BUILD_DIR_ERROR, SUCCESS, UNKNOWN_ERROR, VIRTUALENV_NOT_FOUND, ) from pip._internal.exceptions import ( BadCommand, CommandError, InstallationError, PreviousBuildDirError, UninstallationError, ) from pip._internal.utils.deprecation import deprecated from pip._internal.utils.logging import BrokenStdoutLoggingError, setup_logging from pip._internal.utils.misc import get_prog from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.virtualenv import running_under_virtualenv if MYPY_CHECK_RUNNING: from typing import List, Tuple, Any from optparse import Values __all__ = ['Command'] logger = logging.getLogger(__name__) class Command(CommandContextMixIn): usage = None # type: str ignore_require_venv = False # type: bool def __init__(self, name, summary, isolated=False): # type: (str, str, bool) -> None super(Command, self).__init__() parser_kw = { 'usage': self.usage, 'prog': '%s %s' % (get_prog(), name), 'formatter': UpdatingDefaultsHelpFormatter(), 'add_help_option': False, 'name': name, 'description': self.__doc__, 'isolated': isolated, } self.name = name self.summary = summary self.parser = ConfigOptionParser(**parser_kw) # Commands should add options to this option group optgroup_name = '%s Options' % self.name.capitalize() self.cmd_opts = optparse.OptionGroup(self.parser, optgroup_name) # Add the general options gen_opts = cmdoptions.make_option_group( cmdoptions.general_group, self.parser, ) self.parser.add_option_group(gen_opts) def handle_pip_version_check(self, options): # type: (Values) -> None """ This is a no-op so that commands by default do not do the pip version check. """ # Make sure we do the pip version check if the index_group options # are present. assert not hasattr(options, 'no_index') def run(self, options, args): # type: (Values, List[Any]) -> Any raise NotImplementedError def parse_args(self, args): # type: (List[str]) -> Tuple # factored out for testability return self.parser.parse_args(args) def main(self, args): # type: (List[str]) -> int try: with self.main_context(): return self._main(args) finally: logging.shutdown() def _main(self, args): # type: (List[str]) -> int options, args = self.parse_args(args) # Set verbosity so that it can be used elsewhere. self.verbosity = options.verbose - options.quiet level_number = setup_logging( verbosity=self.verbosity, no_color=options.no_color, user_log_file=options.log, ) if sys.version_info[:2] == (2, 7): message = ( "A future version of pip will drop support for Python 2.7. " "More details about Python 2 support in pip, can be found at " "https://pip.pypa.io/en/latest/development/release-process/#python-2-support" # noqa ) if platform.python_implementation() == "CPython": message = ( "Python 2.7 will reach the end of its life on January " "1st, 2020. Please upgrade your Python as Python 2.7 " "won't be maintained after that date. " ) + message deprecated(message, replacement=None, gone_in=None) # TODO: Try to get these passing down from the command? # without resorting to os.environ to hold these. # This also affects isolated builds and it should. if options.no_input: os.environ['PIP_NO_INPUT'] = '1' if options.exists_action: os.environ['PIP_EXISTS_ACTION'] = ' '.join(options.exists_action) if options.require_venv and not self.ignore_require_venv: # If a venv is required check if it can really be found if not running_under_virtualenv(): logger.critical( 'Could not find an activated virtualenv (required).' ) sys.exit(VIRTUALENV_NOT_FOUND) try: status = self.run(options, args) # FIXME: all commands should return an exit status # and when it is done, isinstance is not needed anymore if isinstance(status, int): return status except PreviousBuildDirError as exc: logger.critical(str(exc)) logger.debug('Exception information:', exc_info=True) return PREVIOUS_BUILD_DIR_ERROR except (InstallationError, UninstallationError, BadCommand) as exc: logger.critical(str(exc)) logger.debug('Exception information:', exc_info=True) return ERROR except CommandError as exc: logger.critical('%s', exc) logger.debug('Exception information:', exc_info=True) return ERROR except BrokenStdoutLoggingError: # Bypass our logger and write any remaining messages to stderr # because stdout no longer works. print('ERROR: Pipe to stdout was broken', file=sys.stderr) if level_number <= logging.DEBUG: traceback.print_exc(file=sys.stderr) return ERROR except KeyboardInterrupt: logger.critical('Operation cancelled by user') logger.debug('Exception information:', exc_info=True) return ERROR except BaseException: logger.critical('Exception:', exc_info=True) return UNKNOWN_ERROR finally: self.handle_pip_version_check(options) return SUCCESS site-packages/pip/_internal/cli/__pycache__/autocompletion.cpython-38.opt-1.pyc000064400000011561151464452710023454 0ustar00U .e@s`dZddlZddlZddlZddlmZddlmZmZddl m Z ddZ dd Z d d Z dS) zBLogic that powers autocompletion installed by ``pip completion``. N)create_main_parser) commands_dictcreate_command)get_installed_distributionscsdtjkrdStjddd}ttjd}z||dWntk rZdYnXttg}zfdd|Dd }Wntk rd}YnXt}|r:|d krt d|d kö d  }|r>g} }t d dD].}|j |r|j |ddkr||j q|r>|D]}t|q$t dt|} | jjD]8} | jtjkrN| j| jD]} || | jfqlqNdd|d|dDfdd|D}fdd|D}t||| jj} | rt| }dd|D}|D]>} | d }| dr,| d dddkr,|d7}t|qndd|jD}||jdd|D} d r|D]$} | jtjkrt| j| j7qtnt|||} | rt| tdfddDt ddS)z?Entry Point for completion of main and subcommand options. ZPIP_AUTO_COMPLETENZ COMP_WORDSZ COMP_CWORDcsg|]}|kr|qSr).0w) subcommandsrD/usr/lib/python3.8/site-packages/pip/_internal/cli/autocompletion.py !sz autocomplete..rhelp)ZshowZ uninstall-T)Z local_onlycSsg|]}|ddqS)=r)splitr xrrr r Dscs g|]\}}|kr||fqSrr)r rv) prev_optsrr r Escs"g|]\}}|r||fqSr startswith)r krcurrentrr r Gs css|]}|dfVqdS)rNr)r optrrr Pszautocomplete..z--rcSsg|] }|jqSr) option_list)r irrr r Zscss|]}|D] }|Vq qdSNr)r itorrr r\s csg|]}|r|qSrrrrrr r gs )osenvironrint IndexErrorlistrrsysexitrlowerrkeyappendprintrparserZoption_list_allroptparse SUPPRESS_HELPZ _long_optsZ _short_optsnargsget_path_completion_typeauto_complete_pathsZ option_groupsrjoin)cwordscwordZoptionsZsubcommand_namer/Zshould_list_installedZ installedZlcZdistZ subcommandrZopt_strcompletion_typeZoptionZ opt_labeloptsr)rrr r autocompletes              r:cCs|dks||ddsdS|D]n}|jtjkr4q"t|dD]L}||ddd|krB|jrtdd|jdDrB|jSqBq"dS) aLGet the type of path completion (``file``, ``dir``, ``path`` or None) :param cwords: same as the environmental variable ``COMP_WORDS`` :param cword: same as the environmental variable ``COMP_CWORD`` :param opts: The available options to check :return: path completion type (``file``, ``dir``, ``path`` or None) rrN/rrcss|]}|dkVqdS))pathfiledirNrrrrr rzsz+get_path_completion_type..)rrr0r1strrmetavarany)r6r7r9rr"rrr r3ks  r3c#stj|\}tj|}t|tjs.dStjfddt|D}|D]`}tj||}tjtj||}|dkrtj |r|VqVtj |rVtj|dVqVdS)aoIf ``completion_type`` is ``file`` or ``path``, list all regular files and directories starting with ``current``; otherwise only list directories starting with ``current``. :param current: The word to be completed :param completion_type: path completion type(`file`, `path` or `dir`)i :return: A generator of regular files and/or directories Nc3s$|]}tj|r|VqdSr )r$r<normcaserrfilenamerr rsz&auto_complete_paths..r>r) r$r<rabspathaccessR_OKrBlistdirr5isfileisdir)rr8Z directoryZ current_pathZ file_listfrZ comp_filerrCr r4s    r4)__doc__r0r$r)Zpip._internal.cli.main_parserrZpip._internal.commandsrrZpip._internal.utils.miscrr:r3r4rrrr s  [site-packages/pip/_internal/cli/__pycache__/command_context.cpython-38.opt-1.pyc000064400000002053151464452710023570 0ustar00U .e@s,ddlmZddlmZGdddeZdS))contextmanager) ExitStackcs0eZdZfddZeddZddZZS)CommandContextMixIncs tt|d|_t|_dS)NF)superr__init___in_main_contextr _main_contextself __class__E/usr/lib/python3.8/site-packages/pip/_internal/cli/command_context.pyr szCommandContextMixIn.__init__c cs0d|_z|j dVW5QRXW5d|_XdS)NTF)rrr r r r main_contexts z CommandContextMixIn.main_contextcCs |j|S)N)r enter_context)r Zcontext_providerr r rrsz!CommandContextMixIn.enter_context)__name__ __module__ __qualname__rrrr __classcell__r r r rr s  rN) contextlibrZpip._vendor.contextlib2robjectrr r r rs  site-packages/pip/_internal/cli/__pycache__/command_context.cpython-38.pyc000064400000002135151464452710022632 0ustar00U .e@s,ddlmZddlmZGdddeZdS))contextmanager) ExitStackcs0eZdZfddZeddZddZZS)CommandContextMixIncs tt|d|_t|_dS)NF)superr__init___in_main_contextr _main_contextself __class__E/usr/lib/python3.8/site-packages/pip/_internal/cli/command_context.pyr szCommandContextMixIn.__init__c cs:|jr td|_z|j dVW5QRXW5d|_XdS)NTF)rAssertionErrorrr r r r main_contexts  z CommandContextMixIn.main_contextcCs|js t|j|S)N)rrr enter_context)r Zcontext_providerr r rrs z!CommandContextMixIn.enter_context)__name__ __module__ __qualname__rrrr __classcell__r r r rr s  rN) contextlibrZpip._vendor.contextlib2robjectrr r r rs  site-packages/pip/_internal/cli/__pycache__/main_parser.cpython-38.pyc000064400000004157151464452710021756 0ustar00U .e @sdZddlZddlZddlmZddlmZmZddlm Z m Z ddl m Z ddl mZmZddlmZer|dd lmZmZd d gZd d Zd d ZdS)z=A single place for constructing and exposing the main parser N) cmdoptions)ConfigOptionParserUpdatingDefaultsHelpFormatter) commands_dictget_similar_commands) CommandError)get_pip_versionget_prog)MYPY_CHECK_RUNNING)TupleListcreate_main_parser parse_commandcCstddtdtd}tf|}|t|_ttj|}| |d|_ dgddt D}d ||_|S) z6Creates and returns the main parser for pip's CLI z %prog [options]Fglobal)ZusageZadd_help_optionZ formatternameprogTcSsg|]\}}d||jfqS)z%-27s %s)Zsummary).0rZ command_inforA/usr/lib/python3.8/site-packages/pip/_internal/cli/main_parser.py 2sz&create_main_parser.. )rr rZdisable_interspersed_argsrversionrZmake_option_groupZ general_groupZadd_option_groupmainritemsjoin description)Z parser_kwparserZgen_optsrrrrr s"    cCst}||\}}|jr>tj|jtjtjt|rZ|ddkrjt |dkrj| t|d}|t krt |}d|g}|r| d|td||dd}||||fS)Nrhelpzunknown command "%s"zmaybe you meant "%s"z - )r parse_argsrsysstdoutwriteoslinesepexitlenZ print_helprrappendrrremove)argsrZgeneral_optionsZ args_elseZcmd_nameZguessmsgZcmd_argsrrrr;s&    )__doc__r$r!Zpip._internal.clirZpip._internal.cli.parserrrZpip._internal.commandsrrZpip._internal.exceptionsrZpip._internal.utils.miscrr Zpip._internal.utils.typingr typingr r __all__r rrrrrs   #site-packages/pip/_internal/cli/__pycache__/req_command.cpython-38.pyc000064400000016506151464452710021744 0ustar00U .ec,@sXdZddlZddlmZddlmZddlmZddlm Z ddl m Z ddl m Z dd lmZdd lmZdd lmZdd lmZmZmZdd lmZddlmZmZddlmZddlm Z e r"ddl!m"Z"ddl#m$Z$m%Z%m&Z&ddl'm(Z(ddl)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0GdddeZ1Gdddee1Z2Gddde2Z3dS)aContains the Command base classes that depend on PipSession. The classes in this module are in a separate module so the commands not needing download / PackageFinder capability don't unnecessarily import the PackageFinder machinery and all its vendored dependencies, etc. N)partial)Command)CommandContextMixIn) CommandError) PackageFinder)Resolver)SelectionPreferences) PipSession)RequirementPreparer)install_req_from_editableinstall_req_from_lineinstall_req_from_req_string)parse_requirements)make_link_collectorpip_self_version_check)normalize_path)MYPY_CHECK_RUNNING)Values)ListOptionalTuple) WheelCache) TargetPython)RequirementSet)RequirementTracker) TempDirectorycs>eZdZdZfddZeddZddZd d d ZZ S) SessionCommandMixinzE A class mixin for command classes needing _build_session(). cstt|d|_dS)N)superr__init___session)self __class__A/usr/lib/python3.8/site-packages/pip/_internal/cli/req_command.pyr2szSessionCommandMixin.__init__cCsLg}t|dds*t|dd}|r*||t|dd}|rD|||pJdS)z7Return a list of index urls from user-provided options.no_indexFZ index_urlNZextra_index_urls)getattrappendextend)clsoptions index_urlsZurlZurlsr#r#r$_get_index_urls6s     z#SessionCommandMixin._get_index_urlscCs"|jdkr||||_|jS)zGet a default-managed session.N)r enter_context_build_session)r r*r#r#r$get_default_sessionDs z'SessionCommandMixin.get_default_sessionNcCst|jrttj|jdnd|dk r*|n|j|j||d}|j rN|j |_ |j r\|j |_ |j sf|rz|dk rr|n|j |_ |j r|j |j d|_|j |j_|S)Nhttp)cacheretries trusted_hostsr+)r0Zhttps)r cache_dirrospathjoinr2r3r,ZcertZverifyZ client_certtimeoutproxyZproxiesZno_inputZauthZ prompting)r r*r2r8sessionr#r#r$r.Ks*   z"SessionCommandMixin._build_session)NN) __name__ __module__ __qualname____doc__r classmethodr,r/r. __classcell__r#r#r!r$r-s   rc@seZdZdZddZdS)IndexGroupCommandz Abstract base class for commands with the index_group options. This also corresponds to the commands that permit the pip version check. c CsTt|dst|js|jrdS|j|dtd|jd}|t||W5QRXdS)z Do the pip version check if not disabled. This overrides the default behavior of not doing the check. r%Nr)r2r8)hasattrAssertionErrorZdisable_pip_version_checkr%r.minr8r)r r*r:r#r#r$handle_pip_version_checkzs  z*IndexGroupCommand.handle_pip_version_checkN)r;r<r=r>rFr#r#r#r$rArsrAc @s:eZdZed ddZedddZd d Zdd d ZdS)RequirementCommandNc Cs.|j}|dk stt||j|||j|j|dS)zQ Create a RequirementPreparer instance for the given parameters. N)Z build_dirsrc_dir download_dirwheel_download_dir progress_barbuild_isolation req_tracker)r6rDr rHrKrL)Ztemp_build_dirr*rMrIrJZtemp_build_dir_pathr#r#r$make_requirement_preparers  z,RequirementCommand.make_requirement_preparerFTto-satisfy-onlyc Cs2tt|j|| d} t|||| ||j|||| | d S)zF Create a Resolver instance for the given parameters. )isolated wheel_cache use_pep517) preparerr:findermake_install_req use_user_siteignore_dependenciesignore_installedignore_requires_pythonforce_reinstallupgrade_strategypy_version_info)rr isolated_moderrW) rSr:rTr*rQrVrXrYrZr[rRr\rUr#r#r$ make_resolvers&z RequirementCommand.make_resolverc Cs,|jD].}t|d||||dD]}d|_||qq|D]*} t| d|j|j|d}d|_||q:|jD](} t| |j|j|d}d|_||ql|j D]0}t||||||jdD]}d|_||qq|j |_ |s(|js(|j s(d|j i} |j rt dt| d|j d n t d | dS) z? Marshal cmd line args into a requirement set. T)Z constraintrTr*r:rQN)rPrRrQ)rTr*r:rQrRnamez^You must give at least one requirement to %(name)s (maybe you meant "pip %(name)s %(links)s"?) )ZlinkszLYou must give at least one requirement to %(name)s (see "pip help %(name)s"))Z constraintsrZ is_directZadd_requirementr r]rRZ editablesr Z requirementsZrequire_hashesr_Z find_linksrdictr7) r Zrequirement_setargsr*rTr:rQfilenameZ req_to_addZreqZoptsr#r#r$populate_requirement_setsn        z+RequirementCommand.populate_requirement_setcCs4t||d}td|j|j|j|d}tj|||dS)z Create a package finder appropriate to this requirement command. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. )r*T)Z allow_yankedformat_controlZallow_all_prereleases prefer_binaryrY)link_collectorselection_prefs target_python)rrreZprerfrZcreate)r r*r:rirYrgrhr#r#r$_build_package_finders z(RequirementCommand._build_package_finder)NN)NFTFFrONN)NN)r;r<r= staticmethodrNr^rdrjr#r#r#r$rGs$  &GrG)4r>r5 functoolsrZpip._internal.cli.base_commandrZ!pip._internal.cli.command_contextrZpip._internal.exceptionsrZpip._internal.indexrZpip._internal.legacy_resolverZ$pip._internal.models.selection_prefsrZpip._internal.network.sessionr Z pip._internal.operations.preparer Zpip._internal.req.constructorsr r r Zpip._internal.req.req_filerZ!pip._internal.self_outdated_checkrrZpip._internal.utils.miscrZpip._internal.utils.typingrZoptparsertypingrrrZpip._internal.cacherZ"pip._internal.models.target_pythonrZpip._internal.req.req_setrZpip._internal.req.req_trackerrZpip._internal.utils.temp_dirrrrArGr#r#r#r$s4                   Esite-packages/pip/_internal/cli/__pycache__/parser.cpython-38.pyc000064400000021421151464452710020743 0ustar00U .e%@sdZddlmZddlZddlZddlZddlZddlmZddl m Z ddl m Z ddl mZmZddlmZeeZGd d d ejZGd d d eZGd ddejZGdddeZddZdS)zBase option parser setup)absolute_importN) strtobool) string_types) UNKNOWN_ERROR) ConfigurationConfigurationError)get_terminal_sizec@sReZdZdZddZddZddd Zd d Zd d ZddZ ddZ ddZ dS)PrettyHelpFormatterz4A prettier/less verbose help formatter for optparse.cOs:d|d<d|d<tdd|d<tjj|f||dS)NZmax_help_positionZindent_incrementrwidth)roptparseIndentedHelpFormatter__init__)selfargskwargsr, )_format_option_stringsroptionrrrformat_option_strings!sz)PrettyHelpFormatter.format_option_stringsrrcCs|g}|jr||jd|jr0||jdt|dkrH|d||rr|jp^|j}|||d |S)a Return a comma-separated list of option strings and metavars. :param option: tuple of (short opt, long opt), e.g: ('-f', '--format') :param mvarfmt: metavar format string - evaluated as mvarfmt % metavar :param optsep: separator rr ) Z _short_optsappendZ _long_optsleninsertZ takes_valuemetavardestlowerjoin)rrZmvarfmtZoptsepZoptsr rrrr$s  z*PrettyHelpFormatter._format_option_stringscCs|dkr dS|dS)NZOptionsrz: r)rZheadingrrrformat_heading;sz"PrettyHelpFormatter.format_headingcCsd|t|d}|S)zz Ensure there is only one newline between usage and the first heading if there is no description. z Usage: %s  ) indent_linestextwrapdedent)rZusagemsgrrr format_usage@sz PrettyHelpFormatter.format_usagecCsV|rNt|jdrd}nd}|d}|}|t|d}d||f}|SdSdS)NmainZCommandsZ Description r%z%s: %s r)hasattrparserlstriprstripr&r'r()r descriptionZlabelrrrformat_descriptionHs   z&PrettyHelpFormatter.format_descriptioncCs|r|SdSdS)Nrr)rZepilogrrr format_epilogZsz!PrettyHelpFormatter.format_epilogcs"fdd|dD}d|S)Ncsg|] }|qSrr).0lineindentrr bsz4PrettyHelpFormatter.indent_lines..r,)splitr#)rtextr7Z new_linesrr6rr&asz PrettyHelpFormatter.indent_linesN)rr) __name__ __module__ __qualname____doc__rrrr$r*r2r3r&rrrrr s r c@seZdZdZddZdS)UpdatingDefaultsHelpFormatterzCustom help formatter for use in ConfigOptionParser. This is updates the defaults before expanding them, allowing them to show up correctly in the help listing. cCs(|jdk r|j|jjtj||S)N)r._update_defaultsdefaultsrrexpand_defaultrrrrrBms z,UpdatingDefaultsHelpFormatter.expand_defaultN)r;r<r=r>rBrrrrr?fsr?c@s eZdZddZeddZdS)CustomOptionParsercOs(|j||}|j|j|||S)z*Insert an OptionGroup at a given position.)Zadd_option_group option_groupspopr)ridxrrgrouprrrinsert_option_groupus  z&CustomOptionParser.insert_option_groupcCs*|jdd}|jD]}||jq|S)ztd|q$|dd\}}||kr$||||fq$|D] }||D]\}}||fVqzqndS)Nglobalz:env:cSsi|] }|gqSrr)r4rOrrr szGConfigOptionParser._get_ordered_configuration_items..z7Ignoring configuration key '%s' as it's value is empty..r )rOrQitemsloggerdebugr9r)rZoverride_orderZ section_itemsZ section_keyrZZsectionrYrrr _get_ordered_configuration_itemss z3ConfigOptionParser._get_ordered_configuration_itemsc sHtj_t}D]\}ddkr>qjdkrz t|}Wn,t k rt j|} |YnXnjdkr| }fdd|D}nhjdkr| j}||}jpd}jpi}j||f||n|}||j<q|D]tj|<q&d_|S) zUpdates the given defaults with values from the config files and the environ. Does a little special handling for certain types of options (lists).z--N) store_true store_falsecountrcsg|]}|qSr)r\)r4vrYrrrrr8sz7ConfigOptionParser._update_defaults..callbackr)rValuesrAvaluessetrcZ get_optionactionr ValueErrorinvalid_config_error_messageerrorr9addr!get_opt_stringZ convert_valueZ callback_argsZcallback_kwargsrir\getattr)rrAZ late_evalrZZ error_msgopt_strrrrrhrr@s@         z#ConfigOptionParser._update_defaultsc Cs|jst|jSz|jWn2tk rR}z|tt |W5d}~XYnX| |j }| D]4}| |j}t|trl|}|||||j<qlt|S)zOverriding to make updating the defaults after instantiation of the option parser possible, _update_defaults() does the dirty work.N)Zprocess_default_valuesrrjrArQloadrrXrstrr@copyZ_get_all_optionsgetr! isinstancerrrrU)rerrrArdefaultrtrrrget_default_valuess "   z%ConfigOptionParser.get_default_valuescCs |tj|td|dS)Nz%s )Z print_usagerWstderrrXr)rr)rrrrps zConfigOptionParser.errorN) r;r<r=r>rr\rcr@r|rprrrrrNs 1rNcCs |dkrd||Sd||S)zQReturns a better error message when invalid configuration option is provided.)rdrezo{0} is not a valid value for {1} option, please specify a boolean value like yes/no, true/false or 1/0 instead.z[{0} is not a valid value for {1} option, please specify a numerical value like 1/0 instead.)format)rmrYrZrrrrosro)r>Z __future__rZloggingrrWr'Zdistutils.utilrZpip._vendor.sixrZpip._internal.cli.status_codesrZpip._internal.configurationrrZpip._internal.utils.compatrZ getLoggerr;rarr r?rSrCrNrorrrrs       O wsite-packages/pip/_internal/cli/__pycache__/__init__.cpython-38.pyc000064400000000354151464452710021210 0ustar00U .e@sdZdS)zGSubpackage containing all of pip's command line interface related code N)__doc__rr>/usr/lib/python3.8/site-packages/pip/_internal/cli/__init__.pysite-packages/pip/_internal/cli/__pycache__/main_parser.cpython-38.opt-1.pyc000064400000004157151464452710022715 0ustar00U .e @sdZddlZddlZddlmZddlmZmZddlm Z m Z ddl m Z ddl mZmZddlmZer|dd lmZmZd d gZd d Zd d ZdS)z=A single place for constructing and exposing the main parser N) cmdoptions)ConfigOptionParserUpdatingDefaultsHelpFormatter) commands_dictget_similar_commands) CommandError)get_pip_versionget_prog)MYPY_CHECK_RUNNING)TupleListcreate_main_parser parse_commandcCstddtdtd}tf|}|t|_ttj|}| |d|_ dgddt D}d ||_|S) z6Creates and returns the main parser for pip's CLI z %prog [options]Fglobal)ZusageZadd_help_optionZ formatternameprogTcSsg|]\}}d||jfqS)z%-27s %s)Zsummary).0rZ command_inforA/usr/lib/python3.8/site-packages/pip/_internal/cli/main_parser.py 2sz&create_main_parser.. )rr rZdisable_interspersed_argsrversionrZmake_option_groupZ general_groupZadd_option_groupmainritemsjoin description)Z parser_kwparserZgen_optsrrrrr s"    cCst}||\}}|jr>tj|jtjtjt|rZ|ddkrjt |dkrj| t|d}|t krt |}d|g}|r| d|td||dd}||||fS)Nrhelpzunknown command "%s"zmaybe you meant "%s"z - )r parse_argsrsysstdoutwriteoslinesepexitlenZ print_helprrappendrrremove)argsrZgeneral_optionsZ args_elseZcmd_nameZguessmsgZcmd_argsrrrr;s&    )__doc__r$r!Zpip._internal.clirZpip._internal.cli.parserrrZpip._internal.commandsrrZpip._internal.exceptionsrZpip._internal.utils.miscrr Zpip._internal.utils.typingr typingr r __all__r rrrrrs   #site-packages/pip/_internal/cli/__pycache__/status_codes.cpython-38.pyc000064400000000555151464452710022154 0ustar00U .e@s(ddlmZdZdZdZdZdZdZdS))absolute_importN)Z __future__rSUCCESSZERRORZ UNKNOWN_ERRORZVIRTUALENV_NOT_FOUNDZPREVIOUS_BUILD_DIR_ERRORZNO_MATCHES_FOUNDr r B/usr/lib/python3.8/site-packages/pip/_internal/cli/status_codes.pys site-packages/pip/_internal/cli/__pycache__/cmdoptions.cpython-38.opt-1.pyc000064400000045506151464452710022577 0ustar00U .eh@sdZddlmZddlZddlZddlZddlmZddlm Z ddl m Z m Z m Z ddlmZddlmZdd lmZmZdd lmZdd lmZdd lmZdd lmZddlmZddlmZerddl m!Z!m"Z"m#Z#m$Z$m%Z%ddl m&Z&m'Z'ddl(m)Z)e*e+Z,ddZ-ddZ.dddZ/dddZ0e e dddddd Z1e e d!d"d#dd$d%Z2e e d&d'd(d#de d%Z3e e d)d*d+d,dd-d%Z4e e d.d/d#dd0d%Z5e e d1d2d3d#d4d Z6e e d5d6d7d,dd8d%Z7e e d9d:d;e8e9d:e9d?d@Z;e e dAdBdCdDdEdFdGZe e dPdQdRdSdTdOZ?e e dUdVdWdXdYdZd[d\Z@e e d]d^dLdMe dOZAd_d`ZBe e dadbdLdEdcddZCe e dedfdLddEdgdhZDe e didjdkdldmejEdndoZFdpdqZGe e drdsd#ddtd%ZHdudvZIdwdxZJdydzZKd{d|ZLd}d~ZMe e ddddddeddo ZNddZOddZPddZQddZRddZSe e ddddddoZTddZUddZVe e ddddeVdLdedd ZWe e ddddddoZXe e ddddddoZYddZZddZ[ddZ\e e ddedddZ]ddZ^e e ddde^ddZ_e e dddd#ddd%Z`e e ddddddddGZae e ddd#dd Zbe e dddddd%ZcddĄZde e ddd#ddd%Zee e dddedde dɍZfe e ddddddύZge e ddddddύZhe e dd#dddՍZie e dd#dddՍZje e ddd#ddd%Zke e dddd#e d Zldd߄Zme e dddemdddZne e ddd#ddd%Zoe e ddEddd ZpddZqde1e2e3e4e6e7ee?e@eAeBeJeCeDe]e_eke5gdZrdeFeGeHeIgdZsdS)aC shared options and groups The principle here is to define options once, but *not* instantiate them globally. One reason being that options with action='append' can carry state between parses. pip parses general options twice internally, and shouldn't pass on state. To be consistent, all options will follow this design. )absolute_importN) strtobool)partial) SUPPRESS_HELPOption OptionGroup)dedent) CommandError)USER_CACHE_DIRget_src_prefix) FormatControl)PyPI) TargetPython) STRONG_HASHES)MYPY_CHECK_RUNNING) BAR_TYPES)AnyCallableDictOptionalTuple) OptionParserValues)ConfigOptionParsercCs.d||}td|}||dS)z Raise an option parsing error using parser.error(). Args: parser: an OptionParser instance. option: an Option instance. msg: the error text. z {} error: {} N)formattextwrapZfilljoinspliterror)parseroptionmsgr#@/usr/lib/python3.8/site-packages/pip/_internal/cli/cmdoptions.pyraise_option_error)s r%cCs,t||d}|dD]}||q|S)z Return an OptionGroup object group -- assumed to be dict with 'name' and 'options' keys parser -- an optparse Parser nameoptions)r add_option)groupr Z option_groupr!r#r#r$make_option_group7s r*csPdkr |fdd}dddg}tt||rL|j}|tjddd dS) zDisable wheels if per-setup.py call options are set. :param options: The OptionParser options to update. :param check_options: The options to check, if not supplied defaults to options. Ncs t|dSN)getattr)n check_optionsr#r$getnameOsz+check_install_build_global..getnameZ build_optionsglobal_optionsinstall_optionszeDisabling all use of wheels due to the use of --build-options / --global-options / --install-options.) stacklevel)anymapformat_controlZdisallow_binarieswarningswarn)r'r/r0namesZcontrolr#r.r$check_install_build_globalDs  r;FcCsbt|j|j|j|jg}ttdh}|j|ko6|j }|rH|rHt d|r^|r^|j s^t ddS)zFunction for determining if custom platform options are allowed. :param options: The OptionParser options. :param check_target: Whether or not to check if --target is being used. z:all:zWhen restricting platform and interpreter constraints using --python-version, --platform, --abi, or --implementation, either --no-deps must be set, or --only-binary=:all: must be set and --no-binary must not be set (or must be set to :none:).zQCan not use any platform or abi specific options unless installing via '--target'N) r5python_versionplatformabiimplementationr setr7ignore_dependenciesr Z target_dir)r'Z check_targetZdist_restriction_setZ binary_onlyZsdist_dependencies_allowedr#r#r$check_dist_restriction[s&  rBz-hz--helphelpz Show help.)destactionrCz --isolated isolated_mode store_truezSRun pip in an isolated mode, ignoring environment variables and user configuration.rDrEdefaultrCz--require-virtualenvz--require-venvZ require_venvz-vz --verboseverbosecountzDGive more output. Option is additive, and can be used up to 3 times.z --no-colorno_colorzSuppress colored outputz-Vz --versionversionzShow version and exit.z-qz--quietquietzGive less output. Option is additive, and can be used up to 3 times (corresponding to WARNING, ERROR, and CRITICAL logging levels).z--progress-bar progress_barchoiceZonz*Specify type of progress to be displayed [|z] (default: %default))rDtypechoicesrIrCz--logz --log-filez --local-loglogpathz Path to a verbose appending log.)rDmetavarrCz --no-inputno_inputz--proxyproxystrz/src". The default for global installs is "/src".cCs t||jS)zGet a format_control object.)r,rD)valuesr!r#r#r$_get_format_controlsrycCs"t|j|}t||j|jdSr+)ryrxr handle_mutual_excludes no_binary only_binaryr!opt_strvaluer Zexistingr#r#r$_handle_no_binarys  rcCs"t|j|}t||j|jdSr+)ryrxr rzr|r{r}r#r#r$_handle_only_binarys  rc Cs$ttt}tdddtd|ddS)Nz --no-binaryr7callbackrYa^Do not use binary packages. Can be supplied multiple times, and each time adds to the existing value. Accepts either :all: to disable all binary packages, :none: to empty the set, or one or more package names with commas between them (no colons). Note that some packages are tricky to compile and may fail to install when this option is used on them.rDrErrRrIrC)r r@rrr7r#r#r$r{sr{c Cs$ttt}tdddtd|ddS)Nz --only-binaryr7rrYaGDo not use source packages. Can be supplied multiple times, and each time adds to the existing value. Accepts either :all: to disable all source packages, :none: to empty the set, or one or more package names with commas between them. Packages without binary distributions will fail to install when this option is used on them.r)r r@rrrr#r#r$r|sr|z --platformr=z[Only use wheels compatible with . Defaults to the platform of the running system.cCs|sdS|d}t|dkr"dSt|dkrV|d}t|dkrV|d|ddg}ztdd |D}Wntk rYd SX|dfS) z Convert a version string like "3", "37", or "3.7.3" into a tuple of ints. :return: A 2-tuple (version_info, error_msg), where `error_msg` is non-None if and only if there was a parsing error. )NN.)r#z'at most three version parts are allowedrNcss|]}t|VqdSr+)r\).0partr#r#r$ sz*_convert_python_version..)r#z$each version part must be an integer)rlentuple ValueError)rparts version_infor#r#r$_convert_python_versions    rcCs:t|\}}|dk r.d||}t|||d||j_dS)z3 Handle a provided --python-version value. Nz(invalid --python-version value: {!r}: {}r!r")rrr%rxr<)r!r~rr rZ error_msgr"r#r#r$_handle_python_version s rz--python-versionr<ra The Python interpreter version to use for wheel and "Requires-Python" compatibility checks. Defaults to a version derived from the running interpreter. The version can be specified using up to three dot-separated integers (e.g. "3" for 3.0.0, "3.7" for 3.7.0, or "3.7.3"). A major-minor version can also be given as a string without dots (e.g. "37" for 3.7.0). )rDrVrErrRrIrCz--implementationr?zOnly use wheels compatible with Python implementation , e.g. 'pp', 'jy', 'cp', or 'ip'. If not specified, then the current interpreter implementation is used. Use 'py' to force implementation-agnostic wheels.z--abir>zOnly use wheels compatible with Python abi , e.g. 'pypy_41'. If not specified, then the current interpreter abi tag is used. Generally you will need to specify --implementation, --platform, and --python-version when using this option.cCs4|t|t|t|tdSr+)r(r=r<r?r>)Zcmd_optsr#r#r$add_target_python_optionsIs   rcCst|j|j|j|jd}|S)N)r=Zpy_version_infor>r?)rr=r<r>r?)r'Z target_pythonr#r#r$make_target_pythonQsrcCstddddddS)Nz--prefer-binary prefer_binaryrGFz8Prefer older binary packages over newer source packages.rHrir#r#r#r$r]srz --cache-dir cache_dirzStore the cache data in .)rDrIrVrCc CsV|dk rJz t|Wn4tk rH}zt||t|dW5d}~XYnXd|j_dS)z Process a value provided for the --no-cache-dir option. This is an optparse.Option callback for the --no-cache-dir option. NrF)rrr%rYrxr)r!optrr excr#r#r$_handle_no_cache_dirrs  $ rz--no-cache-dirzDisable the cache.)rDrErrCz --no-depsz--no-dependenciesrAz#Don't install package dependencies.z-bz--buildz --build-dirz--build-directory build_dira>Directory to unpack packages into and build in. Note that an initial build still takes place in a temporary directory. The location of temporary directories can be controlled by setting the TMPDIR environment variable (TEMP on Windows) appropriately. When passed, build directories are not cleaned in case of failures.z--ignore-requires-pythonignore_requires_pythonz'Ignore the Requires-Python information.z--no-build-isolationZbuild_isolationZ store_falseTzDisable isolation when building a modern source distribution. Build dependencies specified by PEP 518 must be already installed if this option is used.cCs&|dk rd}t|||dd|j_dS)z Process a value provided for the --no-use-pep517 option. This is an optparse.Option callback for the no_use_pep517 option. Na0A value was passed for --no-use-pep517, probably using either the PIP_NO_USE_PEP517 environment variable or the "no-use-pep517" config file option. Use an appropriate value of the PIP_USE_PEP517 environment variable or the "use-pep517" config file option instead. rF)r%rx use_pep517)r!rrr r"r#r#r$_handle_no_use_pep517s rz --use-pep517rz^Use PEP 517 for building source distributions (use --no-use-pep517 to force legacy behaviour).z--no-use-pep517)rDrErrIrCz--install-optionr2rhr'a"Extra arguments to be supplied to the setup.py install command (use like --install-option="--install-scripts=/usr/local/bin"). Use multiple --install-option options to pass multiple options to setup.py install. If you are using an option with a directory path, be sure to use absolute path.)rDrErVrCz--global-optionr1zTExtra global options to be supplied to the setup.py call before the install command.z --no-cleanz!Don't clean up build directories.)rErIrCz--prezYInclude pre-release and development versions. By default, pip only finds stable versions.z--disable-pip-version-checkdisable_pip_version_checkz{Don't periodically check PyPI to determine whether a new version of pip is available for download. Implied with --no-index.z-Zz--always-unzip always_unzipcCs|jjsi|j_z|dd\}}Wn"tk rF|d|YnX|tkrh|d|dtf|jj|g|dS)zkGiven a value spelled "algo:digest", append the digest to a list pointed to in a dict by the algo name.:rzTArguments to %s must be a hash name followed by a value, like --hash=sha256:abcde...z&Allowed hash algorithms for %s are %s.z, N) rxhashesrrrrr setdefaultrh)r!r~rr ZalgoZdigestr#r#r$_handle_merge_hash)s  rz--hashrstringzgVerify that the package's archive matches this hash before installing. Example: --hash=sha256:abcdef...)rDrErrRrCz--require-hashesrequire_hasheszRequire a hash to check each requirement against, for repeatable installs. This option is implied when any package in a requirements file has a --hash option.z--pathz^Restrict to the specified installation path for listing packages (can be used multiple times).cCs|jr|js|jrtddS)Nz2Cannot combine '--path' with '--user' or '--local')rUuserZlocalr )r'r#r#r$check_list_path_option_srzGeneral Options)r&r'zPackage Index Options)N)F)t__doc__Z __future__rZloggingrr8Zdistutils.utilr functoolsrZoptparserrrrZpip._internal.exceptionsr Zpip._internal.locationsr r Z#pip._internal.models.format_controlr Zpip._internal.models.indexr Z"pip._internal.models.target_pythonrZpip._internal.utils.hashesrZpip._internal.utils.typingrZpip._internal.utils.uirtypingrrrrrrrZpip._internal.cli.parserrZ getLogger__name__Zloggerr%r*r;rBZhelp_rFZrequire_virtualenvrJrLrMrNlistkeysrrOrTrWrXr[r^rarbrjrkZ simple_urlrlrnrorprrrsrurvsrcryrrr{r|r=rrr<r?r>rrrrrZno_cacheZno_depsrrZno_build_isolationrrZ no_use_pep517r2r1Zno_cleanZprerrrhashrZ list_pathrZ general_groupZ index_groupr#r#r#r$sB               ,                                         site-packages/pip/_internal/cli/__pycache__/autocompletion.cpython-38.pyc000064400000011561151464452710022515 0ustar00U .e@s`dZddlZddlZddlZddlmZddlmZmZddl m Z ddZ dd Z d d Z dS) zBLogic that powers autocompletion installed by ``pip completion``. N)create_main_parser) commands_dictcreate_command)get_installed_distributionscsdtjkrdStjddd}ttjd}z||dWntk rZdYnXttg}zfdd|Dd }Wntk rd}YnXt}|r:|d krt d|d kö d  }|r>g} }t d dD].}|j |r|j |ddkr||j q|r>|D]}t|q$t dt|} | jjD]8} | jtjkrN| j| jD]} || | jfqlqNdd|d|dDfdd|D}fdd|D}t||| jj} | rt| }dd|D}|D]>} | d }| dr,| d dddkr,|d7}t|qndd|jD}||jdd|D} d r|D]$} | jtjkrt| j| j7qtnt|||} | rt| tdfddDt ddS)z?Entry Point for completion of main and subcommand options. ZPIP_AUTO_COMPLETENZ COMP_WORDSZ COMP_CWORDcsg|]}|kr|qSr).0w) subcommandsrD/usr/lib/python3.8/site-packages/pip/_internal/cli/autocompletion.py !sz autocomplete..rhelp)ZshowZ uninstall-T)Z local_onlycSsg|]}|ddqS)=r)splitr xrrr r Dscs g|]\}}|kr||fqSrr)r rv) prev_optsrr r Escs"g|]\}}|r||fqSr startswith)r krcurrentrr r Gs css|]}|dfVqdS)rNr)r optrrr Pszautocomplete..z--rcSsg|] }|jqSr) option_list)r irrr r Zscss|]}|D] }|Vq qdSNr)r itorrr r\s csg|]}|r|qSrrrrrr r gs )osenvironrint IndexErrorlistrrsysexitrlowerrkeyappendprintrparserZoption_list_allroptparse SUPPRESS_HELPZ _long_optsZ _short_optsnargsget_path_completion_typeauto_complete_pathsZ option_groupsrjoin)cwordscwordZoptionsZsubcommand_namer/Zshould_list_installedZ installedZlcZdistZ subcommandrZopt_strcompletion_typeZoptionZ opt_labeloptsr)rrr r autocompletes              r:cCs|dks||ddsdS|D]n}|jtjkr4q"t|dD]L}||ddd|krB|jrtdd|jdDrB|jSqBq"dS) aLGet the type of path completion (``file``, ``dir``, ``path`` or None) :param cwords: same as the environmental variable ``COMP_WORDS`` :param cword: same as the environmental variable ``COMP_CWORD`` :param opts: The available options to check :return: path completion type (``file``, ``dir``, ``path`` or None) rrN/rrcss|]}|dkVqdS))pathfiledirNrrrrr rzsz+get_path_completion_type..)rrr0r1strrmetavarany)r6r7r9rr"rrr r3ks  r3c#stj|\}tj|}t|tjs.dStjfddt|D}|D]`}tj||}tjtj||}|dkrtj |r|VqVtj |rVtj|dVqVdS)aoIf ``completion_type`` is ``file`` or ``path``, list all regular files and directories starting with ``current``; otherwise only list directories starting with ``current``. :param current: The word to be completed :param completion_type: path completion type(`file`, `path` or `dir`)i :return: A generator of regular files and/or directories Nc3s$|]}tj|r|VqdSr )r$r<normcaserrfilenamerr rsz&auto_complete_paths..r>r) r$r<rabspathaccessR_OKrBlistdirr5isfileisdir)rr8Z directoryZ current_pathZ file_listfrZ comp_filerrCr r4s    r4)__doc__r0r$r)Zpip._internal.cli.main_parserrZpip._internal.commandsrrZpip._internal.utils.miscrr:r3r4rrrr s  [site-packages/pip/_internal/cli/__pycache__/parser.cpython-38.opt-1.pyc000064400000021365151464452710021711 0ustar00U .e%@sdZddlmZddlZddlZddlZddlZddlmZddl m Z ddl m Z ddl mZmZddlmZeeZGd d d ejZGd d d eZGd ddejZGdddeZddZdS)zBase option parser setup)absolute_importN) strtobool) string_types) UNKNOWN_ERROR) ConfigurationConfigurationError)get_terminal_sizec@sReZdZdZddZddZddd Zd d Zd d ZddZ ddZ ddZ dS)PrettyHelpFormatterz4A prettier/less verbose help formatter for optparse.cOs:d|d<d|d<tdd|d<tjj|f||dS)NZmax_help_positionZindent_incrementrwidth)roptparseIndentedHelpFormatter__init__)selfargskwargsr, )_format_option_stringsroptionrrrformat_option_strings!sz)PrettyHelpFormatter.format_option_stringsrrcCs|g}|jr||jd|jr0||jdt|dkrH|d||rr|jp^|j}|||d |S)a Return a comma-separated list of option strings and metavars. :param option: tuple of (short opt, long opt), e.g: ('-f', '--format') :param mvarfmt: metavar format string - evaluated as mvarfmt % metavar :param optsep: separator rr ) Z _short_optsappendZ _long_optsleninsertZ takes_valuemetavardestlowerjoin)rrZmvarfmtZoptsepZoptsr rrrr$s  z*PrettyHelpFormatter._format_option_stringscCs|dkr dS|dS)NZOptionsrz: r)rZheadingrrrformat_heading;sz"PrettyHelpFormatter.format_headingcCsd|t|d}|S)zz Ensure there is only one newline between usage and the first heading if there is no description. z Usage: %s  ) indent_linestextwrapdedent)rZusagemsgrrr format_usage@sz PrettyHelpFormatter.format_usagecCsV|rNt|jdrd}nd}|d}|}|t|d}d||f}|SdSdS)NmainZCommandsZ Description r%z%s: %s r)hasattrparserlstriprstripr&r'r()r descriptionZlabelrrrformat_descriptionHs   z&PrettyHelpFormatter.format_descriptioncCs|r|SdSdS)Nrr)rZepilogrrr format_epilogZsz!PrettyHelpFormatter.format_epilogcs"fdd|dD}d|S)Ncsg|] }|qSrr).0lineindentrr bsz4PrettyHelpFormatter.indent_lines..r,)splitr#)rtextr7Z new_linesrr6rr&asz PrettyHelpFormatter.indent_linesN)rr) __name__ __module__ __qualname____doc__rrrr$r*r2r3r&rrrrr s r c@seZdZdZddZdS)UpdatingDefaultsHelpFormatterzCustom help formatter for use in ConfigOptionParser. This is updates the defaults before expanding them, allowing them to show up correctly in the help listing. cCs(|jdk r|j|jjtj||S)N)r._update_defaultsdefaultsrrexpand_defaultrrrrrBms z,UpdatingDefaultsHelpFormatter.expand_defaultN)r;r<r=r>rBrrrrr?fsr?c@s eZdZddZeddZdS)CustomOptionParsercOs(|j||}|j|j|||S)z*Insert an OptionGroup at a given position.)Zadd_option_group option_groupspopr)ridxrrgrouprrrinsert_option_groupus  z&CustomOptionParser.insert_option_groupcCs*|jdd}|jD]}||jq|S)ztd|q$|dd\}}||kr$||||fq$|D] }||D]\}}||fVqzqndS)Nglobalz:env:cSsi|] }|gqSrr)r4rOrrr szGConfigOptionParser._get_ordered_configuration_items..z7Ignoring configuration key '%s' as it's value is empty..r )rOrQitemsloggerdebugr9r)rZoverride_orderZ section_itemsZ section_keyrYZsectionrXrrr _get_ordered_configuration_itemss z3ConfigOptionParser._get_ordered_configuration_itemsc sHtj_t}D]\}ddkr>qjdkrz t|}Wn,t k rt j|} |YnXnjdkr| }fdd|D}nhjdkr| j}||}jpd}jpi}j||f||n|}||j<q|D]tj|<q&d_|S) zUpdates the given defaults with values from the config files and the environ. Does a little special handling for certain types of options (lists).z--N) store_true store_falsecountrcsg|]}|qSr)r[)r4vrXrrrrr8sz7ConfigOptionParser._update_defaults..callbackr)rValuesrAvaluessetrbZ get_optionactionr ValueErrorinvalid_config_error_messageerrorr9addr!get_opt_stringZ convert_valueZ callback_argsZcallback_kwargsrhr[getattr)rrAZ late_evalrYZ error_msgopt_strrrrrgrr@s@         z#ConfigOptionParser._update_defaultsc Cs|jst|jSz|jWn2tk rR}z|tt |W5d}~XYnX| |j }| D]4}| |j}t|trl|}|||||j<qlt|S)zOverriding to make updating the defaults after instantiation of the option parser possible, _update_defaults() does the dirty work.N)Zprocess_default_valuesrrirArQloadrrWrstrr@copyZ_get_all_optionsgetr! isinstancerrqrT)rerrrArdefaultrsrrrget_default_valuess "   z%ConfigOptionParser.get_default_valuescCs |tj|td|dS)Nz%s )Z print_usagerVstderrrWr)rr)rrrros zConfigOptionParser.errorN) r;r<r=r>rr[rbr@r{rorrrrrNs 1rNcCs |dkrd||Sd||S)zQReturns a better error message when invalid configuration option is provided.)rcrdzo{0} is not a valid value for {1} option, please specify a boolean value like yes/no, true/false or 1/0 instead.z[{0} is not a valid value for {1} option, please specify a numerical value like 1/0 instead.)format)rlrXrYrrrrnsrn)r>Z __future__rZloggingrrVr'Zdistutils.utilrZpip._vendor.sixrZpip._internal.cli.status_codesrZpip._internal.configurationrrZpip._internal.utils.compatrZ getLoggerr;r`rr r?rRrCrNrnrrrrs       O wsite-packages/pip/_internal/cli/__pycache__/req_command.cpython-38.opt-1.pyc000064400000016404151464452710022700 0ustar00U .ec,@sXdZddlZddlmZddlmZddlmZddlm Z ddl m Z ddl m Z dd lmZdd lmZdd lmZdd lmZmZmZdd lmZddlmZmZddlmZddlm Z e r"ddl!m"Z"ddl#m$Z$m%Z%m&Z&ddl'm(Z(ddl)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0GdddeZ1Gdddee1Z2Gddde2Z3dS)aContains the Command base classes that depend on PipSession. The classes in this module are in a separate module so the commands not needing download / PackageFinder capability don't unnecessarily import the PackageFinder machinery and all its vendored dependencies, etc. N)partial)Command)CommandContextMixIn) CommandError) PackageFinder)Resolver)SelectionPreferences) PipSession)RequirementPreparer)install_req_from_editableinstall_req_from_lineinstall_req_from_req_string)parse_requirements)make_link_collectorpip_self_version_check)normalize_path)MYPY_CHECK_RUNNING)Values)ListOptionalTuple) WheelCache) TargetPython)RequirementSet)RequirementTracker) TempDirectorycs>eZdZdZfddZeddZddZd d d ZZ S) SessionCommandMixinzE A class mixin for command classes needing _build_session(). cstt|d|_dS)N)superr__init___session)self __class__A/usr/lib/python3.8/site-packages/pip/_internal/cli/req_command.pyr2szSessionCommandMixin.__init__cCsLg}t|dds*t|dd}|r*||t|dd}|rD|||pJdS)z7Return a list of index urls from user-provided options.no_indexFZ index_urlNZextra_index_urls)getattrappendextend)clsoptions index_urlsZurlZurlsr#r#r$_get_index_urls6s     z#SessionCommandMixin._get_index_urlscCs"|jdkr||||_|jS)zGet a default-managed session.N)r enter_context_build_session)r r*r#r#r$get_default_sessionDs z'SessionCommandMixin.get_default_sessionNcCst|jrttj|jdnd|dk r*|n|j|j||d}|j rN|j |_ |j r\|j |_ |j sf|rz|dk rr|n|j |_ |j r|j |j d|_|j |j_|S)Nhttp)cacheretries trusted_hostsr+)r0Zhttps)r cache_dirrospathjoinr2r3r,ZcertZverifyZ client_certtimeoutproxyZproxiesZno_inputZauthZ prompting)r r*r2r8sessionr#r#r$r.Ks*   z"SessionCommandMixin._build_session)NN) __name__ __module__ __qualname____doc__r classmethodr,r/r. __classcell__r#r#r!r$r-s   rc@seZdZdZddZdS)IndexGroupCommandz Abstract base class for commands with the index_group options. This also corresponds to the commands that permit the pip version check. c CsF|js |jrdS|j|dtd|jd}|t||W5QRXdS)z Do the pip version check if not disabled. This overrides the default behavior of not doing the check. Nr)r2r8)Zdisable_pip_version_checkr%r.minr8r)r r*r:r#r#r$handle_pip_version_checkzs  z*IndexGroupCommand.handle_pip_version_checkN)r;r<r=r>rDr#r#r#r$rArsrAc @s:eZdZed ddZedddZd d Zdd d ZdS)RequirementCommandNc Cs"|j}t||j|||j|j|dS)zQ Create a RequirementPreparer instance for the given parameters. )Z build_dirsrc_dir download_dirwheel_download_dir progress_barbuild_isolation req_tracker)r6r rFrIrJ)Ztemp_build_dirr*rKrGrHZtemp_build_dir_pathr#r#r$make_requirement_preparers z,RequirementCommand.make_requirement_preparerFTto-satisfy-onlyc Cs2tt|j|| d} t|||| ||j|||| | d S)zF Create a Resolver instance for the given parameters. )isolated wheel_cache use_pep517) preparerr:findermake_install_req use_user_siteignore_dependenciesignore_installedignore_requires_pythonforce_reinstallupgrade_strategypy_version_info)rr isolated_moderrU) rQr:rRr*rOrTrVrWrXrYrPrZrSr#r#r$ make_resolvers&z RequirementCommand.make_resolverc Cs,|jD].}t|d||||dD]}d|_||qq|D]*} t| d|j|j|d}d|_||q:|jD](} t| |j|j|d}d|_||ql|j D]0}t||||||jdD]}d|_||qq|j |_ |s(|js(|j s(d|j i} |j rt dt| d|j d n t d | dS) z? Marshal cmd line args into a requirement set. T)Z constraintrRr*r:rON)rNrPrO)rRr*r:rOrPnamez^You must give at least one requirement to %(name)s (maybe you meant "pip %(name)s %(links)s"?) )ZlinkszLYou must give at least one requirement to %(name)s (see "pip help %(name)s"))Z constraintsrZ is_directZadd_requirementr r[rPZ editablesr Z requirementsZrequire_hashesr]Z find_linksrdictr7) r Zrequirement_setargsr*rRr:rOfilenameZ req_to_addZreqZoptsr#r#r$populate_requirement_setsn        z+RequirementCommand.populate_requirement_setcCs4t||d}td|j|j|j|d}tj|||dS)z Create a package finder appropriate to this requirement command. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. )r*T)Z allow_yankedformat_controlZallow_all_prereleases prefer_binaryrW)link_collectorselection_prefs target_python)rrrcZprerdrZcreate)r r*r:rgrWrerfr#r#r$_build_package_finders z(RequirementCommand._build_package_finder)NN)NFTFFrMNN)NN)r;r<r= staticmethodrLr\rbrhr#r#r#r$rEs$  &GrE)4r>r5 functoolsrZpip._internal.cli.base_commandrZ!pip._internal.cli.command_contextrZpip._internal.exceptionsrZpip._internal.indexrZpip._internal.legacy_resolverZ$pip._internal.models.selection_prefsrZpip._internal.network.sessionr Z pip._internal.operations.preparer Zpip._internal.req.constructorsr r r Zpip._internal.req.req_filerZ!pip._internal.self_outdated_checkrrZpip._internal.utils.miscrZpip._internal.utils.typingrZoptparsertypingrrrZpip._internal.cacherZ"pip._internal.models.target_pythonrZpip._internal.req.req_setrZpip._internal.req.req_trackerrZpip._internal.utils.temp_dirrrrArEr#r#r#r$s4                   Esite-packages/pip/_internal/cli/__pycache__/base_command.cpython-38.opt-1.pyc000064400000011603151464452710023017 0ustar00U .eh@s6dZddlmZmZddlZddlZddlZddlZddlZddl Z ddl Z ddl m Z ddl mZddlmZmZddlmZmZmZmZmZddlmZmZmZmZmZdd lmZdd l m!Z!m"Z"dd l#m$Z$dd l%m&Z&dd l'm(Z(e&rddl)m*Z*m+Z+m,Z,ddlm-Z-dgZ.e/e0Z1GdddeZ2dS)z(Base Command class, and related routines)absolute_importprint_functionN) cmdoptions)CommandContextMixIn)ConfigOptionParserUpdatingDefaultsHelpFormatter)ERRORPREVIOUS_BUILD_DIR_ERRORSUCCESS UNKNOWN_ERRORVIRTUALENV_NOT_FOUND) BadCommand CommandErrorInstallationErrorPreviousBuildDirErrorUninstallationError) deprecated)BrokenStdoutLoggingError setup_logging)get_prog)MYPY_CHECK_RUNNING)running_under_virtualenv)ListTupleAny)ValuesCommandcsNeZdZdZdZdfdd ZddZddZd d Zd d Z d dZ Z S)rNFcstt||jdt|ftd||j|d}||_||_t f||_ d|j }t |j ||_ttj|j }|j |dS)Nz%s %sF)usageprogZ formatterZadd_help_optionname descriptionisolatedz %s Options)superr__init__rrr__doc__rsummaryrparser capitalizeoptparseZ OptionGroupZcmd_optsrZmake_option_groupZ general_groupZadd_option_group)selfrr%r!Z parser_kwZ optgroup_nameZgen_opts __class__B/usr/lib/python3.8/site-packages/pip/_internal/cli/base_command.pyr#4s&   zCommand.__init__cCsdS)zf This is a no-op so that commands by default do not do the pip version check. Nr,)r)optionsr,r,r-handle_pip_version_checkPsz Command.handle_pip_version_checkcCstdSN)NotImplementedError)r)r.argsr,r,r-runZsz Command.runcCs |j|Sr0)r& parse_argsr)r2r,r,r-r4^szCommand.parse_argsc Cs>z.|||W5QRW SQRXW5tXdSr0)loggingZshutdownZ main_context_mainr5r,r,r-maincs $z Command.mainc Cs||\}}|j|j|_t|j|j|jd}tjdddkrhd}t dkrZd|}t |ddd|j rxdt jd <|jrd |jt jd <|jr|jststd ttzz(|||}t|tr|WWSWntk r2}z.tt|tjd ddtWYWRSd}~XYn>t t!t"fk r}z.tt|tjd ddt#WYWSd}~XYnt$k r}z*td|tjd ddt#WYWSd}~XYnt%k rt&dtj'd|t(j)krt*j+tj'dt#YWfSt,k rDtdtjd ddt#YW2St-k rntjdddt.YWSXW5||Xt/S)N) verbosityno_colorZ user_log_file)r;zA future version of pip will drop support for Python 2.7. More details about Python 2 support in pip, can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-supportZCPythonzPython 2.7 will reach the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 won't be maintained after that date. )Z replacementZgone_in1Z PIP_NO_INPUT ZPIP_EXISTS_ACTIONz2Could not find an activated virtualenv (required).zException information:T)exc_infoz%sz ERROR: Pipe to stdout was broken)filezOperation cancelled by userz Exception:)0r4verbosequietr9rr:logsys version_infoplatformZpython_implementationrZno_inputosenvironZ exists_actionjoinZ require_venvignore_require_venvrloggerZcriticalexitr r/r3 isinstanceintrstrdebugr rrr rrrprintstderrr6DEBUG traceback print_excKeyboardInterrupt BaseExceptionr r )r)r2r.Z level_numbermessageZstatusexcr,r,r-r7ksn             z Command._main)F) __name__ __module__ __qualname__rrJr#r/r3r4r8r7 __classcell__r,r,r*r-r0s )3r$Z __future__rrr6Zlogging.configr(rGrFrDrTZpip._internal.clirZ!pip._internal.cli.command_contextrZpip._internal.cli.parserrrZpip._internal.cli.status_codesrr r r r Zpip._internal.exceptionsr rrrrZpip._internal.utils.deprecationrZpip._internal.utils.loggingrrZpip._internal.utils.miscrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrtypingrrrr__all__Z getLoggerrZrKrr,r,r,r-s0        site-packages/pip/_internal/cli/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000354151464452710022147 0ustar00U .e@sdZdS)zGSubpackage containing all of pip's command line interface related code N)__doc__rr>/usr/lib/python3.8/site-packages/pip/_internal/cli/__init__.pysite-packages/pip/_internal/cli/__pycache__/cmdoptions.cpython-38.pyc000064400000045506151464452710021640 0ustar00U .eh@sdZddlmZddlZddlZddlZddlmZddlm Z ddl m Z m Z m Z ddlmZddlmZdd lmZmZdd lmZdd lmZdd lmZdd lmZddlmZddlmZerddl m!Z!m"Z"m#Z#m$Z$m%Z%ddl m&Z&m'Z'ddl(m)Z)e*e+Z,ddZ-ddZ.dddZ/dddZ0e e dddddd Z1e e d!d"d#dd$d%Z2e e d&d'd(d#de d%Z3e e d)d*d+d,dd-d%Z4e e d.d/d#dd0d%Z5e e d1d2d3d#d4d Z6e e d5d6d7d,dd8d%Z7e e d9d:d;e8e9d:e9d?d@Z;e e dAdBdCdDdEdFdGZe e dPdQdRdSdTdOZ?e e dUdVdWdXdYdZd[d\Z@e e d]d^dLdMe dOZAd_d`ZBe e dadbdLdEdcddZCe e dedfdLddEdgdhZDe e didjdkdldmejEdndoZFdpdqZGe e drdsd#ddtd%ZHdudvZIdwdxZJdydzZKd{d|ZLd}d~ZMe e ddddddeddo ZNddZOddZPddZQddZRddZSe e ddddddoZTddZUddZVe e ddddeVdLdedd ZWe e ddddddoZXe e ddddddoZYddZZddZ[ddZ\e e ddedddZ]ddZ^e e ddde^ddZ_e e dddd#ddd%Z`e e ddddddddGZae e ddd#dd Zbe e dddddd%ZcddĄZde e ddd#ddd%Zee e dddedde dɍZfe e ddddddύZge e ddddddύZhe e dd#dddՍZie e dd#dddՍZje e ddd#ddd%Zke e dddd#e d Zldd߄Zme e dddemdddZne e ddd#ddd%Zoe e ddEddd ZpddZqde1e2e3e4e6e7ee?e@eAeBeJeCeDe]e_eke5gdZrdeFeGeHeIgdZsdS)aC shared options and groups The principle here is to define options once, but *not* instantiate them globally. One reason being that options with action='append' can carry state between parses. pip parses general options twice internally, and shouldn't pass on state. To be consistent, all options will follow this design. )absolute_importN) strtobool)partial) SUPPRESS_HELPOption OptionGroup)dedent) CommandError)USER_CACHE_DIRget_src_prefix) FormatControl)PyPI) TargetPython) STRONG_HASHES)MYPY_CHECK_RUNNING) BAR_TYPES)AnyCallableDictOptionalTuple) OptionParserValues)ConfigOptionParsercCs.d||}td|}||dS)z Raise an option parsing error using parser.error(). Args: parser: an OptionParser instance. option: an Option instance. msg: the error text. z {} error: {} N)formattextwrapZfilljoinspliterror)parseroptionmsgr#@/usr/lib/python3.8/site-packages/pip/_internal/cli/cmdoptions.pyraise_option_error)s r%cCs,t||d}|dD]}||q|S)z Return an OptionGroup object group -- assumed to be dict with 'name' and 'options' keys parser -- an optparse Parser nameoptions)r add_option)groupr Z option_groupr!r#r#r$make_option_group7s r*csPdkr |fdd}dddg}tt||rL|j}|tjddd dS) zDisable wheels if per-setup.py call options are set. :param options: The OptionParser options to update. :param check_options: The options to check, if not supplied defaults to options. Ncs t|dSN)getattr)n check_optionsr#r$getnameOsz+check_install_build_global..getnameZ build_optionsglobal_optionsinstall_optionszeDisabling all use of wheels due to the use of --build-options / --global-options / --install-options.) stacklevel)anymapformat_controlZdisallow_binarieswarningswarn)r'r/r0namesZcontrolr#r.r$check_install_build_globalDs  r;FcCsbt|j|j|j|jg}ttdh}|j|ko6|j }|rH|rHt d|r^|r^|j s^t ddS)zFunction for determining if custom platform options are allowed. :param options: The OptionParser options. :param check_target: Whether or not to check if --target is being used. z:all:zWhen restricting platform and interpreter constraints using --python-version, --platform, --abi, or --implementation, either --no-deps must be set, or --only-binary=:all: must be set and --no-binary must not be set (or must be set to :none:).zQCan not use any platform or abi specific options unless installing via '--target'N) r5python_versionplatformabiimplementationr setr7ignore_dependenciesr Z target_dir)r'Z check_targetZdist_restriction_setZ binary_onlyZsdist_dependencies_allowedr#r#r$check_dist_restriction[s&  rBz-hz--helphelpz Show help.)destactionrCz --isolated isolated_mode store_truezSRun pip in an isolated mode, ignoring environment variables and user configuration.rDrEdefaultrCz--require-virtualenvz--require-venvZ require_venvz-vz --verboseverbosecountzDGive more output. Option is additive, and can be used up to 3 times.z --no-colorno_colorzSuppress colored outputz-Vz --versionversionzShow version and exit.z-qz--quietquietzGive less output. Option is additive, and can be used up to 3 times (corresponding to WARNING, ERROR, and CRITICAL logging levels).z--progress-bar progress_barchoiceZonz*Specify type of progress to be displayed [|z] (default: %default))rDtypechoicesrIrCz--logz --log-filez --local-loglogpathz Path to a verbose appending log.)rDmetavarrCz --no-inputno_inputz--proxyproxystrz/src". The default for global installs is "/src".cCs t||jS)zGet a format_control object.)r,rD)valuesr!r#r#r$_get_format_controlsrycCs"t|j|}t||j|jdSr+)ryrxr handle_mutual_excludes no_binary only_binaryr!opt_strvaluer Zexistingr#r#r$_handle_no_binarys  rcCs"t|j|}t||j|jdSr+)ryrxr rzr|r{r}r#r#r$_handle_only_binarys  rc Cs$ttt}tdddtd|ddS)Nz --no-binaryr7callbackrYa^Do not use binary packages. Can be supplied multiple times, and each time adds to the existing value. Accepts either :all: to disable all binary packages, :none: to empty the set, or one or more package names with commas between them (no colons). Note that some packages are tricky to compile and may fail to install when this option is used on them.rDrErrRrIrC)r r@rrr7r#r#r$r{sr{c Cs$ttt}tdddtd|ddS)Nz --only-binaryr7rrYaGDo not use source packages. Can be supplied multiple times, and each time adds to the existing value. Accepts either :all: to disable all source packages, :none: to empty the set, or one or more package names with commas between them. Packages without binary distributions will fail to install when this option is used on them.r)r r@rrrr#r#r$r|sr|z --platformr=z[Only use wheels compatible with . Defaults to the platform of the running system.cCs|sdS|d}t|dkr"dSt|dkrV|d}t|dkrV|d|ddg}ztdd |D}Wntk rYd SX|dfS) z Convert a version string like "3", "37", or "3.7.3" into a tuple of ints. :return: A 2-tuple (version_info, error_msg), where `error_msg` is non-None if and only if there was a parsing error. )NN.)r#z'at most three version parts are allowedrNcss|]}t|VqdSr+)r\).0partr#r#r$ sz*_convert_python_version..)r#z$each version part must be an integer)rlentuple ValueError)rparts version_infor#r#r$_convert_python_versions    rcCs:t|\}}|dk r.d||}t|||d||j_dS)z3 Handle a provided --python-version value. Nz(invalid --python-version value: {!r}: {}r!r")rrr%rxr<)r!r~rr rZ error_msgr"r#r#r$_handle_python_version s rz--python-versionr<ra The Python interpreter version to use for wheel and "Requires-Python" compatibility checks. Defaults to a version derived from the running interpreter. The version can be specified using up to three dot-separated integers (e.g. "3" for 3.0.0, "3.7" for 3.7.0, or "3.7.3"). A major-minor version can also be given as a string without dots (e.g. "37" for 3.7.0). )rDrVrErrRrIrCz--implementationr?zOnly use wheels compatible with Python implementation , e.g. 'pp', 'jy', 'cp', or 'ip'. If not specified, then the current interpreter implementation is used. Use 'py' to force implementation-agnostic wheels.z--abir>zOnly use wheels compatible with Python abi , e.g. 'pypy_41'. If not specified, then the current interpreter abi tag is used. Generally you will need to specify --implementation, --platform, and --python-version when using this option.cCs4|t|t|t|tdSr+)r(r=r<r?r>)Zcmd_optsr#r#r$add_target_python_optionsIs   rcCst|j|j|j|jd}|S)N)r=Zpy_version_infor>r?)rr=r<r>r?)r'Z target_pythonr#r#r$make_target_pythonQsrcCstddddddS)Nz--prefer-binary prefer_binaryrGFz8Prefer older binary packages over newer source packages.rHrir#r#r#r$r]srz --cache-dir cache_dirzStore the cache data in .)rDrIrVrCc CsV|dk rJz t|Wn4tk rH}zt||t|dW5d}~XYnXd|j_dS)z Process a value provided for the --no-cache-dir option. This is an optparse.Option callback for the --no-cache-dir option. NrF)rrr%rYrxr)r!optrr excr#r#r$_handle_no_cache_dirrs  $ rz--no-cache-dirzDisable the cache.)rDrErrCz --no-depsz--no-dependenciesrAz#Don't install package dependencies.z-bz--buildz --build-dirz--build-directory build_dira>Directory to unpack packages into and build in. Note that an initial build still takes place in a temporary directory. The location of temporary directories can be controlled by setting the TMPDIR environment variable (TEMP on Windows) appropriately. When passed, build directories are not cleaned in case of failures.z--ignore-requires-pythonignore_requires_pythonz'Ignore the Requires-Python information.z--no-build-isolationZbuild_isolationZ store_falseTzDisable isolation when building a modern source distribution. Build dependencies specified by PEP 518 must be already installed if this option is used.cCs&|dk rd}t|||dd|j_dS)z Process a value provided for the --no-use-pep517 option. This is an optparse.Option callback for the no_use_pep517 option. Na0A value was passed for --no-use-pep517, probably using either the PIP_NO_USE_PEP517 environment variable or the "no-use-pep517" config file option. Use an appropriate value of the PIP_USE_PEP517 environment variable or the "use-pep517" config file option instead. rF)r%rx use_pep517)r!rrr r"r#r#r$_handle_no_use_pep517s rz --use-pep517rz^Use PEP 517 for building source distributions (use --no-use-pep517 to force legacy behaviour).z--no-use-pep517)rDrErrIrCz--install-optionr2rhr'a"Extra arguments to be supplied to the setup.py install command (use like --install-option="--install-scripts=/usr/local/bin"). Use multiple --install-option options to pass multiple options to setup.py install. If you are using an option with a directory path, be sure to use absolute path.)rDrErVrCz--global-optionr1zTExtra global options to be supplied to the setup.py call before the install command.z --no-cleanz!Don't clean up build directories.)rErIrCz--prezYInclude pre-release and development versions. By default, pip only finds stable versions.z--disable-pip-version-checkdisable_pip_version_checkz{Don't periodically check PyPI to determine whether a new version of pip is available for download. Implied with --no-index.z-Zz--always-unzip always_unzipcCs|jjsi|j_z|dd\}}Wn"tk rF|d|YnX|tkrh|d|dtf|jj|g|dS)zkGiven a value spelled "algo:digest", append the digest to a list pointed to in a dict by the algo name.:rzTArguments to %s must be a hash name followed by a value, like --hash=sha256:abcde...z&Allowed hash algorithms for %s are %s.z, N) rxhashesrrrrr setdefaultrh)r!r~rr ZalgoZdigestr#r#r$_handle_merge_hash)s  rz--hashrstringzgVerify that the package's archive matches this hash before installing. Example: --hash=sha256:abcdef...)rDrErrRrCz--require-hashesrequire_hasheszRequire a hash to check each requirement against, for repeatable installs. This option is implied when any package in a requirements file has a --hash option.z--pathz^Restrict to the specified installation path for listing packages (can be used multiple times).cCs|jr|js|jrtddS)Nz2Cannot combine '--path' with '--user' or '--local')rUuserZlocalr )r'r#r#r$check_list_path_option_srzGeneral Options)r&r'zPackage Index Options)N)F)t__doc__Z __future__rZloggingrr8Zdistutils.utilr functoolsrZoptparserrrrZpip._internal.exceptionsr Zpip._internal.locationsr r Z#pip._internal.models.format_controlr Zpip._internal.models.indexr Z"pip._internal.models.target_pythonrZpip._internal.utils.hashesrZpip._internal.utils.typingrZpip._internal.utils.uirtypingrrrrrrrZpip._internal.cli.parserrZ getLogger__name__Zloggerr%r*r;rBZhelp_rFZrequire_virtualenvrJrLrMrNlistkeysrrOrTrWrXr[r^rarbrjrkZ simple_urlrlrnrorprrrsrurvsrcryrrr{r|r=rrr<r?r>rrrrrZno_cacheZno_depsrrZno_build_isolationrrZ no_use_pep517r2r1Zno_cleanZprerrrhashrZ list_pathrZ general_groupZ index_groupr#r#r#r$sB               ,                                         site-packages/pip/_internal/cli/__pycache__/base_command.cpython-38.pyc000064400000011661151464452710022064 0ustar00U .eh@s6dZddlmZmZddlZddlZddlZddlZddlZddl Z ddl Z ddl m Z ddl mZddlmZmZddlmZmZmZmZmZddlmZmZmZmZmZdd lmZdd l m!Z!m"Z"dd l#m$Z$dd l%m&Z&dd l'm(Z(e&rddl)m*Z*m+Z+m,Z,ddlm-Z-dgZ.e/e0Z1GdddeZ2dS)z(Base Command class, and related routines)absolute_importprint_functionN) cmdoptions)CommandContextMixIn)ConfigOptionParserUpdatingDefaultsHelpFormatter)ERRORPREVIOUS_BUILD_DIR_ERRORSUCCESS UNKNOWN_ERRORVIRTUALENV_NOT_FOUND) BadCommand CommandErrorInstallationErrorPreviousBuildDirErrorUninstallationError) deprecated)BrokenStdoutLoggingError setup_logging)get_prog)MYPY_CHECK_RUNNING)running_under_virtualenv)ListTupleAny)ValuesCommandcsNeZdZdZdZdfdd ZddZddZd d Zd d Z d dZ Z S)rNFcstt||jdt|ftd||j|d}||_||_t f||_ d|j }t |j ||_ttj|j }|j |dS)Nz%s %sF)usageprogZ formatterZadd_help_optionname descriptionisolatedz %s Options)superr__init__rrr__doc__rsummaryrparser capitalizeoptparseZ OptionGroupZcmd_optsrZmake_option_groupZ general_groupZadd_option_group)selfrr%r!Z parser_kwZ optgroup_nameZgen_opts __class__B/usr/lib/python3.8/site-packages/pip/_internal/cli/base_command.pyr#4s&   zCommand.__init__cCst|drtdS)zf This is a no-op so that commands by default do not do the pip version check. Zno_indexN)hasattrAssertionError)r)optionsr,r,r-handle_pip_version_checkPsz Command.handle_pip_version_checkcCstdSN)NotImplementedError)r)r0argsr,r,r-runZsz Command.runcCs |j|Sr2)r& parse_argsr)r4r,r,r-r6^szCommand.parse_argsc Cs>z.|||W5QRW SQRXW5tXdSr2)loggingZshutdownZ main_context_mainr7r,r,r-maincs $z Command.mainc Cs||\}}|j|j|_t|j|j|jd}tjdddkrhd}t dkrZd|}t |ddd|j rxdt jd <|jrd |jt jd <|jr|jststd ttzz(|||}t|tr|WWSWntk r2}z.tt|tjd ddtWYWRSd}~XYn>t t!t"fk r}z.tt|tjd ddt#WYWSd}~XYnt$k r}z*td|tjd ddt#WYWSd}~XYnt%k rt&dtj'd|t(j)krt*j+tj'dt#YWfSt,k rDtdtjd ddt#YW2St-k rntjdddt.YWSXW5||Xt/S)N) verbosityno_colorZ user_log_file)r=zA future version of pip will drop support for Python 2.7. More details about Python 2 support in pip, can be found at https://pip.pypa.io/en/latest/development/release-process/#python-2-supportZCPythonzPython 2.7 will reach the end of its life on January 1st, 2020. Please upgrade your Python as Python 2.7 won't be maintained after that date. )Z replacementZgone_in1Z PIP_NO_INPUT ZPIP_EXISTS_ACTIONz2Could not find an activated virtualenv (required).zException information:T)exc_infoz%sz ERROR: Pipe to stdout was broken)filezOperation cancelled by userz Exception:)0r6verbosequietr;rr<logsys version_infoplatformZpython_implementationrZno_inputosenvironZ exists_actionjoinZ require_venvignore_require_venvrloggerZcriticalexitr r1r5 isinstanceintrstrdebugr rrr rrrprintstderrr8DEBUG traceback print_excKeyboardInterrupt BaseExceptionr r )r)r4r0Z level_numbermessageZstatusexcr,r,r-r9ksn             z Command._main)F) __name__ __module__ __qualname__rrLr#r1r5r6r:r9 __classcell__r,r,r*r-r0s )3r$Z __future__rrr8Zlogging.configr(rIrHrFrVZpip._internal.clirZ!pip._internal.cli.command_contextrZpip._internal.cli.parserrrZpip._internal.cli.status_codesrr r r r Zpip._internal.exceptionsr rrrrZpip._internal.utils.deprecationrZpip._internal.utils.loggingrrZpip._internal.utils.miscrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrtypingrrrr__all__Z getLoggerr\rMrr,r,r,r-s0        site-packages/pip/_internal/cli/__pycache__/status_codes.cpython-38.opt-1.pyc000064400000000555151464452710023113 0ustar00U .e@s(ddlmZdZdZdZdZdZdZdS))absolute_importN)Z __future__rSUCCESSZERRORZ UNKNOWN_ERRORZVIRTUALENV_NOT_FOUNDZPREVIOUS_BUILD_DIR_ERRORZNO_MATCHES_FOUNDr r B/usr/lib/python3.8/site-packages/pip/_internal/cli/status_codes.pys site-packages/pip/_internal/cli/status_codes.py000064400000000234151464452710015660 0ustar00from __future__ import absolute_import SUCCESS = 0 ERROR = 1 UNKNOWN_ERROR = 2 VIRTUALENV_NOT_FOUND = 3 PREVIOUS_BUILD_DIR_ERROR = 4 NO_MATCHES_FOUND = 23 site-packages/pip/_internal/cli/command_context.py000064400000001434151464452710016345 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from contextlib import contextmanager from pip._vendor.contextlib2 import ExitStack class CommandContextMixIn(object): def __init__(self): super(CommandContextMixIn, self).__init__() self._in_main_context = False self._main_context = ExitStack() @contextmanager def main_context(self): assert not self._in_main_context self._in_main_context = True try: with self._main_context: yield finally: self._in_main_context = False def enter_context(self, context_provider): assert self._in_main_context return self._main_context.enter_context(context_provider) site-packages/pip/_internal/cli/__init__.py000064400000000204151464452710014714 0ustar00"""Subpackage containing all of pip's command line interface related code """ # This file intentionally does not import submodules site-packages/pip/_internal/cli/req_command.py000064400000026143151464452710015454 0ustar00"""Contains the Command base classes that depend on PipSession. The classes in this module are in a separate module so the commands not needing download / PackageFinder capability don't unnecessarily import the PackageFinder machinery and all its vendored dependencies, etc. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import os from functools import partial from pip._internal.cli.base_command import Command from pip._internal.cli.command_context import CommandContextMixIn from pip._internal.exceptions import CommandError from pip._internal.index import PackageFinder from pip._internal.legacy_resolve import Resolver from pip._internal.models.selection_prefs import SelectionPreferences from pip._internal.network.session import PipSession from pip._internal.operations.prepare import RequirementPreparer from pip._internal.req.constructors import ( install_req_from_editable, install_req_from_line, install_req_from_req_string, ) from pip._internal.req.req_file import parse_requirements from pip._internal.self_outdated_check import ( make_link_collector, pip_self_version_check, ) from pip._internal.utils.misc import normalize_path from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from optparse import Values from typing import List, Optional, Tuple from pip._internal.cache import WheelCache from pip._internal.models.target_python import TargetPython from pip._internal.req.req_set import RequirementSet from pip._internal.req.req_tracker import RequirementTracker from pip._internal.utils.temp_dir import TempDirectory class SessionCommandMixin(CommandContextMixIn): """ A class mixin for command classes needing _build_session(). """ def __init__(self): super(SessionCommandMixin, self).__init__() self._session = None # Optional[PipSession] @classmethod def _get_index_urls(cls, options): """Return a list of index urls from user-provided options.""" index_urls = [] if not getattr(options, "no_index", False): url = getattr(options, "index_url", None) if url: index_urls.append(url) urls = getattr(options, "extra_index_urls", None) if urls: index_urls.extend(urls) # Return None rather than an empty list return index_urls or None def get_default_session(self, options): # type: (Values) -> PipSession """Get a default-managed session.""" if self._session is None: self._session = self.enter_context(self._build_session(options)) return self._session def _build_session(self, options, retries=None, timeout=None): # type: (Values, Optional[int], Optional[int]) -> PipSession session = PipSession( cache=( normalize_path(os.path.join(options.cache_dir, "http")) if options.cache_dir else None ), retries=retries if retries is not None else options.retries, trusted_hosts=options.trusted_hosts, index_urls=self._get_index_urls(options), ) # Handle custom ca-bundles from the user if options.cert: session.verify = options.cert # Handle SSL client certificate if options.client_cert: session.cert = options.client_cert # Handle timeouts if options.timeout or timeout: session.timeout = ( timeout if timeout is not None else options.timeout ) # Handle configured proxies if options.proxy: session.proxies = { "http": options.proxy, "https": options.proxy, } # Determine if we can prompt the user for authentication or not session.auth.prompting = not options.no_input return session class IndexGroupCommand(Command, SessionCommandMixin): """ Abstract base class for commands with the index_group options. This also corresponds to the commands that permit the pip version check. """ def handle_pip_version_check(self, options): # type: (Values) -> None """ Do the pip version check if not disabled. This overrides the default behavior of not doing the check. """ # Make sure the index_group options are present. assert hasattr(options, 'no_index') if options.disable_pip_version_check or options.no_index: return # Otherwise, check if we're using the latest version of pip available. session = self._build_session( options, retries=0, timeout=min(5, options.timeout) ) with session: pip_self_version_check(session, options) class RequirementCommand(IndexGroupCommand): @staticmethod def make_requirement_preparer( temp_build_dir, # type: TempDirectory options, # type: Values req_tracker, # type: RequirementTracker download_dir=None, # type: str wheel_download_dir=None, # type: str ): # type: (...) -> RequirementPreparer """ Create a RequirementPreparer instance for the given parameters. """ temp_build_dir_path = temp_build_dir.path assert temp_build_dir_path is not None return RequirementPreparer( build_dir=temp_build_dir_path, src_dir=options.src_dir, download_dir=download_dir, wheel_download_dir=wheel_download_dir, progress_bar=options.progress_bar, build_isolation=options.build_isolation, req_tracker=req_tracker, ) @staticmethod def make_resolver( preparer, # type: RequirementPreparer session, # type: PipSession finder, # type: PackageFinder options, # type: Values wheel_cache=None, # type: Optional[WheelCache] use_user_site=False, # type: bool ignore_installed=True, # type: bool ignore_requires_python=False, # type: bool force_reinstall=False, # type: bool upgrade_strategy="to-satisfy-only", # type: str use_pep517=None, # type: Optional[bool] py_version_info=None # type: Optional[Tuple[int, ...]] ): # type: (...) -> Resolver """ Create a Resolver instance for the given parameters. """ make_install_req = partial( install_req_from_req_string, isolated=options.isolated_mode, wheel_cache=wheel_cache, use_pep517=use_pep517, ) return Resolver( preparer=preparer, session=session, finder=finder, make_install_req=make_install_req, use_user_site=use_user_site, ignore_dependencies=options.ignore_dependencies, ignore_installed=ignore_installed, ignore_requires_python=ignore_requires_python, force_reinstall=force_reinstall, upgrade_strategy=upgrade_strategy, py_version_info=py_version_info ) def populate_requirement_set( self, requirement_set, # type: RequirementSet args, # type: List[str] options, # type: Values finder, # type: PackageFinder session, # type: PipSession wheel_cache, # type: Optional[WheelCache] ): # type: (...) -> None """ Marshal cmd line args into a requirement set. """ # NOTE: As a side-effect, options.require_hashes and # requirement_set.require_hashes may be updated for filename in options.constraints: for req_to_add in parse_requirements( filename, constraint=True, finder=finder, options=options, session=session, wheel_cache=wheel_cache): req_to_add.is_direct = True requirement_set.add_requirement(req_to_add) for req in args: req_to_add = install_req_from_line( req, None, isolated=options.isolated_mode, use_pep517=options.use_pep517, wheel_cache=wheel_cache ) req_to_add.is_direct = True requirement_set.add_requirement(req_to_add) for req in options.editables: req_to_add = install_req_from_editable( req, isolated=options.isolated_mode, use_pep517=options.use_pep517, wheel_cache=wheel_cache ) req_to_add.is_direct = True requirement_set.add_requirement(req_to_add) for filename in options.requirements: for req_to_add in parse_requirements( filename, finder=finder, options=options, session=session, wheel_cache=wheel_cache, use_pep517=options.use_pep517): req_to_add.is_direct = True requirement_set.add_requirement(req_to_add) # If --require-hashes was a line in a requirements file, tell # RequirementSet about it: requirement_set.require_hashes = options.require_hashes if not (args or options.editables or options.requirements): opts = {'name': self.name} if options.find_links: raise CommandError( 'You must give at least one requirement to %(name)s ' '(maybe you meant "pip %(name)s %(links)s"?)' % dict(opts, links=' '.join(options.find_links))) else: raise CommandError( 'You must give at least one requirement to %(name)s ' '(see "pip help %(name)s")' % opts) def _build_package_finder( self, options, # type: Values session, # type: PipSession target_python=None, # type: Optional[TargetPython] ignore_requires_python=None, # type: Optional[bool] ): # type: (...) -> PackageFinder """ Create a package finder appropriate to this requirement command. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. """ link_collector = make_link_collector(session, options=options) selection_prefs = SelectionPreferences( allow_yanked=True, format_control=options.format_control, allow_all_prereleases=options.pre, prefer_binary=options.prefer_binary, ignore_requires_python=ignore_requires_python, ) return PackageFinder.create( link_collector=link_collector, selection_prefs=selection_prefs, target_python=target_python, ) site-packages/pip/_internal/models/candidate.py000064400000002375151464452710015620 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from pip._vendor.packaging.version import parse as parse_version from pip._internal.utils.models import KeyBasedCompareMixin from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from pip._vendor.packaging.version import _BaseVersion from pip._internal.models.link import Link from typing import Any class InstallationCandidate(KeyBasedCompareMixin): """Represents a potential "candidate" for installation. """ def __init__(self, project, version, link): # type: (Any, str, Link) -> None self.project = project self.version = parse_version(version) # type: _BaseVersion self.link = link super(InstallationCandidate, self).__init__( key=(self.project, self.version, self.link), defining_class=InstallationCandidate ) def __repr__(self): # type: () -> str return "".format( self.project, self.version, self.link, ) def __str__(self): return '{!r} candidate (version {} at {})'.format( self.project, self.version, self.link, ) site-packages/pip/_internal/models/link.py000064400000015314151464452710014636 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import os import posixpath import re from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.utils.filetypes import WHEEL_EXTENSION from pip._internal.utils.misc import ( redact_auth_from_url, split_auth_from_netloc, splitext, ) from pip._internal.utils.models import KeyBasedCompareMixin from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import path_to_url, url_to_path if MYPY_CHECK_RUNNING: from typing import Optional, Text, Tuple, Union from pip._internal.collector import HTMLPage from pip._internal.utils.hashes import Hashes class Link(KeyBasedCompareMixin): """Represents a parsed link from a Package Index's simple URL """ def __init__( self, url, # type: str comes_from=None, # type: Optional[Union[str, HTMLPage]] requires_python=None, # type: Optional[str] yanked_reason=None, # type: Optional[Text] ): # type: (...) -> None """ :param url: url of the resource pointed to (href of the link) :param comes_from: instance of HTMLPage where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by a data-requires-python attribute in the HTML link tag, as described in PEP 503. :param yanked_reason: the reason the file has been yanked, if the file has been yanked, or None if the file hasn't been yanked. This is the value of the "data-yanked" attribute, if present, in a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. """ # url can be a UNC windows share if url.startswith('\\\\'): url = path_to_url(url) self._parsed_url = urllib_parse.urlsplit(url) # Store the url as a private attribute to prevent accidentally # trying to set a new value. self._url = url self.comes_from = comes_from self.requires_python = requires_python if requires_python else None self.yanked_reason = yanked_reason super(Link, self).__init__(key=url, defining_class=Link) def __str__(self): if self.requires_python: rp = ' (requires-python:%s)' % self.requires_python else: rp = '' if self.comes_from: return '%s (from %s)%s' % (redact_auth_from_url(self._url), self.comes_from, rp) else: return redact_auth_from_url(str(self._url)) def __repr__(self): return '' % self @property def url(self): # type: () -> str return self._url @property def filename(self): # type: () -> str path = self.path.rstrip('/') name = posixpath.basename(path) if not name: # Make sure we don't leak auth information if the netloc # includes a username and password. netloc, user_pass = split_auth_from_netloc(self.netloc) return netloc name = urllib_parse.unquote(name) assert name, ('URL %r produced no filename' % self._url) return name @property def file_path(self): # type: () -> str return url_to_path(self.url) @property def scheme(self): # type: () -> str return self._parsed_url.scheme @property def netloc(self): # type: () -> str """ This can contain auth information. """ return self._parsed_url.netloc @property def path(self): # type: () -> str return urllib_parse.unquote(self._parsed_url.path) def splitext(self): # type: () -> Tuple[str, str] return splitext(posixpath.basename(self.path.rstrip('/'))) @property def ext(self): # type: () -> str return self.splitext()[1] @property def url_without_fragment(self): # type: () -> str scheme, netloc, path, query, fragment = self._parsed_url return urllib_parse.urlunsplit((scheme, netloc, path, query, None)) _egg_fragment_re = re.compile(r'[#&]egg=([^&]*)') @property def egg_fragment(self): # type: () -> Optional[str] match = self._egg_fragment_re.search(self._url) if not match: return None return match.group(1) _subdirectory_fragment_re = re.compile(r'[#&]subdirectory=([^&]*)') @property def subdirectory_fragment(self): # type: () -> Optional[str] match = self._subdirectory_fragment_re.search(self._url) if not match: return None return match.group(1) _hash_re = re.compile( r'(sha1|sha224|sha384|sha256|sha512|md5)=([a-f0-9]+)' ) @property def hash(self): # type: () -> Optional[str] match = self._hash_re.search(self._url) if match: return match.group(2) return None @property def hash_name(self): # type: () -> Optional[str] match = self._hash_re.search(self._url) if match: return match.group(1) return None @property def show_url(self): # type: () -> Optional[str] return posixpath.basename(self._url.split('#', 1)[0].split('?', 1)[0]) @property def is_file(self): # type: () -> bool return self.scheme == 'file' def is_existing_dir(self): # type: () -> bool return self.is_file and os.path.isdir(self.file_path) @property def is_wheel(self): # type: () -> bool return self.ext == WHEEL_EXTENSION @property def is_vcs(self): # type: () -> bool from pip._internal.vcs import vcs return self.scheme in vcs.all_schemes @property def is_yanked(self): # type: () -> bool return self.yanked_reason is not None @property def has_hash(self): return self.hash_name is not None def is_hash_allowed(self, hashes): # type: (Optional[Hashes]) -> bool """ Return True if the link has a hash and it is allowed. """ if hashes is None or not self.has_hash: return False # Assert non-None so mypy knows self.hash_name and self.hash are str. assert self.hash_name is not None assert self.hash is not None return hashes.is_hash_allowed(self.hash_name, hex_digest=self.hash) site-packages/pip/_internal/models/__pycache__/link.cpython-38.opt-1.pyc000064400000014617151464452710022070 0ustar00U .e@sddlZddlZddlZddlmZddlmZddlm Z m Z m Z ddl m Z ddlmZddlmZmZerddlmZmZmZmZdd lmZdd lmZGd d d e ZdS) N)parse)WHEEL_EXTENSION)redact_auth_from_urlsplit_auth_from_netlocsplitext)KeyBasedCompareMixin)MYPY_CHECK_RUNNING) path_to_url url_to_path)OptionalTextTupleUnion)HTMLPage)Hashescs@eZdZdZd6fdd ZddZddZed d Zed d Z ed dZ eddZ eddZ eddZ ddZeddZeddZedZeddZedZedd Zed!Zed"d#Zed$d%Zed&d'Zed(d)Zd*d+Zed,d-Zed.d/Zed0d1Zed2d3Z d4d5Z!Z"S)7Linkz?Represents a parsed link from a Package Index's simple URL NcsV|drt|}t||_||_||_|r2|nd|_||_t t |j |t ddS)a` :param url: url of the resource pointed to (href of the link) :param comes_from: instance of HTMLPage where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by a data-requires-python attribute in the HTML link tag, as described in PEP 503. :param yanked_reason: the reason the file has been yanked, if the file has been yanked, or None if the file hasn't been yanked. This is the value of the "data-yanked" attribute, if present, in a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. z\\N)keyZdefining_class) startswithr urllib_parseZurlsplit _parsed_url_url comes_fromrequires_python yanked_reasonsuperr__init__)selfurlrrr __class__=/usr/lib/python3.8/site-packages/pip/_internal/models/link.pyrs  z Link.__init__cCsD|jrd|j}nd}|jr2dt|j|j|fStt|jSdS)Nz (requires-python:%s)z%s (from %s)%s)rrrrstr)rZrpr r r!__str__Es  z Link.__str__cCsd|S)Nz r rr r r!__repr__Psz Link.__repr__cCs|jSN)rr%r r r!rSszLink.urlcCs:|jd}t|}|s,t|j\}}|St|}|SN/)pathrstrip posixpathbasenamernetlocrunquote)rr*namer.Z user_passr r r!filenameXs   z Link.filenamecCs t|jSr')r rr%r r r! file_pathgszLink.file_pathcCs|jjSr')rschemer%r r r!r3lsz Link.schemecCs|jjS)z4 This can contain auth information. )rr.r%r r r!r.qsz Link.netloccCst|jjSr')rr/rr*r%r r r!r*ysz Link.pathcCstt|jdSr()rr,r-r*r+r%r r r!r~sz Link.splitextcCs |dSN)rr%r r r!extszLink.extcCs$|j\}}}}}t||||dfSr')rrZ urlunsplit)rr3r.r*ZqueryZfragmentr r r!url_without_fragmentszLink.url_without_fragmentz[#&]egg=([^&]*)cCs |j|j}|sdS|dSr4)_egg_fragment_researchrgrouprmatchr r r! egg_fragmentszLink.egg_fragmentz[#&]subdirectory=([^&]*)cCs |j|j}|sdS|dSr4)_subdirectory_fragment_rer9rr:r;r r r!subdirectory_fragmentszLink.subdirectory_fragmentz2(sha1|sha224|sha384|sha256|sha512|md5)=([a-f0-9]+)cCs |j|j}|r|dSdS)N_hash_rer9rr:r;r r r!hashs z Link.hashcCs |j|j}|r|dSdSr4rAr;r r r! hash_names zLink.hash_namecCs$t|jddddddS)N#r5r?)r,r-rsplitr%r r r!show_urlsz Link.show_urlcCs |jdkS)Nfile)r3r%r r r!is_filesz Link.is_filecCs|jotj|jSr')rJosr*isdirr2r%r r r!is_existing_dirszLink.is_existing_dircCs |jtkSr')r6rr%r r r!is_wheelsz Link.is_wheelcCsddlm}|j|jkS)Nr)vcs)Zpip._internal.vcsrOr3Z all_schemes)rrOr r r!is_vcss z Link.is_vcscCs |jdk Sr')rr%r r r! is_yankedszLink.is_yankedcCs |jdk Sr')rDr%r r r!has_hashsz Link.has_hashcCs$|dks|jsdS|j|j|jdS)zG Return True if the link has a hash and it is allowed. NF)Z hex_digest)rRis_hash_allowedrDrC)rZhashesr r r!rSszLink.is_hash_allowed)NNN)#__name__ __module__ __qualname____doc__rr$r&propertyrr1r2r3r.r*rr6r7recompiler8r=r>r?rBrCrDrHrJrMrNrPrQrRrS __classcell__r r rr!rsf'                     r)rKr,rYZpip._vendor.six.moves.urllibrrZpip._internal.utils.filetypesrZpip._internal.utils.miscrrrZpip._internal.utils.modelsrZpip._internal.utils.typingrZpip._internal.utils.urlsr r typingr r r rZpip._internal.collectorrZpip._internal.utils.hashesrrr r r r!s      site-packages/pip/_internal/models/__pycache__/index.cpython-38.pyc000064400000002172151464452710021274 0ustar00U .e$@s8ddlmZGdddeZedddZedddZd S) )parsecs(eZdZdZfddZddZZS) PackageIndexzGRepresents a Package Index and provides easier access to endpoints csDtt|||_t|j|_|d|_|d|_ ||_ dS)NZsimpleZpypi) superr__init__url urllib_parseZurlsplitZnetloc _url_for_pathZ simple_urlZpypi_urlfile_storage_domain)selfrr  __class__>/usr/lib/python3.8/site-packages/pip/_internal/models/index.pyrs   zPackageIndex.__init__cCst|j|S)N)rZurljoinr)r pathr r rrszPackageIndex._url_for_path)__name__ __module__ __qualname____doc__rr __classcell__r r r rrs rzhttps://pypi.org/zfiles.pythonhosted.org)r zhttps://test.pypi.org/ztest-files.pythonhosted.orgN)Zpip._vendor.six.moves.urllibrrobjectrZPyPIZTestPyPIr r r rs site-packages/pip/_internal/models/__pycache__/candidate.cpython-38.opt-1.pyc000064400000002662151464452710023044 0ustar00U .e@s`ddlmZddlmZddlmZerLddlmZddlm Z ddl m Z GdddeZ d S) )parse)KeyBasedCompareMixin)MYPY_CHECK_RUNNING) _BaseVersion)Link)Anycs0eZdZdZfddZddZddZZS)InstallationCandidatez9Represents a potential "candidate" for installation. cs:||_t||_||_tt|j|j|j|jftddS)N)keyZdefining_class)project parse_versionversionlinksuperr__init__)selfr r r  __class__B/usr/lib/python3.8/site-packages/pip/_internal/models/candidate.pyrs  zInstallationCandidate.__init__cCsd|j|j|jS)Nz)formatr r r rrrr__repr__s zInstallationCandidate.__repr__cCsd|j|j|jS)Nz!{!r} candidate (version {} at {})rrrrr__str__$s zInstallationCandidate.__str__)__name__ __module__ __qualname____doc__rrr __classcell__rrrrrs rN) Zpip._vendor.packaging.versionrr Zpip._internal.utils.modelsrZpip._internal.utils.typingrrZpip._internal.models.linkrtypingrrrrrrs      site-packages/pip/_internal/models/__pycache__/target_python.cpython-38.opt-1.pyc000064400000006200151464452710024007 0ustar00U .e@shddlZddlmZmZddlmZddlmZerTddlm Z m Z m Z ddlm Z Gddde ZdS) N) get_supportedversion_info_to_nodot)normalize_version_info)MYPY_CHECK_RUNNING)ListOptionalTuple) Pep425Tagc@s*eZdZdZd ddZddZddZdS) TargetPythonzx Encapsulates the properties of a Python interpreter one is targeting for a package install, download, etc. NcCsf||_|dkrtjdd}nt|}dtt|dd}||_||_||_ ||_ ||_ d|_ dS)a :param platform: A string or None. If None, searches for packages that are supported by the current system. Otherwise, will find packages that can be built on the platform passed in. These packages will only be downloaded for distribution: they will not be built locally. :param py_version_info: An optional tuple of ints representing the Python version information to use (e.g. `sys.version_info[:3]`). This can have length 1, 2, or 3 when provided. :param abi: A string or None. This is passed to pep425tags.py's get_supported() function as is. :param implementation: A string or None. This is passed to pep425tags.py's get_supported() function as is. N.) _given_py_version_infosys version_inforjoinmapstrabiimplementationplatform py_versionpy_version_info _valid_tags)selfrrrrrrF/usr/lib/python3.8/site-packages/pip/_internal/models/target_python.py__init__szTargetPython.__init__cCsZd}|jdk r$ddd|jD}d|jfd|fd|jfd|jfg}d d d|DS) zD Format the given, non-None attributes for display. Nr css|]}t|VqdS)N)r).0partrrr Csz,TargetPython.format_given..rrrr css&|]\}}|dk rd||VqdS)Nz{}={!r})format)rkeyvaluerrrr Ms)rrrrr)rZdisplay_versionZ key_valuesrrr format_given<s   zTargetPython.format_givencCsJ|jdkrD|j}|dkrd}n t|g}t||j|j|jd}||_|jS)z Return the supported PEP 425 tags to check wheel candidates against. The tags are returned in order of preference (most preferred first). N)versionsrrimpl)rrrrrrr)rrr&Ztagsrrrget_tagsRs  zTargetPython.get_tags)NNNN)__name__ __module__ __qualname____doc__rr%r(rrrrr s )r )rZpip._internal.pep425tagsrrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingrrrr objectr rrrrs   site-packages/pip/_internal/models/__pycache__/__init__.cpython-38.pyc000064400000000340151464452710021717 0ustar00U .e?@sdZdS)z8A package that contains models that represent entities. N)__doc__rrA/usr/lib/python3.8/site-packages/pip/_internal/models/__init__.pysite-packages/pip/_internal/models/__pycache__/candidate.cpython-38.pyc000064400000002662151464452710022105 0ustar00U .e@s`ddlmZddlmZddlmZerLddlmZddlm Z ddl m Z GdddeZ d S) )parse)KeyBasedCompareMixin)MYPY_CHECK_RUNNING) _BaseVersion)Link)Anycs0eZdZdZfddZddZddZZS)InstallationCandidatez9Represents a potential "candidate" for installation. cs:||_t||_||_tt|j|j|j|jftddS)N)keyZdefining_class)project parse_versionversionlinksuperr__init__)selfr r r  __class__B/usr/lib/python3.8/site-packages/pip/_internal/models/candidate.pyrs  zInstallationCandidate.__init__cCsd|j|j|jS)Nz)formatr r r rrrr__repr__s zInstallationCandidate.__repr__cCsd|j|j|jS)Nz!{!r} candidate (version {} at {})rrrrr__str__$s zInstallationCandidate.__str__)__name__ __module__ __qualname____doc__rrr __classcell__rrrrrs rN) Zpip._vendor.packaging.versionrr Zpip._internal.utils.modelsrZpip._internal.utils.typingrrZpip._internal.models.linkrtypingrrrrrrs      site-packages/pip/_internal/models/__pycache__/format_control.cpython-38.pyc000064400000004560151464452710023220 0ustar00U .e @sPddlmZddlmZddlmZers   site-packages/pip/_internal/models/__pycache__/link.cpython-38.pyc000064400000014775151464452710021136 0ustar00U .e@sddlZddlZddlZddlmZddlmZddlm Z m Z m Z ddl m Z ddlmZddlmZmZerddlmZmZmZmZdd lmZdd lmZGd d d e ZdS) N)parse)WHEEL_EXTENSION)redact_auth_from_urlsplit_auth_from_netlocsplitext)KeyBasedCompareMixin)MYPY_CHECK_RUNNING) path_to_url url_to_path)OptionalTextTupleUnion)HTMLPage)Hashescs@eZdZdZd6fdd ZddZddZed d Zed d Z ed dZ eddZ eddZ eddZ ddZeddZeddZedZeddZedZedd Zed!Zed"d#Zed$d%Zed&d'Zed(d)Zd*d+Zed,d-Zed.d/Zed0d1Zed2d3Z d4d5Z!Z"S)7Linkz?Represents a parsed link from a Package Index's simple URL NcsV|drt|}t||_||_||_|r2|nd|_||_t t |j |t ddS)a` :param url: url of the resource pointed to (href of the link) :param comes_from: instance of HTMLPage where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by a data-requires-python attribute in the HTML link tag, as described in PEP 503. :param yanked_reason: the reason the file has been yanked, if the file has been yanked, or None if the file hasn't been yanked. This is the value of the "data-yanked" attribute, if present, in a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. z\\N)keyZdefining_class) startswithr urllib_parseZurlsplit _parsed_url_url comes_fromrequires_python yanked_reasonsuperr__init__)selfurlrrr __class__=/usr/lib/python3.8/site-packages/pip/_internal/models/link.pyrs  z Link.__init__cCsD|jrd|j}nd}|jr2dt|j|j|fStt|jSdS)Nz (requires-python:%s)z%s (from %s)%s)rrrrstr)rZrpr r r!__str__Es  z Link.__str__cCsd|S)Nz r rr r r!__repr__Psz Link.__repr__cCs|jSN)rr%r r r!rSszLink.urlcCsL|jd}t|}|s,t|j\}}|St|}|sHtd|j |S)N/zURL %r produced no filename) pathrstrip posixpathbasenamernetlocrunquoteAssertionErrorr)rr)namer-Z user_passr r r!filenameXs   z Link.filenamecCs t|jSr')r rr%r r r! file_pathgszLink.file_pathcCs|jjSr')rschemer%r r r!r3lsz Link.schemecCs|jjS)z4 This can contain auth information. )rr-r%r r r!r-qsz Link.netloccCst|jjSr')rr.rr)r%r r r!r)ysz Link.pathcCstt|jdS)Nr()rr+r,r)r*r%r r r!r~sz Link.splitextcCs |dSN)rr%r r r!extszLink.extcCs$|j\}}}}}t||||dfSr')rrZ urlunsplit)rr3r-r)ZqueryZfragmentr r r!url_without_fragmentszLink.url_without_fragmentz[#&]egg=([^&]*)cCs |j|j}|sdS|dSr4)_egg_fragment_researchrgrouprmatchr r r! egg_fragmentszLink.egg_fragmentz[#&]subdirectory=([^&]*)cCs |j|j}|sdS|dSr4)_subdirectory_fragment_rer9rr:r;r r r!subdirectory_fragmentszLink.subdirectory_fragmentz2(sha1|sha224|sha384|sha256|sha512|md5)=([a-f0-9]+)cCs |j|j}|r|dSdS)N_hash_rer9rr:r;r r r!hashs z Link.hashcCs |j|j}|r|dSdSr4rAr;r r r! hash_names zLink.hash_namecCs$t|jddddddS)N#r5r?)r+r,rsplitr%r r r!show_urlsz Link.show_urlcCs |jdkS)Nfile)r3r%r r r!is_filesz Link.is_filecCs|jotj|jSr')rJosr)isdirr2r%r r r!is_existing_dirszLink.is_existing_dircCs |jtkSr')r6rr%r r r!is_wheelsz Link.is_wheelcCsddlm}|j|jkS)Nr)vcs)Zpip._internal.vcsrOr3Z all_schemes)rrOr r r!is_vcss z Link.is_vcscCs |jdk Sr')rr%r r r! is_yankedszLink.is_yankedcCs |jdk Sr')rDr%r r r!has_hashsz Link.has_hashcCs@|dks|jsdS|jdk s t|jdk s.t|j|j|jdS)zG Return True if the link has a hash and it is allowed. NF)Z hex_digest)rRrDr/rCis_hash_allowed)rZhashesr r r!rSs zLink.is_hash_allowed)NNN)#__name__ __module__ __qualname____doc__rr$r&propertyrr1r2r3r-r)rr6r7recompiler8r=r>r?rBrCrDrHrJrMrNrPrQrRrS __classcell__r r rr!rsf'                     r)rKr+rYZpip._vendor.six.moves.urllibrrZpip._internal.utils.filetypesrZpip._internal.utils.miscrrrZpip._internal.utils.modelsrZpip._internal.utils.typingrZpip._internal.utils.urlsr r typingr r r rZpip._internal.collectorrZpip._internal.utils.hashesrrr r r r!s      site-packages/pip/_internal/models/__pycache__/selection_prefs.cpython-38.pyc000064400000003074151464452710023353 0ustar00U .et@s<ddlmZer(ddlmZddlmZGdddeZdS))MYPY_CHECK_RUNNING)Optional) FormatControlc@seZdZdZdddZdS)SelectionPreferenceszd Encapsulates the candidate selection preferences for downloading and installing files. FNcCs.|dkr d}||_||_||_||_||_dS)awCreate a SelectionPreferences object. :param allow_yanked: Whether files marked as yanked (in the sense of PEP 592) are permitted to be candidates for install. :param format_control: A FormatControl object or None. Used to control the selection of source packages / binary packages when consulting the index and links. :param prefer_binary: Whether to prefer an old, but valid, binary dist over a new source dist. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. NF) allow_yankedallow_all_prereleasesformat_control prefer_binaryignore_requires_python)selfrrrr r r H/usr/lib/python3.8/site-packages/pip/_internal/models/selection_prefs.py__init__szSelectionPreferences.__init__)FNFN)__name__ __module__ __qualname____doc__rr r r r rs  rN)Zpip._internal.utils.typingrtypingrZ#pip._internal.models.format_controlrobjectrr r r r s   site-packages/pip/_internal/models/__pycache__/format_control.cpython-38.opt-1.pyc000064400000004560151464452710024157 0ustar00U .e @sPddlmZddlmZddlmZers   site-packages/pip/_internal/models/__pycache__/index.cpython-38.opt-1.pyc000064400000002172151464452710022233 0ustar00U .e$@s8ddlmZGdddeZedddZedddZd S) )parsecs(eZdZdZfddZddZZS) PackageIndexzGRepresents a Package Index and provides easier access to endpoints csDtt|||_t|j|_|d|_|d|_ ||_ dS)NZsimpleZpypi) superr__init__url urllib_parseZurlsplitZnetloc _url_for_pathZ simple_urlZpypi_urlfile_storage_domain)selfrr  __class__>/usr/lib/python3.8/site-packages/pip/_internal/models/index.pyrs   zPackageIndex.__init__cCst|j|S)N)rZurljoinr)r pathr r rrszPackageIndex._url_for_path)__name__ __module__ __qualname____doc__rr __classcell__r r r rrs rzhttps://pypi.org/zfiles.pythonhosted.org)r zhttps://test.pypi.org/ztest-files.pythonhosted.orgN)Zpip._vendor.six.moves.urllibrrobjectrZPyPIZTestPyPIr r r rs site-packages/pip/_internal/models/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000340151464452710022656 0ustar00U .e?@sdZdS)z8A package that contains models that represent entities. N)__doc__rrA/usr/lib/python3.8/site-packages/pip/_internal/models/__init__.pysite-packages/pip/_internal/models/__pycache__/target_python.cpython-38.pyc000064400000006200151464452710023050 0ustar00U .e@shddlZddlmZmZddlmZddlmZerTddlm Z m Z m Z ddlm Z Gddde ZdS) N) get_supportedversion_info_to_nodot)normalize_version_info)MYPY_CHECK_RUNNING)ListOptionalTuple) Pep425Tagc@s*eZdZdZd ddZddZddZdS) TargetPythonzx Encapsulates the properties of a Python interpreter one is targeting for a package install, download, etc. NcCsf||_|dkrtjdd}nt|}dtt|dd}||_||_||_ ||_ ||_ d|_ dS)a :param platform: A string or None. If None, searches for packages that are supported by the current system. Otherwise, will find packages that can be built on the platform passed in. These packages will only be downloaded for distribution: they will not be built locally. :param py_version_info: An optional tuple of ints representing the Python version information to use (e.g. `sys.version_info[:3]`). This can have length 1, 2, or 3 when provided. :param abi: A string or None. This is passed to pep425tags.py's get_supported() function as is. :param implementation: A string or None. This is passed to pep425tags.py's get_supported() function as is. N.) _given_py_version_infosys version_inforjoinmapstrabiimplementationplatform py_versionpy_version_info _valid_tags)selfrrrrrrF/usr/lib/python3.8/site-packages/pip/_internal/models/target_python.py__init__szTargetPython.__init__cCsZd}|jdk r$ddd|jD}d|jfd|fd|jfd|jfg}d d d|DS) zD Format the given, non-None attributes for display. Nr css|]}t|VqdS)N)r).0partrrr Csz,TargetPython.format_given..rrrr css&|]\}}|dk rd||VqdS)Nz{}={!r})format)rkeyvaluerrrr Ms)rrrrr)rZdisplay_versionZ key_valuesrrr format_given<s   zTargetPython.format_givencCsJ|jdkrD|j}|dkrd}n t|g}t||j|j|jd}||_|jS)z Return the supported PEP 425 tags to check wheel candidates against. The tags are returned in order of preference (most preferred first). N)versionsrrimpl)rrrrrrr)rrr&Ztagsrrrget_tagsRs  zTargetPython.get_tags)NNNN)__name__ __module__ __qualname____doc__rr%r(rrrrr s )r )rZpip._internal.pep425tagsrrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingrrrr objectr rrrrs   site-packages/pip/_internal/models/__pycache__/selection_prefs.cpython-38.opt-1.pyc000064400000003074151464452710024312 0ustar00U .et@s<ddlmZer(ddlmZddlmZGdddeZdS))MYPY_CHECK_RUNNING)Optional) FormatControlc@seZdZdZdddZdS)SelectionPreferenceszd Encapsulates the candidate selection preferences for downloading and installing files. FNcCs.|dkr d}||_||_||_||_||_dS)awCreate a SelectionPreferences object. :param allow_yanked: Whether files marked as yanked (in the sense of PEP 592) are permitted to be candidates for install. :param format_control: A FormatControl object or None. Used to control the selection of source packages / binary packages when consulting the index and links. :param prefer_binary: Whether to prefer an old, but valid, binary dist over a new source dist. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. NF) allow_yankedallow_all_prereleasesformat_control prefer_binaryignore_requires_python)selfrrrr r r H/usr/lib/python3.8/site-packages/pip/_internal/models/selection_prefs.py__init__szSelectionPreferences.__init__)FNFN)__name__ __module__ __qualname____doc__rr r r r rs  rN)Zpip._internal.utils.typingrtypingrZ#pip._internal.models.format_controlrobjectrr r r r s   site-packages/pip/_internal/models/__pycache__/search_scope.cpython-38.pyc000064400000006263151464452710022630 0ustar00U .e@sddlZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddl mZmZddlmZer|ddlmZeeZGd d d eZdS) N)canonicalize_name)parse)PyPI)HAS_TLS)normalize_pathredact_auth_from_url)MYPY_CHECK_RUNNING)Listc@s4eZdZdZeddZddZddZdd Zd S) SearchScopezF Encapsulates the locations that pip is configured to search. cCs~g}|D]0}|dr.t|}tj|r.|}||qtsrt||D]&}t |}|j dkrJt dqrqJ|||dS)zQ Create a SearchScope object after normalizing the `find_links`. ~Zhttpszipip is configured with locations that require TLS/SSL, however the ssl module in Python is not available. find_links index_urls) startswithrospathexistsappendr itertoolschain urllib_parseZurlparseZschemeloggerZwarning)clsr rZbuilt_find_linkslinkZnew_linkZparsedrE/usr/lib/python3.8/site-packages/pip/_internal/models/search_scope.pycreates&     zSearchScope.createcCs||_||_dSNr )selfr rrrr__init__GszSearchScope.__init__cCslg}|jr:|jtjgkr:|dddd|jD|jrb|dddd|jDd|S)NzLooking in indexes: {}z, css|]}t|VqdSrr.0urlrrr Usz6SearchScope.get_formatted_locations..zLooking in links: {}css|]}t|VqdSrr r!rrrr$Zs )rrZ simple_urlrformatjoinr )rlinesrrrget_formatted_locationsPsz#SearchScope.get_formatted_locationscs fddfdd|jDS)zReturns the locations found via self.index_urls Checks the url_name on the main (first in the list) index and use this url_name to produce all locations cs,t|tt}|ds(|d}|S)N/) posixpathr'rZquoterendswith)r#Zloc) project_namerrmkurl_pypi_urlgs  z.mkurl_pypi_urlcsg|] }|qSrrr!)r.rr tsz8SearchScope.get_index_urls_locations..)r)rr-r)r.r-rget_index_urls_locations_s z$SearchScope.get_index_urls_locationsN) __name__ __module__ __qualname____doc__ classmethodrrr)r0rrrrr s  ( r )rZloggingrr+Zpip._vendor.packaging.utilsrZpip._vendor.six.moves.urllibrrZpip._internal.models.indexrZpip._internal.utils.compatrZpip._internal.utils.miscrrZpip._internal.utils.typingrtypingr Z getLoggerr1robjectr rrrrs       site-packages/pip/_internal/models/__pycache__/search_scope.cpython-38.opt-1.pyc000064400000006263151464452710023567 0ustar00U .e@sddlZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddl mZmZddlmZer|ddlmZeeZGd d d eZdS) N)canonicalize_name)parse)PyPI)HAS_TLS)normalize_pathredact_auth_from_url)MYPY_CHECK_RUNNING)Listc@s4eZdZdZeddZddZddZdd Zd S) SearchScopezF Encapsulates the locations that pip is configured to search. cCs~g}|D]0}|dr.t|}tj|r.|}||qtsrt||D]&}t |}|j dkrJt dqrqJ|||dS)zQ Create a SearchScope object after normalizing the `find_links`. ~Zhttpszipip is configured with locations that require TLS/SSL, however the ssl module in Python is not available. find_links index_urls) startswithrospathexistsappendr itertoolschain urllib_parseZurlparseZschemeloggerZwarning)clsr rZbuilt_find_linkslinkZnew_linkZparsedrE/usr/lib/python3.8/site-packages/pip/_internal/models/search_scope.pycreates&     zSearchScope.createcCs||_||_dSNr )selfr rrrr__init__GszSearchScope.__init__cCslg}|jr:|jtjgkr:|dddd|jD|jrb|dddd|jDd|S)NzLooking in indexes: {}z, css|]}t|VqdSrr.0urlrrr Usz6SearchScope.get_formatted_locations..zLooking in links: {}css|]}t|VqdSrr r!rrrr$Zs )rrZ simple_urlrformatjoinr )rlinesrrrget_formatted_locationsPsz#SearchScope.get_formatted_locationscs fddfdd|jDS)zReturns the locations found via self.index_urls Checks the url_name on the main (first in the list) index and use this url_name to produce all locations cs,t|tt}|ds(|d}|S)N/) posixpathr'rZquoterendswith)r#Zloc) project_namerrmkurl_pypi_urlgs  z.mkurl_pypi_urlcsg|] }|qSrrr!)r.rr tsz8SearchScope.get_index_urls_locations..)r)rr-r)r.r-rget_index_urls_locations_s z$SearchScope.get_index_urls_locationsN) __name__ __module__ __qualname____doc__ classmethodrrr)r0rrrrr s  ( r )rZloggingrr+Zpip._vendor.packaging.utilsrZpip._vendor.six.moves.urllibrrZpip._internal.models.indexrZpip._internal.utils.compatrZpip._internal.utils.miscrrZpip._internal.utils.typingrtypingr Z getLoggerr1robjectr rrrrs       site-packages/pip/_internal/models/format_control.py000064400000005040151464452710016724 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from pip._vendor.packaging.utils import canonicalize_name from pip._internal.exceptions import CommandError from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional, Set, FrozenSet class FormatControl(object): """Helper for managing formats from which a package can be installed. """ def __init__(self, no_binary=None, only_binary=None): # type: (Optional[Set], Optional[Set]) -> None if no_binary is None: no_binary = set() if only_binary is None: only_binary = set() self.no_binary = no_binary self.only_binary = only_binary def __eq__(self, other): return self.__dict__ == other.__dict__ def __ne__(self, other): return not self.__eq__(other) def __repr__(self): return "{}({}, {})".format( self.__class__.__name__, self.no_binary, self.only_binary ) @staticmethod def handle_mutual_excludes(value, target, other): # type: (str, Optional[Set], Optional[Set]) -> None if value.startswith('-'): raise CommandError( "--no-binary / --only-binary option requires 1 argument." ) new = value.split(',') while ':all:' in new: other.clear() target.clear() target.add(':all:') del new[:new.index(':all:') + 1] # Without a none, we want to discard everything as :all: covers it if ':none:' not in new: return for name in new: if name == ':none:': target.clear() continue name = canonicalize_name(name) other.discard(name) target.add(name) def get_allowed_formats(self, canonical_name): # type: (str) -> FrozenSet result = {"binary", "source"} if canonical_name in self.only_binary: result.discard('source') elif canonical_name in self.no_binary: result.discard('binary') elif ':all:' in self.only_binary: result.discard('source') elif ':all:' in self.no_binary: result.discard('binary') return frozenset(result) def disallow_binaries(self): # type: () -> None self.handle_mutual_excludes( ':all:', self.no_binary, self.only_binary, ) site-packages/pip/_internal/models/search_scope.py000064400000007603151464452710016341 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import itertools import logging import os import posixpath from pip._vendor.packaging.utils import canonicalize_name from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.models.index import PyPI from pip._internal.utils.compat import HAS_TLS from pip._internal.utils.misc import normalize_path, redact_auth_from_url from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import List logger = logging.getLogger(__name__) class SearchScope(object): """ Encapsulates the locations that pip is configured to search. """ @classmethod def create( cls, find_links, # type: List[str] index_urls, # type: List[str] ): # type: (...) -> SearchScope """ Create a SearchScope object after normalizing the `find_links`. """ # Build find_links. If an argument starts with ~, it may be # a local file relative to a home directory. So try normalizing # it and if it exists, use the normalized version. # This is deliberately conservative - it might be fine just to # blindly normalize anything starting with a ~... built_find_links = [] # type: List[str] for link in find_links: if link.startswith('~'): new_link = normalize_path(link) if os.path.exists(new_link): link = new_link built_find_links.append(link) # If we don't have TLS enabled, then WARN if anyplace we're looking # relies on TLS. if not HAS_TLS: for link in itertools.chain(index_urls, built_find_links): parsed = urllib_parse.urlparse(link) if parsed.scheme == 'https': logger.warning( 'pip is configured with locations that require ' 'TLS/SSL, however the ssl module in Python is not ' 'available.' ) break return cls( find_links=built_find_links, index_urls=index_urls, ) def __init__( self, find_links, # type: List[str] index_urls, # type: List[str] ): # type: (...) -> None self.find_links = find_links self.index_urls = index_urls def get_formatted_locations(self): # type: () -> str lines = [] if self.index_urls and self.index_urls != [PyPI.simple_url]: lines.append( 'Looking in indexes: {}'.format(', '.join( redact_auth_from_url(url) for url in self.index_urls)) ) if self.find_links: lines.append( 'Looking in links: {}'.format(', '.join( redact_auth_from_url(url) for url in self.find_links)) ) return '\n'.join(lines) def get_index_urls_locations(self, project_name): # type: (str) -> List[str] """Returns the locations found via self.index_urls Checks the url_name on the main (first in the list) index and use this url_name to produce all locations """ def mkurl_pypi_url(url): loc = posixpath.join( url, urllib_parse.quote(canonicalize_name(project_name))) # For maximum compatibility with easy_install, ensure the path # ends in a trailing slash. Although this isn't in the spec # (and PyPI can handle it without the slash) some other index # implementations might break if they relied on easy_install's # behavior. if not loc.endswith('/'): loc = loc + '/' return loc return [mkurl_pypi_url(url) for url in self.index_urls] site-packages/pip/_internal/models/selection_prefs.py000064400000003564151464452710017071 0ustar00from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional from pip._internal.models.format_control import FormatControl class SelectionPreferences(object): """ Encapsulates the candidate selection preferences for downloading and installing files. """ # Don't include an allow_yanked default value to make sure each call # site considers whether yanked releases are allowed. This also causes # that decision to be made explicit in the calling code, which helps # people when reading the code. def __init__( self, allow_yanked, # type: bool allow_all_prereleases=False, # type: bool format_control=None, # type: Optional[FormatControl] prefer_binary=False, # type: bool ignore_requires_python=None, # type: Optional[bool] ): # type: (...) -> None """Create a SelectionPreferences object. :param allow_yanked: Whether files marked as yanked (in the sense of PEP 592) are permitted to be candidates for install. :param format_control: A FormatControl object or None. Used to control the selection of source packages / binary packages when consulting the index and links. :param prefer_binary: Whether to prefer an old, but valid, binary dist over a new source dist. :param ignore_requires_python: Whether to ignore incompatible "Requires-Python" values in links. Defaults to False. """ if ignore_requires_python is None: ignore_requires_python = False self.allow_yanked = allow_yanked self.allow_all_prereleases = allow_all_prereleases self.format_control = format_control self.prefer_binary = prefer_binary self.ignore_requires_python = ignore_requires_python site-packages/pip/_internal/models/__init__.py000064400000000077151464452710015440 0ustar00"""A package that contains models that represent entities. """ site-packages/pip/_internal/models/target_python.py000064400000007354151464452710016575 0ustar00import sys from pip._internal.pep425tags import get_supported, version_info_to_nodot from pip._internal.utils.misc import normalize_version_info from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import List, Optional, Tuple from pip._internal.pep425tags import Pep425Tag class TargetPython(object): """ Encapsulates the properties of a Python interpreter one is targeting for a package install, download, etc. """ def __init__( self, platform=None, # type: Optional[str] py_version_info=None, # type: Optional[Tuple[int, ...]] abi=None, # type: Optional[str] implementation=None, # type: Optional[str] ): # type: (...) -> None """ :param platform: A string or None. If None, searches for packages that are supported by the current system. Otherwise, will find packages that can be built on the platform passed in. These packages will only be downloaded for distribution: they will not be built locally. :param py_version_info: An optional tuple of ints representing the Python version information to use (e.g. `sys.version_info[:3]`). This can have length 1, 2, or 3 when provided. :param abi: A string or None. This is passed to pep425tags.py's get_supported() function as is. :param implementation: A string or None. This is passed to pep425tags.py's get_supported() function as is. """ # Store the given py_version_info for when we call get_supported(). self._given_py_version_info = py_version_info if py_version_info is None: py_version_info = sys.version_info[:3] else: py_version_info = normalize_version_info(py_version_info) py_version = '.'.join(map(str, py_version_info[:2])) self.abi = abi self.implementation = implementation self.platform = platform self.py_version = py_version self.py_version_info = py_version_info # This is used to cache the return value of get_tags(). self._valid_tags = None # type: Optional[List[Pep425Tag]] def format_given(self): # type: () -> str """ Format the given, non-None attributes for display. """ display_version = None if self._given_py_version_info is not None: display_version = '.'.join( str(part) for part in self._given_py_version_info ) key_values = [ ('platform', self.platform), ('version_info', display_version), ('abi', self.abi), ('implementation', self.implementation), ] return ' '.join( '{}={!r}'.format(key, value) for key, value in key_values if value is not None ) def get_tags(self): # type: () -> List[Pep425Tag] """ Return the supported PEP 425 tags to check wheel candidates against. The tags are returned in order of preference (most preferred first). """ if self._valid_tags is None: # Pass versions=None if no py_version_info was given since # versions=None uses special default logic. py_version_info = self._given_py_version_info if py_version_info is None: versions = None else: versions = [version_info_to_nodot(py_version_info)] tags = get_supported( versions=versions, platform=self.platform, abi=self.abi, impl=self.implementation, ) self._valid_tags = tags return self._valid_tags site-packages/pip/_internal/models/index.py000064400000002044151464452710015004 0ustar00from pip._vendor.six.moves.urllib import parse as urllib_parse class PackageIndex(object): """Represents a Package Index and provides easier access to endpoints """ def __init__(self, url, file_storage_domain): # type: (str, str) -> None super(PackageIndex, self).__init__() self.url = url self.netloc = urllib_parse.urlsplit(url).netloc self.simple_url = self._url_for_path('simple') self.pypi_url = self._url_for_path('pypi') # This is part of a temporary hack used to block installs of PyPI # packages which depend on external urls only necessary until PyPI can # block such packages themselves self.file_storage_domain = file_storage_domain def _url_for_path(self, path): # type: (str) -> str return urllib_parse.urljoin(self.url, path) PyPI = PackageIndex( 'https://pypi.org/', file_storage_domain='files.pythonhosted.org' ) TestPyPI = PackageIndex( 'https://test.pypi.org/', file_storage_domain='test-files.pythonhosted.org' ) site-packages/pip/_internal/collector.py000064400000043127151464452710014407 0ustar00""" The main purpose of this module is to expose LinkCollector.collect_links(). """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import cgi import itertools import logging import mimetypes import os from collections import OrderedDict from pip._vendor import html5lib, requests from pip._vendor.distlib.compat import unescape from pip._vendor.requests.exceptions import HTTPError, RetryError, SSLError from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._vendor.six.moves.urllib import request as urllib_request from pip._internal.models.link import Link from pip._internal.utils.filetypes import ARCHIVE_EXTENSIONS from pip._internal.utils.misc import redact_auth_from_url from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import path_to_url, url_to_path from pip._internal.vcs import is_url, vcs if MYPY_CHECK_RUNNING: from typing import ( Callable, Dict, Iterable, List, MutableMapping, Optional, Sequence, Tuple, Union, ) import xml.etree.ElementTree from pip._vendor.requests import Response from pip._internal.models.search_scope import SearchScope from pip._internal.network.session import PipSession HTMLElement = xml.etree.ElementTree.Element ResponseHeaders = MutableMapping[str, str] logger = logging.getLogger(__name__) def _match_vcs_scheme(url): # type: (str) -> Optional[str] """Look for VCS schemes in the URL. Returns the matched VCS scheme, or None if there's no match. """ for scheme in vcs.schemes: if url.lower().startswith(scheme) and url[len(scheme)] in '+:': return scheme return None def _is_url_like_archive(url): # type: (str) -> bool """Return whether the URL looks like an archive. """ filename = Link(url).filename for bad_ext in ARCHIVE_EXTENSIONS: if filename.endswith(bad_ext): return True return False class _NotHTML(Exception): def __init__(self, content_type, request_desc): # type: (str, str) -> None super(_NotHTML, self).__init__(content_type, request_desc) self.content_type = content_type self.request_desc = request_desc def _ensure_html_header(response): # type: (Response) -> None """Check the Content-Type header to ensure the response contains HTML. Raises `_NotHTML` if the content type is not text/html. """ content_type = response.headers.get("Content-Type", "") if not content_type.lower().startswith("text/html"): raise _NotHTML(content_type, response.request.method) class _NotHTTP(Exception): pass def _ensure_html_response(url, session): # type: (str, PipSession) -> None """Send a HEAD request to the URL, and ensure the response contains HTML. Raises `_NotHTTP` if the URL is not available for a HEAD request, or `_NotHTML` if the content type is not text/html. """ scheme, netloc, path, query, fragment = urllib_parse.urlsplit(url) if scheme not in {'http', 'https'}: raise _NotHTTP() resp = session.head(url, allow_redirects=True) resp.raise_for_status() _ensure_html_header(resp) def _get_html_response(url, session): # type: (str, PipSession) -> Response """Access an HTML page with GET, and return the response. This consists of three parts: 1. If the URL looks suspiciously like an archive, send a HEAD first to check the Content-Type is HTML, to avoid downloading a large file. Raise `_NotHTTP` if the content type cannot be determined, or `_NotHTML` if it is not HTML. 2. Actually perform the request. Raise HTTP exceptions on network failures. 3. Check the Content-Type header to make sure we got HTML, and raise `_NotHTML` otherwise. """ if _is_url_like_archive(url): _ensure_html_response(url, session=session) logger.debug('Getting page %s', redact_auth_from_url(url)) resp = session.get( url, headers={ "Accept": "text/html", # We don't want to blindly returned cached data for # /simple/, because authors generally expecting that # twine upload && pip install will function, but if # they've done a pip install in the last ~10 minutes # it won't. Thus by setting this to zero we will not # blindly use any cached data, however the benefit of # using max-age=0 instead of no-cache, is that we will # still support conditional requests, so we will still # minimize traffic sent in cases where the page hasn't # changed at all, we will just always incur the round # trip for the conditional GET now instead of only # once per 10 minutes. # For more information, please see pypa/pip#5670. "Cache-Control": "max-age=0", }, ) resp.raise_for_status() # The check for archives above only works if the url ends with # something that looks like an archive. However that is not a # requirement of an url. Unless we issue a HEAD request on every # url we cannot know ahead of time for sure if something is HTML # or not. However we can check after we've downloaded it. _ensure_html_header(resp) return resp def _get_encoding_from_headers(headers): # type: (ResponseHeaders) -> Optional[str] """Determine if we have any encoding information in our headers. """ if headers and "Content-Type" in headers: content_type, params = cgi.parse_header(headers["Content-Type"]) if "charset" in params: return params['charset'] return None def _determine_base_url(document, page_url): # type: (HTMLElement, str) -> str """Determine the HTML document's base URL. This looks for a ```` tag in the HTML document. If present, its href attribute denotes the base URL of anchor tags in the document. If there is no such tag (or if it does not have a valid href attribute), the HTML file's URL is used as the base URL. :param document: An HTML document representation. The current implementation expects the result of ``html5lib.parse()``. :param page_url: The URL of the HTML document. """ for base in document.findall(".//base"): href = base.get("href") if href is not None: return href return page_url def _clean_link(url): # type: (str) -> str """Makes sure a link is fully encoded. That is, if a ' ' shows up in the link, it will be rewritten to %20 (while not over-quoting % or other characters).""" # Split the URL into parts according to the general structure # `scheme://netloc/path;parameters?query#fragment`. Note that the # `netloc` can be empty and the URI will then refer to a local # filesystem path. result = urllib_parse.urlparse(url) # In both cases below we unquote prior to quoting to make sure # nothing is double quoted. if result.netloc == "": # On Windows the path part might contain a drive letter which # should not be quoted. On Linux where drive letters do not # exist, the colon should be quoted. We rely on urllib.request # to do the right thing here. path = urllib_request.pathname2url( urllib_request.url2pathname(result.path)) else: # In addition to the `/` character we protect `@` so that # revision strings in VCS URLs are properly parsed. path = urllib_parse.quote(urllib_parse.unquote(result.path), safe="/@") return urllib_parse.urlunparse(result._replace(path=path)) def _create_link_from_element( anchor, # type: HTMLElement page_url, # type: str base_url, # type: str ): # type: (...) -> Optional[Link] """ Convert an anchor element in a simple repository page to a Link. """ href = anchor.get("href") if not href: return None url = _clean_link(urllib_parse.urljoin(base_url, href)) pyrequire = anchor.get('data-requires-python') pyrequire = unescape(pyrequire) if pyrequire else None yanked_reason = anchor.get('data-yanked') if yanked_reason: # This is a unicode string in Python 2 (and 3). yanked_reason = unescape(yanked_reason) link = Link( url, comes_from=page_url, requires_python=pyrequire, yanked_reason=yanked_reason, ) return link def parse_links(page): # type: (HTMLPage) -> Iterable[Link] """ Parse an HTML document, and yield its anchor elements as Link objects. """ document = html5lib.parse( page.content, transport_encoding=page.encoding, namespaceHTMLElements=False, ) url = page.url base_url = _determine_base_url(document, url) for anchor in document.findall(".//a"): link = _create_link_from_element( anchor, page_url=url, base_url=base_url, ) if link is None: continue yield link class HTMLPage(object): """Represents one page, along with its URL""" def __init__( self, content, # type: bytes encoding, # type: Optional[str] url, # type: str ): # type: (...) -> None """ :param encoding: the encoding to decode the given content. :param url: the URL from which the HTML was downloaded. """ self.content = content self.encoding = encoding self.url = url def __str__(self): return redact_auth_from_url(self.url) def _handle_get_page_fail( link, # type: Link reason, # type: Union[str, Exception] meth=None # type: Optional[Callable[..., None]] ): # type: (...) -> None if meth is None: meth = logger.debug meth("Could not fetch URL %s: %s - skipping", link, reason) def _make_html_page(response): # type: (Response) -> HTMLPage encoding = _get_encoding_from_headers(response.headers) return HTMLPage(response.content, encoding=encoding, url=response.url) def _get_html_page(link, session=None): # type: (Link, Optional[PipSession]) -> Optional[HTMLPage] if session is None: raise TypeError( "_get_html_page() missing 1 required keyword argument: 'session'" ) url = link.url.split('#', 1)[0] # Check for VCS schemes that do not support lookup as web pages. vcs_scheme = _match_vcs_scheme(url) if vcs_scheme: logger.debug('Cannot look at %s URL %s', vcs_scheme, link) return None # Tack index.html onto file:// URLs that point to directories scheme, _, path, _, _, _ = urllib_parse.urlparse(url) if (scheme == 'file' and os.path.isdir(urllib_request.url2pathname(path))): # add trailing slash if not present so urljoin doesn't trim # final segment if not url.endswith('/'): url += '/' url = urllib_parse.urljoin(url, 'index.html') logger.debug(' file: URL is directory, getting %s', url) try: resp = _get_html_response(url, session=session) except _NotHTTP: logger.debug( 'Skipping page %s because it looks like an archive, and cannot ' 'be checked by HEAD.', link, ) except _NotHTML as exc: logger.debug( 'Skipping page %s because the %s request got Content-Type: %s', link, exc.request_desc, exc.content_type, ) except HTTPError as exc: _handle_get_page_fail(link, exc) except RetryError as exc: _handle_get_page_fail(link, exc) except SSLError as exc: reason = "There was a problem confirming the ssl certificate: " reason += str(exc) _handle_get_page_fail(link, reason, meth=logger.info) except requests.ConnectionError as exc: _handle_get_page_fail(link, "connection error: %s" % exc) except requests.Timeout: _handle_get_page_fail(link, "timed out") else: return _make_html_page(resp) return None def _remove_duplicate_links(links): # type: (Iterable[Link]) -> List[Link] """ Return a list of links, with duplicates removed and ordering preserved. """ # We preserve the ordering when removing duplicates because we can. return list(OrderedDict.fromkeys(links)) def group_locations(locations, expand_dir=False): # type: (Sequence[str], bool) -> Tuple[List[str], List[str]] """ Divide a list of locations into two groups: "files" (archives) and "urls." :return: A pair of lists (files, urls). """ files = [] urls = [] # puts the url for the given file path into the appropriate list def sort_path(path): url = path_to_url(path) if mimetypes.guess_type(url, strict=False)[0] == 'text/html': urls.append(url) else: files.append(url) for url in locations: is_local_path = os.path.exists(url) is_file_url = url.startswith('file:') if is_local_path or is_file_url: if is_local_path: path = url else: path = url_to_path(url) if os.path.isdir(path): if expand_dir: path = os.path.realpath(path) for item in os.listdir(path): sort_path(os.path.join(path, item)) elif is_file_url: urls.append(url) else: logger.warning( "Path '{0}' is ignored: " "it is a directory.".format(path), ) elif os.path.isfile(path): sort_path(path) else: logger.warning( "Url '%s' is ignored: it is neither a file " "nor a directory.", url, ) elif is_url(url): # Only add url with clear scheme urls.append(url) else: logger.warning( "Url '%s' is ignored. It is either a non-existing " "path or lacks a specific scheme.", url, ) return files, urls class CollectedLinks(object): """ Encapsulates all the Link objects collected by a call to LinkCollector.collect_links(), stored separately as-- (1) links from the configured file locations, (2) links from the configured find_links, and (3) a dict mapping HTML page url to links from that page. """ def __init__( self, files, # type: List[Link] find_links, # type: List[Link] pages, # type: Dict[str, List[Link]] ): # type: (...) -> None """ :param files: Links from file locations. :param find_links: Links from find_links. :param pages: A dict mapping HTML page url to links from that page. """ self.files = files self.find_links = find_links self.pages = pages class LinkCollector(object): """ Responsible for collecting Link objects from all configured locations, making network requests as needed. The class's main method is its collect_links() method. """ def __init__( self, session, # type: PipSession search_scope, # type: SearchScope ): # type: (...) -> None self.search_scope = search_scope self.session = session @property def find_links(self): # type: () -> List[str] return self.search_scope.find_links def _get_pages(self, locations): # type: (Iterable[Link]) -> Iterable[HTMLPage] """ Yields (page, page_url) from the given locations, skipping locations that have errors. """ for location in locations: page = _get_html_page(location, session=self.session) if page is None: continue yield page def collect_links(self, project_name): # type: (str) -> CollectedLinks """Find all available links for the given project name. :return: All the Link objects (unfiltered), as a CollectedLinks object. """ search_scope = self.search_scope index_locations = search_scope.get_index_urls_locations(project_name) index_file_loc, index_url_loc = group_locations(index_locations) fl_file_loc, fl_url_loc = group_locations( self.find_links, expand_dir=True, ) file_links = [ Link(url) for url in itertools.chain(index_file_loc, fl_file_loc) ] # We trust every directly linked archive in find_links find_link_links = [Link(url, '-f') for url in self.find_links] # We trust every url that the user has given us whether it was given # via --index-url or --find-links. # We want to filter out anything that does not have a secure origin. url_locations = [ link for link in itertools.chain( (Link(url) for url in index_url_loc), (Link(url) for url in fl_url_loc), ) if self.session.is_secure_origin(link) ] url_locations = _remove_duplicate_links(url_locations) lines = [ '{} location(s) to search for versions of {}:'.format( len(url_locations), project_name, ), ] for link in url_locations: lines.append('* {}'.format(link)) logger.debug('\n'.join(lines)) pages_links = {} for page in self._get_pages(url_locations): pages_links[page.url] = list(parse_links(page)) return CollectedLinks( files=file_links, find_links=find_link_links, pages=pages_links, ) site-packages/pip/_internal/locations.py000064400000012446151464452710014414 0ustar00"""Locations where we look for configs, install stuff, etc""" # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import os import os.path import platform import site import sys import sysconfig from distutils import sysconfig as distutils_sysconfig from distutils.command.install import SCHEME_KEYS # type: ignore from pip._internal.utils import appdirs from pip._internal.utils.compat import WINDOWS from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.virtualenv import running_under_virtualenv if MYPY_CHECK_RUNNING: from typing import Any, Union, Dict, List, Optional # Application Directories USER_CACHE_DIR = appdirs.user_cache_dir("pip") def get_major_minor_version(): # type: () -> str """ Return the major-minor version of the current Python as a string, e.g. "3.7" or "3.10". """ return '{}.{}'.format(*sys.version_info) def get_src_prefix(): if running_under_virtualenv(): src_prefix = os.path.join(sys.prefix, 'src') else: # FIXME: keep src in cwd for now (it is not a temporary folder) try: src_prefix = os.path.join(os.getcwd(), 'src') except OSError: # In case the current working directory has been renamed or deleted sys.exit( "The folder you are executing pip from can no longer be found." ) # under macOS + virtualenv sys.prefix is not properly resolved # it is something like /path/to/python/bin/.. return os.path.abspath(src_prefix) # FIXME doesn't account for venv linked to global site-packages site_packages = sysconfig.get_path("purelib") # type: Optional[str] # This is because of a bug in PyPy's sysconfig module, see # https://bitbucket.org/pypy/pypy/issues/2506/sysconfig-returns-incorrect-paths # for more information. if platform.python_implementation().lower() == "pypy": site_packages = distutils_sysconfig.get_python_lib() try: # Use getusersitepackages if this is present, as it ensures that the # value is initialised properly. user_site = site.getusersitepackages() except AttributeError: user_site = site.USER_SITE if WINDOWS: bin_py = os.path.join(sys.prefix, 'Scripts') bin_user = os.path.join(user_site, 'Scripts') # buildout uses 'bin' on Windows too? if not os.path.exists(bin_py): bin_py = os.path.join(sys.prefix, 'bin') bin_user = os.path.join(user_site, 'bin') else: bin_py = os.path.join(sys.prefix, 'bin') bin_user = os.path.join(user_site, 'bin') # Forcing to use /usr/local/bin for standard macOS framework installs # Also log to ~/Library/Logs/ for use with the Console.app log viewer if sys.platform[:6] == 'darwin' and sys.prefix[:16] == '/System/Library/': bin_py = '/usr/local/bin' def distutils_scheme(dist_name, user=False, home=None, root=None, isolated=False, prefix=None): # type:(str, bool, str, str, bool, str) -> dict """ Return a distutils install scheme """ from distutils.dist import Distribution scheme = {} if isolated: extra_dist_args = {"script_args": ["--no-user-cfg"]} else: extra_dist_args = {} dist_args = {'name': dist_name} # type: Dict[str, Union[str, List[str]]] dist_args.update(extra_dist_args) d = Distribution(dist_args) # Ignoring, typeshed issue reported python/typeshed/issues/2567 d.parse_config_files() # NOTE: Ignoring type since mypy can't find attributes on 'Command' i = d.get_command_obj('install', create=True) # type: Any assert i is not None # NOTE: setting user or home has the side-effect of creating the home dir # or user base for installations during finalize_options() # ideally, we'd prefer a scheme class that has no side-effects. assert not (user and prefix), "user={} prefix={}".format(user, prefix) assert not (home and prefix), "home={} prefix={}".format(home, prefix) i.user = user or i.user if user or home: i.prefix = "" i.prefix = prefix or i.prefix i.home = home or i.home i.root = root or i.root i.finalize_options() for key in SCHEME_KEYS: scheme[key] = getattr(i, 'install_' + key) # install_lib specified in setup.cfg should install *everything* # into there (i.e. it takes precedence over both purelib and # platlib). Note, i.install_lib is *always* set after # finalize_options(); we only want to override here if the user # has explicitly requested it hence going back to the config # Ignoring, typeshed issue reported python/typeshed/issues/2567 if 'install_lib' in d.get_option_dict('install'): # type: ignore scheme.update(dict(purelib=i.install_lib, platlib=i.install_lib)) if running_under_virtualenv(): scheme['headers'] = os.path.join( sys.prefix, 'include', 'site', 'python{}'.format(get_major_minor_version()), dist_name, ) if root is not None: path_no_drive = os.path.splitdrive( os.path.abspath(scheme["headers"]))[1] scheme["headers"] = os.path.join( root, path_no_drive[1:], ) return scheme site-packages/pip/_internal/__pycache__/cache.cpython-38.opt-1.pyc000064400000015772151464452710020716 0ustar00U .e @sdZddlZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddl mZddlmZdd lmZmZerdd lmZmZmZmZdd lmZdd lmZeeZGd dde Z!Gddde!Z"Gddde"Z#Gddde!Z$dS)zCache Management N)canonicalize_name)Link) expanduser) TempDirectory)MYPY_CHECK_RUNNING) path_to_url)InvalidWheelFilenameWheel)OptionalSetListAny) FormatControl) Pep425TagcsPeZdZdZfddZddZddZdd Zd d Zd d Z ddZ Z S)CacheaAn abstract class - provides cache directories for data from links :param cache_dir: The root of the cache. :param format_control: An object of FormatControl class to limit binaries being read from the cache. :param allowed_formats: which formats of files the cache should store. ('binary' and 'source' are the only allowed values) cs8tt||rt|nd|_||_||_ddh}dS)Nsourcebinary)superr__init__r cache_dirformat_controlallowed_formats)selfrrrZ_valid_formats __class__7/usr/lib/python3.8/site-packages/pip/_internal/cache.pyr(s zCache.__init__cCs|jg}|jdk r4|jdk r4|d|j|jgd|}t|}|dd|dd|dd|ddg}|S)zEGet parts of part that must be os.path.joined with cache_dir N=#) Zurl_without_fragmentZ hash_namehashappendjoinhashlibZsha224encodeZ hexdigest)rlinkZ key_partsZkey_urlZhashedpartsrrr_get_cache_path_parts2s ,zCache._get_cache_path_partsc Cs|j p| p| }|rgSt|}|j|}|j|s@gS||}z t|WSt k r}z$|j t j t j hkrgWYSW5d}~XYnXdSN) rrrZget_allowed_formatsr intersectionget_path_for_linkoslistdirOSErrorerrnoZENOENTZENOTDIR)rr' package_nameZ can_not_cacheZcanonical_nameZformatsrooterrrrr_get_candidatesLs(    zCache._get_candidatescCs tdS)z>Return a directory to store cached items in for link. NNotImplementedErrorrr'rrrr,eszCache.get_path_for_linkcCs tdS)zaReturns a link to a cached item if it exists, otherwise returns the passed link. Nr5)rr'r1supported_tagsrrrgetks z Cache.getcCs$||}tj||}tt|Sr*)r,r-pathr$rr)rr' candidater2r:rrr_link_for_candidatews zCache._link_for_candidatecCsdSr*rrrrrcleanup~sz Cache.cleanup) __name__ __module__ __qualname____doc__rr)r4r,r9r<r> __classcell__rrrrrs  rcs0eZdZdZfddZddZddZZS)SimpleWheelCachez+A cache of wheels for future installs. cstt|||dhdSNr)rrDrrrrrrrrs  zSimpleWheelCache.__init__cCs ||}tjj|jdf|S)aReturn a directory to store cached wheels for link Because there are M wheels for any one sdist, we provide a directory to cache them in, and then consult that directory when looking up cache hits. We only insert things into the cache if they have plausible version numbers, so that we don't contaminate the cache with things that were not unique. E.g. ./package might have dozens of installs done for it and build a version of 0.0...and if we built and cached a wheel, we'd end up using the same wheel even if the source has been edited. :param link: The link of the sdist for which this will cache wheels. Zwheels)r)r-r:r$r)rr'r(rrrr,s z"SimpleWheelCache.get_path_for_linkc Csxg}|||D]J}z t|}Wntk r8YqYnX||sFq||||fq|sd|S||t|dS)N)r4r rZ supportedr#Zsupport_index_minr<min)rr'r1r8Z candidatesZ wheel_nameZwheelrrrr9s    zSimpleWheelCache.get)r?r@rArBrr,r9rCrrrrrDs rDcs(eZdZdZfddZddZZS)EphemWheelCachezGA SimpleWheelCache that creates it's own temporary cache directory cs&tdd|_tt||jj|dS)Nzephem-wheel-cache)Zkind)r _temp_dirrrIrr:)rrrrrrs   zEphemWheelCache.__init__cCs|jdSr*)rJr>r=rrrr>szEphemWheelCache.cleanup)r?r@rArBrr>rCrrrrrIs rIcs@eZdZdZfddZddZddZdd Zd d ZZ S) WheelCachezWraps EphemWheelCache and SimpleWheelCache into a single Cache This Cache allows for gracefully degradation, using the ephem wheel cache when a certain link is not found in the simple wheel cache first. cs0tt|||dht|||_t||_dSrE)rrKrrD _wheel_cacherI _ephem_cacherFrrrrs  zWheelCache.__init__cCs |j|Sr*)rLr,r7rrrr,szWheelCache.get_path_for_linkcCs |j|Sr*)rMr,r7rrrget_ephem_path_for_linksz"WheelCache.get_ephem_path_for_linkcCs0|jj|||d}||k r|S|jj|||dS)N)r'r1r8)rLr9rM)rr'r1r8Zretvalrrrr9szWheelCache.getcCs|j|jdSr*)rLr>rMr=rrrr>s zWheelCache.cleanup) r?r@rArBrr,rNr9r>rCrrrrrKs  rK)%rBr0r%Zloggingr-Zpip._vendor.packaging.utilsrZpip._internal.models.linkrZpip._internal.utils.compatrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.urlsrZpip._internal.wheelrr typingr r r r Zpip._internal.indexrZpip._internal.pep425tagsrZ getLoggerr?ZloggerobjectrrDrIrKrrrrs(         f:site-packages/pip/_internal/__pycache__/index.cpython-38.pyc000064400000061347151464452710020022 0ustar00U .e@s(dZddlmZddlZddlZddlmZddlmZddl m Z ddl m Z mZmZmZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZddlmZddlm Z ddl!m"Z"ddl#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*e$rddl+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1m2Z2m3Z3ddl m4Z4ddl5m6Z6ddl7m8Z8ddl9m:Z:ddl;mZ>e3e2de2e?e@ffZAe2e?e?e?e4eAe/e?fZBdddgZCeDeEZFd1d!d"ZGGd#d$d$eHZId%d&ZJGd'd(d(eHZKGd)ddeHZLGd*d+d+eHZMGd,ddeHZNd-d.ZOd/d0ZPdS)2z!Routines related to PyPI, indexes)absolute_importN) specifiers)canonicalize_name)parse)BestVersionAlreadyInstalledDistributionNotFoundInvalidWheelFilenameUnsupportedWheel)InstallationCandidate) FormatControl)Link)SelectionPreferences) TargetPython)WHEEL_EXTENSION) indent_log) build_netloc)check_requires_python)MYPY_CHECK_RUNNING)SUPPORTED_EXTENSIONS) url_to_path)Wheel) FrozenSetIterableListOptionalSetTextTupleUnion) _BaseVersion) LinkCollector) SearchScope)InstallRequirement) Pep425Tag)Hashesr BestCandidateResult PackageFinderFcCs~zt|j|d}Wn&tjk r8td|j|YnBX|szdtt|}|shtd||j|dStd||j|dS)aa Return whether the given Python version is compatible with a link's "Requires-Python" value. :param version_info: A 3-tuple of ints representing the Python major-minor-micro version to check. :param ignore_requires_python: Whether to ignore the "Requires-Python" value if the given Python version isn't compatible. ) version_infoz2Ignoring invalid Requires-Python (%r) for link: %s.z4Link requires a different Python (%s not in: %r): %sFzBIgnoring failed Requires-Python check (%s not in: %r) for link: %sT) rZrequires_pythonrZInvalidSpecifierloggerdebugjoinmapstr)linkr(ignore_requires_pythonZ is_compatibleversionr%r%7/usr/lib/python3.8/site-packages/pip/_internal/index.py_check_link_requires_python;s8  r3c@s,eZdZdZedZdddZddZdS) LinkEvaluatorzD Responsible for evaluating links for a particular project. z-py([123]\.?[0-9]?)$NcCs4|dkr d}||_||_||_||_||_||_dS)a :param project_name: The user supplied package name. :param canonical_name: The canonical package name. :param formats: The formats allowed for this package. Should be a set with 'binary' or 'source' or both in it. :param target_python: The target Python interpreter to use when evaluating link compatibility. This is used, for example, to check wheel compatibility, as well as when checking the Python version, e.g. the Python version embedded in a link filename (or egg fragment) and against an HTML link's optional PEP 503 "data-requires-python" attribute. :param allow_yanked: Whether files marked as yanked (in the sense of PEP 592) are permitted to be candidates for install. :param ignore_requires_python: Whether to ignore incompatible PEP 503 "data-requires-python" values in HTML links. Defaults to False. NF) _allow_yanked_canonical_name_ignore_requires_python_formats_target_python project_name)selfr:canonical_nameformats target_python allow_yankedr0r%r%r2__init__rszLinkEvaluator.__init__c Csd}|jr(|js(|jpd}dd|fS|jr<|j}|j}n|\}}|sPdS|tkrddd|fSd|jkr|t krd|j }d|fSd |j kr|d krd S|t kr,zt |j }Wntk rYd SXt|j|jkrd |j }d|fS|j}||s&|}dd|}d|fS|j}d|jkrP|t krPdd|j fS|sbt||j}|svdd|j fS|j|} | r|d| }| d} | |jjkrdSt||jj|j d} | sdSt!"d||d|fS)aG Determine whether a link is a candidate for installation. :return: A tuple (is_candidate, result), where `result` is (1) a version string if `is_candidate` is True, and (2) if `is_candidate` is False, an optional string to log the reason the link fails to qualify. N Fzyanked for reason: {})Fz not a filezunsupported archive format: %sZbinaryzNo binaries permitted for %sZmacosx10z.zip)Fz macosx10 one)Fzinvalid wheel filenamezwrong project name (not %s)z"none of the wheel's tags match: {}, sourcezNo sources permitted for %szMissing project version for %s)FzPython version is incorrect)r(r0)FNzFound link %s, version: %sT)# is_yankedr5 yanked_reasonformat egg_fragmentextsplitextrr8rr:pathrfilenamerrnamer6r9get_tags supportedZget_formatted_file_tagsr,r1_extract_version_from_fragment_py_version_researchstartgroup py_versionr3Zpy_version_infor7r*r+) r;r/r1reasonZegg_inforIwheelsupported_tagsZ file_tagsmatchrUZsupports_pythonr%r%r2 evaluate_linksx            zLinkEvaluator.evaluate_link)N) __name__ __module__ __qualname____doc__recompilerQr@rZr%r%r%r2r4fs  &r4c Cs|stdt||t|Sg}g}d}|D]>}|j}|js@n"|j|drV|d7}n ||q.||q.|rx|}nt|}t|t|krd} n dt|d dd |D} td t|||j |t||| |S) a Filter out candidates whose hashes aren't allowed, and return a new list of candidates. If at least one candidate has an allowed hash, then all candidates with either an allowed hash or no hash specified are returned. Otherwise, the given candidates are returned. Including the candidates with no hash specified when there is a match allows a warning to be logged if there is a more preferred candidate with no hash specified. Returning all candidates in the case of no matches lets pip report the hash of the candidate that would otherwise have been installed (e.g. permitting the user to more easily update their requirements file with the desired hash). zJGiven no hashes to check %s links for project %r: discarding no candidatesr)hashesrDzdiscarding no candidateszdiscarding {} non-matches: {}z css|]}t|jVqdSN)r.r/.0 candidater%r%r2 +sz*filter_unallowed_hashes..zPChecked %s links for project %r against %s hashes (%s matches, %s no digest): %s) r*r+lenlistr/Zhas_hashis_hash_allowedappendrGr,Z digest_count) candidatesrar:Zmatches_or_no_digestZ non_matchesZ match_countrer/ZfilteredZdiscard_messager%r%r2filter_unallowed_hashessL      rlc@seZdZdZdddZdS)CandidatePreferenceszk Encapsulates some of the preferences for filtering and sorting InstallationCandidate objects. FcCs||_||_dS)zR :param allow_all_prereleases: Whether to allow all pre-releases. N)allow_all_prereleases prefer_binary)r;rornr%r%r2r@Cs zCandidatePreferences.__init__N)FF)r[r\r]r^r@r%r%r%r2rm<srmc@s(eZdZdZddZddZddZdS) r&zA collection of candidates, returned by `PackageFinder.find_best_candidate`. This class is only intended to be instantiated by CandidateEvaluator's `compute_best_candidate()` method. cCsHt|t|kst|dkr&|r2tn ||ks2t||_||_||_dS)a :param candidates: A sequence of all available candidates found. :param applicable_candidates: The applicable candidates. :param best_candidate: The most preferred candidate found, or None if no applicable candidates were found. N)setAssertionError_applicable_candidates _candidatesbest_candidater;rkapplicable_candidatesrtr%r%r2r@Ws   zBestCandidateResult.__init__cCs t|jS)z(Iterate through all candidates. )iterrsr;r%r%r2iter_allpszBestCandidateResult.iter_allcCs t|jS)z3Iterate through the applicable candidates. )rwrrrxr%r%r2iter_applicablevsz#BestCandidateResult.iter_applicableN)r[r\r]r^r@ryrzr%r%r%r2r&Psc@sHeZdZdZedddZdddZdd Zd d Zd d Z ddZ dS)CandidateEvaluatorzm Responsible for filtering and sorting candidates for installation based on what tags are valid. NFcCs:|dkrt}|dkrt}|}|||||||dS)aCreate a CandidateEvaluator object. :param target_python: The target Python interpreter to use when checking compatibility. If None (the default), a TargetPython object will be constructed from the running Python. :param specifier: An optional object implementing `filter` (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable versions. :param hashes: An optional collection of allowed hashes. N)r:rX specifierrornra)rrZ SpecifierSetrN)clsr:r>rornr|rarXr%r%r2createszCandidateEvaluator.createcCs(||_||_||_||_||_||_dS)z :param supported_tags: The PEP 425 tags supported by the target Python in order of preference (most preferred first). N)_allow_all_prereleases_hashes_prefer_binary _project_name _specifier_supported_tags)r;r:rXr|rornrar%r%r2r@s zCandidateEvaluator.__init__csV|jpd}|j}dd|jdd|D|dDfdd|D}t||j|jd S) zM Return the applicable candidates from a list of candidates. NcSsh|] }t|qSr%)r.)rdvr%r%r2 sz?CandidateEvaluator.get_applicable_candidates..css|]}t|jVqdSrbr.r1rdcr%r%r2rfsz?CandidateEvaluator.get_applicable_candidates..)Z prereleasescsg|]}t|jkr|qSr%rrZversionsr%r2 sz@CandidateEvaluator.get_applicable_candidates..)rkrar:)rrfilterrlrr)r;rkZallow_prereleasesr|rvr%rr2get_applicable_candidatess   z,CandidateEvaluator.get_applicable_candidatesc Cs|j}t|}d}d}|j}|jrt|j}||sDtd|j|jrNd}| | }|j dk rt d|j } | } t| d| df}n| }t||j} dt|j} | | ||j||fS)a) Function to pass as the `key` argument to a call to sorted() to sort InstallationCandidates by preference. Returns a tuple such that tuples sorting as greater using Python's default comparison operator are more preferred. The preference is as follows: First and foremost, candidates with allowed (matching) hashes are always preferred over candidates without matching hashes. This is because e.g. if the only candidate with an allowed hash is yanked, we still want to use that candidate. Second, excepting hash considerations, candidates that have been yanked (in the sense of PEP 592) are always less preferred than candidates that haven't been yanked. Then: If not finding wheels, they are sorted by version only. If finding wheels, then the sort order is by version, then: 1. existing installs 2. wheels ordered via Wheel.support_index_min(self._supported_tags) 3. source archives If prefer_binary was set, then all wheels are sorted above sources. Note: it was considered to embed this logic into the Link comparison operators, but then different sdist links with the same version, would have to be considered equal r%rzB%s is not a supported wheel for this platform. It can't be sorted.rDNz ^(\d+)(.*)$)rrgr/Zis_wheelrrLrOr rZsupport_index_min build_tagr_rYgroupsintrirrEr1) r;reZ valid_tagsZ support_numrZbinary_preferencer/rWZprirYZbuild_tag_groupsZhas_allowed_hashZ yank_valuer%r%r2 _sort_keys<    zCandidateEvaluator._sort_keycCsH|sdSt||jd}|j}|jrD|jp*d}dj||d}t||S)zy Return the best candidate per the instance's sort order, or None if no candidate is acceptable. NkeyrAzqThe candidate selected for download or install is a yanked version: {candidate} Reason for being yanked: {reason})rerV)maxrr/rErFrGr*Zwarning)r;rkrtr/rVmsgr%r%r2sort_best_candidates   z&CandidateEvaluator.sort_best_candidatecCs"||}||}t|||dS)zF Compute and return a `BestCandidateResult` instance. )rvrt)rrr&rur%r%r2compute_best_candidate<s  z)CandidateEvaluator.compute_best_candidate)NFFNN)FFN) r[r\r]r^ classmethodr~r@rrrrr%r%r%r2r{}s  ) $<r{c@seZdZdZd&ddZed'ddZeddZej d dZed d Z ed d Z eddZ eddZ ddZddZddZddZddZddZddZd(d d!Zd)d"d#Zd$d%ZdS)*r'zThis finds packages. This is meant to match easy_install's technique for looking for packages, by reading pages and looking for appropriate links. NcCsP|dkrt}|pttt}||_||_||_||_||_||_t|_ dS)a This constructor is primarily meant to be used by the create() class method and from tests. :param format_control: A FormatControl object, used to control the selection of source packages / binary packages when consulting the index and links. :param candidate_prefs: Options to use when creating a CandidateEvaluator object. N) rmr rpr5_candidate_prefsr7_link_collectorr9format_control _logged_links)r;link_collectorr>r?rcandidate_prefsr0r%r%r2r@VszPackageFinder.__init__cCs8|dkrt}t|j|jd}|||||j|j|jdS)afCreate a PackageFinder. :param selection_prefs: The candidate selection preferences, as a SelectionPreferences object. :param target_python: The target Python interpreter to use when checking compatibility. If None (the default), a TargetPython object will be constructed from the running Python. N)rorn)rrr>r?rr0)rrmrornr?rr0)r}rZselection_prefsr>rr%r%r2r~~szPackageFinder.createcCs|jjSrbr search_scoperxr%r%r2rszPackageFinder.search_scopecCs ||j_dSrbr)r;rr%r%r2rscCs|jjSrb)r find_linksrxr%r%r2rszPackageFinder.find_linkscCs|jjSrb)r index_urlsrxr%r%r2rszPackageFinder.index_urlsccs|jjjD]}t|Vq dSrb)rZsessionZpip_trusted_originsr)r;Z host_portr%r%r2 trusted_hostsszPackageFinder.trusted_hostscCs|jjSrbrrnrxr%r%r2rnsz#PackageFinder.allow_all_prereleasescCs d|j_dS)NTrrxr%r%r2set_allow_all_prereleasessz'PackageFinder.set_allow_all_prereleasescCs.t|}|j|}t||||j|j|jdS)N)r:r<r=r>r?r0)rrZget_allowed_formatsr4r9r5r7)r;r:r<r=r%r%r2make_link_evaluators z!PackageFinder.make_link_evaluatorcCsPgg}}t}|D]2}||kr|||jr<||q||q||S)z Returns elements of links in order, non-egg links first, egg links second, while eliminating duplicates )rpaddrHrj)r;linksZeggsZno_eggsseenr/r%r%r2 _sort_linkss    zPackageFinder._sort_linkscCs(||jkr$td|||j|dS)NzSkipping link: %s: %s)rr*r+r)r;r/rVr%r%r2_log_skipped_links zPackageFinder._log_skipped_linkcCs<||\}}|s(|r$|j||ddSt|j|t|dS)z If the link is a candidate for install, convert it to an InstallationCandidate and return it. Otherwise, return None. )rVN)Zprojectr/r1)rZrr r:r.)r;link_evaluatorr/Z is_candidateresultr%r%r2get_install_candidatesz#PackageFinder.get_install_candidatecCs6g}||D]"}|||}|dk r||q|S)zU Convert links that are candidates to InstallationCandidate objects. N)rrrj)r;rrrkr/rer%r%r2evaluate_linkss   zPackageFinder.evaluate_linksc Cs|j|}||}|j||jd}g}|jD]>\}}td|t |j||d}| |W5QRXq4|j||j d} | r| j ddtdd dd| D| ||S) aFind all available InstallationCandidate for project_name This checks index_urls and find_links. All versions found are returned as an InstallationCandidate list. See LinkEvaluator.evaluate_link() for details on which files are accepted. )rzAnalyzing links from page %sT)reversezLocal files found: %srBcSsg|]}t|jjqSr%)rr/Zurlrcr%r%r2r2sz5PackageFinder.find_all_candidates..)rZ collect_linksrrrZpagesitemsr*r+rextendfilessortr,) r;r:Zcollected_linksrZfind_links_versionsZ page_versionsZpage_urlZ page_linksZ new_versionsZ file_versionsr%r%r2find_all_candidates s8      z!PackageFinder.find_all_candidatescCs"|j}tj||j|j|j||dS)z3Create a CandidateEvaluator object to use. )r:r>rornr|ra)rr{r~r9rorn)r;r:r|rarr%r%r2make_candidate_evaluator;s z&PackageFinder.make_candidate_evaluatorcCs$||}|j|||d}||S)aFind matches for the given project and specifier. :param specifier: An optional object implementing `filter` (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable versions. :return: A `BestCandidateResult` instance. )r:r|ra)rrr)r;r:r|rarkZcandidate_evaluatorr%r%r2find_best_candidateNs z!PackageFinder.find_best_candidatec Cs|jdd}|j|j|j|d}|j}d}|jdk r@t|jj}dd}|dkrz|dkrzt d||| t d|d}|r|dks|j|krd }|s|dk r|rt d |nt d ||jdS|rt d ||| tt d |j|| |jS)zTry to find a Link matching req Expects req, an InstallRequirement and upgrade, a boolean Returns a Link if found, Raises DistributionNotFound or BestVersionAlreadyInstalled otherwise F)Ztrust_internet)r|raNcSs dtdd|DtdpdS)NrBcSsh|]}t|jqSr%rrr%r%r2r}szKPackageFinder.find_requirement.._format_versions..rZnone)r,sorted parse_version)Z cand_iterr%r%r2_format_versionsws  z8PackageFinder.find_requirement.._format_versionszNCould not find a version that satisfies the requirement %s (from versions: %s)z%No matching distribution found for %sTzLExisting installed version (%s) is most up-to-date and satisfies requirementzUExisting installed version (%s) satisfies requirement (most up-to-date version is %s)z=Installed version (%s) is most up-to-date (past versions: %s)z)Using version %s (newest of versions: %s))rarrMr|rtZ satisfied_byrr1r*Zcriticalryrr+rzrr/) r;ZreqZupgraderaZbest_candidate_resultrtZinstalled_versionrZbest_installedr%r%r2find_requirementesh        zPackageFinder.find_requirement)NNN)N)NN)NN)r[r\r]r^r@rr~propertyrsetterrrrrnrrrrrrrrrrr%r%r%r2r'OsD  (         1  cCsLt|D].\}}|dkrqt|d||kr|Sqtd||dS)aFind the separator's index based on the package's canonical name. :param fragment: A + filename "fragment" (stem) or egg fragment. :param canonical_name: The package's canonical name. This function is needed since the canonicalized name does not necessarily have the same length as the egg info's name part. An example:: >>> fragment = 'foo__bar-1.0' >>> canonical_name = 'foo-bar' >>> _find_name_version_sep(fragment, canonical_name) 8 -Nz{} does not match {}) enumerater ValueErrorrG)fragmentr<irr%r%r2_find_name_version_seps  rcCsBzt||d}Wntk r(YdSX||d}|s>dS|S)zParse the version string from a + filename "fragment" (stem) or egg fragment. :param fragment: The string to parse. E.g. foo-2.1 :param canonical_name: The canonicalized name of the package this belongs to. rDN)rr)rr<Z version_startr1r%r%r2rPs  rP)F)Qr^Z __future__rZloggingr_Zpip._vendor.packagingrZpip._vendor.packaging.utilsrZpip._vendor.packaging.versionrrZpip._internal.exceptionsrrrr Zpip._internal.models.candidater Z#pip._internal.models.format_controlr Zpip._internal.models.linkr Z$pip._internal.models.selection_prefsr Z"pip._internal.models.target_pythonrZpip._internal.utils.filetypesrZpip._internal.utils.loggingrZpip._internal.utils.miscrZpip._internal.utils.packagingrZpip._internal.utils.typingrZpip._internal.utils.unpackingrZpip._internal.utils.urlsrZpip._internal.wheelrtypingrrrrrrrrrZpip._internal.collectorr Z!pip._internal.models.search_scoper!Zpip._internal.reqr"Zpip._internal.pep425tagsr#Zpip._internal.utils.hashesr$rr.ZBuildTagZCandidateSortingKey__all__Z getLoggerr[r*r3objectr4rlrmr&r{r'rrPr%r%r%r2s^                 (         + K-Shsite-packages/pip/_internal/__pycache__/self_outdated_check.cpython-38.pyc000064400000012622151464452710022662 0ustar00U .e@s^ddlmZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z ddlmZddlmZddlmZddlmZdd lmZdd lmZdd lmZmZmZdd lmZm Z m!Z!dd l"m#Z#ddl$m%Z%e%rddl&Z&ddl&m'Z'ddl(m)Z)m*Z*m+Z+m,Z,ddl-m.Z.dZ/e0e1Z2dddZ3ddZ4Gddde5Z6ddZ7ddZ8dS))absolute_importN) pkg_resources)version) ensure_binary) LinkCollector) PackageFinder) SearchScope)SelectionPreferences)WINDOWS)adjacent_tmp_filecheck_path_ownerreplace) ensure_dirget_installed_versionredact_auth_from_url) get_installer)MYPY_CHECK_RUNNING)Values)AnyDictTextUnion) PipSessionz%Y-%m-%dT%H:%M:%SZFcCs`|jg|j}|jr8|s8tdddd|Dg}|jp@g}tj||d}t ||d}|S)z :param session: The Session to use to make requests. :param suppress_no_index: Whether to ignore the --no-index option when constructing the SearchScope object. zIgnoring indexes: %s,css|]}t|VqdSN)r).0ZurlrE/usr/lib/python3.8/site-packages/pip/_internal/self_outdated_check.py @sz&make_link_collector..) find_links index_urls)session search_scope) Z index_urlZextra_index_urlsZno_indexloggerdebugjoinrrcreater)r!optionssuppress_no_indexr rr"link_collectorrrrmake_link_collector1s    r*cCst|}t|}|Sr)rhashlibZsha224Z hexdigest)keyZ key_bytesnamerrr_get_statefile_namePsr.c@s(eZdZddZeddZddZdS)SelfCheckStatec Csni|_d|_|rjtj|dt|j|_z&t|j}t ||_W5QRXWnt t t fk rhYnXdS)NZ selfcheck) statestatefile_pathospathr%r.r,openjsonloadIOError ValueErrorKeyError)self cache_dirZ statefilerrr__init__Xs zSelfCheckState.__init__cCstjSr)sysprefix)r:rrrr,jszSelfCheckState.keyc Cs|js dSttj|js dSttj|j|j|t|d}t j |ddd}t |j}| t |W5QRXzt|j|jWntk rYnXdS)N)r, last_check pypi_versionT)r:)Z sort_keysZ separators)r1r r2r3dirnamerr,strftimeSELFCHECK_DATE_FMTr5dumpsr writerr r-OSError)r:r@ current_timer0textfrrrsavens  zSelfCheckState.saveN)__name__ __module__ __qualname__r<propertyr,rKrrrrr/Ws r/cCs6zt|}dt|kWStjk r0YdSXdS)zChecks whether pkg was installed by pip This is used not to display the upgrade message when pip is in fact installed by system package manager, such as dnf on Fedora. pipFN)rZget_distributionrZDistributionNotFound)ZpkgZdistrrrwas_installed_by_pips  rQcCsXtd}|sdSt|}d}zt|jd}tj}d|jkrzd|jkrztj|jdt }|| dkrz|jd}|dkrt ||dd}t d d d } t j|| d } | dj} | dkrWdSt| j}|||t|} || ko|j| jkotd} | s WdStrd }nd}td |||Wn$tk rRtjdddYnXdS)zCheck for an update for pip. Limit the frequency of checks to once per week. State is stored either in the active virtualenv or in the user's USER_CACHE_DIR keyed off the prefix of the pip script path. rPN)r;r?r@i: T)r'r(F)Z allow_yankedZallow_all_prereleases)r)selection_prefsz python -m pipzYou are using pip version %s; however, version %s is available. You should consider upgrading via the '%s install --upgrade pip' command.z5There was an error checking the latest version of pip)exc_info)rpackaging_versionparser/r;datetimeZutcnowr0strptimerDZ total_secondsr*r rr&Zfind_best_candidatebest_candidatestrrrKZ base_versionrQr r#Zwarning Exceptionr$)r!r'Zinstalled_versionZ pip_versionr@r0rHr?r)rRfinderrXZremote_versionZlocal_version_is_olderZpip_cmdrrrpip_self_version_checksp         r\)F)9Z __future__rrVr+r5ZloggingZos.pathr2r=Z pip._vendorrZpip._vendor.packagingrrTZpip._vendor.sixrZpip._internal.collectorrZpip._internal.indexrZ!pip._internal.models.search_scoperZ$pip._internal.models.selection_prefsr Zpip._internal.utils.compatr Zpip._internal.utils.filesystemr r r Zpip._internal.utils.miscrrrZpip._internal.utils.packagingrZpip._internal.utils.typingrZoptparsertypingrrrrZpip._internal.network.sessionrrDZ getLoggerrLr#r*r.objectr/rQr\rrrrs>               ;site-packages/pip/_internal/__pycache__/download.cpython-38.opt-1.pyc000064400000026231151464452710021452 0ustar00U .eEQ@sddlmZddlZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z m Z ddlmZddlmZddlmZmZddlmZdd lmZdd lmZdd lmZdd lmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'dd l(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/ddl0m1Z1ddl2m3Z3e+rddl4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:ddl;mZ>ddl?m@Z@ddlAmBZBered0d ZSd1d2ZTd3d4ZUd?d5dZVd@d6d!ZWd7d#ZXd8d"ZYd9d:ZZd;d<Z[dS)A)absolute_importN)requests)CONTENT_CHUNK_SIZEResponse)PY2)parse) HashMismatchInstallationError)PyPI) PipSession) auto_decode) copy2_fixed) ask_path_exists backup_dirconsume display_path format_sizehide_urlpath_to_displayrmtreesplitext) TempDirectory)MYPY_CHECK_RUNNING)DownloadProgressProvider) unpack_file)get_url_scheme)vcs)IOCallableListOptionalTextTuple) TypedDict)Link)Hashes)VersionControlCopytreeKwargsignoresymlinksF)Ztotal) copy_functionr)Zignore_dangling_symlinksr*get_file_contentunpack_vcs_linkunpack_file_urlunpack_http_url unpack_urlparse_content_dispositionsanitize_content_filenamec Cs6|dkrtdt|}|dkr>||}||j|jfS|dkr|rd|drdtd||f|ddd}| d d }t |}|r| dd|d dd}t |}|d rd |d }|}z&t|d }t|}W5QRXWn4tk r,} ztd t| W5d} ~ XYnX||fS)a*Gets the content of a file; it may be a filename, file: URL, or http: URL. Returns (location, content). Content is unicode. :param url: File path or url. :param comes_from: Origin description of requirements. :param session: Instance of pip.download.PipSession. NzAget_file_content() missing 1 required keyword argument: 'session')httpZhttpsfiler3z6Requirements file %s references URL %s, which is local:\/|rbz$Could not open requirements file: %s) TypeErrorrgetraise_for_statusurltext startswithr splitreplace_url_slash_drive_rematchgroup urllib_parseZunquotelstripopenr readIOErrorstr) r>Z comes_fromsessionschemeresppathrDfZcontentexcrR:/usr/lib/python3.8/site-packages/pip/_internal/download.pyr,UsB         z /*([a-z])\|cCs t|}|j|t|jddS)N)r>)_get_used_vcs_backendunpackrr>)linklocation vcs_backendrRrRrSr-scCs$tjD]}|j|jkr|SqdS)z1 Return a VersionControl object or None. N)rZbackendsrMZschemes)rVrXrRrRrSrTs   rTcOs|SNrR)iterableargskwargsrRrRrS_progress_indicatorsr]c s6ztjd}Wntttfk r0d}YnXtdd}ttj krRd}n&|r\d}n|dkrjd}n|std}nd}|j }fdd} fd d } t } |j t j kr|} n|j} |rt||d } |rtd | t|n td | n|rtd| n td | | | | tt} |r*|| nt| dS)Nzcontent-lengthrZ from_cacheFi@Tc3sTz jj|ddD] }|VqWn.tk rNj|}|sBqJ|Vq0YnXdS)NF)Zdecode_content)rawstreamAttributeErrorrI)Z chunk_sizechunk)rNrRrS resp_reads   z _download_url..resp_readc3s|D]}||VqdSrY)write)Zchunksra) content_filerRrSwritten_chunkss z%_download_url..written_chunks)maxzDownloading %s (%s)zDownloading %szUsing cached %s)intheaders ValueErrorKeyErrorr;getattrloggerZgetEffectiveLevelloggingINFOshow_urlr]Znetlocr Zurl_without_fragmentrinforrZcheck_against_chunksr)rNrVrdhashes progress_barZ total_lengthZ cached_respZ show_progressrorbreZprogress_indicatorr>Zdownloaded_chunksrR)rdrNrS _download_urlsP   %    rscCsd}tj||j}tj|rtdt|d}|dkr@d}nj|dkrdtdt|t |nF|dkrt |}td t|t|t ||n|d krt d |rt ||td t|dS) NTz8The file %s exists. (i)gnore, (w)ipe, (b)ackup, (a)abort)iwbartFruz Deleting %srvzBacking up %s to %srwzSaved %s)osrOjoinfilenameexistsrrrlwarningremovershutilZmovesysexitcopyrp)r{rWrVrZdownload_locationresponseZ dest_filerRrRrS _copy_files6    ronc Cs|dkrtdtddv}d}|r0t|||}|rH|}t|d} nt|||j||\}} t||| |r~|s~t||||st |W5QRXdS)Nz@unpack_http_url() missing 1 required keyword argument: 'session'rU)Zkindr) r;r_check_download_dir mimetypes guess_type_download_http_urlrOrrryunlink) rVrW download_dirrLrqrrtemp_diralready_downloaded_path from_path content_typerRrRrSr/s2    c CsTzt||Wn@tjk rN}z tdt|t|t|W5d}~XYnXdS)zCopying special files is not supported, but as a convenience to users we skip errors copying them. This supports tools that may create e.g. socket files in the project source directory. z>Ignoring special file error '%s' encountered copying %s to %s.N)r rZSpecialFileErrorrlr}rKr)srcdesterRrRrS_copy2_ignoring_special_filesJsrcs`tj|}tj|tj|fdd}t|dd}tsLt|d<tj |f|dS)Ncs6g}|kr|ddg7}tj|kr2|g7}|S)Nz.toxz.nox)ryrOabspath)dnamesZskippedsourceZtarget_basenameZtarget_dirnamerRrSr)es   z!_copy_source_tree..ignoreTr(r+) ryrOrbasenamedirnamedictrrrZcopytree)rtargetZtarget_abspathr)r\rRrrS_copy_source_tree_s    rcCs|j}|r>tj|r"t|t|||r:tddS|rL| |d}|r`t |||}|rj|}n|}t |d}t ||||r|st|||dS)zUnpack link into location. If download_dir is provided and link points to a file, make a copy of the link file inside download_dir. z*Link is a directory, ignoring download_dirNr) file_pathZis_existing_dirryrOisdirrrrlrpcheck_against_pathrrrrr)rVrWrrqZ link_pathrrrrRrRrSr.~s.      cCsP|jrt||n:|jr*t||||dn"|dkr8t}t||||||ddS)avUnpack link. If link is a VCS link: if only_download, export into download_dir and ignore location else unpack into location for other types of link: - unpack into location - if download_dir, copy the file into download_dir - if only_download, mark location for deletion :param hashes: A Hashes object, one of whose embedded hashes must match, or HashMismatch will be raised. If the Hashes is empty, no matches are required, and unhashable types of requirements (like VCS ones, which would ordinarily raise HashUnsupported) are allowed. )rqN)rqrr)Zis_vcsr-is_filer.r r/)rVrWrrLrqrrrRrRrSr0s cCs tj|S)zJ Sanitize the "filename" value from a Content-Disposition header. )ryrOr)r{rRrRrSr2scCs,t|\}}|d}|r$t|}|p*|S)z Parse the "filename" value from a Content-Disposition header, and return the default filename if the result is empty. r{)cgiZ parse_headerr<r2)content_dispositionZdefault_filenameZ_typeZparamsr{rRrRrSr1s  c Cs*|jddd}z |j|ddidd}|Wn8tjk rj}ztd|jj |W5d }~XYnX|j d d }|j } |j d } | rt | | } t | d} | st|} | r| | 7} | s|j|jkrtj |jd} | r| | 7} tj|| } t| d } t||| ||W5QRX| |fS)z6Download link url into temp_dir using provided session#r6rzAccept-EncodingZidentityT)rhr_zHTTP error %s while getting %sNz content-typezcontent-dispositionwb)r>rAr<r=rZ HTTPErrorrlZcriticalrZ status_coderhr{r1rrZguess_extensionryrOrzrHrs)rVrLrrqrrZ target_urlrNrQrr{rZextrrdrRrRrSrsB       rcCsptj||j}tj|s dStd||rlz||Wn,tk rjt d|t |YdSX|S)z Check download_dir for previously downloaded file with correct hash If a correct file is found return its path else None NzFile was already downloaded %sz;Previously-downloaded file %s has bad hash. Re-downloading.) ryrOrzr{r|rlrprrr}r)rVrrqZ download_pathrRrRrSr7s   r)NN)NNNr)NN)NNNr)\Z __future__rrrmrryrerrZ pip._vendorrZpip._vendor.requests.modelsrrZpip._vendor.sixrZpip._vendor.six.moves.urllibrrFZpip._internal.exceptionsrr Zpip._internal.models.indexr Zpip._internal.network.sessionr Zpip._internal.utils.encodingr Zpip._internal.utils.filesystemr Zpip._internal.utils.miscrrrrrrrrrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.uirZpip._internal.utils.unpackingrZpip._internal.utils.urlsrZpip._internal.vcsrtypingrrrr r!r"Zmypy_extensionsr#Zpip._internal.models.linkr$Zpip._internal.utils.hashesr%Z pip._internal.vcs.versioncontrolr&rKboolr'__all__Z getLogger__name__rlr,compileIrCr-rTr]rsrr/rrr.r0r2r1rrrRrRrRrSs        ,               0 d -" 8 .?site-packages/pip/_internal/__pycache__/download.cpython-38.pyc000064400000026267151464452710020524 0ustar00U .eEQ@sddlmZddlZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z m Z ddlmZddlmZddlmZmZddlmZdd lmZdd lmZdd lmZdd lmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'dd l(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/ddl0m1Z1ddl2m3Z3e+rddl4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:ddl;mZ>ddl?m@Z@ddlAmBZBered0d ZSd1d2ZTd3d4ZUd?d5dZVd@d6d!ZWd7d#ZXd8d"ZYd9d:ZZd;d<Z[dS)A)absolute_importN)requests)CONTENT_CHUNK_SIZEResponse)PY2)parse) HashMismatchInstallationError)PyPI) PipSession) auto_decode) copy2_fixed) ask_path_exists backup_dirconsume display_path format_sizehide_urlpath_to_displayrmtreesplitext) TempDirectory)MYPY_CHECK_RUNNING)DownloadProgressProvider) unpack_file)get_url_scheme)vcs)IOCallableListOptionalTextTuple) TypedDict)Link)Hashes)VersionControlCopytreeKwargsignoresymlinksF)Ztotal) copy_functionr)Zignore_dangling_symlinksr*get_file_contentunpack_vcs_linkunpack_file_urlunpack_http_url unpack_urlparse_content_dispositionsanitize_content_filenamec Cs6|dkrtdt|}|dkr>||}||j|jfS|dkr|rd|drdtd||f|ddd}| d d }t |}|r| dd|d dd}t |}|d rd |d }|}z&t|d }t|}W5QRXWn4tk r,} ztd t| W5d} ~ XYnX||fS)a*Gets the content of a file; it may be a filename, file: URL, or http: URL. Returns (location, content). Content is unicode. :param url: File path or url. :param comes_from: Origin description of requirements. :param session: Instance of pip.download.PipSession. NzAget_file_content() missing 1 required keyword argument: 'session')httpZhttpsfiler3z6Requirements file %s references URL %s, which is local:\/|rbz$Could not open requirements file: %s) TypeErrorrgetraise_for_statusurltext startswithr splitreplace_url_slash_drive_rematchgroup urllib_parseZunquotelstripopenr readIOErrorstr) r>Z comes_fromsessionschemeresppathrDfZcontentexcrR:/usr/lib/python3.8/site-packages/pip/_internal/download.pyr,UsB         z /*([a-z])\|cCs,t|}|dk st|j|t|jddS)N)r>)_get_used_vcs_backendAssertionErrorunpackrr>)linklocation vcs_backendrRrRrSr-s cCs$tjD]}|j|jkr|SqdS)z1 Return a VersionControl object or None. N)rZbackendsrMZschemes)rWrYrRrRrSrTs   rTcOs|SNrR)iterableargskwargsrRrRrS_progress_indicatorsr^c s6ztjd}Wntttfk r0d}YnXtdd}ttj krRd}n&|r\d}n|dkrjd}n|std}nd}|j }fdd} fd d } t } |j t j kr|} n|j} |rt||d } |rtd | t|n td | n|rtd| n td | | | | tt} |r*|| nt| dS)Nzcontent-lengthrZ from_cacheFi@Tc3sTz jj|ddD] }|VqWn.tk rNj|}|sBqJ|Vq0YnXdS)NF)Zdecode_content)rawstreamAttributeErrorrI)Z chunk_sizechunk)rNrRrS resp_reads   z _download_url..resp_readc3s|D]}||VqdSrZ)write)Zchunksrb) content_filerRrSwritten_chunkss z%_download_url..written_chunks)maxzDownloading %s (%s)zDownloading %szUsing cached %s)intheaders ValueErrorKeyErrorr;getattrloggerZgetEffectiveLevelloggingINFOshow_urlr^Znetlocr Zurl_without_fragmentrinforrZcheck_against_chunksr)rNrWrehashes progress_barZ total_lengthZ cached_respZ show_progressrprcrfZprogress_indicatorr>Zdownloaded_chunksrR)rerNrS _download_urlsP   %    rtcCsd}tj||j}tj|rtdt|d}|dkr@d}nj|dkrdtdt|t |nF|dkrt |}td t|t|t ||n|d krt d |rt ||td t|dS) NTz8The file %s exists. (i)gnore, (w)ipe, (b)ackup, (a)abort)iwbaruFrvz Deleting %srwzBacking up %s to %srxzSaved %s)osrOjoinfilenameexistsrrrmwarningremovershutilZmovesysexitcopyrq)r|rXrWrZdownload_locationresponseZ dest_filerRrRrS _copy_files6    ronc Cs|dkrtdtddv}d}|r0t|||}|rH|}t|d} nt|||j||\}} t||| |r~|s~t||||st |W5QRXdS)Nz@unpack_http_url() missing 1 required keyword argument: 'session'rV)Zkindr) r;r_check_download_dir mimetypes guess_type_download_http_urlrOrrrzunlink) rWrX download_dirrLrrrstemp_diralready_downloaded_path from_path content_typerRrRrSr/s2    c CsTzt||Wn@tjk rN}z tdt|t|t|W5d}~XYnXdS)zCopying special files is not supported, but as a convenience to users we skip errors copying them. This supports tools that may create e.g. socket files in the project source directory. z>Ignoring special file error '%s' encountered copying %s to %s.N)r rZSpecialFileErrorrmr~rKr)srcdesterRrRrS_copy2_ignoring_special_filesJsrcs`tj|}tj|tj|fdd}t|dd}tsLt|d<tj |f|dS)Ncs6g}|kr|ddg7}tj|kr2|g7}|S)Nz.toxz.nox)rzrOabspath)dnamesZskippedsourceZtarget_basenameZtarget_dirnamerRrSr)es   z!_copy_source_tree..ignoreTr(r+) rzrOrbasenamedirnamedictrrrZcopytree)rtargetZtarget_abspathr)r]rRrrS_copy_source_tree_s    rcCs|j}|r>tj|r"t|t|||r:tddS|rL| |d}|r`t |||}|rj|}n|}t |d}t ||||r|st|||dS)zUnpack link into location. If download_dir is provided and link points to a file, make a copy of the link file inside download_dir. z*Link is a directory, ignoring download_dirNr) file_pathZis_existing_dirrzrOisdirrrrmrqcheck_against_pathrrrrr)rWrXrrrZ link_pathrrrrRrRrSr.~s.      cCsP|jrt||n:|jr*t||||dn"|dkr8t}t||||||ddS)avUnpack link. If link is a VCS link: if only_download, export into download_dir and ignore location else unpack into location for other types of link: - unpack into location - if download_dir, copy the file into download_dir - if only_download, mark location for deletion :param hashes: A Hashes object, one of whose embedded hashes must match, or HashMismatch will be raised. If the Hashes is empty, no matches are required, and unhashable types of requirements (like VCS ones, which would ordinarily raise HashUnsupported) are allowed. )rrN)rrrs)Zis_vcsr-is_filer.r r/)rWrXrrLrrrsrRrRrSr0s cCs tj|S)zJ Sanitize the "filename" value from a Content-Disposition header. )rzrOr)r|rRrRrSr2scCs,t|\}}|d}|r$t|}|p*|S)z Parse the "filename" value from a Content-Disposition header, and return the default filename if the result is empty. r|)cgiZ parse_headerr<r2)content_dispositionZdefault_filenameZ_typeZparamsr|rRrRrSr1s  c Cs*|jddd}z |j|ddidd}|Wn8tjk rj}ztd|jj |W5d }~XYnX|j d d }|j } |j d } | rt | | } t | d} | st|} | r| | 7} | s|j|jkrtj |jd} | r| | 7} tj|| } t| d } t||| ||W5QRX| |fS)z6Download link url into temp_dir using provided session#r6rzAccept-EncodingZidentityT)rir`zHTTP error %s while getting %sNz content-typezcontent-dispositionwb)r>rAr<r=rZ HTTPErrorrmZcriticalrZ status_coderir|r1rrZguess_extensionrzrOr{rHrt)rWrLrrrrsZ target_urlrNrQrr|rZextrrerRrRrSrsB       rcCsptj||j}tj|s dStd||rlz||Wn,tk rjt d|t |YdSX|S)z Check download_dir for previously downloaded file with correct hash If a correct file is found return its path else None NzFile was already downloaded %sz;Previously-downloaded file %s has bad hash. Re-downloading.) rzrOr{r|r}rmrqrrr~r)rWrrrZ download_pathrRrRrSr7s   r)NN)NNNr)NN)NNNr)\Z __future__rrrnrrzrerrZ pip._vendorrZpip._vendor.requests.modelsrrZpip._vendor.sixrZpip._vendor.six.moves.urllibrrFZpip._internal.exceptionsrr Zpip._internal.models.indexr Zpip._internal.network.sessionr Zpip._internal.utils.encodingr Zpip._internal.utils.filesystemr Zpip._internal.utils.miscrrrrrrrrrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.uirZpip._internal.utils.unpackingrZpip._internal.utils.urlsrZpip._internal.vcsrtypingrrrr r!r"Zmypy_extensionsr#Zpip._internal.models.linkr$Zpip._internal.utils.hashesr%Z pip._internal.vcs.versioncontrolr&rKboolr'__all__Z getLogger__name__rmr,compileIrCr-rTr^rtrr/rrr.r0r2r1rrrRrRrRrSs        ,               0 d -" 8 .?site-packages/pip/_internal/__pycache__/configuration.cpython-38.opt-1.pyc000064400000024475151464452710022522 0ustar00U .e7@sdZddlZddlZddlZddlZddlmZddlmZm Z ddl m Z ddl m Z mZddlmZmZddlmZerdd lmZmZmZmZmZmZmZejZed eZee Z!d d Z"d dZ#eddddddZ$e rdndZ%ddZ&Gddde'Z(dS)a Configuration management setup Some terminology: - name As written in config files. - value Value associated with a name - key Name combined with it's section (section.name) - variant A single word describing where the configuration key-value pair came from N) configparser)ConfigurationError!ConfigurationFileCouldNotBeLoaded)appdirs)WINDOWS expanduser) ensure_direnum)MYPY_CHECK_RUNNING)AnyDictIterableListNewTypeOptionalTupleKindcCs*|dd}|dr&|dd}|S)zFMake a name consistent regardless of source (environment or file) _-z--N)lowerreplace startswith)namer?/usr/lib/python3.8/site-packages/pip/_internal/configuration.py_normalize_name.s  rcCs&d|krd|}t||ddS)N.zbKey does not contain dot separated section and key. Perhaps you wanted to use 'global.{}' instead?)formatrsplit)rZ error_messagerrr_disassemble_key8sr!userglobalZsiteenvzenv-var)USERGLOBALSITEENVENV_VARzpip.inizpip.confcCspddtdD}tjtjt}tjtdt r8dndt}tjt dt}t j |t j |gt j||giS)NcSsg|]}tj|tqSr)ospathjoinCONFIG_BASENAME).0r+rrr Qsz+get_configuration_files..Zpip~z.pip)rZsite_config_dirsr*r+r,sysprefixr-rrZuser_config_dirkindsr&r'r%)Zglobal_config_filesZsite_config_fileZlegacy_config_fileZnew_config_filerrrget_configuration_filesPs( r4cseZdZdZd)fdd ZddZddZd d Zd d Zd dZ ddZ ddZ ddZ e ddZddZddZddZddZdd Zd!d"Zd#d$Zd%d&Zd'd(ZZS)* ConfigurationaHandles management of configuration. Provides an interface to accessing and managing configuration files. This class converts provides an API that takes "section.key-name" style keys and stores the value associated with it as "key-name" under the section "section". This allows for a clean interface wherein the both the section and the key-name are preserved in an easy to manage form in the configuration files and the data stored is also nice. Nc stt|tjtjtjdg}||krJtdd t t |dd||_ ||_ tjtjtjtjtjg|_ddg|_dd|jD|_dd|jD|_g|_dS) Nz5Got invalid value for load_only - should be one of {}z, versionhelpcSsi|] }|gqSrrr.variantrrr sz*Configuration.__init__..cSsi|] }|iqSrrr9rrrr;s)superr5__init__r3r%r&r'rrr,mapreprisolated load_onlyr(r)_override_order_ignore_env_names_parsers_config_modified_parsers)selfr@rAZ_valid_load_only __class__rrr=ts0 zConfiguration.__init__cCs||js|dS)zELoads configuration from configuration files and environment N)_load_config_filesr@_load_environment_varsrGrrrloadszConfiguration.loadcCs*z|dWStk r$YdSXdS)z@Returns the file with highest priority in configuration rN)_get_parser_to_modify IndexErrorrLrrrget_file_to_editszConfiguration.get_file_to_editcCs |jS)z`Returns key-value pairs like dict.items() representing the loaded configuration ) _dictionaryitemsrLrrrrRszConfiguration.itemscCs4z |j|WStk r.td|YnXdS)z,Get a value from the configuration. No such key - {}N)rQKeyErrorrr)rGkeyrrr get_values zConfiguration.get_valuecCsj||\}}|dk rJt|\}}||s<||||||||j|j|<|||dS)z-Modify a value in the configuration. N) _ensure_have_load_onlyrNr! has_sectionZ add_sectionsetrErA_mark_as_modified)rGrUvaluefnameparsersectionrrrr set_values    zConfiguration.set_valuec Cs|||j|jkr&td||\}}|dk rt|\}}d}||r`|||}|rt | |}z t |}Wnt k rd}YnX|dkr| ||||ntd|j|j|=dS)z,Unset a value in the configuration. rSNFz4Fatal Internal error [id=1]. Please report as a bug.)rWrErArrrNr!rXZ remove_optioniterrRnext StopIterationZremove_sectionrZ) rGrUr\r]r^rZmodified_somethingZ section_itervalrrr unset_values,       zConfiguration.unset_valuec CsX||jD]D\}}td|ttj|t|d}| |W5QRXqdS)z*Save the current in-memory state. z Writing to %swN) rWrFloggerinforr*r+dirnameopenwrite)rGr\r]frrrsaves   zConfiguration.savecCs$|jdkrtdtd|jdS)Nz'Needed a specific file to be modifying.z$Will be working with %s variant only)rArrfdebugrLrrrrWs z$Configuration._ensure_have_load_onlycCs$i}|jD]}||j|q |S)zs6   $     site-packages/pip/_internal/__pycache__/locations.cpython-38.opt-1.pyc000064400000005766151464452710021650 0ustar00U .e&@sdZddlmZddlZddlZddlZddlZddlZddlZddl mZ ddl m Z ddl mZddlmZddlmZdd lmZerdd lmZmZmZmZmZed Zd d ZddZedZ e!"dkre #Z z e$Z%Wne&k r ej'Z%YnXer`ej()ej*dZ+ej()e%dZ,ej(-e+sej()ej*dZ+ej()e%dZ,nJej()ej*dZ+ej()e%dZ,ejdddkrej*dddkrdZ+dddZ.dS)z7Locations where we look for configs, install stuff, etc)absolute_importN) sysconfig) SCHEME_KEYS)appdirs)WINDOWS)MYPY_CHECK_RUNNING)running_under_virtualenv)AnyUnionDictListOptionalZpipcCs djtjS)ze Return the major-minor version of the current Python as a string, e.g. "3.7" or "3.10". z{}.{})formatsys version_inforr;/usr/lib/python3.8/site-packages/pip/_internal/locations.pyget_major_minor_versionsrcCsZtrtjtjd}n6ztjtd}Wntk rLtdYnXtj |S)Nsrcz=The folder you are executing pip from can no longer be found.) rospathjoinrprefixgetcwdOSErrorexitabspath)Z src_prefixrrrget_src_prefix(s rpurelibZpypyZScriptsbindarwinz/System/Library/z/usr/local/binFcCsHddlm}i}|r ddgi}ni}d|i} | ||| } | | jddd} |p\| j| _|sh|rnd | _|pv| j| _|p| j| _|p| j| _| t D]} t | d | || <qd | dkr|t | j| jd trDtjtjd ddt||d<|dk rDtjtj|dd} tj|| dd|d<|S)z+ Return a distutils install scheme r) DistributionZ script_argsz --no-user-cfgnameZinstallT)ZcreateZinstall_ install_lib)rZplatlibZincludesitezpython{}ZheadersN)Zdistutils.distr#updateZparse_config_filesZget_command_objuserrhomerootZfinalize_optionsrgetattrZget_option_dictdictr&rrrrrrr splitdriver)Z dist_namer*r+r,isolatedrr#ZschemeZextra_dist_argsZ dist_argsdikeyZ path_no_driverrrdistutils_scheme[sP           r4)FNNFN)/__doc__Z __future__rrZos.pathplatformr'rrZ distutilsZdistutils_sysconfigZdistutils.command.installrZpip._internal.utilsrZpip._internal.utils.compatrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrtypingr r r r r Zuser_cache_dirZUSER_CACHE_DIRrrZget_pathZ site_packagesZpython_implementationlowerZget_python_libgetusersitepackages user_siteAttributeError USER_SITErrrZbin_pyZbin_userexistsr4rrrrsN            (site-packages/pip/_internal/__pycache__/main.cpython-38.pyc000064400000002371151464452710017627 0ustar00U .eO@sdZddlmZddlZddlZddlZddlZddlmZddl m Z ddl m Z ddl mZddlmZeeZd d d ZdS) z Primary application entrypoint. )absolute_importN) autocomplete) parse_command)create_command)PipError) deprecationc Cs|dkrtjdd}ttzt|\}}WnJtk r~}z,tjd|tjt j t dW5d}~XYnXzt t jdWn0t jk r}ztd|W5d}~XYnXt|d|kd}||S)Nz ERROR: %sz%Ignoring error %s when setting localez --isolated)isolated)sysargvrZinstall_warning_loggerrrrstderrwriteoslinesepexitlocale setlocaleLC_ALLErrorloggerdebugrmain)argsZcmd_nameZcmd_argsexceZcommandr6/usr/lib/python3.8/site-packages/pip/_internal/main.pyrs r)N)__doc__Z __future__rrZloggingrr Z pip._internal.cli.autocompletionrZpip._internal.cli.main_parserrZpip._internal.commandsrZpip._internal.exceptionsrZpip._internal.utilsrZ getLogger__name__rrrrrrs       site-packages/pip/_internal/__pycache__/legacy_resolve.cpython-38.opt-1.pyc000064400000024134151464452710022646 0ustar00U .eRC@sFdZddlZddlZddlmZddlmZddlmZddl m Z m Z m Z m Z mZddlmZddlmZmZmZmZdd lmZmZdd lmZerdd lmZmZmZmZm Z m!Z!dd l"m#Z#dd l$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/ee0e-ge-fZ1e2e3Z4dddZ5Gddde6Z7dS)ayDependency Resolution The dependency resolution in pip is performed as follows: for top-level requirements: a. only one spec allowed per project, regardless of conflicts or not. otherwise a "double requirement" exception is raised b. they override sub-dependency requirements. for sub-dependencies a. "first found, wins" (where the order is breadth first) N) defaultdict)chain) specifiers)BestVersionAlreadyInstalledDistributionNotFound HashError HashErrorsUnsupportedPythonVersion) indent_log)dist_in_install_pathdist_in_usersite ensure_dirnormalize_version_info)check_requires_pythonget_requires_python)MYPY_CHECK_RUNNING)Callable DefaultDictListOptionalSetTuple) pkg_resources)AbstractDistribution) PipSession) PackageFinder)RequirementPreparer)InstallRequirement)RequirementSetFc Cst|}zt||d}Wn:tjk rR}ztd|j|WYdSd}~XYnX|r\dSdtt |}|rt d|j||dSt d |j||dS)a Check whether the given Python version is compatible with a distribution's "Requires-Python" value. :param version_info: A 3-tuple of ints representing the Python major-minor-micro version to check. :param ignore_requires_python: Whether to ignore the "Requires-Python" value if the given Python version isn't compatible. :raises UnsupportedPythonVersion: When the given Python version isn't compatible. ) version_infoz-Package %r has an invalid Requires-Python: %sN.zBIgnoring failed Requires-Python check for package %r: %s not in %rz8Package {!r} requires a different Python: {} not in {!r}) rrrZInvalidSpecifierloggerwarningZ project_namejoinmapstrdebugr format)distrignore_requires_pythonZrequires_pythonZ is_compatibleexcversionr,@/usr/lib/python3.8/site-packages/pip/_internal/legacy_resolve.py_check_dist_requires_python>s> r.csdeZdZdZdddhZdfdd Zdd Zd d Zd d ZddZ ddZ ddZ ddZ Z S)ResolverzResolves which packages need to be installed/uninstalled to perform the requested operation without breaking the requirements of any package. eageronly-if-neededto-satisfy-onlyNc stt|| dkr&tjdd} nt| } | |_||_||_||_ d|_ | |_ | |_ ||_ ||_||_||_||_tt|_dS)N)superr/__init__sysrr_py_version_infopreparerfindersessionrequire_hashesupgrade_strategyforce_reinstallignore_dependenciesignore_installedr) use_user_site_make_install_reqrlist_discovered_dependencies) selfr8r:r9Zmake_install_reqr@r>r?r)r=r<Zpy_version_info __class__r,r-r5us$zResolver.__init__c Cs|jjrt|jj|jt|j}|jp>tdd|D|_|j j }| }|r`t |g}t}t||D]N}z||||Wqttk r}z||_||W5d}~XYqtXqt|r|dS)aResolve what operations need to be done As a side-effect of this method, the packages (and their dependencies) are downloaded, unpacked and prepared for installation. This preparation is done by ``pip.operations.prepare``. Once PyPI has static dependency metadata available, it would be possible to move the preparation to become a step separated from dependency resolution. css|] }|jVqdSN)Zhas_hash_options).0reqr,r,r- sz#Resolver.resolve..N)r8Zwheel_download_dirr Zunnamed_requirementsrB requirementsvaluesr;anyr9 search_scopeZget_formatted_locationsr!inforrextend _resolve_onerrIappend) rDrequirement_setZ root_reqsrNZ locationsZdiscovered_reqsZ hash_errorsrIr*r,r,r-resolves2     zResolver.resolvecCs&|jdkrdS|jdkrdS|jSdS)Nr2Fr0T)r< is_directrDrIr,r,r-_is_upgrade_alloweds   zResolver._is_upgrade_allowedcCs,|jrt|jr"t|jr"|j|_d|_dS)z4 Set a requirement to be installed. N)r@r satisfied_byr Zconflicts_withrVr,r,r-_set_req_to_reinstallszResolver._set_req_to_reinstallcCs|jr dS||j|js dS|jr4||dS||sP|jdkrLdSdS|jsz|j j |ddWn(t k rYdSt k rYnX||dS)aCheck if req_to_install should be skipped. This will check if the req is installed, and whether we should upgrade or reinstall it, taking into account all the relevant user options. After calling this req_to_install will only have satisfied_by set to None if the req_to_install is to be upgraded/reinstalled etc. Any other value will be a dist recording the current thing installed that satisfies the requirement. Note that for vcs urls and the like we can't assess skipping in this routine - we simply identify that we need to pull the thing down, then later on it is pulled down and introspected to assess upgrade/ reinstalls etc. :return: A text reason for why it was skipped, or None. Nr1z#already satisfied, skipping upgradezalready satisfiedT)Zupgradezalready up-to-date) r?check_if_existsr@rXr=rYrWr<linkr9Zfind_requirementrr)rDreq_to_installr,r,r-_check_skip_installeds*     zResolver._check_skip_installedcCs|jr|j||j|j|jS||}|jr@|j||j|S| |}| |j||j|j ||j |j|j}|j s||j|jr|jdkp|jp|j p|jjdk}|r||n td||S)zzTakes a InstallRequirement and returns a single AbstractDist representing a prepared variant of the same. r2filez.add_reqN)rbz!Installing extra requirements: %r,z"%s does not provide the extra '%s')rc) constraintZpreparedZreqs_to_cleanuprRraZget_pkg_resources_distributionr.r7r)r Zhas_requirementrdrUrer>Zextrasr!r&r#sortedsetr"Zrequiresr_rXZsuccessfully_downloaded) rDrSr\r`r(rhZmissing_requestedZmissingZavailable_requestedrfr,rgr-rQ[sV     zResolver._resolve_onecs8gtfdd|jD] }|q&S)zCreate the installation order. The installation order is topological - requirements are installed before the requiring thing. We break cycles at an arbitrary point, and make no other guarantees. csN|js|krdS|jrdS|j|jD] }|q2|dSrG)rXrjaddrCrdrR)rIZdeporderZ ordered_reqsschedulerDr,r-rps  z1Resolver.get_installation_order..schedule)rlrKrL)rDZreq_setZ install_reqr,rnr-get_installation_orders   zResolver.get_installation_order)N)__name__ __module__ __qualname____doc__Z_allowed_strategiesr5rTrWrYr]rarQrq __classcell__r,r,rEr-r/ns +3  5<Zr/)F)8ruZloggingr6 collectionsr itertoolsrZpip._vendor.packagingrZpip._internal.exceptionsrrrrr Zpip._internal.utils.loggingr Zpip._internal.utils.miscr r r rZpip._internal.utils.packagingrrZpip._internal.utils.typingrtypingrrrrrrZ pip._vendorrZpip._internal.distributionsrZpip._internal.network.sessionrZpip._internal.indexrZ pip._internal.operations.preparerZpip._internal.req.req_installrZpip._internal.req.req_setrr%ZInstallRequirementProviderZ getLoggerrrr!r.objectr/r,r,r,r-s4                0site-packages/pip/_internal/__pycache__/pyproject.cpython-38.opt-1.pyc000064400000006076151464452710021667 0ustar00U .eZ@sddlmZddlZddlZddlZddlmZmZddlm Z ddl m Z e rhddl m Z mZmZmZddZd d Zd d ZdS) )absolute_importN)pytomlsix)InstallationError)MYPY_CHECK_RUNNING)AnyTupleOptionalListcCst|totdd|DS)Ncss|]}t|tjVqdS)N) isinstancerZ string_types).0itemr;/usr/lib/python3.8/site-packages/pip/_internal/pyproject.py sz"_is_list_of_str..)r listall)objrrr_is_list_of_strs rcCs2tj|d}tjr.t|tjr.|t }|S)Nzpyproject.toml) ospathjoinrZPY2r Z text_typeencodesysgetfilesystemencoding)Zunpacked_source_directoryrrrrmake_pyproject_pathsrc Cs<tj|}tj|}|rLtj|dd}t|}W5QRX|d}nd}|rr|sr|dk rl|sltdd}n<|rd|kr|dk r|std |dd}n |dkr|}|sdS|dkrd d gd d }d } d|krt| j |dd|d} t | st| j |dd|d} g} | dkr2d } d d g} | | | fS)aLoad the pyproject.toml file. Parameters: use_pep517 - Has the user requested PEP 517 processing? None means the user hasn't explicitly specified. pyproject_toml - Location of the project's pyproject.toml file setup_py - Location of the project's setup.py file req_name - The name of the requirement we're processing (for error reporting) Returns: None if we should use the legacy code path, otherwise a tuple ( requirements from pyproject.toml, name of PEP 517 backend, requirements we should check are installed after setting up the build environment ) zutf-8)encodingz build-systemNzIDisabling PEP 517 processing is invalid: project does not have a setup.pyT build-backendzbDisabling PEP 517 processing is invalid: project specifies a build backend of {} in pyproject.tomlzsetuptools>=40.8.0Zwheelz setuptools.build_meta:__legacy__)requiresrzO{package} has a pyproject.toml file that does not comply with PEP 518: {reason}rz]it has a 'build-system' table but not 'build-system.requires' which is mandatory in the table)packagereasonz1'build-system.requires' is not a list of strings.) rrisfileioopenrloadgetrformatr) Z use_pep517Zpyproject_tomlZsetup_pyZreq_nameZ has_pyprojectZ has_setupfZpp_tomlZ build_systemZerror_templaterZbackendZcheckrrrload_pyproject_toml#sb          r()Z __future__rr"rrZ pip._vendorrrZpip._internal.exceptionsrZpip._internal.utils.typingrtypingrrr r rrr(rrrrs    site-packages/pip/_internal/__pycache__/__init__.cpython-38.pyc000064400000000320151464452710020432 0ustar00U .eP@s ddlZdS)N)Z*pip._internal.utils.inject_securetransportZpiprr:/usr/lib/python3.8/site-packages/pip/_internal/__init__.pysite-packages/pip/_internal/__pycache__/collector.cpython-38.opt-1.pyc000064400000033422151464452710021631 0ustar00U .eWF@s dZddlZddlZddlZddlZddlZddlmZddlm Z m Z ddl m Z ddl mZmZmZddlmZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZm Z ddl!m"Z"m#Z#erDddl$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-ddl.Z/ddl0m1Z1ddl2m3Z3ddl4m5Z5e/j6j7j8Z9e)e:e:fZ;eddZ?ddZ@GdddeAZBddZCGdddeAZDddZEdd ZFd!d"ZGd#d$ZHd%d&ZId'd(ZJd)d*ZKGd+d,d,eLZMdd6d7ZRGd8d9d9eLZSGd:d;d;eLZTdS)?zM The main purpose of this module is to expose LinkCollector.collect_links(). N) OrderedDict)html5librequests)unescape) HTTPError RetryErrorSSLError)parse)requestLink)ARCHIVE_EXTENSIONS)redact_auth_from_url)MYPY_CHECK_RUNNING) path_to_url url_to_path)is_urlvcs) CallableDictIterableListMutableMappingOptionalSequenceTupleUnion)Response) SearchScope) PipSessioncCs6tjD]*}||r|t|dkr|SqdS)zgLook for VCS schemes in the URL. Returns the matched VCS scheme, or None if there's no match. z+:N)rZschemeslower startswithlen)urlschemer%;/usr/lib/python3.8/site-packages/pip/_internal/collector.py_match_vcs_scheme/s  r'cCs(t|j}tD]}||rdSqdS)z2Return whether the URL looks like an archive. TF)r filenamer endswith)r#r(Zbad_extr%r%r&_is_url_like_archive;s   r*cseZdZfddZZS)_NotHTMLcs"tt|||||_||_dSN)superr+__init__ content_type request_desc)selfr/r0 __class__r%r&r.Gsz_NotHTML.__init__)__name__ __module__ __qualname__r. __classcell__r%r%r2r&r+Fsr+cCs.|jdd}|ds*t||jjdS)zCheck the Content-Type header to ensure the response contains HTML. Raises `_NotHTML` if the content type is not text/html. Content-Type text/htmlN)headersgetr r!r+r method)responser/r%r%r&_ensure_html_headerNsr?c@s eZdZdS)_NotHTTPN)r4r5r6r%r%r%r&r@Ysr@cCsDt|\}}}}}|dkr"t|j|dd}|t|dS)zSend a HEAD request to the URL, and ensure the response contains HTML. Raises `_NotHTTP` if the URL is not available for a HEAD request, or `_NotHTML` if the content type is not text/html. >httphttpsT)Zallow_redirectsN) urllib_parseZurlsplitr@headraise_for_statusr?)r#sessionr$netlocpathZqueryZfragmentrespr%r%r&_ensure_html_response]s rJcCsLt|rt||dtdt||j|dddd}|t||S)aAccess an HTML page with GET, and return the response. This consists of three parts: 1. If the URL looks suspiciously like an archive, send a HEAD first to check the Content-Type is HTML, to avoid downloading a large file. Raise `_NotHTTP` if the content type cannot be determined, or `_NotHTML` if it is not HTML. 2. Actually perform the request. Raise HTTP exceptions on network failures. 3. Check the Content-Type header to make sure we got HTML, and raise `_NotHTML` otherwise. rFzGetting page %sr:z max-age=0)ZAcceptz Cache-Control)r;)r*rJloggerdebugrr<rEr?)r#rFrIr%r%r&_get_html_responsens rNcCs2|r.d|kr.t|d\}}d|kr.|dSdS)zBDetermine if we have any encoding information in our headers. r8charsetN)cgiZ parse_header)r;r/Zparamsr%r%r&_get_encoding_from_headerss  rQcCs.|dD]}|d}|dk r |Sq |S)aDetermine the HTML document's base URL. This looks for a ```` tag in the HTML document. If present, its href attribute denotes the base URL of anchor tags in the document. If there is no such tag (or if it does not have a valid href attribute), the HTML file's URL is used as the base URL. :param document: An HTML document representation. The current implementation expects the result of ``html5lib.parse()``. :param page_url: The URL of the HTML document. z.//basehrefN)findallr<)documentpage_urlbaserRr%r%r&_determine_base_urls   rWcCsPt|}|jdkr(tt|j}ntjt|jdd}t |j |dS)zMakes sure a link is fully encoded. That is, if a ' ' shows up in the link, it will be rewritten to %20 (while not over-quoting % or other characters).r9z/@)Zsafe)rH) rCurlparserGurllib_requestZ pathname2url url2pathnamerHZquoteZunquoteZ urlunparse_replace)r#resultrHr%r%r& _clean_links   r]cCsf|d}|sdStt||}|d}|r8t|nd}|d}|rRt|}t||||d}|S)zJ Convert an anchor element in a simple repository page to a Link. rRNzdata-requires-pythonz data-yanked)Z comes_fromZrequires_python yanked_reason)r<r]rCurljoinrr )anchorrUbase_urlrRr#Z pyrequirer^linkr%r%r&_create_link_from_elements   rcccsVtj|j|jdd}|j}t||}|dD]"}t|||d}|dkrJq.|Vq.dS)zP Parse an HTML document, and yield its anchor elements as Link objects. F)Ztransport_encodingZnamespaceHTMLElementsz.//a)rUraN)rr contentencodingr#rWrSrc)pagerTr#rar`rbr%r%r& parse_linkss  rgc@s eZdZdZddZddZdS)HTMLPagez'Represents one page, along with its URLcCs||_||_||_dS)z :param encoding: the encoding to decode the given content. :param url: the URL from which the HTML was downloaded. N)rdrer#)r1rdrer#r%r%r&r.s zHTMLPage.__init__cCs t|jSr,)rr#r1r%r%r&__str__$szHTMLPage.__str__N)r4r5r6__doc__r.rjr%r%r%r&rhsrhcCs|dkrtj}|d||dS)Nz%Could not fetch URL %s: %s - skipping)rLrM)rbreasonmethr%r%r&_handle_get_page_fail(srncCst|j}t|j||jdS)N)rer#)rQr;rhrdr#)r>rer%r%r&_make_html_page3s roc Cs|dkrtd|jddd}t|}|r@td||dSt|\}}}}}}|dkrtj t |r| ds|d7}t|d}td |zt||d }WnDtk rtd |Yn,tk r}ztd ||j|jW5d}~XYntk r0}zt||W5d}~XYntk r\}zt||W5d}~XYntk r}z$d } | t|7} t|| tjdW5d}~XYn\tjk r}zt|d|W5d}~XYn*tjk rt|dYn Xt|SdS)Nz?_get_html_page() missing 1 required keyword argument: 'session'#rzCannot look at %s URL %sfile/z index.htmlz# file: URL is directory, getting %srKzQSkipping page %s because it looks like an archive, and cannot be checked by HEAD.z.sort_pathzfile:z)Path '{0}' is ignored: it is a directory.z:Url '%s' is ignored: it is neither a file nor a directory.zQUrl '%s' is ignored. It is either a non-existing path or lacks a specific scheme.)rvrHexistsr!rrwrealpathlistdirjoinrrLZwarningformatisfiler) locations expand_dirrr#Z is_local_pathZ is_file_urlrHitemr%rr&group_locationsxsF        rc@seZdZdZddZdS)CollectedLinksa Encapsulates all the Link objects collected by a call to LinkCollector.collect_links(), stored separately as-- (1) links from the configured file locations, (2) links from the configured find_links, and (3) a dict mapping HTML page url to links from that page. cCs||_||_||_dS)z :param files: Links from file locations. :param find_links: Links from find_links. :param pages: A dict mapping HTML page url to links from that page. Nr find_linkspages)r1rrrr%r%r&r.s zCollectedLinks.__init__N)r4r5r6rkr.r%r%r%r&rs rc@s4eZdZdZddZeddZddZdd Zd S) LinkCollectorz Responsible for collecting Link objects from all configured locations, making network requests as needed. The class's main method is its collect_links() method. cCs||_||_dSr,) search_scoperF)r1rFrr%r%r&r.szLinkCollector.__init__cCs|jjSr,)rrrir%r%r&rszLinkCollector.find_linksccs,|D]"}t||jd}|dkr q|VqdS)zp Yields (page, page_url) from the given locations, skipping locations that have errors. rKN)r}rF)r1rlocationrfr%r%r& _get_pagess zLinkCollector._get_pagescsj}||}t|\}}tjdd\}}ddt||D}ddjD} fddtdd|Dd d|DD} t| } d t| |g} | D]} | d | qt d | i} | D]}tt|| |j<qt|| | d S)zFind all available links for the given project name. :return: All the Link objects (unfiltered), as a CollectedLinks object. T)rcSsg|] }t|qSr%r .0r#r%r%r& sz/LinkCollector.collect_links..cSsg|]}t|dqS)z-fr rr%r%r&rscsg|]}j|r|qSr%)rFZis_secure_origin)rrbrir%r&r s css|]}t|VqdSr,r rr%r%r& sz.LinkCollector.collect_links..css|]}t|VqdSr,r rr%r%r&r sz,{} location(s) to search for versions of {}:z* {} r)rZget_index_urls_locationsrr itertoolschainrrr"rrLrMrrr~rgr#r)r1Z project_namerZindex_locationsZindex_file_locZ index_url_locZ fl_file_locZ fl_url_locZ file_linksZfind_link_linksZ url_locationslinesrbZ pages_linksrfr%rir& collect_linkssD       zLinkCollector.collect_linksN) r4r5r6rkr.propertyrrrr%r%r%r&rs    r)N)N)F)UrkrPrZloggingrrv collectionsrZ pip._vendorrrZpip._vendor.distlib.compatrZpip._vendor.requests.exceptionsrrrZpip._vendor.six.moves.urllibr rCr rYZpip._internal.models.linkr Zpip._internal.utils.filetypesr Zpip._internal.utils.miscrZpip._internal.utils.typingrZpip._internal.utils.urlsrrZpip._internal.vcsrrtypingrrrrrrrrrZxml.etree.ElementTreeZxmlZpip._vendor.requestsrZ!pip._internal.models.search_scoperZpip._internal.network.sessionrZetreeZ ElementTreeZElementZ HTMLElementrxZResponseHeadersZ getLoggerr4rLr'r* Exceptionr+r?r@rJrNrQrWr]rcrgobjectrhrnror}rrrrr%r%r%r&s^        ,         3    6 ;site-packages/pip/_internal/__pycache__/configuration.cpython-38.pyc000064400000024625151464452710021560 0ustar00U .e7@sdZddlZddlZddlZddlZddlmZddlmZm Z ddl m Z ddl m Z mZddlmZmZddlmZerdd lmZmZmZmZmZmZmZejZed eZee Z!d d Z"d dZ#eddddddZ$e rdndZ%ddZ&Gddde'Z(dS)a Configuration management setup Some terminology: - name As written in config files. - value Value associated with a name - key Name combined with it's section (section.name) - variant A single word describing where the configuration key-value pair came from N) configparser)ConfigurationError!ConfigurationFileCouldNotBeLoaded)appdirs)WINDOWS expanduser) ensure_direnum)MYPY_CHECK_RUNNING)AnyDictIterableListNewTypeOptionalTupleKindcCs*|dd}|dr&|dd}|S)zFMake a name consistent regardless of source (environment or file) _-z--N)lowerreplace startswith)namer?/usr/lib/python3.8/site-packages/pip/_internal/configuration.py_normalize_name.s  rcCs&d|krd|}t||ddS)N.zbKey does not contain dot separated section and key. Perhaps you wanted to use 'global.{}' instead?)formatrsplit)rZ error_messagerrr_disassemble_key8sr!userglobalZsiteenvzenv-var)USERGLOBALSITEENVENV_VARzpip.inizpip.confcCspddtdD}tjtjt}tjtdt r8dndt}tjt dt}t j |t j |gt j||giS)NcSsg|]}tj|tqSr)ospathjoinCONFIG_BASENAME).0r+rrr Qsz+get_configuration_files..Zpip~z.pip)rZsite_config_dirsr*r+r,sysprefixr-rrZuser_config_dirkindsr&r'r%)Zglobal_config_filesZsite_config_fileZlegacy_config_fileZnew_config_filerrrget_configuration_filesPs( r4cseZdZdZd)fdd ZddZddZd d Zd d Zd dZ ddZ ddZ ddZ e ddZddZddZddZddZdd Zd!d"Zd#d$Zd%d&Zd'd(ZZS)* ConfigurationaHandles management of configuration. Provides an interface to accessing and managing configuration files. This class converts provides an API that takes "section.key-name" style keys and stores the value associated with it as "key-name" under the section "section". This allows for a clean interface wherein the both the section and the key-name are preserved in an easy to manage form in the configuration files and the data stored is also nice. Nc stt|tjtjtjdg}||krJtdd t t |dd||_ ||_ tjtjtjtjtjg|_ddg|_dd|jD|_dd|jD|_g|_dS) Nz5Got invalid value for load_only - should be one of {}z, versionhelpcSsi|] }|gqSrrr.variantrrr sz*Configuration.__init__..cSsi|] }|iqSrrr9rrrr;s)superr5__init__r3r%r&r'rrr,mapreprisolated load_onlyr(r)_override_order_ignore_env_names_parsers_config_modified_parsers)selfr@rAZ_valid_load_only __class__rrr=ts0 zConfiguration.__init__cCs||js|dS)zELoads configuration from configuration files and environment N)_load_config_filesr@_load_environment_varsrGrrrloadszConfiguration.loadcCs<|jdk stdz|dWStk r6YdSXdS)z@Returns the file with highest priority in configuration Nz)Need to be specified a file to be editingr)rAAssertionError_get_parser_to_modify IndexErrorrLrrrget_file_to_edits zConfiguration.get_file_to_editcCs |jS)z`Returns key-value pairs like dict.items() representing the loaded configuration ) _dictionaryitemsrLrrrrSszConfiguration.itemscCs4z |j|WStk r.td|YnXdS)z,Get a value from the configuration. No such key - {}N)rRKeyErrorrr)rGkeyrrr get_values zConfiguration.get_valuecCsj||\}}|dk rJt|\}}||s<||||||||j|j|<|||dS)z-Modify a value in the configuration. N) _ensure_have_load_onlyrOr! has_sectionZ add_sectionsetrErA_mark_as_modified)rGrVvaluefnameparsersectionrrrr set_values    zConfiguration.set_valuec Cs|||j|jkr&td||\}}|dk rt|\}}d}||r`|||}|rt | |}z t |}Wnt k rd}YnX|dkr| ||||ntd|j|j|=dS)z,Unset a value in the configuration. rTNFz4Fatal Internal error [id=1]. Please report as a bug.)rXrErArrrOr!rYZ remove_optioniterrSnext StopIterationZremove_sectionr[) rGrVr]r^r_rZmodified_somethingZ section_itervalrrr unset_values,       zConfiguration.unset_valuec CsX||jD]D\}}td|ttj|t|d}| |W5QRXqdS)z*Save the current in-memory state. z Writing to %swN) rXrFloggerinforr*r+dirnameopenwrite)rGr]r^frrrsaves   zConfiguration.savecCs$|jdkrtdtd|jdS)Nz'Needed a specific file to be modifying.z$Will be working with %s variant only)rArrgdebugrLrrrrXs z$Configuration._ensure_have_load_onlycCs$i}|jD]}||j|q |S)zs6   $     site-packages/pip/_internal/__pycache__/main.cpython-38.opt-1.pyc000064400000002371151464452710020566 0ustar00U .eO@sdZddlmZddlZddlZddlZddlZddlmZddl m Z ddl m Z ddl mZddlmZeeZd d d ZdS) z Primary application entrypoint. )absolute_importN) autocomplete) parse_command)create_command)PipError) deprecationc Cs|dkrtjdd}ttzt|\}}WnJtk r~}z,tjd|tjt j t dW5d}~XYnXzt t jdWn0t jk r}ztd|W5d}~XYnXt|d|kd}||S)Nz ERROR: %sz%Ignoring error %s when setting localez --isolated)isolated)sysargvrZinstall_warning_loggerrrrstderrwriteoslinesepexitlocale setlocaleLC_ALLErrorloggerdebugrmain)argsZcmd_nameZcmd_argsexceZcommandr6/usr/lib/python3.8/site-packages/pip/_internal/main.pyrs r)N)__doc__Z __future__rrZloggingrr Z pip._internal.cli.autocompletionrZpip._internal.cli.main_parserrZpip._internal.commandsrZpip._internal.exceptionsrZpip._internal.utilsrZ getLogger__name__rrrrrrs       site-packages/pip/_internal/__pycache__/build_env.cpython-38.opt-1.pyc000064400000016416151464452710021616 0ustar00U .e]@sdZddlZddlZddlZddlZddlmZddlmZddl m Z ddl m Z m Z mZddlmZddlmZdd lmZdd lmZdd lmZerdd lmZmZmZmZmZdd l m!Z!e"e#Z$GdddZ%Gddde&Z'Gddde'Z(dS)z;Build Environment used for isolation during sdist building N) OrderedDict)get_python_lib) get_paths) RequirementVersionConflict WorkingSet)__file__)call_subprocess) TempDirectory)MYPY_CHECK_RUNNING) open_spinner)TupleSetIterableOptionalList) PackageFinderc@seZdZddZdS)_PrefixcCsj||_d|_ttjdkrdnd||ddd|_td|d}td|d}||kr\|g|_n ||g|_dS) NFntZ posix_prefix)baseZplatbase)varsZscripts) plat_specificprefixT)pathsetuprosnamebin_dirrlib_dirs)selfrZpurelibZplatlibr ;/usr/lib/python3.8/site-packages/pip/_internal/build_env.py__init__!s   z_Prefix.__init__N)__name__ __module__ __qualname__r"r r r r!rsrc@s@eZdZdZddZddZddZdd Zd d Zd d Z dS)BuildEnvironmentzFCreates and manages an isolated environment to install build deps c stdd_tfdddD_g_g_ttjD] }j |j j |j qBddt dd t d d fD}tjjjd _tjjstjttjjd d "}|tdj|jdW5QRXdS)Nz build-env)Zkindc3s(|] }|ttjjj|fVqdSN)rrrjoin _temp_dir.0rrr r! ;sz,BuildEnvironment.__init__..)ZnormalZoverlaycSsh|]}tj|qSr )rrnormcase)r+siter r r! Isz,BuildEnvironment.__init__..F)rTr/zsitecustomize.pywa import os, site, sys # First, drop system-sites related paths. original_sys_path = sys.path[:] known_paths = set() for path in {system_sites!r}: site.addsitedir(path, known_paths=known_paths) system_paths = set( os.path.normcase(path) for path in sys.path[len(original_sys_path):] ) original_sys_path = [ path for path in original_sys_path if os.path.normcase(path) not in system_paths ] sys.path = original_sys_path # Second, add lib directories. # ensuring .pth file are processed. for path in {lib_dirs!r}: assert not path in sys.path site.addsitedir(path) ) system_sitesr)r r)r _prefixes _bin_dirs _lib_dirsreversedlistvaluesappendrextendrrrrr( _site_direxistsmkdiropenwritetextwrapdedentformat)rrr2fpr r,r!r"7s0    zBuildEnvironment.__init__cCsndddD|_|jdd}|jd}|r>||tj|jg}tjtj |dtj |ddS)NcSsi|]}|tj|dqSr')renvirongetr*r r r! osz.BuildEnvironment.__enter__..)PATHZPYTHONNOUSERSITEZ PYTHONPATHrG1) _save_envr4r:splitrpathsepr;rDupdater()rrZold_pathZ pythonpathr r r! __enter__ns   zBuildEnvironment.__enter__cCs:|jD]*\}}|dkr*tj|dq |tj|<q dSr')rIitemsrrDpop)rexc_typeexc_valexc_tbZvarname old_valuer r r!__exit__szBuildEnvironment.__exit__cCs|jdSr')r)cleanupr,r r r!rUszBuildEnvironment.cleanupc Cst}t}|rt|j}|D]p}z"|t|dkrB||Wqtk r}z*|t|j d t|j dfW5d}~XYqXq||fS)zReturn 2 sets: - conflicting requirements: set of (installed, wanted) reqs tuples - missing requirements: set of reqs Nr) setrr5findrparseaddrstrargsZas_requirement)rZreqsZmissingZ conflictingZwsZreqer r r!check_requirementss  z#BuildEnvironment.check_requirementsc CsV|j|}d|_|sdStjtjtdddd|jdg}t t j krP| ddD]:}t |j|}|d |d d d t|pd hfqT|j} | r|d| dg| ddD]} |d| gqn | d|jD]} |d| gq|jD]} |d| gq|jr| d| d ||t|} t|| dW5QRXdS)NTZinstallz--ignore-installedz --no-userz--prefixz--no-warn-script-locationz-v)Z no_binaryZ only_binaryz--_-,z:none:z-irrVz--extra-index-urlz --no-indexz --find-linksz--trusted-hostz--pre)spinner)r3rsys executablerrdirname pip_locationloggerZgetEffectiveLevelloggingDEBUGr9getattrformat_controlr:replacer(sorted index_urlsZ find_linksZ trusted_hostsZallow_all_prereleasesr r )rfinder requirementsZprefix_as_stringmessagerr\rkZformatsrnZ extra_indexlinkZhostrbr r r!install_requirementssH           z%BuildEnvironment.install_requirementsN) r#r$r%__doc__r"rMrTrUr^rsr r r r!r&3s7r&c@s8eZdZdZddZddZddZdd Zd d Zd S) NoOpBuildEnvironmentz5A no-op drop-in replacement for BuildEnvironment cCsdSr'r r,r r r!r"szNoOpBuildEnvironment.__init__cCsdSr'r r,r r r!rMszNoOpBuildEnvironment.__enter__cCsdSr'r )rrPrQrRr r r!rTszNoOpBuildEnvironment.__exit__cCsdSr'r r,r r r!rUszNoOpBuildEnvironment.cleanupcCs tdSr')NotImplementedError)rrorprrqr r r!rssz)NoOpBuildEnvironment.install_requirementsN) r#r$r%rtr"rMrTrUrsr r r r!rus ru))rtrhrrcr@ collectionsrZdistutils.sysconfigrZ sysconfigrZpip._vendor.pkg_resourcesrrrZpiprrfZpip._internal.utils.subprocessr Zpip._internal.utils.temp_dirr Zpip._internal.utils.typingr Zpip._internal.utils.uir typingr rrrrZpip._internal.indexrZ getLoggerr#rgrobjectr&rur r r r!s*          site-packages/pip/_internal/__pycache__/exceptions.cpython-38.pyc000064400000030302151464452710021057 0ustar00U .e(@sdZddlmZddlmZmZmZddlmZddl m Z e rdddl m Z ddl mZddlmZGd d d eZGd d d eZGd ddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdd d eZGd!d"d"eZGd#d$d$eZGd%d&d&eZ Gd'd(d(e Z!Gd)d*d*e Z"Gd+d,d,e Z#Gd-d.d.e Z$Gd/d0d0e Z%Gd1d2d2eZ&Gd3d4d4eZ'd5S)6z"Exceptions used throughout package)absolute_import)chaingroupbyrepeat) iteritems)MYPY_CHECK_RUNNING)Optional) Distribution)InstallRequirementc@seZdZdZdS)PipErrorzBase pip exceptionN__name__ __module__ __qualname____doc__rrnz$HashErrors.__str__..)keycSs|jSr' __class__r,rrrr.or/css|]}|VqdSr')body.0r-rrr qsz%HashErrors.__str__.. )r(sortrr)headextendjoin)rlinesclsZ errors_of_clsrrrrls zHashErrors.__str__cCs t|jSr')boolr(rrrr __nonzero__uszHashErrors.__nonzero__cCs|Sr')r?rrrr__bool__xszHashErrors.__bool__N) r rrrrr)rr?r@rrrrr&cs  r&c@s0eZdZdZdZdZddZddZdd ZdS) HashErrora A failure to verify a package against known-good hashes :cvar order: An int sorting hash exception classes by difficulty of recovery (lower being harder), so the user doesn't bother fretting about unpinned packages when he has deeper issues, like VCS dependencies, to deal with. Also keeps error reports in a deterministic order. :cvar head: A section heading for display above potentially many exceptions of this kind :ivar req: The InstallRequirement that triggered this error. This is pasted on after the exception is instantiated, because it's not typically available earlier. NcCs d|S)a)Return a summary of me for display under the heading. This default implementation simply prints a description of the triggering requirement. :param req: The InstallRequirement that provoked this error, with populate_link() having already been called z %s)_requirement_namerrrrr3s zHashError.bodycCsd|j|fS)Nz%s %s)r9r3rrrrrszHashError.__str__cCs|jrt|jSdS)zReturn a description of the requirement that triggered me. This default implementation returns long description of the req, with line numbers unknown package)reqstrrrrrrCszHashError._requirement_name) r rrrrEr9r3rrCrrrrrA|s  rAc@seZdZdZdZdZdS)VcsHashUnsupporteduA hash was provided for a version-control-system-based requirement, but we don't have a method for hashing those.rzlCan't verify hashes for these requirements because we don't have a way to hash version control repositories:Nr rrrr+r9rrrrrGsrGc@seZdZdZdZdZdS)DirectoryUrlHashUnsupportedrHzUCan't verify hashes for these file:// requirements because they point to directories:NrIrrrrrJsrJc@s(eZdZdZdZdZddZddZdS) HashMissingz2A hash was needed for a requirement but is absent.awHashes are required in --require-hashes mode, but they are missing from some requirements. Here is a list of those requirements along with the hashes their downloaded archives actually had. Add lines like these to your requirements files to prevent tampering. (If you did not enable --require-hashes manually, note that it turns on automatically when any package has a hash.)cCs ||_dS)zq :param gotten_hash: The hash of the (possibly malicious) archive we just downloaded N) gotten_hash)rrNrrrrszHashMissing.__init__cCsHddlm}d}|jr4|jjr&|jjn t|jdd}d|p.hash_then_orc3s|]}dt|fVqdS)z Expected %s %sN)nextr4prefixrrr6sz0HashMismatch._hash_comparison..z Got %s r7)rrVr:r)rWZ hexdigestr;)rrZr<rYZ expectedsrr\rrXs  zHashMismatch._hash_comparisonN) r rrrr+r9rr3rXrrrrrTs  rTc@seZdZdZdS)UnsupportedPythonVersionzMUnsupported python version according to Requires-Python package metadata.Nr rrrrr^sr^cs*eZdZdZdfdd ZddZZS) !ConfigurationFileCouldNotBeLoadedz=When there are errors while loading a configuration file could not be loadedNcs&tt||||_||_||_dSr')superr_rreasonfnamer*)rrbrcr*r1rrr(sz*ConfigurationFileCouldNotBeLoaded.__init__cCsB|jdk rd|j}n|jdk s&td|jj}d|j|S)Nz in {}.z. {} zConfiguration file {}{})rcrr*AssertionErrormessagerb)rZ message_partrrrr.s  z)ConfigurationFileCouldNotBeLoaded.__str__)r`NN)r rrrrr __classcell__rrr1rr_$sr_N)(rZ __future__r itertoolsrrrZpip._vendor.sixrZpip._internal.utils.typingrtypingrZpip._vendor.pkg_resourcesr Zpip._internal.req.req_installr Exceptionr rrrrrrr r!r"r#r$r%r&rArGrJrLrRrTr^r_rrrrs<      ,  % 7site-packages/pip/_internal/__pycache__/pyproject.cpython-38.pyc000064400000006154151464452710020725 0ustar00U .eZ@sddlmZddlZddlZddlZddlmZmZddlm Z ddl m Z e rhddl m Z mZmZmZddZd d Zd d ZdS) )absolute_importN)pytomlsix)InstallationError)MYPY_CHECK_RUNNING)AnyTupleOptionalListcCst|totdd|DS)Ncss|]}t|tjVqdS)N) isinstancerZ string_types).0itemr;/usr/lib/python3.8/site-packages/pip/_internal/pyproject.py sz"_is_list_of_str..)r listall)objrrr_is_list_of_strs rcCs2tj|d}tjr.t|tjr.|t }|S)Nzpyproject.toml) ospathjoinrZPY2r Z text_typeencodesysgetfilesystemencoding)Zunpacked_source_directoryrrrrmake_pyproject_pathsrc CsVtj|}tj|}|rLtj|dd}t|}W5QRX|d}nd}|rr|sr|dk rl|sltdd}n<|rd|kr|dk r|std |dd}n |dkr|}|dk st |sdS|dkrd d gd d }|dk st d } d|krt| j |dd|d} t | s(t| j |dd|d} g} | dkrLd } d d g} | | | fS)aLoad the pyproject.toml file. Parameters: use_pep517 - Has the user requested PEP 517 processing? None means the user hasn't explicitly specified. pyproject_toml - Location of the project's pyproject.toml file setup_py - Location of the project's setup.py file req_name - The name of the requirement we're processing (for error reporting) Returns: None if we should use the legacy code path, otherwise a tuple ( requirements from pyproject.toml, name of PEP 517 backend, requirements we should check are installed after setting up the build environment ) zutf-8)encodingz build-systemNzIDisabling PEP 517 processing is invalid: project does not have a setup.pyT build-backendzbDisabling PEP 517 processing is invalid: project specifies a build backend of {} in pyproject.tomlzsetuptools>=40.8.0Zwheelz setuptools.build_meta:__legacy__)requiresrzO{package} has a pyproject.toml file that does not comply with PEP 518: {reason}rz]it has a 'build-system' table but not 'build-system.requires' which is mandatory in the table)packagereasonz1'build-system.requires' is not a list of strings.) rrisfileioopenrloadgetrformatAssertionErrorr) Z use_pep517Zpyproject_tomlZsetup_pyZreq_nameZ has_pyprojectZ has_setupfZpp_tomlZ build_systemZerror_templaterZbackendZcheckrrrload_pyproject_toml#sf             r))Z __future__rr"rrZ pip._vendorrrZpip._internal.exceptionsrZpip._internal.utils.typingrtypingrrr r rrr)rrrrs    site-packages/pip/_internal/__pycache__/index.cpython-38.opt-1.pyc000064400000061246151464452710020757 0ustar00U .e@s(dZddlmZddlZddlZddlmZddlmZddl m Z ddl m Z mZmZmZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZddlmZddlm Z ddl!m"Z"ddl#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*e$rddl+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1m2Z2m3Z3ddl m4Z4ddl5m6Z6ddl7m8Z8ddl9m:Z:ddl;mZ>e3e2de2e?e@ffZAe2e?e?e?e4eAe/e?fZBdddgZCeDeEZFd1d!d"ZGGd#d$d$eHZId%d&ZJGd'd(d(eHZKGd)ddeHZLGd*d+d+eHZMGd,ddeHZNd-d.ZOd/d0ZPdS)2z!Routines related to PyPI, indexes)absolute_importN) specifiers)canonicalize_name)parse)BestVersionAlreadyInstalledDistributionNotFoundInvalidWheelFilenameUnsupportedWheel)InstallationCandidate) FormatControl)Link)SelectionPreferences) TargetPython)WHEEL_EXTENSION) indent_log) build_netloc)check_requires_python)MYPY_CHECK_RUNNING)SUPPORTED_EXTENSIONS) url_to_path)Wheel) FrozenSetIterableListOptionalSetTextTupleUnion) _BaseVersion) LinkCollector) SearchScope)InstallRequirement) Pep425Tag)Hashesr BestCandidateResult PackageFinderFcCs~zt|j|d}Wn&tjk r8td|j|YnBX|szdtt|}|shtd||j|dStd||j|dS)aa Return whether the given Python version is compatible with a link's "Requires-Python" value. :param version_info: A 3-tuple of ints representing the Python major-minor-micro version to check. :param ignore_requires_python: Whether to ignore the "Requires-Python" value if the given Python version isn't compatible. ) version_infoz2Ignoring invalid Requires-Python (%r) for link: %s.z4Link requires a different Python (%s not in: %r): %sFzBIgnoring failed Requires-Python check (%s not in: %r) for link: %sT) rZrequires_pythonrZInvalidSpecifierloggerdebugjoinmapstr)linkr(ignore_requires_pythonZ is_compatibleversionr%r%7/usr/lib/python3.8/site-packages/pip/_internal/index.py_check_link_requires_python;s8  r3c@s,eZdZdZedZdddZddZdS) LinkEvaluatorzD Responsible for evaluating links for a particular project. z-py([123]\.?[0-9]?)$NcCs4|dkr d}||_||_||_||_||_||_dS)a :param project_name: The user supplied package name. :param canonical_name: The canonical package name. :param formats: The formats allowed for this package. Should be a set with 'binary' or 'source' or both in it. :param target_python: The target Python interpreter to use when evaluating link compatibility. This is used, for example, to check wheel compatibility, as well as when checking the Python version, e.g. the Python version embedded in a link filename (or egg fragment) and against an HTML link's optional PEP 503 "data-requires-python" attribute. :param allow_yanked: Whether files marked as yanked (in the sense of PEP 592) are permitted to be candidates for install. :param ignore_requires_python: Whether to ignore incompatible PEP 503 "data-requires-python" values in HTML links. Defaults to False. NF) _allow_yanked_canonical_name_ignore_requires_python_formats_target_python project_name)selfr:canonical_nameformats target_python allow_yankedr0r%r%r2__init__rszLinkEvaluator.__init__c Csd}|jr(|js(|jpd}dd|fS|jr<|j}|j}n|\}}|sPdS|tkrddd|fSd|jkr|t krd|j }d|fSd |j kr|d krd S|t kr,zt |j }Wntk rYd SXt|j|jkrd |j }d|fS|j}||s&|}dd|}d|fS|j}d|jkrP|t krPdd|j fS|sbt||j}|svdd|j fS|j|} | r|d| }| d} | |jjkrdSt||jj|j d} | sdSt!"d||d|fS)aG Determine whether a link is a candidate for installation. :return: A tuple (is_candidate, result), where `result` is (1) a version string if `is_candidate` is True, and (2) if `is_candidate` is False, an optional string to log the reason the link fails to qualify. N Fzyanked for reason: {})Fz not a filezunsupported archive format: %sZbinaryzNo binaries permitted for %sZmacosx10z.zip)Fz macosx10 one)Fzinvalid wheel filenamezwrong project name (not %s)z"none of the wheel's tags match: {}, sourcezNo sources permitted for %szMissing project version for %s)FzPython version is incorrect)r(r0)FNzFound link %s, version: %sT)# is_yankedr5 yanked_reasonformat egg_fragmentextsplitextrr8rr:pathrfilenamerrnamer6r9get_tags supportedZget_formatted_file_tagsr,r1_extract_version_from_fragment_py_version_researchstartgroup py_versionr3Zpy_version_infor7r*r+) r;r/r1reasonZegg_inforIwheelsupported_tagsZ file_tagsmatchrUZsupports_pythonr%r%r2 evaluate_linksx            zLinkEvaluator.evaluate_link)N) __name__ __module__ __qualname____doc__recompilerQr@rZr%r%r%r2r4fs  &r4c Cs|stdt||t|Sg}g}d}|D]>}|j}|js@n"|j|drV|d7}n ||q.||q.|rx|}nt|}t|t|krd} n dt|d dd |D} td t|||j |t||| |S) a Filter out candidates whose hashes aren't allowed, and return a new list of candidates. If at least one candidate has an allowed hash, then all candidates with either an allowed hash or no hash specified are returned. Otherwise, the given candidates are returned. Including the candidates with no hash specified when there is a match allows a warning to be logged if there is a more preferred candidate with no hash specified. Returning all candidates in the case of no matches lets pip report the hash of the candidate that would otherwise have been installed (e.g. permitting the user to more easily update their requirements file with the desired hash). zJGiven no hashes to check %s links for project %r: discarding no candidatesr)hashesrDzdiscarding no candidateszdiscarding {} non-matches: {}z css|]}t|jVqdSN)r.r/.0 candidater%r%r2 +sz*filter_unallowed_hashes..zPChecked %s links for project %r against %s hashes (%s matches, %s no digest): %s) r*r+lenlistr/Zhas_hashis_hash_allowedappendrGr,Z digest_count) candidatesrar:Zmatches_or_no_digestZ non_matchesZ match_countrer/ZfilteredZdiscard_messager%r%r2filter_unallowed_hashessL      rlc@seZdZdZdddZdS)CandidatePreferenceszk Encapsulates some of the preferences for filtering and sorting InstallationCandidate objects. FcCs||_||_dS)zR :param allow_all_prereleases: Whether to allow all pre-releases. N)allow_all_prereleases prefer_binary)r;rornr%r%r2r@Cs zCandidatePreferences.__init__N)FF)r[r\r]r^r@r%r%r%r2rm<srmc@s(eZdZdZddZddZddZdS) r&zA collection of candidates, returned by `PackageFinder.find_best_candidate`. This class is only intended to be instantiated by CandidateEvaluator's `compute_best_candidate()` method. cCs |dkr n||_||_||_dS)a :param candidates: A sequence of all available candidates found. :param applicable_candidates: The applicable candidates. :param best_candidate: The most preferred candidate found, or None if no applicable candidates were found. N)_applicable_candidates _candidatesbest_candidater;rkapplicable_candidatesrrr%r%r2r@Ws zBestCandidateResult.__init__cCs t|jS)z(Iterate through all candidates. )iterrqr;r%r%r2iter_allpszBestCandidateResult.iter_allcCs t|jS)z3Iterate through the applicable candidates. )rurprvr%r%r2iter_applicablevsz#BestCandidateResult.iter_applicableN)r[r\r]r^r@rwrxr%r%r%r2r&Psc@sHeZdZdZedddZdddZdd Zd d Zd d Z ddZ dS)CandidateEvaluatorzm Responsible for filtering and sorting candidates for installation based on what tags are valid. NFcCs:|dkrt}|dkrt}|}|||||||dS)aCreate a CandidateEvaluator object. :param target_python: The target Python interpreter to use when checking compatibility. If None (the default), a TargetPython object will be constructed from the running Python. :param specifier: An optional object implementing `filter` (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable versions. :param hashes: An optional collection of allowed hashes. N)r:rX specifierrornra)rrZ SpecifierSetrN)clsr:r>rornrzrarXr%r%r2createszCandidateEvaluator.createcCs(||_||_||_||_||_||_dS)z :param supported_tags: The PEP 425 tags supported by the target Python in order of preference (most preferred first). N)_allow_all_prereleases_hashes_prefer_binary _project_name _specifier_supported_tags)r;r:rXrzrornrar%r%r2r@s zCandidateEvaluator.__init__csV|jpd}|j}dd|jdd|D|dDfdd|D}t||j|jd S) zM Return the applicable candidates from a list of candidates. NcSsh|] }t|qSr%)r.)rdvr%r%r2 sz?CandidateEvaluator.get_applicable_candidates..css|]}t|jVqdSrbr.r1rdcr%r%r2rfsz?CandidateEvaluator.get_applicable_candidates..)Z prereleasescsg|]}t|jkr|qSr%rrZversionsr%r2 sz@CandidateEvaluator.get_applicable_candidates..)rkrar:)r}rfilterrlr~r)r;rkZallow_prereleasesrzrtr%rr2get_applicable_candidatess   z,CandidateEvaluator.get_applicable_candidatesc Cs|j}t|}d}d}|j}|jrt|j}||sDtd|j|jrNd}| | }|j dk rt d|j } | } t| d| df}n| }t||j} dt|j} | | ||j||fS)a) Function to pass as the `key` argument to a call to sorted() to sort InstallationCandidates by preference. Returns a tuple such that tuples sorting as greater using Python's default comparison operator are more preferred. The preference is as follows: First and foremost, candidates with allowed (matching) hashes are always preferred over candidates without matching hashes. This is because e.g. if the only candidate with an allowed hash is yanked, we still want to use that candidate. Second, excepting hash considerations, candidates that have been yanked (in the sense of PEP 592) are always less preferred than candidates that haven't been yanked. Then: If not finding wheels, they are sorted by version only. If finding wheels, then the sort order is by version, then: 1. existing installs 2. wheels ordered via Wheel.support_index_min(self._supported_tags) 3. source archives If prefer_binary was set, then all wheels are sorted above sources. Note: it was considered to embed this logic into the Link comparison operators, but then different sdist links with the same version, would have to be considered equal r%rzB%s is not a supported wheel for this platform. It can't be sorted.rDNz ^(\d+)(.*)$)rrgr/Zis_wheelrrLrOr rZsupport_index_min build_tagr_rYgroupsintrir~rEr1) r;reZ valid_tagsZ support_numrZbinary_preferencer/rWZprirYZbuild_tag_groupsZhas_allowed_hashZ yank_valuer%r%r2 _sort_keys<    zCandidateEvaluator._sort_keycCsH|sdSt||jd}|j}|jrD|jp*d}dj||d}t||S)zy Return the best candidate per the instance's sort order, or None if no candidate is acceptable. NkeyrAzqThe candidate selected for download or install is a yanked version: {candidate} Reason for being yanked: {reason})rerV)maxrr/rErFrGr*Zwarning)r;rkrrr/rVmsgr%r%r2sort_best_candidates   z&CandidateEvaluator.sort_best_candidatecCs"||}||}t|||dS)zF Compute and return a `BestCandidateResult` instance. )rtrr)rrr&rsr%r%r2compute_best_candidate<s  z)CandidateEvaluator.compute_best_candidate)NFFNN)FFN) r[r\r]r^ classmethodr|r@rrrrr%r%r%r2ry}s  ) $<ryc@seZdZdZd&ddZed'ddZeddZej d dZed d Z ed d Z eddZ eddZ ddZddZddZddZddZddZddZd(d d!Zd)d"d#Zd$d%ZdS)*r'zThis finds packages. This is meant to match easy_install's technique for looking for packages, by reading pages and looking for appropriate links. NcCsP|dkrt}|pttt}||_||_||_||_||_||_t|_ dS)a This constructor is primarily meant to be used by the create() class method and from tests. :param format_control: A FormatControl object, used to control the selection of source packages / binary packages when consulting the index and links. :param candidate_prefs: Options to use when creating a CandidateEvaluator object. N) rmr setr5_candidate_prefsr7_link_collectorr9format_control _logged_links)r;link_collectorr>r?rcandidate_prefsr0r%r%r2r@VszPackageFinder.__init__cCs8|dkrt}t|j|jd}|||||j|j|jdS)afCreate a PackageFinder. :param selection_prefs: The candidate selection preferences, as a SelectionPreferences object. :param target_python: The target Python interpreter to use when checking compatibility. If None (the default), a TargetPython object will be constructed from the running Python. N)rorn)rrr>r?rr0)rrmrornr?rr0)r{rZselection_prefsr>rr%r%r2r|~szPackageFinder.createcCs|jjSrbr search_scopervr%r%r2rszPackageFinder.search_scopecCs ||j_dSrbr)r;rr%r%r2rscCs|jjSrb)r find_linksrvr%r%r2rszPackageFinder.find_linkscCs|jjSrb)r index_urlsrvr%r%r2rszPackageFinder.index_urlsccs|jjjD]}t|Vq dSrb)rZsessionZpip_trusted_originsr)r;Z host_portr%r%r2 trusted_hostsszPackageFinder.trusted_hostscCs|jjSrbrrnrvr%r%r2rnsz#PackageFinder.allow_all_prereleasescCs d|j_dS)NTrrvr%r%r2set_allow_all_prereleasessz'PackageFinder.set_allow_all_prereleasescCs.t|}|j|}t||||j|j|jdS)N)r:r<r=r>r?r0)rrZget_allowed_formatsr4r9r5r7)r;r:r<r=r%r%r2make_link_evaluators z!PackageFinder.make_link_evaluatorcCsPgg}}t}|D]2}||kr|||jr<||q||q||S)z Returns elements of links in order, non-egg links first, egg links second, while eliminating duplicates )raddrHrj)r;linksZeggsZno_eggsseenr/r%r%r2 _sort_linkss    zPackageFinder._sort_linkscCs(||jkr$td|||j|dS)NzSkipping link: %s: %s)rr*r+r)r;r/rVr%r%r2_log_skipped_links zPackageFinder._log_skipped_linkcCs<||\}}|s(|r$|j||ddSt|j|t|dS)z If the link is a candidate for install, convert it to an InstallationCandidate and return it. Otherwise, return None. )rVN)Zprojectr/r1)rZrr r:r.)r;link_evaluatorr/Z is_candidateresultr%r%r2get_install_candidatesz#PackageFinder.get_install_candidatecCs6g}||D]"}|||}|dk r||q|S)zU Convert links that are candidates to InstallationCandidate objects. N)rrrj)r;rrrkr/rer%r%r2evaluate_linkss   zPackageFinder.evaluate_linksc Cs|j|}||}|j||jd}g}|jD]>\}}td|t |j||d}| |W5QRXq4|j||j d} | r| j ddtdd dd| D| ||S) aFind all available InstallationCandidate for project_name This checks index_urls and find_links. All versions found are returned as an InstallationCandidate list. See LinkEvaluator.evaluate_link() for details on which files are accepted. )rzAnalyzing links from page %sT)reversezLocal files found: %srBcSsg|]}t|jjqSr%)rr/Zurlrcr%r%r2r2sz5PackageFinder.find_all_candidates..)rZ collect_linksrrrZpagesitemsr*r+rextendfilessortr,) r;r:Zcollected_linksrZfind_links_versionsZ page_versionsZpage_urlZ page_linksZ new_versionsZ file_versionsr%r%r2find_all_candidates s8      z!PackageFinder.find_all_candidatescCs"|j}tj||j|j|j||dS)z3Create a CandidateEvaluator object to use. )r:r>rornrzra)rryr|r9rorn)r;r:rzrarr%r%r2make_candidate_evaluator;s z&PackageFinder.make_candidate_evaluatorcCs$||}|j|||d}||S)aFind matches for the given project and specifier. :param specifier: An optional object implementing `filter` (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable versions. :return: A `BestCandidateResult` instance. )r:rzra)rrr)r;r:rzrarkZcandidate_evaluatorr%r%r2find_best_candidateNs z!PackageFinder.find_best_candidatec Cs|jdd}|j|j|j|d}|j}d}|jdk r@t|jj}dd}|dkrz|dkrzt d||| t d|d}|r|dks|j|krd }|s|dk r|rt d |nt d ||jdS|rt d ||| tt d |j|| |jS)zTry to find a Link matching req Expects req, an InstallRequirement and upgrade, a boolean Returns a Link if found, Raises DistributionNotFound or BestVersionAlreadyInstalled otherwise F)Ztrust_internet)rzraNcSs dtdd|DtdpdS)NrBcSsh|]}t|jqSr%rrr%r%r2r}szKPackageFinder.find_requirement.._format_versions..rZnone)r,sorted parse_version)Z cand_iterr%r%r2_format_versionsws  z8PackageFinder.find_requirement.._format_versionszNCould not find a version that satisfies the requirement %s (from versions: %s)z%No matching distribution found for %sTzLExisting installed version (%s) is most up-to-date and satisfies requirementzUExisting installed version (%s) satisfies requirement (most up-to-date version is %s)z=Installed version (%s) is most up-to-date (past versions: %s)z)Using version %s (newest of versions: %s))rarrMrzrrZ satisfied_byrr1r*Zcriticalrwrr+rxrr/) r;ZreqZupgraderaZbest_candidate_resultrrZinstalled_versionrZbest_installedr%r%r2find_requirementesh        zPackageFinder.find_requirement)NNN)N)NN)NN)r[r\r]r^r@rr|propertyrsetterrrrrnrrrrrrrrrrr%r%r%r2r'OsD  (         1  cCsLt|D].\}}|dkrqt|d||kr|Sqtd||dS)aFind the separator's index based on the package's canonical name. :param fragment: A + filename "fragment" (stem) or egg fragment. :param canonical_name: The package's canonical name. This function is needed since the canonicalized name does not necessarily have the same length as the egg info's name part. An example:: >>> fragment = 'foo__bar-1.0' >>> canonical_name = 'foo-bar' >>> _find_name_version_sep(fragment, canonical_name) 8 -Nz{} does not match {}) enumerater ValueErrorrG)fragmentr<irr%r%r2_find_name_version_seps  rcCsBzt||d}Wntk r(YdSX||d}|s>dS|S)zParse the version string from a + filename "fragment" (stem) or egg fragment. :param fragment: The string to parse. E.g. foo-2.1 :param canonical_name: The canonicalized name of the package this belongs to. rDN)rr)rr<Z version_startr1r%r%r2rPs  rP)F)Qr^Z __future__rZloggingr_Zpip._vendor.packagingrZpip._vendor.packaging.utilsrZpip._vendor.packaging.versionrrZpip._internal.exceptionsrrrr Zpip._internal.models.candidater Z#pip._internal.models.format_controlr Zpip._internal.models.linkr Z$pip._internal.models.selection_prefsr Z"pip._internal.models.target_pythonrZpip._internal.utils.filetypesrZpip._internal.utils.loggingrZpip._internal.utils.miscrZpip._internal.utils.packagingrZpip._internal.utils.typingrZpip._internal.utils.unpackingrZpip._internal.utils.urlsrZpip._internal.wheelrtypingrrrrrrrrrZpip._internal.collectorr Z!pip._internal.models.search_scoper!Zpip._internal.reqr"Zpip._internal.pep425tagsr#Zpip._internal.utils.hashesr$rr.ZBuildTagZCandidateSortingKey__all__Z getLoggerr[r*r3objectr4rlrmr&ryr'rrPr%r%r%r2s^                 (         + K-Shsite-packages/pip/_internal/__pycache__/pep425tags.cpython-38.opt-1.pyc000064400000023072151464452710021541 0ustar00U .eE>@sTdZddlmZddlZddlZddlZddlZddlZddl Z ddl Z ddl m Z ddl ZddlmZddlmZerddlmZmZmZmZmZmZmZeeeefZeeZe dZ!d d Z"d d Z#d dZ$ddZ%ddZ&ddZ'd-ddZ(ddZ)ddZ*ddZ+ddZ,d d!Z-d"d#Z.d$d%Z/d&d'Z0d(d)Z1d.d+d,Z2e'Z3dS)/z2Generate and work with PEP 425 Compatibility Tags.)absolute_importN) OrderedDict)get_extension_suffixes)MYPY_CHECK_RUNNING)TupleCallableListOptionalUnionDictSetz(.+)_(\d+)_(\d+)_(.+)c CsLz t|WStk rF}ztd|tWYdSd}~XYnXdS)Nz{}) sysconfigget_config_varIOErrorwarningswarnformatRuntimeWarning)varerConfig variable '%s' is unset, Python ABI tag may be incorrect)rloggerdebug)rZfallbackexpectedrvalrrrget_flagYsr2cstd}t}d}|s|dkrttdrd}d}d}|dktddd d rRd }tjd krvtd fdd d rvd}tjdkrtddd ddrd}d|t|||f}n:|r|drd|dd}n|r| dd dd}|S)zXReturn the ABI tag based on SOABI (if available) or emulate SOABI (CPython 2, PyPy).ZSOABIN>rr maxunicoder!rZPy_DEBUGcSs ttdS)NZgettotalrefcount)rrrrrrtzget_abi_tag..)rd)Z WITH_PYMALLOCcsS)NrrZ is_cpythonrrr4xr5m)r7r7ZPy_UNICODE_SIZEcSs tjdkS)Ni)rr3rrrrr4{r5)r0ruz %s%s%s%s%szcpython--r*._) rr rrr2r&r)rsplitreplace)Zsoabiimplabir6r:r<rr9r get_abi_tagfsB   rDcCs tjdkS)Ni)rmaxsizerrrr_is_running_32bitsrFcCstjdkr^t\}}}|d}|dkr6tr6d}n|dkrHtrHd}d|d|d |Stj dd  d d }|d krtrd }|S)z0Return our platform name 'win32', 'linux_x86_64'darwinr>x86_64i386ppc64ppczmacosx_{}_{}_{}rr*r?r= linux_x86_64 linux_i686) rrZmac_verr@rFr distutilsutil get_platformrA)releaser?machineZ split_verresultrrrrPs  rPc Cs$tdkrdSz&ttjd}|d}W5QRXWntttfk rPYdSX|dksft|dkrjdSt |t rdd|D}ndd|D}|dd d d d d gk}||d ddgkM}||dddgkM}||ddddgkM}||dddgkM}||dddd @d kM}|S)N linux_armv7lFrb(cSsg|] }t|qSr)ord).0crrr sz"is_linux_armhf..cSsg|]}|qSrr)rXbrrrrZsrr;ELFr*'%&) rPopenr executablereadrOSError TypeErrorlen isinstancer%)fZelf_header_rawZ elf_headerrSrrris_linux_armhfs&  roc CsNtdkrdSzddl}t|jWSttfk r:YnXtjjj ddS)NrLrMFrr"r`) rP _manylinuxboolZmanylinux1_compatible ImportErrorAttributeErrorpip _internalutilsglibchave_compatible_glibcrqrrris_manylinux1_compatibles  r{c CsNtdkrdSzddl}t|jWSttfk r:YnXtjjj ddS)NrpFrr" ) rPrqrrZmanylinux2010_compatiblersrtrurvrwrxryrzrrris_manylinux2010_compatibles  r}c Csdt}|dkrdS|dkr$ts$dSzddl}t|jWSttfk rPYnXtjj j ddS)N> linux_s390xrMrL linux_ppc64lerT linux_aarch64 linux_ppc64FrTrr") rProrqrrZmanylinux2014_compatiblersrtrurvrwrxry)rrqrrris_manylinux2014_compatibles rcsrg}fddtddddg|||r8||D]&}||kr<|||r<||q<|d|S)zReturn a list of supported arches (including group arches) for the given major, minor and machine architecture of an macOS machine. cs||dkr||fdkS|dkr(||fdkS|dkr<||fdkS|dkrP||fdkS|krx|D]}|||r`dSq`dS) NrK) r`rJrI)rr;rHTFr)r+r,archgarch_supports_archgroupsrrrs      z)get_darwin_arches.._supports_arch)Zfat)rIrK)Zintel)rHrI)Zfat64)rHrJ)Zfat32)rHrIrKZ universal)rappend)r+r,rRarchesrrrrget_darwin_archess%    rc CsFg}|dd}t|dddD] }|dtt||fq |S)Nr!)rangerr#r$r%)r&versionsr+r,rrr!get_all_minor_versions_as_strings>s  rFcCsDg}|dkrt}t|}|p"t}g}|p0t}|rD|g|dd<t}tD]$} | drP|| dddqP| t t || d|s|pt } | d\} } } | d r>t| }|r6|\}}}}d ||}g}ttt|dD]0}tt|||D]}| |||fqqn| g}n| d kr~| g}| d kr| d | | | d| | n| d kr| d| | g}nn|dkrg}tr| d | | tr| d | | tr| d| | | | n| g}|D].}|D]"} | d||df|| fqq |ddD]F}|dkr\q|D]*}|D]} | d||f|| fqhq`qH|D]"} | d|ddd| fq| d||dfddf| d||ddfddft|D]B\}}| d|fddf|dkr| d|dddfq|S)acReturn a list of supported tags for each version specified in `versions`. :param versions: a list of string versions, of the form ["33", "32"], or None. The first version will be assumed to support our ABI. :param platform: specify the exact platform you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. :param abi: specify the exact abi you want valid tags for, or None. If None, use the local interpreter abi. Nrz.abir>r"r*Znoner?Zmacosxz {}_{}_%i_%sZ manylinux2014>i686rHZ manylinux2010Z manylinux1z%s%s>3130zpy%sany)r(rr rDsetrraddr@extendsortedlistrrP partition _osx_arch_patmatchrrreversedrintrrr}r{ enumerate)rZnoarchrrBrCZ supportedr&ZabisZabi3ssuffixrZ arch_prefixZarch_sepZ arch_suffixrnamer+r,Z actual_archZtplrr:aversionirrr get_supportedHs~              $ $   r)TT)NFNNN)4__doc__Z __future__rZdistutils.utilrNZloggingrrerr r collectionsrZpip._internal.utils.glibcruZpip._internal.utils.compatrZpip._internal.utils.typingrtypingrrrr r r r r%Z Pep425TagZ getLogger__name__r.compilerrr r'r)r(r-r2rDrFrPror{r}rrrrZimplementation_tagrrrrsP    $     !?  ysite-packages/pip/_internal/__pycache__/cache.cpython-38.pyc000064400000016045151464452710017751 0ustar00U .e @sdZddlZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddl mZddlmZdd lmZmZerdd lmZmZmZmZdd lmZdd lmZeeZGd dde Z!Gddde!Z"Gddde"Z#Gddde!Z$dS)zCache Management N)canonicalize_name)Link) expanduser) TempDirectory)MYPY_CHECK_RUNNING) path_to_url)InvalidWheelFilenameWheel)OptionalSetListAny) FormatControl) Pep425TagcsPeZdZdZfddZddZddZdd Zd d Zd d Z ddZ Z S)CacheaAn abstract class - provides cache directories for data from links :param cache_dir: The root of the cache. :param format_control: An object of FormatControl class to limit binaries being read from the cache. :param allowed_formats: which formats of files the cache should store. ('binary' and 'source' are the only allowed values) csLtt||rt|nd|_||_||_ddh}|j||ksHtdS)Nsourcebinary) superr__init__r cache_dirformat_controlallowed_formatsunionAssertionError)selfrrrZ_valid_formats __class__7/usr/lib/python3.8/site-packages/pip/_internal/cache.pyr(s zCache.__init__cCs|jg}|jdk r4|jdk r4|d|j|jgd|}t|}|dd|dd|dd|ddg}|S)zEGet parts of part that must be os.path.joined with cache_dir N=#) Zurl_without_fragmentZ hash_namehashappendjoinhashlibZsha224encodeZ hexdigest)rlinkZ key_partsZkey_urlZhashedpartsrrr_get_cache_path_parts2s ,zCache._get_cache_path_partsc Cs|j p| p| }|rgSt|}|j|}|j|s@gS||}z t|WSt k r}z$|j t j t j hkrgWYSW5d}~XYnXdSN) rrrZget_allowed_formatsr intersectionget_path_for_linkoslistdirOSErrorerrnoZENOENTZENOTDIR)rr) package_nameZ can_not_cacheZcanonical_nameZformatsrooterrrrr_get_candidatesLs(    zCache._get_candidatescCs tdS)z>Return a directory to store cached items in for link. NNotImplementedErrorrr)rrrr.eszCache.get_path_for_linkcCs tdS)zaReturns a link to a cached item if it exists, otherwise returns the passed link. Nr7)rr)r3supported_tagsrrrgetks z Cache.getcCs$||}tj||}tt|Sr,)r.r/pathr&rr)rr) candidater4r<rrr_link_for_candidatews zCache._link_for_candidatecCsdSr,rrrrrcleanup~sz Cache.cleanup) __name__ __module__ __qualname____doc__rr+r6r.r;r>r@ __classcell__rrrrrs  rcs0eZdZdZfddZddZddZZS)SimpleWheelCachez+A cache of wheels for future installs. cstt|||dhdSNr)rrFrrrrrrrrs  zSimpleWheelCache.__init__cCs ||}tjj|jdf|S)aReturn a directory to store cached wheels for link Because there are M wheels for any one sdist, we provide a directory to cache them in, and then consult that directory when looking up cache hits. We only insert things into the cache if they have plausible version numbers, so that we don't contaminate the cache with things that were not unique. E.g. ./package might have dozens of installs done for it and build a version of 0.0...and if we built and cached a wheel, we'd end up using the same wheel even if the source has been edited. :param link: The link of the sdist for which this will cache wheels. Zwheels)r+r/r<r&r)rr)r*rrrr.s z"SimpleWheelCache.get_path_for_linkc Csxg}|||D]J}z t|}Wntk r8YqYnX||sFq||||fq|sd|S||t|dS)N)r6r rZ supportedr%Zsupport_index_minr>min)rr)r3r:Z candidatesZ wheel_nameZwheelrrrr;s    zSimpleWheelCache.get)rArBrCrDrr.r;rErrrrrFs rFcs(eZdZdZfddZddZZS)EphemWheelCachezGA SimpleWheelCache that creates it's own temporary cache directory cs&tdd|_tt||jj|dS)Nzephem-wheel-cache)Zkind)r _temp_dirrrKrr<)rrrrrrs   zEphemWheelCache.__init__cCs|jdSr,)rLr@r?rrrr@szEphemWheelCache.cleanup)rArBrCrDrr@rErrrrrKs rKcs@eZdZdZfddZddZddZdd Zd d ZZ S) WheelCachezWraps EphemWheelCache and SimpleWheelCache into a single Cache This Cache allows for gracefully degradation, using the ephem wheel cache when a certain link is not found in the simple wheel cache first. cs0tt|||dht|||_t||_dSrG)rrMrrF _wheel_cacherK _ephem_cacherHrrrrs  zWheelCache.__init__cCs |j|Sr,)rNr.r9rrrr.szWheelCache.get_path_for_linkcCs |j|Sr,)rOr.r9rrrget_ephem_path_for_linksz"WheelCache.get_ephem_path_for_linkcCs0|jj|||d}||k r|S|jj|||dS)N)r)r3r:)rNr;rO)rr)r3r:Zretvalrrrr;szWheelCache.getcCs|j|jdSr,)rNr@rOr?rrrr@s zWheelCache.cleanup) rArBrCrDrr.rPr;r@rErrrrrMs  rM)%rDr2r'Zloggingr/Zpip._vendor.packaging.utilsrZpip._internal.models.linkrZpip._internal.utils.compatrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.urlsrZpip._internal.wheelrr typingr r r r Zpip._internal.indexrZpip._internal.pep425tagsrZ getLoggerrAZloggerobjectrrFrKrMrrrrs(         f:site-packages/pip/_internal/__pycache__/wheel.cpython-38.opt-1.pyc000064400000066275151464452710020763 0ustar00U .eH@s4dZddlmZddlZddlZddlZddlZddlZddlZ ddl Z ddl Z ddl Z ddl Z ddlZddlmZddlmZddlmZddlmZddlmZdd lmZdd lmZdd lmZdd lm Z m!Z!m"Z"dd l#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-m.Z.m/Z/ddl0m1Z1ddl2m3Z3m4Z4m5Z5m6Z6ddl7m8Z8ddl9m:Z:ddl;mZ>ddl?m@Z@e:rddlAmBZBmCZCmDZDmEZEmFZFmGZGmHZHmIZImJZJmKZKmLZLmMZMddlNmOZOddlPmQZQddlRmSZSddlTmUZUddlVmWZWeGeXdfZYeLeQgeZfZ[d Z\e]e^Z_d!d"Z`dXd$d%ZadYd&d'Zbd(d)Zcd*d+Zdd,d-Zee fd.e jgZhd/d0Zid1d2Zjd3d4Zkd5d6Zld7d8ZmGd9d:d:enZod;d<ZpGd=d>d>eZqdZdAdBZrdCdDZsdEdFZtdGdHZuGdIdJdJevZwe fdKe jxfdLdMZydNdOZzdPdQZ{dRdSZ|dTdUZ}GdVdWdWevZ~dS)[zH Support for installing and building the "wheel" binary package format. )absolute_importN)urlsafe_b64encode)Parser) pkg_resources) ScriptMaker)get_export_entry)canonicalize_name)StringIO) pep425tags)InstallationErrorInvalidWheelFilenameUnsupportedWheel)distutils_schemeget_major_minor_version)Link) indent_log)has_delete_marker_file)captured_stdout ensure_dir read_chunks)make_setuptools_shim_args) LOG_DIVIDERcall_subprocessformat_command_argsrunner_with_spinner_message) TempDirectory)MYPY_CHECK_RUNNING) open_spinner) unpack_file) path_to_url) DictListOptionalSequenceMappingTupleIOTextAnyIterableCallableSet) Requirement)InstallRequirement)RequirementPreparer) WheelCache) Pep425Tag.)rcCstj||tjjdS)N/)ospathrelpathreplacesep)srcpr:7/usr/lib/python3.8/site-packages/pip/_internal/wheel.pynormpathOsr<c CsRt}d}t|d.}t||dD]}|t|7}||q$W5QRX||fS)z5Return (hash, length) for path using hashlib.sha256()rrb)size)hashlibZsha256openrlenupdate)r4 blocksizehlengthfblockr:r:r; hash_fileSs  rIcCs6t||\}}dt|dd}|t|fS)z?Return (encoded_digest, length) for path using hashlib.sha256()zsha256=latin1=)rIrdigestdecoderstripstr)r4rDrErFrLr:r:r;rehash_srPcCs6tjddkri}d}n ddi}d}t|||f|S)Nrbnewline)sys version_inforA)namemodenlbinr:r:r; open_for_csvjs r[cCs|d}||d<d|S)zBReplace the Python tag in a wheel file name with a new value. -)splitjoin)Z wheelnameZnew_tagpartsr:r:r;replace_python_tagus rac Cstj|rt|dR}|}|ds8W5QRdStjt }d|tj d}| }W5QRXt|d}| || |W5QRXdSdS) zLReplace #!python with #!/path/to/python Return True if file was changed.r>s#!pythonFs#!asciiwbTN) r3r4isfilerAreadline startswithrU executableencodegetfilesystemencodinglinesepreadwrite)r4Zscript firstlineZexenamerestr:r:r; fix_script~s     rozX^(?P(?P.+?)(-(?P.+?))?) \.dist-info$c Cs|dd}t|D]p}t|}|r|d|krttj||d6}|D]*}| }|dkrPW5QRdSqPW5QRXqdS)zP Return True if the extracted wheel in wheeldir should go into purelib. r\_rWWHEELzroot-is-purelib: trueTF) r6r3listdir dist_info_rematchgrouprAr4r_lowerrN)rWwheeldirZ name_foldeditemrtwheelliner:r:r;root_is_purelibs    r{c stj|siifSt|8}t}|D]}|||dq(|dW5QRXtj |}| di}| di}ddt fdd| D}t fd d| D}||fS) N rZconsole_scriptsZ gui_scriptscSst|dddS)zRget the string representation of EntryPoint, remove space and split on '=' rTrK)rOr6r^)sr:r:r; _split_epsz"get_entrypoints.._split_epc3s|]}|VqdSNr:.0vrr:r; sz"get_entrypoints..c3s|]}|VqdSrr:rrr:r;rs)r3r4existsrAr rlstripseekrZ EntryPointZ parse_mapgetdictvalues)filenamefpdatarzZ entry_pointsconsoleguir:rr;get_entrypointss      rc sJ|sdStt}|D]*}tj|}tj|}|||qddtj dd tj D tj tjtjfdd|D}|sdSg}|D]b\}}t|}t|dkrd |d } n$d d |dd d|d } | d| |qd} t|dkr0| | dn| | dd|S)zDetermine if any scripts are not on PATH and format a warning. Returns a warning message if one or more scripts are not on PATH, otherwise None. NcSs g|]}tj|tjqSr:)r3r4normcaserNr7)rir:r:r; sz5message_about_scripts_not_on_PATH..PATHrTcs&i|]\}}tj|kr||qSr:)r3r4r)r parent_dirscriptsZ not_warn_dirsr:r; sz5message_about_scripts_not_on_PATH..r1z script {} isrzscripts {} are, z and z.The {} installed in '{}' which is not on PATH.zeConsider adding {} to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.zthis directoryzthese directoriesr|) collections defaultdictsetr3r4dirnamebasenameaddenvironrr^pathsepappendrrUrgitemssortedrBformatr_) rZgrouped_by_dirdestfilerZ script_nameZwarn_for msg_linesZ dir_scriptsZsorted_scriptsZ start_textZ last_line_fmtr:rr;!message_about_scripts_not_on_PATHsH     rcCst|dddS)a; Return the given rows of a RECORD file in sorted order. Each row is a 3-tuple (path, hash, size) and corresponds to a record of a RECORD file (see PEP 376 and PEP 427 for details). For the rows passed to this function, the size can be an integer as an int or string, or the empty string. cSstdd|DS)Ncss|]}t|VqdSr)rO)rxr:r:r;rsz3sorted_outrows....)tuple)rowr:r:r;z sorted_outrows..)key)r)outrowsr:r:r;sorted_outrowssrc Csg}|D]v}t|dkr(td|t|}|d}|||}||d<||krpt|\} } | |d<| |d<|t|q|D]*} t| \} } |t | || t | fq|D]} ||| ddfq|S)z_ :param installed: A map from archive RECORD path to installation RECORD path. rQz,RECORD line has more than three elements: {}rr1rT) rBloggerwarningrlistpoprPrrr<rO) Z old_csv_rows installedchanged generatedlib_dirZinstalled_rowsrZold_pathnew_pathrLrFrGr:r:r;get_csv_rows_for_installeds*     rc@s eZdZdS)MissingCallableSuffixN)__name__ __module__ __qualname__r:r:r:r;r=srcCs*t|}|dk r&|jdkr&tt|dSr)rsuffixrrO) specificationentryr:r:r;_raise_for_invalid_entrypointAsrcseZdZdfdd ZZS)PipScriptMakerNcst|tt|||Sr)rsuperrmake)selfrZoptions __class__r:r;rHszPipScriptMaker.make)N)rrrr __classcell__r:r:rr;rGsrFTc - sL|st|||||| d}t| r,|dn|dgg tjjtjj} itg} |rt4} t t dt j | dddW5QRXW5QRXt | d- fdd d.fd d }|| dtjd d}t|\fdd}D]j}d }d }ttj |D]F}d }|dkrPt}|}tj ||} ||}|| |d||dq6qtd |d}d|_dh|_d|_g}dd }|r4dtjkr|d|tjdddkr|dtjd |f|dt|fddD}|D] }|=q&dd }|rdtjkr`|d||dt|fddD}|D] }|=q|dd  Dd!d D}g}z.|!|}| || |!|d"diWn>t"k r.}z|j#d } t$d#%| W5d }~XYnX| rRt&|}!|!d k rRt '|!tjd d$}"tjd d%}#t(|#d&}$|$)d'W5QRXt*+|#|"| |"tjd d(}%tjd d)}&t,|%d*\}'t,|&d+F}(t-.|'})t/|)| d,}*t-0|(}+t1|*D]},|+2|,qW5QRXW5QRXt*+|&|%d S)/zInstall a wheel)userhomerootisolatedprefixZpurelibZplatlibignoreT)forcequietFcs.t|}t|}||<|r*|dS)z6Map archive RECORD paths to installation RECORD paths.N)r<r)srcfilerZmodifiedZoldpathnewpath)rrrrwr:r;record_installeds   z*move_wheel_files..record_installedNcst|t|D]\}}}|t|dtjj}tj||} |rf|tjjdd drfq|D]d} tj||| } |r|dkr| dr | qjqj|rj| drjt |  t j rj | qj|D]} |r|| rqtj|| } tj||| }t| tj|r$t|t| |t| }ttdrZt||j|jft| tjrt| }|jtjBtjBtjB}t||d}|r||}| ||qqdS)Nr1rz.datarTz .dist-infoutimeF)rr3walkrBlstripr4r7r_r^endswithrrrfrWrunlinkshutilZcopyfilestathasattrrst_atimest_mtimeaccessX_OKst_modeS_IXUSRS_IXGRPS_IXOTHchmod)sourcedestZis_basefixerfilterdirZsubdirsfilesZbasedirZdestdirr~Z destsubdirrGrrstZ permissionsr) data_dirsinfo_dirrreqr:r;clobbersP          z!move_wheel_files..clobberrzentry_points.txtcsh|dr|dd}n<|dr8|dd}n |drT|dd}n|}|kpf|kS)Nz.exez -script.pyiz.pya)rvr)rWZ matchname)rrr:r;is_entrypoint_wrappersz/move_wheel_files..is_entrypoint_wrapperr)rrrTZpipZENSUREPIP_OPTIONSzpip = Z altinstallz pip%s = %scSsg|]}td|r|qS)zpip(\d(\.\d)?)?$rertrkr:r:r;r0s z$move_wheel_files..Z easy_installzeasy_install = zeasy_install-%s = %scSsg|]}td|r|qS)zeasy_install(-\d\.\d)?$rrr:r:r;r@s css|]}d|VqdS)%s = %sNr:rZkvr:r:r;rGsz#move_wheel_files..cSsg|] }d|qS)rr:rr:r:r;rKsrzInvalid script entry point: {} for req: {} - A callable suffix is required. Cf https://packaging.python.org/en/latest/distributing.html#console-scripts for more information.Z INSTALLERz INSTALLER.piprcspip ZRECORDz RECORD.piprzw+)rrrr)F)NN)3rr{rNr3r4r7rrwarningscatch_warningsfilterwarnings compileall compile_dirrdebuggetvaluer_rrrrorrZvariantsZset_moderrrrrUrVrextendrZ make_multiplerargsr rrrrArlrmover[csvreaderrwriterrZwriterow)-rWrrwrrrZ pycompileZschemerrZwarn_script_locationrrstdoutrZep_filerZdatadirrrZsubdirrZmakerZscripts_to_generateZ pip_scriptZpip_eprZeasy_install_scriptZeasy_install_epZgui_scripts_to_generateZgenerated_console_scriptsermsgZ installerZtemp_installerZinstaller_filerecordZ temp_recordZ record_inZ record_outrrr rr:) rrrrrrrrrrwr;move_wheel_filesMs    $F   #                     "rcCsrzVddtd|Dd}|d}t|}|d}ttt| d}|WSt k rlYdSXdS)z Return the Wheel-Version of an extracted wheel, if possible. Otherwise, return None if we couldn't parse / extract it. cSsg|]}|qSr:r:)rdr:r:r;rsz!wheel_version..Nrrqz Wheel-Version.) rZ find_on_pathZ get_metadatarZparsestrrrmapintr^ Exception) source_dirZdistZ wheel_dataversionr:r:r; wheel_versions   rcCsb|std||dtdkr>td|dtt|fn |tkr^tddtt|dS)a Raises errors or warns if called with an incompatible Wheel-Version. Pip should refuse to install a Wheel-Version that's a major series ahead of what it's compatible with (e.g 2.0 > 1.1); and warn when installing a version only minor version ahead (e.g 1.2 > 1.1). version: a 2-tuple representing a Wheel-Version (Major, Minor) name: name of wheel or package to raise exception about :raises UnsupportedWheel: when an incompatible Wheel-Version is given z(%s is in an unsupported or invalid wheelrzB%s's Wheel-Version (%s) is not compatible with this version of piprz*Installing from a newer Wheel-Version (%s)N)r VERSION_COMPATIBLEr_rrOrr)rrWr:r:r;check_compatibilitysrcCs d|S)z Format three tags in the form "--". :param file_tag: A 3-tuple of tags (python_tag, abi_tag, platform_tag). r\)r_)Zfile_tagr:r:r; format_tagsrc@s>eZdZdZedejZddZddZ ddZ d d Z d S) Wheelz A wheel filez^(?P(?P.+?)-(?P.*?)) ((-(?P\d[^-]*?))?-(?P.+?)-(?P.+?)-(?P.+?) \.whl|\.dist-info)$csj|}|std||_|ddd_|ddd_|d_|d d_ |d  d_ |d  d_ fd d j D_ d S)zX :raises InvalidWheelFilename: when the filename is invalid for a wheel z!%s is not a valid wheel filename.rWrpr\ZverbuildZpyverrZabiZplatcs.h|]&}jD]}jD]}|||fqqqSr:)abisplats)rryzrr:r; s z!Wheel.__init__..N) wheel_file_rertr rrur6rWrZ build_tagr^Z pyversionsrr file_tags)rrZ wheel_infor:r r;__init__s   zWheel.__init__cCstdd|jDS)zF Return the wheel's tags as a sorted list of strings. css|]}t|VqdSr)rrtagr:r:r;rsz0Wheel.get_formatted_file_tags..)rr#r r:r:r;get_formatted_file_tagsszWheel.get_formatted_file_tagscstfdd|jDS)a Return the lowest index that one of the wheel's file_tag combinations achieves in the given list of supported tags. For example, if there are 8 supported tags and one of the file tags is first in the list, then return 0. :param tags: the PEP 425 tags to check the wheel against, in order with most preferred first. :raises ValueError: If none of the wheel's file tags match one of the supported tags. c3s |]}|kr|VqdSr)indexr%tagsr:r;rsz*Wheel.support_index_min..)minr#rr*r:r)r;support_index_minszWheel.support_index_mincCs|j| S)z Return whether the wheel is compatible with one of the given tags. :param tags: the PEP 425 tags to check the wheel against. )r# isdisjointr,r:r:r; supportedszWheel.supportedN) rrr__doc__rcompileVERBOSEr"r$r'r-r/r:r:r:r;rsrz([a-z0-9_.]+)-([a-z0-9_.!+-]+)cCst||S)zjDetermine whether the string looks like an egg_info. :param s: The string to parse. E.g. foo-2.1 )boolsearch)r~Z _egg_info_rer:r:r;_contains_egg_infosr5cCs|jr dS|jr&|s"td|jdS|s.dS|js:|js>dS||sXtd|jdS|jrj|jjrjdS|j}| \}}|rt |rdSdS)a[ Return whether to build an InstallRequirement object using the ephemeral cache. :param cache_available: whether a cache directory is available for the should_unpack=True case. :return: True or False to build the requirement with ephem_cache=True or False, respectively; or None not to build the requirement. Nz(Skipping %s, due to already being wheel.FzCSkipping wheel build for %s, due to binaries being disabled for it.T) Z constraintZis_wheelrinforWZeditablerlinkZis_vcssplitextr5)r should_unpackcache_availablecheck_binary_allowedr7baseZextr:r:r;should_use_ephemeral_cache s4   r=cCs^t|}d|}|s |d7}n:ttjkr8|d7}n"|dsJ|d7}|d|t7}|S)z1 Format command information for logging. zCommand arguments: {} zCommand output: Nonez'Command output: [use --verbose to show]r|zCommand output: {}{})rrrZgetEffectiveLevelloggingDEBUGrr) command_argscommand_outputZ command_desctextr:r:r;format_command_resultGs    rCcCsxt|}|s4d|j}|t||7}t|dSt|dkrfd|j|}|t||7}t|tj ||dS)zH Return the path to the wheel in the temporary build directory. z1Legacy build of wheel for {!r} created no files. Nr1zZLegacy build of wheel for {!r} created more than one file. Filenames (choosing first): {} r) rrrWrCrrrBr3r4r_)namestemp_dirrr@rAr r:r:r;get_legacy_build_wheel_path^s$    rFcCsdS)NTr:)rpr:r:r; _always_true~srGc@s\eZdZdZdddZdddZddd Zd d Zdd d ZdddZ ddZ dddZ dS) WheelBuilderz#Build wheels from a RequirementSet.NFcCsD|dkr t}||_||_|j|_|p&g|_|p0g|_||_||_dSr) rGpreparer wheel_cacheZwheel_download_dir _wheel_dir build_optionsglobal_optionsr;no_clean)rrIrJrLrMr;rNr:r:r;r$s   zWheelBuilder.__init__c Cs.|j|j|||dW5QRSQRXdS)ziBuild one wheel. :return: The filename of the built wheel, or None if the build failed.  python_tagN)Z build_env_build_one_inside_env)rr output_dirrPr:r:r; _build_oneszWheelBuilder._build_onec Cstdd}|jr|j}n|j}|||j|d}|dk rtj|}tj||}zNt|\} } t ||t d|j || | t d||WW5QRStk rYnX||W5QRdSQRXdS)Nry)ZkindrOz3Created wheel for %s: filename=%s size=%d sha256=%szStored in directory: %s)rZ use_pep517_build_one_pep517_build_one_legacyr4r3rr_rIrrrr6rWZ hexdigestr _clean_one) rrrRrPrEZbuilder wheel_path wheel_nameZ dest_pathZ wheel_hashrFr:r:r;rQs.      z"WheelBuilder._build_one_inside_envcCst|j|jddS)NT)rMZunbuffered_output)rZ setup_py_pathrM)rrr:r:r;_base_setup_argss zWheelBuilder._base_setup_argsc Cs|jrtd|jfdSz~td|td|j}|j}|||j ||j d}W5QRX|rt ||}t t j||t j|||}Wn$tk rtd|jYdSXt j||S)zBuild one InstallRequirement using the PEP 517 build process. Returns path to wheel if successfully built. Otherwise, returns None. zGCannot build wheel for %s using PEP 517 when --build-options is presentNDestination directory: %szBuilding wheel for {} (PEP 517))metadata_directoryFailed building wheel for %s)rLrerrorrWrrrZpep517_backendZsubprocess_runnerZ build_wheelr[rar3renamer4r_r)rrtempdrPZrunnerZbackendrXnew_namer:r:r;rTs6      zWheelBuilder._build_one_pep517c Cs||}d|jf}t|}td||dd|g|j}|dk rT|d|g7}zt||j|d}Wn8tk r| dt d |jYW5QRdSXt |} t | ||||d } | W5QRSQRXdS) zBuild one InstallRequirement using the "legacy" build process. Returns path to wheel if successfully built. Otherwise, returns None. z Building wheel for %s (setup.py)rZZ bdist_wheelz-dNz --python-tag)cwdspinnerr]r\)rDrErr@rA)rYrWrrrrLrZunpacked_source_directoryrZfinishr]r3rrrF) rrr_rP base_argsZ spin_messagerbZ wheel_argsoutputrDrWr:r:r;rUs8         zWheelBuilder._build_one_legacycCsb||}td|j|ddg}zt||jdWdStk r\td|jYdSXdS)NzRunning setup.py clean for %sZcleanz--all)raTz Failed cleaning build dir for %sF)rYrr6rWrrrr])rrrcZ clean_argsr:r:r;rVs  zWheelBuilder._clean_onec Csg}t|jj}|D]\}t||||jd}|dkr4q|r\|rL|j|j}qb|j|j}n|j}| ||fq|szgSt dd dd|Dd}|rt j}tgg} } |D]\}}z t|WnFtk r} z&t d|j| | |WYqW5d} ~ XYnX|j|||d} | r| ||r|jrXt|jsXtd |||jj|_tt| |_t|jj|jq| |qW5QRX| rt d d d d| D| rt d d dd| D| S)aBuild wheels. :param should_unpack: If True, after building the wheel, unpack it and replace the sdist with the unpacked version in preparation for installation. :return: True if all the wheels built correctly. )r9r:r;Nz*Building wheels for collected packages: %srcSsg|]\}}|jqSr:rW)rrrpr:r:r;r\sz&WheelBuilder.build..z Building wheel for %s failed: %srOzbad source dir - missing markerzSuccessfully built %sr}cSsg|] }|jqSr:rerrr:r:r;rszFailed to build %scSsg|] }|jqSr:rerfr:r:r;rs) r3rJ cache_dirr=r;Zget_ephem_path_for_linkr7Zget_path_for_linkrKrrr6r_r Zimplementation_tagrrOSErrorrrWrSrrAssertionErrorZremove_temporary_sourceZensure_build_locationrIZ build_dirrrrZ file_path) rZ requirementsr9Zbuildsetr:rZ ephem_cacherRrPZ build_successZ build_failurer Z wheel_filer:r:r;r#s      zWheelBuilder.build)NNNF)N)N)N)N)F) rrrr0r$rSrQrYrTrUrVrr:r:r:r;rHs   ' %rH)r=)r=)FNNTNFNT)r0Z __future__rrrrr@r>Zos.pathr3rrrrUrbase64rZ email.parserrZ pip._vendorrZpip._vendor.distlib.scriptsrZpip._vendor.distlib.utilrZpip._vendor.packaging.utilsrZpip._vendor.sixr Z pip._internalr Zpip._internal.exceptionsr r r Zpip._internal.locationsrrZpip._internal.models.linkrZpip._internal.utils.loggingrZ pip._internal.utils.marker_filesrZpip._internal.utils.miscrrrZ$pip._internal.utils.setuptools_buildrZpip._internal.utils.subprocessrrrrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.uirZpip._internal.utils.unpackingrZpip._internal.utils.urlsrtypingr r!r"r#r$r%r&r'r(r)r*r+Z"pip._vendor.packaging.requirementsr,Zpip._internal.req.req_installr-Z pip._internal.operations.preparer.Zpip._internal.cacher/Zpip._internal.pep425tagsr0rOZInstalledCSVRowr3ZBinaryAllowedPredicaterZ getLoggerrrr<rIrPr[raror1r2rsr{rrrrrrrrrrrrobjectrIr5r=rCrFrGrHr:r:r:r;s                  8           =$  4 J : site-packages/pip/_internal/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000320151464452710021371 0ustar00U .eP@s ddlZdS)N)Z*pip._internal.utils.inject_securetransportZpiprr:/usr/lib/python3.8/site-packages/pip/_internal/__init__.pysite-packages/pip/_internal/__pycache__/exceptions.cpython-38.opt-1.pyc000064400000030242151464452710022021 0ustar00U .e(@sdZddlmZddlmZmZmZddlmZddl m Z e rdddl m Z ddl mZddlmZGd d d eZGd d d eZGd ddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdddeZGdd d eZGd!d"d"eZGd#d$d$eZGd%d&d&eZ Gd'd(d(e Z!Gd)d*d*e Z"Gd+d,d,e Z#Gd-d.d.e Z$Gd/d0d0e Z%Gd1d2d2eZ&Gd3d4d4eZ'd5S)6z"Exceptions used throughout package)absolute_import)chaingroupbyrepeat) iteritems)MYPY_CHECK_RUNNING)Optional) Distribution)InstallRequirementc@seZdZdZdS)PipErrorzBase pip exceptionN__name__ __module__ __qualname____doc__rrnz$HashErrors.__str__..)keycSs|jSr' __class__r,rrrr.or/css|]}|VqdSr')body.0r-rrr qsz%HashErrors.__str__.. )r(sortrr)headextendjoin)rlinesclsZ errors_of_clsrrrrls zHashErrors.__str__cCs t|jSr')boolr(rrrr __nonzero__uszHashErrors.__nonzero__cCs|Sr')r?rrrr__bool__xszHashErrors.__bool__N) r rrrrr)rr?r@rrrrr&cs  r&c@s0eZdZdZdZdZddZddZdd ZdS) HashErrora A failure to verify a package against known-good hashes :cvar order: An int sorting hash exception classes by difficulty of recovery (lower being harder), so the user doesn't bother fretting about unpinned packages when he has deeper issues, like VCS dependencies, to deal with. Also keeps error reports in a deterministic order. :cvar head: A section heading for display above potentially many exceptions of this kind :ivar req: The InstallRequirement that triggered this error. This is pasted on after the exception is instantiated, because it's not typically available earlier. NcCs d|S)a)Return a summary of me for display under the heading. This default implementation simply prints a description of the triggering requirement. :param req: The InstallRequirement that provoked this error, with populate_link() having already been called z %s)_requirement_namerrrrr3s zHashError.bodycCsd|j|fS)Nz%s %s)r9r3rrrrrszHashError.__str__cCs|jrt|jSdS)zReturn a description of the requirement that triggered me. This default implementation returns long description of the req, with line numbers unknown package)reqstrrrrrrCszHashError._requirement_name) r rrrrEr9r3rrCrrrrrA|s  rAc@seZdZdZdZdZdS)VcsHashUnsupporteduA hash was provided for a version-control-system-based requirement, but we don't have a method for hashing those.rzlCan't verify hashes for these requirements because we don't have a way to hash version control repositories:Nr rrrr+r9rrrrrGsrGc@seZdZdZdZdZdS)DirectoryUrlHashUnsupportedrHzUCan't verify hashes for these file:// requirements because they point to directories:NrIrrrrrJsrJc@s(eZdZdZdZdZddZddZdS) HashMissingz2A hash was needed for a requirement but is absent.awHashes are required in --require-hashes mode, but they are missing from some requirements. Here is a list of those requirements along with the hashes their downloaded archives actually had. Add lines like these to your requirements files to prevent tampering. (If you did not enable --require-hashes manually, note that it turns on automatically when any package has a hash.)cCs ||_dS)zq :param gotten_hash: The hash of the (possibly malicious) archive we just downloaded N) gotten_hash)rrNrrrrszHashMissing.__init__cCsHddlm}d}|jr4|jjr&|jjn t|jdd}d|p.hash_then_orc3s|]}dt|fVqdS)z Expected %s %sN)nextr4prefixrrr6sz0HashMismatch._hash_comparison..z Got %s r7)rrVr:r)rWZ hexdigestr;)rrZr<rYZ expectedsrr\rrXs  zHashMismatch._hash_comparisonN) r rrrr+r9rr3rXrrrrrTs  rTc@seZdZdZdS)UnsupportedPythonVersionzMUnsupported python version according to Requires-Python package metadata.Nr rrrrr^sr^cs*eZdZdZdfdd ZddZZS) !ConfigurationFileCouldNotBeLoadedz=When there are errors while loading a configuration file could not be loadedNcs&tt||||_||_||_dSr')superr_rreasonfnamer*)rrbrcr*r1rrr(sz*ConfigurationFileCouldNotBeLoaded.__init__cCs4|jdk rd|j}nd|jj}d|j|S)Nz in {}.z. {} zConfiguration file {}{})rcrr*messagerb)rZ message_partrrrr.s z)ConfigurationFileCouldNotBeLoaded.__str__)r`NN)r rrrrr __classcell__rrr1rr_$sr_N)(rZ __future__r itertoolsrrrZpip._vendor.sixrZpip._internal.utils.typingrtypingrZpip._vendor.pkg_resourcesr Zpip._internal.req.req_installr Exceptionr rrrrrrr r!r"r#r$r%r&rArGrJrLrRrTr^r_rrrrs<      ,  % 7site-packages/pip/_internal/__pycache__/self_outdated_check.cpython-38.opt-1.pyc000064400000012622151464452710023621 0ustar00U .e@s^ddlmZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z ddlmZddlmZddlmZddlmZdd lmZdd lmZdd lmZmZmZdd lmZm Z m!Z!dd l"m#Z#ddl$m%Z%e%rddl&Z&ddl&m'Z'ddl(m)Z)m*Z*m+Z+m,Z,ddl-m.Z.dZ/e0e1Z2dddZ3ddZ4Gddde5Z6ddZ7ddZ8dS))absolute_importN) pkg_resources)version) ensure_binary) LinkCollector) PackageFinder) SearchScope)SelectionPreferences)WINDOWS)adjacent_tmp_filecheck_path_ownerreplace) ensure_dirget_installed_versionredact_auth_from_url) get_installer)MYPY_CHECK_RUNNING)Values)AnyDictTextUnion) PipSessionz%Y-%m-%dT%H:%M:%SZFcCs`|jg|j}|jr8|s8tdddd|Dg}|jp@g}tj||d}t ||d}|S)z :param session: The Session to use to make requests. :param suppress_no_index: Whether to ignore the --no-index option when constructing the SearchScope object. zIgnoring indexes: %s,css|]}t|VqdSN)r).0ZurlrE/usr/lib/python3.8/site-packages/pip/_internal/self_outdated_check.py @sz&make_link_collector..) find_links index_urls)session search_scope) Z index_urlZextra_index_urlsZno_indexloggerdebugjoinrrcreater)r!optionssuppress_no_indexr rr"link_collectorrrrmake_link_collector1s    r*cCst|}t|}|Sr)rhashlibZsha224Z hexdigest)keyZ key_bytesnamerrr_get_statefile_namePsr.c@s(eZdZddZeddZddZdS)SelfCheckStatec Csni|_d|_|rjtj|dt|j|_z&t|j}t ||_W5QRXWnt t t fk rhYnXdS)NZ selfcheck) statestatefile_pathospathr%r.r,openjsonloadIOError ValueErrorKeyError)self cache_dirZ statefilerrr__init__Xs zSelfCheckState.__init__cCstjSr)sysprefix)r:rrrr,jszSelfCheckState.keyc Cs|js dSttj|js dSttj|j|j|t|d}t j |ddd}t |j}| t |W5QRXzt|j|jWntk rYnXdS)N)r, last_check pypi_versionT)r:)Z sort_keysZ separators)r1r r2r3dirnamerr,strftimeSELFCHECK_DATE_FMTr5dumpsr writerr r-OSError)r:r@ current_timer0textfrrrsavens  zSelfCheckState.saveN)__name__ __module__ __qualname__r<propertyr,rKrrrrr/Ws r/cCs6zt|}dt|kWStjk r0YdSXdS)zChecks whether pkg was installed by pip This is used not to display the upgrade message when pip is in fact installed by system package manager, such as dnf on Fedora. pipFN)rZget_distributionrZDistributionNotFound)ZpkgZdistrrrwas_installed_by_pips  rQcCsXtd}|sdSt|}d}zt|jd}tj}d|jkrzd|jkrztj|jdt }|| dkrz|jd}|dkrt ||dd}t d d d } t j|| d } | dj} | dkrWdSt| j}|||t|} || ko|j| jkotd} | s WdStrd }nd}td |||Wn$tk rRtjdddYnXdS)zCheck for an update for pip. Limit the frequency of checks to once per week. State is stored either in the active virtualenv or in the user's USER_CACHE_DIR keyed off the prefix of the pip script path. rPN)r;r?r@i: T)r'r(F)Z allow_yankedZallow_all_prereleases)r)selection_prefsz python -m pipzYou are using pip version %s; however, version %s is available. You should consider upgrading via the '%s install --upgrade pip' command.z5There was an error checking the latest version of pip)exc_info)rpackaging_versionparser/r;datetimeZutcnowr0strptimerDZ total_secondsr*r rr&Zfind_best_candidatebest_candidatestrrrKZ base_versionrQr r#Zwarning Exceptionr$)r!r'Zinstalled_versionZ pip_versionr@r0rHr?r)rRfinderrXZremote_versionZlocal_version_is_olderZpip_cmdrrrpip_self_version_checksp         r\)F)9Z __future__rrVr+r5ZloggingZos.pathr2r=Z pip._vendorrZpip._vendor.packagingrrTZpip._vendor.sixrZpip._internal.collectorrZpip._internal.indexrZ!pip._internal.models.search_scoperZ$pip._internal.models.selection_prefsr Zpip._internal.utils.compatr Zpip._internal.utils.filesystemr r r Zpip._internal.utils.miscrrrZpip._internal.utils.packagingrZpip._internal.utils.typingrZoptparsertypingrrrrZpip._internal.network.sessionrrDZ getLoggerrLr#r*r.objectr/rQr\rrrrs>               ;site-packages/pip/_internal/__pycache__/collector.cpython-38.pyc000064400000033422151464452710020672 0ustar00U .eWF@s dZddlZddlZddlZddlZddlZddlmZddlm Z m Z ddl m Z ddl mZmZmZddlmZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZm Z ddl!m"Z"m#Z#erDddl$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-ddl.Z/ddl0m1Z1ddl2m3Z3ddl4m5Z5e/j6j7j8Z9e)e:e:fZ;eddZ?ddZ@GdddeAZBddZCGdddeAZDddZEdd ZFd!d"ZGd#d$ZHd%d&ZId'd(ZJd)d*ZKGd+d,d,eLZMdd6d7ZRGd8d9d9eLZSGd:d;d;eLZTdS)?zM The main purpose of this module is to expose LinkCollector.collect_links(). N) OrderedDict)html5librequests)unescape) HTTPError RetryErrorSSLError)parse)requestLink)ARCHIVE_EXTENSIONS)redact_auth_from_url)MYPY_CHECK_RUNNING) path_to_url url_to_path)is_urlvcs) CallableDictIterableListMutableMappingOptionalSequenceTupleUnion)Response) SearchScope) PipSessioncCs6tjD]*}||r|t|dkr|SqdS)zgLook for VCS schemes in the URL. Returns the matched VCS scheme, or None if there's no match. z+:N)rZschemeslower startswithlen)urlschemer%;/usr/lib/python3.8/site-packages/pip/_internal/collector.py_match_vcs_scheme/s  r'cCs(t|j}tD]}||rdSqdS)z2Return whether the URL looks like an archive. TF)r filenamer endswith)r#r(Zbad_extr%r%r&_is_url_like_archive;s   r*cseZdZfddZZS)_NotHTMLcs"tt|||||_||_dSN)superr+__init__ content_type request_desc)selfr/r0 __class__r%r&r.Gsz_NotHTML.__init__)__name__ __module__ __qualname__r. __classcell__r%r%r2r&r+Fsr+cCs.|jdd}|ds*t||jjdS)zCheck the Content-Type header to ensure the response contains HTML. Raises `_NotHTML` if the content type is not text/html. Content-Type text/htmlN)headersgetr r!r+r method)responser/r%r%r&_ensure_html_headerNsr?c@s eZdZdS)_NotHTTPN)r4r5r6r%r%r%r&r@Ysr@cCsDt|\}}}}}|dkr"t|j|dd}|t|dS)zSend a HEAD request to the URL, and ensure the response contains HTML. Raises `_NotHTTP` if the URL is not available for a HEAD request, or `_NotHTML` if the content type is not text/html. >httphttpsT)Zallow_redirectsN) urllib_parseZurlsplitr@headraise_for_statusr?)r#sessionr$netlocpathZqueryZfragmentrespr%r%r&_ensure_html_response]s rJcCsLt|rt||dtdt||j|dddd}|t||S)aAccess an HTML page with GET, and return the response. This consists of three parts: 1. If the URL looks suspiciously like an archive, send a HEAD first to check the Content-Type is HTML, to avoid downloading a large file. Raise `_NotHTTP` if the content type cannot be determined, or `_NotHTML` if it is not HTML. 2. Actually perform the request. Raise HTTP exceptions on network failures. 3. Check the Content-Type header to make sure we got HTML, and raise `_NotHTML` otherwise. rFzGetting page %sr:z max-age=0)ZAcceptz Cache-Control)r;)r*rJloggerdebugrr<rEr?)r#rFrIr%r%r&_get_html_responsens rNcCs2|r.d|kr.t|d\}}d|kr.|dSdS)zBDetermine if we have any encoding information in our headers. r8charsetN)cgiZ parse_header)r;r/Zparamsr%r%r&_get_encoding_from_headerss  rQcCs.|dD]}|d}|dk r |Sq |S)aDetermine the HTML document's base URL. This looks for a ```` tag in the HTML document. If present, its href attribute denotes the base URL of anchor tags in the document. If there is no such tag (or if it does not have a valid href attribute), the HTML file's URL is used as the base URL. :param document: An HTML document representation. The current implementation expects the result of ``html5lib.parse()``. :param page_url: The URL of the HTML document. z.//basehrefN)findallr<)documentpage_urlbaserRr%r%r&_determine_base_urls   rWcCsPt|}|jdkr(tt|j}ntjt|jdd}t |j |dS)zMakes sure a link is fully encoded. That is, if a ' ' shows up in the link, it will be rewritten to %20 (while not over-quoting % or other characters).r9z/@)Zsafe)rH) rCurlparserGurllib_requestZ pathname2url url2pathnamerHZquoteZunquoteZ urlunparse_replace)r#resultrHr%r%r& _clean_links   r]cCsf|d}|sdStt||}|d}|r8t|nd}|d}|rRt|}t||||d}|S)zJ Convert an anchor element in a simple repository page to a Link. rRNzdata-requires-pythonz data-yanked)Z comes_fromZrequires_python yanked_reason)r<r]rCurljoinrr )anchorrUbase_urlrRr#Z pyrequirer^linkr%r%r&_create_link_from_elements   rcccsVtj|j|jdd}|j}t||}|dD]"}t|||d}|dkrJq.|Vq.dS)zP Parse an HTML document, and yield its anchor elements as Link objects. F)Ztransport_encodingZnamespaceHTMLElementsz.//a)rUraN)rr contentencodingr#rWrSrc)pagerTr#rar`rbr%r%r& parse_linkss  rgc@s eZdZdZddZddZdS)HTMLPagez'Represents one page, along with its URLcCs||_||_||_dS)z :param encoding: the encoding to decode the given content. :param url: the URL from which the HTML was downloaded. N)rdrer#)r1rdrer#r%r%r&r.s zHTMLPage.__init__cCs t|jSr,)rr#r1r%r%r&__str__$szHTMLPage.__str__N)r4r5r6__doc__r.rjr%r%r%r&rhsrhcCs|dkrtj}|d||dS)Nz%Could not fetch URL %s: %s - skipping)rLrM)rbreasonmethr%r%r&_handle_get_page_fail(srncCst|j}t|j||jdS)N)rer#)rQr;rhrdr#)r>rer%r%r&_make_html_page3s roc Cs|dkrtd|jddd}t|}|r@td||dSt|\}}}}}}|dkrtj t |r| ds|d7}t|d}td |zt||d }WnDtk rtd |Yn,tk r}ztd ||j|jW5d}~XYntk r0}zt||W5d}~XYntk r\}zt||W5d}~XYntk r}z$d } | t|7} t|| tjdW5d}~XYn\tjk r}zt|d|W5d}~XYn*tjk rt|dYn Xt|SdS)Nz?_get_html_page() missing 1 required keyword argument: 'session'#rzCannot look at %s URL %sfile/z index.htmlz# file: URL is directory, getting %srKzQSkipping page %s because it looks like an archive, and cannot be checked by HEAD.z.sort_pathzfile:z)Path '{0}' is ignored: it is a directory.z:Url '%s' is ignored: it is neither a file nor a directory.zQUrl '%s' is ignored. It is either a non-existing path or lacks a specific scheme.)rvrHexistsr!rrwrealpathlistdirjoinrrLZwarningformatisfiler) locations expand_dirrr#Z is_local_pathZ is_file_urlrHitemr%rr&group_locationsxsF        rc@seZdZdZddZdS)CollectedLinksa Encapsulates all the Link objects collected by a call to LinkCollector.collect_links(), stored separately as-- (1) links from the configured file locations, (2) links from the configured find_links, and (3) a dict mapping HTML page url to links from that page. cCs||_||_||_dS)z :param files: Links from file locations. :param find_links: Links from find_links. :param pages: A dict mapping HTML page url to links from that page. Nr find_linkspages)r1rrrr%r%r&r.s zCollectedLinks.__init__N)r4r5r6rkr.r%r%r%r&rs rc@s4eZdZdZddZeddZddZdd Zd S) LinkCollectorz Responsible for collecting Link objects from all configured locations, making network requests as needed. The class's main method is its collect_links() method. cCs||_||_dSr,) search_scoperF)r1rFrr%r%r&r.szLinkCollector.__init__cCs|jjSr,)rrrir%r%r&rszLinkCollector.find_linksccs,|D]"}t||jd}|dkr q|VqdS)zp Yields (page, page_url) from the given locations, skipping locations that have errors. rKN)r}rF)r1rlocationrfr%r%r& _get_pagess zLinkCollector._get_pagescsj}||}t|\}}tjdd\}}ddt||D}ddjD} fddtdd|Dd d|DD} t| } d t| |g} | D]} | d | qt d | i} | D]}tt|| |j<qt|| | d S)zFind all available links for the given project name. :return: All the Link objects (unfiltered), as a CollectedLinks object. T)rcSsg|] }t|qSr%r .0r#r%r%r& sz/LinkCollector.collect_links..cSsg|]}t|dqS)z-fr rr%r%r&rscsg|]}j|r|qSr%)rFZis_secure_origin)rrbrir%r&r s css|]}t|VqdSr,r rr%r%r& sz.LinkCollector.collect_links..css|]}t|VqdSr,r rr%r%r&r sz,{} location(s) to search for versions of {}:z* {} r)rZget_index_urls_locationsrr itertoolschainrrr"rrLrMrrr~rgr#r)r1Z project_namerZindex_locationsZindex_file_locZ index_url_locZ fl_file_locZ fl_url_locZ file_linksZfind_link_linksZ url_locationslinesrbZ pages_linksrfr%rir& collect_linkssD       zLinkCollector.collect_linksN) r4r5r6rkr.propertyrrrr%r%r%r&rs    r)N)N)F)UrkrPrZloggingrrv collectionsrZ pip._vendorrrZpip._vendor.distlib.compatrZpip._vendor.requests.exceptionsrrrZpip._vendor.six.moves.urllibr rCr rYZpip._internal.models.linkr Zpip._internal.utils.filetypesr Zpip._internal.utils.miscrZpip._internal.utils.typingrZpip._internal.utils.urlsrrZpip._internal.vcsrrtypingrrrrrrrrrZxml.etree.ElementTreeZxmlZpip._vendor.requestsrZ!pip._internal.models.search_scoperZpip._internal.network.sessionrZetreeZ ElementTreeZElementZ HTMLElementrxZResponseHeadersZ getLoggerr4rLr'r* Exceptionr+r?r@rJrNrQrWr]rcrgobjectrhrnror}rrrrr%r%r%r&s^        ,         3    6 ;site-packages/pip/_internal/__pycache__/locations.cpython-38.pyc000064400000006160151464452710020676 0ustar00U .e&@sdZddlmZddlZddlZddlZddlZddlZddlZddl mZ ddl m Z ddl mZddlmZddlmZdd lmZerdd lmZmZmZmZmZed Zd d ZddZedZ e!"dkre #Z z e$Z%Wne&k r ej'Z%YnXer`ej()ej*dZ+ej()e%dZ,ej(-e+sej()ej*dZ+ej()e%dZ,nJej()ej*dZ+ej()e%dZ,ejdddkrej*dddkrdZ+dddZ.dS)z7Locations where we look for configs, install stuff, etc)absolute_importN) sysconfig) SCHEME_KEYS)appdirs)WINDOWS)MYPY_CHECK_RUNNING)running_under_virtualenv)AnyUnionDictListOptionalZpipcCs djtjS)ze Return the major-minor version of the current Python as a string, e.g. "3.7" or "3.10". z{}.{})formatsys version_inforr;/usr/lib/python3.8/site-packages/pip/_internal/locations.pyget_major_minor_versionsrcCsZtrtjtjd}n6ztjtd}Wntk rLtdYnXtj |S)Nsrcz=The folder you are executing pip from can no longer be found.) rospathjoinrprefixgetcwdOSErrorexitabspath)Z src_prefixrrrget_src_prefix(s rpurelibZpypyZScriptsbindarwinz/System/Library/z/usr/local/binFcCsddlm}i}|r ddgi}ni}d|i} | ||| } | | jddd} | d k s`t|rx|rxtd |||r|rtd |||p| j| _|s|rd | _|p| j| _|p| j | _ |p| j | _ | t D]} t | d | || <qd| dkr|t| j| jdtrtjtjdddt||d<|d k rtjtj|dd} tj|| dd |d<|S)z+ Return a distutils install scheme r) DistributionZ script_argsz --no-user-cfgnameZinstallT)ZcreateNzuser={} prefix={}zhome={} prefix={}Zinstall_ install_lib)rZplatlibZincludesitezpython{}Zheaders)Zdistutils.distr#updateZparse_config_filesZget_command_objAssertionErrorruserrhomerootZfinalize_optionsrgetattrZget_option_dictdictr&rrrrrr splitdriver)Z dist_namer+r,r-isolatedrr#ZschemeZextra_dist_argsZ dist_argsdikeyZ path_no_driverrrdistutils_scheme[sV            r5)FNNFN)/__doc__Z __future__rrZos.pathplatformr'rrZ distutilsZdistutils_sysconfigZdistutils.command.installrZpip._internal.utilsrZpip._internal.utils.compatrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrtypingr r r r r Zuser_cache_dirZUSER_CACHE_DIRrrZget_pathZ site_packagesZpython_implementationlowerZget_python_libgetusersitepackages user_siteAttributeError USER_SITErrrZbin_pyZbin_userexistsr5rrrrsN            (site-packages/pip/_internal/__pycache__/build_env.cpython-38.pyc000064400000016454151464452710020661 0ustar00U .e]@sdZddlZddlZddlZddlZddlmZddlmZddl m Z ddl m Z m Z mZddlmZddlmZdd lmZdd lmZdd lmZerdd lmZmZmZmZmZdd l m!Z!e"e#Z$GdddZ%Gddde&Z'Gddde'Z(dS)z;Build Environment used for isolation during sdist building N) OrderedDict)get_python_lib) get_paths) RequirementVersionConflict WorkingSet)__file__)call_subprocess) TempDirectory)MYPY_CHECK_RUNNING) open_spinner)TupleSetIterableOptionalList) PackageFinderc@seZdZddZdS)_PrefixcCsj||_d|_ttjdkrdnd||ddd|_td|d}td|d}||kr\|g|_n ||g|_dS) NFntZ posix_prefix)baseZplatbase)varsZscripts) plat_specificprefixT)pathsetuprosnamebin_dirrlib_dirs)selfrZpurelibZplatlibr ;/usr/lib/python3.8/site-packages/pip/_internal/build_env.py__init__!s   z_Prefix.__init__N)__name__ __module__ __qualname__r"r r r r!rsrc@s@eZdZdZddZddZddZdd Zd d Zd d Z dS)BuildEnvironmentzFCreates and manages an isolated environment to install build deps c stdd_tfdddD_g_g_ttjD] }j |j j |j qBddt dd t d d fD}tjjjd _tjjstjttjjd d "}|tdj|jdW5QRXdS)Nz build-env)Zkindc3s(|] }|ttjjj|fVqdSN)rrrjoin _temp_dir.0rrr r! ;sz,BuildEnvironment.__init__..)ZnormalZoverlaycSsh|]}tj|qSr )rrnormcase)r+siter r r! Isz,BuildEnvironment.__init__..F)rTr/zsitecustomize.pywa import os, site, sys # First, drop system-sites related paths. original_sys_path = sys.path[:] known_paths = set() for path in {system_sites!r}: site.addsitedir(path, known_paths=known_paths) system_paths = set( os.path.normcase(path) for path in sys.path[len(original_sys_path):] ) original_sys_path = [ path for path in original_sys_path if os.path.normcase(path) not in system_paths ] sys.path = original_sys_path # Second, add lib directories. # ensuring .pth file are processed. for path in {lib_dirs!r}: assert not path in sys.path site.addsitedir(path) ) system_sitesr)r r)r _prefixes _bin_dirs _lib_dirsreversedlistvaluesappendrextendrrrrr( _site_direxistsmkdiropenwritetextwrapdedentformat)rrr2fpr r,r!r"7s0    zBuildEnvironment.__init__cCsndddD|_|jdd}|jd}|r>||tj|jg}tjtj |dtj |ddS)NcSsi|]}|tj|dqSr')renvirongetr*r r r! osz.BuildEnvironment.__enter__..)PATHZPYTHONNOUSERSITEZ PYTHONPATHrG1) _save_envr4r:splitrpathsepr;rDupdater()rrZold_pathZ pythonpathr r r! __enter__ns   zBuildEnvironment.__enter__cCs:|jD]*\}}|dkr*tj|dq |tj|<q dSr')rIitemsrrDpop)rexc_typeexc_valexc_tbZvarname old_valuer r r!__exit__szBuildEnvironment.__exit__cCs|jdSr')r)cleanupr,r r r!rUszBuildEnvironment.cleanupc Cst}t}|rt|j}|D]p}z"|t|dkrB||Wqtk r}z*|t|j d t|j dfW5d}~XYqXq||fS)zReturn 2 sets: - conflicting requirements: set of (installed, wanted) reqs tuples - missing requirements: set of reqs Nr) setrr5findrparseaddrstrargsZas_requirement)rZreqsZmissingZ conflictingZwsZreqer r r!check_requirementss  z#BuildEnvironment.check_requirementsc Csb|j|}|jrtd|_|s"dStjtjtdddd|jdg}t t j krZ| ddD]:}t|j|}|d |d d d t|pd hfq^|j} | r|d| dg| ddD]} |d| gqn | d|jD]} |d| gq|jD]} |d| gq|jr*| d| d ||t|} t|| dW5QRXdS)NTZinstallz--ignore-installedz --no-userz--prefixz--no-warn-script-locationz-v)Z no_binaryZ only_binaryz--_-,z:none:z-irrVz--extra-index-urlz --no-indexz --find-linksz--trusted-hostz--pre)spinner)r3rAssertionErrorsys executablerrdirname pip_locationloggerZgetEffectiveLevelloggingDEBUGr9getattrformat_controlr:replacer(sorted index_urlsZ find_linksZ trusted_hostsZallow_all_prereleasesr r )rfinder requirementsZprefix_as_stringmessagerr\rlZformatsroZ extra_indexlinkZhostrbr r r!install_requirementssJ            z%BuildEnvironment.install_requirementsN) r#r$r%__doc__r"rMrTrUr^rtr r r r!r&3s7r&c@s8eZdZdZddZddZddZdd Zd d Zd S) NoOpBuildEnvironmentz5A no-op drop-in replacement for BuildEnvironment cCsdSr'r r,r r r!r"szNoOpBuildEnvironment.__init__cCsdSr'r r,r r r!rMszNoOpBuildEnvironment.__enter__cCsdSr'r )rrPrQrRr r r!rTszNoOpBuildEnvironment.__exit__cCsdSr'r r,r r r!rUszNoOpBuildEnvironment.cleanupcCs tdSr')NotImplementedError)rrprqrrrr r r!rtsz)NoOpBuildEnvironment.install_requirementsN) r#r$r%rur"rMrTrUrtr r r r!rvs rv))rurirrdr@ collectionsrZdistutils.sysconfigrZ sysconfigrZpip._vendor.pkg_resourcesrrrZpiprrgZpip._internal.utils.subprocessr Zpip._internal.utils.temp_dirr Zpip._internal.utils.typingr Zpip._internal.utils.uir typingr rrrrZpip._internal.indexrZ getLoggerr#rhrobjectr&rvr r r r!s*          site-packages/pip/_internal/__pycache__/wheel.cpython-38.pyc000064400000066650151464452710020021 0ustar00U .eH@s4dZddlmZddlZddlZddlZddlZddlZddlZ ddl Z ddl Z ddl Z ddl Z ddlZddlmZddlmZddlmZddlmZddlmZdd lmZdd lmZdd lmZdd lm Z m!Z!m"Z"dd l#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-m.Z.m/Z/ddl0m1Z1ddl2m3Z3m4Z4m5Z5m6Z6ddl7m8Z8ddl9m:Z:ddl;mZ>ddl?m@Z@e:rddlAmBZBmCZCmDZDmEZEmFZFmGZGmHZHmIZImJZJmKZKmLZLmMZMddlNmOZOddlPmQZQddlRmSZSddlTmUZUddlVmWZWeGeXdfZYeLeQgeZfZ[d Z\e]e^Z_d!d"Z`dXd$d%ZadYd&d'Zbd(d)Zcd*d+Zdd,d-Zee fd.e jgZhd/d0Zid1d2Zjd3d4Zkd5d6Zld7d8ZmGd9d:d:enZod;d<ZpGd=d>d>eZqdZdAdBZrdCdDZsdEdFZtdGdHZuGdIdJdJevZwe fdKe jxfdLdMZydNdOZzdPdQZ{dRdSZ|dTdUZ}GdVdWdWevZ~dS)[zH Support for installing and building the "wheel" binary package format. )absolute_importN)urlsafe_b64encode)Parser) pkg_resources) ScriptMaker)get_export_entry)canonicalize_name)StringIO) pep425tags)InstallationErrorInvalidWheelFilenameUnsupportedWheel)distutils_schemeget_major_minor_version)Link) indent_log)has_delete_marker_file)captured_stdout ensure_dir read_chunks)make_setuptools_shim_args) LOG_DIVIDERcall_subprocessformat_command_argsrunner_with_spinner_message) TempDirectory)MYPY_CHECK_RUNNING) open_spinner) unpack_file) path_to_url) DictListOptionalSequenceMappingTupleIOTextAnyIterableCallableSet) Requirement)InstallRequirement)RequirementPreparer) WheelCache) Pep425Tag.)rcCstj||tjjdS)N/)ospathrelpathreplacesep)srcpr:7/usr/lib/python3.8/site-packages/pip/_internal/wheel.pynormpathOsr<c CsRt}d}t|d.}t||dD]}|t|7}||q$W5QRX||fS)z5Return (hash, length) for path using hashlib.sha256()rrb)size)hashlibZsha256openrlenupdate)r4 blocksizehlengthfblockr:r:r; hash_fileSs  rIcCs6t||\}}dt|dd}|t|fS)z?Return (encoded_digest, length) for path using hashlib.sha256()zsha256=latin1=)rIrdigestdecoderstripstr)r4rDrErFrLr:r:r;rehash_srPcCs6tjddkri}d}n ddi}d}t|||f|S)Nrbnewline)sys version_inforA)namemodenlbinr:r:r; open_for_csvjs r[cCs|d}||d<d|S)zBReplace the Python tag in a wheel file name with a new value. -)splitjoin)Z wheelnameZnew_tagpartsr:r:r;replace_python_tagus rac Cstj|rt|dR}|}|ds8W5QRdStjt }d|tj d}| }W5QRXt|d}| || |W5QRXdSdS) zLReplace #!python with #!/path/to/python Return True if file was changed.r>s#!pythonFs#!asciiwbTN) r3r4isfilerAreadline startswithrU executableencodegetfilesystemencodinglinesepreadwrite)r4Zscript firstlineZexenamerestr:r:r; fix_script~s     rozX^(?P(?P.+?)(-(?P.+?))?) \.dist-info$c Cs|dd}t|D]p}t|}|r|d|krttj||d6}|D]*}| }|dkrPW5QRdSqPW5QRXqdS)zP Return True if the extracted wheel in wheeldir should go into purelib. r\_rWWHEELzroot-is-purelib: trueTF) r6r3listdir dist_info_rematchgrouprAr4r_lowerrN)rWwheeldirZ name_foldeditemrtwheelliner:r:r;root_is_purelibs    r{c stj|siifSt|8}t}|D]}|||dq(|dW5QRXtj |}| di}| di}ddt fdd| D}t fd d| D}||fS) N rZconsole_scriptsZ gui_scriptscSst|dddS)zRget the string representation of EntryPoint, remove space and split on '=' rTrK)rOr6r^)sr:r:r; _split_epsz"get_entrypoints.._split_epc3s|]}|VqdSNr:.0vrr:r; sz"get_entrypoints..c3s|]}|VqdSrr:rrr:r;rs)r3r4existsrAr rlstripseekrZ EntryPointZ parse_mapgetdictvalues)filenamefpdatarzZ entry_pointsconsoleguir:rr;get_entrypointss      rc sJ|sdStt}|D]*}tj|}tj|}|||qddtj dd tj D tj tjtjfdd|D}|sdSg}|D]b\}}t|}t|dkrd |d } n$d d |dd d|d } | d| |qd} t|dkr0| | dn| | dd|S)zDetermine if any scripts are not on PATH and format a warning. Returns a warning message if one or more scripts are not on PATH, otherwise None. NcSs g|]}tj|tjqSr:)r3r4normcaserNr7)rir:r:r; sz5message_about_scripts_not_on_PATH..PATHrTcs&i|]\}}tj|kr||qSr:)r3r4r)r parent_dirscriptsZ not_warn_dirsr:r; sz5message_about_scripts_not_on_PATH..r1z script {} isrzscripts {} are, z and z.The {} installed in '{}' which is not on PATH.zeConsider adding {} to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.zthis directoryzthese directoriesr|) collections defaultdictsetr3r4dirnamebasenameaddenvironrr^pathsepappendrrUrgitemssortedrBformatr_) rZgrouped_by_dirdestfilerZ script_nameZwarn_for msg_linesZ dir_scriptsZsorted_scriptsZ start_textZ last_line_fmtr:rr;!message_about_scripts_not_on_PATHsH     rcCst|dddS)a; Return the given rows of a RECORD file in sorted order. Each row is a 3-tuple (path, hash, size) and corresponds to a record of a RECORD file (see PEP 376 and PEP 427 for details). For the rows passed to this function, the size can be an integer as an int or string, or the empty string. cSstdd|DS)Ncss|]}t|VqdSr)rO)rxr:r:r;rsz3sorted_outrows....)tuple)rowr:r:r;z sorted_outrows..)key)r)outrowsr:r:r;sorted_outrowssrc Csg}|D]v}t|dkr(td|t|}|d}|||}||d<||krpt|\} } | |d<| |d<|t|q|D]*} t| \} } |t | || t | fq|D]} ||| ddfq|S)z_ :param installed: A map from archive RECORD path to installation RECORD path. rQz,RECORD line has more than three elements: {}rr1rT) rBloggerwarningrlistpoprPrrr<rO) Z old_csv_rows installedchanged generatedlib_dirZinstalled_rowsrZold_pathnew_pathrLrFrGr:r:r;get_csv_rows_for_installeds*     rc@s eZdZdS)MissingCallableSuffixN)__name__ __module__ __qualname__r:r:r:r;r=srcCs*t|}|dk r&|jdkr&tt|dSr)rsuffixrrO) specificationentryr:r:r;_raise_for_invalid_entrypointAsrcseZdZdfdd ZZS)PipScriptMakerNcst|tt|||Sr)rsuperrmake)selfrZoptions __class__r:r;rHszPipScriptMaker.make)N)rrrr __classcell__r:r:rr;rGsrFTc - s\|st|||||| d}t| r,|dn|dgg tjjtjj} itg} |rt4} t t dt j | dddW5QRXW5QRXt | d. fdd d/fd d }|| dstd tjdd}t|\fdd}D]j}d }d }ttj |D]F}d }|dkr`t}|}tj ||} ||}|| |d||dqFq&td |d}d|_dh|_d|_g}dd }|rDdtjkr|d|tjdddkr|dtjd|f|dt|fddD}|D] }|=q6dd }|rdtjkrp|d||dt|fddD}|D] }|=q| d d!!Dd"d!D}g}z.|"|}| || |"|d#diWn>t#k r>}z|j$d} t%d$&| W5d }~XYnX| rbt'|}!|!d k rbt (|!tjdd%}"tjdd&}#t)|#d'}$|$*d(W5QRXt+,|#|"| |"tjdd)}%tjdd*}&t-|%d+\}'t-|&d,F}(t./|'})t0|)| d-}*t.1|(}+t2|*D]},|+3|,q&W5QRXW5QRXt+,|&|%d S)0zInstall a wheel)userhomerootisolatedprefixZpurelibZplatlibignoreT)forcequietFcs.t|}t|}||<|r*|dS)z6Map archive RECORD paths to installation RECORD paths.N)r<r)srcfilerZmodifiedZoldpathnewpath)rrrrwr:r;record_installeds   z*move_wheel_files..record_installedNcst|t|D]\}}}|t|dtjj}tj||} |rf|tjjdd drfq|D]} tj||| } |r|dkr| dr | qjqj|rj| drjt |  t j rjrtd| dd | qj|D]} |r|| rqtj|| } tj||| }t| tj|rFt|t| |t| }ttdr|t||j|jft| tjrt| }|jtjBtjBtjB}t||d }|r||}| ||qqdS) Nr1rz.datarTz .dist-infoz!Multiple .dist-info directories: rutimeF)rr3walkrBlstripr4r7r_r^endswithrrrfrWAssertionErrorrunlinkshutilZcopyfilestathasattrrst_atimest_mtimeaccessX_OKst_modeS_IXUSRS_IXGRPS_IXOTHchmod)sourcedestZis_basefixerfilterdirZsubdirsfilesZbasedirZdestdirr~Z destsubdirrGrrstZ permissionsr) data_dirsinfo_dirrreqr:r;clobbers^         z!move_wheel_files..clobberz!%s .dist-info directory not foundrzentry_points.txtcsh|dr|dd}n<|dr8|dd}n |drT|dd}n|}|kpf|kS)Nz.exez -script.pyiz.pya)rvr)rWZ matchname)rrr:r;is_entrypoint_wrappersz/move_wheel_files..is_entrypoint_wrapperr)rrrTZpipZENSUREPIP_OPTIONSzpip = Z altinstallz pip%s = %scSsg|]}td|r|qS)zpip(\d(\.\d)?)?$rertrkr:r:r;r0s z$move_wheel_files..Z easy_installzeasy_install = zeasy_install-%s = %scSsg|]}td|r|qS)zeasy_install(-\d\.\d)?$rrr:r:r;r@s css|]}d|VqdS)%s = %sNr:rZkvr:r:r;rGsz#move_wheel_files..cSsg|] }d|qS)rr:rr:r:r;rKsrzInvalid script entry point: {} for req: {} - A callable suffix is required. Cf https://packaging.python.org/en/latest/distributing.html#console-scripts for more information.Z INSTALLERz INSTALLER.piprcspip ZRECORDz RECORD.piprzw+)rrrr)F)NN)4rr{rNr3r4r7rrwarningscatch_warningsfilterwarnings compileall compile_dirrdebuggetvaluerr_rrrrorrZvariantsZset_moderrrrrUrVrextendrZ make_multiplerargsr rrrrArlrmover[csvreaderrwriterrZwriterow)-rWrrwrrrZ pycompileZschemerrZwarn_script_locationrrstdoutrZep_filerZdatadirrrZsubdirrZmakerZscripts_to_generateZ pip_scriptZpip_eprZeasy_install_scriptZeasy_install_epZgui_scripts_to_generateZgenerated_console_scriptsermsgZ installerZtemp_installerZinstaller_filerecordZ temp_recordZ record_inZ record_outr rr rr:) rrrrrrrrrrwr;move_wheel_filesMs    $F   #                     "rcCsrzVddtd|Dd}|d}t|}|d}ttt| d}|WSt k rlYdSXdS)z Return the Wheel-Version of an extracted wheel, if possible. Otherwise, return None if we couldn't parse / extract it. cSsg|]}|qSr:r:)rdr:r:r;rsz!wheel_version..Nrrqz Wheel-Version.) rZ find_on_pathZ get_metadatarZparsestrrrmapintr^ Exception) source_dirZdistZ wheel_dataversionr:r:r; wheel_versions   rcCsb|std||dtdkr>td|dtt|fn |tkr^tddtt|dS)a Raises errors or warns if called with an incompatible Wheel-Version. Pip should refuse to install a Wheel-Version that's a major series ahead of what it's compatible with (e.g 2.0 > 1.1); and warn when installing a version only minor version ahead (e.g 1.2 > 1.1). version: a 2-tuple representing a Wheel-Version (Major, Minor) name: name of wheel or package to raise exception about :raises UnsupportedWheel: when an incompatible Wheel-Version is given z(%s is in an unsupported or invalid wheelrzB%s's Wheel-Version (%s) is not compatible with this version of piprz*Installing from a newer Wheel-Version (%s)N)r VERSION_COMPATIBLEr_rrOrr)rrWr:r:r;check_compatibilitysrcCs d|S)z Format three tags in the form "--". :param file_tag: A 3-tuple of tags (python_tag, abi_tag, platform_tag). r\)r_)Zfile_tagr:r:r; format_tagsrc@s>eZdZdZedejZddZddZ ddZ d d Z d S) Wheelz A wheel filez^(?P(?P.+?)-(?P.*?)) ((-(?P\d[^-]*?))?-(?P.+?)-(?P.+?)-(?P.+?) \.whl|\.dist-info)$csj|}|std||_|ddd_|ddd_|d_|d d_ |d  d_ |d  d_ fd d j D_ d S)zX :raises InvalidWheelFilename: when the filename is invalid for a wheel z!%s is not a valid wheel filename.rWrpr\ZverbuildZpyverrZabiZplatcs.h|]&}jD]}jD]}|||fqqqSr:)abisplats)rryzrr:r; s z!Wheel.__init__..N) wheel_file_rertr rrur6rWrZ build_tagr^Z pyversionsrr file_tags)rrZ wheel_infor:r!r;__init__s   zWheel.__init__cCstdd|jDS)zF Return the wheel's tags as a sorted list of strings. css|]}t|VqdSr)rrtagr:r:r;rsz0Wheel.get_formatted_file_tags..)rr$r!r:r:r;get_formatted_file_tagsszWheel.get_formatted_file_tagscstfdd|jDS)a Return the lowest index that one of the wheel's file_tag combinations achieves in the given list of supported tags. For example, if there are 8 supported tags and one of the file tags is first in the list, then return 0. :param tags: the PEP 425 tags to check the wheel against, in order with most preferred first. :raises ValueError: If none of the wheel's file tags match one of the supported tags. c3s |]}|kr|VqdSr)indexr&tagsr:r;rsz*Wheel.support_index_min..)minr$rr+r:r*r;support_index_minszWheel.support_index_mincCs|j| S)z Return whether the wheel is compatible with one of the given tags. :param tags: the PEP 425 tags to check the wheel against. )r$ isdisjointr-r:r:r; supportedszWheel.supportedN) rrr__doc__rcompileVERBOSEr#r%r(r.r0r:r:r:r;rsrz([a-z0-9_.]+)-([a-z0-9_.!+-]+)cCst||S)zjDetermine whether the string looks like an egg_info. :param s: The string to parse. E.g. foo-2.1 )boolsearch)r~Z _egg_info_rer:r:r;_contains_egg_infosr6cCs|jr dS|jr&|s"td|jdS|s.dS|js:|js>dS||sXtd|jdS|jrj|jjrjdS|j}| \}}|rt |rdSdS)a[ Return whether to build an InstallRequirement object using the ephemeral cache. :param cache_available: whether a cache directory is available for the should_unpack=True case. :return: True or False to build the requirement with ephem_cache=True or False, respectively; or None not to build the requirement. Nz(Skipping %s, due to already being wheel.FzCSkipping wheel build for %s, due to binaries being disabled for it.T) Z constraintis_wheelrinforWZeditablerlinkZis_vcssplitextr6)r should_unpackcache_availablecheck_binary_allowedr9baseZextr:r:r;should_use_ephemeral_cache s4   r?cCs^t|}d|}|s |d7}n:ttjkr8|d7}n"|dsJ|d7}|d|t7}|S)z1 Format command information for logging. zCommand arguments: {} zCommand output: Nonez'Command output: [use --verbose to show]r|zCommand output: {}{})rrrZgetEffectiveLevelloggingDEBUGrr) command_argscommand_outputZ command_desctextr:r:r;format_command_resultGs    rEcCsxt|}|s4d|j}|t||7}t|dSt|dkrfd|j|}|t||7}t|tj ||dS)zH Return the path to the wheel in the temporary build directory. z1Legacy build of wheel for {!r} created no files. Nr1zZLegacy build of wheel for {!r} created more than one file. Filenames (choosing first): {} r) rrrWrErrrBr3r4r_)namestemp_dirrrBrCr r:r:r;get_legacy_build_wheel_path^s$    rHcCsdS)NTr:)rpr:r:r; _always_true~srIc@s\eZdZdZdddZdddZddd Zd d Zdd d ZdddZ ddZ dddZ dS) WheelBuilderz#Build wheels from a RequirementSet.NFcCsD|dkr t}||_||_|j|_|p&g|_|p0g|_||_||_dSr) rIpreparer wheel_cacheZwheel_download_dir _wheel_dir build_optionsglobal_optionsr=no_clean)rrKrLrNrOr=rPr:r:r;r%s   zWheelBuilder.__init__c Cs.|j|j|||dW5QRSQRXdS)ziBuild one wheel. :return: The filename of the built wheel, or None if the build failed.  python_tagN)Z build_env_build_one_inside_env)rr output_dirrRr:r:r; _build_oneszWheelBuilder._build_onec Cstdd}|jr|j}n|j}|||j|d}|dk rtj|}tj||}zNt|\} } t ||t d|j || | t d||WW5QRStk rYnX||W5QRdSQRXdS)Nry)ZkindrQz3Created wheel for %s: filename=%s size=%d sha256=%szStored in directory: %s)rZ use_pep517_build_one_pep517_build_one_legacyr4r3rr_rIrrrr8rWZ hexdigestr _clean_one) rrrTrRrGZbuilder wheel_path wheel_nameZ dest_pathZ wheel_hashrFr:r:r;rSs.      z"WheelBuilder._build_one_inside_envcCst|j|jddS)NT)rOZunbuffered_output)rZ setup_py_pathrO)rrr:r:r;_base_setup_argss zWheelBuilder._base_setup_argsc Cs|jdk st|jr*td|jfdSz~td|td|j}|j }| ||j ||jd}W5QRX|rt ||}t t j||t j|||}Wn$tk rtd|jYdSXt j||S)zBuild one InstallRequirement using the PEP 517 build process. Returns path to wheel if successfully built. Otherwise, returns None. NzGCannot build wheel for %s using PEP 517 when --build-options is presentDestination directory: %szBuilding wheel for {} (PEP 517))metadata_directoryFailed building wheel for %s)r]rrNrerrorrWrrrZpep517_backendZsubprocess_runnerZ build_wheelrar3renamer4r_r)rrtempdrRZrunnerZbackendrZnew_namer:r:r;rVs8      zWheelBuilder._build_one_pep517c Cs||}d|jf}t|}td||dd|g|j}|dk rT|d|g7}zt||j|d}Wn8tk r| dt d |jYW5QRdSXt |} t | ||||d } | W5QRSQRXdS) zBuild one InstallRequirement using the "legacy" build process. Returns path to wheel if successfully built. Otherwise, returns None. z Building wheel for %s (setup.py)r\Z bdist_wheelz-dNz --python-tag)cwdspinnerr_r^)rFrGrrBrC)r[rWrrrrNrZunpacked_source_directoryrZfinishr_r3rrrH) rrrarR base_argsZ spin_messagerdZ wheel_argsoutputrFrYr:r:r;rWs8         zWheelBuilder._build_one_legacycCsb||}td|j|ddg}zt||jdWdStk r\td|jYdSXdS)NzRunning setup.py clean for %sZcleanz--all)rcTz Failed cleaning build dir for %sF)r[rr8rWrrrr_)rrreZ clean_argsr:r:r;rXs  zWheelBuilder._clean_onec Cs|r |jr|s|jstg}t|jj}|D]\}t||||jd}|dkrLq,|rt|rd|j|j}qz|j |j}n|j}| ||fq,|sgSt dd dd|Dd}|rtj}tgg} } |D]\}}z t|WnFtk r.} z&t d|j| | |WYqW5d} ~ XYnX|j|||d} | r| ||r|jrrt|jsrtd |||jj|_tt| |_|jjstt|jj |jq| |qW5QRX| rt d d d d| D| rt d d dd| D| S)aBuild wheels. :param should_unpack: If True, after building the wheel, unpack it and replace the sdist with the unpacked version in preparation for installation. :return: True if all the wheels built correctly. )r;r<r=Nz*Building wheels for collected packages: %srcSsg|]\}}|jqSr:rW)rrrpr:r:r;r\sz&WheelBuilder.build..z Building wheel for %s failed: %srQzbad source dir - missing markerzSuccessfully built %sr}cSsg|] }|jqSr:rgrrr:r:r;rszFailed to build %scSsg|] }|jqSr:rgrhr:r:r;rs)!rMrr4rL cache_dirr?r=Zget_ephem_path_for_linkr9Zget_path_for_linkrrr8r_r Zimplementation_tagrrOSErrorrrWrUrrZremove_temporary_sourceZensure_build_locationrKZ build_dirrrr7rZ file_path) rZ requirementsr;Zbuildsetr<rZ ephem_cacherTrRZ build_successZ build_failurer Z wheel_filer:r:r;r#s       zWheelBuilder.build)NNNF)N)N)N)N)F) rrrr1r%rUrSr[rVrWrXrr:r:r:r;rJs   ' %rJ)r=)r=)FNNTNFNT)r1Z __future__rrrrr@r@Zos.pathr3rrrrUrbase64rZ email.parserrZ pip._vendorrZpip._vendor.distlib.scriptsrZpip._vendor.distlib.utilrZpip._vendor.packaging.utilsrZpip._vendor.sixr Z pip._internalr Zpip._internal.exceptionsr r r Zpip._internal.locationsrrZpip._internal.models.linkrZpip._internal.utils.loggingrZ pip._internal.utils.marker_filesrZpip._internal.utils.miscrrrZ$pip._internal.utils.setuptools_buildrZpip._internal.utils.subprocessrrrrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.uirZpip._internal.utils.unpackingrZpip._internal.utils.urlsrtypingr r!r"r#r$r%r&r'r(r)r*r+Z"pip._vendor.packaging.requirementsr,Zpip._internal.req.req_installr-Z pip._internal.operations.preparer.Zpip._internal.cacher/Zpip._internal.pep425tagsr0rOZInstalledCSVRowr4ZBinaryAllowedPredicaterZ getLoggerrrr<rIrPr[raror2r3rsr{rrrrrrrrrrrrobjectrIr6r?rErHrIrJr:r:r:r;s                  8           =$  4 J : site-packages/pip/_internal/__pycache__/legacy_resolve.cpython-38.pyc000064400000024404151464452710021707 0ustar00U .eRC@sFdZddlZddlZddlmZddlmZddlmZddl m Z m Z m Z m Z mZddlmZddlmZmZmZmZdd lmZmZdd lmZerdd lmZmZmZmZm Z m!Z!dd l"m#Z#dd l$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/ee0e-ge-fZ1e2e3Z4dddZ5Gddde6Z7dS)ayDependency Resolution The dependency resolution in pip is performed as follows: for top-level requirements: a. only one spec allowed per project, regardless of conflicts or not. otherwise a "double requirement" exception is raised b. they override sub-dependency requirements. for sub-dependencies a. "first found, wins" (where the order is breadth first) N) defaultdict)chain) specifiers)BestVersionAlreadyInstalledDistributionNotFound HashError HashErrorsUnsupportedPythonVersion) indent_log)dist_in_install_pathdist_in_usersite ensure_dirnormalize_version_info)check_requires_pythonget_requires_python)MYPY_CHECK_RUNNING)Callable DefaultDictListOptionalSetTuple) pkg_resources)AbstractDistribution) PipSession) PackageFinder)RequirementPreparer)InstallRequirement)RequirementSetFc Cst|}zt||d}Wn:tjk rR}ztd|j|WYdSd}~XYnX|r\dSdtt |}|rt d|j||dSt d |j||dS)a Check whether the given Python version is compatible with a distribution's "Requires-Python" value. :param version_info: A 3-tuple of ints representing the Python major-minor-micro version to check. :param ignore_requires_python: Whether to ignore the "Requires-Python" value if the given Python version isn't compatible. :raises UnsupportedPythonVersion: When the given Python version isn't compatible. ) version_infoz-Package %r has an invalid Requires-Python: %sN.zBIgnoring failed Requires-Python check for package %r: %s not in %rz8Package {!r} requires a different Python: {} not in {!r}) rrrZInvalidSpecifierloggerwarningZ project_namejoinmapstrdebugr format)distrignore_requires_pythonZrequires_pythonZ is_compatibleexcversionr,@/usr/lib/python3.8/site-packages/pip/_internal/legacy_resolve.py_check_dist_requires_python>s> r.csdeZdZdZdddhZdfdd Zdd Zd d Zd d ZddZ ddZ ddZ ddZ Z S)ResolverzResolves which packages need to be installed/uninstalled to perform the requested operation without breaking the requirements of any package. eageronly-if-neededto-satisfy-onlyNc stt|| |jkst| dkr4tjdd} nt| } | |_||_ ||_ ||_ d|_ | |_ | |_||_||_||_||_||_tt|_dS)N)superr/__init___allowed_strategiesAssertionErrorsysrr_py_version_infopreparerfindersessionrequire_hashesupgrade_strategyforce_reinstallignore_dependenciesignore_installedr) use_user_site_make_install_reqrlist_discovered_dependencies) selfr:r<r;Zmake_install_reqrBr@rAr)r?r>Zpy_version_info __class__r,r-r5us&zResolver.__init__c Cs|jjrt|jj|jt|j}|jp>tdd|D|_|j j }| }|r`t |g}t}t||D]N}z||||Wqttk r}z||_||W5d}~XYqtXqt|r|dS)aResolve what operations need to be done As a side-effect of this method, the packages (and their dependencies) are downloaded, unpacked and prepared for installation. This preparation is done by ``pip.operations.prepare``. Once PyPI has static dependency metadata available, it would be possible to move the preparation to become a step separated from dependency resolution. css|] }|jVqdSN)Zhas_hash_options).0reqr,r,r- sz#Resolver.resolve..N)r:Zwheel_download_dirr Zunnamed_requirementsrD requirementsvaluesr=anyr; search_scopeZget_formatted_locationsr!inforrextend _resolve_onerrKappend) rFrequirement_setZ root_reqsrPZ locationsZdiscovered_reqsZ hash_errorsrKr*r,r,r-resolves2     zResolver.resolvecCs4|jdkrdS|jdkrdS|jdks*t|jSdS)Nr2Fr0Tr1)r>r7 is_directrFrKr,r,r-_is_upgrade_alloweds   zResolver._is_upgrade_allowedcCs,|jrt|jr"t|jr"|j|_d|_dS)z4 Set a requirement to be installed. N)rBr satisfied_byr Zconflicts_withrXr,r,r-_set_req_to_reinstallszResolver._set_req_to_reinstallcCs|jr dS||j|js dS|jr4||dS||sP|jdkrLdSdS|jsz|j j |ddWn(t k rYdSt k rYnX||dS)aCheck if req_to_install should be skipped. This will check if the req is installed, and whether we should upgrade or reinstall it, taking into account all the relevant user options. After calling this req_to_install will only have satisfied_by set to None if the req_to_install is to be upgraded/reinstalled etc. Any other value will be a dist recording the current thing installed that satisfies the requirement. Note that for vcs urls and the like we can't assess skipping in this routine - we simply identify that we need to pull the thing down, then later on it is pulled down and introspected to assess upgrade/ reinstalls etc. :return: A text reason for why it was skipped, or None. Nr1z#already satisfied, skipping upgradezalready satisfiedT)Zupgradezalready up-to-date) rAcheck_if_existsrBrZr?r[rYr>linkr;Zfind_requirementrr)rFreq_to_installr,r,r-_check_skip_installeds*     zResolver._check_skip_installedcCs|jdk std|jr0|j||j|j|jS|jdks>t||}|jr`|j ||j|S| |}| |j||j|j ||j |j|j}|js||j|jr|jdkp|jp|jp|jjdk}|r||n td||S)zzTakes a InstallRequirement and returns a single AbstractDist representing a prepared variant of the same. Nz9require_hashes should have been set in Resolver.resolve()r2filezr?r]Zschemer[r!rQ)rFrKZ skip_reasonZupgrade_allowed abstract_distZ should_modifyr,r,r-_get_abstract_dist_forsV        zResolver._get_abstract_dist_forc s4js jrgSd_j}|}t|jjdgfdd}t  j sd_ j ddjs jrtddjttjt|j}|D]}td ||qtt|jtj@}||D]} || |d qjs&js&jW5QRXS) zxPrepare a single requirements file. :return: A list of additional InstallRequirements to also install. T)rr)csPt|}j}j|||d\}}|rB|rBj|||dS)N)parent_req_nameextras_requested)rCr%nameadd_requirementrErTrR)subreqreZsub_install_reqrdZ to_scan_againZ add_to_parentZ more_reqsr^rUrFr,r-add_req}s  z&Resolver._resolve_one..add_reqN)rdz!Installing extra requirements: %r,z"%s does not provide the extra '%s')re) constraintZpreparedZreqs_to_cleanuprTrcZget_pkg_resources_distributionr.r9r)r Zhas_requirementrfrWrgr@Zextrasr!r&r#sortedsetr"ZrequiresrarZZsuccessfully_downloaded) rFrUr^rbr(rjZmissing_requestedZmissingZavailable_requestedrhr,rir-rS[sV     zResolver._resolve_onecs8gtfdd|jD] }|q&S)zCreate the installation order. The installation order is topological - requirements are installed before the requiring thing. We break cycles at an arbitrary point, and make no other guarantees. csN|js|krdS|jrdS|j|jD] }|q2|dSrI)rZrladdrErfrT)rKZdeporderZ ordered_reqsschedulerFr,r-rrs  z1Resolver.get_installation_order..schedule)rnrMrN)rFZreq_setZ install_reqr,rpr-get_installation_orders   zResolver.get_installation_order)N)__name__ __module__ __qualname____doc__r6r5rVrYr[r_rcrSrs __classcell__r,r,rGr-r/ns +3  5<Zr/)F)8rwZloggingr8 collectionsr itertoolsrZpip._vendor.packagingrZpip._internal.exceptionsrrrrr Zpip._internal.utils.loggingr Zpip._internal.utils.miscr r r rZpip._internal.utils.packagingrrZpip._internal.utils.typingrtypingrrrrrrZ pip._vendorrZpip._internal.distributionsrZpip._internal.network.sessionrZpip._internal.indexrZ pip._internal.operations.preparerZpip._internal.req.req_installrZpip._internal.req.req_setrr%ZInstallRequirementProviderZ getLoggerrtr!r.objectr/r,r,r,r-s4                0site-packages/pip/_internal/__pycache__/pep425tags.cpython-38.pyc000064400000023072151464452710020602 0ustar00U .eE>@sTdZddlmZddlZddlZddlZddlZddlZddl Z ddl Z ddl m Z ddl ZddlmZddlmZerddlmZmZmZmZmZmZmZeeeefZeeZe dZ!d d Z"d d Z#d dZ$ddZ%ddZ&ddZ'd-ddZ(ddZ)ddZ*ddZ+ddZ,d d!Z-d"d#Z.d$d%Z/d&d'Z0d(d)Z1d.d+d,Z2e'Z3dS)/z2Generate and work with PEP 425 Compatibility Tags.)absolute_importN) OrderedDict)get_extension_suffixes)MYPY_CHECK_RUNNING)TupleCallableListOptionalUnionDictSetz(.+)_(\d+)_(\d+)_(.+)c CsLz t|WStk rF}ztd|tWYdSd}~XYnXdS)Nz{}) sysconfigget_config_varIOErrorwarningswarnformatRuntimeWarning)varerConfig variable '%s' is unset, Python ABI tag may be incorrect)rloggerdebug)rZfallbackexpectedrvalrrrget_flagYsr2cstd}t}d}|s|dkrttdrd}d}d}|dktddd d rRd }tjd krvtd fdd d rvd}tjdkrtddd ddrd}d|t|||f}n:|r|drd|dd}n|r| dd dd}|S)zXReturn the ABI tag based on SOABI (if available) or emulate SOABI (CPython 2, PyPy).ZSOABIN>rr maxunicoder!rZPy_DEBUGcSs ttdS)NZgettotalrefcount)rrrrrrtzget_abi_tag..)rd)Z WITH_PYMALLOCcsS)NrrZ is_cpythonrrr4xr5m)r7r7ZPy_UNICODE_SIZEcSs tjdkS)Ni)rr3rrrrr4{r5)r0ruz %s%s%s%s%szcpython--r*._) rr rrr2r&r)rsplitreplace)Zsoabiimplabir6r:r<rr9r get_abi_tagfsB   rDcCs tjdkS)Ni)rmaxsizerrrr_is_running_32bitsrFcCstjdkr^t\}}}|d}|dkr6tr6d}n|dkrHtrHd}d|d|d |Stj dd  d d }|d krtrd }|S)z0Return our platform name 'win32', 'linux_x86_64'darwinr>x86_64i386ppc64ppczmacosx_{}_{}_{}rr*r?r= linux_x86_64 linux_i686) rrZmac_verr@rFr distutilsutil get_platformrA)releaser?machineZ split_verresultrrrrPs  rPc Cs$tdkrdSz&ttjd}|d}W5QRXWntttfk rPYdSX|dksft|dkrjdSt |t rdd|D}ndd|D}|dd d d d d gk}||d ddgkM}||dddgkM}||ddddgkM}||dddgkM}||dddd @d kM}|S)N linux_armv7lFrb(cSsg|] }t|qSr)ord).0crrr sz"is_linux_armhf..cSsg|]}|qSrr)rXbrrrrZsrr;ELFr*'%&) rPopenr executablereadrOSError TypeErrorlen isinstancer%)fZelf_header_rawZ elf_headerrSrrris_linux_armhfs&  roc CsNtdkrdSzddl}t|jWSttfk r:YnXtjjj ddS)NrMrLFrr"r`) rP _manylinuxboolZmanylinux1_compatible ImportErrorAttributeErrorpip _internalutilsglibchave_compatible_glibcrqrrris_manylinux1_compatibles  r{c CsNtdkrdSzddl}t|jWSttfk r:YnXtjjj ddS)NrpFrr" ) rPrqrrZmanylinux2010_compatiblersrtrurvrwrxryrzrrris_manylinux2010_compatibles  r}c Csdt}|dkrdS|dkr$ts$dSzddl}t|jWSttfk rPYnXtjj j ddS)N>rM linux_s390x linux_ppc64lerLrT linux_aarch64 linux_ppc64FrTrr") rProrqrrZmanylinux2014_compatiblersrtrurvrwrxry)rrqrrris_manylinux2014_compatibles rcsrg}fddtddddg|||r8||D]&}||kr<|||r<||q<|d|S)zReturn a list of supported arches (including group arches) for the given major, minor and machine architecture of an macOS machine. cs||dkr||fdkS|dkr(||fdkS|dkr<||fdkS|dkrP||fdkS|krx|D]}|||r`dSq`dS) NrK) r`rJrI)rr;rHTFr)r+r,archgarch_supports_archgroupsrrrs      z)get_darwin_arches.._supports_arch)Zfat)rIrK)Zintel)rHrI)Zfat64)rHrJ)Zfat32)rHrIrKZ universal)rappend)r+r,rRarchesrrrrget_darwin_archess%    rc CsFg}|dd}t|dddD] }|dtt||fq |S)Nr!)rangerr#r$r%)r&versionsr+r,rrr!get_all_minor_versions_as_strings>s  rFcCsDg}|dkrt}t|}|p"t}g}|p0t}|rD|g|dd<t}tD]$} | drP|| dddqP| t t || d|s|pt } | d\} } } | d r>t| }|r6|\}}}}d ||}g}ttt|dD]0}tt|||D]}| |||fqqn| g}n| d kr~| g}| d kr| d | | | d| | n| d kr| d| | g}nn|dkrg}tr| d | | tr| d | | tr| d| | | | n| g}|D].}|D]"} | d||df|| fqq |ddD]F}|dkr\q|D]*}|D]} | d||f|| fqhq`qH|D]"} | d|ddd| fq| d||dfddf| d||ddfddft|D]B\}}| d|fddf|dkr| d|dddfq|S)acReturn a list of supported tags for each version specified in `versions`. :param versions: a list of string versions, of the form ["33", "32"], or None. The first version will be assumed to support our ABI. :param platform: specify the exact platform you want valid tags for, or None. If None, use the local system platform. :param impl: specify the exact implementation you want valid tags for, or None. If None, use the local interpreter impl. :param abi: specify the exact abi you want valid tags for, or None. If None, use the local interpreter abi. Nrz.abir>r"r*Znoner?Zmacosxz {}_{}_%i_%sZ manylinux2014>rHi686Z manylinux2010Z manylinux1z%s%s>3130zpy%sany)r(rr rDsetrraddr@extendsortedlistrrP partition _osx_arch_patmatchrrreversedrintrrr}r{ enumerate)rZnoarchrrBrCZ supportedr&ZabisZabi3ssuffixrZ arch_prefixZarch_sepZ arch_suffixrnamer+r,Z actual_archZtplrr:aversionirrr get_supportedHs~              $ $   r)TT)NFNNN)4__doc__Z __future__rZdistutils.utilrNZloggingrrerr r collectionsrZpip._internal.utils.glibcruZpip._internal.utils.compatrZpip._internal.utils.typingrtypingrrrr r r r r%Z Pep425TagZ getLogger__name__r.compilerrr r'r)r(r-r2rDrFrPror{r}rrrrZimplementation_tagrrrrsP    $     !?  ysite-packages/pip/_internal/exceptions.py000064400000024007151464452710014576 0ustar00"""Exceptions used throughout package""" # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import from itertools import chain, groupby, repeat from pip._vendor.six import iteritems from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional from pip._vendor.pkg_resources import Distribution from pip._internal.req.req_install import InstallRequirement class PipError(Exception): """Base pip exception""" class ConfigurationError(PipError): """General exception in configuration""" class InstallationError(PipError): """General exception during installation""" class UninstallationError(PipError): """General exception during uninstallation""" class NoneMetadataError(PipError): """ Raised when accessing "METADATA" or "PKG-INFO" metadata for a pip._vendor.pkg_resources.Distribution object and `dist.has_metadata('METADATA')` returns True but `dist.get_metadata('METADATA')` returns None (and similarly for "PKG-INFO"). """ def __init__(self, dist, metadata_name): # type: (Distribution, str) -> None """ :param dist: A Distribution object. :param metadata_name: The name of the metadata being accessed (can be "METADATA" or "PKG-INFO"). """ self.dist = dist self.metadata_name = metadata_name def __str__(self): # type: () -> str # Use `dist` in the error message because its stringification # includes more information, like the version and location. return ( 'None {} metadata found for distribution: {}'.format( self.metadata_name, self.dist, ) ) class DistributionNotFound(InstallationError): """Raised when a distribution cannot be found to satisfy a requirement""" class RequirementsFileParseError(InstallationError): """Raised when a general error occurs parsing a requirements file line.""" class BestVersionAlreadyInstalled(PipError): """Raised when the most up-to-date version of a package is already installed.""" class BadCommand(PipError): """Raised when virtualenv or a command is not found""" class CommandError(PipError): """Raised when there is an error in command-line arguments""" class PreviousBuildDirError(PipError): """Raised when there's a previous conflicting build directory""" class InvalidWheelFilename(InstallationError): """Invalid wheel filename.""" class UnsupportedWheel(InstallationError): """Unsupported wheel.""" class HashErrors(InstallationError): """Multiple HashError instances rolled into one for reporting""" def __init__(self): self.errors = [] def append(self, error): self.errors.append(error) def __str__(self): lines = [] self.errors.sort(key=lambda e: e.order) for cls, errors_of_cls in groupby(self.errors, lambda e: e.__class__): lines.append(cls.head) lines.extend(e.body() for e in errors_of_cls) if lines: return '\n'.join(lines) def __nonzero__(self): return bool(self.errors) def __bool__(self): return self.__nonzero__() class HashError(InstallationError): """ A failure to verify a package against known-good hashes :cvar order: An int sorting hash exception classes by difficulty of recovery (lower being harder), so the user doesn't bother fretting about unpinned packages when he has deeper issues, like VCS dependencies, to deal with. Also keeps error reports in a deterministic order. :cvar head: A section heading for display above potentially many exceptions of this kind :ivar req: The InstallRequirement that triggered this error. This is pasted on after the exception is instantiated, because it's not typically available earlier. """ req = None # type: Optional[InstallRequirement] head = '' def body(self): """Return a summary of me for display under the heading. This default implementation simply prints a description of the triggering requirement. :param req: The InstallRequirement that provoked this error, with populate_link() having already been called """ return ' %s' % self._requirement_name() def __str__(self): return '%s\n%s' % (self.head, self.body()) def _requirement_name(self): """Return a description of the requirement that triggered me. This default implementation returns long description of the req, with line numbers """ return str(self.req) if self.req else 'unknown package' class VcsHashUnsupported(HashError): """A hash was provided for a version-control-system-based requirement, but we don't have a method for hashing those.""" order = 0 head = ("Can't verify hashes for these requirements because we don't " "have a way to hash version control repositories:") class DirectoryUrlHashUnsupported(HashError): """A hash was provided for a version-control-system-based requirement, but we don't have a method for hashing those.""" order = 1 head = ("Can't verify hashes for these file:// requirements because they " "point to directories:") class HashMissing(HashError): """A hash was needed for a requirement but is absent.""" order = 2 head = ('Hashes are required in --require-hashes mode, but they are ' 'missing from some requirements. Here is a list of those ' 'requirements along with the hashes their downloaded archives ' 'actually had. Add lines like these to your requirements files to ' 'prevent tampering. (If you did not enable --require-hashes ' 'manually, note that it turns on automatically when any package ' 'has a hash.)') def __init__(self, gotten_hash): """ :param gotten_hash: The hash of the (possibly malicious) archive we just downloaded """ self.gotten_hash = gotten_hash def body(self): # Dodge circular import. from pip._internal.utils.hashes import FAVORITE_HASH package = None if self.req: # In the case of URL-based requirements, display the original URL # seen in the requirements file rather than the package name, # so the output can be directly copied into the requirements file. package = (self.req.original_link if self.req.original_link # In case someone feeds something downright stupid # to InstallRequirement's constructor. else getattr(self.req, 'req', None)) return ' %s --hash=%s:%s' % (package or 'unknown package', FAVORITE_HASH, self.gotten_hash) class HashUnpinned(HashError): """A requirement had a hash specified but was not pinned to a specific version.""" order = 3 head = ('In --require-hashes mode, all requirements must have their ' 'versions pinned with ==. These do not:') class HashMismatch(HashError): """ Distribution file hash values don't match. :ivar package_name: The name of the package that triggered the hash mismatch. Feel free to write to this after the exception is raise to improve its error message. """ order = 4 head = ('THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS ' 'FILE. If you have updated the package versions, please update ' 'the hashes. Otherwise, examine the package contents carefully; ' 'someone may have tampered with them.') def __init__(self, allowed, gots): """ :param allowed: A dict of algorithm names pointing to lists of allowed hex digests :param gots: A dict of algorithm names pointing to hashes we actually got from the files under suspicion """ self.allowed = allowed self.gots = gots def body(self): return ' %s:\n%s' % (self._requirement_name(), self._hash_comparison()) def _hash_comparison(self): """ Return a comparison of actual and expected hash values. Example:: Expected sha256 abcdeabcdeabcdeabcdeabcdeabcdeabcdeabcdeabcde or 123451234512345123451234512345123451234512345 Got bcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdef """ def hash_then_or(hash_name): # For now, all the decent hashes have 6-char names, so we can get # away with hard-coding space literals. return chain([hash_name], repeat(' or')) lines = [] for hash_name, expecteds in iteritems(self.allowed): prefix = hash_then_or(hash_name) lines.extend((' Expected %s %s' % (next(prefix), e)) for e in expecteds) lines.append(' Got %s\n' % self.gots[hash_name].hexdigest()) return '\n'.join(lines) class UnsupportedPythonVersion(InstallationError): """Unsupported python version according to Requires-Python package metadata.""" class ConfigurationFileCouldNotBeLoaded(ConfigurationError): """When there are errors while loading a configuration file """ def __init__(self, reason="could not be loaded", fname=None, error=None): super(ConfigurationFileCouldNotBeLoaded, self).__init__(error) self.reason = reason self.fname = fname self.error = error def __str__(self): if self.fname is not None: message_part = " in {}.".format(self.fname) else: assert self.error is not None message_part = ".\n{}\n".format(self.error.message) return "Configuration file {}{}".format(self.reason, message_part) site-packages/pip/_internal/wheel.py000064400000124110151464452710013515 0ustar00""" Support for installing and building the "wheel" binary package format. """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import collections import compileall import csv import hashlib import logging import os.path import re import shutil import stat import sys import warnings from base64 import urlsafe_b64encode from email.parser import Parser from pip._vendor import pkg_resources from pip._vendor.distlib.scripts import ScriptMaker from pip._vendor.distlib.util import get_export_entry from pip._vendor.packaging.utils import canonicalize_name from pip._vendor.six import StringIO from pip._internal import pep425tags from pip._internal.exceptions import ( InstallationError, InvalidWheelFilename, UnsupportedWheel, ) from pip._internal.locations import distutils_scheme, get_major_minor_version from pip._internal.models.link import Link from pip._internal.utils.logging import indent_log from pip._internal.utils.marker_files import has_delete_marker_file from pip._internal.utils.misc import captured_stdout, ensure_dir, read_chunks from pip._internal.utils.setuptools_build import make_setuptools_shim_args from pip._internal.utils.subprocess import ( LOG_DIVIDER, call_subprocess, format_command_args, runner_with_spinner_message, ) from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.ui import open_spinner from pip._internal.utils.unpacking import unpack_file from pip._internal.utils.urls import path_to_url if MYPY_CHECK_RUNNING: from typing import ( Dict, List, Optional, Sequence, Mapping, Tuple, IO, Text, Any, Iterable, Callable, Set, ) from pip._vendor.packaging.requirements import Requirement from pip._internal.req.req_install import InstallRequirement from pip._internal.operations.prepare import ( RequirementPreparer ) from pip._internal.cache import WheelCache from pip._internal.pep425tags import Pep425Tag InstalledCSVRow = Tuple[str, ...] BinaryAllowedPredicate = Callable[[InstallRequirement], bool] VERSION_COMPATIBLE = (1, 0) logger = logging.getLogger(__name__) def normpath(src, p): return os.path.relpath(src, p).replace(os.path.sep, '/') def hash_file(path, blocksize=1 << 20): # type: (str, int) -> Tuple[Any, int] """Return (hash, length) for path using hashlib.sha256()""" h = hashlib.sha256() length = 0 with open(path, 'rb') as f: for block in read_chunks(f, size=blocksize): length += len(block) h.update(block) return (h, length) # type: ignore def rehash(path, blocksize=1 << 20): # type: (str, int) -> Tuple[str, str] """Return (encoded_digest, length) for path using hashlib.sha256()""" h, length = hash_file(path, blocksize) digest = 'sha256=' + urlsafe_b64encode( h.digest() ).decode('latin1').rstrip('=') # unicode/str python2 issues return (digest, str(length)) # type: ignore def open_for_csv(name, mode): # type: (str, Text) -> IO if sys.version_info[0] < 3: nl = {} # type: Dict[str, Any] bin = 'b' else: nl = {'newline': ''} # type: Dict[str, Any] bin = '' return open(name, mode + bin, **nl) def replace_python_tag(wheelname, new_tag): # type: (str, str) -> str """Replace the Python tag in a wheel file name with a new value. """ parts = wheelname.split('-') parts[-3] = new_tag return '-'.join(parts) def fix_script(path): # type: (str) -> Optional[bool] """Replace #!python with #!/path/to/python Return True if file was changed.""" # XXX RECORD hashes will need to be updated if os.path.isfile(path): with open(path, 'rb') as script: firstline = script.readline() if not firstline.startswith(b'#!python'): return False exename = sys.executable.encode(sys.getfilesystemencoding()) firstline = b'#!' + exename + os.linesep.encode("ascii") rest = script.read() with open(path, 'wb') as script: script.write(firstline) script.write(rest) return True return None dist_info_re = re.compile(r"""^(?P(?P.+?)(-(?P.+?))?) \.dist-info$""", re.VERBOSE) def root_is_purelib(name, wheeldir): # type: (str, str) -> bool """ Return True if the extracted wheel in wheeldir should go into purelib. """ name_folded = name.replace("-", "_") for item in os.listdir(wheeldir): match = dist_info_re.match(item) if match and match.group('name') == name_folded: with open(os.path.join(wheeldir, item, 'WHEEL')) as wheel: for line in wheel: line = line.lower().rstrip() if line == "root-is-purelib: true": return True return False def get_entrypoints(filename): # type: (str) -> Tuple[Dict[str, str], Dict[str, str]] if not os.path.exists(filename): return {}, {} # This is done because you can pass a string to entry_points wrappers which # means that they may or may not be valid INI files. The attempt here is to # strip leading and trailing whitespace in order to make them valid INI # files. with open(filename) as fp: data = StringIO() for line in fp: data.write(line.strip()) data.write("\n") data.seek(0) # get the entry points and then the script names entry_points = pkg_resources.EntryPoint.parse_map(data) console = entry_points.get('console_scripts', {}) gui = entry_points.get('gui_scripts', {}) def _split_ep(s): """get the string representation of EntryPoint, remove space and split on '='""" return str(s).replace(" ", "").split("=") # convert the EntryPoint objects into strings with module:function console = dict(_split_ep(v) for v in console.values()) gui = dict(_split_ep(v) for v in gui.values()) return console, gui def message_about_scripts_not_on_PATH(scripts): # type: (Sequence[str]) -> Optional[str] """Determine if any scripts are not on PATH and format a warning. Returns a warning message if one or more scripts are not on PATH, otherwise None. """ if not scripts: return None # Group scripts by the path they were installed in grouped_by_dir = collections.defaultdict(set) # type: Dict[str, Set[str]] for destfile in scripts: parent_dir = os.path.dirname(destfile) script_name = os.path.basename(destfile) grouped_by_dir[parent_dir].add(script_name) # We don't want to warn for directories that are on PATH. not_warn_dirs = [ os.path.normcase(i).rstrip(os.sep) for i in os.environ.get("PATH", "").split(os.pathsep) ] # If an executable sits with sys.executable, we don't warn for it. # This covers the case of venv invocations without activating the venv. not_warn_dirs.append(os.path.normcase(os.path.dirname(sys.executable))) warn_for = { parent_dir: scripts for parent_dir, scripts in grouped_by_dir.items() if os.path.normcase(parent_dir) not in not_warn_dirs } # type: Dict[str, Set[str]] if not warn_for: return None # Format a message msg_lines = [] for parent_dir, dir_scripts in warn_for.items(): sorted_scripts = sorted(dir_scripts) # type: List[str] if len(sorted_scripts) == 1: start_text = "script {} is".format(sorted_scripts[0]) else: start_text = "scripts {} are".format( ", ".join(sorted_scripts[:-1]) + " and " + sorted_scripts[-1] ) msg_lines.append( "The {} installed in '{}' which is not on PATH." .format(start_text, parent_dir) ) last_line_fmt = ( "Consider adding {} to PATH or, if you prefer " "to suppress this warning, use --no-warn-script-location." ) if len(msg_lines) == 1: msg_lines.append(last_line_fmt.format("this directory")) else: msg_lines.append(last_line_fmt.format("these directories")) # Returns the formatted multiline message return "\n".join(msg_lines) def sorted_outrows(outrows): # type: (Iterable[InstalledCSVRow]) -> List[InstalledCSVRow] """ Return the given rows of a RECORD file in sorted order. Each row is a 3-tuple (path, hash, size) and corresponds to a record of a RECORD file (see PEP 376 and PEP 427 for details). For the rows passed to this function, the size can be an integer as an int or string, or the empty string. """ # Normally, there should only be one row per path, in which case the # second and third elements don't come into play when sorting. # However, in cases in the wild where a path might happen to occur twice, # we don't want the sort operation to trigger an error (but still want # determinism). Since the third element can be an int or string, we # coerce each element to a string to avoid a TypeError in this case. # For additional background, see-- # https://github.com/pypa/pip/issues/5868 return sorted(outrows, key=lambda row: tuple(str(x) for x in row)) def get_csv_rows_for_installed( old_csv_rows, # type: Iterable[List[str]] installed, # type: Dict[str, str] changed, # type: set generated, # type: List[str] lib_dir, # type: str ): # type: (...) -> List[InstalledCSVRow] """ :param installed: A map from archive RECORD path to installation RECORD path. """ installed_rows = [] # type: List[InstalledCSVRow] for row in old_csv_rows: if len(row) > 3: logger.warning( 'RECORD line has more than three elements: {}'.format(row) ) # Make a copy because we are mutating the row. row = list(row) old_path = row[0] new_path = installed.pop(old_path, old_path) row[0] = new_path if new_path in changed: digest, length = rehash(new_path) row[1] = digest row[2] = length installed_rows.append(tuple(row)) for f in generated: digest, length = rehash(f) installed_rows.append((normpath(f, lib_dir), digest, str(length))) for f in installed: installed_rows.append((installed[f], '', '')) return installed_rows class MissingCallableSuffix(Exception): pass def _raise_for_invalid_entrypoint(specification): entry = get_export_entry(specification) if entry is not None and entry.suffix is None: raise MissingCallableSuffix(str(entry)) class PipScriptMaker(ScriptMaker): def make(self, specification, options=None): _raise_for_invalid_entrypoint(specification) return super(PipScriptMaker, self).make(specification, options) def move_wheel_files( name, # type: str req, # type: Requirement wheeldir, # type: str user=False, # type: bool home=None, # type: Optional[str] root=None, # type: Optional[str] pycompile=True, # type: bool scheme=None, # type: Optional[Mapping[str, str]] isolated=False, # type: bool prefix=None, # type: Optional[str] warn_script_location=True # type: bool ): # type: (...) -> None """Install a wheel""" # TODO: Investigate and break this up. # TODO: Look into moving this into a dedicated class for representing an # installation. if not scheme: scheme = distutils_scheme( name, user=user, home=home, root=root, isolated=isolated, prefix=prefix, ) if root_is_purelib(name, wheeldir): lib_dir = scheme['purelib'] else: lib_dir = scheme['platlib'] info_dir = [] # type: List[str] data_dirs = [] source = wheeldir.rstrip(os.path.sep) + os.path.sep # Record details of the files moved # installed = files copied from the wheel to the destination # changed = files changed while installing (scripts #! line typically) # generated = files newly generated during the install (script wrappers) installed = {} # type: Dict[str, str] changed = set() generated = [] # type: List[str] # Compile all of the pyc files that we're going to be installing if pycompile: with captured_stdout() as stdout: with warnings.catch_warnings(): warnings.filterwarnings('ignore') compileall.compile_dir(source, force=True, quiet=True) logger.debug(stdout.getvalue()) def record_installed(srcfile, destfile, modified=False): """Map archive RECORD paths to installation RECORD paths.""" oldpath = normpath(srcfile, wheeldir) newpath = normpath(destfile, lib_dir) installed[oldpath] = newpath if modified: changed.add(destfile) def clobber(source, dest, is_base, fixer=None, filter=None): ensure_dir(dest) # common for the 'include' path for dir, subdirs, files in os.walk(source): basedir = dir[len(source):].lstrip(os.path.sep) destdir = os.path.join(dest, basedir) if is_base and basedir.split(os.path.sep, 1)[0].endswith('.data'): continue for s in subdirs: destsubdir = os.path.join(dest, basedir, s) if is_base and basedir == '' and destsubdir.endswith('.data'): data_dirs.append(s) continue elif (is_base and s.endswith('.dist-info') and canonicalize_name(s).startswith( canonicalize_name(req.name))): assert not info_dir, ('Multiple .dist-info directories: ' + destsubdir + ', ' + ', '.join(info_dir)) info_dir.append(destsubdir) for f in files: # Skip unwanted files if filter and filter(f): continue srcfile = os.path.join(dir, f) destfile = os.path.join(dest, basedir, f) # directory creation is lazy and after the file filtering above # to ensure we don't install empty dirs; empty dirs can't be # uninstalled. ensure_dir(destdir) # copyfile (called below) truncates the destination if it # exists and then writes the new contents. This is fine in most # cases, but can cause a segfault if pip has loaded a shared # object (e.g. from pyopenssl through its vendored urllib3) # Since the shared object is mmap'd an attempt to call a # symbol in it will then cause a segfault. Unlinking the file # allows writing of new contents while allowing the process to # continue to use the old copy. if os.path.exists(destfile): os.unlink(destfile) # We use copyfile (not move, copy, or copy2) to be extra sure # that we are not moving directories over (copyfile fails for # directories) as well as to ensure that we are not copying # over any metadata because we want more control over what # metadata we actually copy over. shutil.copyfile(srcfile, destfile) # Copy over the metadata for the file, currently this only # includes the atime and mtime. st = os.stat(srcfile) if hasattr(os, "utime"): os.utime(destfile, (st.st_atime, st.st_mtime)) # If our file is executable, then make our destination file # executable. if os.access(srcfile, os.X_OK): st = os.stat(srcfile) permissions = ( st.st_mode | stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH ) os.chmod(destfile, permissions) changed = False if fixer: changed = fixer(destfile) record_installed(srcfile, destfile, changed) clobber(source, lib_dir, True) assert info_dir, "%s .dist-info directory not found" % req # Get the defined entry points ep_file = os.path.join(info_dir[0], 'entry_points.txt') console, gui = get_entrypoints(ep_file) def is_entrypoint_wrapper(name): # EP, EP.exe and EP-script.py are scripts generated for # entry point EP by setuptools if name.lower().endswith('.exe'): matchname = name[:-4] elif name.lower().endswith('-script.py'): matchname = name[:-10] elif name.lower().endswith(".pya"): matchname = name[:-4] else: matchname = name # Ignore setuptools-generated scripts return (matchname in console or matchname in gui) for datadir in data_dirs: fixer = None filter = None for subdir in os.listdir(os.path.join(wheeldir, datadir)): fixer = None if subdir == 'scripts': fixer = fix_script filter = is_entrypoint_wrapper source = os.path.join(wheeldir, datadir, subdir) dest = scheme[subdir] clobber(source, dest, False, fixer=fixer, filter=filter) maker = PipScriptMaker(None, scheme['scripts']) # Ensure old scripts are overwritten. # See https://github.com/pypa/pip/issues/1800 maker.clobber = True # Ensure we don't generate any variants for scripts because this is almost # never what somebody wants. # See https://bitbucket.org/pypa/distlib/issue/35/ maker.variants = {''} # This is required because otherwise distlib creates scripts that are not # executable. # See https://bitbucket.org/pypa/distlib/issue/32/ maker.set_mode = True scripts_to_generate = [] # Special case pip and setuptools to generate versioned wrappers # # The issue is that some projects (specifically, pip and setuptools) use # code in setup.py to create "versioned" entry points - pip2.7 on Python # 2.7, pip3.3 on Python 3.3, etc. But these entry points are baked into # the wheel metadata at build time, and so if the wheel is installed with # a *different* version of Python the entry points will be wrong. The # correct fix for this is to enhance the metadata to be able to describe # such versioned entry points, but that won't happen till Metadata 2.0 is # available. # In the meantime, projects using versioned entry points will either have # incorrect versioned entry points, or they will not be able to distribute # "universal" wheels (i.e., they will need a wheel per Python version). # # Because setuptools and pip are bundled with _ensurepip and virtualenv, # we need to use universal wheels. So, as a stopgap until Metadata 2.0, we # override the versioned entry points in the wheel and generate the # correct ones. This code is purely a short-term measure until Metadata 2.0 # is available. # # To add the level of hack in this section of code, in order to support # ensurepip this code will look for an ``ENSUREPIP_OPTIONS`` environment # variable which will control which version scripts get installed. # # ENSUREPIP_OPTIONS=altinstall # - Only pipX.Y and easy_install-X.Y will be generated and installed # ENSUREPIP_OPTIONS=install # - pipX.Y, pipX, easy_install-X.Y will be generated and installed. Note # that this option is technically if ENSUREPIP_OPTIONS is set and is # not altinstall # DEFAULT # - The default behavior is to install pip, pipX, pipX.Y, easy_install # and easy_install-X.Y. pip_script = console.pop('pip', None) if pip_script: if "ENSUREPIP_OPTIONS" not in os.environ: scripts_to_generate.append('pip = ' + pip_script) if os.environ.get("ENSUREPIP_OPTIONS", "") != "altinstall": scripts_to_generate.append( 'pip%s = %s' % (sys.version_info[0], pip_script) ) scripts_to_generate.append( 'pip%s = %s' % (get_major_minor_version(), pip_script) ) # Delete any other versioned pip entry points pip_ep = [k for k in console if re.match(r'pip(\d(\.\d)?)?$', k)] for k in pip_ep: del console[k] easy_install_script = console.pop('easy_install', None) if easy_install_script: if "ENSUREPIP_OPTIONS" not in os.environ: scripts_to_generate.append( 'easy_install = ' + easy_install_script ) scripts_to_generate.append( 'easy_install-%s = %s' % ( get_major_minor_version(), easy_install_script ) ) # Delete any other versioned easy_install entry points easy_install_ep = [ k for k in console if re.match(r'easy_install(-\d\.\d)?$', k) ] for k in easy_install_ep: del console[k] # Generate the console and GUI entry points specified in the wheel scripts_to_generate.extend( '%s = %s' % kv for kv in console.items() ) gui_scripts_to_generate = [ '%s = %s' % kv for kv in gui.items() ] generated_console_scripts = [] # type: List[str] try: generated_console_scripts = maker.make_multiple(scripts_to_generate) generated.extend(generated_console_scripts) generated.extend( maker.make_multiple(gui_scripts_to_generate, {'gui': True}) ) except MissingCallableSuffix as e: entry = e.args[0] raise InstallationError( "Invalid script entry point: {} for req: {} - A callable " "suffix is required. Cf https://packaging.python.org/en/" "latest/distributing.html#console-scripts for more " "information.".format(entry, req) ) if warn_script_location: msg = message_about_scripts_not_on_PATH(generated_console_scripts) if msg is not None: logger.warning(msg) # Record pip as the installer installer = os.path.join(info_dir[0], 'INSTALLER') temp_installer = os.path.join(info_dir[0], 'INSTALLER.pip') with open(temp_installer, 'wb') as installer_file: installer_file.write(b'pip\n') shutil.move(temp_installer, installer) generated.append(installer) # Record details of all files installed record = os.path.join(info_dir[0], 'RECORD') temp_record = os.path.join(info_dir[0], 'RECORD.pip') with open_for_csv(record, 'r') as record_in: with open_for_csv(temp_record, 'w+') as record_out: reader = csv.reader(record_in) outrows = get_csv_rows_for_installed( reader, installed=installed, changed=changed, generated=generated, lib_dir=lib_dir, ) writer = csv.writer(record_out) # Sort to simplify testing. for row in sorted_outrows(outrows): writer.writerow(row) shutil.move(temp_record, record) def wheel_version(source_dir): # type: (Optional[str]) -> Optional[Tuple[int, ...]] """ Return the Wheel-Version of an extracted wheel, if possible. Otherwise, return None if we couldn't parse / extract it. """ try: dist = [d for d in pkg_resources.find_on_path(None, source_dir)][0] wheel_data = dist.get_metadata('WHEEL') wheel_data = Parser().parsestr(wheel_data) version = wheel_data['Wheel-Version'].strip() version = tuple(map(int, version.split('.'))) return version except Exception: return None def check_compatibility(version, name): # type: (Optional[Tuple[int, ...]], str) -> None """ Raises errors or warns if called with an incompatible Wheel-Version. Pip should refuse to install a Wheel-Version that's a major series ahead of what it's compatible with (e.g 2.0 > 1.1); and warn when installing a version only minor version ahead (e.g 1.2 > 1.1). version: a 2-tuple representing a Wheel-Version (Major, Minor) name: name of wheel or package to raise exception about :raises UnsupportedWheel: when an incompatible Wheel-Version is given """ if not version: raise UnsupportedWheel( "%s is in an unsupported or invalid wheel" % name ) if version[0] > VERSION_COMPATIBLE[0]: raise UnsupportedWheel( "%s's Wheel-Version (%s) is not compatible with this version " "of pip" % (name, '.'.join(map(str, version))) ) elif version > VERSION_COMPATIBLE: logger.warning( 'Installing from a newer Wheel-Version (%s)', '.'.join(map(str, version)), ) def format_tag(file_tag): # type: (Tuple[str, ...]) -> str """ Format three tags in the form "--". :param file_tag: A 3-tuple of tags (python_tag, abi_tag, platform_tag). """ return '-'.join(file_tag) class Wheel(object): """A wheel file""" # TODO: Maybe move the class into the models sub-package # TODO: Maybe move the install code into this class wheel_file_re = re.compile( r"""^(?P(?P.+?)-(?P.*?)) ((-(?P\d[^-]*?))?-(?P.+?)-(?P.+?)-(?P.+?) \.whl|\.dist-info)$""", re.VERBOSE ) def __init__(self, filename): # type: (str) -> None """ :raises InvalidWheelFilename: when the filename is invalid for a wheel """ wheel_info = self.wheel_file_re.match(filename) if not wheel_info: raise InvalidWheelFilename( "%s is not a valid wheel filename." % filename ) self.filename = filename self.name = wheel_info.group('name').replace('_', '-') # we'll assume "_" means "-" due to wheel naming scheme # (https://github.com/pypa/pip/issues/1150) self.version = wheel_info.group('ver').replace('_', '-') self.build_tag = wheel_info.group('build') self.pyversions = wheel_info.group('pyver').split('.') self.abis = wheel_info.group('abi').split('.') self.plats = wheel_info.group('plat').split('.') # All the tag combinations from this file self.file_tags = { (x, y, z) for x in self.pyversions for y in self.abis for z in self.plats } def get_formatted_file_tags(self): # type: () -> List[str] """ Return the wheel's tags as a sorted list of strings. """ return sorted(format_tag(tag) for tag in self.file_tags) def support_index_min(self, tags): # type: (List[Pep425Tag]) -> int """ Return the lowest index that one of the wheel's file_tag combinations achieves in the given list of supported tags. For example, if there are 8 supported tags and one of the file tags is first in the list, then return 0. :param tags: the PEP 425 tags to check the wheel against, in order with most preferred first. :raises ValueError: If none of the wheel's file tags match one of the supported tags. """ return min(tags.index(tag) for tag in self.file_tags if tag in tags) def supported(self, tags): # type: (List[Pep425Tag]) -> bool """ Return whether the wheel is compatible with one of the given tags. :param tags: the PEP 425 tags to check the wheel against. """ return not self.file_tags.isdisjoint(tags) def _contains_egg_info( s, _egg_info_re=re.compile(r'([a-z0-9_.]+)-([a-z0-9_.!+-]+)', re.I)): """Determine whether the string looks like an egg_info. :param s: The string to parse. E.g. foo-2.1 """ return bool(_egg_info_re.search(s)) def should_use_ephemeral_cache( req, # type: InstallRequirement should_unpack, # type: bool cache_available, # type: bool check_binary_allowed, # type: BinaryAllowedPredicate ): # type: (...) -> Optional[bool] """ Return whether to build an InstallRequirement object using the ephemeral cache. :param cache_available: whether a cache directory is available for the should_unpack=True case. :return: True or False to build the requirement with ephem_cache=True or False, respectively; or None not to build the requirement. """ if req.constraint: # never build requirements that are merely constraints return None if req.is_wheel: if not should_unpack: logger.info( 'Skipping %s, due to already being wheel.', req.name, ) return None if not should_unpack: # i.e. pip wheel, not pip install; # return False, knowing that the caller will never cache # in this case anyway, so this return merely means "build it". # TODO improve this behavior return False if req.editable or not req.source_dir: return None if not check_binary_allowed(req): logger.info( "Skipping wheel build for %s, due to binaries " "being disabled for it.", req.name, ) return None if req.link and req.link.is_vcs: # VCS checkout. Build wheel just for this run. return True link = req.link base, ext = link.splitext() if cache_available and _contains_egg_info(base): return False # Otherwise, build the wheel just for this run using the ephemeral # cache since we are either in the case of e.g. a local directory, or # no cache directory is available to use. return True def format_command_result( command_args, # type: List[str] command_output, # type: str ): # type: (...) -> str """ Format command information for logging. """ command_desc = format_command_args(command_args) text = 'Command arguments: {}\n'.format(command_desc) if not command_output: text += 'Command output: None' elif logger.getEffectiveLevel() > logging.DEBUG: text += 'Command output: [use --verbose to show]' else: if not command_output.endswith('\n'): command_output += '\n' text += 'Command output:\n{}{}'.format(command_output, LOG_DIVIDER) return text def get_legacy_build_wheel_path( names, # type: List[str] temp_dir, # type: str req, # type: InstallRequirement command_args, # type: List[str] command_output, # type: str ): # type: (...) -> Optional[str] """ Return the path to the wheel in the temporary build directory. """ # Sort for determinism. names = sorted(names) if not names: msg = ( 'Legacy build of wheel for {!r} created no files.\n' ).format(req.name) msg += format_command_result(command_args, command_output) logger.warning(msg) return None if len(names) > 1: msg = ( 'Legacy build of wheel for {!r} created more than one file.\n' 'Filenames (choosing first): {}\n' ).format(req.name, names) msg += format_command_result(command_args, command_output) logger.warning(msg) return os.path.join(temp_dir, names[0]) def _always_true(_): return True class WheelBuilder(object): """Build wheels from a RequirementSet.""" def __init__( self, preparer, # type: RequirementPreparer wheel_cache, # type: WheelCache build_options=None, # type: Optional[List[str]] global_options=None, # type: Optional[List[str]] check_binary_allowed=None, # type: Optional[BinaryAllowedPredicate] no_clean=False # type: bool ): # type: (...) -> None if check_binary_allowed is None: # Binaries allowed by default. check_binary_allowed = _always_true self.preparer = preparer self.wheel_cache = wheel_cache self._wheel_dir = preparer.wheel_download_dir self.build_options = build_options or [] self.global_options = global_options or [] self.check_binary_allowed = check_binary_allowed self.no_clean = no_clean def _build_one(self, req, output_dir, python_tag=None): """Build one wheel. :return: The filename of the built wheel, or None if the build failed. """ # Install build deps into temporary directory (PEP 518) with req.build_env: return self._build_one_inside_env(req, output_dir, python_tag=python_tag) def _build_one_inside_env(self, req, output_dir, python_tag=None): with TempDirectory(kind="wheel") as temp_dir: if req.use_pep517: builder = self._build_one_pep517 else: builder = self._build_one_legacy wheel_path = builder(req, temp_dir.path, python_tag=python_tag) if wheel_path is not None: wheel_name = os.path.basename(wheel_path) dest_path = os.path.join(output_dir, wheel_name) try: wheel_hash, length = hash_file(wheel_path) shutil.move(wheel_path, dest_path) logger.info('Created wheel for %s: ' 'filename=%s size=%d sha256=%s', req.name, wheel_name, length, wheel_hash.hexdigest()) logger.info('Stored in directory: %s', output_dir) return dest_path except Exception: pass # Ignore return, we can't do anything else useful. self._clean_one(req) return None def _base_setup_args(self, req): # NOTE: Eventually, we'd want to also -S to the flags here, when we're # isolating. Currently, it breaks Python in virtualenvs, because it # relies on site.py to find parts of the standard library outside the # virtualenv. return make_setuptools_shim_args( req.setup_py_path, global_options=self.global_options, unbuffered_output=True ) def _build_one_pep517(self, req, tempd, python_tag=None): """Build one InstallRequirement using the PEP 517 build process. Returns path to wheel if successfully built. Otherwise, returns None. """ assert req.metadata_directory is not None if self.build_options: # PEP 517 does not support --build-options logger.error('Cannot build wheel for %s using PEP 517 when ' '--build-options is present' % (req.name,)) return None try: logger.debug('Destination directory: %s', tempd) runner = runner_with_spinner_message( 'Building wheel for {} (PEP 517)'.format(req.name) ) backend = req.pep517_backend with backend.subprocess_runner(runner): wheel_name = backend.build_wheel( tempd, metadata_directory=req.metadata_directory, ) if python_tag: # General PEP 517 backends don't necessarily support # a "--python-tag" option, so we rename the wheel # file directly. new_name = replace_python_tag(wheel_name, python_tag) os.rename( os.path.join(tempd, wheel_name), os.path.join(tempd, new_name) ) # Reassign to simplify the return at the end of function wheel_name = new_name except Exception: logger.error('Failed building wheel for %s', req.name) return None return os.path.join(tempd, wheel_name) def _build_one_legacy(self, req, tempd, python_tag=None): """Build one InstallRequirement using the "legacy" build process. Returns path to wheel if successfully built. Otherwise, returns None. """ base_args = self._base_setup_args(req) spin_message = 'Building wheel for %s (setup.py)' % (req.name,) with open_spinner(spin_message) as spinner: logger.debug('Destination directory: %s', tempd) wheel_args = base_args + ['bdist_wheel', '-d', tempd] \ + self.build_options if python_tag is not None: wheel_args += ["--python-tag", python_tag] try: output = call_subprocess( wheel_args, cwd=req.unpacked_source_directory, spinner=spinner, ) except Exception: spinner.finish("error") logger.error('Failed building wheel for %s', req.name) return None names = os.listdir(tempd) wheel_path = get_legacy_build_wheel_path( names=names, temp_dir=tempd, req=req, command_args=wheel_args, command_output=output, ) return wheel_path def _clean_one(self, req): base_args = self._base_setup_args(req) logger.info('Running setup.py clean for %s', req.name) clean_args = base_args + ['clean', '--all'] try: call_subprocess(clean_args, cwd=req.source_dir) return True except Exception: logger.error('Failed cleaning build dir for %s', req.name) return False def build( self, requirements, # type: Iterable[InstallRequirement] should_unpack=False # type: bool ): # type: (...) -> List[InstallRequirement] """Build wheels. :param should_unpack: If True, after building the wheel, unpack it and replace the sdist with the unpacked version in preparation for installation. :return: True if all the wheels built correctly. """ # pip install uses should_unpack=True. # pip install never provides a _wheel_dir. # pip wheel uses should_unpack=False. # pip wheel always provides a _wheel_dir (via the preparer). assert ( (should_unpack and not self._wheel_dir) or (not should_unpack and self._wheel_dir) ) buildset = [] cache_available = bool(self.wheel_cache.cache_dir) for req in requirements: ephem_cache = should_use_ephemeral_cache( req, should_unpack=should_unpack, cache_available=cache_available, check_binary_allowed=self.check_binary_allowed, ) if ephem_cache is None: continue # Determine where the wheel should go. if should_unpack: if ephem_cache: output_dir = self.wheel_cache.get_ephem_path_for_link( req.link ) else: output_dir = self.wheel_cache.get_path_for_link(req.link) else: output_dir = self._wheel_dir buildset.append((req, output_dir)) if not buildset: return [] # TODO by @pradyunsg # Should break up this method into 2 separate methods. # Build the wheels. logger.info( 'Building wheels for collected packages: %s', ', '.join([req.name for (req, _) in buildset]), ) python_tag = None if should_unpack: python_tag = pep425tags.implementation_tag with indent_log(): build_success, build_failure = [], [] for req, output_dir in buildset: try: ensure_dir(output_dir) except OSError as e: logger.warning( "Building wheel for %s failed: %s", req.name, e, ) build_failure.append(req) continue wheel_file = self._build_one( req, output_dir, python_tag=python_tag, ) if wheel_file: build_success.append(req) if should_unpack: # XXX: This is mildly duplicative with prepare_files, # but not close enough to pull out to a single common # method. # The code below assumes temporary source dirs - # prevent it doing bad things. if ( req.source_dir and not has_delete_marker_file(req.source_dir) ): raise AssertionError( "bad source dir - missing marker") # Delete the source we built the wheel from req.remove_temporary_source() # set the build directory again - name is known from # the work prepare_files did. req.source_dir = req.ensure_build_location( self.preparer.build_dir ) # Update the link for this. req.link = Link(path_to_url(wheel_file)) assert req.link.is_wheel # extract the wheel into the dir unpack_file(req.link.file_path, req.source_dir) else: build_failure.append(req) # notify success/failure if build_success: logger.info( 'Successfully built %s', ' '.join([req.name for req in build_success]), ) if build_failure: logger.info( 'Failed to build %s', ' '.join([req.name for req in build_failure]), ) # Return a list of requirements that failed to build return build_failure site-packages/pip/_internal/self_outdated_check.py000064400000017376151464452710016407 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import datetime import hashlib import json import logging import os.path import sys from pip._vendor import pkg_resources from pip._vendor.packaging import version as packaging_version from pip._vendor.six import ensure_binary from pip._internal.collector import LinkCollector from pip._internal.index import PackageFinder from pip._internal.models.search_scope import SearchScope from pip._internal.models.selection_prefs import SelectionPreferences from pip._internal.utils.compat import WINDOWS from pip._internal.utils.filesystem import ( adjacent_tmp_file, check_path_owner, replace, ) from pip._internal.utils.misc import ( ensure_dir, get_installed_version, redact_auth_from_url, ) from pip._internal.utils.packaging import get_installer from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: import optparse from optparse import Values from typing import Any, Dict, Text, Union from pip._internal.network.session import PipSession SELFCHECK_DATE_FMT = "%Y-%m-%dT%H:%M:%SZ" logger = logging.getLogger(__name__) def make_link_collector( session, # type: PipSession options, # type: Values suppress_no_index=False, # type: bool ): # type: (...) -> LinkCollector """ :param session: The Session to use to make requests. :param suppress_no_index: Whether to ignore the --no-index option when constructing the SearchScope object. """ index_urls = [options.index_url] + options.extra_index_urls if options.no_index and not suppress_no_index: logger.debug( 'Ignoring indexes: %s', ','.join(redact_auth_from_url(url) for url in index_urls), ) index_urls = [] # Make sure find_links is a list before passing to create(). find_links = options.find_links or [] search_scope = SearchScope.create( find_links=find_links, index_urls=index_urls, ) link_collector = LinkCollector(session=session, search_scope=search_scope) return link_collector def _get_statefile_name(key): # type: (Union[str, Text]) -> str key_bytes = ensure_binary(key) name = hashlib.sha224(key_bytes).hexdigest() return name class SelfCheckState(object): def __init__(self, cache_dir): # type: (str) -> None self.state = {} # type: Dict[str, Any] self.statefile_path = None # Try to load the existing state if cache_dir: self.statefile_path = os.path.join( cache_dir, "selfcheck", _get_statefile_name(self.key) ) try: with open(self.statefile_path) as statefile: self.state = json.load(statefile) except (IOError, ValueError, KeyError): # Explicitly suppressing exceptions, since we don't want to # error out if the cache file is invalid. pass @property def key(self): return sys.prefix def save(self, pypi_version, current_time): # type: (str, datetime.datetime) -> None # If we do not have a path to cache in, don't bother saving. if not self.statefile_path: return # Check to make sure that we own the directory if not check_path_owner(os.path.dirname(self.statefile_path)): return # Now that we've ensured the directory is owned by this user, we'll go # ahead and make sure that all our directories are created. ensure_dir(os.path.dirname(self.statefile_path)) state = { # Include the key so it's easy to tell which pip wrote the # file. "key": self.key, "last_check": current_time.strftime(SELFCHECK_DATE_FMT), "pypi_version": pypi_version, } text = json.dumps(state, sort_keys=True, separators=(",", ":")) with adjacent_tmp_file(self.statefile_path) as f: f.write(ensure_binary(text)) try: # Since we have a prefix-specific state file, we can just # overwrite whatever is there, no need to check. replace(f.name, self.statefile_path) except OSError: # Best effort. pass def was_installed_by_pip(pkg): # type: (str) -> bool """Checks whether pkg was installed by pip This is used not to display the upgrade message when pip is in fact installed by system package manager, such as dnf on Fedora. """ try: dist = pkg_resources.get_distribution(pkg) return "pip" == get_installer(dist) except pkg_resources.DistributionNotFound: return False def pip_self_version_check(session, options): # type: (PipSession, optparse.Values) -> None """Check for an update for pip. Limit the frequency of checks to once per week. State is stored either in the active virtualenv or in the user's USER_CACHE_DIR keyed off the prefix of the pip script path. """ installed_version = get_installed_version("pip") if not installed_version: return pip_version = packaging_version.parse(installed_version) pypi_version = None try: state = SelfCheckState(cache_dir=options.cache_dir) current_time = datetime.datetime.utcnow() # Determine if we need to refresh the state if "last_check" in state.state and "pypi_version" in state.state: last_check = datetime.datetime.strptime( state.state["last_check"], SELFCHECK_DATE_FMT ) if (current_time - last_check).total_seconds() < 7 * 24 * 60 * 60: pypi_version = state.state["pypi_version"] # Refresh the version if we need to or just see if we need to warn if pypi_version is None: # Lets use PackageFinder to see what the latest pip version is link_collector = make_link_collector( session, options=options, suppress_no_index=True, ) # Pass allow_yanked=False so we don't suggest upgrading to a # yanked version. selection_prefs = SelectionPreferences( allow_yanked=False, allow_all_prereleases=False, # Explicitly set to False ) finder = PackageFinder.create( link_collector=link_collector, selection_prefs=selection_prefs, ) best_candidate = finder.find_best_candidate("pip").best_candidate if best_candidate is None: return pypi_version = str(best_candidate.version) # save that we've performed a check state.save(pypi_version, current_time) remote_version = packaging_version.parse(pypi_version) local_version_is_older = ( pip_version < remote_version and pip_version.base_version != remote_version.base_version and was_installed_by_pip('pip') ) # Determine if our pypi_version is older if not local_version_is_older: return # Advise "python -m pip" on Windows to avoid issues # with overwriting pip.exe. if WINDOWS: pip_cmd = "python -m pip" else: pip_cmd = "pip" logger.warning( "You are using pip version %s; however, version %s is " "available.\nYou should consider upgrading via the " "'%s install --upgrade pip' command.", pip_version, pypi_version, pip_cmd ) except Exception: logger.debug( "There was an error checking the latest version of pip", exc_info=True, ) site-packages/pip/_internal/download.py000064400000050505151464452710014226 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import cgi import logging import mimetypes import os import re import shutil import sys from pip._vendor import requests from pip._vendor.requests.models import CONTENT_CHUNK_SIZE, Response from pip._vendor.six import PY2 from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.exceptions import HashMismatch, InstallationError from pip._internal.models.index import PyPI from pip._internal.network.session import PipSession from pip._internal.utils.encoding import auto_decode from pip._internal.utils.filesystem import copy2_fixed from pip._internal.utils.misc import ( ask_path_exists, backup_dir, consume, display_path, format_size, hide_url, path_to_display, rmtree, splitext, ) from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.ui import DownloadProgressProvider from pip._internal.utils.unpacking import unpack_file from pip._internal.utils.urls import get_url_scheme from pip._internal.vcs import vcs if MYPY_CHECK_RUNNING: from typing import ( IO, Callable, List, Optional, Text, Tuple, ) from mypy_extensions import TypedDict from pip._internal.models.link import Link from pip._internal.utils.hashes import Hashes from pip._internal.vcs.versioncontrol import VersionControl if PY2: CopytreeKwargs = TypedDict( 'CopytreeKwargs', { 'ignore': Callable[[str, List[str]], List[str]], 'symlinks': bool, }, total=False, ) else: CopytreeKwargs = TypedDict( 'CopytreeKwargs', { 'copy_function': Callable[[str, str], None], 'ignore': Callable[[str, List[str]], List[str]], 'ignore_dangling_symlinks': bool, 'symlinks': bool, }, total=False, ) __all__ = ['get_file_content', 'unpack_vcs_link', 'unpack_file_url', 'unpack_http_url', 'unpack_url', 'parse_content_disposition', 'sanitize_content_filename'] logger = logging.getLogger(__name__) def get_file_content(url, comes_from=None, session=None): # type: (str, Optional[str], Optional[PipSession]) -> Tuple[str, Text] """Gets the content of a file; it may be a filename, file: URL, or http: URL. Returns (location, content). Content is unicode. :param url: File path or url. :param comes_from: Origin description of requirements. :param session: Instance of pip.download.PipSession. """ if session is None: raise TypeError( "get_file_content() missing 1 required keyword argument: 'session'" ) scheme = get_url_scheme(url) if scheme in ['http', 'https']: # FIXME: catch some errors resp = session.get(url) resp.raise_for_status() return resp.url, resp.text elif scheme == 'file': if comes_from and comes_from.startswith('http'): raise InstallationError( 'Requirements file %s references URL %s, which is local' % (comes_from, url)) path = url.split(':', 1)[1] path = path.replace('\\', '/') match = _url_slash_drive_re.match(path) if match: path = match.group(1) + ':' + path.split('|', 1)[1] path = urllib_parse.unquote(path) if path.startswith('/'): path = '/' + path.lstrip('/') url = path try: with open(url, 'rb') as f: content = auto_decode(f.read()) except IOError as exc: raise InstallationError( 'Could not open requirements file: %s' % str(exc) ) return url, content _url_slash_drive_re = re.compile(r'/*([a-z])\|', re.I) def unpack_vcs_link(link, location): # type: (Link, str) -> None vcs_backend = _get_used_vcs_backend(link) assert vcs_backend is not None vcs_backend.unpack(location, url=hide_url(link.url)) def _get_used_vcs_backend(link): # type: (Link) -> Optional[VersionControl] """ Return a VersionControl object or None. """ for vcs_backend in vcs.backends: if link.scheme in vcs_backend.schemes: return vcs_backend return None def _progress_indicator(iterable, *args, **kwargs): return iterable def _download_url( resp, # type: Response link, # type: Link content_file, # type: IO hashes, # type: Optional[Hashes] progress_bar # type: str ): # type: (...) -> None try: total_length = int(resp.headers['content-length']) except (ValueError, KeyError, TypeError): total_length = 0 cached_resp = getattr(resp, "from_cache", False) if logger.getEffectiveLevel() > logging.INFO: show_progress = False elif cached_resp: show_progress = False elif total_length > (40 * 1000): show_progress = True elif not total_length: show_progress = True else: show_progress = False show_url = link.show_url def resp_read(chunk_size): try: # Special case for urllib3. for chunk in resp.raw.stream( chunk_size, # We use decode_content=False here because we don't # want urllib3 to mess with the raw bytes we get # from the server. If we decompress inside of # urllib3 then we cannot verify the checksum # because the checksum will be of the compressed # file. This breakage will only occur if the # server adds a Content-Encoding header, which # depends on how the server was configured: # - Some servers will notice that the file isn't a # compressible file and will leave the file alone # and with an empty Content-Encoding # - Some servers will notice that the file is # already compressed and will leave the file # alone and will add a Content-Encoding: gzip # header # - Some servers won't notice anything at all and # will take a file that's already been compressed # and compress it again and set the # Content-Encoding: gzip header # # By setting this not to decode automatically we # hope to eliminate problems with the second case. decode_content=False): yield chunk except AttributeError: # Standard file-like object. while True: chunk = resp.raw.read(chunk_size) if not chunk: break yield chunk def written_chunks(chunks): for chunk in chunks: content_file.write(chunk) yield chunk progress_indicator = _progress_indicator if link.netloc == PyPI.netloc: url = show_url else: url = link.url_without_fragment if show_progress: # We don't show progress on cached responses progress_indicator = DownloadProgressProvider(progress_bar, max=total_length) if total_length: logger.info("Downloading %s (%s)", url, format_size(total_length)) else: logger.info("Downloading %s", url) elif cached_resp: logger.info("Using cached %s", url) else: logger.info("Downloading %s", url) downloaded_chunks = written_chunks( progress_indicator( resp_read(CONTENT_CHUNK_SIZE), CONTENT_CHUNK_SIZE ) ) if hashes: hashes.check_against_chunks(downloaded_chunks) else: consume(downloaded_chunks) def _copy_file(filename, location, link): copy = True download_location = os.path.join(location, link.filename) if os.path.exists(download_location): response = ask_path_exists( 'The file %s exists. (i)gnore, (w)ipe, (b)ackup, (a)abort' % display_path(download_location), ('i', 'w', 'b', 'a')) if response == 'i': copy = False elif response == 'w': logger.warning('Deleting %s', display_path(download_location)) os.remove(download_location) elif response == 'b': dest_file = backup_dir(download_location) logger.warning( 'Backing up %s to %s', display_path(download_location), display_path(dest_file), ) shutil.move(download_location, dest_file) elif response == 'a': sys.exit(-1) if copy: shutil.copy(filename, download_location) logger.info('Saved %s', display_path(download_location)) def unpack_http_url( link, # type: Link location, # type: str download_dir=None, # type: Optional[str] session=None, # type: Optional[PipSession] hashes=None, # type: Optional[Hashes] progress_bar="on" # type: str ): # type: (...) -> None if session is None: raise TypeError( "unpack_http_url() missing 1 required keyword argument: 'session'" ) with TempDirectory(kind="unpack") as temp_dir: # If a download dir is specified, is the file already downloaded there? already_downloaded_path = None if download_dir: already_downloaded_path = _check_download_dir(link, download_dir, hashes) if already_downloaded_path: from_path = already_downloaded_path content_type = mimetypes.guess_type(from_path)[0] else: # let's download to a tmp dir from_path, content_type = _download_http_url(link, session, temp_dir.path, hashes, progress_bar) # unpack the archive to the build dir location. even when only # downloading archives, they have to be unpacked to parse dependencies unpack_file(from_path, location, content_type) # a download dir is specified; let's copy the archive there if download_dir and not already_downloaded_path: _copy_file(from_path, download_dir, link) if not already_downloaded_path: os.unlink(from_path) def _copy2_ignoring_special_files(src, dest): # type: (str, str) -> None """Copying special files is not supported, but as a convenience to users we skip errors copying them. This supports tools that may create e.g. socket files in the project source directory. """ try: copy2_fixed(src, dest) except shutil.SpecialFileError as e: # SpecialFileError may be raised due to either the source or # destination. If the destination was the cause then we would actually # care, but since the destination directory is deleted prior to # copy we ignore all of them assuming it is caused by the source. logger.warning( "Ignoring special file error '%s' encountered copying %s to %s.", str(e), path_to_display(src), path_to_display(dest), ) def _copy_source_tree(source, target): # type: (str, str) -> None target_abspath = os.path.abspath(target) target_basename = os.path.basename(target_abspath) target_dirname = os.path.dirname(target_abspath) def ignore(d, names): skipped = [] # type: List[str] if d == source: # Pulling in those directories can potentially be very slow, # exclude the following directories if they appear in the top # level dir (and only it). # See discussion at https://github.com/pypa/pip/pull/6770 skipped += ['.tox', '.nox'] if os.path.abspath(d) == target_dirname: # Prevent an infinite recursion if the target is in source. # This can happen when TMPDIR is set to ${PWD}/... # and we copy PWD to TMPDIR. skipped += [target_basename] return skipped kwargs = dict(ignore=ignore, symlinks=True) # type: CopytreeKwargs if not PY2: # Python 2 does not support copy_function, so we only ignore # errors on special file copy in Python 3. kwargs['copy_function'] = _copy2_ignoring_special_files shutil.copytree(source, target, **kwargs) def unpack_file_url( link, # type: Link location, # type: str download_dir=None, # type: Optional[str] hashes=None # type: Optional[Hashes] ): # type: (...) -> None """Unpack link into location. If download_dir is provided and link points to a file, make a copy of the link file inside download_dir. """ link_path = link.file_path # If it's a url to a local directory if link.is_existing_dir(): if os.path.isdir(location): rmtree(location) _copy_source_tree(link_path, location) if download_dir: logger.info('Link is a directory, ignoring download_dir') return # If --require-hashes is off, `hashes` is either empty, the # link's embedded hash, or MissingHashes; it is required to # match. If --require-hashes is on, we are satisfied by any # hash in `hashes` matching: a URL-based or an option-based # one; no internet-sourced hash will be in `hashes`. if hashes: hashes.check_against_path(link_path) # If a download dir is specified, is the file already there and valid? already_downloaded_path = None if download_dir: already_downloaded_path = _check_download_dir(link, download_dir, hashes) if already_downloaded_path: from_path = already_downloaded_path else: from_path = link_path content_type = mimetypes.guess_type(from_path)[0] # unpack the archive to the build dir location. even when only downloading # archives, they have to be unpacked to parse dependencies unpack_file(from_path, location, content_type) # a download dir is specified and not already downloaded if download_dir and not already_downloaded_path: _copy_file(from_path, download_dir, link) def unpack_url( link, # type: Link location, # type: str download_dir=None, # type: Optional[str] session=None, # type: Optional[PipSession] hashes=None, # type: Optional[Hashes] progress_bar="on" # type: str ): # type: (...) -> None """Unpack link. If link is a VCS link: if only_download, export into download_dir and ignore location else unpack into location for other types of link: - unpack into location - if download_dir, copy the file into download_dir - if only_download, mark location for deletion :param hashes: A Hashes object, one of whose embedded hashes must match, or HashMismatch will be raised. If the Hashes is empty, no matches are required, and unhashable types of requirements (like VCS ones, which would ordinarily raise HashUnsupported) are allowed. """ # non-editable vcs urls if link.is_vcs: unpack_vcs_link(link, location) # file urls elif link.is_file: unpack_file_url(link, location, download_dir, hashes=hashes) # http urls else: if session is None: session = PipSession() unpack_http_url( link, location, download_dir, session, hashes=hashes, progress_bar=progress_bar ) def sanitize_content_filename(filename): # type: (str) -> str """ Sanitize the "filename" value from a Content-Disposition header. """ return os.path.basename(filename) def parse_content_disposition(content_disposition, default_filename): # type: (str, str) -> str """ Parse the "filename" value from a Content-Disposition header, and return the default filename if the result is empty. """ _type, params = cgi.parse_header(content_disposition) filename = params.get('filename') if filename: # We need to sanitize the filename to prevent directory traversal # in case the filename contains ".." path parts. filename = sanitize_content_filename(filename) return filename or default_filename def _download_http_url( link, # type: Link session, # type: PipSession temp_dir, # type: str hashes, # type: Optional[Hashes] progress_bar # type: str ): # type: (...) -> Tuple[str, str] """Download link url into temp_dir using provided session""" target_url = link.url.split('#', 1)[0] try: resp = session.get( target_url, # We use Accept-Encoding: identity here because requests # defaults to accepting compressed responses. This breaks in # a variety of ways depending on how the server is configured. # - Some servers will notice that the file isn't a compressible # file and will leave the file alone and with an empty # Content-Encoding # - Some servers will notice that the file is already # compressed and will leave the file alone and will add a # Content-Encoding: gzip header # - Some servers won't notice anything at all and will take # a file that's already been compressed and compress it again # and set the Content-Encoding: gzip header # By setting this to request only the identity encoding We're # hoping to eliminate the third case. Hopefully there does not # exist a server which when given a file will notice it is # already compressed and that you're not asking for a # compressed file and will then decompress it before sending # because if that's the case I don't think it'll ever be # possible to make this work. headers={"Accept-Encoding": "identity"}, stream=True, ) resp.raise_for_status() except requests.HTTPError as exc: logger.critical( "HTTP error %s while getting %s", exc.response.status_code, link, ) raise content_type = resp.headers.get('content-type', '') filename = link.filename # fallback # Have a look at the Content-Disposition header for a better guess content_disposition = resp.headers.get('content-disposition') if content_disposition: filename = parse_content_disposition(content_disposition, filename) ext = splitext(filename)[1] # type: Optional[str] if not ext: ext = mimetypes.guess_extension(content_type) if ext: filename += ext if not ext and link.url != resp.url: ext = os.path.splitext(resp.url)[1] if ext: filename += ext file_path = os.path.join(temp_dir, filename) with open(file_path, 'wb') as content_file: _download_url(resp, link, content_file, hashes, progress_bar) return file_path, content_type def _check_download_dir(link, download_dir, hashes): # type: (Link, str, Optional[Hashes]) -> Optional[str] """ Check download_dir for previously downloaded file with correct hash If a correct file is found return its path else None """ download_path = os.path.join(download_dir, link.filename) if not os.path.exists(download_path): return None # If already downloaded, does its hash match? logger.info('File was already downloaded %s', download_path) if hashes: try: hashes.check_against_path(download_path) except HashMismatch: logger.warning( 'Previously-downloaded file %s has bad hash. ' 'Re-downloading.', download_path ) os.unlink(download_path) return None return download_path site-packages/pip/_internal/req/constructors.py000064400000034064151464452710015760 0ustar00"""Backing implementation for InstallRequirement's various constructors The idea here is that these formed a major chunk of InstallRequirement's size so, moving them and support code dedicated to them outside of that class helps creates for better understandability for the rest of the code. These are meant to be used elsewhere within pip to create instances of InstallRequirement. """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False import logging import os import re from pip._vendor.packaging.markers import Marker from pip._vendor.packaging.requirements import InvalidRequirement, Requirement from pip._vendor.packaging.specifiers import Specifier from pip._vendor.pkg_resources import RequirementParseError, parse_requirements from pip._internal.exceptions import InstallationError from pip._internal.models.index import PyPI, TestPyPI from pip._internal.models.link import Link from pip._internal.pyproject import make_pyproject_path from pip._internal.req.req_install import InstallRequirement from pip._internal.utils.filetypes import ARCHIVE_EXTENSIONS from pip._internal.utils.misc import is_installable_dir, splitext from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import path_to_url from pip._internal.vcs import is_url, vcs from pip._internal.wheel import Wheel if MYPY_CHECK_RUNNING: from typing import ( Any, Dict, Optional, Set, Tuple, Union, ) from pip._internal.cache import WheelCache __all__ = [ "install_req_from_editable", "install_req_from_line", "parse_editable" ] logger = logging.getLogger(__name__) operators = Specifier._operators.keys() def is_archive_file(name): # type: (str) -> bool """Return True if `name` is a considered as an archive file.""" ext = splitext(name)[1].lower() if ext in ARCHIVE_EXTENSIONS: return True return False def _strip_extras(path): # type: (str) -> Tuple[str, Optional[str]] m = re.match(r'^(.+)(\[[^\]]+\])$', path) extras = None if m: path_no_extras = m.group(1) extras = m.group(2) else: path_no_extras = path return path_no_extras, extras def convert_extras(extras): # type: (Optional[str]) -> Set[str] if not extras: return set() return Requirement("placeholder" + extras.lower()).extras def parse_editable(editable_req): # type: (str) -> Tuple[Optional[str], str, Optional[Set[str]]] """Parses an editable requirement into: - a requirement name - an URL - extras - editable options Accepted requirements: svn+http://blahblah@rev#egg=Foobar[baz]&subdirectory=version_subdir .[some_extra] """ url = editable_req # If a file path is specified with extras, strip off the extras. url_no_extras, extras = _strip_extras(url) if os.path.isdir(url_no_extras): if not os.path.exists(os.path.join(url_no_extras, 'setup.py')): msg = ( 'File "setup.py" not found. Directory cannot be installed ' 'in editable mode: {}'.format(os.path.abspath(url_no_extras)) ) pyproject_path = make_pyproject_path(url_no_extras) if os.path.isfile(pyproject_path): msg += ( '\n(A "pyproject.toml" file was found, but editable ' 'mode currently requires a setup.py based build.)' ) raise InstallationError(msg) # Treating it as code that has already been checked out url_no_extras = path_to_url(url_no_extras) if url_no_extras.lower().startswith('file:'): package_name = Link(url_no_extras).egg_fragment if extras: return ( package_name, url_no_extras, Requirement("placeholder" + extras.lower()).extras, ) else: return package_name, url_no_extras, None for version_control in vcs: if url.lower().startswith('%s:' % version_control): url = '%s+%s' % (version_control, url) break if '+' not in url: raise InstallationError( '{} is not a valid editable requirement. ' 'It should either be a path to a local project or a VCS URL ' '(beginning with svn+, git+, hg+, or bzr+).'.format(editable_req) ) vc_type = url.split('+', 1)[0].lower() if not vcs.get_backend(vc_type): error_message = 'For --editable=%s only ' % editable_req + \ ', '.join([backend.name + '+URL' for backend in vcs.backends]) + \ ' is currently supported' raise InstallationError(error_message) package_name = Link(url).egg_fragment if not package_name: raise InstallationError( "Could not detect requirement name for '%s', please specify one " "with #egg=your_package_name" % editable_req ) return package_name, url, None def deduce_helpful_msg(req): # type: (str) -> str """Returns helpful msg in case requirements file does not exist, or cannot be parsed. :params req: Requirements file path """ msg = "" if os.path.exists(req): msg = " It does exist." # Try to parse and check if it is a requirements file. try: with open(req, 'r') as fp: # parse first line only next(parse_requirements(fp.read())) msg += " The argument you provided " + \ "(%s) appears to be a" % (req) + \ " requirements file. If that is the" + \ " case, use the '-r' flag to install" + \ " the packages specified within it." except RequirementParseError: logger.debug("Cannot parse '%s' as requirements \ file" % (req), exc_info=True) else: msg += " File '%s' does not exist." % (req) return msg class RequirementParts(object): def __init__( self, requirement, # type: Optional[Requirement] link, # type: Optional[Link] markers, # type: Optional[Marker] extras, # type: Set[str] ): self.requirement = requirement self.link = link self.markers = markers self.extras = extras def parse_req_from_editable(editable_req): # type: (str) -> RequirementParts name, url, extras_override = parse_editable(editable_req) if name is not None: try: req = Requirement(name) except InvalidRequirement: raise InstallationError("Invalid requirement: '%s'" % name) else: req = None link = Link(url) return RequirementParts(req, link, None, extras_override) # ---- The actual constructors follow ---- def install_req_from_editable( editable_req, # type: str comes_from=None, # type: Optional[str] use_pep517=None, # type: Optional[bool] isolated=False, # type: bool options=None, # type: Optional[Dict[str, Any]] wheel_cache=None, # type: Optional[WheelCache] constraint=False # type: bool ): # type: (...) -> InstallRequirement parts = parse_req_from_editable(editable_req) source_dir = parts.link.file_path if parts.link.scheme == 'file' else None return InstallRequirement( parts.requirement, comes_from, source_dir=source_dir, editable=True, link=parts.link, constraint=constraint, use_pep517=use_pep517, isolated=isolated, options=options if options else {}, wheel_cache=wheel_cache, extras=parts.extras, ) def _looks_like_path(name): # type: (str) -> bool """Checks whether the string "looks like" a path on the filesystem. This does not check whether the target actually exists, only judge from the appearance. Returns true if any of the following conditions is true: * a path separator is found (either os.path.sep or os.path.altsep); * a dot is found (which represents the current directory). """ if os.path.sep in name: return True if os.path.altsep is not None and os.path.altsep in name: return True if name.startswith("."): return True return False def _get_url_from_path(path, name): # type: (str, str) -> str """ First, it checks whether a provided path is an installable directory (e.g. it has a setup.py). If it is, returns the path. If false, check if the path is an archive file (such as a .whl). The function checks if the path is a file. If false, if the path has an @, it will treat it as a PEP 440 URL requirement and return the path. """ if _looks_like_path(name) and os.path.isdir(path): if is_installable_dir(path): return path_to_url(path) raise InstallationError( "Directory %r is not installable. Neither 'setup.py' " "nor 'pyproject.toml' found." % name ) if not is_archive_file(path): return None if os.path.isfile(path): return path_to_url(path) urlreq_parts = name.split('@', 1) if len(urlreq_parts) >= 2 and not _looks_like_path(urlreq_parts[0]): # If the path contains '@' and the part before it does not look # like a path, try to treat it as a PEP 440 URL req instead. return None logger.warning( 'Requirement %r looks like a filename, but the ' 'file does not exist', name ) return path_to_url(path) def parse_req_from_line(name, line_source): # type: (str, Optional[str]) -> RequirementParts if is_url(name): marker_sep = '; ' else: marker_sep = ';' if marker_sep in name: name, markers_as_string = name.split(marker_sep, 1) markers_as_string = markers_as_string.strip() if not markers_as_string: markers = None else: markers = Marker(markers_as_string) else: markers = None name = name.strip() req_as_string = None path = os.path.normpath(os.path.abspath(name)) link = None extras_as_string = None if is_url(name): link = Link(name) else: p, extras_as_string = _strip_extras(path) url = _get_url_from_path(p, name) if url is not None: link = Link(url) # it's a local file, dir, or url if link: # Handle relative file URLs if link.scheme == 'file' and re.search(r'\.\./', link.url): link = Link( path_to_url(os.path.normpath(os.path.abspath(link.path)))) # wheel file if link.is_wheel: wheel = Wheel(link.filename) # can raise InvalidWheelFilename req_as_string = "%s==%s" % (wheel.name, wheel.version) else: # set the req to the egg fragment. when it's not there, this # will become an 'unnamed' requirement req_as_string = link.egg_fragment # a requirement specifier else: req_as_string = name extras = convert_extras(extras_as_string) def with_source(text): if not line_source: return text return '{} (from {})'.format(text, line_source) if req_as_string is not None: try: req = Requirement(req_as_string) except InvalidRequirement: if os.path.sep in req_as_string: add_msg = "It looks like a path." add_msg += deduce_helpful_msg(req_as_string) elif ('=' in req_as_string and not any(op in req_as_string for op in operators)): add_msg = "= is not a valid operator. Did you mean == ?" else: add_msg = '' msg = with_source( 'Invalid requirement: {!r}'.format(req_as_string) ) if add_msg: msg += '\nHint: {}'.format(add_msg) raise InstallationError(msg) else: req = None return RequirementParts(req, link, markers, extras) def install_req_from_line( name, # type: str comes_from=None, # type: Optional[Union[str, InstallRequirement]] use_pep517=None, # type: Optional[bool] isolated=False, # type: bool options=None, # type: Optional[Dict[str, Any]] wheel_cache=None, # type: Optional[WheelCache] constraint=False, # type: bool line_source=None, # type: Optional[str] ): # type: (...) -> InstallRequirement """Creates an InstallRequirement from a name, which might be a requirement, directory containing 'setup.py', filename, or URL. :param line_source: An optional string describing where the line is from, for logging purposes in case of an error. """ parts = parse_req_from_line(name, line_source) return InstallRequirement( parts.requirement, comes_from, link=parts.link, markers=parts.markers, use_pep517=use_pep517, isolated=isolated, options=options if options else {}, wheel_cache=wheel_cache, constraint=constraint, extras=parts.extras, ) def install_req_from_req_string( req_string, # type: str comes_from=None, # type: Optional[InstallRequirement] isolated=False, # type: bool wheel_cache=None, # type: Optional[WheelCache] use_pep517=None # type: Optional[bool] ): # type: (...) -> InstallRequirement try: req = Requirement(req_string) except InvalidRequirement: raise InstallationError("Invalid requirement: '%s'" % req_string) domains_not_allowed = [ PyPI.file_storage_domain, TestPyPI.file_storage_domain, ] if (req.url and comes_from and comes_from.link and comes_from.link.netloc in domains_not_allowed): # Explicitly disallow pypi packages that depend on external urls raise InstallationError( "Packages installed from PyPI cannot depend on packages " "which are not also hosted on PyPI.\n" "%s depends on %s " % (comes_from.name, req) ) return InstallRequirement( req, comes_from, isolated=isolated, wheel_cache=wheel_cache, use_pep517=use_pep517 ) site-packages/pip/_internal/req/req_install.py000064400000107433151464452710015526 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import atexit import logging import os import shutil import sys import sysconfig import zipfile from distutils.util import change_root from pip._vendor import pkg_resources, six from pip._vendor.packaging.requirements import Requirement from pip._vendor.packaging.utils import canonicalize_name from pip._vendor.packaging.version import Version from pip._vendor.packaging.version import parse as parse_version from pip._vendor.pep517.wrappers import Pep517HookCaller from pip._internal import pep425tags, wheel from pip._internal.build_env import NoOpBuildEnvironment from pip._internal.exceptions import InstallationError from pip._internal.models.link import Link from pip._internal.operations.generate_metadata import get_metadata_generator from pip._internal.pyproject import load_pyproject_toml, make_pyproject_path from pip._internal.req.req_uninstall import UninstallPathSet from pip._internal.utils.compat import native_str from pip._internal.utils.hashes import Hashes from pip._internal.utils.logging import indent_log from pip._internal.utils.marker_files import ( PIP_DELETE_MARKER_FILENAME, has_delete_marker_file, ) from pip._internal.utils.misc import ( _make_build_dir, ask_path_exists, backup_dir, display_path, dist_in_install_path, dist_in_site_packages, dist_in_usersite, ensure_dir, get_installed_version, hide_url, redact_auth_from_url, rmtree, ) from pip._internal.utils.packaging import get_metadata from pip._internal.utils.setuptools_build import make_setuptools_shim_args from pip._internal.utils.subprocess import ( call_subprocess, runner_with_spinner_message, ) from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.virtualenv import running_under_virtualenv from pip._internal.vcs import vcs if MYPY_CHECK_RUNNING: from typing import ( Any, Dict, Iterable, List, Optional, Sequence, Union, ) from pip._internal.build_env import BuildEnvironment from pip._internal.cache import WheelCache from pip._internal.index import PackageFinder from pip._vendor.pkg_resources import Distribution from pip._vendor.packaging.specifiers import SpecifierSet from pip._vendor.packaging.markers import Marker logger = logging.getLogger(__name__) class InstallRequirement(object): """ Represents something that may be installed later on, may have information about where to fetch the relevant requirement and also contains logic for installing the said requirement. """ def __init__( self, req, # type: Optional[Requirement] comes_from, # type: Optional[Union[str, InstallRequirement]] source_dir=None, # type: Optional[str] editable=False, # type: bool link=None, # type: Optional[Link] markers=None, # type: Optional[Marker] use_pep517=None, # type: Optional[bool] isolated=False, # type: bool options=None, # type: Optional[Dict[str, Any]] wheel_cache=None, # type: Optional[WheelCache] constraint=False, # type: bool extras=() # type: Iterable[str] ): # type: (...) -> None assert req is None or isinstance(req, Requirement), req self.req = req self.comes_from = comes_from self.constraint = constraint if source_dir is None: self.source_dir = None # type: Optional[str] else: self.source_dir = os.path.normpath(os.path.abspath(source_dir)) self.editable = editable self._wheel_cache = wheel_cache if link is None and req and req.url: # PEP 508 URL requirement link = Link(req.url) self.link = self.original_link = link if extras: self.extras = extras elif req: self.extras = { pkg_resources.safe_extra(extra) for extra in req.extras } else: self.extras = set() if markers is None and req: markers = req.marker self.markers = markers # This holds the pkg_resources.Distribution object if this requirement # is already available: self.satisfied_by = None # This hold the pkg_resources.Distribution object if this requirement # conflicts with another installed distribution: self.conflicts_with = None # Temporary build location self._temp_build_dir = None # type: Optional[TempDirectory] # Used to store the global directory where the _temp_build_dir should # have been created. Cf move_to_correct_build_directory method. self._ideal_build_dir = None # type: Optional[str] # Set to True after successful installation self.install_succeeded = None # type: Optional[bool] self.options = options if options else {} # Set to True after successful preparation of this requirement self.prepared = False self.is_direct = False self.isolated = isolated self.build_env = NoOpBuildEnvironment() # type: BuildEnvironment # For PEP 517, the directory where we request the project metadata # gets stored. We need this to pass to build_wheel, so the backend # can ensure that the wheel matches the metadata (see the PEP for # details). self.metadata_directory = None # type: Optional[str] # The static build requirements (from pyproject.toml) self.pyproject_requires = None # type: Optional[List[str]] # Build requirements that we will check are available self.requirements_to_check = [] # type: List[str] # The PEP 517 backend we should use to build the project self.pep517_backend = None # type: Optional[Pep517HookCaller] # Are we using PEP 517 for this requirement? # After pyproject.toml has been loaded, the only valid values are True # and False. Before loading, None is valid (meaning "use the default"). # Setting an explicit value before loading pyproject.toml is supported, # but after loading this flag should be treated as read only. self.use_pep517 = use_pep517 def __str__(self): # type: () -> str if self.req: s = str(self.req) if self.link: s += ' from %s' % redact_auth_from_url(self.link.url) elif self.link: s = redact_auth_from_url(self.link.url) else: s = '' if self.satisfied_by is not None: s += ' in %s' % display_path(self.satisfied_by.location) if self.comes_from: if isinstance(self.comes_from, six.string_types): comes_from = self.comes_from # type: Optional[str] else: comes_from = self.comes_from.from_path() if comes_from: s += ' (from %s)' % comes_from return s def __repr__(self): # type: () -> str return '<%s object: %s editable=%r>' % ( self.__class__.__name__, str(self), self.editable) def format_debug(self): # type: () -> str """An un-tested helper for getting state, for debugging. """ attributes = vars(self) names = sorted(attributes) state = ( "{}={!r}".format(attr, attributes[attr]) for attr in sorted(names) ) return '<{name} object: {{{state}}}>'.format( name=self.__class__.__name__, state=", ".join(state), ) def populate_link(self, finder, upgrade, require_hashes): # type: (PackageFinder, bool, bool) -> None """Ensure that if a link can be found for this, that it is found. Note that self.link may still be None - if Upgrade is False and the requirement is already installed. If require_hashes is True, don't use the wheel cache, because cached wheels, always built locally, have different hashes than the files downloaded from the index server and thus throw false hash mismatches. Furthermore, cached wheels at present have undeterministic contents due to file modification times. """ if self.link is None: self.link = finder.find_requirement(self, upgrade) if self._wheel_cache is not None and not require_hashes: old_link = self.link supported_tags = pep425tags.get_supported() self.link = self._wheel_cache.get( link=self.link, package_name=self.name, supported_tags=supported_tags, ) if old_link != self.link: logger.debug('Using cached wheel link: %s', self.link) # Things that are valid for all kinds of requirements? @property def name(self): # type: () -> Optional[str] if self.req is None: return None return native_str(pkg_resources.safe_name(self.req.name)) @property def specifier(self): # type: () -> SpecifierSet return self.req.specifier @property def is_pinned(self): # type: () -> bool """Return whether I am pinned to an exact version. For example, some-package==1.2 is pinned; some-package>1.2 is not. """ specifiers = self.specifier return (len(specifiers) == 1 and next(iter(specifiers)).operator in {'==', '==='}) @property def installed_version(self): # type: () -> Optional[str] return get_installed_version(self.name) def match_markers(self, extras_requested=None): # type: (Optional[Iterable[str]]) -> bool if not extras_requested: # Provide an extra to safely evaluate the markers # without matching any extra extras_requested = ('',) if self.markers is not None: return any( self.markers.evaluate({'extra': extra}) for extra in extras_requested) else: return True @property def has_hash_options(self): # type: () -> bool """Return whether any known-good hashes are specified as options. These activate --require-hashes mode; hashes specified as part of a URL do not. """ return bool(self.options.get('hashes', {})) def hashes(self, trust_internet=True): # type: (bool) -> Hashes """Return a hash-comparer that considers my option- and URL-based hashes to be known-good. Hashes in URLs--ones embedded in the requirements file, not ones downloaded from an index server--are almost peers with ones from flags. They satisfy --require-hashes (whether it was implicitly or explicitly activated) but do not activate it. md5 and sha224 are not allowed in flags, which should nudge people toward good algos. We always OR all hashes together, even ones from URLs. :param trust_internet: Whether to trust URL-based (#md5=...) hashes downloaded from the internet, as by populate_link() """ good_hashes = self.options.get('hashes', {}).copy() link = self.link if trust_internet else self.original_link if link and link.hash: good_hashes.setdefault(link.hash_name, []).append(link.hash) return Hashes(good_hashes) def from_path(self): # type: () -> Optional[str] """Format a nice indicator to show where this "comes from" """ if self.req is None: return None s = str(self.req) if self.comes_from: if isinstance(self.comes_from, six.string_types): comes_from = self.comes_from else: comes_from = self.comes_from.from_path() if comes_from: s += '->' + comes_from return s def ensure_build_location(self, build_dir): # type: (str) -> str assert build_dir is not None if self._temp_build_dir is not None: assert self._temp_build_dir.path return self._temp_build_dir.path if self.req is None: # for requirement via a path to a directory: the name of the # package is not available yet so we create a temp directory # Once run_egg_info will have run, we'll be able to fix it via # move_to_correct_build_directory(). # Some systems have /tmp as a symlink which confuses custom # builds (such as numpy). Thus, we ensure that the real path # is returned. self._temp_build_dir = TempDirectory(kind="req-build") self._ideal_build_dir = build_dir return self._temp_build_dir.path if self.editable: name = self.name.lower() else: name = self.name # FIXME: Is there a better place to create the build_dir? (hg and bzr # need this) if not os.path.exists(build_dir): logger.debug('Creating directory %s', build_dir) _make_build_dir(build_dir) return os.path.join(build_dir, name) def move_to_correct_build_directory(self): # type: () -> None """Move self._temp_build_dir to "self._ideal_build_dir/self.req.name" For some requirements (e.g. a path to a directory), the name of the package is not available until we run egg_info, so the build_location will return a temporary directory and store the _ideal_build_dir. This is only called to "fix" the build directory after generating metadata. """ if self.source_dir is not None: return assert self.req is not None assert self._temp_build_dir assert ( self._ideal_build_dir is not None and self._ideal_build_dir.path # type: ignore ) old_location = self._temp_build_dir self._temp_build_dir = None # checked inside ensure_build_location # Figure out the correct place to put the files. new_location = self.ensure_build_location(self._ideal_build_dir) if os.path.exists(new_location): raise InstallationError( 'A package already exists in %s; please remove it to continue' % display_path(new_location) ) # Move the files to the correct location. logger.debug( 'Moving package %s from %s to new location %s', self, display_path(old_location.path), display_path(new_location), ) shutil.move(old_location.path, new_location) # Update directory-tracking variables, to be in line with new_location self.source_dir = os.path.normpath(os.path.abspath(new_location)) self._temp_build_dir = TempDirectory( path=new_location, kind="req-install", ) # Correct the metadata directory, if it exists if self.metadata_directory: old_meta = self.metadata_directory rel = os.path.relpath(old_meta, start=old_location.path) new_meta = os.path.join(new_location, rel) new_meta = os.path.normpath(os.path.abspath(new_meta)) self.metadata_directory = new_meta # Done with any "move built files" work, since have moved files to the # "ideal" build location. Setting to None allows to clearly flag that # no more moves are needed. self._ideal_build_dir = None def remove_temporary_source(self): # type: () -> None """Remove the source files from this requirement, if they are marked for deletion""" if self.source_dir and has_delete_marker_file(self.source_dir): logger.debug('Removing source in %s', self.source_dir) rmtree(self.source_dir) self.source_dir = None if self._temp_build_dir: self._temp_build_dir.cleanup() self._temp_build_dir = None self.build_env.cleanup() def check_if_exists(self, use_user_site): # type: (bool) -> bool """Find an installed distribution that satisfies or conflicts with this requirement, and set self.satisfied_by or self.conflicts_with appropriately. """ if self.req is None: return False try: # get_distribution() will resolve the entire list of requirements # anyway, and we've already determined that we need the requirement # in question, so strip the marker so that we don't try to # evaluate it. no_marker = Requirement(str(self.req)) no_marker.marker = None self.satisfied_by = pkg_resources.get_distribution(str(no_marker)) if self.editable and self.satisfied_by: self.conflicts_with = self.satisfied_by # when installing editables, nothing pre-existing should ever # satisfy self.satisfied_by = None return True except pkg_resources.DistributionNotFound: return False except pkg_resources.VersionConflict: existing_dist = pkg_resources.get_distribution( self.req.name ) if use_user_site: if dist_in_usersite(existing_dist): self.conflicts_with = existing_dist elif (running_under_virtualenv() and dist_in_site_packages(existing_dist)): raise InstallationError( "Will not install to the user site because it will " "lack sys.path precedence to %s in %s" % (existing_dist.project_name, existing_dist.location) ) elif dist_in_install_path(existing_dist): self.conflicts_with = existing_dist return True # Things valid for wheels @property def is_wheel(self): # type: () -> bool if not self.link: return False return self.link.is_wheel def move_wheel_files( self, wheeldir, # type: str root=None, # type: Optional[str] home=None, # type: Optional[str] prefix=None, # type: Optional[str] warn_script_location=True, # type: bool use_user_site=False, # type: bool pycompile=True # type: bool ): # type: (...) -> None wheel.move_wheel_files( self.name, self.req, wheeldir, user=use_user_site, home=home, root=root, prefix=prefix, pycompile=pycompile, isolated=self.isolated, warn_script_location=warn_script_location, ) # Things valid for sdists @property def unpacked_source_directory(self): # type: () -> str return os.path.join( self.source_dir, self.link and self.link.subdirectory_fragment or '') @property def setup_py_path(self): # type: () -> str assert self.source_dir, "No source dir for %s" % self setup_py = os.path.join(self.unpacked_source_directory, 'setup.py') # Python2 __file__ should not be unicode if six.PY2 and isinstance(setup_py, six.text_type): setup_py = setup_py.encode(sys.getfilesystemencoding()) return setup_py @property def pyproject_toml_path(self): # type: () -> str assert self.source_dir, "No source dir for %s" % self return make_pyproject_path(self.unpacked_source_directory) def load_pyproject_toml(self): # type: () -> None """Load the pyproject.toml file. After calling this routine, all of the attributes related to PEP 517 processing for this requirement have been set. In particular, the use_pep517 attribute can be used to determine whether we should follow the PEP 517 or legacy (setup.py) code path. """ pyproject_toml_data = load_pyproject_toml( self.use_pep517, self.pyproject_toml_path, self.setup_py_path, str(self) ) if pyproject_toml_data is None: self.use_pep517 = False return self.use_pep517 = True requires, backend, check = pyproject_toml_data self.requirements_to_check = check self.pyproject_requires = requires self.pep517_backend = Pep517HookCaller( self.unpacked_source_directory, backend ) def prepare_metadata(self): # type: () -> None """Ensure that project metadata is available. Under PEP 517, call the backend hook to prepare the metadata. Under legacy processing, call setup.py egg-info. """ assert self.source_dir metadata_generator = get_metadata_generator(self) with indent_log(): self.metadata_directory = metadata_generator(self) if not self.req: if isinstance(parse_version(self.metadata["Version"]), Version): op = "==" else: op = "===" self.req = Requirement( "".join([ self.metadata["Name"], op, self.metadata["Version"], ]) ) self.move_to_correct_build_directory() else: metadata_name = canonicalize_name(self.metadata["Name"]) if canonicalize_name(self.req.name) != metadata_name: logger.warning( 'Generating metadata for package %s ' 'produced metadata for project name %s. Fix your ' '#egg=%s fragments.', self.name, metadata_name, self.name ) self.req = Requirement(metadata_name) def prepare_pep517_metadata(self): # type: () -> str assert self.pep517_backend is not None # NOTE: This needs to be refactored to stop using atexit metadata_tmpdir = TempDirectory(kind="modern-metadata") atexit.register(metadata_tmpdir.cleanup) metadata_dir = metadata_tmpdir.path with self.build_env: # Note that Pep517HookCaller implements a fallback for # prepare_metadata_for_build_wheel, so we don't have to # consider the possibility that this hook doesn't exist. runner = runner_with_spinner_message("Preparing wheel metadata") backend = self.pep517_backend with backend.subprocess_runner(runner): distinfo_dir = backend.prepare_metadata_for_build_wheel( metadata_dir ) return os.path.join(metadata_dir, distinfo_dir) @property def metadata(self): # type: () -> Any if not hasattr(self, '_metadata'): self._metadata = get_metadata(self.get_dist()) return self._metadata def get_dist(self): # type: () -> Distribution """Return a pkg_resources.Distribution for this requirement""" dist_dir = self.metadata_directory.rstrip(os.sep) # Determine the correct Distribution object type. if dist_dir.endswith(".egg-info"): dist_cls = pkg_resources.Distribution else: assert dist_dir.endswith(".dist-info") dist_cls = pkg_resources.DistInfoDistribution # Build a PathMetadata object, from path to metadata. :wink: base_dir, dist_dir_name = os.path.split(dist_dir) dist_name = os.path.splitext(dist_dir_name)[0] metadata = pkg_resources.PathMetadata(base_dir, dist_dir) return dist_cls( base_dir, project_name=dist_name, metadata=metadata, ) def assert_source_matches_version(self): # type: () -> None assert self.source_dir version = self.metadata['version'] if self.req.specifier and version not in self.req.specifier: logger.warning( 'Requested %s, but installing version %s', self, version, ) else: logger.debug( 'Source in %s has version %s, which satisfies requirement %s', display_path(self.source_dir), version, self, ) # For both source distributions and editables def ensure_has_source_dir(self, parent_dir): # type: (str) -> None """Ensure that a source_dir is set. This will create a temporary build dir if the name of the requirement isn't known yet. :param parent_dir: The ideal pip parent_dir for the source_dir. Generally src_dir for editables and build_dir for sdists. :return: self.source_dir """ if self.source_dir is None: self.source_dir = self.ensure_build_location(parent_dir) # For editable installations def install_editable( self, install_options, # type: List[str] global_options=(), # type: Sequence[str] prefix=None # type: Optional[str] ): # type: (...) -> None logger.info('Running setup.py develop for %s', self.name) if prefix: prefix_param = ['--prefix={}'.format(prefix)] install_options = list(install_options) + prefix_param base_cmd = make_setuptools_shim_args( self.setup_py_path, global_options=global_options, no_user_config=self.isolated ) with indent_log(): with self.build_env: call_subprocess( base_cmd + ['develop', '--no-deps'] + list(install_options), cwd=self.unpacked_source_directory, ) self.install_succeeded = True def update_editable(self, obtain=True): # type: (bool) -> None if not self.link: logger.debug( "Cannot update repository at %s; repository location is " "unknown", self.source_dir, ) return assert self.editable assert self.source_dir if self.link.scheme == 'file': # Static paths don't get updated return assert '+' in self.link.url, "bad url: %r" % self.link.url vc_type, url = self.link.url.split('+', 1) vcs_backend = vcs.get_backend(vc_type) if vcs_backend: hidden_url = hide_url(self.link.url) if obtain: vcs_backend.obtain(self.source_dir, url=hidden_url) else: vcs_backend.export(self.source_dir, url=hidden_url) else: assert 0, ( 'Unexpected version control type (in %s): %s' % (self.link, vc_type)) # Top-level Actions def uninstall(self, auto_confirm=False, verbose=False, use_user_site=False): # type: (bool, bool, bool) -> Optional[UninstallPathSet] """ Uninstall the distribution currently satisfying this requirement. Prompts before removing or modifying files unless ``auto_confirm`` is True. Refuses to delete or modify files outside of ``sys.prefix`` - thus uninstallation within a virtual environment can only modify that virtual environment, even if the virtualenv is linked to global site-packages. """ if not self.check_if_exists(use_user_site): logger.warning("Skipping %s as it is not installed.", self.name) return None dist = self.satisfied_by or self.conflicts_with uninstalled_pathset = UninstallPathSet.from_dist(dist) uninstalled_pathset.remove(auto_confirm, verbose) return uninstalled_pathset def _clean_zip_name(self, name, prefix): # only used by archive. # type: (str, str) -> str assert name.startswith(prefix + os.path.sep), ( "name %r doesn't start with prefix %r" % (name, prefix) ) name = name[len(prefix) + 1:] name = name.replace(os.path.sep, '/') return name def _get_archive_name(self, path, parentdir, rootdir): # type: (str, str, str) -> str path = os.path.join(parentdir, path) name = self._clean_zip_name(path, rootdir) return self.name + '/' + name def archive(self, build_dir): # type: (str) -> None """Saves archive to provided build_dir. Used for saving downloaded VCS requirements as part of `pip download`. """ assert self.source_dir create_archive = True archive_name = '%s-%s.zip' % (self.name, self.metadata["version"]) archive_path = os.path.join(build_dir, archive_name) if os.path.exists(archive_path): response = ask_path_exists( 'The file %s exists. (i)gnore, (w)ipe, (b)ackup, (a)bort ' % display_path(archive_path), ('i', 'w', 'b', 'a')) if response == 'i': create_archive = False elif response == 'w': logger.warning('Deleting %s', display_path(archive_path)) os.remove(archive_path) elif response == 'b': dest_file = backup_dir(archive_path) logger.warning( 'Backing up %s to %s', display_path(archive_path), display_path(dest_file), ) shutil.move(archive_path, dest_file) elif response == 'a': sys.exit(-1) if not create_archive: return zip_output = zipfile.ZipFile( archive_path, 'w', zipfile.ZIP_DEFLATED, allowZip64=True, ) with zip_output: dir = os.path.normcase( os.path.abspath(self.unpacked_source_directory) ) for dirpath, dirnames, filenames in os.walk(dir): if 'pip-egg-info' in dirnames: dirnames.remove('pip-egg-info') for dirname in dirnames: dir_arcname = self._get_archive_name( dirname, parentdir=dirpath, rootdir=dir, ) zipdir = zipfile.ZipInfo(dir_arcname + '/') zipdir.external_attr = 0x1ED << 16 # 0o755 zip_output.writestr(zipdir, '') for filename in filenames: if filename == PIP_DELETE_MARKER_FILENAME: continue file_arcname = self._get_archive_name( filename, parentdir=dirpath, rootdir=dir, ) filename = os.path.join(dirpath, filename) zip_output.write(filename, file_arcname) logger.info('Saved %s', display_path(archive_path)) def install( self, install_options, # type: List[str] global_options=None, # type: Optional[Sequence[str]] root=None, # type: Optional[str] home=None, # type: Optional[str] prefix=None, # type: Optional[str] warn_script_location=True, # type: bool use_user_site=False, # type: bool pycompile=True # type: bool ): # type: (...) -> None global_options = global_options if global_options is not None else [] if self.editable: self.install_editable( install_options, global_options, prefix=prefix, ) return if self.is_wheel: version = wheel.wheel_version(self.source_dir) wheel.check_compatibility(version, self.name) self.move_wheel_files( self.source_dir, root=root, prefix=prefix, home=home, warn_script_location=warn_script_location, use_user_site=use_user_site, pycompile=pycompile, ) self.install_succeeded = True return # Extend the list of global and install options passed on to # the setup.py call with the ones from the requirements file. # Options specified in requirements file override those # specified on the command line, since the last option given # to setup.py is the one that is used. global_options = list(global_options) + \ self.options.get('global_options', []) install_options = list(install_options) + \ self.options.get('install_options', []) with TempDirectory(kind="record") as temp_dir: record_filename = os.path.join(temp_dir.path, 'install-record.txt') install_args = self.get_install_args( global_options, record_filename, root, prefix, pycompile, ) runner = runner_with_spinner_message( "Running setup.py install for {}".format(self.name) ) with indent_log(), self.build_env: runner( cmd=install_args + install_options, cwd=self.unpacked_source_directory, ) if not os.path.exists(record_filename): logger.debug('Record file %s not found', record_filename) return self.install_succeeded = True def prepend_root(path): # type: (str) -> str if root is None or not os.path.isabs(path): return path else: return change_root(root, path) with open(record_filename) as f: for line in f: directory = os.path.dirname(line) if directory.endswith('.egg-info'): egg_info_dir = prepend_root(directory) break else: logger.warning( 'Could not find .egg-info directory in install record' ' for %s', self, ) # FIXME: put the record somewhere return new_lines = [] with open(record_filename) as f: for line in f: filename = line.strip() if os.path.isdir(filename): filename += os.path.sep new_lines.append( os.path.relpath(prepend_root(filename), egg_info_dir) ) new_lines.sort() ensure_dir(egg_info_dir) inst_files_path = os.path.join(egg_info_dir, 'installed-files.txt') with open(inst_files_path, 'w') as f: f.write('\n'.join(new_lines) + '\n') def get_install_args( self, global_options, # type: Sequence[str] record_filename, # type: str root, # type: Optional[str] prefix, # type: Optional[str] pycompile # type: bool ): # type: (...) -> List[str] install_args = make_setuptools_shim_args( self.setup_py_path, global_options=global_options, no_user_config=self.isolated, unbuffered_output=True ) install_args += ['install', '--record', record_filename] install_args += ['--single-version-externally-managed'] if root is not None: install_args += ['--root', root] if prefix is not None: install_args += ['--prefix', prefix] if pycompile: install_args += ["--compile"] else: install_args += ["--no-compile"] if running_under_virtualenv(): py_ver_str = 'python' + sysconfig.get_python_version() install_args += ['--install-headers', os.path.join(sys.prefix, 'include', 'site', py_ver_str, self.name)] return install_args site-packages/pip/_internal/req/req_set.py000064400000017704151464452710014654 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False from __future__ import absolute_import import logging from collections import OrderedDict from pip._vendor.packaging.utils import canonicalize_name from pip._internal import pep425tags from pip._internal.exceptions import InstallationError from pip._internal.utils.logging import indent_log from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.wheel import Wheel if MYPY_CHECK_RUNNING: from typing import Dict, Iterable, List, Optional, Tuple from pip._internal.req.req_install import InstallRequirement logger = logging.getLogger(__name__) class RequirementSet(object): def __init__(self, require_hashes=False, check_supported_wheels=True): # type: (bool, bool) -> None """Create a RequirementSet. """ self.requirements = OrderedDict() # type: Dict[str, InstallRequirement] # noqa: E501 self.require_hashes = require_hashes self.check_supported_wheels = check_supported_wheels self.unnamed_requirements = [] # type: List[InstallRequirement] self.successfully_downloaded = [] # type: List[InstallRequirement] self.reqs_to_cleanup = [] # type: List[InstallRequirement] def __str__(self): # type: () -> str requirements = sorted( (req for req in self.requirements.values() if not req.comes_from), key=lambda req: canonicalize_name(req.name), ) return ' '.join(str(req.req) for req in requirements) def __repr__(self): # type: () -> str requirements = sorted( self.requirements.values(), key=lambda req: canonicalize_name(req.name), ) format_string = '<{classname} object; {count} requirement(s): {reqs}>' return format_string.format( classname=self.__class__.__name__, count=len(requirements), reqs=', '.join(str(req.req) for req in requirements), ) def add_unnamed_requirement(self, install_req): # type: (InstallRequirement) -> None assert not install_req.name self.unnamed_requirements.append(install_req) def add_named_requirement(self, install_req): # type: (InstallRequirement) -> None assert install_req.name project_name = canonicalize_name(install_req.name) self.requirements[project_name] = install_req def add_requirement( self, install_req, # type: InstallRequirement parent_req_name=None, # type: Optional[str] extras_requested=None # type: Optional[Iterable[str]] ): # type: (...) -> Tuple[List[InstallRequirement], Optional[InstallRequirement]] # noqa: E501 """Add install_req as a requirement to install. :param parent_req_name: The name of the requirement that needed this added. The name is used because when multiple unnamed requirements resolve to the same name, we could otherwise end up with dependency links that point outside the Requirements set. parent_req must already be added. Note that None implies that this is a user supplied requirement, vs an inferred one. :param extras_requested: an iterable of extras used to evaluate the environment markers. :return: Additional requirements to scan. That is either [] if the requirement is not applicable, or [install_req] if the requirement is applicable and has just been added. """ # If the markers do not match, ignore this requirement. if not install_req.match_markers(extras_requested): logger.info( "Ignoring %s: markers '%s' don't match your environment", install_req.name, install_req.markers, ) return [], None # If the wheel is not supported, raise an error. # Should check this after filtering out based on environment markers to # allow specifying different wheels based on the environment/OS, in a # single requirements file. if install_req.link and install_req.link.is_wheel: wheel = Wheel(install_req.link.filename) tags = pep425tags.get_supported() if (self.check_supported_wheels and not wheel.supported(tags)): raise InstallationError( "%s is not a supported wheel on this platform." % wheel.filename ) # This next bit is really a sanity check. assert install_req.is_direct == (parent_req_name is None), ( "a direct req shouldn't have a parent and also, " "a non direct req should have a parent" ) # Unnamed requirements are scanned again and the requirement won't be # added as a dependency until after scanning. if not install_req.name: self.add_unnamed_requirement(install_req) return [install_req], None try: existing_req = self.get_requirement(install_req.name) except KeyError: existing_req = None has_conflicting_requirement = ( parent_req_name is None and existing_req and not existing_req.constraint and existing_req.extras == install_req.extras and existing_req.req.specifier != install_req.req.specifier ) if has_conflicting_requirement: raise InstallationError( "Double requirement given: %s (already in %s, name=%r)" % (install_req, existing_req, install_req.name) ) # When no existing requirement exists, add the requirement as a # dependency and it will be scanned again after. if not existing_req: self.add_named_requirement(install_req) # We'd want to rescan this requirement later return [install_req], install_req # Assume there's no need to scan, and that we've already # encountered this for scanning. if install_req.constraint or not existing_req.constraint: return [], existing_req does_not_satisfy_constraint = ( install_req.link and not ( existing_req.link and install_req.link.path == existing_req.link.path ) ) if does_not_satisfy_constraint: self.reqs_to_cleanup.append(install_req) raise InstallationError( "Could not satisfy constraints for '%s': " "installation from path or url cannot be " "constrained to a version" % install_req.name, ) # If we're now installing a constraint, mark the existing # object for real installation. existing_req.constraint = False existing_req.extras = tuple(sorted( set(existing_req.extras) | set(install_req.extras) )) logger.debug( "Setting %s extras to: %s", existing_req, existing_req.extras, ) # Return the existing requirement for addition to the parent and # scanning again. return [existing_req], existing_req def has_requirement(self, name): # type: (str) -> bool project_name = canonicalize_name(name) return ( project_name in self.requirements and not self.requirements[project_name].constraint ) def get_requirement(self, name): # type: (str) -> InstallRequirement project_name = canonicalize_name(name) if project_name in self.requirements: return self.requirements[project_name] raise KeyError("No project with the name %r" % name) def cleanup_files(self): # type: () -> None """Clean up files, remove builds.""" logger.debug('Cleaning up...') with indent_log(): for req in self.reqs_to_cleanup: req.remove_temporary_source() site-packages/pip/_internal/req/req_file.py000064400000033726151464452710015002 0ustar00""" Requirements file parsing """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False from __future__ import absolute_import import optparse import os import re import shlex import sys from pip._vendor.six.moves import filterfalse from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.cli import cmdoptions from pip._internal.download import get_file_content from pip._internal.exceptions import RequirementsFileParseError from pip._internal.models.search_scope import SearchScope from pip._internal.req.constructors import ( install_req_from_editable, install_req_from_line, ) from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ( Any, Callable, Iterator, List, NoReturn, Optional, Text, Tuple, ) from pip._internal.req import InstallRequirement from pip._internal.cache import WheelCache from pip._internal.index import PackageFinder from pip._internal.network.session import PipSession ReqFileLines = Iterator[Tuple[int, Text]] __all__ = ['parse_requirements'] SCHEME_RE = re.compile(r'^(http|https|file):', re.I) COMMENT_RE = re.compile(r'(^|\s+)#.*$') # Matches environment variable-style values in '${MY_VARIABLE_1}' with the # variable name consisting of only uppercase letters, digits or the '_' # (underscore). This follows the POSIX standard defined in IEEE Std 1003.1, # 2013 Edition. ENV_VAR_RE = re.compile(r'(?P\$\{(?P[A-Z0-9_]+)\})') SUPPORTED_OPTIONS = [ cmdoptions.constraints, cmdoptions.editable, cmdoptions.requirements, cmdoptions.no_index, cmdoptions.index_url, cmdoptions.find_links, cmdoptions.extra_index_url, cmdoptions.always_unzip, cmdoptions.no_binary, cmdoptions.only_binary, cmdoptions.pre, cmdoptions.trusted_host, cmdoptions.require_hashes, ] # type: List[Callable[..., optparse.Option]] # options to be passed to requirements SUPPORTED_OPTIONS_REQ = [ cmdoptions.install_options, cmdoptions.global_options, cmdoptions.hash, ] # type: List[Callable[..., optparse.Option]] # the 'dest' string values SUPPORTED_OPTIONS_REQ_DEST = [str(o().dest) for o in SUPPORTED_OPTIONS_REQ] def parse_requirements( filename, # type: str finder=None, # type: Optional[PackageFinder] comes_from=None, # type: Optional[str] options=None, # type: Optional[optparse.Values] session=None, # type: Optional[PipSession] constraint=False, # type: bool wheel_cache=None, # type: Optional[WheelCache] use_pep517=None # type: Optional[bool] ): # type: (...) -> Iterator[InstallRequirement] """Parse a requirements file and yield InstallRequirement instances. :param filename: Path or url of requirements file. :param finder: Instance of pip.index.PackageFinder. :param comes_from: Origin description of requirements. :param options: cli options. :param session: Instance of pip.download.PipSession. :param constraint: If true, parsing a constraint file rather than requirements file. :param wheel_cache: Instance of pip.wheel.WheelCache :param use_pep517: Value of the --use-pep517 option. """ if session is None: raise TypeError( "parse_requirements() missing 1 required keyword argument: " "'session'" ) _, content = get_file_content( filename, comes_from=comes_from, session=session ) lines_enum = preprocess(content, options) for line_number, line in lines_enum: req_iter = process_line(line, filename, line_number, finder, comes_from, options, session, wheel_cache, use_pep517=use_pep517, constraint=constraint) for req in req_iter: yield req def preprocess(content, options): # type: (Text, Optional[optparse.Values]) -> ReqFileLines """Split, filter, and join lines, and return a line iterator :param content: the content of the requirements file :param options: cli options """ lines_enum = enumerate(content.splitlines(), start=1) # type: ReqFileLines lines_enum = join_lines(lines_enum) lines_enum = ignore_comments(lines_enum) lines_enum = skip_regex(lines_enum, options) lines_enum = expand_env_variables(lines_enum) return lines_enum def process_line( line, # type: Text filename, # type: str line_number, # type: int finder=None, # type: Optional[PackageFinder] comes_from=None, # type: Optional[str] options=None, # type: Optional[optparse.Values] session=None, # type: Optional[PipSession] wheel_cache=None, # type: Optional[WheelCache] use_pep517=None, # type: Optional[bool] constraint=False, # type: bool ): # type: (...) -> Iterator[InstallRequirement] """Process a single requirements line; This can result in creating/yielding requirements, or updating the finder. For lines that contain requirements, the only options that have an effect are from SUPPORTED_OPTIONS_REQ, and they are scoped to the requirement. Other options from SUPPORTED_OPTIONS may be present, but are ignored. For lines that do not contain requirements, the only options that have an effect are from SUPPORTED_OPTIONS. Options from SUPPORTED_OPTIONS_REQ may be present, but are ignored. These lines may contain multiple options (although our docs imply only one is supported), and all our parsed and affect the finder. :param constraint: If True, parsing a constraints file. :param options: OptionParser options that we may update """ parser = build_parser(line) defaults = parser.get_default_values() defaults.index_url = None if finder: defaults.format_control = finder.format_control args_str, options_str = break_args_options(line) # Prior to 2.7.3, shlex cannot deal with unicode entries if sys.version_info < (2, 7, 3): # https://github.com/python/mypy/issues/1174 options_str = options_str.encode('utf8') # type: ignore # https://github.com/python/mypy/issues/1174 opts, _ = parser.parse_args( shlex.split(options_str), defaults) # type: ignore # preserve for the nested code path line_comes_from = '%s %s (line %s)' % ( '-c' if constraint else '-r', filename, line_number, ) # yield a line requirement if args_str: isolated = options.isolated_mode if options else False if options: cmdoptions.check_install_build_global(options, opts) # get the options that apply to requirements req_options = {} for dest in SUPPORTED_OPTIONS_REQ_DEST: if dest in opts.__dict__ and opts.__dict__[dest]: req_options[dest] = opts.__dict__[dest] line_source = 'line {} of {}'.format(line_number, filename) yield install_req_from_line( args_str, comes_from=line_comes_from, use_pep517=use_pep517, isolated=isolated, options=req_options, wheel_cache=wheel_cache, constraint=constraint, line_source=line_source, ) # yield an editable requirement elif opts.editables: isolated = options.isolated_mode if options else False yield install_req_from_editable( opts.editables[0], comes_from=line_comes_from, use_pep517=use_pep517, constraint=constraint, isolated=isolated, wheel_cache=wheel_cache ) # parse a nested requirements file elif opts.requirements or opts.constraints: if opts.requirements: req_path = opts.requirements[0] nested_constraint = False else: req_path = opts.constraints[0] nested_constraint = True # original file is over http if SCHEME_RE.search(filename): # do a url join so relative paths work req_path = urllib_parse.urljoin(filename, req_path) # original file and nested file are paths elif not SCHEME_RE.search(req_path): # do a join so relative paths work req_path = os.path.join(os.path.dirname(filename), req_path) # TODO: Why not use `comes_from='-r {} (line {})'` here as well? parsed_reqs = parse_requirements( req_path, finder, comes_from, options, session, constraint=nested_constraint, wheel_cache=wheel_cache ) for req in parsed_reqs: yield req # percolate hash-checking option upward elif opts.require_hashes: options.require_hashes = opts.require_hashes # set finder options elif finder: find_links = finder.find_links index_urls = finder.index_urls if opts.index_url: index_urls = [opts.index_url] if opts.no_index is True: index_urls = [] if opts.extra_index_urls: index_urls.extend(opts.extra_index_urls) if opts.find_links: # FIXME: it would be nice to keep track of the source # of the find_links: support a find-links local path # relative to a requirements file. value = opts.find_links[0] req_dir = os.path.dirname(os.path.abspath(filename)) relative_to_reqs_file = os.path.join(req_dir, value) if os.path.exists(relative_to_reqs_file): value = relative_to_reqs_file find_links.append(value) search_scope = SearchScope( find_links=find_links, index_urls=index_urls, ) finder.search_scope = search_scope if opts.pre: finder.set_allow_all_prereleases() for host in opts.trusted_hosts or []: source = 'line {} of {}'.format(line_number, filename) session.add_trusted_host(host, source=source) def break_args_options(line): # type: (Text) -> Tuple[str, Text] """Break up the line into an args and options string. We only want to shlex (and then optparse) the options, not the args. args can contain markers which are corrupted by shlex. """ tokens = line.split(' ') args = [] options = tokens[:] for token in tokens: if token.startswith('-') or token.startswith('--'): break else: args.append(token) options.pop(0) return ' '.join(args), ' '.join(options) # type: ignore def build_parser(line): # type: (Text) -> optparse.OptionParser """ Return a parser for parsing requirement lines """ parser = optparse.OptionParser(add_help_option=False) option_factories = SUPPORTED_OPTIONS + SUPPORTED_OPTIONS_REQ for option_factory in option_factories: option = option_factory() parser.add_option(option) # By default optparse sys.exits on parsing errors. We want to wrap # that in our own exception. def parser_exit(self, msg): # type: (Any, str) -> NoReturn # add offending line msg = 'Invalid requirement: %s\n%s' % (line, msg) raise RequirementsFileParseError(msg) # NOTE: mypy disallows assigning to a method # https://github.com/python/mypy/issues/2427 parser.exit = parser_exit # type: ignore return parser def join_lines(lines_enum): # type: (ReqFileLines) -> ReqFileLines """Joins a line ending in '\' with the previous line (except when following comments). The joined line takes on the index of the first line. """ primary_line_number = None new_line = [] # type: List[Text] for line_number, line in lines_enum: if not line.endswith('\\') or COMMENT_RE.match(line): if COMMENT_RE.match(line): # this ensures comments are always matched later line = ' ' + line if new_line: new_line.append(line) yield primary_line_number, ''.join(new_line) new_line = [] else: yield line_number, line else: if not new_line: primary_line_number = line_number new_line.append(line.strip('\\')) # last line contains \ if new_line: yield primary_line_number, ''.join(new_line) # TODO: handle space after '\'. def ignore_comments(lines_enum): # type: (ReqFileLines) -> ReqFileLines """ Strips comments and filter empty lines. """ for line_number, line in lines_enum: line = COMMENT_RE.sub('', line) line = line.strip() if line: yield line_number, line def skip_regex(lines_enum, options): # type: (ReqFileLines, Optional[optparse.Values]) -> ReqFileLines """ Skip lines that match '--skip-requirements-regex' pattern Note: the regex pattern is only built once """ skip_regex = options.skip_requirements_regex if options else None if skip_regex: pattern = re.compile(skip_regex) lines_enum = filterfalse(lambda e: pattern.search(e[1]), lines_enum) return lines_enum def expand_env_variables(lines_enum): # type: (ReqFileLines) -> ReqFileLines """Replace all environment variables that can be retrieved via `os.getenv`. The only allowed format for environment variables defined in the requirement file is `${MY_VARIABLE_1}` to ensure two things: 1. Strings that contain a `$` aren't accidentally (partially) expanded. 2. Ensure consistency across platforms for requirement files. These points are the result of a discussion on the `github pull request #3514 `_. Valid characters in variable names follow the `POSIX standard `_ and are limited to uppercase letter, digits and the `_` (underscore). """ for line_number, line in lines_enum: for env_var, var_name in ENV_VAR_RE.findall(line): value = os.getenv(var_name) if not value: continue line = line.replace(env_var, value) yield line_number, line site-packages/pip/_internal/req/__pycache__/req_tracker.cpython-38.pyc000064400000006205151464452710021774 0ustar00U .e{ @sddlmZddlZddlZddlZddlZddlZddlmZddl m Z e rddl m Z ddl mZmZmZmZddlmZddlmZeeZGd d d eZdS) )absolute_importN) TempDirectory)MYPY_CHECK_RUNNING) TracebackType)IteratorOptionalSetType)InstallRequirement)Linkc@sReZdZddZddZddZddZd d Zd d Zd dZ e j ddZ dS)RequirementTrackercCsjtjd|_|jdkrJtddd|_|jj|_tjd<td|jnd|_td|jt |_ dS)NZPIP_REQ_TRACKERFz req-tracker)deleteZkindzCreated requirements tracker %rz Re-using requirements tracker %r) osenvironget_rootr _temp_dirpathloggerdebugset_entriesselfrA/usr/lib/python3.8/site-packages/pip/_internal/req/req_tracker.py__init__s zRequirementTracker.__init__cCs|SNrrrrr __enter__&szRequirementTracker.__enter__cCs |dSr)cleanup)rexc_typeZexc_valZexc_tbrrr__exit__*szRequirementTracker.__exit__cCs$t|j}tj|j|Sr) hashlibZsha224Zurl_without_fragmentencodeZ hexdigestrrjoinr)rlinkZhashedrrr _entry_path3szRequirementTracker._entry_pathc Cs|j}t|}||}z,t|}td||fW5QRXWnztk r}z\|jtjkrd||j ksrt t|d}| |W5QRX|j |t d||jW5d}~XYnXdS)Nz%s is already being built: %swzAdded %s to build tracker %r)r%strr&open LookupErrorreadIOErrorerrnoZENOENTrAssertionErrorwriteaddrrr)rreqr%infoZ entry_pathfperrrr08s       zRequirementTracker.addcCs6|j}|j|t||td||jdS)Nz Removed %s from build tracker %r) r%rremoverunlinkr&rrr)rr1r%rrrr5Ks zRequirementTracker.removecCsNt|jD]}||q |jdk }|r2|jtd|r@dnd|jdS)Nz%s build tracker %rZRemovedZCleaned)rrr5rrrrr)rr1r5rrrrRs    zRequirementTracker.cleanupccs||dV||dSr)r0r5)rr1rrrtrack]s zRequirementTracker.trackN) __name__ __module__ __qualname__rrr!r&r0r5r contextlibcontextmanagerr7rrrrr s   r )Z __future__rr;r-r"ZloggingrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrtypesrtypingrrrr Zpip._internal.req.req_installr Zpip._internal.models.linkr Z getLoggerr8robjectr rrrrs       site-packages/pip/_internal/req/__pycache__/req_tracker.cpython-38.opt-1.pyc000064400000006145151464452710022736 0ustar00U .e{ @sddlmZddlZddlZddlZddlZddlZddlmZddl m Z e rddl m Z ddl mZmZmZmZddlmZddlmZeeZGd d d eZdS) )absolute_importN) TempDirectory)MYPY_CHECK_RUNNING) TracebackType)IteratorOptionalSetType)InstallRequirement)Linkc@sReZdZddZddZddZddZd d Zd d Zd dZ e j ddZ dS)RequirementTrackercCsjtjd|_|jdkrJtddd|_|jj|_tjd<td|jnd|_td|jt |_ dS)NZPIP_REQ_TRACKERFz req-tracker)deleteZkindzCreated requirements tracker %rz Re-using requirements tracker %r) osenvironget_rootr _temp_dirpathloggerdebugset_entriesselfrA/usr/lib/python3.8/site-packages/pip/_internal/req/req_tracker.py__init__s zRequirementTracker.__init__cCs|SNrrrrr __enter__&szRequirementTracker.__enter__cCs |dSr)cleanup)rexc_typeZexc_valZexc_tbrrr__exit__*szRequirementTracker.__exit__cCs$t|j}tj|j|Sr) hashlibZsha224Zurl_without_fragmentencodeZ hexdigestrrjoinr)rlinkZhashedrrr _entry_path3szRequirementTracker._entry_pathc Cs|j}t|}||}z,t|}td||fW5QRXWnltk r}zN|jtjkrdt|d}| |W5QRX|j |t d||jW5d}~XYnXdS)Nz%s is already being built: %swzAdded %s to build tracker %r)r%strr&open LookupErrorreadIOErrorerrnoZENOENTwriteraddrrr)rreqr%infoZ entry_pathfperrrr/8s      zRequirementTracker.addcCs6|j}|j|t||td||jdS)Nz Removed %s from build tracker %r) r%rremoverunlinkr&rrr)rr0r%rrrr4Ks zRequirementTracker.removecCsNt|jD]}||q |jdk }|r2|jtd|r@dnd|jdS)Nz%s build tracker %rZRemovedZCleaned)rrr4rrrrr)rr0r4rrrrRs    zRequirementTracker.cleanupccs||dV||dSr)r/r4)rr0rrrtrack]s zRequirementTracker.trackN) __name__ __module__ __qualname__rrr!r&r/r4r contextlibcontextmanagerr6rrrrr s   r )Z __future__rr:r-r"ZloggingrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrtypesrtypingrrrr Zpip._internal.req.req_installr Zpip._internal.models.linkr Z getLoggerr7robjectr rrrrs       site-packages/pip/_internal/req/__pycache__/req_install.cpython-38.pyc000064400000056654151464452710022024 0ustar00U .e@sLddlmZddlZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z m Z ddlmZddlmZddlmZddlmZdd lmZdd lmZmZdd lmZdd lmZdd lm Z ddl!m"Z"ddl#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/m0Z0ddl1m2Z2m3Z3m4Z4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;mm?Z?ddl@mAZAddlBmCZCmDZDddlEmFZFddlGmHZHddlImJZJddlKmLZLeHr.ddlMmNZNmOZOmPZPmQZQmRZRmSZSmTZTddlmUZUddlVmWZWdd lXmYZYdd!lZm[Z[dd"l\m]Z]dd#l^m_Z_e`eaZbGd$d%d%ecZddS)&)absolute_importN) change_root) pkg_resourcessix) Requirement)canonicalize_name)Version)parse)Pep517HookCaller) pep425tagswheel)NoOpBuildEnvironment)InstallationError)Link)get_metadata_generator)load_pyproject_tomlmake_pyproject_path)UninstallPathSet) native_str)Hashes) indent_log)PIP_DELETE_MARKER_FILENAMEhas_delete_marker_file) _make_build_dirask_path_exists backup_dir display_pathdist_in_install_pathdist_in_site_packagesdist_in_usersite ensure_dirget_installed_versionhide_urlredact_auth_from_urlrmtree) get_metadata)make_setuptools_shim_args)call_subprocessrunner_with_spinner_message) TempDirectory)MYPY_CHECK_RUNNING)running_under_virtualenv)vcs)AnyDictIterableListOptionalSequenceUnion)BuildEnvironment) WheelCache) PackageFinder) Distribution) SpecifierSet)Markerc @speZdZdZdPddZddZd d Zd d Zd dZe ddZ e ddZ e ddZ e ddZ dQddZe ddZdRddZddZd d!Zd"d#Zd$d%Zd&d'Ze d(d)ZdSd*d+Ze d,d-Ze d.d/Ze d0d1Zd2d3Zd4d5Zd6d7Ze d8d9Zd:d;Zdd?Z!dTd@dAZ"dUdBdCZ#dVdDdEZ$dFdGZ%dHdIZ&dJdKZ'dWdLdMZ(dNdOZ)dS)XInstallRequirementz Represents something that may be installed later on, may have information about where to fetch the relevant requirement and also contains logic for installing the said requirement. NFc Cs2|dkst|tst|||_||_| |_|dkrwsz.InstallRequirement.__init__..F)% isinstancerAssertionErrorreq comes_from constraint source_dirospathnormpathabspatheditable _wheel_cacheurlrlink original_linkextrassetmarkermarkers satisfied_byconflicts_with_temp_build_dir_ideal_build_dirinstall_succeededoptionsZpreparedZ is_directisolatedr build_envmetadata_directorypyproject_requiresrequirements_to_checkpep517_backend use_pep517) selfrCrDrFrKrNrSr`rZrYZ wheel_cacherErPr;r;r?__init__TsJ    zInstallRequirement.__init__cCs|jr,t|j}|jrD|dt|jj7}n|jr@t|jj}nd}|jdk rb|dt|jj7}|jrt |jt j r~|j}n |j }|r|d|7}|S)Nz from %szz in %sz (from %s)) rCstrrNr#rMrTrlocationrDrAr string_types from_pathrasrDr;r;r?__str__s     zInstallRequirement.__str__cCsd|jjt||jfS)Nz<%s object: %s editable=%r>) __class____name__rcrKrar;r;r?__repr__s zInstallRequirement.__repr__cs>t|t}fddt|D}dj|jjd|dS)z>An un-tested helper for getting state, for debugging. c3s|]}d||VqdS)z{}={!r}N)format)r=attrZ attributesr;r? sz2InstallRequirement.format_debug..z<{name} object: {{{state}}}>z, )namestate)varssortedrnrjrkjoin)ranamesrsr;rpr? format_debugs zInstallRequirement.format_debugcCsh|jdkr||||_|jdk rd|sd|j}t}|jj|j|j|d|_||jkrdtd|jdS)aEnsure that if a link can be found for this, that it is found. Note that self.link may still be None - if Upgrade is False and the requirement is already installed. If require_hashes is True, don't use the wheel cache, because cached wheels, always built locally, have different hashes than the files downloaded from the index server and thus throw false hash mismatches. Furthermore, cached wheels at present have undeterministic contents due to file modification times. N)rNZ package_namesupported_tagszUsing cached wheel link: %s) rNZfind_requirementrLr Z get_supportedgetrrloggerdebug)rafinderZupgradeZrequire_hashesZold_linkryr;r;r? populate_links  z InstallRequirement.populate_linkcCs |jdkrdStt|jjSN)rCrrZ safe_namerrrlr;r;r?rrs zInstallRequirement.namecCs|jjSr)rC specifierrlr;r;r?rszInstallRequirement.specifiercCs$|j}t|dko"tt|jdkS)zReturn whether I am pinned to an exact version. For example, some-package==1.2 is pinned; some-package>1.2 is not. >=====)rlennextiteroperator)raZ specifiersr;r;r? is_pinneds zInstallRequirement.is_pinnedcCs t|jSr)r!rrrlr;r;r?installed_versionsz$InstallRequirement.installed_versioncs0|sd}jdk r(tfdd|DSdSdS)N)c3s|]}jd|iVqdS)r>N)rSZevaluater<rlr;r?rqsz3InstallRequirement.match_markers..T)rSany)raZextras_requestedr;rlr? match_markers s  z InstallRequirement.match_markerscCst|jdiS)zReturn whether any known-good hashes are specified as options. These activate --require-hashes mode; hashes specified as part of a URL do not. hashes)boolrYrzrlr;r;r?has_hash_optionss z#InstallRequirement.has_hash_optionsTcCsJ|jdi}|r|jn|j}|rB|jrB||jg|jt |S)aReturn a hash-comparer that considers my option- and URL-based hashes to be known-good. Hashes in URLs--ones embedded in the requirements file, not ones downloaded from an index server--are almost peers with ones from flags. They satisfy --require-hashes (whether it was implicitly or explicitly activated) but do not activate it. md5 and sha224 are not allowed in flags, which should nudge people toward good algos. We always OR all hashes together, even ones from URLs. :param trust_internet: Whether to trust URL-based (#md5=...) hashes downloaded from the internet, as by populate_link() r) rYrzcopyrNrOhash setdefaultZ hash_nameappendr)raZtrust_internetZ good_hashesrNr;r;r?r#s  zInstallRequirement.hashescCsR|jdkrdSt|j}|jrNt|jtjr4|j}n |j}|rN|d|7}|S)z@Format a nice indicator to show where this "comes from" Nz->)rCrcrDrArrerfrgr;r;r?rf9s    zInstallRequirement.from_pathcCs|dk s t|jdk r*|jjs"t|jjS|jdkrNtdd|_||_|jjS|jr`|j}n|j}t j |st d|t |t j||S)Nz req-buildkindzCreating directory %s)rBrVrHrCr)rWrKrrlowerrGexistsr{r|rrv)ra build_dirrrr;r;r?ensure_build_locationIs        z(InstallRequirement.ensure_build_locationcCs|jdk rdS|jdk st|js&t|jdk r8|jjsCannot update repository at %s; repository location is unknownfile+z bad url: %rr)rMrz+Unexpected version control type (in %s): %s)rNr{r|rFrKrBZschemerMrr,Z get_backendr"obtainZexport)rarZvc_typerMZ vcs_backendZ hidden_urlr;r;r?update_editables.     z"InstallRequirement.update_editablecCsB||std|jdS|jp&|j}t|}||||S)a Uninstall the distribution currently satisfying this requirement. Prompts before removing or modifying files unless ``auto_confirm`` is True. Refuses to delete or modify files outside of ``sys.prefix`` - thus uninstallation within a virtual environment can only modify that virtual environment, even if the virtualenv is linked to global site-packages. z#Skipping %s as it is not installed.N) rr{rrrrTrUrZ from_distremove)raZ auto_confirmverboserZdistZuninstalled_pathsetr;r;r? uninstalls    zInstallRequirement.uninstallcCsJ||tjjs"td||f|t|dd}|tjjd}|S)Nz$name %r doesn't start with prefix %rr/) startswithrGrHrrBrreplace)rarrrr;r;r?_clean_zip_names  z"InstallRequirement._clean_zip_namecCs(tj||}|||}|jd|S)Nr)rGrHrvrrr)rarH parentdirrootdirrrr;r;r?_get_archive_names z$InstallRequirement._get_archive_namec Cs|js td}d|j|jdf}tj||}tj|rtdt |d}|dkr\d}nj|dkrt d t |t |nF|d krt |}t d t |t |t||n|d krtd |sdStj|dtjdd}|tjtj|j}t|D]\} } } d| kr&| d| D]6} |j| | |d} t| d}d|_||dq*| D]>}|tkrxqf|j|| |d}tj| |}|||qfqW5QRXt dt |dS)z}Saves archive to provided build_dir. Used for saving downloaded VCS requirements as part of `pip download`. Tz %s-%s.ziprz8The file %s exists. (i)gnore, (w)ipe, (b)ackup, (a)bort )iwbarFrz Deleting %srzBacking up %s to %srN)Z allowZip64z pip-egg-info)rrrirzSaved %s) rFrBrrrrGrHrvrrrr{rrrrrrexitzipfileZZipFileZ ZIP_DEFLATEDnormcaserJrwalkrZZipInfoZ external_attrZwritestrrwriter)rarZcreate_archiveZ archive_nameZ archive_pathZresponseZ dest_fileZ zip_outputdirdirpathZdirnames filenamesdirnameZ dir_arcnameZzipdirfilenameZ file_arcnamer;r;r?archivesz        zInstallRequirement.archivec  sh|dk r |ng}|jr*|j|||ddS|jrnt|j} t| |j|j|j|||||dd|_ dSt ||j dg}t ||j dg}t dd} tj| jd} ||| ||} td |j} t*|j| | ||jd W5QRXW5QRXtj| s8td | W5QRdSd|_ fd d }t| \}|D],}tj|}|drX||}qqXtd|W5QRW5QRdSW5QRXg}t| L}|D]@}|}tj|r|tjj 7}|!tj"|||qW5QRX|#t$|tj|d}t|d}|%d|dW5QRXW5QRXdS)N)r)rrrrrrTrrrecordrzinstall-record.txtzRunning setup.py install for {})cmdrzRecord file %s not foundcs&dkstj|s|St|SdSr)rGrHisabsr)rHrr;r? prepend_rootsz0InstallRequirement.install..prepend_rootrz;Could not find .egg-info directory in install record for %szinstalled-files.txtr )&rKrrr Z wheel_versionrFZcheck_compatibilityrrrrXrrYrzr)rGrHrvget_install_argsr(rnrr[rrr{r|openrrrstripisdirrrrsortr r)rarrrrrrrrrZtemp_dirrecord_filename install_argsrrflineZ directoryZ egg_info_dirZ new_linesrZinst_files_pathr;rr?installEs           "   zInstallRequirement.installc Cst|j||jdd}|dd|g7}|dg7}|dk r@|d|g7}|dk rT|d|g7}|rd|dg7}n |d g7}trd t}|d tjt j d d ||j g7}|S)NT)rrZunbuffered_outputrz--recordz#--single-version-externally-managedz--rootz--prefixz --compilez --no-compilepythonz--install-headersZincludeZsite) r&rrZr+ sysconfigZget_python_versionrGrHrvrrrr)rarrrrrrZ py_ver_strr;r;r?rs0       z#InstallRequirement.get_install_args) NFNNNFNNFr;)N)T)NNNTFT)r;N)T)FFF)NNNNTFT)*rk __module__ __qualname____doc__rbrirmrxr~propertyrrrrrrrrrfrrrrrrrrrrrrrrrrrrrrrrrrr;r;r;r?r:Ms  W      8 +     %     B `r:)eZ __future__rrZloggingrGrrrrZdistutils.utilrZ pip._vendorrrZ"pip._vendor.packaging.requirementsrZpip._vendor.packaging.utilsrZpip._vendor.packaging.versionrr rZpip._vendor.pep517.wrappersr Z pip._internalr r Zpip._internal.build_envr Zpip._internal.exceptionsrZpip._internal.models.linkrZ*pip._internal.operations.generate_metadatarZpip._internal.pyprojectrrZpip._internal.req.req_uninstallrZpip._internal.utils.compatrZpip._internal.utils.hashesrZpip._internal.utils.loggingrZ pip._internal.utils.marker_filesrrZpip._internal.utils.miscrrrrrrrr r!r"r#r$Zpip._internal.utils.packagingr%Z$pip._internal.utils.setuptools_buildr&Zpip._internal.utils.subprocessr'r(Zpip._internal.utils.temp_dirr)Zpip._internal.utils.typingr*Zpip._internal.utils.virtualenvr+Zpip._internal.vcsr,typingr-r.r/r0r1r2r3r4Zpip._internal.cacher5Zpip._internal.indexr6Zpip._vendor.pkg_resourcesr7Z pip._vendor.packaging.specifiersr8Zpip._vendor.packaging.markersr9Z getLoggerrkr{objectr:r;r;r;r?sV               8      $       site-packages/pip/_internal/req/__pycache__/constructors.cpython-38.pyc000064400000024164151464452710022246 0ustar00U .e48@sdZddlZddlZddlZddlmZddlmZmZddl m Z ddl m Z m Z ddlmZddlmZmZdd lmZdd lmZdd lmZdd lmZdd lmZmZddlmZddl m!Z!ddl"m#Z#m$Z$ddl%m&Z&erddl'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-ddl.m/Z/dddgZ0e1e2Z3e j45Z6ddZ7ddZ8ddZ9ddZ:ddZ;Gd d!d!e<Z=d"d#Z>d/d%dZ?d&d'Z@d(d)ZAd*d+ZBd0d,dZCd1d-d.ZDdS)2a~Backing implementation for InstallRequirement's various constructors The idea here is that these formed a major chunk of InstallRequirement's size so, moving them and support code dedicated to them outside of that class helps creates for better understandability for the rest of the code. These are meant to be used elsewhere within pip to create instances of InstallRequirement. N)Marker)InvalidRequirement Requirement) Specifier)RequirementParseErrorparse_requirements)InstallationError)PyPITestPyPI)Link)make_pyproject_path)InstallRequirement)ARCHIVE_EXTENSIONS)is_installable_dirsplitext)MYPY_CHECK_RUNNING) path_to_url)is_urlvcs)Wheel)AnyDictOptionalSetTupleUnion) WheelCacheinstall_req_from_editableinstall_req_from_lineparse_editablecCs t|d}|tkrdSdS)z9Return True if `name` is a considered as an archive file.TF)rlowerr)nameZextr#B/usr/lib/python3.8/site-packages/pip/_internal/req/constructors.pyis_archive_file4sr%cCs6td|}d}|r*|d}|d}n|}||fS)Nz^(.+)(\[[^\]]+\])$r )rematchgroup)pathmextrasZpath_no_extrasr#r#r$ _strip_extras=s   r-cCs|s tStd|jS)N placeholder)setrr!r,)r,r#r#r$convert_extrasJsr0c Cs`|}t|\}}tj|rptjtj|dshdtj|}t|}tj |r`|d7}t |t |}| drt|j}|r||td| jfS||dfStD]&}| d|rd||f}qqd |krt d ||d d d  }t|s:d |dddtjDd} t | t|j}|sVt d|||dfS)aParses an editable requirement into: - a requirement name - an URL - extras - editable options Accepted requirements: svn+http://blahblah@rev#egg=Foobar[baz]&subdirectory=version_subdir .[some_extra] zsetup.pyzMFile "setup.py" not found. Directory cannot be installed in editable mode: {}zb (A "pyproject.toml" file was found, but editable mode currently requires a setup.py based build.)zfile:r.Nz%s:z%s+%s+z{} is not a valid editable requirement. It should either be a path to a local project or a VCS URL (beginning with svn+, git+, hg+, or bzr+).r rzFor --editable=%s only z, cSsg|]}|jdqS)z+URLr").0Zbackendr#r#r$ sz"parse_editable..z is currently supportedzZCould not detect requirement name for '%s', please specify one with #egg=your_package_name)r-osr*isdirexistsjoinformatabspathr isfilerrr! startswithr egg_fragmentrr,rsplitZ get_backendZbackends) editable_requrlZ url_no_extrasr,msgZpyproject_pathZ package_nameZversion_controlZvc_typeZ error_messager#r#r$rQsb          c Csd}tj|rd}zFt|d2}tt||dd|ddd7}W5QRXWqtk rtj d |d d YqXn |d |7}|S) zReturns helpful msg in case requirements file does not exist, or cannot be parsed. :params req: Requirements file path z It does exist.rz The argument you provided z(%s) appears to be az" requirements file. If that is thez# case, use the '-r' flag to installz" the packages specified within it.z2Cannot parse '%s' as requirements fileT)exc_infoz File '%s' does not exist.) r5r*r7opennextrreadrloggerdebug)reqrAfpr#r#r$deduce_helpful_msgs.   rLc@seZdZddZdS)RequirementPartscCs||_||_||_||_dSN) requirementlinkmarkersr,)selfrOrPrQr,r#r#r$__init__szRequirementParts.__init__N)__name__ __module__ __qualname__rSr#r#r#r$rMsrMcCs`t|\}}}|dk rFz t|}WqJtk rBtd|YqJXnd}t|}t||d|S)NInvalid requirement: '%s')rrrrr rM)r?r"r@Zextras_overriderJrPr#r#r$parse_req_from_editables rXFc CsLt|}|jjdkr|jjnd}t|j||d|j||||r>|ni||jd S)NfileT) source_dirZeditablerP constraint use_pep517isolatedoptions wheel_cacher,)rXrPschemeZ file_pathr rOr,) r? comes_fromr\r]r^r_r[partsrZr#r#r$rs  cCs>tjj|krdStjjdk r,tjj|kr,dS|dr:dSdS)akChecks whether the string "looks like" a path on the filesystem. This does not check whether the target actually exists, only judge from the appearance. Returns true if any of the following conditions is true: * a path separator is found (either os.path.sep or os.path.altsep); * a dot is found (which represents the current directory). TN.F)r5r*sepaltsepr<r2r#r#r$_looks_like_paths  rfcCst|r0tj|r0t|r$t|Std|t|slenrHZwarning)r*r"Z urlreq_partsr#r#r$_get_url_from_path s(   ricst|rd}nd}||krF||d\}}|}|s.with_sourcezIt looks like a path.=c3s|]}|kVqdSrNr#)r3op) req_as_stringr#r$ jsz&parse_req_from_line..z,= is not a valid operator. Did you mean == ?rBzInvalid requirement: {!r}z Hint: {}) rr>striprr5r*normpathr:r r-rir`r'searchr@rZis_wheelrfilenamer"versionr=r0rrrdrLany operatorsr9rrM)r"rlZ marker_sepZmarkers_as_stringrQr*rPZextras_as_stringpr@Zwheelr,rmrJZadd_msgrAr#)rlrpr$parse_req_from_line+sj         rzc Cs6t||}t|j||j|j|||r&|ni|||jd S)aCreates an InstallRequirement from a name, which might be a requirement, directory containing 'setup.py', filename, or URL. :param line_source: An optional string describing where the line is from, for logging purposes in case of an error. )rPrQr\r]r^r_r[r,)rzr rOrPrQr,) r"rar\r]r^r_r[rlrbr#r#r$rzs  cCszz t|}Wn tk r,td|YnXtjtjg}|jrh|rh|jrh|jj|krhtd|j |ft |||||dS)NrWzkPackages installed from PyPI cannot depend on packages which are not also hosted on PyPI. %s depends on %s )r]r_r\) rrrr Zfile_storage_domainr r@rPZnetlocr"r )Z req_stringrar]r_r\rJZdomains_not_allowedr#r#r$install_req_from_req_strings,  r{)NNFNNF)NNFNNFN)NFNN)E__doc__Zloggingr5r'Zpip._vendor.packaging.markersrZ"pip._vendor.packaging.requirementsrrZ pip._vendor.packaging.specifiersrZpip._vendor.pkg_resourcesrrZpip._internal.exceptionsrZpip._internal.models.indexr r Zpip._internal.models.linkr Zpip._internal.pyprojectr Zpip._internal.req.req_installr Zpip._internal.utils.filetypesrZpip._internal.utils.miscrrZpip._internal.utils.typingrZpip._internal.utils.urlsrZpip._internal.vcsrrZpip._internal.wheelrtypingrrrrrrZpip._internal.cacher__all__Z getLoggerrTrHZ _operatorskeysrxr%r-r0rrLobjectrMrXrrfrirzrr{r#r#r#r$sr                J "Q site-packages/pip/_internal/req/__pycache__/req_set.cpython-38.pyc000064400000013635151464452710021141 0ustar00U .e@sddlmZddlZddlmZddlmZddlmZddl m Z ddl m Z ddl mZdd lmZerdd lmZmZmZmZmZdd lmZeeZGd d d eZdS))absolute_importN) OrderedDict)canonicalize_name) pep425tags)InstallationError) indent_log)MYPY_CHECK_RUNNING)Wheel)DictIterableListOptionalTuple)InstallRequirementc@sXeZdZdddZddZddZd d Zd d ZdddZddZ ddZ ddZ d S)RequirementSetFTcCs*t|_||_||_g|_g|_g|_dS)z!Create a RequirementSet. N)r requirementsrequire_hashescheck_supported_wheelsunnamed_requirementsZsuccessfully_downloadedreqs_to_cleanup)selfrrr=/usr/lib/python3.8/site-packages/pip/_internal/req/req_set.py__init__s zRequirementSet.__init__cCs4tdd|jDddd}ddd|DS)Ncss|]}|js|VqdSN)Z comes_from.0reqrrr +sz)RequirementSet.__str__..cSs t|jSrrnamerrrr,z(RequirementSet.__str__..key css|]}t|jVqdSrstrrrrrrr.s)sortedrvaluesjoin)rrrrr__str__(s zRequirementSet.__str__cCsBt|jddd}d}|j|jjt|ddd|DdS) NcSs t|jSrrr!rrrr"4r#z)RequirementSet.__repr__..r$z4<{classname} object; {count} requirement(s): {reqs}>z, css|]}t|jVqdSrr'rrrrr;sz*RequirementSet.__repr__..)Z classnamecountZreqs)r)rr*format __class____name__lenr+)rr format_stringrrr__repr__0szRequirementSet.__repr__cCs|jr t|j|dSr)r AssertionErrorrappend)r install_reqrrradd_unnamed_requirement>s z&RequirementSet.add_unnamed_requirementcCs"|js tt|j}||j|<dSr)r r4rr)rr6 project_namerrradd_named_requirementCs  z$RequirementSet.add_named_requirementNc Cs||s$td|j|jgdfS|jrd|jjrdt|jj}t }|j rd| |sdt d|j|j|dkksztd|js|||gdfSz||j}Wntk rd}YnX|dko|o|j o|j|jko|jj|jjk}|rt d|||jf|s"|||g|fS|js2|js:g|fS|joZ|joX|jj|jjk }|r||j|t d|jd|_ttt|jt|jB|_td||j|g|fS) a&Add install_req as a requirement to install. :param parent_req_name: The name of the requirement that needed this added. The name is used because when multiple unnamed requirements resolve to the same name, we could otherwise end up with dependency links that point outside the Requirements set. parent_req must already be added. Note that None implies that this is a user supplied requirement, vs an inferred one. :param extras_requested: an iterable of extras used to evaluate the environment markers. :return: Additional requirements to scan. That is either [] if the requirement is not applicable, or [install_req] if the requirement is applicable and has just been added. z6Ignoring %s: markers '%s' don't match your environmentNz-%s is not a supported wheel on this platform.zTa direct req shouldn't have a parent and also, a non direct req should have a parentz5Double requirement given: %s (already in %s, name=%r)zhCould not satisfy constraints for '%s': installation from path or url cannot be constrained to a versionFzSetting %s extras to: %s)Z match_markersloggerinfor ZmarkerslinkZis_wheelr filenamerZ get_supportedrZ supportedrZ is_directr4r7get_requirementKeyError constraintZextrasrZ specifierr9pathrr5tupler)setdebug) rr6Zparent_req_nameZextras_requestedZwheelZtagsZ existing_reqZhas_conflicting_requirementZdoes_not_satisfy_constraintrrradd_requirementJs          zRequirementSet.add_requirementcCs t|}||jko|j|j Sr)rrr@rr r8rrrhas_requirements  zRequirementSet.has_requirementcCs,t|}||jkr|j|Std|dS)NzNo project with the name %r)rrr?rFrrrr>s  zRequirementSet.get_requirementc Cs4tdt|jD] }|qW5QRXdS)zClean up files, remove builds.zCleaning up...N)r:rDrrZremove_temporary_source)rrrrr cleanup_filess  zRequirementSet.cleanup_files)FT)NN) r0 __module__ __qualname__rr,r3r7r9rErGr>rHrrrrrs   p  r)Z __future__rZlogging collectionsrZpip._vendor.packaging.utilsrZ pip._internalrZpip._internal.exceptionsrZpip._internal.utils.loggingrZpip._internal.utils.typingrZpip._internal.wheelr typingr r r r rZpip._internal.req.req_installrZ getLoggerr0r:objectrrrrrs          site-packages/pip/_internal/req/__pycache__/req_install.cpython-38.opt-1.pyc000064400000055511151464452710022752 0ustar00U .e@sLddlmZddlZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z m Z ddlmZddlmZddlmZddlmZdd lmZdd lmZmZdd lmZdd lmZdd lm Z ddl!m"Z"ddl#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/m0Z0ddl1m2Z2m3Z3m4Z4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;mm?Z?ddl@mAZAddlBmCZCmDZDddlEmFZFddlGmHZHddlImJZJddlKmLZLeHr.ddlMmNZNmOZOmPZPmQZQmRZRmSZSmTZTddlmUZUddlVmWZWdd lXmYZYdd!lZm[Z[dd"l\m]Z]dd#l^m_Z_e`eaZbGd$d%d%ecZddS)&)absolute_importN) change_root) pkg_resourcessix) Requirement)canonicalize_name)Version)parse)Pep517HookCaller) pep425tagswheel)NoOpBuildEnvironment)InstallationError)Link)get_metadata_generator)load_pyproject_tomlmake_pyproject_path)UninstallPathSet) native_str)Hashes) indent_log)PIP_DELETE_MARKER_FILENAMEhas_delete_marker_file) _make_build_dirask_path_exists backup_dir display_pathdist_in_install_pathdist_in_site_packagesdist_in_usersite ensure_dirget_installed_versionhide_urlredact_auth_from_urlrmtree) get_metadata)make_setuptools_shim_args)call_subprocessrunner_with_spinner_message) TempDirectory)MYPY_CHECK_RUNNING)running_under_virtualenv)vcs)AnyDictIterableListOptionalSequenceUnion)BuildEnvironment) WheelCache) PackageFinder) Distribution) SpecifierSet)Markerc @speZdZdZdPddZddZd d Zd d Zd dZe ddZ e ddZ e ddZ e ddZ dQddZe ddZdRddZddZd d!Zd"d#Zd$d%Zd&d'Ze d(d)ZdSd*d+Ze d,d-Ze d.d/Ze d0d1Zd2d3Zd4d5Zd6d7Ze d8d9Zd:d;Zdd?Z!dTd@dAZ"dUdBdCZ#dVdDdEZ$dFdGZ%dHdIZ&dJdKZ'dWdLdMZ(dNdOZ)dS)XInstallRequirementz Represents something that may be installed later on, may have information about where to fetch the relevant requirement and also contains logic for installing the said requirement. NFc Cs||_||_| |_|dkr"d|_ntjtj||_||_| |_ |dkr`|r`|j r`t |j }||_ |_ | rx| |_n |rdd|jD|_nt|_|dkr|r|j}||_d|_d|_d|_d|_d|_| r| ni|_d|_d|_||_t|_d|_d|_g|_d|_ ||_!dS)NcSsh|]}t|qSr;)rZ safe_extra.0extrar;r;A/usr/lib/python3.8/site-packages/pip/_internal/req/req_install.py wsz.InstallRequirement.__init__..F)"req comes_from constraint source_dirospathnormpathabspatheditable _wheel_cacheurlrlink original_linkextrassetmarkermarkers satisfied_byconflicts_with_temp_build_dir_ideal_build_dirinstall_succeededoptionsZpreparedZ is_directisolatedr build_envmetadata_directorypyproject_requiresrequirements_to_checkpep517_backend use_pep517) selfrArBrDrIrLrQr^rXrWZ wheel_cacherCrNr;r;r?__init__TsH    zInstallRequirement.__init__cCs|jr,t|j}|jrD|dt|jj7}n|jr@t|jj}nd}|jdk rb|dt|jj7}|jrt |jt j r~|j}n |j }|r|d|7}|S)Nz from %szz in %sz (from %s)) rAstrrLr#rKrRrlocationrB isinstancer string_types from_pathr_srBr;r;r?__str__s     zInstallRequirement.__str__cCsd|jjt||jfS)Nz<%s object: %s editable=%r>) __class____name__rarIr_r;r;r?__repr__s zInstallRequirement.__repr__cs>t|t}fddt|D}dj|jjd|dS)z>An un-tested helper for getting state, for debugging. c3s|]}d||VqdS)z{}={!r}N)format)r=attrZ attributesr;r? sz2InstallRequirement.format_debug..z<{name} object: {{{state}}}>z, )namestate)varssortedrmrirjjoin)r_namesrrr;ror? format_debugs zInstallRequirement.format_debugcCsh|jdkr||||_|jdk rd|sd|j}t}|jj|j|j|d|_||jkrdtd|jdS)aEnsure that if a link can be found for this, that it is found. Note that self.link may still be None - if Upgrade is False and the requirement is already installed. If require_hashes is True, don't use the wheel cache, because cached wheels, always built locally, have different hashes than the files downloaded from the index server and thus throw false hash mismatches. Furthermore, cached wheels at present have undeterministic contents due to file modification times. N)rLZ package_namesupported_tagszUsing cached wheel link: %s) rLZfind_requirementrJr Z get_supportedgetrqloggerdebug)r_finderZupgradeZrequire_hashesZold_linkrxr;r;r? populate_links  z InstallRequirement.populate_linkcCs |jdkrdStt|jjSN)rArrZ safe_namerqrkr;r;r?rqs zInstallRequirement.namecCs|jjSr~)rA specifierrkr;r;r?rszInstallRequirement.specifiercCs$|j}t|dko"tt|jdkS)zReturn whether I am pinned to an exact version. For example, some-package==1.2 is pinned; some-package>1.2 is not. >=====)rlennextiteroperator)r_Z specifiersr;r;r? is_pinneds zInstallRequirement.is_pinnedcCs t|jSr~)r!rqrkr;r;r?installed_versionsz$InstallRequirement.installed_versioncs0|sd}jdk r(tfdd|DSdSdS)N)c3s|]}jd|iVqdS)r>N)rQZevaluater<rkr;r?rpsz3InstallRequirement.match_markers..T)rQany)r_Zextras_requestedr;rkr? match_markers s  z InstallRequirement.match_markerscCst|jdiS)zReturn whether any known-good hashes are specified as options. These activate --require-hashes mode; hashes specified as part of a URL do not. hashes)boolrWryrkr;r;r?has_hash_optionss z#InstallRequirement.has_hash_optionsTcCsJ|jdi}|r|jn|j}|rB|jrB||jg|jt |S)aReturn a hash-comparer that considers my option- and URL-based hashes to be known-good. Hashes in URLs--ones embedded in the requirements file, not ones downloaded from an index server--are almost peers with ones from flags. They satisfy --require-hashes (whether it was implicitly or explicitly activated) but do not activate it. md5 and sha224 are not allowed in flags, which should nudge people toward good algos. We always OR all hashes together, even ones from URLs. :param trust_internet: Whether to trust URL-based (#md5=...) hashes downloaded from the internet, as by populate_link() r) rWrycopyrLrMhash setdefaultZ hash_nameappendr)r_Ztrust_internetZ good_hashesrLr;r;r?r#s  zInstallRequirement.hashescCsR|jdkrdSt|j}|jrNt|jtjr4|j}n |j}|rN|d|7}|S)z@Format a nice indicator to show where this "comes from" Nz->)rArarBrcrrdrerfr;r;r?re9s    zInstallRequirement.from_pathcCs||jdk r|jjS|jdkr6tdd|_||_|jjS|jrH|j}n|j}tj |snt d|t |tj ||S)Nz req-buildkindzCreating directory %s)rTrFrAr)rUrIrqlowerrEexistsrzr{rru)r_ build_dirrqr;r;r?ensure_build_locationIs      z(InstallRequirement.ensure_build_locationcCs|jdk rdS|j}d|_||j}tj|rBtdt|t d|t|jt|t |j|tj tj||_t|dd|_|jr|j}tjj||jd}tj||}tj tj|}||_d|_dS)aMove self._temp_build_dir to "self._ideal_build_dir/self.req.name" For some requirements (e.g. a path to a directory), the name of the package is not available until we run egg_info, so the build_location will return a temporary directory and store the _ideal_build_dir. This is only called to "fix" the build directory after generating metadata. NzCannot update repository at %s; repository location is unknownfile+r)rK) rLrzr{rDZschemerKrr,Z get_backendr"obtainZexport)r_rZvc_typerKZ vcs_backendZ hidden_urlr;r;r?update_editables    z"InstallRequirement.update_editablecCsB||std|jdS|jp&|j}t|}||||S)a Uninstall the distribution currently satisfying this requirement. Prompts before removing or modifying files unless ``auto_confirm`` is True. Refuses to delete or modify files outside of ``sys.prefix`` - thus uninstallation within a virtual environment can only modify that virtual environment, even if the virtualenv is linked to global site-packages. z#Skipping %s as it is not installed.N) rrzrrqrRrSrZ from_distremove)r_Z auto_confirmverboserZdistZuninstalled_pathsetr;r;r? uninstalls    zInstallRequirement.uninstallcCs(|t|dd}|tjjd}|S)Nr/)rreplacerErFr)r_rqrr;r;r?_clean_zip_namesz"InstallRequirement._clean_zip_namecCs(tj||}|||}|jd|S)Nr)rErFrurrq)r_rF parentdirrootdirrqr;r;r?_get_archive_names z$InstallRequirement._get_archive_namec Csd}d|j|jdf}tj||}tj|rtdt|d}|dkrRd}nj|dkrvt d t|t |nF|d krt |}t d t|t|t ||n|d krtd |sdStj|dtjdd}|tjtj|j}t|D]\} } } d| kr| d| D]6} |j| | |d} t| d}d|_||dq | D]>}|tkrnq\|j|| |d}tj| |}|||q\qW5QRXtdt|dS)z}Saves archive to provided build_dir. Used for saving downloaded VCS requirements as part of `pip download`. Tz %s-%s.ziprz8The file %s exists. (i)gnore, (w)ipe, (b)ackup, (a)bort )iwbarFrz Deleting %srzBacking up %s to %srN)Z allowZip64z pip-egg-info)rrrirzSaved %s)rqrrErFrurrrrzrrrrrrexitzipfileZZipFileZ ZIP_DEFLATEDnormcaserHrwalkrZZipInfoZ external_attrZwritestrrwriter)r_rZcreate_archiveZ archive_nameZ archive_pathZresponseZ dest_fileZ zip_outputdirdirpathZdirnames filenamesdirnameZ dir_arcnameZzipdirfilenameZ file_arcnamer;r;r?archivesx       zInstallRequirement.archivec  sh|dk r |ng}|jr*|j|||ddS|jrnt|j} t| |j|j|j|||||dd|_ dSt ||j dg}t ||j dg}t dd} tj| jd} ||| ||} td |j} t*|j| | ||jd W5QRXW5QRXtj| s8td | W5QRdSd|_ fd d }t| \}|D],}tj|}|drX||}qqXtd|W5QRW5QRdSW5QRXg}t| L}|D]@}|}tj|r|tjj 7}|!tj"|||qW5QRX|#t$|tj|d}t|d}|%d|dW5QRXW5QRXdS)N)r)rrrrrrTrrrecordrzinstall-record.txtzRunning setup.py install for {})cmdrzRecord file %s not foundcs&dkstj|s|St|SdSr~)rErFisabsr)rFrr;r? prepend_rootsz0InstallRequirement.install..prepend_rootrz;Could not find .egg-info directory in install record for %szinstalled-files.txtr )&rIrrr Z wheel_versionrDZcheck_compatibilityrqrrVrrWryr)rErFruget_install_argsr(rmrrYrrrzr{openrrrstripisdirrrrsortr r)r_rrrrrrrrrZtemp_dirrecord_filename install_argsrrflineZ directoryZ egg_info_dirZ new_linesrZinst_files_pathr;rr?installEs           "   zInstallRequirement.installc Cst|j||jdd}|dd|g7}|dg7}|dk r@|d|g7}|dk rT|d|g7}|rd|dg7}n |d g7}trd t}|d tjt j d d ||j g7}|S)NT)rrZunbuffered_outputrz--recordz#--single-version-externally-managedz--rootz--prefixz --compilez --no-compilepythonz--install-headersZincludeZsite) r&rrXr+ sysconfigZget_python_versionrErFrurrrq)r_rrrrrrZ py_ver_strr;r;r?rs0       z#InstallRequirement.get_install_args) NFNNNFNNFr;)N)T)NNNTFT)r;N)T)FFF)NNNNTFT)*rj __module__ __qualname____doc__r`rhrlrwr}propertyrqrrrrrrrerrrrrrrrrrrrrrrrrrrrrrrrr;r;r;r?r:Ms  W      8 +     %     B `r:)eZ __future__rrZloggingrErrrrZdistutils.utilrZ pip._vendorrrZ"pip._vendor.packaging.requirementsrZpip._vendor.packaging.utilsrZpip._vendor.packaging.versionrr rZpip._vendor.pep517.wrappersr Z pip._internalr r Zpip._internal.build_envr Zpip._internal.exceptionsrZpip._internal.models.linkrZ*pip._internal.operations.generate_metadatarZpip._internal.pyprojectrrZpip._internal.req.req_uninstallrZpip._internal.utils.compatrZpip._internal.utils.hashesrZpip._internal.utils.loggingrZ pip._internal.utils.marker_filesrrZpip._internal.utils.miscrrrrrrrr r!r"r#r$Zpip._internal.utils.packagingr%Z$pip._internal.utils.setuptools_buildr&Zpip._internal.utils.subprocessr'r(Zpip._internal.utils.temp_dirr)Zpip._internal.utils.typingr*Zpip._internal.utils.virtualenvr+Zpip._internal.vcsr,typingr-r.r/r0r1r2r3r4Zpip._internal.cacher5Zpip._internal.indexr6Zpip._vendor.pkg_resourcesr7Z pip._vendor.packaging.specifiersr8Zpip._vendor.packaging.markersr9Z getLoggerrjrzobjectr:r;r;r;r?sV               8      $       site-packages/pip/_internal/req/__pycache__/req_uninstall.cpython-38.pyc000064400000042030151464452710022346 0ustar00U .e&\@s|ddlmZddlZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl m Z mZddlmZmZmZddlmZddlmZmZmZmZmZmZmZmZmZdd lm Z m!Z!dd l"m#Z#e#r dd l$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-dd l.m/Z/e0e1Z2d dZ3ddZ4e4ddZ5ddZ6ddZ7ddZ8Gddde9Z:Gddde9Z;Gddde9Z.unique) functoolswraps)r3r4r(r2r)_unique=sr7ccstt|d}|D]t}tj|j|d}|V|drtj |\}}|dd}tj||d}|Vtj||d}|VqdS)a Yield all the uninstallation paths for dist based on RECORD-without-.py[co] Yield paths to all the files in RECORD. For each .py file in RECORD, add the .pyc and .pyo in the same directory. UninstallPathSet.add() takes care of the __pycache__ .py[co]. ZRECORDr.pyN.pyc.pyo) csvreaderr Zget_metadata_linesr!r"r#locationendswithsplit)r%rrowr"Zdnr3baser(r(r)uninstallation_pathsJs   rDcsJtjjt}t|tdD]*tfdd|D}|s|q|S)zCompact a path set to contain the minimal number of paths necessary to contain all paths in the set. If /a/path/ and /a/path/to/a/file.txt are both in the set, leave only the shorter path.keyc3s:|]2}|do0t|dkVqdS)*N) startswithrstriplen).0Z shortpathr"sepr(r) lszcompact..)r!r"rMr,sortedrJanyr-)pathsZ short_pathsZ should_skipr(rLr)compactbs rRc stdd|D}t|}ttdd|Dtd}t}dd|D]tfdd|DrfqJt}t}tD]B\}}|fdd|D|fd d|Dq|||sJ| || tj qJtt |j ||BS) zReturns a set containing the paths that need to be renamed. This set may include directories when the original sequence of paths included every file on disk. css|]}tj||fVqdSr+)r!r"normcaserKpr(r(r)rN}sz&compress_for_rename..css|]}tj|dVqdS)rN)r!r"r@rTr(r(r)rNsrEcWstjtjj|Sr+)r!r"rSr#)ar(r(r) norm_joinsz&compress_for_rename..norm_joinc3s |]}tj|VqdSr+)r!r"rSrH)rKw)rootr(r)rNsc3s|]}|VqdSr+r()rKddirnamerWrYr(r)rNsc3s|]}|VqdSr+r()rKfr[r(r)rNs)dictr,rOvaluesrJrPr!walkupdatedifference_updater-rMmap __getitem__) rQZcase_mapZ remainingZ uncheckedZ wildcardsZ all_filesZ all_subdirsZsubdirsfilesr(r[r)compress_for_renamevs6    rfc Cst|}t}t}t}|D]>}|dr.q|ds@d|krR|tj|||qtttjj|}t|}|D]d}t |D]T\}} } | D]D} | drqtj || } tj | rtj| |kr|| qqq||dd|DB}||fS)asReturns a tuple of 2 sets of which paths to display to user The first set contains paths that would be deleted. Files of a package are not added and the top-level directory of the package has a '*' added at the end - to signify that all it's contents are removed. The second set contains files that would have been skipped in the above folders. r:z __init__.py .dist-infocSsh|]}tj|dqS)rG)r!r"r#)rKfolderr(r(r) sz.compress_for_output_listing..) r,r?r-r!r"r\rcrSrRr`r#isfile) rQ will_remove will_skipZfoldersrer"Z_normcased_filesrhdirpath_ZdirfilesZfnameZfile_r(r(r)compress_for_output_listings4     roc@sLeZdZdZddZddZddZdd Zd d Zd d Z e ddZ dS)StashedUninstallPathSetzWA set of file rename operations to stash files while tentatively uninstalling them.cCsi|_g|_dSr+) _save_dirs_movesselfr(r(r)__init__sz StashedUninstallPathSet.__init__cCsDz t|}Wntk r*tdd}YnX||jtj|<|jS)zStashes a directory. Directories are stashed adjacent to their original location if possible, or else moved/copied into the user's temp dir. uninstallZkind)rOSErrorrrqr!r"rS)rtr"save_dirr(r(r)_get_directory_stashs  z,StashedUninstallPathSet._get_directory_stashcCstj|}tj|d}}d}||krfz|j|}WqWntk rPYnXtj||}}q"tj|}tdd}||j|<tj||}|r|tjjkrtj |j|S|jS)zStashes a file. If no root has been provided, one will be created for the directory in the user's temp directory.Nrvrw) r!r"rSr\rqKeyErrorrrelpathcurdirr#)rtr"headZold_headryr|r(r(r)_get_file_stashs"     z'StashedUninstallPathSet._get_file_stashcCsltj|otj| }|r*||}n ||}|j||f|r^tj|r^t|t |||S)zStashes the directory or file and returns its new location. Handle symlinks as files to avoid modifying the symlink targets. ) r!r"isdirislinkrzrrrr$rmdirr)rtr"Z path_is_dirnew_pathr(r(r)stashs    zStashedUninstallPathSet.stashcCs,|jD]\}}|q g|_i|_dS)z0Commits the uninstall by removing stashed files.N)rqitemsZcleanuprr)rtrnryr(r(r)commits zStashedUninstallPathSet.commitc Cs|jD]}tjd|q|jD]\}}zTtd||tj|sPtj|r\t |ntj |rpt |t ||Wq t k r}ztd|td|W5d}~XYq Xq |dS)z2Undoes the uninstall by moving stashed files back.Moving to %s from %szReplacing %s from %szFailed to restore %sz Exception: %sN)r)rrlogginginfologgerdebugr!r"rjrunlinkrrrrxerrorr)rtrUrr"Zexr(r(r)rollback&s     z StashedUninstallPathSet.rollbackcCs t|jSr+)boolrrrsr(r(r) can_rollback:sz$StashedUninstallPathSet.can_rollbackN) __name__ __module__ __qualname____doc__rurzrrrrpropertyrr(r(r(r)rps rpc@s^eZdZdZddZddZddZdd Zdd d Zd dZ ddZ ddZ e ddZ dS)UninstallPathSetzMA set of file paths to be removed in the uninstallation of a requirement.cCs(t|_t|_i|_||_t|_dSr+)r,rQ_refusepthr%rp _moved_paths)rtr%r(r(r)ruCs zUninstallPathSet.__init__cCst|S)zs Return True if the given path is one we are permitted to remove/modify, False otherwise. )r)rtr"r(r(r) _permittedKszUninstallPathSet._permittedcCstj|\}}tjt|tj|}tj|s:dS||rR|j |n |j |tj |ddkrt r| t |dS)Nr8)r!r"r@r#rrSexistsrrQr-rsplitextr r)rtr"r~tailr(r(r)r-Ts   zUninstallPathSet.addcCsLt|}||r<||jkr*t||j|<|j||n |j|dSr+)rrrUninstallPthEntriesr-r)rtpth_fileentryr(r(r)add_pthhs   zUninstallPathSet.add_pthFc Cs|jstd|jjdS|jjd|jj}td|tp|sP||r|j}t |j}t t |D]}| |t d|ql|jD] }|qtd|W5QRXdS)z[Remove paths in ``self.paths`` with confirmation (unless ``auto_confirm`` is True).z7Can't uninstall '%s'. No files were found to uninstall.N-zUninstalling %s:zRemoving file or directory %szSuccessfully uninstalled %s)rQrrr% project_nameversionr _allowed_to_proceedrrfrOrRrrrr_remove)rtZ auto_confirmverboseZdist_name_versionZmovedZ for_renamer"rr(r(r)rrs&    zUninstallPathSet.removecCsndd}|st|j\}}nt|j}t}|d||d||d|j|r`|dt|jtddd kS) zIDisplay which files would be deleted and prompt for confirmation c SsD|sdSt|t"tt|D]}t|q&W5QRXdSr+)rrr rOrR)msgrQr"r(r(r)_displays  z6UninstallPathSet._allowed_to_proceed.._displayz Would remove:z+Would not remove (might be manually added):z%Would not remove (outside of prefix):zWill actually move:zProceed (y/n)? )ynr)rorQr,rrfr )rtrrrkrlr(r(r)rs     z$UninstallPathSet._allowed_to_proceedcCsR|jjstd|jjdStd|jj|j|j D] }|q@dS)z1Rollback the changes previously made by remove().z'Can't roll back %s; was not uninstalledNzRolling back uninstall of %s) rrrrr%rrrrr_)rtrr(r(r)rs zUninstallPathSet.rollbackcCs|jdS)z?Remove temporary save dir: rollback will no longer be possible.N)rrrsr(r(r)rszUninstallPathSet.commitc st|j}t|s.td|j|tj||S|ddt dt dhDkrhtd|j|||S||}t |}d t |j}|jotj|j}t|jdd}|r|jd r|j|s||j|d r|d D]&}tjtj|j|} || qn|d r|d rB|d ngfd d|d DD]J} tj|j| } || || d|| d|| dqbn2|rtd |jn|jdr"||jtj|jd} tjtj|jd} || d| n|rR|jdrRt |D]} || q>n|rt!|d} tj"| #$}W5QRX||jkst%d||j|jf||tjtj|d} || |jnt&d||j|drT|'drT|(dD]L}t)|rt*}nt+}|tj||t,r|tj||dqg}|j-dd}|.D]}|/t0||dql|j-d d}|.D]}|/t0||d!q|D]}||q|S)"Nz1Not uninstalling %s at %s, outside environment %scSsh|] }|r|qSr(r(rTr(r(r)risz-UninstallPathSet.from_dist..ZstdlibZ platstdlibzsz.UninstallPathSet.from_dist..r8r:r;zCannot uninstall {!r}. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall.z.eggrzeasy-install.pthz./rgrAz;Egg-link %s does not match installed location of %s (at %s)z)Not sure how to uninstall: %s - Check: %sZscriptsz.batconsole_scripts)groupF gui_scriptsT)1rr>rrrrFsysprefix sysconfigZget_pathrformatrZ to_filenamerZegg_infor!r"rgetattrZ _providerr?r-Z has_metadataZ get_metadata splitlinesnormpathr#rr@r\rrDopenrSreadlinestripAssertionErrorrZmetadata_isdirZmetadata_listdirr rrrZ get_entry_mapkeysextendr*)clsr%Z dist_pathr'Zdevelop_egg_linkZdevelop_egg_link_egg_infoZegg_info_existsZdistutils_egg_infoZinstalled_filer"Z top_level_pkgZeasy_install_eggZeasy_install_pthfhZ link_pointerZscriptr&Z_scripts_to_removernamersr(rr) from_dists                       zUninstallPathSet.from_distN)FF)rrrrrurr-rrrrr classmethodrr(r(r(r)r@s  rc@s,eZdZddZddZddZddZd S) rcCs0tj|std|||_t|_d|_dS)Nz.Cannot remove entries from nonexistent file %s)r!r"rjrfiler,entries _saved_lines)rtrr(r(r)ruIs zUninstallPthEntries.__init__cCs<tj|}tr,tj|ds,|dd}|j|dS)Nr\/)r!r"rSr splitdrivereplacerr-)rtrr(r(r)r-Ss  zUninstallPthEntries.addc Cstd|jt|jd}|}||_W5QRXtdd|DrLd}nd}|r~|d|ds~|d|d|d<|j D]>}z$td || ||dWqt k rYqXqt|jd }| |W5QRXdS) NzRemoving pth entries from %s:rbcss|]}d|kVqdS)s Nr()rKliner(r(r)rNjsz-UninstallPthEntries.remove..z  zutf-8zRemoving entry: %swb) rrrr readlinesrrPr?encoderr ValueError writelines)rtrlinesZendlinerr(r(r)rcs"  zUninstallPthEntries.removec CsR|jdkrtd|jdStd|jt|jd}||jW5QRXdS)Nz.Cannot roll back changes to %s, none were madeFz!Rolling %s back to previous staterT)rrrrrrr)rtrr(r(r)rzs zUninstallPthEntries.rollbackN)rrrrur-rrr(r(r(r)rHs r)=Z __future__rr<r5rr!rrZ pip._vendorrZpip._internal.exceptionsrZpip._internal.locationsrrZpip._internal.utils.compatrrr Zpip._internal.utils.loggingr Zpip._internal.utils.miscr r r rrrrrrZpip._internal.utils.temp_dirrrZpip._internal.utils.typingrtypingrrrrrrrrrZpip._vendor.pkg_resourcesr Z getLoggerrrr*r7rDrRrfroobjectrprrr(r(r(r)s<    ,  ,    (3o site-packages/pip/_internal/req/__pycache__/req_file.cpython-38.pyc000064400000022300151464452710021252 0ustar00U .e7 @sdZddlmZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl mZddlmZddlmZdd lmZdd lmZmZdd lmZer dd lmZmZmZmZmZm Z m!Z!m"Z"dd l#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*ee"e+e!fZ,dgZ-e.dej/Z0e.dZ1e.dZ2ej3ej4ej5ej6ej7ej8ej9ej:ej;ejej?g Z@ejAejBejCgZDddeDDZEd)ddZFddZGd*ddZHddZIdd ZJd!d"ZKd#d$ZLd%d&ZMd'd(ZNdS)+z Requirements file parsing )absolute_importN) filterfalse)parse) cmdoptions)get_file_contentRequirementsFileParseError) SearchScope)install_req_from_editableinstall_req_from_line)MYPY_CHECK_RUNNING)AnyCallableIteratorListNoReturnOptionalTextTuple)InstallRequirement) WheelCache) PackageFinder) PipSessionparse_requirementsz^(http|https|file):z (^|\s+)#.*$z#(?P\$\{(?P[A-Z0-9_]+)\})cCsg|]}t|jqS)strdest).0orr>/usr/lib/python3.8/site-packages/pip/_internal/req/req_file.py Ksr Fc csj|dkrtdt|||d\}} t| |} | D]4\} } t| || |||||||d } | D] }|VqXq0dS)a1Parse a requirements file and yield InstallRequirement instances. :param filename: Path or url of requirements file. :param finder: Instance of pip.index.PackageFinder. :param comes_from: Origin description of requirements. :param options: cli options. :param session: Instance of pip.download.PipSession. :param constraint: If true, parsing a constraint file rather than requirements file. :param wheel_cache: Instance of pip.wheel.WheelCache :param use_pep517: Value of the --use-pep517 option. NzCparse_requirements() missing 1 required keyword argument: 'session') comes_fromsession) use_pep517 constraint) TypeErrorr preprocess process_line)filenamefinderr!optionsr"r$ wheel_cacher#_content lines_enum line_numberlineZreq_iterreqrrrrNs*    cCs6t|dd}t|}t|}t||}t|}|S)zSplit, filter, and join lines, and return a line iterator :param content: the content of the requirements file :param options: cli options )start) enumerate splitlines join_linesignore_comments skip_regexexpand_env_variables)r-r*r.rrrr&ys  r&c ! cst|} | } d| _|r"|j| _t|\} } tjdkrB| d} | t | | \}}d| rbdnd||f}| r|r||j nd}|rt ||i}tD]&}||jkr|j|r|j|||<qd||}t| |||||| |d Vn|jr |r|j nd}t|jd ||| ||d Vn|js0|jr|jrH|jd }d}n|jd }d }t|rpt||}n"t|stjtj||}t|||||||d }|D] }|Vqn|jr|j|_n|r|j }|j!}|jr|jg}|j"d krg}|j#r|$|j#|j rd|j d }tjtj%|}tj||}tj&|rZ|}|'|t(||d}||_)|j*r|+|j,pgD] }d||} |j-|| dqdS)a#Process a single requirements line; This can result in creating/yielding requirements, or updating the finder. For lines that contain requirements, the only options that have an effect are from SUPPORTED_OPTIONS_REQ, and they are scoped to the requirement. Other options from SUPPORTED_OPTIONS may be present, but are ignored. For lines that do not contain requirements, the only options that have an effect are from SUPPORTED_OPTIONS. Options from SUPPORTED_OPTIONS_REQ may be present, but are ignored. These lines may contain multiple options (although our docs imply only one is supported), and all our parsed and affect the finder. :param constraint: If True, parsing a constraints file. :param options: OptionParser options that we may update N)utf8z%s %s (line %s)z-cz-rFz line {} of {})r!r#isolatedr*r+r$ line_sourcer)r!r#r$r>r+T)r$r+) find_links index_urls)source). build_parserZget_default_values index_urlZformat_controlbreak_args_optionssys version_infoencode parse_argsshlexsplitZ isolated_moderZcheck_install_build_globalSUPPORTED_OPTIONS_REQ_DEST__dict__formatr Z editablesr requirements constraints SCHEME_REsearch urllib_parseZurljoinospathjoindirnamerrequire_hashesr@rAno_indexZextra_index_urlsextendabspathexistsappendr search_scopepreZset_allow_all_prereleasesZ trusted_hostsZadd_trusted_host)!r0r(r/r)r!r*r"r+r#r$parserdefaultsZargs_strZ options_strZoptsr,Zline_comes_fromr>Z req_optionsrr?Zreq_pathZnested_constraintZ parsed_reqsr1r@rAvalueZreq_dirZrelative_to_reqs_filer^ZhostrBrrrr's                   r'cCsf|d}g}|dd}|D]2}|ds6|dr<qRq|||dqd|d|fS)zBreak up the line into an args and options string. We only want to shlex (and then optparse) the options, not the args. args can contain markers which are corrupted by shlex.  N-z--r)rK startswithr]poprV)r0tokensargsr*tokenrrrrEs    rEcsDtjdd}tt}|D]}|}||qfdd}||_|S)z7 Return a parser for parsing requirement lines F)Zadd_help_optioncsd|f}t|dS)NzInvalid requirement: %s %sr)selfmsgr0rr parser_exit6s z!build_parser..parser_exit)optparseZ OptionParserSUPPORTED_OPTIONSSUPPORTED_OPTIONS_REQZ add_optionexit)r0r`Zoption_factoriesZoption_factoryZoptionrmrrlrrC(s   rCccsd}g}|D]v\}}|dr(t|rjt|r:d|}|r^|||d|fVg}q||fVq |sr|}||dq |r|d|fVdS)zJoins a line ending in '' with the previous line (except when following comments). The joined line takes on the index of the first line. N\rc)endswith COMMENT_REmatchr]rVstrip)r.Zprimary_line_numberZnew_liner/r0rrrr6Bs     r6ccs4|D]*\}}td|}|}|r||fVqdS)z1 Strips comments and filter empty lines. rsN)rusubrw)r.r/r0rrrr7`s   r7cs2|r |jnd}|r.t|tfdd|}|S)zs Skip lines that match '--skip-requirements-regex' pattern Note: the regex pattern is only built once Ncs|dS)Nr2)rR)epatternrrvzskip_regex..)Zskip_requirements_regexrecompiler)r.r*r8rrzrr8ls  r8ccsL|D]B\}}t|D]$\}}t|}|s.q|||}q||fVqdS)aReplace all environment variables that can be retrieved via `os.getenv`. The only allowed format for environment variables defined in the requirement file is `${MY_VARIABLE_1}` to ensure two things: 1. Strings that contain a `$` aren't accidentally (partially) expanded. 2. Ensure consistency across platforms for requirement files. These points are the result of a discussion on the `github pull request #3514 `_. Valid characters in variable names follow the `POSIX standard `_ and are limited to uppercase letter, digits and the `_` (underscore). N) ENV_VAR_REfindallrTgetenvreplace)r.r/r0Zenv_varZvar_namerbrrrr9zs  r9)NNNNFNN)NNNNNNF)O__doc__Z __future__rrnrTr~rJrFZpip._vendor.six.movesrZpip._vendor.six.moves.urllibrrSZpip._internal.clirZpip._internal.downloadrZpip._internal.exceptionsrZ!pip._internal.models.search_scoper Zpip._internal.req.constructorsr r Zpip._internal.utils.typingr typingr rrrrrrrZpip._internal.reqrZpip._internal.cacherZpip._internal.indexrZpip._internal.network.sessionrintZ ReqFileLines__all__rIrQrurrPZeditablerOrYrDr@Zextra_index_urlZ always_unzipZ no_binaryZ only_binaryr_Z trusted_hostrXroZinstall_optionsZglobal_optionshashrprLrr&r'rErCr6r7r8r9rrrrs        (       +  site-packages/pip/_internal/req/__pycache__/__init__.cpython-38.pyc000064400000003244151464452710021231 0ustar00U .e @sddlmZddlZddlmZddlmZddlmZddl m Z ddl m Z erhdd l mZmZmZd d d d gZeeZddd ZdS))absolute_importN) indent_log)MYPY_CHECK_RUNNING)parse_requirements)InstallRequirement)RequirementSet)AnyListSequencerrrinstall_given_reqsc Os|r tdddd|Dt|D]}|jrbtd|jt|jdd}W5QRXz|j||f||Wn0tk r|jo|j }|r| YnX|jo|j}|r| | q,W5QRX|S)zu Install everything in the given list. (to be called after having downloaded and unpacked the packages) z!Installing collected packages: %sz, cSsg|] }|jqSr )name).0Zreqr r >/usr/lib/python3.8/site-packages/pip/_internal/req/__init__.py +sz&install_given_reqs..zFound existing installation: %sT)Z auto_confirm) loggerinfojoinrZconflicts_withZ uninstallZinstall ExceptionZinstall_succeededZrollbackZcommitZremove_temporary_source) Z to_installZinstall_optionsZglobal_optionsargskwargsZ requirementZuninstalled_pathsetZshould_rollbackZ should_commitr r rr sN)r )Z __future__rZloggingZpip._internal.utils.loggingrZpip._internal.utils.typingrZreq_filerZ req_installrZreq_setrtypingr r r __all__Z getLogger__name__rr r r r rs       site-packages/pip/_internal/req/__pycache__/constructors.cpython-38.opt-1.pyc000064400000024164151464452710023205 0ustar00U .e48@sdZddlZddlZddlZddlmZddlmZmZddl m Z ddl m Z m Z ddlmZddlmZmZdd lmZdd lmZdd lmZdd lmZdd lmZmZddlmZddl m!Z!ddl"m#Z#m$Z$ddl%m&Z&erddl'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-ddl.m/Z/dddgZ0e1e2Z3e j45Z6ddZ7ddZ8ddZ9ddZ:ddZ;Gd d!d!e<Z=d"d#Z>d/d%dZ?d&d'Z@d(d)ZAd*d+ZBd0d,dZCd1d-d.ZDdS)2a~Backing implementation for InstallRequirement's various constructors The idea here is that these formed a major chunk of InstallRequirement's size so, moving them and support code dedicated to them outside of that class helps creates for better understandability for the rest of the code. These are meant to be used elsewhere within pip to create instances of InstallRequirement. N)Marker)InvalidRequirement Requirement) Specifier)RequirementParseErrorparse_requirements)InstallationError)PyPITestPyPI)Link)make_pyproject_path)InstallRequirement)ARCHIVE_EXTENSIONS)is_installable_dirsplitext)MYPY_CHECK_RUNNING) path_to_url)is_urlvcs)Wheel)AnyDictOptionalSetTupleUnion) WheelCacheinstall_req_from_editableinstall_req_from_lineparse_editablecCs t|d}|tkrdSdS)z9Return True if `name` is a considered as an archive file.TF)rlowerr)nameZextr#B/usr/lib/python3.8/site-packages/pip/_internal/req/constructors.pyis_archive_file4sr%cCs6td|}d}|r*|d}|d}n|}||fS)Nz^(.+)(\[[^\]]+\])$r )rematchgroup)pathmextrasZpath_no_extrasr#r#r$ _strip_extras=s   r-cCs|s tStd|jS)N placeholder)setrr!r,)r,r#r#r$convert_extrasJsr0c Cs`|}t|\}}tj|rptjtj|dshdtj|}t|}tj |r`|d7}t |t |}| drt|j}|r||td| jfS||dfStD]&}| d|rd||f}qqd |krt d ||d d d  }t|s:d |dddtjDd} t | t|j}|sVt d|||dfS)aParses an editable requirement into: - a requirement name - an URL - extras - editable options Accepted requirements: svn+http://blahblah@rev#egg=Foobar[baz]&subdirectory=version_subdir .[some_extra] zsetup.pyzMFile "setup.py" not found. Directory cannot be installed in editable mode: {}zb (A "pyproject.toml" file was found, but editable mode currently requires a setup.py based build.)zfile:r.Nz%s:z%s+%s+z{} is not a valid editable requirement. It should either be a path to a local project or a VCS URL (beginning with svn+, git+, hg+, or bzr+).r rzFor --editable=%s only z, cSsg|]}|jdqS)z+URLr").0Zbackendr#r#r$ sz"parse_editable..z is currently supportedzZCould not detect requirement name for '%s', please specify one with #egg=your_package_name)r-osr*isdirexistsjoinformatabspathr isfilerrr! startswithr egg_fragmentrr,rsplitZ get_backendZbackends) editable_requrlZ url_no_extrasr,msgZpyproject_pathZ package_nameZversion_controlZvc_typeZ error_messager#r#r$rQsb          c Csd}tj|rd}zFt|d2}tt||dd|ddd7}W5QRXWqtk rtj d |d d YqXn |d |7}|S) zReturns helpful msg in case requirements file does not exist, or cannot be parsed. :params req: Requirements file path z It does exist.rz The argument you provided z(%s) appears to be az" requirements file. If that is thez# case, use the '-r' flag to installz" the packages specified within it.z2Cannot parse '%s' as requirements fileT)exc_infoz File '%s' does not exist.) r5r*r7opennextrreadrloggerdebug)reqrAfpr#r#r$deduce_helpful_msgs.   rLc@seZdZddZdS)RequirementPartscCs||_||_||_||_dSN) requirementlinkmarkersr,)selfrOrPrQr,r#r#r$__init__szRequirementParts.__init__N)__name__ __module__ __qualname__rSr#r#r#r$rMsrMcCs`t|\}}}|dk rFz t|}WqJtk rBtd|YqJXnd}t|}t||d|S)NInvalid requirement: '%s')rrrrr rM)r?r"r@Zextras_overriderJrPr#r#r$parse_req_from_editables rXFc CsLt|}|jjdkr|jjnd}t|j||d|j||||r>|ni||jd S)NfileT) source_dirZeditablerP constraint use_pep517isolatedoptions wheel_cacher,)rXrPschemeZ file_pathr rOr,) r? comes_fromr\r]r^r_r[partsrZr#r#r$rs  cCs>tjj|krdStjjdk r,tjj|kr,dS|dr:dSdS)akChecks whether the string "looks like" a path on the filesystem. This does not check whether the target actually exists, only judge from the appearance. Returns true if any of the following conditions is true: * a path separator is found (either os.path.sep or os.path.altsep); * a dot is found (which represents the current directory). TN.F)r5r*sepaltsepr<r2r#r#r$_looks_like_paths  rfcCst|r0tj|r0t|r$t|Std|t|slenrHZwarning)r*r"Z urlreq_partsr#r#r$_get_url_from_path s(   ricst|rd}nd}||krF||d\}}|}|s.with_sourcezIt looks like a path.=c3s|]}|kVqdSrNr#)r3op) req_as_stringr#r$ jsz&parse_req_from_line..z,= is not a valid operator. Did you mean == ?rBzInvalid requirement: {!r}z Hint: {}) rr>striprr5r*normpathr:r r-rir`r'searchr@rZis_wheelrfilenamer"versionr=r0rrrdrLany operatorsr9rrM)r"rlZ marker_sepZmarkers_as_stringrQr*rPZextras_as_stringpr@Zwheelr,rmrJZadd_msgrAr#)rlrpr$parse_req_from_line+sj         rzc Cs6t||}t|j||j|j|||r&|ni|||jd S)aCreates an InstallRequirement from a name, which might be a requirement, directory containing 'setup.py', filename, or URL. :param line_source: An optional string describing where the line is from, for logging purposes in case of an error. )rPrQr\r]r^r_r[r,)rzr rOrPrQr,) r"rar\r]r^r_r[rlrbr#r#r$rzs  cCszz t|}Wn tk r,td|YnXtjtjg}|jrh|rh|jrh|jj|krhtd|j |ft |||||dS)NrWzkPackages installed from PyPI cannot depend on packages which are not also hosted on PyPI. %s depends on %s )r]r_r\) rrrr Zfile_storage_domainr r@rPZnetlocr"r )Z req_stringrar]r_r\rJZdomains_not_allowedr#r#r$install_req_from_req_strings,  r{)NNFNNF)NNFNNFN)NFNN)E__doc__Zloggingr5r'Zpip._vendor.packaging.markersrZ"pip._vendor.packaging.requirementsrrZ pip._vendor.packaging.specifiersrZpip._vendor.pkg_resourcesrrZpip._internal.exceptionsrZpip._internal.models.indexr r Zpip._internal.models.linkr Zpip._internal.pyprojectr Zpip._internal.req.req_installr Zpip._internal.utils.filetypesrZpip._internal.utils.miscrrZpip._internal.utils.typingrZpip._internal.utils.urlsrZpip._internal.vcsrrZpip._internal.wheelrtypingrrrrrrZpip._internal.cacher__all__Z getLoggerrTrHZ _operatorskeysrxr%r-r0rrLobjectrMrXrrfrirzrr{r#r#r#r$sr                J "Q site-packages/pip/_internal/req/__pycache__/req_file.cpython-38.opt-1.pyc000064400000022300151464452710022211 0ustar00U .e7 @sdZddlmZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl mZddlmZddlmZdd lmZdd lmZmZdd lmZer dd lmZmZmZmZmZm Z m!Z!m"Z"dd l#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*ee"e+e!fZ,dgZ-e.dej/Z0e.dZ1e.dZ2ej3ej4ej5ej6ej7ej8ej9ej:ej;ejej?g Z@ejAejBejCgZDddeDDZEd)ddZFddZGd*ddZHddZIdd ZJd!d"ZKd#d$ZLd%d&ZMd'd(ZNdS)+z Requirements file parsing )absolute_importN) filterfalse)parse) cmdoptions)get_file_contentRequirementsFileParseError) SearchScope)install_req_from_editableinstall_req_from_line)MYPY_CHECK_RUNNING)AnyCallableIteratorListNoReturnOptionalTextTuple)InstallRequirement) WheelCache) PackageFinder) PipSessionparse_requirementsz^(http|https|file):z (^|\s+)#.*$z#(?P\$\{(?P[A-Z0-9_]+)\})cCsg|]}t|jqS)strdest).0orr>/usr/lib/python3.8/site-packages/pip/_internal/req/req_file.py Ksr Fc csj|dkrtdt|||d\}} t| |} | D]4\} } t| || |||||||d } | D] }|VqXq0dS)a1Parse a requirements file and yield InstallRequirement instances. :param filename: Path or url of requirements file. :param finder: Instance of pip.index.PackageFinder. :param comes_from: Origin description of requirements. :param options: cli options. :param session: Instance of pip.download.PipSession. :param constraint: If true, parsing a constraint file rather than requirements file. :param wheel_cache: Instance of pip.wheel.WheelCache :param use_pep517: Value of the --use-pep517 option. NzCparse_requirements() missing 1 required keyword argument: 'session') comes_fromsession) use_pep517 constraint) TypeErrorr preprocess process_line)filenamefinderr!optionsr"r$ wheel_cacher#_content lines_enum line_numberlineZreq_iterreqrrrrNs*    cCs6t|dd}t|}t|}t||}t|}|S)zSplit, filter, and join lines, and return a line iterator :param content: the content of the requirements file :param options: cli options )start) enumerate splitlines join_linesignore_comments skip_regexexpand_env_variables)r-r*r.rrrr&ys  r&c ! cst|} | } d| _|r"|j| _t|\} } tjdkrB| d} | t | | \}}d| rbdnd||f}| r|r||j nd}|rt ||i}tD]&}||jkr|j|r|j|||<qd||}t| |||||| |d Vn|jr |r|j nd}t|jd ||| ||d Vn|js0|jr|jrH|jd }d}n|jd }d }t|rpt||}n"t|stjtj||}t|||||||d }|D] }|Vqn|jr|j|_n|r|j }|j!}|jr|jg}|j"d krg}|j#r|$|j#|j rd|j d }tjtj%|}tj||}tj&|rZ|}|'|t(||d}||_)|j*r|+|j,pgD] }d||} |j-|| dqdS)a#Process a single requirements line; This can result in creating/yielding requirements, or updating the finder. For lines that contain requirements, the only options that have an effect are from SUPPORTED_OPTIONS_REQ, and they are scoped to the requirement. Other options from SUPPORTED_OPTIONS may be present, but are ignored. For lines that do not contain requirements, the only options that have an effect are from SUPPORTED_OPTIONS. Options from SUPPORTED_OPTIONS_REQ may be present, but are ignored. These lines may contain multiple options (although our docs imply only one is supported), and all our parsed and affect the finder. :param constraint: If True, parsing a constraints file. :param options: OptionParser options that we may update N)utf8z%s %s (line %s)z-cz-rFz line {} of {})r!r#isolatedr*r+r$ line_sourcer)r!r#r$r>r+T)r$r+) find_links index_urls)source). build_parserZget_default_values index_urlZformat_controlbreak_args_optionssys version_infoencode parse_argsshlexsplitZ isolated_moderZcheck_install_build_globalSUPPORTED_OPTIONS_REQ_DEST__dict__formatr Z editablesr requirements constraints SCHEME_REsearch urllib_parseZurljoinospathjoindirnamerrequire_hashesr@rAno_indexZextra_index_urlsextendabspathexistsappendr search_scopepreZset_allow_all_prereleasesZ trusted_hostsZadd_trusted_host)!r0r(r/r)r!r*r"r+r#r$parserdefaultsZargs_strZ options_strZoptsr,Zline_comes_fromr>Z req_optionsrr?Zreq_pathZnested_constraintZ parsed_reqsr1r@rAvalueZreq_dirZrelative_to_reqs_filer^ZhostrBrrrr's                   r'cCsf|d}g}|dd}|D]2}|ds6|dr<qRq|||dqd|d|fS)zBreak up the line into an args and options string. We only want to shlex (and then optparse) the options, not the args. args can contain markers which are corrupted by shlex.  N-z--r)rK startswithr]poprV)r0tokensargsr*tokenrrrrEs    rEcsDtjdd}tt}|D]}|}||qfdd}||_|S)z7 Return a parser for parsing requirement lines F)Zadd_help_optioncsd|f}t|dS)NzInvalid requirement: %s %sr)selfmsgr0rr parser_exit6s z!build_parser..parser_exit)optparseZ OptionParserSUPPORTED_OPTIONSSUPPORTED_OPTIONS_REQZ add_optionexit)r0r`Zoption_factoriesZoption_factoryZoptionrmrrlrrC(s   rCccsd}g}|D]v\}}|dr(t|rjt|r:d|}|r^|||d|fVg}q||fVq |sr|}||dq |r|d|fVdS)zJoins a line ending in '' with the previous line (except when following comments). The joined line takes on the index of the first line. N\rc)endswith COMMENT_REmatchr]rVstrip)r.Zprimary_line_numberZnew_liner/r0rrrr6Bs     r6ccs4|D]*\}}td|}|}|r||fVqdS)z1 Strips comments and filter empty lines. rsN)rusubrw)r.r/r0rrrr7`s   r7cs2|r |jnd}|r.t|tfdd|}|S)zs Skip lines that match '--skip-requirements-regex' pattern Note: the regex pattern is only built once Ncs|dS)Nr2)rR)epatternrrvzskip_regex..)Zskip_requirements_regexrecompiler)r.r*r8rrzrr8ls  r8ccsL|D]B\}}t|D]$\}}t|}|s.q|||}q||fVqdS)aReplace all environment variables that can be retrieved via `os.getenv`. The only allowed format for environment variables defined in the requirement file is `${MY_VARIABLE_1}` to ensure two things: 1. Strings that contain a `$` aren't accidentally (partially) expanded. 2. Ensure consistency across platforms for requirement files. These points are the result of a discussion on the `github pull request #3514 `_. Valid characters in variable names follow the `POSIX standard `_ and are limited to uppercase letter, digits and the `_` (underscore). N) ENV_VAR_REfindallrTgetenvreplace)r.r/r0Zenv_varZvar_namerbrrrr9zs  r9)NNNNFNN)NNNNNNF)O__doc__Z __future__rrnrTr~rJrFZpip._vendor.six.movesrZpip._vendor.six.moves.urllibrrSZpip._internal.clirZpip._internal.downloadrZpip._internal.exceptionsrZ!pip._internal.models.search_scoper Zpip._internal.req.constructorsr r Zpip._internal.utils.typingr typingr rrrrrrrZpip._internal.reqrZpip._internal.cacherZpip._internal.indexrZpip._internal.network.sessionrintZ ReqFileLines__all__rIrQrurrPZeditablerOrYrDr@Zextra_index_urlZ always_unzipZ no_binaryZ only_binaryr_Z trusted_hostrXroZinstall_optionsZglobal_optionshashrprLrr&r'rErCr6r7r8r9rrrrs        (       +  site-packages/pip/_internal/req/__pycache__/__init__.cpython-38.opt-1.pyc000064400000003244151464452710022170 0ustar00U .e @sddlmZddlZddlmZddlmZddlmZddl m Z ddl m Z erhdd l mZmZmZd d d d gZeeZddd ZdS))absolute_importN) indent_log)MYPY_CHECK_RUNNING)parse_requirements)InstallRequirement)RequirementSet)AnyListSequencerrrinstall_given_reqsc Os|r tdddd|Dt|D]}|jrbtd|jt|jdd}W5QRXz|j||f||Wn0tk r|jo|j }|r| YnX|jo|j}|r| | q,W5QRX|S)zu Install everything in the given list. (to be called after having downloaded and unpacked the packages) z!Installing collected packages: %sz, cSsg|] }|jqSr )name).0Zreqr r >/usr/lib/python3.8/site-packages/pip/_internal/req/__init__.py +sz&install_given_reqs..zFound existing installation: %sT)Z auto_confirm) loggerinfojoinrZconflicts_withZ uninstallZinstall ExceptionZinstall_succeededZrollbackZcommitZremove_temporary_source) Z to_installZinstall_optionsZglobal_optionsargskwargsZ requirementZuninstalled_pathsetZshould_rollbackZ should_commitr r rr sN)r )Z __future__rZloggingZpip._internal.utils.loggingrZpip._internal.utils.typingrZreq_filerZ req_installrZreq_setrtypingr r r __all__Z getLogger__name__rr r r r rs       site-packages/pip/_internal/req/__pycache__/req_uninstall.cpython-38.opt-1.pyc000064400000041635151464452710023317 0ustar00U .e&\@s|ddlmZddlZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl m Z mZddlmZmZmZddlmZddlmZmZmZmZmZmZmZmZmZdd lm Z m!Z!dd l"m#Z#e#r dd l$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-dd l.m/Z/e0e1Z2d dZ3ddZ4e4ddZ5ddZ6ddZ7ddZ8Gddde9Z:Gddde9Z;Gddde9Z.unique) functoolswraps)r3r4r(r2r)_unique=sr7ccstt|d}|D]t}tj|j|d}|V|drtj |\}}|dd}tj||d}|Vtj||d}|VqdS)a Yield all the uninstallation paths for dist based on RECORD-without-.py[co] Yield paths to all the files in RECORD. For each .py file in RECORD, add the .pyc and .pyo in the same directory. UninstallPathSet.add() takes care of the __pycache__ .py[co]. ZRECORDr.pyN.pyc.pyo) csvreaderr Zget_metadata_linesr!r"r#locationendswithsplit)r%rrowr"Zdnr3baser(r(r)uninstallation_pathsJs   rDcsJtjjt}t|tdD]*tfdd|D}|s|q|S)zCompact a path set to contain the minimal number of paths necessary to contain all paths in the set. If /a/path/ and /a/path/to/a/file.txt are both in the set, leave only the shorter path.keyc3s:|]2}|do0t|dkVqdS)*N) startswithrstriplen).0Z shortpathr"sepr(r) lszcompact..)r!r"rMr,sortedrJanyr-)pathsZ short_pathsZ should_skipr(rLr)compactbs rRc stdd|D}t|}ttdd|Dtd}t}dd|D]tfdd|DrfqJt}t}tD]B\}}|fdd|D|fd d|Dq|||sJ| || tj qJtt |j ||BS) zReturns a set containing the paths that need to be renamed. This set may include directories when the original sequence of paths included every file on disk. css|]}tj||fVqdSr+)r!r"normcaserKpr(r(r)rN}sz&compress_for_rename..css|]}tj|dVqdS)rN)r!r"r@rTr(r(r)rNsrEcWstjtjj|Sr+)r!r"rSr#)ar(r(r) norm_joinsz&compress_for_rename..norm_joinc3s |]}tj|VqdSr+)r!r"rSrH)rKw)rootr(r)rNsc3s|]}|VqdSr+r()rKddirnamerWrYr(r)rNsc3s|]}|VqdSr+r()rKfr[r(r)rNs)dictr,rOvaluesrJrPr!walkupdatedifference_updater-rMmap __getitem__) rQZcase_mapZ remainingZ uncheckedZ wildcardsZ all_filesZ all_subdirsZsubdirsfilesr(r[r)compress_for_renamevs6    rfc Cst|}t}t}t}|D]>}|dr.q|ds@d|krR|tj|||qtttjj|}t|}|D]d}t |D]T\}} } | D]D} | drqtj || } tj | rtj| |kr|| qqq||dd|DB}||fS)asReturns a tuple of 2 sets of which paths to display to user The first set contains paths that would be deleted. Files of a package are not added and the top-level directory of the package has a '*' added at the end - to signify that all it's contents are removed. The second set contains files that would have been skipped in the above folders. r:z __init__.py .dist-infocSsh|]}tj|dqS)rG)r!r"r#)rKfolderr(r(r) sz.compress_for_output_listing..) r,r?r-r!r"r\rcrSrRr`r#isfile) rQ will_remove will_skipZfoldersrer"Z_normcased_filesrhdirpath_ZdirfilesZfnameZfile_r(r(r)compress_for_output_listings4     roc@sLeZdZdZddZddZddZdd Zd d Zd d Z e ddZ dS)StashedUninstallPathSetzWA set of file rename operations to stash files while tentatively uninstalling them.cCsi|_g|_dSr+) _save_dirs_movesselfr(r(r)__init__sz StashedUninstallPathSet.__init__cCsDz t|}Wntk r*tdd}YnX||jtj|<|jS)zStashes a directory. Directories are stashed adjacent to their original location if possible, or else moved/copied into the user's temp dir. uninstallZkind)rOSErrorrrqr!r"rS)rtr"save_dirr(r(r)_get_directory_stashs  z,StashedUninstallPathSet._get_directory_stashcCstj|}tj|d}}d}||krfz|j|}WqWntk rPYnXtj||}}q"tj|}tdd}||j|<tj||}|r|tjjkrtj |j|S|jS)zStashes a file. If no root has been provided, one will be created for the directory in the user's temp directory.Nrvrw) r!r"rSr\rqKeyErrorrrelpathcurdirr#)rtr"headZold_headryr|r(r(r)_get_file_stashs"     z'StashedUninstallPathSet._get_file_stashcCsltj|otj| }|r*||}n ||}|j||f|r^tj|r^t|t |||S)zStashes the directory or file and returns its new location. Handle symlinks as files to avoid modifying the symlink targets. ) r!r"isdirislinkrzrrrr$rmdirr)rtr"Z path_is_dirnew_pathr(r(r)stashs    zStashedUninstallPathSet.stashcCs,|jD]\}}|q g|_i|_dS)z0Commits the uninstall by removing stashed files.N)rqitemsZcleanuprr)rtrnryr(r(r)commits zStashedUninstallPathSet.commitc Cs|jD]}tjd|q|jD]\}}zTtd||tj|sPtj|r\t |ntj |rpt |t ||Wq t k r}ztd|td|W5d}~XYq Xq |dS)z2Undoes the uninstall by moving stashed files back.Moving to %s from %szReplacing %s from %szFailed to restore %sz Exception: %sN)r)rrlogginginfologgerdebugr!r"rjrunlinkrrrrxerrorr)rtrUrr"Zexr(r(r)rollback&s     z StashedUninstallPathSet.rollbackcCs t|jSr+)boolrrrsr(r(r) can_rollback:sz$StashedUninstallPathSet.can_rollbackN) __name__ __module__ __qualname____doc__rurzrrrrpropertyrr(r(r(r)rps rpc@s^eZdZdZddZddZddZdd Zdd d Zd dZ ddZ ddZ e ddZ dS)UninstallPathSetzMA set of file paths to be removed in the uninstallation of a requirement.cCs(t|_t|_i|_||_t|_dSr+)r,rQ_refusepthr%rp _moved_paths)rtr%r(r(r)ruCs zUninstallPathSet.__init__cCst|S)zs Return True if the given path is one we are permitted to remove/modify, False otherwise. )r)rtr"r(r(r) _permittedKszUninstallPathSet._permittedcCstj|\}}tjt|tj|}tj|s:dS||rR|j |n |j |tj |ddkrt r| t |dS)Nr8)r!r"r@r#rrSexistsrrQr-rsplitextr r)rtr"r~tailr(r(r)r-Ts   zUninstallPathSet.addcCsLt|}||r<||jkr*t||j|<|j||n |j|dSr+)rrrUninstallPthEntriesr-r)rtpth_fileentryr(r(r)add_pthhs   zUninstallPathSet.add_pthFc Cs|jstd|jjdS|jjd|jj}td|tp|sP||r|j}t |j}t t |D]}| |t d|ql|jD] }|qtd|W5QRXdS)z[Remove paths in ``self.paths`` with confirmation (unless ``auto_confirm`` is True).z7Can't uninstall '%s'. No files were found to uninstall.N-zUninstalling %s:zRemoving file or directory %szSuccessfully uninstalled %s)rQrrr% project_nameversionr _allowed_to_proceedrrfrOrRrrrr_remove)rtZ auto_confirmverboseZdist_name_versionZmovedZ for_renamer"rr(r(r)rrs&    zUninstallPathSet.removecCsndd}|st|j\}}nt|j}t}|d||d||d|j|r`|dt|jtddd kS) zIDisplay which files would be deleted and prompt for confirmation c SsD|sdSt|t"tt|D]}t|q&W5QRXdSr+)rrr rOrR)msgrQr"r(r(r)_displays  z6UninstallPathSet._allowed_to_proceed.._displayz Would remove:z+Would not remove (might be manually added):z%Would not remove (outside of prefix):zWill actually move:zProceed (y/n)? )ynr)rorQr,rrfr )rtrrrkrlr(r(r)rs     z$UninstallPathSet._allowed_to_proceedcCsR|jjstd|jjdStd|jj|j|j D] }|q@dS)z1Rollback the changes previously made by remove().z'Can't roll back %s; was not uninstalledNzRolling back uninstall of %s) rrrrr%rrrrr_)rtrr(r(r)rs zUninstallPathSet.rollbackcCs|jdS)z?Remove temporary save dir: rollback will no longer be possible.N)rrrsr(r(r)rszUninstallPathSet.commitc st|j}t|s.td|j|tj||S|ddt dt dhDkrhtd|j|||S||}t |}d t |j}|jotj|j}t|jdd}|r|jd r|j|s||j|d r|d D]&}tjtj|j|} || qn|d r|d rB|d ngfd d|d DD]J} tj|j| } || || d|| d|| dqbn|rtd |jn|jdr ||jtj|jd} tjtj|jd} || d| n|rP|jdrPt |D]} || q.ZstdlibZ platstdlibzsz.UninstallPathSet.from_dist..r8r:r;zCannot uninstall {!r}. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall.z.eggrzeasy-install.pthz./rgrAz)Not sure how to uninstall: %s - Check: %sZscriptsz.batconsole_scripts)groupF gui_scriptsT)0rr>rrrrFsysprefix sysconfigZget_pathrformatrZ to_filenamerZegg_infor!r"rgetattrZ _providerr?r-Z has_metadataZ get_metadata splitlinesnormpathr#rr@r\rrDopenrSreadlinestriprZmetadata_isdirZmetadata_listdirr rrrZ get_entry_mapkeysextendr*)clsr%Z dist_pathr'Zdevelop_egg_linkZdevelop_egg_link_egg_infoZegg_info_existsZdistutils_egg_infoZinstalled_filer"Z top_level_pkgZeasy_install_eggZeasy_install_pthfhZ link_pointerZscriptr&Z_scripts_to_removernamersr(rr) from_dists                      zUninstallPathSet.from_distN)FF)rrrrrurr-rrrrr classmethodrr(r(r(r)r@s  rc@s,eZdZddZddZddZddZd S) rcCs0tj|std|||_t|_d|_dS)Nz.Cannot remove entries from nonexistent file %s)r!r"rjrfiler,entries _saved_lines)rtrr(r(r)ruIs zUninstallPthEntries.__init__cCs<tj|}tr,tj|ds,|dd}|j|dS)Nr\/)r!r"rSr splitdrivereplacerr-)rtrr(r(r)r-Ss  zUninstallPthEntries.addc Cstd|jt|jd}|}||_W5QRXtdd|DrLd}nd}|r~|d|ds~|d|d|d<|j D]>}z$td || ||dWqt k rYqXqt|jd }| |W5QRXdS) NzRemoving pth entries from %s:rbcss|]}d|kVqdS)s Nr()rKliner(r(r)rNjsz-UninstallPthEntries.remove..z  zutf-8zRemoving entry: %swb) rrrr readlinesrrPr?encoderr ValueError writelines)rtrlinesZendlinerr(r(r)rcs"  zUninstallPthEntries.removec CsR|jdkrtd|jdStd|jt|jd}||jW5QRXdS)Nz.Cannot roll back changes to %s, none were madeFz!Rolling %s back to previous staterT)rrrrrrr)rtrr(r(r)rzs zUninstallPthEntries.rollbackN)rrrrur-rrr(r(r(r)rHs r)=Z __future__rr<r5rr!rrZ pip._vendorrZpip._internal.exceptionsrZpip._internal.locationsrrZpip._internal.utils.compatrrr Zpip._internal.utils.loggingr Zpip._internal.utils.miscr r r rrrrrrZpip._internal.utils.temp_dirrrZpip._internal.utils.typingrtypingrrrrrrrrrZpip._vendor.pkg_resourcesr Z getLoggerrrr*r7rDrRrfroobjectrprrr(r(r(r)s<    ,  ,    (3o site-packages/pip/_internal/req/__pycache__/req_set.cpython-38.opt-1.pyc000064400000013347151464452710022100 0ustar00U .e@sddlmZddlZddlmZddlmZddlmZddl m Z ddl m Z ddl mZdd lmZerdd lmZmZmZmZmZdd lmZeeZGd d d eZdS))absolute_importN) OrderedDict)canonicalize_name) pep425tags)InstallationError) indent_log)MYPY_CHECK_RUNNING)Wheel)DictIterableListOptionalTuple)InstallRequirementc@sXeZdZdddZddZddZd d Zd d ZdddZddZ ddZ ddZ d S)RequirementSetFTcCs*t|_||_||_g|_g|_g|_dS)z!Create a RequirementSet. N)r requirementsrequire_hashescheck_supported_wheelsunnamed_requirementsZsuccessfully_downloadedreqs_to_cleanup)selfrrr=/usr/lib/python3.8/site-packages/pip/_internal/req/req_set.py__init__s zRequirementSet.__init__cCs4tdd|jDddd}ddd|DS)Ncss|]}|js|VqdSN)Z comes_from.0reqrrr +sz)RequirementSet.__str__..cSs t|jSrrnamerrrr,z(RequirementSet.__str__..key css|]}t|jVqdSrstrrrrrrr.s)sortedrvaluesjoin)rrrrr__str__(s zRequirementSet.__str__cCsBt|jddd}d}|j|jjt|ddd|DdS) NcSs t|jSrrr!rrrr"4r#z)RequirementSet.__repr__..r$z4<{classname} object; {count} requirement(s): {reqs}>z, css|]}t|jVqdSrr'rrrrr;sz*RequirementSet.__repr__..)Z classnamecountZreqs)r)rr*format __class____name__lenr+)rr format_stringrrr__repr__0szRequirementSet.__repr__cCs|j|dSr)rappend)r install_reqrrradd_unnamed_requirement>sz&RequirementSet.add_unnamed_requirementcCst|j}||j|<dSr)rr r)rr5 project_namerrradd_named_requirementCs z$RequirementSet.add_named_requirementNc Cs||s$td|j|jgdfS|jrd|jjrdt|jj}t }|j rd| |sdt d|j|js~|||gdfSz||j}Wntk rd}YnX|dko|o|j o|j|jko|jj|jjk}|rt d|||jf|s |||g|fS|js|js"g|fS|joB|jo@|jj|jjk }|rd|j|t d|jd|_ttt|jt|jB|_td||j|g|fS)a&Add install_req as a requirement to install. :param parent_req_name: The name of the requirement that needed this added. The name is used because when multiple unnamed requirements resolve to the same name, we could otherwise end up with dependency links that point outside the Requirements set. parent_req must already be added. Note that None implies that this is a user supplied requirement, vs an inferred one. :param extras_requested: an iterable of extras used to evaluate the environment markers. :return: Additional requirements to scan. That is either [] if the requirement is not applicable, or [install_req] if the requirement is applicable and has just been added. z6Ignoring %s: markers '%s' don't match your environmentNz-%s is not a supported wheel on this platform.z5Double requirement given: %s (already in %s, name=%r)zhCould not satisfy constraints for '%s': installation from path or url cannot be constrained to a versionFzSetting %s extras to: %s)Z match_markersloggerinfor ZmarkerslinkZis_wheelr filenamerZ get_supportedrZ supportedrr6get_requirementKeyError constraintZextrasrZ specifierr8pathrr4tupler)setdebug) rr5Zparent_req_nameZextras_requestedZwheelZtagsZ existing_reqZhas_conflicting_requirementZdoes_not_satisfy_constraintrrradd_requirementJs           zRequirementSet.add_requirementcCs t|}||jko|j|j Sr)rrr?rr r7rrrhas_requirements  zRequirementSet.has_requirementcCs,t|}||jkr|j|Std|dS)NzNo project with the name %r)rrr>rErrrr=s  zRequirementSet.get_requirementc Cs4tdt|jD] }|qW5QRXdS)zClean up files, remove builds.zCleaning up...N)r9rCrrZremove_temporary_source)rrrrr cleanup_filess  zRequirementSet.cleanup_files)FT)NN) r0 __module__ __qualname__rr,r3r6r8rDrFr=rGrrrrrs   p  r)Z __future__rZlogging collectionsrZpip._vendor.packaging.utilsrZ pip._internalrZpip._internal.exceptionsrZpip._internal.utils.loggingrZpip._internal.utils.typingrZpip._internal.wheelr typingr r r r rZpip._internal.req.req_installrZ getLoggerr0r9objectrrrrrs          site-packages/pip/_internal/req/req_tracker.py000064400000006173151464452710015512 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False from __future__ import absolute_import import contextlib import errno import hashlib import logging import os from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from types import TracebackType from typing import Iterator, Optional, Set, Type from pip._internal.req.req_install import InstallRequirement from pip._internal.models.link import Link logger = logging.getLogger(__name__) class RequirementTracker(object): def __init__(self): # type: () -> None self._root = os.environ.get('PIP_REQ_TRACKER') if self._root is None: self._temp_dir = TempDirectory(delete=False, kind='req-tracker') self._root = os.environ['PIP_REQ_TRACKER'] = self._temp_dir.path logger.debug('Created requirements tracker %r', self._root) else: self._temp_dir = None logger.debug('Re-using requirements tracker %r', self._root) self._entries = set() # type: Set[InstallRequirement] def __enter__(self): # type: () -> RequirementTracker return self def __exit__( self, exc_type, # type: Optional[Type[BaseException]] exc_val, # type: Optional[BaseException] exc_tb # type: Optional[TracebackType] ): # type: (...) -> None self.cleanup() def _entry_path(self, link): # type: (Link) -> str hashed = hashlib.sha224(link.url_without_fragment.encode()).hexdigest() return os.path.join(self._root, hashed) def add(self, req): # type: (InstallRequirement) -> None link = req.link info = str(req) entry_path = self._entry_path(link) try: with open(entry_path) as fp: # Error, these's already a build in progress. raise LookupError('%s is already being built: %s' % (link, fp.read())) except IOError as e: if e.errno != errno.ENOENT: raise assert req not in self._entries with open(entry_path, 'w') as fp: fp.write(info) self._entries.add(req) logger.debug('Added %s to build tracker %r', req, self._root) def remove(self, req): # type: (InstallRequirement) -> None link = req.link self._entries.remove(req) os.unlink(self._entry_path(link)) logger.debug('Removed %s from build tracker %r', req, self._root) def cleanup(self): # type: () -> None for req in set(self._entries): self.remove(req) remove = self._temp_dir is not None if remove: self._temp_dir.cleanup() logger.debug('%s build tracker %r', 'Removed' if remove else 'Cleaned', self._root) @contextlib.contextmanager def track(self, req): # type: (InstallRequirement) -> Iterator[None] self.add(req) yield self.remove(req) site-packages/pip/_internal/req/req_uninstall.py000064400000056046151464452710016074 0ustar00from __future__ import absolute_import import csv import functools import logging import os import sys import sysconfig from pip._vendor import pkg_resources from pip._internal.exceptions import UninstallationError from pip._internal.locations import bin_py, bin_user from pip._internal.utils.compat import WINDOWS, cache_from_source, uses_pycache from pip._internal.utils.logging import indent_log from pip._internal.utils.misc import ( FakeFile, ask, dist_in_usersite, dist_is_local, egg_link_path, is_local, normalize_path, renames, rmtree, ) from pip._internal.utils.temp_dir import AdjacentTempDirectory, TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ( Any, Callable, Dict, Iterable, Iterator, List, Optional, Set, Tuple, ) from pip._vendor.pkg_resources import Distribution logger = logging.getLogger(__name__) def _script_names(dist, script_name, is_gui): # type: (Distribution, str, bool) -> List[str] """Create the fully qualified name of the files created by {console,gui}_scripts for the given ``dist``. Returns the list of file names """ if dist_in_usersite(dist): bin_dir = bin_user else: bin_dir = bin_py exe_name = os.path.join(bin_dir, script_name) paths_to_remove = [exe_name] if WINDOWS: paths_to_remove.append(exe_name + '.exe') paths_to_remove.append(exe_name + '.exe.manifest') if is_gui: paths_to_remove.append(exe_name + '-script.pyw') else: paths_to_remove.append(exe_name + '-script.py') return paths_to_remove def _unique(fn): # type: (Callable) -> Callable[..., Iterator[Any]] @functools.wraps(fn) def unique(*args, **kw): # type: (Any, Any) -> Iterator[Any] seen = set() # type: Set[Any] for item in fn(*args, **kw): if item not in seen: seen.add(item) yield item return unique @_unique def uninstallation_paths(dist): # type: (Distribution) -> Iterator[str] """ Yield all the uninstallation paths for dist based on RECORD-without-.py[co] Yield paths to all the files in RECORD. For each .py file in RECORD, add the .pyc and .pyo in the same directory. UninstallPathSet.add() takes care of the __pycache__ .py[co]. """ r = csv.reader(FakeFile(dist.get_metadata_lines('RECORD'))) for row in r: path = os.path.join(dist.location, row[0]) yield path if path.endswith('.py'): dn, fn = os.path.split(path) base = fn[:-3] path = os.path.join(dn, base + '.pyc') yield path path = os.path.join(dn, base + '.pyo') yield path def compact(paths): # type: (Iterable[str]) -> Set[str] """Compact a path set to contain the minimal number of paths necessary to contain all paths in the set. If /a/path/ and /a/path/to/a/file.txt are both in the set, leave only the shorter path.""" sep = os.path.sep short_paths = set() # type: Set[str] for path in sorted(paths, key=len): should_skip = any( path.startswith(shortpath.rstrip("*")) and path[len(shortpath.rstrip("*").rstrip(sep))] == sep for shortpath in short_paths ) if not should_skip: short_paths.add(path) return short_paths def compress_for_rename(paths): # type: (Iterable[str]) -> Set[str] """Returns a set containing the paths that need to be renamed. This set may include directories when the original sequence of paths included every file on disk. """ case_map = dict((os.path.normcase(p), p) for p in paths) remaining = set(case_map) unchecked = sorted(set(os.path.split(p)[0] for p in case_map.values()), key=len) wildcards = set() # type: Set[str] def norm_join(*a): # type: (str) -> str return os.path.normcase(os.path.join(*a)) for root in unchecked: if any(os.path.normcase(root).startswith(w) for w in wildcards): # This directory has already been handled. continue all_files = set() # type: Set[str] all_subdirs = set() # type: Set[str] for dirname, subdirs, files in os.walk(root): all_subdirs.update(norm_join(root, dirname, d) for d in subdirs) all_files.update(norm_join(root, dirname, f) for f in files) # If all the files we found are in our remaining set of files to # remove, then remove them from the latter set and add a wildcard # for the directory. if not (all_files - remaining): remaining.difference_update(all_files) wildcards.add(root + os.sep) return set(map(case_map.__getitem__, remaining)) | wildcards def compress_for_output_listing(paths): # type: (Iterable[str]) -> Tuple[Set[str], Set[str]] """Returns a tuple of 2 sets of which paths to display to user The first set contains paths that would be deleted. Files of a package are not added and the top-level directory of the package has a '*' added at the end - to signify that all it's contents are removed. The second set contains files that would have been skipped in the above folders. """ will_remove = set(paths) will_skip = set() # Determine folders and files folders = set() files = set() for path in will_remove: if path.endswith(".pyc"): continue if path.endswith("__init__.py") or ".dist-info" in path: folders.add(os.path.dirname(path)) files.add(path) # probably this one https://github.com/python/mypy/issues/390 _normcased_files = set(map(os.path.normcase, files)) # type: ignore folders = compact(folders) # This walks the tree using os.walk to not miss extra folders # that might get added. for folder in folders: for dirpath, _, dirfiles in os.walk(folder): for fname in dirfiles: if fname.endswith(".pyc"): continue file_ = os.path.join(dirpath, fname) if (os.path.isfile(file_) and os.path.normcase(file_) not in _normcased_files): # We are skipping this file. Add it to the set. will_skip.add(file_) will_remove = files | { os.path.join(folder, "*") for folder in folders } return will_remove, will_skip class StashedUninstallPathSet(object): """A set of file rename operations to stash files while tentatively uninstalling them.""" def __init__(self): # type: () -> None # Mapping from source file root to [Adjacent]TempDirectory # for files under that directory. self._save_dirs = {} # type: Dict[str, TempDirectory] # (old path, new path) tuples for each move that may need # to be undone. self._moves = [] # type: List[Tuple[str, str]] def _get_directory_stash(self, path): # type: (str) -> str """Stashes a directory. Directories are stashed adjacent to their original location if possible, or else moved/copied into the user's temp dir.""" try: save_dir = AdjacentTempDirectory(path) # type: TempDirectory except OSError: save_dir = TempDirectory(kind="uninstall") self._save_dirs[os.path.normcase(path)] = save_dir return save_dir.path def _get_file_stash(self, path): # type: (str) -> str """Stashes a file. If no root has been provided, one will be created for the directory in the user's temp directory.""" path = os.path.normcase(path) head, old_head = os.path.dirname(path), None save_dir = None while head != old_head: try: save_dir = self._save_dirs[head] break except KeyError: pass head, old_head = os.path.dirname(head), head else: # Did not find any suitable root head = os.path.dirname(path) save_dir = TempDirectory(kind='uninstall') self._save_dirs[head] = save_dir relpath = os.path.relpath(path, head) if relpath and relpath != os.path.curdir: return os.path.join(save_dir.path, relpath) return save_dir.path def stash(self, path): # type: (str) -> str """Stashes the directory or file and returns its new location. Handle symlinks as files to avoid modifying the symlink targets. """ path_is_dir = os.path.isdir(path) and not os.path.islink(path) if path_is_dir: new_path = self._get_directory_stash(path) else: new_path = self._get_file_stash(path) self._moves.append((path, new_path)) if (path_is_dir and os.path.isdir(new_path)): # If we're moving a directory, we need to # remove the destination first or else it will be # moved to inside the existing directory. # We just created new_path ourselves, so it will # be removable. os.rmdir(new_path) renames(path, new_path) return new_path def commit(self): # type: () -> None """Commits the uninstall by removing stashed files.""" for _, save_dir in self._save_dirs.items(): save_dir.cleanup() self._moves = [] self._save_dirs = {} def rollback(self): # type: () -> None """Undoes the uninstall by moving stashed files back.""" for p in self._moves: logging.info("Moving to %s\n from %s", *p) for new_path, path in self._moves: try: logger.debug('Replacing %s from %s', new_path, path) if os.path.isfile(new_path) or os.path.islink(new_path): os.unlink(new_path) elif os.path.isdir(new_path): rmtree(new_path) renames(path, new_path) except OSError as ex: logger.error("Failed to restore %s", new_path) logger.debug("Exception: %s", ex) self.commit() @property def can_rollback(self): # type: () -> bool return bool(self._moves) class UninstallPathSet(object): """A set of file paths to be removed in the uninstallation of a requirement.""" def __init__(self, dist): # type: (Distribution) -> None self.paths = set() # type: Set[str] self._refuse = set() # type: Set[str] self.pth = {} # type: Dict[str, UninstallPthEntries] self.dist = dist self._moved_paths = StashedUninstallPathSet() def _permitted(self, path): # type: (str) -> bool """ Return True if the given path is one we are permitted to remove/modify, False otherwise. """ return is_local(path) def add(self, path): # type: (str) -> None head, tail = os.path.split(path) # we normalize the head to resolve parent directory symlinks, but not # the tail, since we only want to uninstall symlinks, not their targets path = os.path.join(normalize_path(head), os.path.normcase(tail)) if not os.path.exists(path): return if self._permitted(path): self.paths.add(path) else: self._refuse.add(path) # __pycache__ files can show up after 'installed-files.txt' is created, # due to imports if os.path.splitext(path)[1] == '.py' and uses_pycache: self.add(cache_from_source(path)) def add_pth(self, pth_file, entry): # type: (str, str) -> None pth_file = normalize_path(pth_file) if self._permitted(pth_file): if pth_file not in self.pth: self.pth[pth_file] = UninstallPthEntries(pth_file) self.pth[pth_file].add(entry) else: self._refuse.add(pth_file) def remove(self, auto_confirm=False, verbose=False): # type: (bool, bool) -> None """Remove paths in ``self.paths`` with confirmation (unless ``auto_confirm`` is True).""" if not self.paths: logger.info( "Can't uninstall '%s'. No files were found to uninstall.", self.dist.project_name, ) return dist_name_version = ( self.dist.project_name + "-" + self.dist.version ) logger.info('Uninstalling %s:', dist_name_version) with indent_log(): if auto_confirm or self._allowed_to_proceed(verbose): moved = self._moved_paths for_rename = compress_for_rename(self.paths) for path in sorted(compact(for_rename)): moved.stash(path) logger.debug('Removing file or directory %s', path) for pth in self.pth.values(): pth.remove() logger.info('Successfully uninstalled %s', dist_name_version) def _allowed_to_proceed(self, verbose): # type: (bool) -> bool """Display which files would be deleted and prompt for confirmation """ def _display(msg, paths): # type: (str, Iterable[str]) -> None if not paths: return logger.info(msg) with indent_log(): for path in sorted(compact(paths)): logger.info(path) if not verbose: will_remove, will_skip = compress_for_output_listing(self.paths) else: # In verbose mode, display all the files that are going to be # deleted. will_remove = set(self.paths) will_skip = set() _display('Would remove:', will_remove) _display('Would not remove (might be manually added):', will_skip) _display('Would not remove (outside of prefix):', self._refuse) if verbose: _display('Will actually move:', compress_for_rename(self.paths)) return ask('Proceed (y/n)? ', ('y', 'n')) == 'y' def rollback(self): # type: () -> None """Rollback the changes previously made by remove().""" if not self._moved_paths.can_rollback: logger.error( "Can't roll back %s; was not uninstalled", self.dist.project_name, ) return logger.info('Rolling back uninstall of %s', self.dist.project_name) self._moved_paths.rollback() for pth in self.pth.values(): pth.rollback() def commit(self): # type: () -> None """Remove temporary save dir: rollback will no longer be possible.""" self._moved_paths.commit() @classmethod def from_dist(cls, dist): # type: (Distribution) -> UninstallPathSet dist_path = normalize_path(dist.location) if not dist_is_local(dist): logger.info( "Not uninstalling %s at %s, outside environment %s", dist.key, dist_path, sys.prefix, ) return cls(dist) if dist_path in {p for p in {sysconfig.get_path("stdlib"), sysconfig.get_path("platstdlib")} if p}: logger.info( "Not uninstalling %s at %s, as it is in the standard library.", dist.key, dist_path, ) return cls(dist) paths_to_remove = cls(dist) develop_egg_link = egg_link_path(dist) develop_egg_link_egg_info = '{}.egg-info'.format( pkg_resources.to_filename(dist.project_name)) egg_info_exists = dist.egg_info and os.path.exists(dist.egg_info) # Special case for distutils installed package distutils_egg_info = getattr(dist._provider, 'path', None) # Uninstall cases order do matter as in the case of 2 installs of the # same package, pip needs to uninstall the currently detected version if (egg_info_exists and dist.egg_info.endswith('.egg-info') and not dist.egg_info.endswith(develop_egg_link_egg_info)): # if dist.egg_info.endswith(develop_egg_link_egg_info), we # are in fact in the develop_egg_link case paths_to_remove.add(dist.egg_info) if dist.has_metadata('installed-files.txt'): for installed_file in dist.get_metadata( 'installed-files.txt').splitlines(): path = os.path.normpath( os.path.join(dist.egg_info, installed_file) ) paths_to_remove.add(path) # FIXME: need a test for this elif block # occurs with --single-version-externally-managed/--record outside # of pip elif dist.has_metadata('top_level.txt'): if dist.has_metadata('namespace_packages.txt'): namespaces = dist.get_metadata('namespace_packages.txt') else: namespaces = [] for top_level_pkg in [ p for p in dist.get_metadata('top_level.txt').splitlines() if p and p not in namespaces]: path = os.path.join(dist.location, top_level_pkg) paths_to_remove.add(path) paths_to_remove.add(path + '.py') paths_to_remove.add(path + '.pyc') paths_to_remove.add(path + '.pyo') elif distutils_egg_info: raise UninstallationError( "Cannot uninstall {!r}. It is a distutils installed project " "and thus we cannot accurately determine which files belong " "to it which would lead to only a partial uninstall.".format( dist.project_name, ) ) elif dist.location.endswith('.egg'): # package installed by easy_install # We cannot match on dist.egg_name because it can slightly vary # i.e. setuptools-0.6c11-py2.6.egg vs setuptools-0.6rc11-py2.6.egg paths_to_remove.add(dist.location) easy_install_egg = os.path.split(dist.location)[1] easy_install_pth = os.path.join(os.path.dirname(dist.location), 'easy-install.pth') paths_to_remove.add_pth(easy_install_pth, './' + easy_install_egg) elif egg_info_exists and dist.egg_info.endswith('.dist-info'): for path in uninstallation_paths(dist): paths_to_remove.add(path) elif develop_egg_link: # develop egg with open(develop_egg_link, 'r') as fh: link_pointer = os.path.normcase(fh.readline().strip()) assert (link_pointer == dist.location), ( 'Egg-link %s does not match installed location of %s ' '(at %s)' % (link_pointer, dist.project_name, dist.location) ) paths_to_remove.add(develop_egg_link) easy_install_pth = os.path.join(os.path.dirname(develop_egg_link), 'easy-install.pth') paths_to_remove.add_pth(easy_install_pth, dist.location) else: logger.debug( 'Not sure how to uninstall: %s - Check: %s', dist, dist.location, ) # find distutils scripts= scripts if dist.has_metadata('scripts') and dist.metadata_isdir('scripts'): for script in dist.metadata_listdir('scripts'): if dist_in_usersite(dist): bin_dir = bin_user else: bin_dir = bin_py paths_to_remove.add(os.path.join(bin_dir, script)) if WINDOWS: paths_to_remove.add(os.path.join(bin_dir, script) + '.bat') # find console_scripts _scripts_to_remove = [] console_scripts = dist.get_entry_map(group='console_scripts') for name in console_scripts.keys(): _scripts_to_remove.extend(_script_names(dist, name, False)) # find gui_scripts gui_scripts = dist.get_entry_map(group='gui_scripts') for name in gui_scripts.keys(): _scripts_to_remove.extend(_script_names(dist, name, True)) for s in _scripts_to_remove: paths_to_remove.add(s) return paths_to_remove class UninstallPthEntries(object): def __init__(self, pth_file): # type: (str) -> None if not os.path.isfile(pth_file): raise UninstallationError( "Cannot remove entries from nonexistent file %s" % pth_file ) self.file = pth_file self.entries = set() # type: Set[str] self._saved_lines = None # type: Optional[List[bytes]] def add(self, entry): # type: (str) -> None entry = os.path.normcase(entry) # On Windows, os.path.normcase converts the entry to use # backslashes. This is correct for entries that describe absolute # paths outside of site-packages, but all the others use forward # slashes. # os.path.splitdrive is used instead of os.path.isabs because isabs # treats non-absolute paths with drive letter markings like c:foo\bar # as absolute paths. It also does not recognize UNC paths if they don't # have more than "\\sever\share". Valid examples: "\\server\share\" or # "\\server\share\folder". Python 2.7.8+ support UNC in splitdrive. if WINDOWS and not os.path.splitdrive(entry)[0]: entry = entry.replace('\\', '/') self.entries.add(entry) def remove(self): # type: () -> None logger.debug('Removing pth entries from %s:', self.file) with open(self.file, 'rb') as fh: # windows uses '\r\n' with py3k, but uses '\n' with py2.x lines = fh.readlines() self._saved_lines = lines if any(b'\r\n' in line for line in lines): endline = '\r\n' else: endline = '\n' # handle missing trailing newline if lines and not lines[-1].endswith(endline.encode("utf-8")): lines[-1] = lines[-1] + endline.encode("utf-8") for entry in self.entries: try: logger.debug('Removing entry: %s', entry) lines.remove((entry + endline).encode("utf-8")) except ValueError: pass with open(self.file, 'wb') as fh: fh.writelines(lines) def rollback(self): # type: () -> bool if self._saved_lines is None: logger.error( 'Cannot roll back changes to %s, none were made', self.file ) return False logger.debug('Rolling %s back to previous state', self.file) with open(self.file, 'wb') as fh: fh.writelines(self._saved_lines) return True site-packages/pip/_internal/req/__init__.py000064400000004643151464452710014747 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False from __future__ import absolute_import import logging from pip._internal.utils.logging import indent_log from pip._internal.utils.typing import MYPY_CHECK_RUNNING from .req_file import parse_requirements from .req_install import InstallRequirement from .req_set import RequirementSet if MYPY_CHECK_RUNNING: from typing import Any, List, Sequence __all__ = [ "RequirementSet", "InstallRequirement", "parse_requirements", "install_given_reqs", ] logger = logging.getLogger(__name__) def install_given_reqs( to_install, # type: List[InstallRequirement] install_options, # type: List[str] global_options=(), # type: Sequence[str] *args, # type: Any **kwargs # type: Any ): # type: (...) -> List[InstallRequirement] """ Install everything in the given list. (to be called after having downloaded and unpacked the packages) """ if to_install: logger.info( 'Installing collected packages: %s', ', '.join([req.name for req in to_install]), ) with indent_log(): for requirement in to_install: if requirement.conflicts_with: logger.info( 'Found existing installation: %s', requirement.conflicts_with, ) with indent_log(): uninstalled_pathset = requirement.uninstall( auto_confirm=True ) try: requirement.install( install_options, global_options, *args, **kwargs ) except Exception: should_rollback = ( requirement.conflicts_with and not requirement.install_succeeded ) # if install did not succeed, rollback previous uninstall if should_rollback: uninstalled_pathset.rollback() raise else: should_commit = ( requirement.conflicts_with and requirement.install_succeeded ) if should_commit: uninstalled_pathset.commit() requirement.remove_temporary_source() return to_install site-packages/pip/_internal/legacy_resolve.py000064400000041522151464452710015421 0ustar00"""Dependency Resolution The dependency resolution in pip is performed as follows: for top-level requirements: a. only one spec allowed per project, regardless of conflicts or not. otherwise a "double requirement" exception is raised b. they override sub-dependency requirements. for sub-dependencies a. "first found, wins" (where the order is breadth first) """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False import logging import sys from collections import defaultdict from itertools import chain from pip._vendor.packaging import specifiers from pip._internal.exceptions import ( BestVersionAlreadyInstalled, DistributionNotFound, HashError, HashErrors, UnsupportedPythonVersion, ) from pip._internal.utils.logging import indent_log from pip._internal.utils.misc import ( dist_in_install_path, dist_in_usersite, ensure_dir, normalize_version_info, ) from pip._internal.utils.packaging import ( check_requires_python, get_requires_python, ) from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Callable, DefaultDict, List, Optional, Set, Tuple from pip._vendor import pkg_resources from pip._internal.distributions import AbstractDistribution from pip._internal.network.session import PipSession from pip._internal.index import PackageFinder from pip._internal.operations.prepare import RequirementPreparer from pip._internal.req.req_install import InstallRequirement from pip._internal.req.req_set import RequirementSet InstallRequirementProvider = Callable[ [str, InstallRequirement], InstallRequirement ] logger = logging.getLogger(__name__) def _check_dist_requires_python( dist, # type: pkg_resources.Distribution version_info, # type: Tuple[int, int, int] ignore_requires_python=False, # type: bool ): # type: (...) -> None """ Check whether the given Python version is compatible with a distribution's "Requires-Python" value. :param version_info: A 3-tuple of ints representing the Python major-minor-micro version to check. :param ignore_requires_python: Whether to ignore the "Requires-Python" value if the given Python version isn't compatible. :raises UnsupportedPythonVersion: When the given Python version isn't compatible. """ requires_python = get_requires_python(dist) try: is_compatible = check_requires_python( requires_python, version_info=version_info, ) except specifiers.InvalidSpecifier as exc: logger.warning( "Package %r has an invalid Requires-Python: %s", dist.project_name, exc, ) return if is_compatible: return version = '.'.join(map(str, version_info)) if ignore_requires_python: logger.debug( 'Ignoring failed Requires-Python check for package %r: ' '%s not in %r', dist.project_name, version, requires_python, ) return raise UnsupportedPythonVersion( 'Package {!r} requires a different Python: {} not in {!r}'.format( dist.project_name, version, requires_python, )) class Resolver(object): """Resolves which packages need to be installed/uninstalled to perform \ the requested operation without breaking the requirements of any package. """ _allowed_strategies = {"eager", "only-if-needed", "to-satisfy-only"} def __init__( self, preparer, # type: RequirementPreparer session, # type: PipSession finder, # type: PackageFinder make_install_req, # type: InstallRequirementProvider use_user_site, # type: bool ignore_dependencies, # type: bool ignore_installed, # type: bool ignore_requires_python, # type: bool force_reinstall, # type: bool upgrade_strategy, # type: str py_version_info=None, # type: Optional[Tuple[int, ...]] ): # type: (...) -> None super(Resolver, self).__init__() assert upgrade_strategy in self._allowed_strategies if py_version_info is None: py_version_info = sys.version_info[:3] else: py_version_info = normalize_version_info(py_version_info) self._py_version_info = py_version_info self.preparer = preparer self.finder = finder self.session = session # This is set in resolve self.require_hashes = None # type: Optional[bool] self.upgrade_strategy = upgrade_strategy self.force_reinstall = force_reinstall self.ignore_dependencies = ignore_dependencies self.ignore_installed = ignore_installed self.ignore_requires_python = ignore_requires_python self.use_user_site = use_user_site self._make_install_req = make_install_req self._discovered_dependencies = \ defaultdict(list) # type: DefaultDict[str, List] def resolve(self, requirement_set): # type: (RequirementSet) -> None """Resolve what operations need to be done As a side-effect of this method, the packages (and their dependencies) are downloaded, unpacked and prepared for installation. This preparation is done by ``pip.operations.prepare``. Once PyPI has static dependency metadata available, it would be possible to move the preparation to become a step separated from dependency resolution. """ # make the wheelhouse if self.preparer.wheel_download_dir: ensure_dir(self.preparer.wheel_download_dir) # If any top-level requirement has a hash specified, enter # hash-checking mode, which requires hashes from all. root_reqs = ( requirement_set.unnamed_requirements + list(requirement_set.requirements.values()) ) self.require_hashes = ( requirement_set.require_hashes or any(req.has_hash_options for req in root_reqs) ) # Display where finder is looking for packages search_scope = self.finder.search_scope locations = search_scope.get_formatted_locations() if locations: logger.info(locations) # Actually prepare the files, and collect any exceptions. Most hash # exceptions cannot be checked ahead of time, because # req.populate_link() needs to be called before we can make decisions # based on link type. discovered_reqs = [] # type: List[InstallRequirement] hash_errors = HashErrors() for req in chain(root_reqs, discovered_reqs): try: discovered_reqs.extend( self._resolve_one(requirement_set, req) ) except HashError as exc: exc.req = req hash_errors.append(exc) if hash_errors: raise hash_errors def _is_upgrade_allowed(self, req): # type: (InstallRequirement) -> bool if self.upgrade_strategy == "to-satisfy-only": return False elif self.upgrade_strategy == "eager": return True else: assert self.upgrade_strategy == "only-if-needed" return req.is_direct def _set_req_to_reinstall(self, req): # type: (InstallRequirement) -> None """ Set a requirement to be installed. """ # Don't uninstall the conflict if doing a user install and the # conflict is not a user install. if ((not self.use_user_site or dist_in_usersite(req.satisfied_by)) and dist_in_install_path(req.satisfied_by)): req.conflicts_with = req.satisfied_by req.satisfied_by = None def _check_skip_installed(self, req_to_install): # type: (InstallRequirement) -> Optional[str] """Check if req_to_install should be skipped. This will check if the req is installed, and whether we should upgrade or reinstall it, taking into account all the relevant user options. After calling this req_to_install will only have satisfied_by set to None if the req_to_install is to be upgraded/reinstalled etc. Any other value will be a dist recording the current thing installed that satisfies the requirement. Note that for vcs urls and the like we can't assess skipping in this routine - we simply identify that we need to pull the thing down, then later on it is pulled down and introspected to assess upgrade/ reinstalls etc. :return: A text reason for why it was skipped, or None. """ if self.ignore_installed: return None req_to_install.check_if_exists(self.use_user_site) if not req_to_install.satisfied_by: return None if self.force_reinstall: self._set_req_to_reinstall(req_to_install) return None if not self._is_upgrade_allowed(req_to_install): if self.upgrade_strategy == "only-if-needed": return 'already satisfied, skipping upgrade' return 'already satisfied' # Check for the possibility of an upgrade. For link-based # requirements we have to pull the tree down and inspect to assess # the version #, so it's handled way down. if not req_to_install.link: try: self.finder.find_requirement(req_to_install, upgrade=True) except BestVersionAlreadyInstalled: # Then the best version is installed. return 'already up-to-date' except DistributionNotFound: # No distribution found, so we squash the error. It will # be raised later when we re-try later to do the install. # Why don't we just raise here? pass self._set_req_to_reinstall(req_to_install) return None def _get_abstract_dist_for(self, req): # type: (InstallRequirement) -> AbstractDistribution """Takes a InstallRequirement and returns a single AbstractDist \ representing a prepared variant of the same. """ assert self.require_hashes is not None, ( "require_hashes should have been set in Resolver.resolve()" ) if req.editable: return self.preparer.prepare_editable_requirement( req, self.require_hashes, self.use_user_site, self.finder, ) # satisfied_by is only evaluated by calling _check_skip_installed, # so it must be None here. assert req.satisfied_by is None skip_reason = self._check_skip_installed(req) if req.satisfied_by: return self.preparer.prepare_installed_requirement( req, self.require_hashes, skip_reason ) upgrade_allowed = self._is_upgrade_allowed(req) # We eagerly populate the link, since that's our "legacy" behavior. req.populate_link(self.finder, upgrade_allowed, self.require_hashes) abstract_dist = self.preparer.prepare_linked_requirement( req, self.session, self.finder, self.require_hashes ) # NOTE # The following portion is for determining if a certain package is # going to be re-installed/upgraded or not and reporting to the user. # This should probably get cleaned up in a future refactor. # req.req is only avail after unpack for URL # pkgs repeat check_if_exists to uninstall-on-upgrade # (#14) if not self.ignore_installed: req.check_if_exists(self.use_user_site) if req.satisfied_by: should_modify = ( self.upgrade_strategy != "to-satisfy-only" or self.force_reinstall or self.ignore_installed or req.link.scheme == 'file' ) if should_modify: self._set_req_to_reinstall(req) else: logger.info( 'Requirement already satisfied (use --upgrade to upgrade):' ' %s', req, ) return abstract_dist def _resolve_one( self, requirement_set, # type: RequirementSet req_to_install # type: InstallRequirement ): # type: (...) -> List[InstallRequirement] """Prepare a single requirements file. :return: A list of additional InstallRequirements to also install. """ # Tell user what we are doing for this requirement: # obtain (editable), skipping, processing (local url), collecting # (remote url or package name) if req_to_install.constraint or req_to_install.prepared: return [] req_to_install.prepared = True # register tmp src for cleanup in case something goes wrong requirement_set.reqs_to_cleanup.append(req_to_install) abstract_dist = self._get_abstract_dist_for(req_to_install) # Parse and return dependencies dist = abstract_dist.get_pkg_resources_distribution() # This will raise UnsupportedPythonVersion if the given Python # version isn't compatible with the distribution's Requires-Python. _check_dist_requires_python( dist, version_info=self._py_version_info, ignore_requires_python=self.ignore_requires_python, ) more_reqs = [] # type: List[InstallRequirement] def add_req(subreq, extras_requested): sub_install_req = self._make_install_req( str(subreq), req_to_install, ) parent_req_name = req_to_install.name to_scan_again, add_to_parent = requirement_set.add_requirement( sub_install_req, parent_req_name=parent_req_name, extras_requested=extras_requested, ) if parent_req_name and add_to_parent: self._discovered_dependencies[parent_req_name].append( add_to_parent ) more_reqs.extend(to_scan_again) with indent_log(): # We add req_to_install before its dependencies, so that we # can refer to it when adding dependencies. if not requirement_set.has_requirement(req_to_install.name): # 'unnamed' requirements will get added here req_to_install.is_direct = True requirement_set.add_requirement( req_to_install, parent_req_name=None, ) if not self.ignore_dependencies: if req_to_install.extras: logger.debug( "Installing extra requirements: %r", ','.join(req_to_install.extras), ) missing_requested = sorted( set(req_to_install.extras) - set(dist.extras) ) for missing in missing_requested: logger.warning( '%s does not provide the extra \'%s\'', dist, missing ) available_requested = sorted( set(dist.extras) & set(req_to_install.extras) ) for subreq in dist.requires(available_requested): add_req(subreq, extras_requested=available_requested) if not req_to_install.editable and not req_to_install.satisfied_by: # XXX: --no-install leads this to report 'Successfully # downloaded' for only non-editable reqs, even though we took # action on them. requirement_set.successfully_downloaded.append(req_to_install) return more_reqs def get_installation_order(self, req_set): # type: (RequirementSet) -> List[InstallRequirement] """Create the installation order. The installation order is topological - requirements are installed before the requiring thing. We break cycles at an arbitrary point, and make no other guarantees. """ # The current implementation, which we may change at any point # installs the user specified things in the order given, except when # dependencies must come earlier to achieve topological order. order = [] ordered_reqs = set() # type: Set[InstallRequirement] def schedule(req): if req.satisfied_by or req in ordered_reqs: return if req.constraint: return ordered_reqs.add(req) for dep in self._discovered_dependencies[req.name]: schedule(dep) order.append(req) for install_req in req_set.requirements.values(): schedule(install_req) return order site-packages/pip/_internal/build_env.py000064400000016535151464452710014373 0ustar00"""Build Environment used for isolation during sdist building """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False import logging import os import sys import textwrap from collections import OrderedDict from distutils.sysconfig import get_python_lib from sysconfig import get_paths from pip._vendor.pkg_resources import Requirement, VersionConflict, WorkingSet from pip import __file__ as pip_location from pip._internal.utils.subprocess import call_subprocess from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.ui import open_spinner if MYPY_CHECK_RUNNING: from typing import Tuple, Set, Iterable, Optional, List from pip._internal.index import PackageFinder logger = logging.getLogger(__name__) class _Prefix: def __init__(self, path): # type: (str) -> None self.path = path self.setup = False self.bin_dir = get_paths( 'nt' if os.name == 'nt' else 'posix_prefix', vars={'base': path, 'platbase': path} )['scripts'] # Note: prefer distutils' sysconfig to get the # library paths so PyPy is correctly supported. purelib = get_python_lib(plat_specific=False, prefix=path) platlib = get_python_lib(plat_specific=True, prefix=path) if purelib == platlib: self.lib_dirs = [purelib] else: self.lib_dirs = [purelib, platlib] class BuildEnvironment(object): """Creates and manages an isolated environment to install build deps """ def __init__(self): # type: () -> None self._temp_dir = TempDirectory(kind="build-env") self._prefixes = OrderedDict(( (name, _Prefix(os.path.join(self._temp_dir.path, name))) for name in ('normal', 'overlay') )) self._bin_dirs = [] # type: List[str] self._lib_dirs = [] # type: List[str] for prefix in reversed(list(self._prefixes.values())): self._bin_dirs.append(prefix.bin_dir) self._lib_dirs.extend(prefix.lib_dirs) # Customize site to: # - ensure .pth files are honored # - prevent access to system site packages system_sites = { os.path.normcase(site) for site in ( get_python_lib(plat_specific=False), get_python_lib(plat_specific=True), ) } self._site_dir = os.path.join(self._temp_dir.path, 'site') if not os.path.exists(self._site_dir): os.mkdir(self._site_dir) with open(os.path.join(self._site_dir, 'sitecustomize.py'), 'w') as fp: fp.write(textwrap.dedent( ''' import os, site, sys # First, drop system-sites related paths. original_sys_path = sys.path[:] known_paths = set() for path in {system_sites!r}: site.addsitedir(path, known_paths=known_paths) system_paths = set( os.path.normcase(path) for path in sys.path[len(original_sys_path):] ) original_sys_path = [ path for path in original_sys_path if os.path.normcase(path) not in system_paths ] sys.path = original_sys_path # Second, add lib directories. # ensuring .pth file are processed. for path in {lib_dirs!r}: assert not path in sys.path site.addsitedir(path) ''' ).format(system_sites=system_sites, lib_dirs=self._lib_dirs)) def __enter__(self): self._save_env = { name: os.environ.get(name, None) for name in ('PATH', 'PYTHONNOUSERSITE', 'PYTHONPATH') } path = self._bin_dirs[:] old_path = self._save_env['PATH'] if old_path: path.extend(old_path.split(os.pathsep)) pythonpath = [self._site_dir] os.environ.update({ 'PATH': os.pathsep.join(path), 'PYTHONNOUSERSITE': '1', 'PYTHONPATH': os.pathsep.join(pythonpath), }) def __exit__(self, exc_type, exc_val, exc_tb): for varname, old_value in self._save_env.items(): if old_value is None: os.environ.pop(varname, None) else: os.environ[varname] = old_value def cleanup(self): # type: () -> None self._temp_dir.cleanup() def check_requirements(self, reqs): # type: (Iterable[str]) -> Tuple[Set[Tuple[str, str]], Set[str]] """Return 2 sets: - conflicting requirements: set of (installed, wanted) reqs tuples - missing requirements: set of reqs """ missing = set() conflicting = set() if reqs: ws = WorkingSet(self._lib_dirs) for req in reqs: try: if ws.find(Requirement.parse(req)) is None: missing.add(req) except VersionConflict as e: conflicting.add((str(e.args[0].as_requirement()), str(e.args[1]))) return conflicting, missing def install_requirements( self, finder, # type: PackageFinder requirements, # type: Iterable[str] prefix_as_string, # type: str message # type: Optional[str] ): # type: (...) -> None prefix = self._prefixes[prefix_as_string] assert not prefix.setup prefix.setup = True if not requirements: return args = [ sys.executable, os.path.dirname(pip_location), 'install', '--ignore-installed', '--no-user', '--prefix', prefix.path, '--no-warn-script-location', ] # type: List[str] if logger.getEffectiveLevel() <= logging.DEBUG: args.append('-v') for format_control in ('no_binary', 'only_binary'): formats = getattr(finder.format_control, format_control) args.extend(('--' + format_control.replace('_', '-'), ','.join(sorted(formats or {':none:'})))) index_urls = finder.index_urls if index_urls: args.extend(['-i', index_urls[0]]) for extra_index in index_urls[1:]: args.extend(['--extra-index-url', extra_index]) else: args.append('--no-index') for link in finder.find_links: args.extend(['--find-links', link]) for host in finder.trusted_hosts: args.extend(['--trusted-host', host]) if finder.allow_all_prereleases: args.append('--pre') args.append('--') args.extend(requirements) with open_spinner(message) as spinner: call_subprocess(args, spinner=spinner) class NoOpBuildEnvironment(BuildEnvironment): """A no-op drop-in replacement for BuildEnvironment """ def __init__(self): pass def __enter__(self): pass def __exit__(self, exc_type, exc_val, exc_tb): pass def cleanup(self): pass def install_requirements(self, finder, requirements, prefix, message): raise NotImplementedError() site-packages/pip/_internal/network/cache.py000064400000004271151464452710015152 0ustar00"""HTTP cache implementation. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import os from contextlib import contextmanager from pip._vendor.cachecontrol.cache import BaseCache from pip._vendor.cachecontrol.caches import FileCache from pip._internal.utils.filesystem import adjacent_tmp_file, replace from pip._internal.utils.misc import ensure_dir from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional @contextmanager def suppressed_cache_errors(): """If we can't access the cache then we can just skip caching and process requests as if caching wasn't enabled. """ try: yield except (OSError, IOError): pass class SafeFileCache(BaseCache): """ A file based cache which is safe to use even when the target directory may not be accessible or writable. """ def __init__(self, directory): # type: (str) -> None assert directory is not None, "Cache directory must not be None." super(SafeFileCache, self).__init__() self.directory = directory def _get_cache_path(self, name): # type: (str) -> str # From cachecontrol.caches.file_cache.FileCache._fn, brought into our # class for backwards-compatibility and to avoid using a non-public # method. hashed = FileCache.encode(name) parts = list(hashed[:5]) + [hashed] return os.path.join(self.directory, *parts) def get(self, key): # type: (str) -> Optional[bytes] path = self._get_cache_path(key) with suppressed_cache_errors(): with open(path, 'rb') as f: return f.read() def set(self, key, value): # type: (str, bytes) -> None path = self._get_cache_path(key) with suppressed_cache_errors(): ensure_dir(os.path.dirname(path)) with adjacent_tmp_file(path) as f: f.write(value) replace(f.name, path) def delete(self, key): # type: (str) -> None path = self._get_cache_path(key) with suppressed_cache_errors(): os.remove(path) site-packages/pip/_internal/network/xmlrpc.py000064400000003075151464452710015415 0ustar00"""xmlrpclib.Transport implementation """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import logging from pip._vendor import requests # NOTE: XMLRPC Client is not annotated in typeshed as on 2017-07-17, which is # why we ignore the type on this import from pip._vendor.six.moves import xmlrpc_client # type: ignore from pip._vendor.six.moves.urllib import parse as urllib_parse logger = logging.getLogger(__name__) class PipXmlrpcTransport(xmlrpc_client.Transport): """Provide a `xmlrpclib.Transport` implementation via a `PipSession` object. """ def __init__(self, index_url, session, use_datetime=False): xmlrpc_client.Transport.__init__(self, use_datetime) index_parts = urllib_parse.urlparse(index_url) self._scheme = index_parts.scheme self._session = session def request(self, host, handler, request_body, verbose=False): parts = (self._scheme, host, handler, None, None, None) url = urllib_parse.urlunparse(parts) try: headers = {'Content-Type': 'text/xml'} response = self._session.post(url, data=request_body, headers=headers, stream=True) response.raise_for_status() self.verbose = verbose return self.parse_response(response.raw) except requests.HTTPError as exc: logger.critical( "HTTP error %s while getting %s", exc.response.status_code, url, ) raise site-packages/pip/_internal/network/auth.py000064400000025557151464452710015062 0ustar00"""Network Authentication Helpers Contains interface (MultiDomainBasicAuth) and associated glue code for providing credentials in the context of network requests. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import logging from pip._vendor.requests.auth import AuthBase, HTTPBasicAuth from pip._vendor.requests.utils import get_netrc_auth from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.utils.misc import ( ask, ask_input, ask_password, remove_auth_from_url, split_auth_netloc_from_url, ) from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from optparse import Values from typing import Dict, Optional, Tuple from pip._internal.vcs.versioncontrol import AuthInfo Credentials = Tuple[str, str, str] logger = logging.getLogger(__name__) try: import keyring # noqa except ImportError: keyring = None except Exception as exc: logger.warning( "Keyring is skipped due to an exception: %s", str(exc), ) keyring = None def get_keyring_auth(url, username): """Return the tuple auth for a given url from keyring.""" if not url or not keyring: return None try: try: get_credential = keyring.get_credential except AttributeError: pass else: logger.debug("Getting credentials from keyring for %s", url) cred = get_credential(url, username) if cred is not None: return cred.username, cred.password return None if username: logger.debug("Getting password from keyring for %s", url) password = keyring.get_password(url, username) if password: return username, password except Exception as exc: logger.warning( "Keyring is skipped due to an exception: %s", str(exc), ) class MultiDomainBasicAuth(AuthBase): def __init__(self, prompting=True, index_urls=None): # type: (bool, Optional[Values]) -> None self.prompting = prompting self.index_urls = index_urls self.passwords = {} # type: Dict[str, AuthInfo] # When the user is prompted to enter credentials and keyring is # available, we will offer to save them. If the user accepts, # this value is set to the credentials they entered. After the # request authenticates, the caller should call # ``save_credentials`` to save these. self._credentials_to_save = None # type: Optional[Credentials] def _get_index_url(self, url): """Return the original index URL matching the requested URL. Cached or dynamically generated credentials may work against the original index URL rather than just the netloc. The provided url should have had its username and password removed already. If the original index url had credentials then they will be included in the return value. Returns None if no matching index was found, or if --no-index was specified by the user. """ if not url or not self.index_urls: return None for u in self.index_urls: prefix = remove_auth_from_url(u).rstrip("/") + "/" if url.startswith(prefix): return u def _get_new_credentials(self, original_url, allow_netrc=True, allow_keyring=True): """Find and return credentials for the specified URL.""" # Split the credentials and netloc from the url. url, netloc, url_user_password = split_auth_netloc_from_url( original_url, ) # Start with the credentials embedded in the url username, password = url_user_password if username is not None and password is not None: logger.debug("Found credentials in url for %s", netloc) return url_user_password # Find a matching index url for this request index_url = self._get_index_url(url) if index_url: # Split the credentials from the url. index_info = split_auth_netloc_from_url(index_url) if index_info: index_url, _, index_url_user_password = index_info logger.debug("Found index url %s", index_url) # If an index URL was found, try its embedded credentials if index_url and index_url_user_password[0] is not None: username, password = index_url_user_password if username is not None and password is not None: logger.debug("Found credentials in index url for %s", netloc) return index_url_user_password # Get creds from netrc if we still don't have them if allow_netrc: netrc_auth = get_netrc_auth(original_url) if netrc_auth: logger.debug("Found credentials in netrc for %s", netloc) return netrc_auth # If we don't have a password and keyring is available, use it. if allow_keyring: # The index url is more specific than the netloc, so try it first kr_auth = ( get_keyring_auth(index_url, username) or get_keyring_auth(netloc, username) ) if kr_auth: logger.debug("Found credentials in keyring for %s", netloc) return kr_auth return username, password def _get_url_and_credentials(self, original_url): """Return the credentials to use for the provided URL. If allowed, netrc and keyring may be used to obtain the correct credentials. Returns (url_without_credentials, username, password). Note that even if the original URL contains credentials, this function may return a different username and password. """ url, netloc, _ = split_auth_netloc_from_url(original_url) # Use any stored credentials that we have for this netloc username, password = self.passwords.get(netloc, (None, None)) if username is None and password is None: # No stored credentials. Acquire new credentials without prompting # the user. (e.g. from netrc, keyring, or the URL itself) username, password = self._get_new_credentials(original_url) if username is not None or password is not None: # Convert the username and password if they're None, so that # this netloc will show up as "cached" in the conditional above. # Further, HTTPBasicAuth doesn't accept None, so it makes sense to # cache the value that is going to be used. username = username or "" password = password or "" # Store any acquired credentials. self.passwords[netloc] = (username, password) assert ( # Credentials were found (username is not None and password is not None) or # Credentials were not found (username is None and password is None) ), "Could not load credentials from url: {}".format(original_url) return url, username, password def __call__(self, req): # Get credentials for this request url, username, password = self._get_url_and_credentials(req.url) # Set the url of the request to the url without any credentials req.url = url if username is not None and password is not None: # Send the basic auth with this request req = HTTPBasicAuth(username, password)(req) # Attach a hook to handle 401 responses req.register_hook("response", self.handle_401) return req # Factored out to allow for easy patching in tests def _prompt_for_password(self, netloc): username = ask_input("User for %s: " % netloc) if not username: return None, None auth = get_keyring_auth(netloc, username) if auth: return auth[0], auth[1], False password = ask_password("Password: ") return username, password, True # Factored out to allow for easy patching in tests def _should_save_password_to_keyring(self): if not keyring: return False return ask("Save credentials to keyring [y/N]: ", ["y", "n"]) == "y" def handle_401(self, resp, **kwargs): # We only care about 401 responses, anything else we want to just # pass through the actual response if resp.status_code != 401: return resp # We are not able to prompt the user so simply return the response if not self.prompting: return resp parsed = urllib_parse.urlparse(resp.url) # Prompt the user for a new username and password username, password, save = self._prompt_for_password(parsed.netloc) # Store the new username and password to use for future requests self._credentials_to_save = None if username is not None and password is not None: self.passwords[parsed.netloc] = (username, password) # Prompt to save the password to keyring if save and self._should_save_password_to_keyring(): self._credentials_to_save = (parsed.netloc, username, password) # Consume content and release the original connection to allow our new # request to reuse the same one. resp.content resp.raw.release_conn() # Add our new username and password to the request req = HTTPBasicAuth(username or "", password or "")(resp.request) req.register_hook("response", self.warn_on_401) # On successful request, save the credentials that were used to # keyring. (Note that if the user responded "no" above, this member # is not set and nothing will be saved.) if self._credentials_to_save: req.register_hook("response", self.save_credentials) # Send our new request new_resp = resp.connection.send(req, **kwargs) new_resp.history.append(resp) return new_resp def warn_on_401(self, resp, **kwargs): """Response callback to warn about incorrect credentials.""" if resp.status_code == 401: logger.warning( '401 Error, Credentials not correct for %s', resp.request.url, ) def save_credentials(self, resp, **kwargs): """Response callback to save credentials on success.""" assert keyring is not None, "should never reach here without keyring" if not keyring: return creds = self._credentials_to_save self._credentials_to_save = None if creds and resp.status_code < 400: try: logger.info('Saving credentials to keyring') keyring.set_password(*creds) except Exception: logger.exception('Failed to save credentials') site-packages/pip/_internal/network/__pycache__/cache.cpython-38.opt-1.pyc000064400000004566151464452710022406 0ustar00U .e@sdZddlZddlmZddlmZddlmZddlm Z m Z ddl m Z ddl mZerhdd lmZed d ZGd d d eZdS)zHTTP cache implementation. N)contextmanager) BaseCache) FileCache)adjacent_tmp_filereplace) ensure_dir)MYPY_CHECK_RUNNING)Optionalc cs(z dVWnttfk r"YnXdS)zvIf we can't access the cache then we can just skip caching and process requests as if caching wasn't enabled. N)OSErrorIOErrorr r ?/usr/lib/python3.8/site-packages/pip/_internal/network/cache.pysuppressed_cache_errorss rcs@eZdZdZfddZddZddZdd Zd d ZZ S) SafeFileCachezw A file based cache which is safe to use even when the target directory may not be accessible or writable. cstt|||_dSN)superr__init__ directory)selfr __class__r r r&szSafeFileCache.__init__cCs4t|}t|dd|g}tjj|jf|S)N)rencodelistospathjoinr)rnameZhashedpartsr r r _get_cache_path,s zSafeFileCache._get_cache_pathc CsR||}t8t|d"}|W5QRW5QRSQRXW5QRXdS)Nrb)rropenread)rkeyrfr r r get5s  zSafeFileCache.getc CsZ||}t@ttj|t|}||W5QRXt|j |W5QRXdSr) rrrrrdirnamerwriterr)rr#valuerr$r r r set<s   zSafeFileCache.setc Cs*||}tt|W5QRXdSr)rrrremove)rr#rr r r deleteGs zSafeFileCache.delete) __name__ __module__ __qualname____doc__rrr%r)r+ __classcell__r r rr r s    r)r/r contextlibrZpip._vendor.cachecontrol.cacherZpip._vendor.cachecontrol.cachesrZpip._internal.utils.filesystemrrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingr rrr r r r s       site-packages/pip/_internal/network/__pycache__/session.cpython-38.opt-1.pyc000064400000022012151464452710023010 0ustar00U .e=@sdZddlZddlZddlZddlZddlZddlZddlZddl Z ddl m Z m Z m Z ddlmZddlmZmZddlmZddlmZddlmZdd lmZdd lmZdd lmZdd l m!Z!dd l"m#Z#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+m,Z,m-Z-ddl.m/Z/ddl0m1Z1e/rhddl2m3Z3m4Z4m5Z5m6Z6m7Z7ddl8m9Z9e6e:e:e5e7e;e:ffZZ?e j@dedddddddgZAdZBddZCd d!ZDGd"d#d#eZEGd$d%d%eZFGd&d'd'e jGZHdS)(zhPipSession and supporting code, containing all pip-specific network request configuration and behavior. N)requestssixurllib3)CacheControlAdapter) BaseAdapter HTTPAdapter)Response)CaseInsensitiveDict)parse)InsecureRequestWarning) __version__)MultiDomainBasicAuth) SafeFileCache)HAS_TLS ipaddressssl)check_path_owner)libc_ver)build_url_from_netlocget_installed_version parse_netloc)MYPY_CHECK_RUNNING) url_to_path)IteratorListOptionalTupleUnion)Linkignore)category)Zhttps*r!)r!Z localhostr!)r!z 127.0.0.0/8r!)r!z::1/128r!)filer!N)Zsshr!r!)Z BUILD_BUILDIDZBUILD_IDZCIZ PIP_IS_CIcCstddtDS)z? Return whether it looks like pip is running under CI. css|]}|tjkVqdSN)osenviron).0namer(A/usr/lib/python3.8/site-packages/pip/_internal/network/session.py asz looks_like_ci..)anyCI_ENVIRONMENT_VARIABLESr(r(r(r) looks_like_ciYsr-cCsBdtdtdtid}|dddkr@t|dd<n|dddkrtjjd krltjd d }ntj}d d d|D|dd<nB|dddkrt|dd<n |dddkrt|dd<tjdrHddl m }t t ddt dddg|}t t ddt ddgt}|r:||d<|rH||d<tjdrztdrzdtdd|d<trt|did<trt|did<trt|d<trtj|d <td!}|d k r||d"<trd#nd |d$<tjd%}|d k r(||d&<d'j|tj|d(d#d)d*S)+z6 Return a string representing the user agent. pip)r'versionr')Z installerpythonimplementationr1ZCPythonr/ZPyPyfinalN.cSsg|] }t|qSr()str)r&xr(r(r) xszuser_agent..ZJythonZ IronPythonZlinuxr)distrocSs|dSNr(r6r(r(r)zuser_agent..idcSs|dSr9r(r;r(r(r)r<r=liblibcr8darwinZmacOSsystemreleaseZcpuZopenssl_versionZ setuptoolssetuptools_versionTZciZPIP_USER_AGENT_USER_DATA user_dataz9{data[installer][name]}/{data[installer][version]} {json}),:)Z separatorsZ sort_keys)datajson) r platformZpython_versionZpython_implementationsyspypy_version_info releaseleveljoin startswith pip._vendorr8dictfilterzipZlinux_distributionrZmac_verrB setdefaultrCmachinerrZOPENSSL_VERSIONrr-r$r%getformatrIdumps)rHrLr8Z distro_infosr@rDrEr(r(r) user_agentdsl           rYc@seZdZdddZddZdS)LocalFSAdapterNc Cst|j}t}d|_|j|_zt|} Wn.tk rZ} zd|_| |_W5d} ~ XYnPXtj j | j dd} t |dp~d} t| | j| d|_t|d|_|jj|_|S) NiT)Zusegmtrz text/plain)z Content-TypezContent-Lengthz Last-Modifiedrb)rurlrZ status_coder$statOSErrorrawemailZutilsZ formatdatest_mtime mimetypesZ guess_typer st_sizeheadersopenclose) selfrequeststreamtimeoutverifycertZproxiespathnameZrespZstatsexcZmodifiedZ content_typer(r(r)sends&    zLocalFSAdapter.sendcCsdSr#r()rhr(r(r)rgszLocalFSAdapter.close)NNNNN)__name__ __module__ __qualname__rprgr(r(r(r)rZs rZc@seZdZddZdS)InsecureHTTPAdaptercCsd|_d|_dS)NZ CERT_NONE)Z cert_reqsZca_certs)rhZconnr]rlrmr(r(r) cert_verifyszInsecureHTTPAdapter.cert_verifyN)rqrrrsrur(r(r(r)rtsrtcsFeZdZdZfddZd ddZddZd d Zfd d ZZ S) PipSessionNc s|dd}|dd}|dg}|dd}tt|j||g|_t|jd<t|d|_t j |d d d d gd d}|rt |st d|d}|rtt||d}n t|d}t|d}||_|d||d||dt|D]} |j| ddqdS)zj :param trusted_hosts: Domains not to emit warnings for when not using HTTPS. retriesrcacheN trusted_hosts index_urlsz User-Agent)rziiiig?)ZtotalZstatus_forcelistZbackoff_factorzThe directory '%s' or its parent directory is not owned by the current user and the cache has been disabled. Please check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag.)rx max_retries)r{zhttps://zhttp://zfile://T)suppress_logging)popsuperrv__init__pip_trusted_originsrYrer ZauthrZRetryrloggerwarningrrrrt_insecure_adaptermountrZadd_trusted_host) rhargskwargsrwrxryrzZsecure_adapterZinsecure_adapterhost __class__r(r)rs@            zPipSession.__init__FcCs|s.d|}|dk r$|d|7}t|t|}||jkrL|j||t|d|j|ds|t|d|jdS)z :param host: It is okay to provide a host that has previously been added. :param source: An optional source string, for logging where the host string came from. zadding trusted host: {!r}Nz (from {})/r:rG) rWrinforrappendrrr)rhrsourcer|msgZ host_portr(r(r)r3s     zPipSession.add_trusted_hostccs<tD] }|Vq|jD] \}}d||dkr.dn|fVqdS)Nr!)SECURE_ORIGINSr)rh secure_originrportr(r(r)iter_secure_originsMszPipSession.iter_secure_originsc Cs tt|}|j|j|j}}}|ddd}|D]}|\}}} ||kr\|dkr\qsV                O!site-packages/pip/_internal/network/__pycache__/__init__.cpython-38.pyc000064400000000324151464452710022127 0ustar00U .e2@sdZdS)z+Contains purely network-related utilities. N)__doc__rrB/usr/lib/python3.8/site-packages/pip/_internal/network/__init__.pysite-packages/pip/_internal/network/__pycache__/xmlrpc.cpython-38.pyc000064400000003046151464452710021701 0ustar00U .e=@sPdZddlZddlmZddlmZddlmZe e Z Gdddej Z dS)z#xmlrpclib.Transport implementation N)requests) xmlrpc_client)parsec@s$eZdZdZdddZd ddZdS) PipXmlrpcTransportzRProvide a `xmlrpclib.Transport` implementation via a `PipSession` object. FcCs*tj||t|}|j|_||_dS)N)r Transport__init__ urllib_parseZurlparseZscheme_scheme_session)selfZ index_urlZsessionZ use_datetimeZ index_partsr @/usr/lib/python3.8/site-packages/pip/_internal/network/xmlrpc.pyrs zPipXmlrpcTransport.__init__c Cs|j||dddf}t|}z8ddi}|jj|||dd}|||_||jWSt j k r} zt d| j j|W5d} ~ XYnXdS)Nz Content-Typeztext/xmlT)dataheadersstreamzHTTP error %s while getting %s)r rZ urlunparser ZpostZraise_for_statusverboseZparse_responserawrZ HTTPErrorloggerZcriticalresponseZ status_code) r ZhostZhandlerZ request_bodyrpartsZurlrrexcr r r requests$  zPipXmlrpcTransport.requestN)F)F)__name__ __module__ __qualname____doc__rrr r r r rs r)rZloggingZ pip._vendorrZpip._vendor.six.movesrZpip._vendor.six.moves.urllibrrZ getLoggerrrrrr r r r s     site-packages/pip/_internal/network/__pycache__/auth.cpython-38.opt-1.pyc000064400000015177151464452710022304 0ustar00U .eo+ @sdZddlZddlmZmZddlmZddlmZ ddl m Z m Z m Z mZmZddlmZerddlmZdd lmZmZmZdd lmZeeeefZeeZz ddlZWnLe k rdZYn6e!k rZ"ze#d ee"dZW5dZ"["XYnXd d Z$GdddeZ%dS)zNetwork Authentication Helpers Contains interface (MultiDomainBasicAuth) and associated glue code for providing credentials in the context of network requests. N)AuthBase HTTPBasicAuth)get_netrc_auth)parse)ask ask_input ask_passwordremove_auth_from_urlsplit_auth_netloc_from_url)MYPY_CHECK_RUNNING)Values)DictOptionalTuple)AuthInfo*Keyring is skipped due to an exception: %sc Cs|rts dSzz tj}Wntk r,Yn4Xtd||||}|dk rZ|j|jfWSWdS|rtd|t||}|r||fWSWn2tk r}zt dt |W5d}~XYnXdS)z3Return the tuple auth for a given url from keyring.Nz'Getting credentials from keyring for %sz$Getting password from keyring for %sr) keyringget_credentialAttributeErrorloggerdebugusernamepasswordZ get_password Exceptionwarningstr)urlrrZcredrexcr>/usr/lib/python3.8/site-packages/pip/_internal/network/auth.pyget_keyring_auth.s,     r c@s`eZdZdddZddZdddZd d Zd d Zd dZddZ ddZ ddZ ddZ dS)MultiDomainBasicAuthTNcCs||_||_i|_d|_dS)N) prompting index_urls passwords_credentials_to_save)selfr"r#rrr__init__MszMultiDomainBasicAuth.__init__cCsB|r |jsdS|jD](}t|dd}||r|SqdS)aReturn the original index URL matching the requested URL. Cached or dynamically generated credentials may work against the original index URL rather than just the netloc. The provided url should have had its username and password removed already. If the original index url had credentials then they will be included in the return value. Returns None if no matching index was found, or if --no-index was specified by the user. N/)r#r rstrip startswith)r&ruprefixrrr_get_index_urlYs   z#MultiDomainBasicAuth._get_index_urlcCst|\}}}|\}}|dk r6|dk r6td||S||} | rft| } | rf| \} } } td| | r| ddk r| \}}|dk r|dk rtd|| S|rt|} | rtd|| S|rt| |pt||}|rtd||S||fS)z2Find and return credentials for the specified URL.NzFound credentials in url for %szFound index url %srz%Found credentials in index url for %sz!Found credentials in netrc for %sz#Found credentials in keyring for %s)r rrr-rr )r& original_urlZ allow_netrcZ allow_keyringrnetlocZurl_user_passwordrrZ index_urlZ index_info_Zindex_url_user_passwordZ netrc_authZkr_authrrr_get_new_credentialsns>         z)MultiDomainBasicAuth._get_new_credentialscCsvt|\}}}|j|d\}}|dkr>|dkr>||\}}|dk sN|dk rl|pTd}|p\d}||f|j|<|||fS)a_Return the credentials to use for the provided URL. If allowed, netrc and keyring may be used to obtain the correct credentials. Returns (url_without_credentials, username, password). Note that even if the original URL contains credentials, this function may return a different username and password. NNN)r r$getr1)r&r.rr/r0rrrrr_get_url_and_credentialss  z-MultiDomainBasicAuth._get_url_and_credentialscCsH||j\}}}||_|dk r6|dk r6t|||}|d|j|S)Nresponse)r5rr register_hook handle_401)r&reqrrrrrr__call__s zMultiDomainBasicAuth.__call__cCsFtd|}|sdSt||}|r4|d|ddfStd}||dfS)Nz User for %s: r2rFz Password: T)rr r)r&r/rZauthrrrr_prompt_for_passwords  z)MultiDomainBasicAuth._prompt_for_passwordcCstsdStdddgdkS)NFz#Save credentials to keyring [y/N]: yn)rr)r&rrr _should_save_password_to_keyringsz5MultiDomainBasicAuth._should_save_password_to_keyringc Ks|jdkr|S|js|St|j}||j\}}}d|_|dk rv|dk rv||f|j|j<|rv| rv|j||f|_|j |j t |pd|pd|j}|d|j|jr|d|j|jj|f|}|j||S)Nr3r6) status_coder" urllib_parseZurlparserr<r/r%r$r?ZcontentrawZ release_connrrequestr7 warn_on_401save_credentialsZ connectionsendhistoryappend) r&respkwargsZparsedrrZsaver9Znew_resprrrr8s(     zMultiDomainBasicAuth.handle_401cKs|jdkrtd|jjdS)z6Response callback to warn about incorrect credentials.r@z)401 Error, Credentials not correct for %sN)rArrrDr)r&rJrKrrrrEs  z MultiDomainBasicAuth.warn_on_401cKs^tsdS|j}d|_|rZ|jdkrZztdtj|Wntk rXtdYnXdS)z1Response callback to save credentials on success.NizSaving credentials to keyringzFailed to save credentials)rr%rArinfoZ set_passwordrZ exception)r&rJrKZcredsrrrrFs z%MultiDomainBasicAuth.save_credentials)TN)TT) __name__ __module__ __qualname__r'r-r1r5r:r<r?r8rErFrrrrr!Ks  2( -r!)&__doc__ZloggingZpip._vendor.requests.authrrZpip._vendor.requests.utilsrZpip._vendor.six.moves.urllibrrBZpip._internal.utils.miscrrrr r Zpip._internal.utils.typingr Zoptparser typingr rrZ pip._internal.vcs.versioncontrolrrZ CredentialsZ getLoggerrMrr ImportErrorrrrr r!rrrrs0        site-packages/pip/_internal/network/__pycache__/xmlrpc.cpython-38.opt-1.pyc000064400000003046151464452710022640 0ustar00U .e=@sPdZddlZddlmZddlmZddlmZe e Z Gdddej Z dS)z#xmlrpclib.Transport implementation N)requests) xmlrpc_client)parsec@s$eZdZdZdddZd ddZdS) PipXmlrpcTransportzRProvide a `xmlrpclib.Transport` implementation via a `PipSession` object. FcCs*tj||t|}|j|_||_dS)N)r Transport__init__ urllib_parseZurlparseZscheme_scheme_session)selfZ index_urlZsessionZ use_datetimeZ index_partsr @/usr/lib/python3.8/site-packages/pip/_internal/network/xmlrpc.pyrs zPipXmlrpcTransport.__init__c Cs|j||dddf}t|}z8ddi}|jj|||dd}|||_||jWSt j k r} zt d| j j|W5d} ~ XYnXdS)Nz Content-Typeztext/xmlT)dataheadersstreamzHTTP error %s while getting %s)r rZ urlunparser ZpostZraise_for_statusverboseZparse_responserawrZ HTTPErrorloggerZcriticalresponseZ status_code) r ZhostZhandlerZ request_bodyrpartsZurlrrexcr r r requests$  zPipXmlrpcTransport.requestN)F)F)__name__ __module__ __qualname____doc__rrr r r r rs r)rZloggingZ pip._vendorrZpip._vendor.six.movesrZpip._vendor.six.moves.urllibrrZ getLoggerrrrrr r r r s     site-packages/pip/_internal/network/__pycache__/auth.cpython-38.pyc000064400000015502151464452710021335 0ustar00U .eo+ @sdZddlZddlmZmZddlmZddlmZ ddl m Z m Z m Z mZmZddlmZerddlmZdd lmZmZmZdd lmZeeeefZeeZz ddlZWnLe k rdZYn6e!k rZ"ze#d ee"dZW5dZ"["XYnXd d Z$GdddeZ%dS)zNetwork Authentication Helpers Contains interface (MultiDomainBasicAuth) and associated glue code for providing credentials in the context of network requests. N)AuthBase HTTPBasicAuth)get_netrc_auth)parse)ask ask_input ask_passwordremove_auth_from_urlsplit_auth_netloc_from_url)MYPY_CHECK_RUNNING)Values)DictOptionalTuple)AuthInfo*Keyring is skipped due to an exception: %sc Cs|rts dSzz tj}Wntk r,Yn4Xtd||||}|dk rZ|j|jfWSWdS|rtd|t||}|r||fWSWn2tk r}zt dt |W5d}~XYnXdS)z3Return the tuple auth for a given url from keyring.Nz'Getting credentials from keyring for %sz$Getting password from keyring for %sr) keyringget_credentialAttributeErrorloggerdebugusernamepasswordZ get_password Exceptionwarningstr)urlrrZcredrexcr>/usr/lib/python3.8/site-packages/pip/_internal/network/auth.pyget_keyring_auth.s,     r c@s`eZdZdddZddZdddZd d Zd d Zd dZddZ ddZ ddZ ddZ dS)MultiDomainBasicAuthTNcCs||_||_i|_d|_dS)N) prompting index_urls passwords_credentials_to_save)selfr"r#rrr__init__MszMultiDomainBasicAuth.__init__cCsB|r |jsdS|jD](}t|dd}||r|SqdS)aReturn the original index URL matching the requested URL. Cached or dynamically generated credentials may work against the original index URL rather than just the netloc. The provided url should have had its username and password removed already. If the original index url had credentials then they will be included in the return value. Returns None if no matching index was found, or if --no-index was specified by the user. N/)r#r rstrip startswith)r&ruprefixrrr_get_index_urlYs   z#MultiDomainBasicAuth._get_index_urlcCst|\}}}|\}}|dk r6|dk r6td||S||} | rft| } | rf| \} } } td| | r| ddk r| \}}|dk r|dk rtd|| S|rt|} | rtd|| S|rt| |pt||}|rtd||S||fS)z2Find and return credentials for the specified URL.NzFound credentials in url for %szFound index url %srz%Found credentials in index url for %sz!Found credentials in netrc for %sz#Found credentials in keyring for %s)r rrr-rr )r& original_urlZ allow_netrcZ allow_keyringrnetlocZurl_user_passwordrrZ index_urlZ index_info_Zindex_url_user_passwordZ netrc_authZkr_authrrr_get_new_credentialsns>         z)MultiDomainBasicAuth._get_new_credentialscCst|\}}}|j|d\}}|dkr>|dkr>||\}}|dk sN|dk rl|pTd}|p\d}||f|j|<|dk r||dk s|dkr|dkstd||||fS)a_Return the credentials to use for the provided URL. If allowed, netrc and keyring may be used to obtain the correct credentials. Returns (url_without_credentials, username, password). Note that even if the original URL contains credentials, this function may return a different username and password. NNNz'Could not load credentials from url: {})r r$getr1AssertionErrorformat)r&r.rr/r0rrrrr_get_url_and_credentialss& z-MultiDomainBasicAuth._get_url_and_credentialscCsH||j\}}}||_|dk r6|dk r6t|||}|d|j|S)Nresponse)r7rr register_hook handle_401)r&reqrrrrrr__call__s zMultiDomainBasicAuth.__call__cCsFtd|}|sdSt||}|r4|d|ddfStd}||dfS)Nz User for %s: r2rFz Password: T)rr r)r&r/rZauthrrrr_prompt_for_passwords  z)MultiDomainBasicAuth._prompt_for_passwordcCstsdStdddgdkS)NFz#Save credentials to keyring [y/N]: yn)rr)r&rrr _should_save_password_to_keyringsz5MultiDomainBasicAuth._should_save_password_to_keyringc Ks|jdkr|S|js|St|j}||j\}}}d|_|dk rv|dk rv||f|j|j<|rv| rv|j||f|_|j |j t |pd|pd|j}|d|j|jr|d|j|jj|f|}|j||S)Nr3r8) status_coder" urllib_parseZurlparserr>r/r%r$rAZcontentrawZ release_connrrequestr9 warn_on_401save_credentialsZ connectionsendhistoryappend) r&respkwargsZparsedrrZsaver;Znew_resprrrr:s(     zMultiDomainBasicAuth.handle_401cKs|jdkrtd|jjdS)z6Response callback to warn about incorrect credentials.rBz)401 Error, Credentials not correct for %sN)rCrrrFr)r&rLrMrrrrGs  z MultiDomainBasicAuth.warn_on_401cKsntdk stdtsdS|j}d|_|rj|jdkrjztdtj|Wntk rhtdYnXdS)z1Response callback to save credentials on success.Nz'should never reach here without keyringizSaving credentials to keyringzFailed to save credentials) rr5r%rCrinfoZ set_passwordrZ exception)r&rLrMZcredsrrrrHs z%MultiDomainBasicAuth.save_credentials)TN)TT) __name__ __module__ __qualname__r'r-r1r7r<r>rAr:rGrHrrrrr!Ks  2( -r!)&__doc__ZloggingZpip._vendor.requests.authrrZpip._vendor.requests.utilsrZpip._vendor.six.moves.urllibrrDZpip._internal.utils.miscrrrr r Zpip._internal.utils.typingr Zoptparser typingr rrZ pip._internal.vcs.versioncontrolrrZ CredentialsZ getLoggerrOrr ImportErrorrrrr r!rrrrs0        site-packages/pip/_internal/network/__pycache__/cache.cpython-38.pyc000064400000004671151464452710021444 0ustar00U .e@sdZddlZddlmZddlmZddlmZddlm Z m Z ddl m Z ddl mZerhdd lmZed d ZGd d d eZdS)zHTTP cache implementation. N)contextmanager) BaseCache) FileCache)adjacent_tmp_filereplace) ensure_dir)MYPY_CHECK_RUNNING)Optionalc cs(z dVWnttfk r"YnXdS)zvIf we can't access the cache then we can just skip caching and process requests as if caching wasn't enabled. N)OSErrorIOErrorr r ?/usr/lib/python3.8/site-packages/pip/_internal/network/cache.pysuppressed_cache_errorss rcs@eZdZdZfddZddZddZdd Zd d ZZ S) SafeFileCachezw A file based cache which is safe to use even when the target directory may not be accessible or writable. cs(|dk stdtt|||_dS)Nz!Cache directory must not be None.)AssertionErrorsuperr__init__ directory)selfr __class__r r r&szSafeFileCache.__init__cCs4t|}t|dd|g}tjj|jf|S)N)rencodelistospathjoinr)rnameZhashedpartsr r r _get_cache_path,s zSafeFileCache._get_cache_pathc CsR||}t8t|d"}|W5QRW5QRSQRXW5QRXdS)Nrb)rropenread)rkeyrfr r r get5s  zSafeFileCache.getc CsZ||}t@ttj|t|}||W5QRXt|j |W5QRXdSN) rrrrrdirnamerwriterr)rr#valuerr$r r r set<s   zSafeFileCache.setc Cs*||}tt|W5QRXdSr&)rrrremove)rr#rr r r deleteGs zSafeFileCache.delete) __name__ __module__ __qualname____doc__rrr%r*r, __classcell__r r rr r s    r)r0r contextlibrZpip._vendor.cachecontrol.cacherZpip._vendor.cachecontrol.cachesrZpip._internal.utils.filesystemrrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingr rrr r r r s       site-packages/pip/_internal/network/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000324151464452710023066 0ustar00U .e2@sdZdS)z+Contains purely network-related utilities. N)__doc__rrB/usr/lib/python3.8/site-packages/pip/_internal/network/__init__.pysite-packages/pip/_internal/network/__pycache__/session.cpython-38.pyc000064400000022012151464452710022051 0ustar00U .e=@sdZddlZddlZddlZddlZddlZddlZddlZddl Z ddl m Z m Z m Z ddlmZddlmZmZddlmZddlmZddlmZdd lmZdd lmZdd lmZdd l m!Z!dd l"m#Z#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+m,Z,m-Z-ddl.m/Z/ddl0m1Z1e/rhddl2m3Z3m4Z4m5Z5m6Z6m7Z7ddl8m9Z9e6e:e:e5e7e;e:ffZZ?e j@dedddddddgZAdZBddZCd d!ZDGd"d#d#eZEGd$d%d%eZFGd&d'd'e jGZHdS)(zhPipSession and supporting code, containing all pip-specific network request configuration and behavior. N)requestssixurllib3)CacheControlAdapter) BaseAdapter HTTPAdapter)Response)CaseInsensitiveDict)parse)InsecureRequestWarning) __version__)MultiDomainBasicAuth) SafeFileCache)HAS_TLS ipaddressssl)check_path_owner)libc_ver)build_url_from_netlocget_installed_version parse_netloc)MYPY_CHECK_RUNNING) url_to_path)IteratorListOptionalTupleUnion)Linkignore)category)Zhttps*r!)r!Z localhostr!)r!z 127.0.0.0/8r!)r!z::1/128r!)filer!N)Zsshr!r!)Z BUILD_BUILDIDZBUILD_IDZCIZ PIP_IS_CIcCstddtDS)z? Return whether it looks like pip is running under CI. css|]}|tjkVqdSN)osenviron).0namer(A/usr/lib/python3.8/site-packages/pip/_internal/network/session.py asz looks_like_ci..)anyCI_ENVIRONMENT_VARIABLESr(r(r(r) looks_like_ciYsr-cCsBdtdtdtid}|dddkr@t|dd<n|dddkrtjjd krltjd d }ntj}d d d|D|dd<nB|dddkrt|dd<n |dddkrt|dd<tjdrHddl m }t t ddt dddg|}t t ddt ddgt}|r:||d<|rH||d<tjdrztdrzdtdd|d<trt|did<trt|did<trt|d<trtj|d <td!}|d k r||d"<trd#nd |d$<tjd%}|d k r(||d&<d'j|tj|d(d#d)d*S)+z6 Return a string representing the user agent. pip)r'versionr')Z installerpythonimplementationr1ZCPythonr/ZPyPyfinalN.cSsg|] }t|qSr()str)r&xr(r(r) xszuser_agent..ZJythonZ IronPythonZlinuxr)distrocSs|dSNr(r6r(r(r)zuser_agent..idcSs|dSr9r(r;r(r(r)r<r=liblibcr8darwinZmacOSsystemreleaseZcpuZopenssl_versionZ setuptoolssetuptools_versionTZciZPIP_USER_AGENT_USER_DATA user_dataz9{data[installer][name]}/{data[installer][version]} {json}),:)Z separatorsZ sort_keys)datajson) r platformZpython_versionZpython_implementationsyspypy_version_info releaseleveljoin startswith pip._vendorr8dictfilterzipZlinux_distributionrZmac_verrB setdefaultrCmachinerrZOPENSSL_VERSIONrr-r$r%getformatrIdumps)rHrLr8Z distro_infosr@rDrEr(r(r) user_agentdsl           rYc@seZdZdddZddZdS)LocalFSAdapterNc Cst|j}t}d|_|j|_zt|} Wn.tk rZ} zd|_| |_W5d} ~ XYnPXtj j | j dd} t |dp~d} t| | j| d|_t|d|_|jj|_|S) NiT)Zusegmtrz text/plain)z Content-TypezContent-Lengthz Last-Modifiedrb)rurlrZ status_coder$statOSErrorrawemailZutilsZ formatdatest_mtime mimetypesZ guess_typer st_sizeheadersopenclose) selfrequeststreamtimeoutverifycertZproxiespathnameZrespZstatsexcZmodifiedZ content_typer(r(r)sends&    zLocalFSAdapter.sendcCsdSr#r()rhr(r(r)rgszLocalFSAdapter.close)NNNNN)__name__ __module__ __qualname__rprgr(r(r(r)rZs rZc@seZdZddZdS)InsecureHTTPAdaptercCsd|_d|_dS)NZ CERT_NONE)Z cert_reqsZca_certs)rhZconnr]rlrmr(r(r) cert_verifyszInsecureHTTPAdapter.cert_verifyN)rqrrrsrur(r(r(r)rtsrtcsFeZdZdZfddZd ddZddZd d Zfd d ZZ S) PipSessionNc s|dd}|dd}|dg}|dd}tt|j||g|_t|jd<t|d|_t j |d d d d gd d}|rt |st d|d}|rtt||d}n t|d}t|d}||_|d||d||dt|D]} |j| ddqdS)zj :param trusted_hosts: Domains not to emit warnings for when not using HTTPS. retriesrcacheN trusted_hosts index_urlsz User-Agent)rziiiig?)ZtotalZstatus_forcelistZbackoff_factorzThe directory '%s' or its parent directory is not owned by the current user and the cache has been disabled. Please check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag.)rx max_retries)r{zhttps://zhttp://zfile://T)suppress_logging)popsuperrv__init__pip_trusted_originsrYrer ZauthrZRetryrloggerwarningrrrrt_insecure_adaptermountrZadd_trusted_host) rhargskwargsrwrxryrzZsecure_adapterZinsecure_adapterhost __class__r(r)rs@            zPipSession.__init__FcCs|s.d|}|dk r$|d|7}t|t|}||jkrL|j||t|d|j|ds|t|d|jdS)z :param host: It is okay to provide a host that has previously been added. :param source: An optional source string, for logging where the host string came from. zadding trusted host: {!r}Nz (from {})/r:rG) rWrinforrappendrrr)rhrsourcer|msgZ host_portr(r(r)r3s     zPipSession.add_trusted_hostccs<tD] }|Vq|jD] \}}d||dkr.dn|fVqdS)Nr!)SECURE_ORIGINSr)rh secure_originrportr(r(r)iter_secure_originsMszPipSession.iter_secure_originsc Cs tt|}|j|j|j}}}|ddd}|D]}|\}}} ||kr\|dkr\qsV                O!site-packages/pip/_internal/network/session.py000064400000036742151464452710015602 0ustar00"""PipSession and supporting code, containing all pip-specific network request configuration and behavior. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import email.utils import json import logging import mimetypes import os import platform import sys import warnings from pip._vendor import requests, six, urllib3 from pip._vendor.cachecontrol import CacheControlAdapter from pip._vendor.requests.adapters import BaseAdapter, HTTPAdapter from pip._vendor.requests.models import Response from pip._vendor.requests.structures import CaseInsensitiveDict from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._vendor.urllib3.exceptions import InsecureRequestWarning from pip import __version__ from pip._internal.network.auth import MultiDomainBasicAuth from pip._internal.network.cache import SafeFileCache # Import ssl from compat so the initial import occurs in only one place. from pip._internal.utils.compat import HAS_TLS, ipaddress, ssl from pip._internal.utils.filesystem import check_path_owner from pip._internal.utils.glibc import libc_ver from pip._internal.utils.misc import ( build_url_from_netloc, get_installed_version, parse_netloc, ) from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import url_to_path if MYPY_CHECK_RUNNING: from typing import ( Iterator, List, Optional, Tuple, Union, ) from pip._internal.models.link import Link SecureOrigin = Tuple[str, str, Optional[Union[int, str]]] logger = logging.getLogger(__name__) # Ignore warning raised when using --trusted-host. warnings.filterwarnings("ignore", category=InsecureRequestWarning) SECURE_ORIGINS = [ # protocol, hostname, port # Taken from Chrome's list of secure origins (See: http://bit.ly/1qrySKC) ("https", "*", "*"), ("*", "localhost", "*"), ("*", "127.0.0.0/8", "*"), ("*", "::1/128", "*"), ("file", "*", None), # ssh is always secure. ("ssh", "*", "*"), ] # type: List[SecureOrigin] # These are environment variables present when running under various # CI systems. For each variable, some CI systems that use the variable # are indicated. The collection was chosen so that for each of a number # of popular systems, at least one of the environment variables is used. # This list is used to provide some indication of and lower bound for # CI traffic to PyPI. Thus, it is okay if the list is not comprehensive. # For more background, see: https://github.com/pypa/pip/issues/5499 CI_ENVIRONMENT_VARIABLES = ( # Azure Pipelines 'BUILD_BUILDID', # Jenkins 'BUILD_ID', # AppVeyor, CircleCI, Codeship, Gitlab CI, Shippable, Travis CI 'CI', # Explicit environment variable. 'PIP_IS_CI', ) def looks_like_ci(): # type: () -> bool """ Return whether it looks like pip is running under CI. """ # We don't use the method of checking for a tty (e.g. using isatty()) # because some CI systems mimic a tty (e.g. Travis CI). Thus that # method doesn't provide definitive information in either direction. return any(name in os.environ for name in CI_ENVIRONMENT_VARIABLES) def user_agent(): """ Return a string representing the user agent. """ data = { "installer": {"name": "pip", "version": __version__}, "python": platform.python_version(), "implementation": { "name": platform.python_implementation(), }, } if data["implementation"]["name"] == 'CPython': data["implementation"]["version"] = platform.python_version() elif data["implementation"]["name"] == 'PyPy': if sys.pypy_version_info.releaselevel == 'final': pypy_version_info = sys.pypy_version_info[:3] else: pypy_version_info = sys.pypy_version_info data["implementation"]["version"] = ".".join( [str(x) for x in pypy_version_info] ) elif data["implementation"]["name"] == 'Jython': # Complete Guess data["implementation"]["version"] = platform.python_version() elif data["implementation"]["name"] == 'IronPython': # Complete Guess data["implementation"]["version"] = platform.python_version() if sys.platform.startswith("linux"): from pip._vendor import distro distro_infos = dict(filter( lambda x: x[1], zip(["name", "version", "id"], distro.linux_distribution()), )) libc = dict(filter( lambda x: x[1], zip(["lib", "version"], libc_ver()), )) if libc: distro_infos["libc"] = libc if distro_infos: data["distro"] = distro_infos if sys.platform.startswith("darwin") and platform.mac_ver()[0]: data["distro"] = {"name": "macOS", "version": platform.mac_ver()[0]} if platform.system(): data.setdefault("system", {})["name"] = platform.system() if platform.release(): data.setdefault("system", {})["release"] = platform.release() if platform.machine(): data["cpu"] = platform.machine() if HAS_TLS: data["openssl_version"] = ssl.OPENSSL_VERSION setuptools_version = get_installed_version("setuptools") if setuptools_version is not None: data["setuptools_version"] = setuptools_version # Use None rather than False so as not to give the impression that # pip knows it is not being run under CI. Rather, it is a null or # inconclusive result. Also, we include some value rather than no # value to make it easier to know that the check has been run. data["ci"] = True if looks_like_ci() else None user_data = os.environ.get("PIP_USER_AGENT_USER_DATA") if user_data is not None: data["user_data"] = user_data return "{data[installer][name]}/{data[installer][version]} {json}".format( data=data, json=json.dumps(data, separators=(",", ":"), sort_keys=True), ) class LocalFSAdapter(BaseAdapter): def send(self, request, stream=None, timeout=None, verify=None, cert=None, proxies=None): pathname = url_to_path(request.url) resp = Response() resp.status_code = 200 resp.url = request.url try: stats = os.stat(pathname) except OSError as exc: resp.status_code = 404 resp.raw = exc else: modified = email.utils.formatdate(stats.st_mtime, usegmt=True) content_type = mimetypes.guess_type(pathname)[0] or "text/plain" resp.headers = CaseInsensitiveDict({ "Content-Type": content_type, "Content-Length": stats.st_size, "Last-Modified": modified, }) resp.raw = open(pathname, "rb") resp.close = resp.raw.close return resp def close(self): pass class InsecureHTTPAdapter(HTTPAdapter): def cert_verify(self, conn, url, verify, cert): conn.cert_reqs = 'CERT_NONE' conn.ca_certs = None class PipSession(requests.Session): timeout = None # type: Optional[int] def __init__(self, *args, **kwargs): """ :param trusted_hosts: Domains not to emit warnings for when not using HTTPS. """ retries = kwargs.pop("retries", 0) cache = kwargs.pop("cache", None) trusted_hosts = kwargs.pop("trusted_hosts", []) # type: List[str] index_urls = kwargs.pop("index_urls", None) super(PipSession, self).__init__(*args, **kwargs) # Namespace the attribute with "pip_" just in case to prevent # possible conflicts with the base class. self.pip_trusted_origins = [] # type: List[Tuple[str, Optional[int]]] # Attach our User Agent to the request self.headers["User-Agent"] = user_agent() # Attach our Authentication handler to the session self.auth = MultiDomainBasicAuth(index_urls=index_urls) # Create our urllib3.Retry instance which will allow us to customize # how we handle retries. retries = urllib3.Retry( # Set the total number of retries that a particular request can # have. total=retries, # A 503 error from PyPI typically means that the Fastly -> Origin # connection got interrupted in some way. A 503 error in general # is typically considered a transient error so we'll go ahead and # retry it. # A 500 may indicate transient error in Amazon S3 # A 520 or 527 - may indicate transient error in CloudFlare status_forcelist=[500, 503, 520, 527], # Add a small amount of back off between failed requests in # order to prevent hammering the service. backoff_factor=0.25, ) # Check to ensure that the directory containing our cache directory # is owned by the user current executing pip. If it does not exist # we will check the parent directory until we find one that does exist. if cache and not check_path_owner(cache): logger.warning( "The directory '%s' or its parent directory is not owned by " "the current user and the cache has been disabled. Please " "check the permissions and owner of that directory. If " "executing pip with sudo, you may want sudo's -H flag.", cache, ) cache = None # We want to _only_ cache responses on securely fetched origins. We do # this because we can't validate the response of an insecurely fetched # origin, and we don't want someone to be able to poison the cache and # require manual eviction from the cache to fix it. if cache: secure_adapter = CacheControlAdapter( cache=SafeFileCache(cache), max_retries=retries, ) else: secure_adapter = HTTPAdapter(max_retries=retries) # Our Insecure HTTPAdapter disables HTTPS validation. It does not # support caching (see above) so we'll use it for all http:// URLs as # well as any https:// host that we've marked as ignoring TLS errors # for. insecure_adapter = InsecureHTTPAdapter(max_retries=retries) # Save this for later use in add_insecure_host(). self._insecure_adapter = insecure_adapter self.mount("https://", secure_adapter) self.mount("http://", insecure_adapter) # Enable file:// urls self.mount("file://", LocalFSAdapter()) for host in trusted_hosts: self.add_trusted_host(host, suppress_logging=True) def add_trusted_host(self, host, source=None, suppress_logging=False): # type: (str, Optional[str], bool) -> None """ :param host: It is okay to provide a host that has previously been added. :param source: An optional source string, for logging where the host string came from. """ if not suppress_logging: msg = 'adding trusted host: {!r}'.format(host) if source is not None: msg += ' (from {})'.format(source) logger.info(msg) host_port = parse_netloc(host) if host_port not in self.pip_trusted_origins: self.pip_trusted_origins.append(host_port) self.mount(build_url_from_netloc(host) + '/', self._insecure_adapter) if not host_port[1]: # Mount wildcard ports for the same host. self.mount( build_url_from_netloc(host) + ':', self._insecure_adapter ) def iter_secure_origins(self): # type: () -> Iterator[SecureOrigin] for secure_origin in SECURE_ORIGINS: yield secure_origin for host, port in self.pip_trusted_origins: yield ('*', host, '*' if port is None else port) def is_secure_origin(self, location): # type: (Link) -> bool # Determine if this url used a secure transport mechanism parsed = urllib_parse.urlparse(str(location)) origin_protocol, origin_host, origin_port = ( parsed.scheme, parsed.hostname, parsed.port, ) # The protocol to use to see if the protocol matches. # Don't count the repository type as part of the protocol: in # cases such as "git+ssh", only use "ssh". (I.e., Only verify against # the last scheme.) origin_protocol = origin_protocol.rsplit('+', 1)[-1] # Determine if our origin is a secure origin by looking through our # hardcoded list of secure origins, as well as any additional ones # configured on this PackageFinder instance. for secure_origin in self.iter_secure_origins(): secure_protocol, secure_host, secure_port = secure_origin if origin_protocol != secure_protocol and secure_protocol != "*": continue try: # We need to do this decode dance to ensure that we have a # unicode object, even on Python 2.x. addr = ipaddress.ip_address( origin_host if ( isinstance(origin_host, six.text_type) or origin_host is None ) else origin_host.decode("utf8") ) network = ipaddress.ip_network( secure_host if isinstance(secure_host, six.text_type) # setting secure_host to proper Union[bytes, str] # creates problems in other places else secure_host.decode("utf8") # type: ignore ) except ValueError: # We don't have both a valid address or a valid network, so # we'll check this origin against hostnames. if ( origin_host and origin_host.lower() != secure_host.lower() and secure_host != "*" ): continue else: # We have a valid address and network, so see if the address # is contained within the network. if addr not in network: continue # Check to see if the port matches. if ( origin_port != secure_port and secure_port != "*" and secure_port is not None ): continue # If we've gotten here, then this origin matches the current # secure origin and we should return True return True # If we've gotten to this point, then the origin isn't secure and we # will not accept it as a valid location to search. We will however # log a warning that we are ignoring it. logger.warning( "The repository located at %s is not a trusted or secure host and " "is being ignored. If this repository is available via HTTPS we " "recommend you use HTTPS instead, otherwise you may silence " "this warning and allow it anyway with '--trusted-host %s'.", origin_host, origin_host, ) return False def request(self, method, url, *args, **kwargs): # Allow setting a default timeout on a session kwargs.setdefault("timeout", self.timeout) # Dispatch the actual request return super(PipSession, self).request(method, url, *args, **kwargs) site-packages/pip/_internal/network/__init__.py000064400000000062151464452710015640 0ustar00"""Contains purely network-related utilities. """ site-packages/pip/_internal/__init__.py000064400000000120151464452710014142 0ustar00#!/usr/bin/env python import pip._internal.utils.inject_securetransport # noqa site-packages/pip/_internal/vcs/subversion.py000064400000030004151464452710015401 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os import re from pip._internal.utils.logging import indent_log from pip._internal.utils.misc import ( display_path, is_console_interactive, rmtree, split_auth_from_netloc, ) from pip._internal.utils.subprocess import make_command from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.vcs.versioncontrol import VersionControl, vcs _svn_xml_url_re = re.compile('url="([^"]+)"') _svn_rev_re = re.compile(r'committed-rev="(\d+)"') _svn_info_xml_rev_re = re.compile(r'\s*revision="(\d+)"') _svn_info_xml_url_re = re.compile(r'(.*)') if MYPY_CHECK_RUNNING: from typing import Optional, Tuple from pip._internal.utils.subprocess import CommandArgs from pip._internal.utils.misc import HiddenText from pip._internal.vcs.versioncontrol import AuthInfo, RevOptions logger = logging.getLogger(__name__) class Subversion(VersionControl): name = 'svn' dirname = '.svn' repo_name = 'checkout' schemes = ('svn', 'svn+ssh', 'svn+http', 'svn+https', 'svn+svn') @classmethod def should_add_vcs_url_prefix(cls, remote_url): return True @staticmethod def get_base_rev_args(rev): return ['-r', rev] @classmethod def get_revision(cls, location): """ Return the maximum revision for all files under a given location """ # Note: taken from setuptools.command.egg_info revision = 0 for base, dirs, files in os.walk(location): if cls.dirname not in dirs: dirs[:] = [] continue # no sense walking uncontrolled subdirs dirs.remove(cls.dirname) entries_fn = os.path.join(base, cls.dirname, 'entries') if not os.path.exists(entries_fn): # FIXME: should we warn? continue dirurl, localrev = cls._get_svn_url_rev(base) if base == location: base = dirurl + '/' # save the root url elif not dirurl or not dirurl.startswith(base): dirs[:] = [] continue # not part of the same svn tree, skip it revision = max(revision, localrev) return revision @classmethod def get_netloc_and_auth(cls, netloc, scheme): """ This override allows the auth information to be passed to svn via the --username and --password options instead of via the URL. """ if scheme == 'ssh': # The --username and --password options can't be used for # svn+ssh URLs, so keep the auth information in the URL. return super(Subversion, cls).get_netloc_and_auth(netloc, scheme) return split_auth_from_netloc(netloc) @classmethod def get_url_rev_and_auth(cls, url): # type: (str) -> Tuple[str, Optional[str], AuthInfo] # hotfix the URL scheme after removing svn+ from svn+ssh:// readd it url, rev, user_pass = super(Subversion, cls).get_url_rev_and_auth(url) if url.startswith('ssh://'): url = 'svn+' + url return url, rev, user_pass @staticmethod def make_rev_args(username, password): # type: (Optional[str], Optional[HiddenText]) -> CommandArgs extra_args = [] # type: CommandArgs if username: extra_args += ['--username', username] if password: extra_args += ['--password', password] return extra_args @classmethod def get_remote_url(cls, location): # In cases where the source is in a subdirectory, not alongside # setup.py we have to look up in the location until we find a real # setup.py orig_location = location while not os.path.exists(os.path.join(location, 'setup.py')): last_location = location location = os.path.dirname(location) if location == last_location: # We've traversed up to the root of the filesystem without # finding setup.py logger.warning( "Could not find setup.py for directory %s (tried all " "parent directories)", orig_location, ) return None return cls._get_svn_url_rev(location)[0] @classmethod def _get_svn_url_rev(cls, location): from pip._internal.exceptions import InstallationError entries_path = os.path.join(location, cls.dirname, 'entries') if os.path.exists(entries_path): with open(entries_path) as f: data = f.read() else: # subversion >= 1.7 does not have the 'entries' file data = '' if (data.startswith('8') or data.startswith('9') or data.startswith('10')): data = list(map(str.splitlines, data.split('\n\x0c\n'))) del data[0][0] # get rid of the '8' url = data[0][3] revs = [int(d[9]) for d in data if len(d) > 9 and d[9]] + [0] elif data.startswith('= 1.7 # Note that using get_remote_call_options is not necessary here # because `svn info` is being run against a local directory. # We don't need to worry about making sure interactive mode # is being used to prompt for passwords, because passwords # are only potentially needed for remote server requests. xml = cls.run_command( ['info', '--xml', location], show_stdout=False, ) url = _svn_info_xml_url_re.search(xml).group(1) revs = [ int(m.group(1)) for m in _svn_info_xml_rev_re.finditer(xml) ] except InstallationError: url, revs = None, [] if revs: rev = max(revs) else: rev = 0 return url, rev @classmethod def is_commit_id_equal(cls, dest, name): """Always assume the versions don't match""" return False def __init__(self, use_interactive=None): # type: (bool) -> None if use_interactive is None: use_interactive = is_console_interactive() self.use_interactive = use_interactive # This member is used to cache the fetched version of the current # ``svn`` client. # Special value definitions: # None: Not evaluated yet. # Empty tuple: Could not parse version. self._vcs_version = None # type: Optional[Tuple[int, ...]] super(Subversion, self).__init__() def call_vcs_version(self): # type: () -> Tuple[int, ...] """Query the version of the currently installed Subversion client. :return: A tuple containing the parts of the version information or ``()`` if the version returned from ``svn`` could not be parsed. :raises: BadCommand: If ``svn`` is not installed. """ # Example versions: # svn, version 1.10.3 (r1842928) # compiled Feb 25 2019, 14:20:39 on x86_64-apple-darwin17.0.0 # svn, version 1.7.14 (r1542130) # compiled Mar 28 2018, 08:49:13 on x86_64-pc-linux-gnu version_prefix = 'svn, version ' version = self.run_command(['--version'], show_stdout=False) if not version.startswith(version_prefix): return () version = version[len(version_prefix):].split()[0] version_list = version.split('.') try: parsed_version = tuple(map(int, version_list)) except ValueError: return () return parsed_version def get_vcs_version(self): # type: () -> Tuple[int, ...] """Return the version of the currently installed Subversion client. If the version of the Subversion client has already been queried, a cached value will be used. :return: A tuple containing the parts of the version information or ``()`` if the version returned from ``svn`` could not be parsed. :raises: BadCommand: If ``svn`` is not installed. """ if self._vcs_version is not None: # Use cached version, if available. # If parsing the version failed previously (empty tuple), # do not attempt to parse it again. return self._vcs_version vcs_version = self.call_vcs_version() self._vcs_version = vcs_version return vcs_version def get_remote_call_options(self): # type: () -> CommandArgs """Return options to be used on calls to Subversion that contact the server. These options are applicable for the following ``svn`` subcommands used in this class. - checkout - export - switch - update :return: A list of command line arguments to pass to ``svn``. """ if not self.use_interactive: # --non-interactive switch is available since Subversion 0.14.4. # Subversion < 1.8 runs in interactive mode by default. return ['--non-interactive'] svn_version = self.get_vcs_version() # By default, Subversion >= 1.8 runs in non-interactive mode if # stdin is not a TTY. Since that is how pip invokes SVN, in # call_subprocess(), pip must pass --force-interactive to ensure # the user can be prompted for a password, if required. # SVN added the --force-interactive option in SVN 1.8. Since # e.g. RHEL/CentOS 7, which is supported until 2024, ships with # SVN 1.7, pip should continue to support SVN 1.7. Therefore, pip # can't safely add the option if the SVN version is < 1.8 (or unknown). if svn_version >= (1, 8): return ['--force-interactive'] return [] def export(self, location, url): # type: (str, HiddenText) -> None """Export the svn repository at the url to the destination location""" url, rev_options = self.get_url_rev_options(url) logger.info('Exporting svn repository %s to %s', url, location) with indent_log(): if os.path.exists(location): # Subversion doesn't like to check out over an existing # directory --force fixes this, but was only added in svn 1.5 rmtree(location) cmd_args = make_command( 'export', self.get_remote_call_options(), rev_options.to_args(), url, location, ) self.run_command(cmd_args, show_stdout=False) def fetch_new(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None rev_display = rev_options.to_display() logger.info( 'Checking out %s%s to %s', url, rev_display, display_path(dest), ) cmd_args = make_command( 'checkout', '-q', self.get_remote_call_options(), rev_options.to_args(), url, dest, ) self.run_command(cmd_args) def switch(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None cmd_args = make_command( 'switch', self.get_remote_call_options(), rev_options.to_args(), url, dest, ) self.run_command(cmd_args) def update(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None cmd_args = make_command( 'update', self.get_remote_call_options(), rev_options.to_args(), dest, ) self.run_command(cmd_args) vcs.register(Subversion) site-packages/pip/_internal/vcs/mercurial.py000064400000011766151464452710015203 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os from pip._vendor.six.moves import configparser from pip._internal.exceptions import BadCommand, InstallationError from pip._internal.utils.misc import display_path from pip._internal.utils.subprocess import make_command from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import path_to_url from pip._internal.vcs.versioncontrol import ( VersionControl, find_path_to_setup_from_repo_root, vcs, ) if MYPY_CHECK_RUNNING: from pip._internal.utils.misc import HiddenText from pip._internal.vcs.versioncontrol import RevOptions logger = logging.getLogger(__name__) class Mercurial(VersionControl): name = 'hg' dirname = '.hg' repo_name = 'clone' schemes = ( 'hg', 'hg+file', 'hg+http', 'hg+https', 'hg+ssh', 'hg+static-http', ) @staticmethod def get_base_rev_args(rev): return [rev] def export(self, location, url): # type: (str, HiddenText) -> None """Export the Hg repository at the url to the destination location""" with TempDirectory(kind="export") as temp_dir: self.unpack(temp_dir.path, url=url) self.run_command( ['archive', location], show_stdout=False, cwd=temp_dir.path ) def fetch_new(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None rev_display = rev_options.to_display() logger.info( 'Cloning hg %s%s to %s', url, rev_display, display_path(dest), ) self.run_command(make_command('clone', '--noupdate', '-q', url, dest)) self.run_command( make_command('update', '-q', rev_options.to_args()), cwd=dest, ) def switch(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None repo_config = os.path.join(dest, self.dirname, 'hgrc') config = configparser.RawConfigParser() try: config.read(repo_config) config.set('paths', 'default', url.secret) with open(repo_config, 'w') as config_file: config.write(config_file) except (OSError, configparser.NoSectionError) as exc: logger.warning( 'Could not switch Mercurial repository to %s: %s', url, exc, ) else: cmd_args = make_command('update', '-q', rev_options.to_args()) self.run_command(cmd_args, cwd=dest) def update(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None self.run_command(['pull', '-q'], cwd=dest) cmd_args = make_command('update', '-q', rev_options.to_args()) self.run_command(cmd_args, cwd=dest) @classmethod def get_remote_url(cls, location): url = cls.run_command( ['showconfig', 'paths.default'], show_stdout=False, cwd=location).strip() if cls._is_local_repository(url): url = path_to_url(url) return url.strip() @classmethod def get_revision(cls, location): """ Return the repository-local changeset revision number, as an integer. """ current_revision = cls.run_command( ['parents', '--template={rev}'], show_stdout=False, cwd=location).strip() return current_revision @classmethod def get_requirement_revision(cls, location): """ Return the changeset identification hash, as a 40-character hexadecimal string """ current_rev_hash = cls.run_command( ['parents', '--template={node}'], show_stdout=False, cwd=location).strip() return current_rev_hash @classmethod def is_commit_id_equal(cls, dest, name): """Always assume the versions don't match""" return False @classmethod def get_subdirectory(cls, location): """ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. """ # find the repo root repo_root = cls.run_command( ['root'], show_stdout=False, cwd=location).strip() if not os.path.isabs(repo_root): repo_root = os.path.abspath(os.path.join(location, repo_root)) return find_path_to_setup_from_repo_root(location, repo_root) @classmethod def controls_location(cls, location): if super(Mercurial, cls).controls_location(location): return True try: cls.run_command( ['identify'], cwd=location, show_stdout=False, on_returncode='raise', log_failed_cmd=False) return True except (BadCommand, InstallationError): return False vcs.register(Mercurial) site-packages/pip/_internal/vcs/bazaar.py000064400000007565151464452710014462 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.utils.misc import display_path, rmtree from pip._internal.utils.subprocess import make_command from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import path_to_url from pip._internal.vcs.versioncontrol import VersionControl, vcs if MYPY_CHECK_RUNNING: from typing import Optional, Tuple from pip._internal.utils.misc import HiddenText from pip._internal.vcs.versioncontrol import AuthInfo, RevOptions logger = logging.getLogger(__name__) class Bazaar(VersionControl): name = 'bzr' dirname = '.bzr' repo_name = 'branch' schemes = ( 'bzr', 'bzr+http', 'bzr+https', 'bzr+ssh', 'bzr+sftp', 'bzr+ftp', 'bzr+lp', ) def __init__(self, *args, **kwargs): super(Bazaar, self).__init__(*args, **kwargs) # This is only needed for python <2.7.5 # Register lp but do not expose as a scheme to support bzr+lp. if getattr(urllib_parse, 'uses_fragment', None): urllib_parse.uses_fragment.extend(['lp']) @staticmethod def get_base_rev_args(rev): return ['-r', rev] def export(self, location, url): # type: (str, HiddenText) -> None """ Export the Bazaar repository at the url to the destination location """ # Remove the location to make sure Bazaar can export it correctly if os.path.exists(location): rmtree(location) url, rev_options = self.get_url_rev_options(url) self.run_command( make_command('export', location, url, rev_options.to_args()), show_stdout=False, ) def fetch_new(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None rev_display = rev_options.to_display() logger.info( 'Checking out %s%s to %s', url, rev_display, display_path(dest), ) cmd_args = ( make_command('branch', '-q', rev_options.to_args(), url, dest) ) self.run_command(cmd_args) def switch(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None self.run_command(make_command('switch', url), cwd=dest) def update(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None cmd_args = make_command('pull', '-q', rev_options.to_args()) self.run_command(cmd_args, cwd=dest) @classmethod def get_url_rev_and_auth(cls, url): # type: (str) -> Tuple[str, Optional[str], AuthInfo] # hotfix the URL scheme after removing bzr+ from bzr+ssh:// readd it url, rev, user_pass = super(Bazaar, cls).get_url_rev_and_auth(url) if url.startswith('ssh://'): url = 'bzr+' + url return url, rev, user_pass @classmethod def get_remote_url(cls, location): urls = cls.run_command(['info'], show_stdout=False, cwd=location) for line in urls.splitlines(): line = line.strip() for x in ('checkout of branch: ', 'parent branch: '): if line.startswith(x): repo = line.split(x)[1] if cls._is_local_repository(repo): return path_to_url(repo) return repo return None @classmethod def get_revision(cls, location): revision = cls.run_command( ['revno'], show_stdout=False, cwd=location, ) return revision.splitlines()[-1] @classmethod def is_commit_id_equal(cls, dest, name): """Always assume the versions don't match""" return False vcs.register(Bazaar) site-packages/pip/_internal/vcs/git.py000064400000032400151464452710013767 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os.path import re from pip._vendor.packaging.version import parse as parse_version from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._vendor.six.moves.urllib import request as urllib_request from pip._internal.exceptions import BadCommand from pip._internal.utils.misc import display_path from pip._internal.utils.subprocess import make_command from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.vcs.versioncontrol import ( RemoteNotFoundError, VersionControl, find_path_to_setup_from_repo_root, vcs, ) if MYPY_CHECK_RUNNING: from typing import Optional, Tuple from pip._internal.utils.misc import HiddenText from pip._internal.vcs.versioncontrol import AuthInfo, RevOptions urlsplit = urllib_parse.urlsplit urlunsplit = urllib_parse.urlunsplit logger = logging.getLogger(__name__) HASH_REGEX = re.compile('^[a-fA-F0-9]{40}$') def looks_like_hash(sha): return bool(HASH_REGEX.match(sha)) class Git(VersionControl): name = 'git' dirname = '.git' repo_name = 'clone' schemes = ( 'git', 'git+http', 'git+https', 'git+ssh', 'git+git', 'git+file', ) # Prevent the user's environment variables from interfering with pip: # https://github.com/pypa/pip/issues/1130 unset_environ = ('GIT_DIR', 'GIT_WORK_TREE') default_arg_rev = 'HEAD' @staticmethod def get_base_rev_args(rev): return [rev] def get_git_version(self): VERSION_PFX = 'git version ' version = self.run_command(['version'], show_stdout=False) if version.startswith(VERSION_PFX): version = version[len(VERSION_PFX):].split()[0] else: version = '' # get first 3 positions of the git version because # on windows it is x.y.z.windows.t, and this parses as # LegacyVersion which always smaller than a Version. version = '.'.join(version.split('.')[:3]) return parse_version(version) @classmethod def get_current_branch(cls, location): """ Return the current branch, or None if HEAD isn't at a branch (e.g. detached HEAD). """ # git-symbolic-ref exits with empty stdout if "HEAD" is a detached # HEAD rather than a symbolic ref. In addition, the -q causes the # command to exit with status code 1 instead of 128 in this case # and to suppress the message to stderr. args = ['symbolic-ref', '-q', 'HEAD'] output = cls.run_command( args, extra_ok_returncodes=(1, ), show_stdout=False, cwd=location, ) ref = output.strip() if ref.startswith('refs/heads/'): return ref[len('refs/heads/'):] return None def export(self, location, url): # type: (str, HiddenText) -> None """Export the Git repository at the url to the destination location""" if not location.endswith('/'): location = location + '/' with TempDirectory(kind="export") as temp_dir: self.unpack(temp_dir.path, url=url) self.run_command( ['checkout-index', '-a', '-f', '--prefix', location], show_stdout=False, cwd=temp_dir.path ) @classmethod def get_revision_sha(cls, dest, rev): """ Return (sha_or_none, is_branch), where sha_or_none is a commit hash if the revision names a remote branch or tag, otherwise None. Args: dest: the repository directory. rev: the revision name. """ # Pass rev to pre-filter the list. output = cls.run_command(['show-ref', rev], cwd=dest, show_stdout=False, on_returncode='ignore') refs = {} # NOTE: We do not use splitlines here since that would split on other # unicode separators, which can be maliciously used to install a # different revision. for line in output.strip().split("\n"): line = line.rstrip("\r") if not line: continue try: sha, ref = line.split(" ", maxsplit=2) except ValueError: # Include the offending line to simplify troubleshooting if # this error ever occurs. raise ValueError('unexpected show-ref line: {!r}'.format(line)) refs[ref] = sha branch_ref = 'refs/remotes/origin/{}'.format(rev) tag_ref = 'refs/tags/{}'.format(rev) sha = refs.get(branch_ref) if sha is not None: return (sha, True) sha = refs.get(tag_ref) return (sha, False) @classmethod def resolve_revision(cls, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> RevOptions """ Resolve a revision to a new RevOptions object with the SHA1 of the branch, tag, or ref if found. Args: rev_options: a RevOptions object. """ rev = rev_options.arg_rev # The arg_rev property's implementation for Git ensures that the # rev return value is always non-None. assert rev is not None sha, is_branch = cls.get_revision_sha(dest, rev) if sha is not None: rev_options = rev_options.make_new(sha) rev_options.branch_name = rev if is_branch else None return rev_options # Do not show a warning for the common case of something that has # the form of a Git commit hash. if not looks_like_hash(rev): logger.warning( "Did not find branch or tag '%s', assuming revision or ref.", rev, ) if not rev.startswith('refs/'): return rev_options # If it looks like a ref, we have to fetch it explicitly. cls.run_command( make_command('fetch', '-q', url, rev_options.to_args()), cwd=dest, ) # Change the revision to the SHA of the ref we fetched sha = cls.get_revision(dest, rev='FETCH_HEAD') rev_options = rev_options.make_new(sha) return rev_options @classmethod def is_commit_id_equal(cls, dest, name): """ Return whether the current commit hash equals the given name. Args: dest: the repository directory. name: a string name. """ if not name: # Then avoid an unnecessary subprocess call. return False return cls.get_revision(dest) == name def fetch_new(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None rev_display = rev_options.to_display() logger.info('Cloning %s%s to %s', url, rev_display, display_path(dest)) self.run_command(make_command('clone', '-q', url, dest)) if rev_options.rev: # Then a specific revision was requested. rev_options = self.resolve_revision(dest, url, rev_options) branch_name = getattr(rev_options, 'branch_name', None) if branch_name is None: # Only do a checkout if the current commit id doesn't match # the requested revision. if not self.is_commit_id_equal(dest, rev_options.rev): cmd_args = make_command( 'checkout', '-q', rev_options.to_args(), ) self.run_command(cmd_args, cwd=dest) elif self.get_current_branch(dest) != branch_name: # Then a specific branch was requested, and that branch # is not yet checked out. track_branch = 'origin/{}'.format(branch_name) cmd_args = [ 'checkout', '-b', branch_name, '--track', track_branch, ] self.run_command(cmd_args, cwd=dest) #: repo may contain submodules self.update_submodules(dest) def switch(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None self.run_command( make_command('config', 'remote.origin.url', url), cwd=dest, ) cmd_args = make_command('checkout', '-q', rev_options.to_args()) self.run_command(cmd_args, cwd=dest) self.update_submodules(dest) def update(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None # First fetch changes from the default remote if self.get_git_version() >= parse_version('1.9.0'): # fetch tags in addition to everything else self.run_command(['fetch', '-q', '--tags'], cwd=dest) else: self.run_command(['fetch', '-q'], cwd=dest) # Then reset to wanted revision (maybe even origin/master) rev_options = self.resolve_revision(dest, url, rev_options) cmd_args = make_command('reset', '--hard', '-q', rev_options.to_args()) self.run_command(cmd_args, cwd=dest) #: update submodules self.update_submodules(dest) @classmethod def get_remote_url(cls, location): """ Return URL of the first remote encountered. Raises RemoteNotFoundError if the repository does not have a remote url configured. """ # We need to pass 1 for extra_ok_returncodes since the command # exits with return code 1 if there are no matching lines. stdout = cls.run_command( ['config', '--get-regexp', r'remote\..*\.url'], extra_ok_returncodes=(1, ), show_stdout=False, cwd=location, ) remotes = stdout.splitlines() try: found_remote = remotes[0] except IndexError: raise RemoteNotFoundError for remote in remotes: if remote.startswith('remote.origin.url '): found_remote = remote break url = found_remote.split(' ')[1] return url.strip() @classmethod def get_revision(cls, location, rev=None): if rev is None: rev = 'HEAD' current_rev = cls.run_command( ['rev-parse', rev], show_stdout=False, cwd=location, ) return current_rev.strip() @classmethod def get_subdirectory(cls, location): """ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. """ # find the repo root git_dir = cls.run_command( ['rev-parse', '--git-dir'], show_stdout=False, cwd=location).strip() if not os.path.isabs(git_dir): git_dir = os.path.join(location, git_dir) repo_root = os.path.abspath(os.path.join(git_dir, '..')) return find_path_to_setup_from_repo_root(location, repo_root) @classmethod def get_url_rev_and_auth(cls, url): # type: (str) -> Tuple[str, Optional[str], AuthInfo] """ Prefixes stub URLs like 'user@hostname:user/repo.git' with 'ssh://'. That's required because although they use SSH they sometimes don't work with a ssh:// scheme (e.g. GitHub). But we need a scheme for parsing. Hence we remove it again afterwards and return it as a stub. """ # Works around an apparent Git bug # (see https://article.gmane.org/gmane.comp.version-control.git/146500) scheme, netloc, path, query, fragment = urlsplit(url) if scheme.endswith('file'): initial_slashes = path[:-len(path.lstrip('/'))] newpath = ( initial_slashes + urllib_request.url2pathname(path) .replace('\\', '/').lstrip('/') ) url = urlunsplit((scheme, netloc, newpath, query, fragment)) after_plus = scheme.find('+') + 1 url = scheme[:after_plus] + urlunsplit( (scheme[after_plus:], netloc, newpath, query, fragment), ) if '://' not in url: assert 'file:' not in url url = url.replace('git+', 'git+ssh://') url, rev, user_pass = super(Git, cls).get_url_rev_and_auth(url) url = url.replace('ssh://', '') else: url, rev, user_pass = super(Git, cls).get_url_rev_and_auth(url) return url, rev, user_pass @classmethod def update_submodules(cls, location): if not os.path.exists(os.path.join(location, '.gitmodules')): return cls.run_command( ['submodule', 'update', '--init', '--recursive', '-q'], cwd=location, ) @classmethod def controls_location(cls, location): if super(Git, cls).controls_location(location): return True try: r = cls.run_command(['rev-parse'], cwd=location, show_stdout=False, on_returncode='ignore', log_failed_cmd=False) return not r except BadCommand: logger.debug("could not determine if %s is under git control " "because git is not available", location) return False vcs.register(Git) site-packages/pip/_internal/vcs/__pycache__/bazaar.cpython-38.pyc000064400000007243151464452710020741 0ustar00U .eu@sddlmZddlZddlZddlmZddlmZm Z ddl m Z ddl m Z ddlmZddlmZmZe rdd lmZmZdd lmZdd lmZmZeeZGd d d eZeedS))absolute_importN)parse) display_pathrmtree) make_command)MYPY_CHECK_RUNNING) path_to_url)VersionControlvcs)OptionalTuple) HiddenText)AuthInfo RevOptionscseZdZdZdZdZdZfddZeddZ d d Z d d Z d dZ ddZ efddZeddZeddZeddZZS)Bazaarbzrz.bzrbranch)rzbzr+httpz bzr+httpszbzr+sshzbzr+sftpzbzr+ftpzbzr+lpcs0tt|j||ttddr,tjdgdS)N uses_fragmentZlp)superr__init__getattr urllib_parserextend)selfargskwargs __class__|}td||t|tdd|||}||dS)NzChecking out %s%s to %sr-q)Z to_displayloggerinforrr(r')rdestr*r+Z rev_displaycmd_argsrrr fetch_new=szBazaar.fetch_newcCs|jtd||ddS)Nswitchcwd)r'r)rr/r*r+rrrr2Ksz Bazaar.switchcCs"tdd|}|j||ddS)NZpullr,r3)rr(r')rr/r*r+r0rrrupdateOsz Bazaar.updatecs2tt||\}}}|dr(d|}|||fS)Nzssh://zbzr+)rrget_url_rev_and_auth startswith)clsr*r Z user_passrrrr6Ts zBazaar.get_url_rev_and_authcCst|jdgd|d}|D]T}|}dD]B}||r*||d}||r`t|S|Sq*qdS)Nr.Fr#r4)zcheckout of branch: zparent branch: )r' splitlinesstripr7splitZ_is_local_repositoryr)r8r)ZurlslinexZreporrrget_remote_url]s   zBazaar.get_remote_urlcCs|jdgd|d}|dS)NZrevnoFr9)r'r;)r8r)Zrevisionrrr get_revisionks zBazaar.get_revisioncCsdS)z&Always assume the versions don't matchFr)r8r/namerrris_commit_id_equalrszBazaar.is_commit_id_equal)__name__ __module__ __qualname__rCdirnameZ repo_nameZschemesr staticmethodr!r"r1r2r5 classmethodr6r@rBrD __classcell__rrrrrs&    r)Z __future__rZloggingr$Zpip._vendor.six.moves.urllibrrZpip._internal.utils.miscrrZpip._internal.utils.subprocessrZpip._internal.utils.typingrZpip._internal.utils.urlsrZ pip._internal.vcs.versioncontrolr r typingr r r rrZ getLoggerrEr-rregisterrrrrs       ^site-packages/pip/_internal/vcs/__pycache__/subversion.cpython-38.opt-1.pyc000064400000020447151464452710022640 0ustar00U .e0@sddlmZddlZddlZddlZddlmZddlmZm Z m Z m Z ddl m Z ddlmZddlmZmZedZed Zed Zed Zerdd lmZmZdd l mZddlmZddlmZmZee Z!GdddeZ"e#e"dS))absolute_importN) indent_log) display_pathis_console_interactivermtreesplit_auth_from_netloc) make_command)MYPY_CHECK_RUNNING)VersionControlvcsz url="([^"]+)"zcommitted-rev="(\d+)"z\s*revision="(\d+)"z(.*))OptionalTuple) CommandArgs) HiddenText)AuthInfo RevOptionscseZdZdZdZdZdZeddZe ddZ ed d Z efd d Z efd dZ e ddZeddZeddZeddZd(fdd ZddZddZddZd d!Zd"d#Zd$d%Zd&d'ZZS)) Subversionsvnz.svncheckout)rzsvn+sshzsvn+httpz svn+httpszsvn+svncCsdS)NT)clsZ remote_urlrr@/usr/lib/python3.8/site-packages/pip/_internal/vcs/subversion.pyshould_add_vcs_url_prefix+sz$Subversion.should_add_vcs_url_prefixcCsd|gS)Nz-rr)revrrrget_base_rev_args/szSubversion.get_base_rev_argsc Csd}t|D]\}}}|j|kr0g|dd<q||jtj||jd}tj|s\q||\}}||kr||d}n|r||sg|dd<qt ||}q|S)zR Return the maximum revision for all files under a given location rNentries/) oswalkdirnameremovepathjoinexists_get_svn_url_rev startswithmax) rlocationZrevisionbasedirsfilesZ entries_fnZdirurlZlocalrevrrr get_revision3s"       zSubversion.get_revisioncs"|dkrtt|||St|S)z This override allows the auth information to be passed to svn via the --username and --password options instead of via the URL. Zssh)superrget_netloc_and_authr)rZnetlocZscheme __class__rrr-OszSubversion.get_netloc_and_authcs2tt||\}}}|dr(d|}|||fS)Nzssh://zsvn+)r,rget_url_rev_and_authr%)rurlrZ user_passr.rrr0\s zSubversion.get_url_rev_and_authcCs(g}|r|d|g7}|r$|d|g7}|S)Nz --usernamez --passwordr)ZusernameZpasswordZ extra_argsrrr make_rev_argses   zSubversion.make_rev_argscCsR|}tjtj|dsD|}tj|}||krtd|dSq||dS)Nzsetup.pyzGCould not find setup.py for directory %s (tried all parent directories)r)rr!r#r"rloggerZwarningr$)rr'Z orig_locationZ last_locationrrrget_remote_urlps zSubversion.get_remote_urlc Cspddlm}tj||jd}tj|rHt|}|}W5QRXnd}| dsj| dsj| drt t t j |d}|dd=|dd }d d |Ddg}n| d rt|}|std ||d}dd t|Ddg}n^z<|jdd|gdd} t| d}dd t| D}Wn |k rRdg}}YnX|rdt|} nd} || fS)Nr)InstallationErrorr89Z10z cSs,g|]$}t|dkr|drt|dqS) )lenint).0drrr s z/Subversion._get_svn_url_rev..zs,           *site-packages/pip/_internal/vcs/__pycache__/subversion.cpython-38.pyc000064400000020447151464452710021701 0ustar00U .e0@sddlmZddlZddlZddlZddlmZddlmZm Z m Z m Z ddl m Z ddlmZddlmZmZedZed Zed Zed Zerdd lmZmZdd l mZddlmZddlmZmZee Z!GdddeZ"e#e"dS))absolute_importN) indent_log) display_pathis_console_interactivermtreesplit_auth_from_netloc) make_command)MYPY_CHECK_RUNNING)VersionControlvcsz url="([^"]+)"zcommitted-rev="(\d+)"z\s*revision="(\d+)"z(.*))OptionalTuple) CommandArgs) HiddenText)AuthInfo RevOptionscseZdZdZdZdZdZeddZe ddZ ed d Z efd d Z efd dZ e ddZeddZeddZeddZd(fdd ZddZddZddZd d!Zd"d#Zd$d%Zd&d'ZZS)) Subversionsvnz.svncheckout)rzsvn+sshzsvn+httpz svn+httpszsvn+svncCsdS)NT)clsZ remote_urlrr@/usr/lib/python3.8/site-packages/pip/_internal/vcs/subversion.pyshould_add_vcs_url_prefix+sz$Subversion.should_add_vcs_url_prefixcCsd|gS)Nz-rr)revrrrget_base_rev_args/szSubversion.get_base_rev_argsc Csd}t|D]\}}}|j|kr0g|dd<q||jtj||jd}tj|s\q||\}}||kr||d}n|r||sg|dd<qt ||}q|S)zR Return the maximum revision for all files under a given location rNentries/) oswalkdirnameremovepathjoinexists_get_svn_url_rev startswithmax) rlocationZrevisionbasedirsfilesZ entries_fnZdirurlZlocalrevrrr get_revision3s"       zSubversion.get_revisioncs"|dkrtt|||St|S)z This override allows the auth information to be passed to svn via the --username and --password options instead of via the URL. Zssh)superrget_netloc_and_authr)rZnetlocZscheme __class__rrr-OszSubversion.get_netloc_and_authcs2tt||\}}}|dr(d|}|||fS)Nzssh://zsvn+)r,rget_url_rev_and_authr%)rurlrZ user_passr.rrr0\s zSubversion.get_url_rev_and_authcCs(g}|r|d|g7}|r$|d|g7}|S)Nz --usernamez --passwordr)ZusernameZpasswordZ extra_argsrrr make_rev_argses   zSubversion.make_rev_argscCsR|}tjtj|dsD|}tj|}||krtd|dSq||dS)Nzsetup.pyzGCould not find setup.py for directory %s (tried all parent directories)r)rr!r#r"rloggerZwarningr$)rr'Z orig_locationZ last_locationrrrget_remote_urlps zSubversion.get_remote_urlc Cspddlm}tj||jd}tj|rHt|}|}W5QRXnd}| dsj| dsj| drt t t j |d}|dd=|dd }d d |Ddg}n| d rt|}|std ||d}dd t|Ddg}n^z<|jdd|gdd} t| d}dd t| D}Wn |k rRdg}}YnX|rdt|} nd} || fS)Nr)InstallationErrorr89Z10z cSs,g|]$}t|dkr|drt|dqS) )lenint).0drrr s z/Subversion._get_svn_url_rev..zs,           *site-packages/pip/_internal/vcs/__pycache__/bazaar.cpython-38.opt-1.pyc000064400000007243151464452710021700 0ustar00U .eu@sddlmZddlZddlZddlmZddlmZm Z ddl m Z ddl m Z ddlmZddlmZmZe rdd lmZmZdd lmZdd lmZmZeeZGd d d eZeedS))absolute_importN)parse) display_pathrmtree) make_command)MYPY_CHECK_RUNNING) path_to_url)VersionControlvcs)OptionalTuple) HiddenText)AuthInfo RevOptionscseZdZdZdZdZdZfddZeddZ d d Z d d Z d dZ ddZ efddZeddZeddZeddZZS)Bazaarbzrz.bzrbranch)rzbzr+httpz bzr+httpszbzr+sshzbzr+sftpzbzr+ftpzbzr+lpcs0tt|j||ttddr,tjdgdS)N uses_fragmentZlp)superr__init__getattr urllib_parserextend)selfargskwargs __class__|}td||t|tdd|||}||dS)NzChecking out %s%s to %sr-q)Z to_displayloggerinforrr(r')rdestr*r+Z rev_displaycmd_argsrrr fetch_new=szBazaar.fetch_newcCs|jtd||ddS)Nswitchcwd)r'r)rr/r*r+rrrr2Ksz Bazaar.switchcCs"tdd|}|j||ddS)NZpullr,r3)rr(r')rr/r*r+r0rrrupdateOsz Bazaar.updatecs2tt||\}}}|dr(d|}|||fS)Nzssh://zbzr+)rrget_url_rev_and_auth startswith)clsr*r Z user_passrrrr6Ts zBazaar.get_url_rev_and_authcCst|jdgd|d}|D]T}|}dD]B}||r*||d}||r`t|S|Sq*qdS)Nr.Fr#r4)zcheckout of branch: zparent branch: )r' splitlinesstripr7splitZ_is_local_repositoryr)r8r)ZurlslinexZreporrrget_remote_url]s   zBazaar.get_remote_urlcCs|jdgd|d}|dS)NZrevnoFr9)r'r;)r8r)Zrevisionrrr get_revisionks zBazaar.get_revisioncCsdS)z&Always assume the versions don't matchFr)r8r/namerrris_commit_id_equalrszBazaar.is_commit_id_equal)__name__ __module__ __qualname__rCdirnameZ repo_nameZschemesr staticmethodr!r"r1r2r5 classmethodr6r@rBrD __classcell__rrrrrs&    r)Z __future__rZloggingr$Zpip._vendor.six.moves.urllibrrZpip._internal.utils.miscrrZpip._internal.utils.subprocessrZpip._internal.utils.typingrZpip._internal.utils.urlsrZ pip._internal.vcs.versioncontrolr r typingr r r rrZ getLoggerrEr-rregisterrrrrs       ^site-packages/pip/_internal/vcs/__pycache__/mercurial.cpython-38.pyc000064400000011430151464452710021455 0ustar00U .e@sddlmZddlZddlZddlmZddlmZmZddl m Z ddl m Z ddl mZddlmZdd lmZdd lmZmZmZerdd l mZdd lmZeeZGd ddeZeedS))absolute_importN) configparser) BadCommandInstallationError) display_path) make_command) TempDirectory)MYPY_CHECK_RUNNING) path_to_url)VersionControl!find_path_to_setup_from_repo_rootvcs) HiddenText) RevOptionscseZdZdZdZdZdZeddZddZ d d Z d d Z d dZ e ddZe ddZe ddZe ddZe ddZe fddZZS) Mercurialhgz.hgclone)rzhg+filezhg+httpzhg+httpszhg+sshzhg+static-httpcCs|gS)N)Zrevrr?/usr/lib/python3.8/site-packages/pip/_internal/vcs/mercurial.pyget_base_rev_args'szMercurial.get_base_rev_argsc Cs@tdd,}|j|j|d|jd|gd|jdW5QRXdS)z?Export the Hg repository at the url to the destination locationexport)Zkind)urlarchiveF show_stdoutcwdN)runpackpath run_command)selflocationrZtemp_dirrrrr+s zMercurial.exportcCsP|}td||t||tddd|||jtdd||ddS)NzCloning hg %s%s to %srz --noupdate-qupdater)Z to_displayloggerinforrrto_args)rdestr rev_optionsZ rev_displayrrr fetch_new5szMercurial.fetch_newc Cstj||jd}t}z>|||dd|jt |d}| |W5QRXWn6t tj fk r}zt d||W5d}~XYn Xtdd|}|j||ddS) NZhgrcpathsdefaultwz/Could not switch Mercurial repository to %s: %sr"r!r#)osrjoindirnamerZRawConfigParserreadsetZsecretopenwriteOSErrorZNoSectionErrorr$Zwarningrr&r) rr'rr(Z repo_configZconfigZ config_fileexccmd_argsrrrswitchDs  zMercurial.switchcCs4|jddg|dtdd|}|j||ddS)NZpullr!r#r")rrr&)rr'rr(r6rrrr"UszMercurial.updatecCs2|jddgd|d}||r*t|}|S)NZ showconfigz paths.defaultFr)rstripZ_is_local_repositoryr )clsr rrrrget_remote_url[s  zMercurial.get_remote_urlcCs|jddgd|d}|S)zW Return the repository-local changeset revision number, as an integer. parentsz--template={rev}Frrr8)r9r Zcurrent_revisionrrr get_revisionds  zMercurial.get_revisioncCs|jddgd|d}|S)zh Return the changeset identification hash, as a 40-character hexadecimal string r;z--template={node}Frr<)r9r Zcurrent_rev_hashrrrget_requirement_revisionns  z"Mercurial.get_requirement_revisioncCsdS)z&Always assume the versions don't matchFr)r9r'namerrris_commit_id_equalyszMercurial.is_commit_id_equalcCsB|jdgd|d}tj|s8tjtj||}t||S)z~ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. rootFr)rr8r-risabsabspathr.r )r9r Z repo_rootrrrget_subdirectory~s  zMercurial.get_subdirectoryc sPtt||rdSz|jdg|ddddWdSttfk rJYdSXdS)NTZidentifyFraise)rrZ on_returncodeZlog_failed_cmd)superrcontrols_locationrrr)r9r  __class__rrrGszMercurial.controls_location)__name__ __module__ __qualname__r?r/Z repo_nameZschemes staticmethodrrr)r7r" classmethodr:r=r>r@rDrG __classcell__rrrHrrs,       r)Z __future__rZloggingr-Zpip._vendor.six.movesrZpip._internal.exceptionsrrZpip._internal.utils.miscrZpip._internal.utils.subprocessrZpip._internal.utils.temp_dirrZpip._internal.utils.typingr Zpip._internal.utils.urlsr Z pip._internal.vcs.versioncontrolr r r rrZ getLoggerrJr$rregisterrrrrs          |site-packages/pip/_internal/vcs/__pycache__/__init__.cpython-38.pyc000064400000000700151464452710021227 0ustar00U .ei@s<ddlZddlZddlZddlZddlmZmZmZm Z dS)N)RemoteNotFoundErroris_urlmake_vcs_requirement_urlvcs) Zpip._internal.vcs.bazaarZpipZpip._internal.vcs.gitZpip._internal.vcs.mercurialZpip._internal.vcs.subversionZ pip._internal.vcs.versioncontrolrrrrrr>/usr/lib/python3.8/site-packages/pip/_internal/vcs/__init__.pyssite-packages/pip/_internal/vcs/__pycache__/versioncontrol.cpython-38.pyc000064400000044035151464452710022567 0ustar00U .e~S@sdZddlmZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl mZddlmZddlmZmZmZmZmZmZdd lmZmZdd lmZdd lmZerdd lm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)dd l*m+Z+ddlm,Z,ddlm-Z-e'e%e.e%e.fZ/dgZ0e1e2Z3ddZ4dddZ5ddZ6Gddde7Z8Gddde9Z:Gddde9Z;e;Z)r&r=r!r(r>r#r#r$__repr__szRevOptions.__repr__cCs|jdkr|jjS|jSN)r(r=default_arg_revr@r#r#r$arg_revs zRevOptions.arg_revcCs0g}|j}|dk r"||j|7}||j7}|S)z< Return the VCS-specific command arguments. N)rDr=get_base_rev_argsr<)r>argsr(r#r#r$to_argss  zRevOptions.to_argscCs|js dSd|jS)Nz (to revision {}))r(r&r@r#r#r$ to_displayszRevOptions.to_displaycCs|jj||jdS)z Make a copy of the current instance, but with a new rev. Args: rev: the name of the revision for the new object. r<)r=make_rev_optionsr<)r>r(r#r#r$make_newszRevOptions.make_new)NN) r8r9r:__doc__r?rApropertyrDrGrIrLr#r#r#r$r;js    r;cs|eZdZiZddddddgZfddZd d Zed d Zed dZ eddZ ddZ ddZ ddZ ddZZS) VcsSupportZsshZgitZhgZbzrZsftpZsvncs:tj|jttddr(tj|jtt|dS)N uses_fragment) urllib_parseZ uses_netlocextendschemesgetattrrPsuperrOr?r@ __class__r#r$r?s zVcsSupport.__init__cCs |jSrB) _registry__iter__r@r#r#r$rYszVcsSupport.__iter__cCst|jSrB)listrXvaluesr@r#r#r$backendsszVcsSupport.backendscCsdd|jDS)NcSsg|] }|jqSr#)r1).0backendr#r#r$ sz'VcsSupport.dirnames..)r\r@r#r#r$dirnamesszVcsSupport.dirnamescCs g}|jD]}||jq |SrB)r\rRrS)r>rSr^r#r#r$r s zVcsSupport.all_schemescCsHt|dstd|jdS|j|jkrD||j|j<td|jdS)Nr!zCannot register VCS %szRegistered VCS backend: %s)hasattrr2r3r8r!rXdebug)r>clsr#r#r$registers   zVcsSupport.registercCs||jkr|j|=dSrB)rXr>r!r#r#r$ unregisters zVcsSupport.unregistercCs6|jD]&}||r td||j|Sq dS)zv Return a VersionControl object if a repository of that type is found at the given directory. zDetermine that %s uses VCS: %sN)rXr[controls_locationr2rbr!)r>r5Z vcs_backendr#r#r$get_backend_for_dirs  zVcsSupport.get_backend_for_dircCs|}|j|S)z9 Return a VersionControl object or None. )lowerrXgetrer#r#r$ get_backendszVcsSupport.get_backend)r8r9r:rXrSr?rYrNr\r`r rdrfrhrk __classcell__r#r#rVr$rOs      rOc @s8eZdZdZdZdZdZdZdZe ddZ e ddZ e dd Z e d d Z ed d Ze d8ddZe ddZddZe ddZe ddZeddZddZeddZe ddZd d!Zd"d#Zd$d%Ze d&d'Zd(d)Zd*d+Ze d,d-Ze d.d/Z e d9d2d3Z!e d4d5Z"e d6d7Z#dS):VersionControlrHr#NcCs|d|j S)z Return whether the vcs prefix (e.g. "git+") should be added to a repository's remote url when used in a requirement. z{}:)ri startswithr&r!)rcZ remote_urlr#r#r$should_add_vcs_url_prefixsz(VersionControl.should_add_vcs_url_prefixcCsdS)z~ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. Nr#rcr5r#r#r$get_subdirectoryszVersionControl.get_subdirectorycCs ||S)zR Return the revision string that should be used in a requirement. ) get_revision)rcrepo_dirr#r#r$get_requirement_revisionsz'VersionControl.get_requirement_revisioncCsV||}|dkrdS||r.d|j|}||}||}t||||d}|S)aC Return the requirement string to use to redownload the files currently at the given repository directory. Args: project_name: the (unescaped) project name. The return value has a form similar to the following: {repository_url}@{revision}#egg={project_name} Nz{}+{})r*)get_remote_urlror&r!rtrqr,)rcrsr)r'Zrevisionr*r+r#r#r$get_src_requirements    z"VersionControl.get_src_requirementcCstdS)z Return the base revision arguments for a vcs command. Args: rev: the name of a revision to install. Cannot be None. NNotImplementedError)r(r#r#r$rE8sz VersionControl.get_base_rev_argscCst|||dS)z Return a RevOptions object. Args: rev: the name of a revision to install. extra_args: a list of extra options. rJ)r;)rcr(r<r#r#r$rKBs zVersionControl.make_rev_optionscCs&tj|\}}|tjjp$t|S)zy posix absolute paths start with os.path.sep, win32 ones start with drive (like c:\folder) )r-r. splitdrivernsepbool)rcZrepoZdrivetailr#r#r$_is_local_repositoryNsz#VersionControl._is_local_repositorycCstdS)z Export the repository at the url to the destination location i.e. only download the files, without vcs informations :param url: the repository URL starting with a vcs prefix. Nrwr>r5urlr#r#r$exportXszVersionControl.exportcCs|dfS)aZ Parse the repository URL's netloc, and return the new netloc to use along with auth information. Args: netloc: the original repository URL netloc. scheme: the repository URL's scheme without the vcs prefix. This is mainly for the Subversion class to override, so that auth information can be provided via the --username and --password options instead of through the URL. For other subclasses like Git without such an option, auth information must stay in the URL. Returns: (netloc, (username, password)). )NNr#)rcnetlocr"r#r#r$get_netloc_and_authbsz"VersionControl.get_netloc_and_authc Cst|\}}}}}d|kr*td||ddd}|||\}}d}d|krf|dd\}}t||||df}|||fS)z Parse the repository URL to use, and return the URL, revision, and auth info to use. Returns: (url, rev, (username, password)). +zvSorry, {!r} is a malformed VCS url. The format is +://, e.g. svn+http://myrepo/svn/MyApp#egg=MyAppN@rH)rQZurlsplit ValueErrorr&splitrrsplitZ urlunsplit) rcrr"rr.ZqueryZfrag user_passr(r#r#r$get_url_rev_and_authus z#VersionControl.get_url_rev_and_authcCsgS)zM Return the RevOptions "extra arguments" to use in obtain(). r#)usernamepasswordr#r#r$ make_rev_argsszVersionControl.make_rev_argsc CsT||j\}}}|\}}d}|dk r.t|}|||}|j||d} t|| fS)z Return the URL and RevOptions object to use in obtain() and in some cases export(), as a tuple (url, rev_options). NrJ)rsecretr rrKr ) r>rZ secret_urlr(rrZsecret_passwordrr< rev_optionsr#r#r$get_url_rev_optionss z"VersionControl.get_url_rev_optionscCst|dS)zi Normalize a URL for comparison by unquoting it and removing any trailing slash. /)rQZunquoterstriprr#r#r$ normalize_urlszVersionControl.normalize_urlcCs||||kS)zV Compare two repo URLs for identity, ignoring incidental differences. )r)rcZurl1Zurl2r#r#r$ compare_urlsszVersionControl.compare_urlscCstdS)z Fetch a revision from a repository, in the case that this is the first fetch from the repository. Args: dest: the directory to fetch the repository to. rev_options: a RevOptions object. Nrwr>destrrr#r#r$ fetch_news zVersionControl.fetch_newcCstdS)z} Switch the repo at ``dest`` to point to ``URL``. Args: rev_options: a RevOptions object. Nrwrr#r#r$switchszVersionControl.switchcCstdS)z Update an already-existing repo to the given ``rev_options``. Args: rev_options: a RevOptions object. Nrwrr#r#r$updateszVersionControl.updatecCstdS)z Return whether the id of the current commit equals the given name. Args: dest: the repository directory. name: a string name. Nrw)rcrr!r#r#r$is_commit_id_equals z!VersionControl.is_commit_id_equalc Cs||\}}tj|s,||||dS|}||r||}|||j rt d|j t|||||jst dt||j |||||n t ddSt d|j|j t||d}nt d||j|j d}t d |j|td |d |d }|d kr$td|dkrXt dt|t|||||dS|dkrt|}t dt||t||||||dS|dkrt d|j t|||||||dS)a Install or update in editable mode the package represented by this VersionControl object. :param dest: the repository directory in which to install or update. :param url: the repository URL starting with a vcs prefix. Nz)%s in %s exists, and has correct URL (%s)zUpdating %s %s%sz$Skipping because already up-to-date.z%s %s in %s exists with URL %s)z%(s)witch, (i)gnore, (w)ipe, (b)ackup )siwbz0Directory %s already exists, and is not a %s %s.)z(i)gnore, (w)ipe, (b)ackup )rrrz+The plan is to install the %s repository %szWhat to do? %srrarz Deleting %srzBacking up %s to %srzSwitching %s %s to %s%s)rr-r.r/rrIis_repository_directoryrurrr2rb repo_nametitler rr(inforr3r!rsysexitr rshutilZmover) r>rrrZ rev_displayZ existing_urlpromptZresponseZdest_dirr#r#r$obtains           zVersionControl.obtaincCs&tj|rt||j||ddS)z Clean up current location and download the url repository (and vcs infos) into location :param url: the repository URL starting with a vcs prefix. rN)r-r.r/r rr~r#r#r$unpack?s zVersionControl.unpackcCstdS)z Return the url used at location Raises RemoteNotFoundError if the repository does not have a remote url configured. Nrwrpr#r#r$ruKszVersionControl.get_remote_urlcCstdS)zR Return the current commit id of the files at the given location. Nrwrpr#r#r$rrUszVersionControl.get_revisionTraisec Cs|t|jf|}z t||||||||j|| d WStk rv} z(| jtjkrdtd|j|jfnW5d} ~ XYnXdS)z Run a VCS subcommand This is simply a wrapper around call_subprocess that adds the VCS command name, and checks that the VCS is available ) on_returncodeextra_ok_returncodes command_desc extra_environ unset_environspinnerlog_failed_cmdzCCannot find command %r - do you have %r installed and in your PATH?N)rr!r rOSErrorerrnoZENOENTr) rccmdZ show_stdoutcwdrrrrrrer#r#r$ run_command\s&  zVersionControl.run_commandcCs,td||j|jtjtj||jS)zL Return whether a directory path is a repository directory. zChecking in %s for %s (%s)...)r2rbr1r!r-r.r/r0)rcr.r#r#r$rs z&VersionControl.is_repository_directorycCs ||S)a6 Check if a location is controlled by the vcs. It is meant to be overridden to implement smarter detection mechanisms for specific vcs. This can do more than is_repository_directory() alone. For example, the Git override checks that Git is actually available. )rrpr#r#r$rgs z VersionControl.controls_location)NN)TNrNNNNT)$r8r9r:r!r1rrSrrC classmethodrorqrtrv staticmethodrErKr}rrrrrrrrrrrrrrurrrrrgr#r#r#r$rmsr                 ]    ' rm)N)>rMZ __future__rrZloggingr-rrZ pip._vendorrZpip._vendor.six.moves.urllibrrQZpip._internal.exceptionsrZpip._internal.utils.compatrZpip._internal.utils.miscrrr r r r Zpip._internal.utils.subprocessr rZpip._internal.utils.typingrZpip._internal.utils.urlsrtypingrrrrrrrrrrZpip._internal.utils.uirrrstrZAuthInfo__all__Z getLoggerr8r2r%r,r6 Exceptionr7objectr;rOrrmr#r#r#r$s<        0     HGsite-packages/pip/_internal/vcs/__pycache__/versioncontrol.cpython-38.opt-1.pyc000064400000044035151464452710023526 0ustar00U .e~S@sdZddlmZddlZddlZddlZddlZddlZddlm Z ddl m Z ddl mZddlmZddlmZmZmZmZmZmZdd lmZmZdd lmZdd lmZerdd lm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)dd l*m+Z+ddlm,Z,ddlm-Z-e'e%e.e%e.fZ/dgZ0e1e2Z3ddZ4dddZ5ddZ6Gddde7Z8Gddde9Z:Gddde9Z;e;Z)r&r=r!r(r>r#r#r$__repr__szRevOptions.__repr__cCs|jdkr|jjS|jSN)r(r=default_arg_revr@r#r#r$arg_revs zRevOptions.arg_revcCs0g}|j}|dk r"||j|7}||j7}|S)z< Return the VCS-specific command arguments. N)rDr=get_base_rev_argsr<)r>argsr(r#r#r$to_argss  zRevOptions.to_argscCs|js dSd|jS)Nz (to revision {}))r(r&r@r#r#r$ to_displayszRevOptions.to_displaycCs|jj||jdS)z Make a copy of the current instance, but with a new rev. Args: rev: the name of the revision for the new object. r<)r=make_rev_optionsr<)r>r(r#r#r$make_newszRevOptions.make_new)NN) r8r9r:__doc__r?rApropertyrDrGrIrLr#r#r#r$r;js    r;cs|eZdZiZddddddgZfddZd d Zed d Zed dZ eddZ ddZ ddZ ddZ ddZZS) VcsSupportZsshZgitZhgZbzrZsftpZsvncs:tj|jttddr(tj|jtt|dS)N uses_fragment) urllib_parseZ uses_netlocextendschemesgetattrrPsuperrOr?r@ __class__r#r$r?s zVcsSupport.__init__cCs |jSrB) _registry__iter__r@r#r#r$rYszVcsSupport.__iter__cCst|jSrB)listrXvaluesr@r#r#r$backendsszVcsSupport.backendscCsdd|jDS)NcSsg|] }|jqSr#)r1).0backendr#r#r$ sz'VcsSupport.dirnames..)r\r@r#r#r$dirnamesszVcsSupport.dirnamescCs g}|jD]}||jq |SrB)r\rRrS)r>rSr^r#r#r$r s zVcsSupport.all_schemescCsHt|dstd|jdS|j|jkrD||j|j<td|jdS)Nr!zCannot register VCS %szRegistered VCS backend: %s)hasattrr2r3r8r!rXdebug)r>clsr#r#r$registers   zVcsSupport.registercCs||jkr|j|=dSrB)rXr>r!r#r#r$ unregisters zVcsSupport.unregistercCs6|jD]&}||r td||j|Sq dS)zv Return a VersionControl object if a repository of that type is found at the given directory. zDetermine that %s uses VCS: %sN)rXr[controls_locationr2rbr!)r>r5Z vcs_backendr#r#r$get_backend_for_dirs  zVcsSupport.get_backend_for_dircCs|}|j|S)z9 Return a VersionControl object or None. )lowerrXgetrer#r#r$ get_backendszVcsSupport.get_backend)r8r9r:rXrSr?rYrNr\r`r rdrfrhrk __classcell__r#r#rVr$rOs      rOc @s8eZdZdZdZdZdZdZdZe ddZ e ddZ e dd Z e d d Z ed d Ze d8ddZe ddZddZe ddZe ddZeddZddZeddZe ddZd d!Zd"d#Zd$d%Ze d&d'Zd(d)Zd*d+Ze d,d-Ze d.d/Z e d9d2d3Z!e d4d5Z"e d6d7Z#dS):VersionControlrHr#NcCs|d|j S)z Return whether the vcs prefix (e.g. "git+") should be added to a repository's remote url when used in a requirement. z{}:)ri startswithr&r!)rcZ remote_urlr#r#r$should_add_vcs_url_prefixsz(VersionControl.should_add_vcs_url_prefixcCsdS)z~ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. Nr#rcr5r#r#r$get_subdirectoryszVersionControl.get_subdirectorycCs ||S)zR Return the revision string that should be used in a requirement. ) get_revision)rcrepo_dirr#r#r$get_requirement_revisionsz'VersionControl.get_requirement_revisioncCsV||}|dkrdS||r.d|j|}||}||}t||||d}|S)aC Return the requirement string to use to redownload the files currently at the given repository directory. Args: project_name: the (unescaped) project name. The return value has a form similar to the following: {repository_url}@{revision}#egg={project_name} Nz{}+{})r*)get_remote_urlror&r!rtrqr,)rcrsr)r'Zrevisionr*r+r#r#r$get_src_requirements    z"VersionControl.get_src_requirementcCstdS)z Return the base revision arguments for a vcs command. Args: rev: the name of a revision to install. Cannot be None. NNotImplementedError)r(r#r#r$rE8sz VersionControl.get_base_rev_argscCst|||dS)z Return a RevOptions object. Args: rev: the name of a revision to install. extra_args: a list of extra options. rJ)r;)rcr(r<r#r#r$rKBs zVersionControl.make_rev_optionscCs&tj|\}}|tjjp$t|S)zy posix absolute paths start with os.path.sep, win32 ones start with drive (like c:\folder) )r-r. splitdrivernsepbool)rcZrepoZdrivetailr#r#r$_is_local_repositoryNsz#VersionControl._is_local_repositorycCstdS)z Export the repository at the url to the destination location i.e. only download the files, without vcs informations :param url: the repository URL starting with a vcs prefix. Nrwr>r5urlr#r#r$exportXszVersionControl.exportcCs|dfS)aZ Parse the repository URL's netloc, and return the new netloc to use along with auth information. Args: netloc: the original repository URL netloc. scheme: the repository URL's scheme without the vcs prefix. This is mainly for the Subversion class to override, so that auth information can be provided via the --username and --password options instead of through the URL. For other subclasses like Git without such an option, auth information must stay in the URL. Returns: (netloc, (username, password)). )NNr#)rcnetlocr"r#r#r$get_netloc_and_authbsz"VersionControl.get_netloc_and_authc Cst|\}}}}}d|kr*td||ddd}|||\}}d}d|krf|dd\}}t||||df}|||fS)z Parse the repository URL to use, and return the URL, revision, and auth info to use. Returns: (url, rev, (username, password)). +zvSorry, {!r} is a malformed VCS url. The format is +://, e.g. svn+http://myrepo/svn/MyApp#egg=MyAppN@rH)rQZurlsplit ValueErrorr&splitrrsplitZ urlunsplit) rcrr"rr.ZqueryZfrag user_passr(r#r#r$get_url_rev_and_authus z#VersionControl.get_url_rev_and_authcCsgS)zM Return the RevOptions "extra arguments" to use in obtain(). r#)usernamepasswordr#r#r$ make_rev_argsszVersionControl.make_rev_argsc CsT||j\}}}|\}}d}|dk r.t|}|||}|j||d} t|| fS)z Return the URL and RevOptions object to use in obtain() and in some cases export(), as a tuple (url, rev_options). NrJ)rsecretr rrKr ) r>rZ secret_urlr(rrZsecret_passwordrr< rev_optionsr#r#r$get_url_rev_optionss z"VersionControl.get_url_rev_optionscCst|dS)zi Normalize a URL for comparison by unquoting it and removing any trailing slash. /)rQZunquoterstriprr#r#r$ normalize_urlszVersionControl.normalize_urlcCs||||kS)zV Compare two repo URLs for identity, ignoring incidental differences. )r)rcZurl1Zurl2r#r#r$ compare_urlsszVersionControl.compare_urlscCstdS)z Fetch a revision from a repository, in the case that this is the first fetch from the repository. Args: dest: the directory to fetch the repository to. rev_options: a RevOptions object. Nrwr>destrrr#r#r$ fetch_news zVersionControl.fetch_newcCstdS)z} Switch the repo at ``dest`` to point to ``URL``. Args: rev_options: a RevOptions object. Nrwrr#r#r$switchszVersionControl.switchcCstdS)z Update an already-existing repo to the given ``rev_options``. Args: rev_options: a RevOptions object. Nrwrr#r#r$updateszVersionControl.updatecCstdS)z Return whether the id of the current commit equals the given name. Args: dest: the repository directory. name: a string name. Nrw)rcrr!r#r#r$is_commit_id_equals z!VersionControl.is_commit_id_equalc Cs||\}}tj|s,||||dS|}||r||}|||j rt d|j t|||||jst dt||j |||||n t ddSt d|j|j t||d}nt d||j|j d}t d |j|td |d |d }|d kr$td|dkrXt dt|t|||||dS|dkrt|}t dt||t||||||dS|dkrt d|j t|||||||dS)a Install or update in editable mode the package represented by this VersionControl object. :param dest: the repository directory in which to install or update. :param url: the repository URL starting with a vcs prefix. Nz)%s in %s exists, and has correct URL (%s)zUpdating %s %s%sz$Skipping because already up-to-date.z%s %s in %s exists with URL %s)z%(s)witch, (i)gnore, (w)ipe, (b)ackup )siwbz0Directory %s already exists, and is not a %s %s.)z(i)gnore, (w)ipe, (b)ackup )rrrz+The plan is to install the %s repository %szWhat to do? %srrarz Deleting %srzBacking up %s to %srzSwitching %s %s to %s%s)rr-r.r/rrIis_repository_directoryrurrr2rb repo_nametitler rr(inforr3r!rsysexitr rshutilZmover) r>rrrZ rev_displayZ existing_urlpromptZresponseZdest_dirr#r#r$obtains           zVersionControl.obtaincCs&tj|rt||j||ddS)z Clean up current location and download the url repository (and vcs infos) into location :param url: the repository URL starting with a vcs prefix. rN)r-r.r/r rr~r#r#r$unpack?s zVersionControl.unpackcCstdS)z Return the url used at location Raises RemoteNotFoundError if the repository does not have a remote url configured. Nrwrpr#r#r$ruKszVersionControl.get_remote_urlcCstdS)zR Return the current commit id of the files at the given location. Nrwrpr#r#r$rrUszVersionControl.get_revisionTraisec Cs|t|jf|}z t||||||||j|| d WStk rv} z(| jtjkrdtd|j|jfnW5d} ~ XYnXdS)z Run a VCS subcommand This is simply a wrapper around call_subprocess that adds the VCS command name, and checks that the VCS is available ) on_returncodeextra_ok_returncodes command_desc extra_environ unset_environspinnerlog_failed_cmdzCCannot find command %r - do you have %r installed and in your PATH?N)rr!r rOSErrorerrnoZENOENTr) rccmdZ show_stdoutcwdrrrrrrer#r#r$ run_command\s&  zVersionControl.run_commandcCs,td||j|jtjtj||jS)zL Return whether a directory path is a repository directory. zChecking in %s for %s (%s)...)r2rbr1r!r-r.r/r0)rcr.r#r#r$rs z&VersionControl.is_repository_directorycCs ||S)a6 Check if a location is controlled by the vcs. It is meant to be overridden to implement smarter detection mechanisms for specific vcs. This can do more than is_repository_directory() alone. For example, the Git override checks that Git is actually available. )rrpr#r#r$rgs z VersionControl.controls_location)NN)TNrNNNNT)$r8r9r:r!r1rrSrrC classmethodrorqrtrv staticmethodrErKr}rrrrrrrrrrrrrrurrrrrgr#r#r#r$rmsr                 ]    ' rm)N)>rMZ __future__rrZloggingr-rrZ pip._vendorrZpip._vendor.six.moves.urllibrrQZpip._internal.exceptionsrZpip._internal.utils.compatrZpip._internal.utils.miscrrr r r r Zpip._internal.utils.subprocessr rZpip._internal.utils.typingrZpip._internal.utils.urlsrtypingrrrrrrrrrrZpip._internal.utils.uirrrstrZAuthInfo__all__Z getLoggerr8r2r%r,r6 Exceptionr7objectr;rOrrmr#r#r#r$s<        0     HGsite-packages/pip/_internal/vcs/__pycache__/git.cpython-38.pyc000064400000022005151464452710020255 0ustar00U .e5@sddlmZddlZddlZddlZddlmZddl mZ ddl m Z ddl mZddlmZddlmZddlmZdd lmZdd lmZmZmZmZerdd lmZmZdd lmZdd lm Z m!Z!e j"Z"e j#Z#e$e%Z&e'dZ(ddZ)GdddeZ*e+e*dS))absolute_importN)parse)request) BadCommand) display_path) make_command) TempDirectory)MYPY_CHECK_RUNNING)RemoteNotFoundErrorVersionControl!find_path_to_setup_from_repo_rootvcs)OptionalTuple) HiddenText)AuthInfo RevOptionsz^[a-fA-F0-9]{40}$cCstt|SN)bool HASH_REGEXmatch)shar9/usr/lib/python3.8/site-packages/pip/_internal/vcs/git.pylooks_like_hash*srcseZdZdZdZdZdZdZdZe ddZ d d Z e d d Z d dZe ddZe ddZe ddZddZddZddZe ddZe d(ddZe d d!Ze fd"d#Ze d$d%Ze fd&d'ZZS))Gitgitz.gitclone)rzgit+httpz git+httpszgit+sshzgit+gitzgit+file)ZGIT_DIRZ GIT_WORK_TREEHEADcCs|gSrrrevrrrget_base_rev_args:szGit.get_base_rev_argscCs\d}|jdgdd}||r8|t|dd}nd}d|ddd}t|S) Nz git version versionF) show_stdoutr.) run_command startswithlensplitjoin parse_version)selfZ VERSION_PFXr"rrrget_git_version>s zGit.get_git_versioncCsBdddg}|j|dd|d}|}|dr>|tddSdS) zl Return the current branch, or None if HEAD isn't at a branch (e.g. detached HEAD). z symbolic-ref-qrFZextra_ok_returncodesr#cwdz refs/heads/N)r'stripr(r))clslocationargsoutputrefrrrget_current_branchKs  zGit.get_current_branchc CsX|ds|d}tdd2}|j|j|d|jdddd|gd |jd W5QRXd S) z@Export the Git repository at the url to the destination location/export)Zkind)urlzcheckout-indexz-az-fz--prefixFr#r3N)endswithrunpackpathr')r-r6r=Ztemp_dirrrrr<`s   z Git.exportc Cs|jd|g|ddd}i}|dD]V}|d}|s)r'r4)r5r6r Z current_revrrrrV%szGit.get_revisioncCsR|jddgd|d}tj|s2tj||}tjtj|d}t||S)z~ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. rjz --git-dirFr>z..)r'r4osrAisabsr+abspathr )r5r6Zgit_dirZ repo_rootrrrget_subdirectory.s  zGit.get_subdirectoryc st|\}}}}}|dr|dt|d }|t|ddd}t|||||f}|dd} |d| t|| d||||f}d|krd|kst |d d }t t | |\}} } |d d }nt t | |\}} } || | fS) a9 Prefixes stub URLs like 'user@hostname:user/repo.git' with 'ssh://'. That's required because although they use SSH they sometimes don't work with a ssh:// scheme (e.g. GitHub). But we need a scheme for parsing. Hence we remove it again afterwards and return it as a stub. fileNr;\+r1z://zfile:zgit+z git+ssh://zssh://r$) urlsplitr?r)lstripurllib_requestZ url2pathnamereplace urlunsplitfindrRsuperrget_url_rev_and_auth) r5r=ZschemeZnetlocrAZqueryZfragmentinitial_slashesnewpathZ after_plusr Z user_pass __class__rrry=s0      zGit.get_url_rev_and_authcCs6tjtj|dsdS|jdddddg|ddS)Nz .gitmodulesZ submodulerez--initz --recursiver/rQ)rkrAexistsr+r')r5r6rrrr^`s  zGit.update_submodulescsZtt||rdSz|jdg|dddd}| WStk rTtd|YdSXdS)NTrjFrB)r3r#rCZlog_failed_cmdzKcould not determine if %s is under git control because git is not available)rxrcontrols_locationr'rrTdebug)r5r6rr|rrriszGit.controls_location)N)__name__ __module__ __qualname__rYdirnameZ repo_nameZschemesZ unset_environZdefault_arg_rev staticmethodr!r. classmethodr:r<rOrXrZr`rcrerirVrnryr^r __classcell__rrr|rr.sB     ( ,     " r),Z __future__rZloggingZos.pathrkreZpip._vendor.packaging.versionrr,Zpip._vendor.six.moves.urllibZ urllib_parserrtZpip._internal.exceptionsrZpip._internal.utils.miscrZpip._internal.utils.subprocessrZpip._internal.utils.temp_dirrZpip._internal.utils.typingr Z pip._internal.vcs.versioncontrolr r r r typingrrrrrrrrvZ getLoggerrrTcompilerrrregisterrrrrs2            Nsite-packages/pip/_internal/vcs/__pycache__/mercurial.cpython-38.opt-1.pyc000064400000011430151464452710022414 0ustar00U .e@sddlmZddlZddlZddlmZddlmZmZddl m Z ddl m Z ddl mZddlmZdd lmZdd lmZmZmZerdd l mZdd lmZeeZGd ddeZeedS))absolute_importN) configparser) BadCommandInstallationError) display_path) make_command) TempDirectory)MYPY_CHECK_RUNNING) path_to_url)VersionControl!find_path_to_setup_from_repo_rootvcs) HiddenText) RevOptionscseZdZdZdZdZdZeddZddZ d d Z d d Z d dZ e ddZe ddZe ddZe ddZe ddZe fddZZS) Mercurialhgz.hgclone)rzhg+filezhg+httpzhg+httpszhg+sshzhg+static-httpcCs|gS)N)Zrevrr?/usr/lib/python3.8/site-packages/pip/_internal/vcs/mercurial.pyget_base_rev_args'szMercurial.get_base_rev_argsc Cs@tdd,}|j|j|d|jd|gd|jdW5QRXdS)z?Export the Hg repository at the url to the destination locationexport)Zkind)urlarchiveF show_stdoutcwdN)runpackpath run_command)selflocationrZtemp_dirrrrr+s zMercurial.exportcCsP|}td||t||tddd|||jtdd||ddS)NzCloning hg %s%s to %srz --noupdate-qupdater)Z to_displayloggerinforrrto_args)rdestr rev_optionsZ rev_displayrrr fetch_new5szMercurial.fetch_newc Cstj||jd}t}z>|||dd|jt |d}| |W5QRXWn6t tj fk r}zt d||W5d}~XYn Xtdd|}|j||ddS) NZhgrcpathsdefaultwz/Could not switch Mercurial repository to %s: %sr"r!r#)osrjoindirnamerZRawConfigParserreadsetZsecretopenwriteOSErrorZNoSectionErrorr$Zwarningrr&r) rr'rr(Z repo_configZconfigZ config_fileexccmd_argsrrrswitchDs  zMercurial.switchcCs4|jddg|dtdd|}|j||ddS)NZpullr!r#r")rrr&)rr'rr(r6rrrr"UszMercurial.updatecCs2|jddgd|d}||r*t|}|S)NZ showconfigz paths.defaultFr)rstripZ_is_local_repositoryr )clsr rrrrget_remote_url[s  zMercurial.get_remote_urlcCs|jddgd|d}|S)zW Return the repository-local changeset revision number, as an integer. parentsz--template={rev}Frrr8)r9r Zcurrent_revisionrrr get_revisionds  zMercurial.get_revisioncCs|jddgd|d}|S)zh Return the changeset identification hash, as a 40-character hexadecimal string r;z--template={node}Frr<)r9r Zcurrent_rev_hashrrrget_requirement_revisionns  z"Mercurial.get_requirement_revisioncCsdS)z&Always assume the versions don't matchFr)r9r'namerrris_commit_id_equalyszMercurial.is_commit_id_equalcCsB|jdgd|d}tj|s8tjtj||}t||S)z~ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. rootFr)rr8r-risabsabspathr.r )r9r Z repo_rootrrrget_subdirectory~s  zMercurial.get_subdirectoryc sPtt||rdSz|jdg|ddddWdSttfk rJYdSXdS)NTZidentifyFraise)rrZ on_returncodeZlog_failed_cmd)superrcontrols_locationrrr)r9r  __class__rrrGszMercurial.controls_location)__name__ __module__ __qualname__r?r/Z repo_nameZschemes staticmethodrrr)r7r" classmethodr:r=r>r@rDrG __classcell__rrrHrrs,       r)Z __future__rZloggingr-Zpip._vendor.six.movesrZpip._internal.exceptionsrrZpip._internal.utils.miscrZpip._internal.utils.subprocessrZpip._internal.utils.temp_dirrZpip._internal.utils.typingr Zpip._internal.utils.urlsr Z pip._internal.vcs.versioncontrolr r r rrZ getLoggerrJr$rregisterrrrrs          |site-packages/pip/_internal/vcs/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000700151464452710022166 0ustar00U .ei@s<ddlZddlZddlZddlZddlmZmZmZm Z dS)N)RemoteNotFoundErroris_urlmake_vcs_requirement_urlvcs) Zpip._internal.vcs.bazaarZpipZpip._internal.vcs.gitZpip._internal.vcs.mercurialZpip._internal.vcs.subversionZ pip._internal.vcs.versioncontrolrrrrrr>/usr/lib/python3.8/site-packages/pip/_internal/vcs/__init__.pyssite-packages/pip/_internal/vcs/__pycache__/git.cpython-38.opt-1.pyc000064400000021715151464452710021223 0ustar00U .e5@sddlmZddlZddlZddlZddlmZddl mZ ddl m Z ddl mZddlmZddlmZddlmZdd lmZdd lmZmZmZmZerdd lmZmZdd lmZdd lm Z m!Z!e j"Z"e j#Z#e$e%Z&e'dZ(ddZ)GdddeZ*e+e*dS))absolute_importN)parse)request) BadCommand) display_path) make_command) TempDirectory)MYPY_CHECK_RUNNING)RemoteNotFoundErrorVersionControl!find_path_to_setup_from_repo_rootvcs)OptionalTuple) HiddenText)AuthInfo RevOptionsz^[a-fA-F0-9]{40}$cCstt|SN)bool HASH_REGEXmatch)shar9/usr/lib/python3.8/site-packages/pip/_internal/vcs/git.pylooks_like_hash*srcseZdZdZdZdZdZdZdZe ddZ d d Z e d d Z d dZe ddZe ddZe ddZddZddZddZe ddZe d(ddZe d d!Ze fd"d#Ze d$d%Ze fd&d'ZZS))Gitgitz.gitclone)rzgit+httpz git+httpszgit+sshzgit+gitzgit+file)ZGIT_DIRZ GIT_WORK_TREEHEADcCs|gSrrrevrrrget_base_rev_args:szGit.get_base_rev_argscCs\d}|jdgdd}||r8|t|dd}nd}d|ddd}t|S) Nz git version versionF) show_stdoutr.) run_command startswithlensplitjoin parse_version)selfZ VERSION_PFXr"rrrget_git_version>s zGit.get_git_versioncCsBdddg}|j|dd|d}|}|dr>|tddSdS) zl Return the current branch, or None if HEAD isn't at a branch (e.g. detached HEAD). z symbolic-ref-qrFZextra_ok_returncodesr#cwdz refs/heads/N)r'stripr(r))clslocationargsoutputrefrrrget_current_branchKs  zGit.get_current_branchc CsX|ds|d}tdd2}|j|j|d|jdddd|gd |jd W5QRXd S) z@Export the Git repository at the url to the destination location/export)Zkind)urlzcheckout-indexz-az-fz--prefixFr#r3N)endswithrunpackpathr')r-r6r=Ztemp_dirrrrr<`s   z Git.exportc Cs|jd|g|ddd}i}|dD]V}|d}|s)r'r4)r5r6r Z current_revrrrrU%szGit.get_revisioncCsR|jddgd|d}tj|s2tj||}tjtj|d}t||S)z~ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. riz --git-dirFr>z..)r'r4osrAisabsr+abspathr )r5r6Zgit_dirZ repo_rootrrrget_subdirectory.s  zGit.get_subdirectoryc st|\}}}}}|dr|dt|d }|t|ddd}t|||||f}|dd} |d| t|| d||||f}d|kr|dd }t t | |\}} } |d d }nt t | |\}} } || | fS) a9 Prefixes stub URLs like 'user@hostname:user/repo.git' with 'ssh://'. That's required because although they use SSH they sometimes don't work with a ssh:// scheme (e.g. GitHub). But we need a scheme for parsing. Hence we remove it again afterwards and return it as a stub. fileNr;\+r1z://zgit+z git+ssh://zssh://r$) urlsplitr?r)lstripurllib_requestZ url2pathnamereplace urlunsplitfindsuperrget_url_rev_and_auth) r5r=ZschemeZnetlocrAZqueryZfragmentinitial_slashesnewpathZ after_plusr Z user_pass __class__rrrx=s.     zGit.get_url_rev_and_authcCs6tjtj|dsdS|jdddddg|ddS)Nz .gitmodulesZ submodulerdz--initz --recursiver/rQ)rjrAexistsr+r')r5r6rrrr]`s  zGit.update_submodulescsZtt||rdSz|jdg|dddd}| WStk rTtd|YdSXdS)NTriFrB)r3r#rCZlog_failed_cmdzKcould not determine if %s is under git control because git is not available)rwrcontrols_locationr'rrSdebug)r5r6rr{rrr~iszGit.controls_location)N)__name__ __module__ __qualname__rXdirnameZ repo_nameZschemesZ unset_environZdefault_arg_rev staticmethodr!r. classmethodr:r<rOrWrYr_rbrdrhrUrmrxr]r~ __classcell__rrr{rr.sB     ( ,     " r),Z __future__rZloggingZos.pathrjreZpip._vendor.packaging.versionrr,Zpip._vendor.six.moves.urllibZ urllib_parserrsZpip._internal.exceptionsrZpip._internal.utils.miscrZpip._internal.utils.subprocessrZpip._internal.utils.temp_dirrZpip._internal.utils.typingr Z pip._internal.vcs.versioncontrolr r r r typingrrrrrrqruZ getLoggerrrScompilerrrregisterrrrrs2            Nsite-packages/pip/_internal/vcs/versioncontrol.py000064400000051576151464452710016311 0ustar00"""Handles all VCS (version control) support""" # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import errno import logging import os import shutil import sys from pip._vendor import pkg_resources from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._internal.exceptions import BadCommand from pip._internal.utils.compat import samefile from pip._internal.utils.misc import ( ask_path_exists, backup_dir, display_path, hide_url, hide_value, rmtree, ) from pip._internal.utils.subprocess import call_subprocess, make_command from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.urls import get_url_scheme if MYPY_CHECK_RUNNING: from typing import ( Any, Dict, Iterable, List, Mapping, Optional, Text, Tuple, Type, Union ) from pip._internal.utils.ui import SpinnerInterface from pip._internal.utils.misc import HiddenText from pip._internal.utils.subprocess import CommandArgs AuthInfo = Tuple[Optional[str], Optional[str]] __all__ = ['vcs'] logger = logging.getLogger(__name__) def is_url(name): # type: (Union[str, Text]) -> bool """ Return true if the name looks like a URL. """ scheme = get_url_scheme(name) if scheme is None: return False return scheme in ['http', 'https', 'file', 'ftp'] + vcs.all_schemes def make_vcs_requirement_url(repo_url, rev, project_name, subdir=None): """ Return the URL for a VCS requirement. Args: repo_url: the remote VCS url, with any needed VCS prefix (e.g. "git+"). project_name: the (unescaped) project name. """ egg_project_name = pkg_resources.to_filename(project_name) req = '{}@{}#egg={}'.format(repo_url, rev, egg_project_name) if subdir: req += '&subdirectory={}'.format(subdir) return req def find_path_to_setup_from_repo_root(location, repo_root): """ Find the path to `setup.py` by searching up the filesystem from `location`. Return the path to `setup.py` relative to `repo_root`. Return None if `setup.py` is in `repo_root` or cannot be found. """ # find setup.py orig_location = location while not os.path.exists(os.path.join(location, 'setup.py')): last_location = location location = os.path.dirname(location) if location == last_location: # We've traversed up to the root of the filesystem without # finding setup.py logger.warning( "Could not find setup.py for directory %s (tried all " "parent directories)", orig_location, ) return None if samefile(repo_root, location): return None return os.path.relpath(location, repo_root) class RemoteNotFoundError(Exception): pass class RevOptions(object): """ Encapsulates a VCS-specific revision to install, along with any VCS install options. Instances of this class should be treated as if immutable. """ def __init__( self, vc_class, # type: Type[VersionControl] rev=None, # type: Optional[str] extra_args=None, # type: Optional[CommandArgs] ): # type: (...) -> None """ Args: vc_class: a VersionControl subclass. rev: the name of the revision to install. extra_args: a list of extra options. """ if extra_args is None: extra_args = [] self.extra_args = extra_args self.rev = rev self.vc_class = vc_class self.branch_name = None # type: Optional[str] def __repr__(self): return ''.format(self.vc_class.name, self.rev) @property def arg_rev(self): # type: () -> Optional[str] if self.rev is None: return self.vc_class.default_arg_rev return self.rev def to_args(self): # type: () -> CommandArgs """ Return the VCS-specific command arguments. """ args = [] # type: CommandArgs rev = self.arg_rev if rev is not None: args += self.vc_class.get_base_rev_args(rev) args += self.extra_args return args def to_display(self): # type: () -> str if not self.rev: return '' return ' (to revision {})'.format(self.rev) def make_new(self, rev): # type: (str) -> RevOptions """ Make a copy of the current instance, but with a new rev. Args: rev: the name of the revision for the new object. """ return self.vc_class.make_rev_options(rev, extra_args=self.extra_args) class VcsSupport(object): _registry = {} # type: Dict[str, VersionControl] schemes = ['ssh', 'git', 'hg', 'bzr', 'sftp', 'svn'] def __init__(self): # type: () -> None # Register more schemes with urlparse for various version control # systems urllib_parse.uses_netloc.extend(self.schemes) # Python >= 2.7.4, 3.3 doesn't have uses_fragment if getattr(urllib_parse, 'uses_fragment', None): urllib_parse.uses_fragment.extend(self.schemes) super(VcsSupport, self).__init__() def __iter__(self): return self._registry.__iter__() @property def backends(self): # type: () -> List[VersionControl] return list(self._registry.values()) @property def dirnames(self): # type: () -> List[str] return [backend.dirname for backend in self.backends] @property def all_schemes(self): # type: () -> List[str] schemes = [] # type: List[str] for backend in self.backends: schemes.extend(backend.schemes) return schemes def register(self, cls): # type: (Type[VersionControl]) -> None if not hasattr(cls, 'name'): logger.warning('Cannot register VCS %s', cls.__name__) return if cls.name not in self._registry: self._registry[cls.name] = cls() logger.debug('Registered VCS backend: %s', cls.name) def unregister(self, name): # type: (str) -> None if name in self._registry: del self._registry[name] def get_backend_for_dir(self, location): # type: (str) -> Optional[VersionControl] """ Return a VersionControl object if a repository of that type is found at the given directory. """ for vcs_backend in self._registry.values(): if vcs_backend.controls_location(location): logger.debug('Determine that %s uses VCS: %s', location, vcs_backend.name) return vcs_backend return None def get_backend(self, name): # type: (str) -> Optional[VersionControl] """ Return a VersionControl object or None. """ name = name.lower() return self._registry.get(name) vcs = VcsSupport() class VersionControl(object): name = '' dirname = '' repo_name = '' # List of supported schemes for this Version Control schemes = () # type: Tuple[str, ...] # Iterable of environment variable names to pass to call_subprocess(). unset_environ = () # type: Tuple[str, ...] default_arg_rev = None # type: Optional[str] @classmethod def should_add_vcs_url_prefix(cls, remote_url): """ Return whether the vcs prefix (e.g. "git+") should be added to a repository's remote url when used in a requirement. """ return not remote_url.lower().startswith('{}:'.format(cls.name)) @classmethod def get_subdirectory(cls, location): """ Return the path to setup.py, relative to the repo root. Return None if setup.py is in the repo root. """ return None @classmethod def get_requirement_revision(cls, repo_dir): """ Return the revision string that should be used in a requirement. """ return cls.get_revision(repo_dir) @classmethod def get_src_requirement(cls, repo_dir, project_name): """ Return the requirement string to use to redownload the files currently at the given repository directory. Args: project_name: the (unescaped) project name. The return value has a form similar to the following: {repository_url}@{revision}#egg={project_name} """ repo_url = cls.get_remote_url(repo_dir) if repo_url is None: return None if cls.should_add_vcs_url_prefix(repo_url): repo_url = '{}+{}'.format(cls.name, repo_url) revision = cls.get_requirement_revision(repo_dir) subdir = cls.get_subdirectory(repo_dir) req = make_vcs_requirement_url(repo_url, revision, project_name, subdir=subdir) return req @staticmethod def get_base_rev_args(rev): """ Return the base revision arguments for a vcs command. Args: rev: the name of a revision to install. Cannot be None. """ raise NotImplementedError @classmethod def make_rev_options(cls, rev=None, extra_args=None): # type: (Optional[str], Optional[CommandArgs]) -> RevOptions """ Return a RevOptions object. Args: rev: the name of a revision to install. extra_args: a list of extra options. """ return RevOptions(cls, rev, extra_args=extra_args) @classmethod def _is_local_repository(cls, repo): # type: (str) -> bool """ posix absolute paths start with os.path.sep, win32 ones start with drive (like c:\\folder) """ drive, tail = os.path.splitdrive(repo) return repo.startswith(os.path.sep) or bool(drive) def export(self, location, url): # type: (str, HiddenText) -> None """ Export the repository at the url to the destination location i.e. only download the files, without vcs informations :param url: the repository URL starting with a vcs prefix. """ raise NotImplementedError @classmethod def get_netloc_and_auth(cls, netloc, scheme): """ Parse the repository URL's netloc, and return the new netloc to use along with auth information. Args: netloc: the original repository URL netloc. scheme: the repository URL's scheme without the vcs prefix. This is mainly for the Subversion class to override, so that auth information can be provided via the --username and --password options instead of through the URL. For other subclasses like Git without such an option, auth information must stay in the URL. Returns: (netloc, (username, password)). """ return netloc, (None, None) @classmethod def get_url_rev_and_auth(cls, url): # type: (str) -> Tuple[str, Optional[str], AuthInfo] """ Parse the repository URL to use, and return the URL, revision, and auth info to use. Returns: (url, rev, (username, password)). """ scheme, netloc, path, query, frag = urllib_parse.urlsplit(url) if '+' not in scheme: raise ValueError( "Sorry, {!r} is a malformed VCS url. " "The format is +://, " "e.g. svn+http://myrepo/svn/MyApp#egg=MyApp".format(url) ) # Remove the vcs prefix. scheme = scheme.split('+', 1)[1] netloc, user_pass = cls.get_netloc_and_auth(netloc, scheme) rev = None if '@' in path: path, rev = path.rsplit('@', 1) url = urllib_parse.urlunsplit((scheme, netloc, path, query, '')) return url, rev, user_pass @staticmethod def make_rev_args(username, password): # type: (Optional[str], Optional[HiddenText]) -> CommandArgs """ Return the RevOptions "extra arguments" to use in obtain(). """ return [] def get_url_rev_options(self, url): # type: (HiddenText) -> Tuple[HiddenText, RevOptions] """ Return the URL and RevOptions object to use in obtain() and in some cases export(), as a tuple (url, rev_options). """ secret_url, rev, user_pass = self.get_url_rev_and_auth(url.secret) username, secret_password = user_pass password = None # type: Optional[HiddenText] if secret_password is not None: password = hide_value(secret_password) extra_args = self.make_rev_args(username, password) rev_options = self.make_rev_options(rev, extra_args=extra_args) return hide_url(secret_url), rev_options @staticmethod def normalize_url(url): # type: (str) -> str """ Normalize a URL for comparison by unquoting it and removing any trailing slash. """ return urllib_parse.unquote(url).rstrip('/') @classmethod def compare_urls(cls, url1, url2): # type: (str, str) -> bool """ Compare two repo URLs for identity, ignoring incidental differences. """ return (cls.normalize_url(url1) == cls.normalize_url(url2)) def fetch_new(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None """ Fetch a revision from a repository, in the case that this is the first fetch from the repository. Args: dest: the directory to fetch the repository to. rev_options: a RevOptions object. """ raise NotImplementedError def switch(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None """ Switch the repo at ``dest`` to point to ``URL``. Args: rev_options: a RevOptions object. """ raise NotImplementedError def update(self, dest, url, rev_options): # type: (str, HiddenText, RevOptions) -> None """ Update an already-existing repo to the given ``rev_options``. Args: rev_options: a RevOptions object. """ raise NotImplementedError @classmethod def is_commit_id_equal(cls, dest, name): """ Return whether the id of the current commit equals the given name. Args: dest: the repository directory. name: a string name. """ raise NotImplementedError def obtain(self, dest, url): # type: (str, HiddenText) -> None """ Install or update in editable mode the package represented by this VersionControl object. :param dest: the repository directory in which to install or update. :param url: the repository URL starting with a vcs prefix. """ url, rev_options = self.get_url_rev_options(url) if not os.path.exists(dest): self.fetch_new(dest, url, rev_options) return rev_display = rev_options.to_display() if self.is_repository_directory(dest): existing_url = self.get_remote_url(dest) if self.compare_urls(existing_url, url.secret): logger.debug( '%s in %s exists, and has correct URL (%s)', self.repo_name.title(), display_path(dest), url, ) if not self.is_commit_id_equal(dest, rev_options.rev): logger.info( 'Updating %s %s%s', display_path(dest), self.repo_name, rev_display, ) self.update(dest, url, rev_options) else: logger.info('Skipping because already up-to-date.') return logger.warning( '%s %s in %s exists with URL %s', self.name, self.repo_name, display_path(dest), existing_url, ) prompt = ('(s)witch, (i)gnore, (w)ipe, (b)ackup ', ('s', 'i', 'w', 'b')) else: logger.warning( 'Directory %s already exists, and is not a %s %s.', dest, self.name, self.repo_name, ) # https://github.com/python/mypy/issues/1174 prompt = ('(i)gnore, (w)ipe, (b)ackup ', # type: ignore ('i', 'w', 'b')) logger.warning( 'The plan is to install the %s repository %s', self.name, url, ) response = ask_path_exists('What to do? %s' % prompt[0], prompt[1]) if response == 'a': sys.exit(-1) if response == 'w': logger.warning('Deleting %s', display_path(dest)) rmtree(dest) self.fetch_new(dest, url, rev_options) return if response == 'b': dest_dir = backup_dir(dest) logger.warning( 'Backing up %s to %s', display_path(dest), dest_dir, ) shutil.move(dest, dest_dir) self.fetch_new(dest, url, rev_options) return # Do nothing if the response is "i". if response == 's': logger.info( 'Switching %s %s to %s%s', self.repo_name, display_path(dest), url, rev_display, ) self.switch(dest, url, rev_options) def unpack(self, location, url): # type: (str, HiddenText) -> None """ Clean up current location and download the url repository (and vcs infos) into location :param url: the repository URL starting with a vcs prefix. """ if os.path.exists(location): rmtree(location) self.obtain(location, url=url) @classmethod def get_remote_url(cls, location): """ Return the url used at location Raises RemoteNotFoundError if the repository does not have a remote url configured. """ raise NotImplementedError @classmethod def get_revision(cls, location): """ Return the current commit id of the files at the given location. """ raise NotImplementedError @classmethod def run_command( cls, cmd, # type: Union[List[str], CommandArgs] show_stdout=True, # type: bool cwd=None, # type: Optional[str] on_returncode='raise', # type: str extra_ok_returncodes=None, # type: Optional[Iterable[int]] command_desc=None, # type: Optional[str] extra_environ=None, # type: Optional[Mapping[str, Any]] spinner=None, # type: Optional[SpinnerInterface] log_failed_cmd=True ): # type: (...) -> Text """ Run a VCS subcommand This is simply a wrapper around call_subprocess that adds the VCS command name, and checks that the VCS is available """ cmd = make_command(cls.name, *cmd) try: return call_subprocess(cmd, show_stdout, cwd, on_returncode=on_returncode, extra_ok_returncodes=extra_ok_returncodes, command_desc=command_desc, extra_environ=extra_environ, unset_environ=cls.unset_environ, spinner=spinner, log_failed_cmd=log_failed_cmd) except OSError as e: # errno.ENOENT = no such file or directory # In other words, the VCS executable isn't available if e.errno == errno.ENOENT: raise BadCommand( 'Cannot find command %r - do you have ' '%r installed and in your ' 'PATH?' % (cls.name, cls.name)) else: raise # re-raise exception if a different error occurred @classmethod def is_repository_directory(cls, path): # type: (str) -> bool """ Return whether a directory path is a repository directory. """ logger.debug('Checking in %s for %s (%s)...', path, cls.dirname, cls.name) return os.path.exists(os.path.join(path, cls.dirname)) @classmethod def controls_location(cls, location): # type: (str) -> bool """ Check if a location is controlled by the vcs. It is meant to be overridden to implement smarter detection mechanisms for specific vcs. This can do more than is_repository_directory() alone. For example, the Git override checks that Git is actually available. """ return cls.is_repository_directory(location) site-packages/pip/_internal/vcs/__init__.py000064400000001151151464452710014742 0ustar00# Expose a limited set of classes and functions so callers outside of # the vcs package don't need to import deeper than `pip._internal.vcs`. # (The test directory and imports protected by MYPY_CHECK_RUNNING may # still need to import from a vcs sub-package.) # Import all vcs modules to register each VCS in the VcsSupport object. import pip._internal.vcs.bazaar import pip._internal.vcs.git import pip._internal.vcs.mercurial import pip._internal.vcs.subversion # noqa: F401 from pip._internal.vcs.versioncontrol import ( # noqa: F401 RemoteNotFoundError, is_url, make_vcs_requirement_url, vcs, ) site-packages/pip/_internal/index.py000064400000110027151464452710013522 0ustar00"""Routines related to PyPI, indexes""" # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import re from pip._vendor.packaging import specifiers from pip._vendor.packaging.utils import canonicalize_name from pip._vendor.packaging.version import parse as parse_version from pip._internal.exceptions import ( BestVersionAlreadyInstalled, DistributionNotFound, InvalidWheelFilename, UnsupportedWheel, ) from pip._internal.models.candidate import InstallationCandidate from pip._internal.models.format_control import FormatControl from pip._internal.models.link import Link from pip._internal.models.selection_prefs import SelectionPreferences from pip._internal.models.target_python import TargetPython from pip._internal.utils.filetypes import WHEEL_EXTENSION from pip._internal.utils.logging import indent_log from pip._internal.utils.misc import build_netloc from pip._internal.utils.packaging import check_requires_python from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.unpacking import SUPPORTED_EXTENSIONS from pip._internal.utils.urls import url_to_path from pip._internal.wheel import Wheel if MYPY_CHECK_RUNNING: from typing import ( FrozenSet, Iterable, List, Optional, Set, Text, Tuple, Union, ) from pip._vendor.packaging.version import _BaseVersion from pip._internal.collector import LinkCollector from pip._internal.models.search_scope import SearchScope from pip._internal.req import InstallRequirement from pip._internal.pep425tags import Pep425Tag from pip._internal.utils.hashes import Hashes BuildTag = Union[Tuple[()], Tuple[int, str]] CandidateSortingKey = ( Tuple[int, int, int, _BaseVersion, BuildTag, Optional[int]] ) __all__ = ['FormatControl', 'BestCandidateResult', 'PackageFinder'] logger = logging.getLogger(__name__) def _check_link_requires_python( link, # type: Link version_info, # type: Tuple[int, int, int] ignore_requires_python=False, # type: bool ): # type: (...) -> bool """ Return whether the given Python version is compatible with a link's "Requires-Python" value. :param version_info: A 3-tuple of ints representing the Python major-minor-micro version to check. :param ignore_requires_python: Whether to ignore the "Requires-Python" value if the given Python version isn't compatible. """ try: is_compatible = check_requires_python( link.requires_python, version_info=version_info, ) except specifiers.InvalidSpecifier: logger.debug( "Ignoring invalid Requires-Python (%r) for link: %s", link.requires_python, link, ) else: if not is_compatible: version = '.'.join(map(str, version_info)) if not ignore_requires_python: logger.debug( 'Link requires a different Python (%s not in: %r): %s', version, link.requires_python, link, ) return False logger.debug( 'Ignoring failed Requires-Python check (%s not in: %r) ' 'for link: %s', version, link.requires_python, link, ) return True class LinkEvaluator(object): """ Responsible for evaluating links for a particular project. """ _py_version_re = re.compile(r'-py([123]\.?[0-9]?)$') # Don't include an allow_yanked default value to make sure each call # site considers whether yanked releases are allowed. This also causes # that decision to be made explicit in the calling code, which helps # people when reading the code. def __init__( self, project_name, # type: str canonical_name, # type: str formats, # type: FrozenSet target_python, # type: TargetPython allow_yanked, # type: bool ignore_requires_python=None, # type: Optional[bool] ): # type: (...) -> None """ :param project_name: The user supplied package name. :param canonical_name: The canonical package name. :param formats: The formats allowed for this package. Should be a set with 'binary' or 'source' or both in it. :param target_python: The target Python interpreter to use when evaluating link compatibility. This is used, for example, to check wheel compatibility, as well as when checking the Python version, e.g. the Python version embedded in a link filename (or egg fragment) and against an HTML link's optional PEP 503 "data-requires-python" attribute. :param allow_yanked: Whether files marked as yanked (in the sense of PEP 592) are permitted to be candidates for install. :param ignore_requires_python: Whether to ignore incompatible PEP 503 "data-requires-python" values in HTML links. Defaults to False. """ if ignore_requires_python is None: ignore_requires_python = False self._allow_yanked = allow_yanked self._canonical_name = canonical_name self._ignore_requires_python = ignore_requires_python self._formats = formats self._target_python = target_python self.project_name = project_name def evaluate_link(self, link): # type: (Link) -> Tuple[bool, Optional[Text]] """ Determine whether a link is a candidate for installation. :return: A tuple (is_candidate, result), where `result` is (1) a version string if `is_candidate` is True, and (2) if `is_candidate` is False, an optional string to log the reason the link fails to qualify. """ version = None if link.is_yanked and not self._allow_yanked: reason = link.yanked_reason or '' # Mark this as a unicode string to prevent "UnicodeEncodeError: # 'ascii' codec can't encode character" in Python 2 when # the reason contains non-ascii characters. return (False, u'yanked for reason: {}'.format(reason)) if link.egg_fragment: egg_info = link.egg_fragment ext = link.ext else: egg_info, ext = link.splitext() if not ext: return (False, 'not a file') if ext not in SUPPORTED_EXTENSIONS: return (False, 'unsupported archive format: %s' % ext) if "binary" not in self._formats and ext == WHEEL_EXTENSION: reason = 'No binaries permitted for %s' % self.project_name return (False, reason) if "macosx10" in link.path and ext == '.zip': return (False, 'macosx10 one') if ext == WHEEL_EXTENSION: try: wheel = Wheel(link.filename) except InvalidWheelFilename: return (False, 'invalid wheel filename') if canonicalize_name(wheel.name) != self._canonical_name: reason = 'wrong project name (not %s)' % self.project_name return (False, reason) supported_tags = self._target_python.get_tags() if not wheel.supported(supported_tags): # Include the wheel's tags in the reason string to # simplify troubleshooting compatibility issues. file_tags = wheel.get_formatted_file_tags() reason = ( "none of the wheel's tags match: {}".format( ', '.join(file_tags) ) ) return (False, reason) version = wheel.version # This should be up by the self.ok_binary check, but see issue 2700. if "source" not in self._formats and ext != WHEEL_EXTENSION: return (False, 'No sources permitted for %s' % self.project_name) if not version: version = _extract_version_from_fragment( egg_info, self._canonical_name, ) if not version: return ( False, 'Missing project version for %s' % self.project_name, ) match = self._py_version_re.search(version) if match: version = version[:match.start()] py_version = match.group(1) if py_version != self._target_python.py_version: return (False, 'Python version is incorrect') supports_python = _check_link_requires_python( link, version_info=self._target_python.py_version_info, ignore_requires_python=self._ignore_requires_python, ) if not supports_python: # Return None for the reason text to suppress calling # _log_skipped_link(). return (False, None) logger.debug('Found link %s, version: %s', link, version) return (True, version) def filter_unallowed_hashes( candidates, # type: List[InstallationCandidate] hashes, # type: Hashes project_name, # type: str ): # type: (...) -> List[InstallationCandidate] """ Filter out candidates whose hashes aren't allowed, and return a new list of candidates. If at least one candidate has an allowed hash, then all candidates with either an allowed hash or no hash specified are returned. Otherwise, the given candidates are returned. Including the candidates with no hash specified when there is a match allows a warning to be logged if there is a more preferred candidate with no hash specified. Returning all candidates in the case of no matches lets pip report the hash of the candidate that would otherwise have been installed (e.g. permitting the user to more easily update their requirements file with the desired hash). """ if not hashes: logger.debug( 'Given no hashes to check %s links for project %r: ' 'discarding no candidates', len(candidates), project_name, ) # Make sure we're not returning back the given value. return list(candidates) matches_or_no_digest = [] # Collect the non-matches for logging purposes. non_matches = [] match_count = 0 for candidate in candidates: link = candidate.link if not link.has_hash: pass elif link.is_hash_allowed(hashes=hashes): match_count += 1 else: non_matches.append(candidate) continue matches_or_no_digest.append(candidate) if match_count: filtered = matches_or_no_digest else: # Make sure we're not returning back the given value. filtered = list(candidates) if len(filtered) == len(candidates): discard_message = 'discarding no candidates' else: discard_message = 'discarding {} non-matches:\n {}'.format( len(non_matches), '\n '.join(str(candidate.link) for candidate in non_matches) ) logger.debug( 'Checked %s links for project %r against %s hashes ' '(%s matches, %s no digest): %s', len(candidates), project_name, hashes.digest_count, match_count, len(matches_or_no_digest) - match_count, discard_message ) return filtered class CandidatePreferences(object): """ Encapsulates some of the preferences for filtering and sorting InstallationCandidate objects. """ def __init__( self, prefer_binary=False, # type: bool allow_all_prereleases=False, # type: bool ): # type: (...) -> None """ :param allow_all_prereleases: Whether to allow all pre-releases. """ self.allow_all_prereleases = allow_all_prereleases self.prefer_binary = prefer_binary class BestCandidateResult(object): """A collection of candidates, returned by `PackageFinder.find_best_candidate`. This class is only intended to be instantiated by CandidateEvaluator's `compute_best_candidate()` method. """ def __init__( self, candidates, # type: List[InstallationCandidate] applicable_candidates, # type: List[InstallationCandidate] best_candidate, # type: Optional[InstallationCandidate] ): # type: (...) -> None """ :param candidates: A sequence of all available candidates found. :param applicable_candidates: The applicable candidates. :param best_candidate: The most preferred candidate found, or None if no applicable candidates were found. """ assert set(applicable_candidates) <= set(candidates) if best_candidate is None: assert not applicable_candidates else: assert best_candidate in applicable_candidates self._applicable_candidates = applicable_candidates self._candidates = candidates self.best_candidate = best_candidate def iter_all(self): # type: () -> Iterable[InstallationCandidate] """Iterate through all candidates. """ return iter(self._candidates) def iter_applicable(self): # type: () -> Iterable[InstallationCandidate] """Iterate through the applicable candidates. """ return iter(self._applicable_candidates) class CandidateEvaluator(object): """ Responsible for filtering and sorting candidates for installation based on what tags are valid. """ @classmethod def create( cls, project_name, # type: str target_python=None, # type: Optional[TargetPython] prefer_binary=False, # type: bool allow_all_prereleases=False, # type: bool specifier=None, # type: Optional[specifiers.BaseSpecifier] hashes=None, # type: Optional[Hashes] ): # type: (...) -> CandidateEvaluator """Create a CandidateEvaluator object. :param target_python: The target Python interpreter to use when checking compatibility. If None (the default), a TargetPython object will be constructed from the running Python. :param specifier: An optional object implementing `filter` (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable versions. :param hashes: An optional collection of allowed hashes. """ if target_python is None: target_python = TargetPython() if specifier is None: specifier = specifiers.SpecifierSet() supported_tags = target_python.get_tags() return cls( project_name=project_name, supported_tags=supported_tags, specifier=specifier, prefer_binary=prefer_binary, allow_all_prereleases=allow_all_prereleases, hashes=hashes, ) def __init__( self, project_name, # type: str supported_tags, # type: List[Pep425Tag] specifier, # type: specifiers.BaseSpecifier prefer_binary=False, # type: bool allow_all_prereleases=False, # type: bool hashes=None, # type: Optional[Hashes] ): # type: (...) -> None """ :param supported_tags: The PEP 425 tags supported by the target Python in order of preference (most preferred first). """ self._allow_all_prereleases = allow_all_prereleases self._hashes = hashes self._prefer_binary = prefer_binary self._project_name = project_name self._specifier = specifier self._supported_tags = supported_tags def get_applicable_candidates( self, candidates, # type: List[InstallationCandidate] ): # type: (...) -> List[InstallationCandidate] """ Return the applicable candidates from a list of candidates. """ # Using None infers from the specifier instead. allow_prereleases = self._allow_all_prereleases or None specifier = self._specifier versions = { str(v) for v in specifier.filter( # We turn the version object into a str here because otherwise # when we're debundled but setuptools isn't, Python will see # packaging.version.Version and # pkg_resources._vendor.packaging.version.Version as different # types. This way we'll use a str as a common data interchange # format. If we stop using the pkg_resources provided specifier # and start using our own, we can drop the cast to str(). (str(c.version) for c in candidates), prereleases=allow_prereleases, ) } # Again, converting version to str to deal with debundling. applicable_candidates = [ c for c in candidates if str(c.version) in versions ] return filter_unallowed_hashes( candidates=applicable_candidates, hashes=self._hashes, project_name=self._project_name, ) def _sort_key(self, candidate): # type: (InstallationCandidate) -> CandidateSortingKey """ Function to pass as the `key` argument to a call to sorted() to sort InstallationCandidates by preference. Returns a tuple such that tuples sorting as greater using Python's default comparison operator are more preferred. The preference is as follows: First and foremost, candidates with allowed (matching) hashes are always preferred over candidates without matching hashes. This is because e.g. if the only candidate with an allowed hash is yanked, we still want to use that candidate. Second, excepting hash considerations, candidates that have been yanked (in the sense of PEP 592) are always less preferred than candidates that haven't been yanked. Then: If not finding wheels, they are sorted by version only. If finding wheels, then the sort order is by version, then: 1. existing installs 2. wheels ordered via Wheel.support_index_min(self._supported_tags) 3. source archives If prefer_binary was set, then all wheels are sorted above sources. Note: it was considered to embed this logic into the Link comparison operators, but then different sdist links with the same version, would have to be considered equal """ valid_tags = self._supported_tags support_num = len(valid_tags) build_tag = () # type: BuildTag binary_preference = 0 link = candidate.link if link.is_wheel: # can raise InvalidWheelFilename wheel = Wheel(link.filename) if not wheel.supported(valid_tags): raise UnsupportedWheel( "%s is not a supported wheel for this platform. It " "can't be sorted." % wheel.filename ) if self._prefer_binary: binary_preference = 1 pri = -(wheel.support_index_min(valid_tags)) if wheel.build_tag is not None: match = re.match(r'^(\d+)(.*)$', wheel.build_tag) build_tag_groups = match.groups() build_tag = (int(build_tag_groups[0]), build_tag_groups[1]) else: # sdist pri = -(support_num) has_allowed_hash = int(link.is_hash_allowed(self._hashes)) yank_value = -1 * int(link.is_yanked) # -1 for yanked. return ( has_allowed_hash, yank_value, binary_preference, candidate.version, build_tag, pri, ) def sort_best_candidate( self, candidates, # type: List[InstallationCandidate] ): # type: (...) -> Optional[InstallationCandidate] """ Return the best candidate per the instance's sort order, or None if no candidate is acceptable. """ if not candidates: return None best_candidate = max(candidates, key=self._sort_key) # Log a warning per PEP 592 if necessary before returning. link = best_candidate.link if link.is_yanked: reason = link.yanked_reason or '' msg = ( # Mark this as a unicode string to prevent # "UnicodeEncodeError: 'ascii' codec can't encode character" # in Python 2 when the reason contains non-ascii characters. u'The candidate selected for download or install is a ' 'yanked version: {candidate}\n' 'Reason for being yanked: {reason}' ).format(candidate=best_candidate, reason=reason) logger.warning(msg) return best_candidate def compute_best_candidate( self, candidates, # type: List[InstallationCandidate] ): # type: (...) -> BestCandidateResult """ Compute and return a `BestCandidateResult` instance. """ applicable_candidates = self.get_applicable_candidates(candidates) best_candidate = self.sort_best_candidate(applicable_candidates) return BestCandidateResult( candidates, applicable_candidates=applicable_candidates, best_candidate=best_candidate, ) class PackageFinder(object): """This finds packages. This is meant to match easy_install's technique for looking for packages, by reading pages and looking for appropriate links. """ def __init__( self, link_collector, # type: LinkCollector target_python, # type: TargetPython allow_yanked, # type: bool format_control=None, # type: Optional[FormatControl] candidate_prefs=None, # type: CandidatePreferences ignore_requires_python=None, # type: Optional[bool] ): # type: (...) -> None """ This constructor is primarily meant to be used by the create() class method and from tests. :param format_control: A FormatControl object, used to control the selection of source packages / binary packages when consulting the index and links. :param candidate_prefs: Options to use when creating a CandidateEvaluator object. """ if candidate_prefs is None: candidate_prefs = CandidatePreferences() format_control = format_control or FormatControl(set(), set()) self._allow_yanked = allow_yanked self._candidate_prefs = candidate_prefs self._ignore_requires_python = ignore_requires_python self._link_collector = link_collector self._target_python = target_python self.format_control = format_control # These are boring links that have already been logged somehow. self._logged_links = set() # type: Set[Link] # Don't include an allow_yanked default value to make sure each call # site considers whether yanked releases are allowed. This also causes # that decision to be made explicit in the calling code, which helps # people when reading the code. @classmethod def create( cls, link_collector, # type: LinkCollector selection_prefs, # type: SelectionPreferences target_python=None, # type: Optional[TargetPython] ): # type: (...) -> PackageFinder """Create a PackageFinder. :param selection_prefs: The candidate selection preferences, as a SelectionPreferences object. :param target_python: The target Python interpreter to use when checking compatibility. If None (the default), a TargetPython object will be constructed from the running Python. """ if target_python is None: target_python = TargetPython() candidate_prefs = CandidatePreferences( prefer_binary=selection_prefs.prefer_binary, allow_all_prereleases=selection_prefs.allow_all_prereleases, ) return cls( candidate_prefs=candidate_prefs, link_collector=link_collector, target_python=target_python, allow_yanked=selection_prefs.allow_yanked, format_control=selection_prefs.format_control, ignore_requires_python=selection_prefs.ignore_requires_python, ) @property def search_scope(self): # type: () -> SearchScope return self._link_collector.search_scope @search_scope.setter def search_scope(self, search_scope): # type: (SearchScope) -> None self._link_collector.search_scope = search_scope @property def find_links(self): # type: () -> List[str] return self._link_collector.find_links @property def index_urls(self): # type: () -> List[str] return self.search_scope.index_urls @property def trusted_hosts(self): # type: () -> Iterable[str] for host_port in self._link_collector.session.pip_trusted_origins: yield build_netloc(*host_port) @property def allow_all_prereleases(self): # type: () -> bool return self._candidate_prefs.allow_all_prereleases def set_allow_all_prereleases(self): # type: () -> None self._candidate_prefs.allow_all_prereleases = True def make_link_evaluator(self, project_name): # type: (str) -> LinkEvaluator canonical_name = canonicalize_name(project_name) formats = self.format_control.get_allowed_formats(canonical_name) return LinkEvaluator( project_name=project_name, canonical_name=canonical_name, formats=formats, target_python=self._target_python, allow_yanked=self._allow_yanked, ignore_requires_python=self._ignore_requires_python, ) def _sort_links(self, links): # type: (Iterable[Link]) -> List[Link] """ Returns elements of links in order, non-egg links first, egg links second, while eliminating duplicates """ eggs, no_eggs = [], [] seen = set() # type: Set[Link] for link in links: if link not in seen: seen.add(link) if link.egg_fragment: eggs.append(link) else: no_eggs.append(link) return no_eggs + eggs def _log_skipped_link(self, link, reason): # type: (Link, Text) -> None if link not in self._logged_links: # Mark this as a unicode string to prevent "UnicodeEncodeError: # 'ascii' codec can't encode character" in Python 2 when # the reason contains non-ascii characters. # Also, put the link at the end so the reason is more visible # and because the link string is usually very long. logger.debug(u'Skipping link: %s: %s', reason, link) self._logged_links.add(link) def get_install_candidate(self, link_evaluator, link): # type: (LinkEvaluator, Link) -> Optional[InstallationCandidate] """ If the link is a candidate for install, convert it to an InstallationCandidate and return it. Otherwise, return None. """ is_candidate, result = link_evaluator.evaluate_link(link) if not is_candidate: if result: self._log_skipped_link(link, reason=result) return None return InstallationCandidate( project=link_evaluator.project_name, link=link, # Convert the Text result to str since InstallationCandidate # accepts str. version=str(result), ) def evaluate_links(self, link_evaluator, links): # type: (LinkEvaluator, Iterable[Link]) -> List[InstallationCandidate] """ Convert links that are candidates to InstallationCandidate objects. """ candidates = [] for link in self._sort_links(links): candidate = self.get_install_candidate(link_evaluator, link) if candidate is not None: candidates.append(candidate) return candidates def find_all_candidates(self, project_name): # type: (str) -> List[InstallationCandidate] """Find all available InstallationCandidate for project_name This checks index_urls and find_links. All versions found are returned as an InstallationCandidate list. See LinkEvaluator.evaluate_link() for details on which files are accepted. """ collected_links = self._link_collector.collect_links(project_name) link_evaluator = self.make_link_evaluator(project_name) find_links_versions = self.evaluate_links( link_evaluator, links=collected_links.find_links, ) page_versions = [] for page_url, page_links in collected_links.pages.items(): logger.debug('Analyzing links from page %s', page_url) with indent_log(): new_versions = self.evaluate_links( link_evaluator, links=page_links, ) page_versions.extend(new_versions) file_versions = self.evaluate_links( link_evaluator, links=collected_links.files, ) if file_versions: file_versions.sort(reverse=True) logger.debug( 'Local files found: %s', ', '.join([ url_to_path(candidate.link.url) for candidate in file_versions ]) ) # This is an intentional priority ordering return file_versions + find_links_versions + page_versions def make_candidate_evaluator( self, project_name, # type: str specifier=None, # type: Optional[specifiers.BaseSpecifier] hashes=None, # type: Optional[Hashes] ): # type: (...) -> CandidateEvaluator """Create a CandidateEvaluator object to use. """ candidate_prefs = self._candidate_prefs return CandidateEvaluator.create( project_name=project_name, target_python=self._target_python, prefer_binary=candidate_prefs.prefer_binary, allow_all_prereleases=candidate_prefs.allow_all_prereleases, specifier=specifier, hashes=hashes, ) def find_best_candidate( self, project_name, # type: str specifier=None, # type: Optional[specifiers.BaseSpecifier] hashes=None, # type: Optional[Hashes] ): # type: (...) -> BestCandidateResult """Find matches for the given project and specifier. :param specifier: An optional object implementing `filter` (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable versions. :return: A `BestCandidateResult` instance. """ candidates = self.find_all_candidates(project_name) candidate_evaluator = self.make_candidate_evaluator( project_name=project_name, specifier=specifier, hashes=hashes, ) return candidate_evaluator.compute_best_candidate(candidates) def find_requirement(self, req, upgrade): # type: (InstallRequirement, bool) -> Optional[Link] """Try to find a Link matching req Expects req, an InstallRequirement and upgrade, a boolean Returns a Link if found, Raises DistributionNotFound or BestVersionAlreadyInstalled otherwise """ hashes = req.hashes(trust_internet=False) best_candidate_result = self.find_best_candidate( req.name, specifier=req.specifier, hashes=hashes, ) best_candidate = best_candidate_result.best_candidate installed_version = None # type: Optional[_BaseVersion] if req.satisfied_by is not None: installed_version = parse_version(req.satisfied_by.version) def _format_versions(cand_iter): # This repeated parse_version and str() conversion is needed to # handle different vendoring sources from pip and pkg_resources. # If we stop using the pkg_resources provided specifier and start # using our own, we can drop the cast to str(). return ", ".join(sorted( {str(c.version) for c in cand_iter}, key=parse_version, )) or "none" if installed_version is None and best_candidate is None: logger.critical( 'Could not find a version that satisfies the requirement %s ' '(from versions: %s)', req, _format_versions(best_candidate_result.iter_all()), ) raise DistributionNotFound( 'No matching distribution found for %s' % req ) best_installed = False if installed_version and ( best_candidate is None or best_candidate.version <= installed_version): best_installed = True if not upgrade and installed_version is not None: if best_installed: logger.debug( 'Existing installed version (%s) is most up-to-date and ' 'satisfies requirement', installed_version, ) else: logger.debug( 'Existing installed version (%s) satisfies requirement ' '(most up-to-date version is %s)', installed_version, best_candidate.version, ) return None if best_installed: # We have an existing version, and its the best version logger.debug( 'Installed version (%s) is most up-to-date (past versions: ' '%s)', installed_version, _format_versions(best_candidate_result.iter_applicable()), ) raise BestVersionAlreadyInstalled logger.debug( 'Using version %s (newest of versions: %s)', best_candidate.version, _format_versions(best_candidate_result.iter_applicable()), ) return best_candidate.link def _find_name_version_sep(fragment, canonical_name): # type: (str, str) -> int """Find the separator's index based on the package's canonical name. :param fragment: A + filename "fragment" (stem) or egg fragment. :param canonical_name: The package's canonical name. This function is needed since the canonicalized name does not necessarily have the same length as the egg info's name part. An example:: >>> fragment = 'foo__bar-1.0' >>> canonical_name = 'foo-bar' >>> _find_name_version_sep(fragment, canonical_name) 8 """ # Project name and version must be separated by one single dash. Find all # occurrences of dashes; if the string in front of it matches the canonical # name, this is the one separating the name and version parts. for i, c in enumerate(fragment): if c != "-": continue if canonicalize_name(fragment[:i]) == canonical_name: return i raise ValueError("{} does not match {}".format(fragment, canonical_name)) def _extract_version_from_fragment(fragment, canonical_name): # type: (str, str) -> Optional[str] """Parse the version string from a + filename "fragment" (stem) or egg fragment. :param fragment: The string to parse. E.g. foo-2.1 :param canonical_name: The canonicalized name of the package this belongs to. """ try: version_start = _find_name_version_sep(fragment, canonical_name) + 1 except ValueError: return None version = fragment[version_start:] if not version: return None return version site-packages/pip/_internal/commands/hash.py000064400000003307151464452710015141 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import hashlib import logging import sys from pip._internal.cli.base_command import Command from pip._internal.cli.status_codes import ERROR from pip._internal.utils.hashes import FAVORITE_HASH, STRONG_HASHES from pip._internal.utils.misc import read_chunks, write_output logger = logging.getLogger(__name__) class HashCommand(Command): """ Compute a hash of a local package archive. These can be used with --hash in a requirements file to do repeatable installs. """ usage = '%prog [options] ...' ignore_require_venv = True def __init__(self, *args, **kw): super(HashCommand, self).__init__(*args, **kw) self.cmd_opts.add_option( '-a', '--algorithm', dest='algorithm', choices=STRONG_HASHES, action='store', default=FAVORITE_HASH, help='The hash algorithm to use: one of %s' % ', '.join(STRONG_HASHES)) self.parser.insert_option_group(0, self.cmd_opts) def run(self, options, args): if not args: self.parser.print_usage(sys.stderr) return ERROR algorithm = options.algorithm for path in args: write_output('%s:\n--hash=%s:%s', path, algorithm, _hash_of_file(path, algorithm)) def _hash_of_file(path, algorithm): """Return the hash digest of a file.""" with open(path, 'rb') as archive: hash = hashlib.new(algorithm) for chunk in read_chunks(archive): hash.update(chunk) return hash.hexdigest() site-packages/pip/_internal/commands/freeze.py000064400000006610151464452710015476 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import sys from pip._internal.cache import WheelCache from pip._internal.cli import cmdoptions from pip._internal.cli.base_command import Command from pip._internal.models.format_control import FormatControl from pip._internal.operations.freeze import freeze from pip._internal.utils.compat import stdlib_pkgs DEV_PKGS = {'pip', 'setuptools', 'distribute', 'wheel'} class FreezeCommand(Command): """ Output installed packages in requirements format. packages are listed in a case-insensitive sorted order. """ usage = """ %prog [options]""" log_streams = ("ext://sys.stderr", "ext://sys.stderr") def __init__(self, *args, **kw): super(FreezeCommand, self).__init__(*args, **kw) self.cmd_opts.add_option( '-r', '--requirement', dest='requirements', action='append', default=[], metavar='file', help="Use the order in the given requirements file and its " "comments when generating output. This option can be " "used multiple times.") self.cmd_opts.add_option( '-f', '--find-links', dest='find_links', action='append', default=[], metavar='URL', help='URL for finding packages, which will be added to the ' 'output.') self.cmd_opts.add_option( '-l', '--local', dest='local', action='store_true', default=False, help='If in a virtualenv that has global access, do not output ' 'globally-installed packages.') self.cmd_opts.add_option( '--user', dest='user', action='store_true', default=False, help='Only output packages installed in user-site.') self.cmd_opts.add_option(cmdoptions.list_path()) self.cmd_opts.add_option( '--all', dest='freeze_all', action='store_true', help='Do not skip these packages in the output:' ' %s' % ', '.join(DEV_PKGS)) self.cmd_opts.add_option( '--exclude-editable', dest='exclude_editable', action='store_true', help='Exclude editable package from output.') self.parser.insert_option_group(0, self.cmd_opts) def run(self, options, args): format_control = FormatControl(set(), set()) wheel_cache = WheelCache(options.cache_dir, format_control) skip = set(stdlib_pkgs) if not options.freeze_all: skip.update(DEV_PKGS) cmdoptions.check_list_path_option(options) freeze_kwargs = dict( requirement=options.requirements, find_links=options.find_links, local_only=options.local, user_only=options.user, paths=options.path, skip_regex=options.skip_requirements_regex, isolated=options.isolated_mode, wheel_cache=wheel_cache, skip=skip, exclude_editable=options.exclude_editable, ) try: for line in freeze(**freeze_kwargs): sys.stdout.write(line + '\n') finally: wheel_cache.cleanup() site-packages/pip/_internal/commands/configuration.py000064400000016072151464452710017070 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import logging import os import subprocess from pip._internal.cli.base_command import Command from pip._internal.cli.status_codes import ERROR, SUCCESS from pip._internal.configuration import ( Configuration, get_configuration_files, kinds, ) from pip._internal.exceptions import PipError from pip._internal.utils.misc import get_prog, write_output logger = logging.getLogger(__name__) class ConfigurationCommand(Command): """Manage local and global configuration. Subcommands: list: List the active configuration (or from the file specified) edit: Edit the configuration file in an editor get: Get the value associated with name set: Set the name=value unset: Unset the value associated with name If none of --user, --global and --site are passed, a virtual environment configuration file is used if one is active and the file exists. Otherwise, all modifications happen on the to the user file by default. """ ignore_require_venv = True usage = """ %prog [] list %prog [] [--editor ] edit %prog [] get name %prog [] set name value %prog [] unset name """ def __init__(self, *args, **kwargs): super(ConfigurationCommand, self).__init__(*args, **kwargs) self.configuration = None self.cmd_opts.add_option( '--editor', dest='editor', action='store', default=None, help=( 'Editor to use to edit the file. Uses VISUAL or EDITOR ' 'environment variables if not provided.' ) ) self.cmd_opts.add_option( '--global', dest='global_file', action='store_true', default=False, help='Use the system-wide configuration file only' ) self.cmd_opts.add_option( '--user', dest='user_file', action='store_true', default=False, help='Use the user configuration file only' ) self.cmd_opts.add_option( '--site', dest='site_file', action='store_true', default=False, help='Use the current environment configuration file only' ) self.parser.insert_option_group(0, self.cmd_opts) def run(self, options, args): handlers = { "list": self.list_values, "edit": self.open_in_editor, "get": self.get_name, "set": self.set_name_value, "unset": self.unset_name } # Determine action if not args or args[0] not in handlers: logger.error("Need an action ({}) to perform.".format( ", ".join(sorted(handlers))) ) return ERROR action = args[0] # Determine which configuration files are to be loaded # Depends on whether the command is modifying. try: load_only = self._determine_file( options, need_value=(action in ["get", "set", "unset", "edit"]) ) except PipError as e: logger.error(e.args[0]) return ERROR # Load a new configuration self.configuration = Configuration( isolated=options.isolated_mode, load_only=load_only ) self.configuration.load() # Error handling happens here, not in the action-handlers. try: handlers[action](options, args[1:]) except PipError as e: logger.error(e.args[0]) return ERROR return SUCCESS def _determine_file(self, options, need_value): file_options = [key for key, value in ( (kinds.USER, options.user_file), (kinds.GLOBAL, options.global_file), (kinds.SITE, options.site_file), ) if value] if not file_options: if not need_value: return None # Default to user, unless there's a site file. elif any( os.path.exists(site_config_file) for site_config_file in get_configuration_files()[kinds.SITE] ): return kinds.SITE else: return kinds.USER elif len(file_options) == 1: return file_options[0] raise PipError( "Need exactly one file to operate upon " "(--user, --site, --global) to perform." ) def list_values(self, options, args): self._get_n_args(args, "list", n=0) for key, value in sorted(self.configuration.items()): write_output("%s=%r", key, value) def get_name(self, options, args): key = self._get_n_args(args, "get [name]", n=1) value = self.configuration.get_value(key) write_output("%s", value) def set_name_value(self, options, args): key, value = self._get_n_args(args, "set [name] [value]", n=2) self.configuration.set_value(key, value) self._save_configuration() def unset_name(self, options, args): key = self._get_n_args(args, "unset [name]", n=1) self.configuration.unset_value(key) self._save_configuration() def open_in_editor(self, options, args): editor = self._determine_editor(options) fname = self.configuration.get_file_to_edit() if fname is None: raise PipError("Could not determine appropriate file.") try: subprocess.check_call([editor, fname]) except subprocess.CalledProcessError as e: raise PipError( "Editor Subprocess exited with exit code {}" .format(e.returncode) ) def _get_n_args(self, args, example, n): """Helper to make sure the command got the right number of arguments """ if len(args) != n: msg = ( 'Got unexpected number of arguments, expected {}. ' '(example: "{} config {}")' ).format(n, get_prog(), example) raise PipError(msg) if n == 1: return args[0] else: return args def _save_configuration(self): # We successfully ran a modifying command. Need to save the # configuration. try: self.configuration.save() except Exception: logger.error( "Unable to save configuration. Please report this as a bug.", exc_info=1 ) raise PipError("Internal Error.") def _determine_editor(self, options): if options.editor is not None: return options.editor elif "VISUAL" in os.environ: return os.environ["VISUAL"] elif "EDITOR" in os.environ: return os.environ["EDITOR"] else: raise PipError("Could not determine editor to use.") site-packages/pip/_internal/commands/check.py000064400000002741151464452710015274 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import logging from pip._internal.cli.base_command import Command from pip._internal.operations.check import ( check_package_set, create_package_set_from_installed, ) from pip._internal.utils.misc import write_output logger = logging.getLogger(__name__) class CheckCommand(Command): """Verify installed packages have compatible dependencies.""" usage = """ %prog [options]""" def run(self, options, args): package_set, parsing_probs = create_package_set_from_installed() missing, conflicting = check_package_set(package_set) for project_name in missing: version = package_set[project_name].version for dependency in missing[project_name]: write_output( "%s %s requires %s, which is not installed.", project_name, version, dependency[0], ) for project_name in conflicting: version = package_set[project_name].version for dep_name, dep_version, req in conflicting[project_name]: write_output( "%s %s has requirement %s, but you have %s %s.", project_name, version, req, dep_name, dep_version, ) if missing or conflicting or parsing_probs: return 1 else: write_output("No broken requirements found.") site-packages/pip/_internal/commands/show.py000064400000015210151464452710015172 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os from email.parser import FeedParser from pip._vendor import pkg_resources from pip._vendor.packaging.utils import canonicalize_name from pip._internal.cli.base_command import Command from pip._internal.cli.status_codes import ERROR, SUCCESS from pip._internal.utils.misc import write_output logger = logging.getLogger(__name__) class ShowCommand(Command): """ Show information about one or more installed packages. The output is in RFC-compliant mail header format. """ usage = """ %prog [options] ...""" ignore_require_venv = True def __init__(self, *args, **kw): super(ShowCommand, self).__init__(*args, **kw) self.cmd_opts.add_option( '-f', '--files', dest='files', action='store_true', default=False, help='Show the full list of installed files for each package.') self.parser.insert_option_group(0, self.cmd_opts) def run(self, options, args): if not args: logger.warning('ERROR: Please provide a package name or names.') return ERROR query = args results = search_packages_info(query) if not print_results( results, list_files=options.files, verbose=options.verbose): return ERROR return SUCCESS def search_packages_info(query): """ Gather details from installed distributions. Print distribution name, version, location, and installed files. Installed files requires a pip generated 'installed-files.txt' in the distributions '.egg-info' directory. """ installed = {} for p in pkg_resources.working_set: installed[canonicalize_name(p.project_name)] = p query_names = [canonicalize_name(name) for name in query] missing = sorted( [name for name, pkg in zip(query, query_names) if pkg not in installed] ) if missing: logger.warning('Package(s) not found: %s', ', '.join(missing)) def get_requiring_packages(package_name): canonical_name = canonicalize_name(package_name) return [ pkg.project_name for pkg in pkg_resources.working_set if canonical_name in [canonicalize_name(required.name) for required in pkg.requires()] ] for dist in [installed[pkg] for pkg in query_names if pkg in installed]: package = { 'name': dist.project_name, 'version': dist.version, 'location': dist.location, 'requires': [dep.project_name for dep in dist.requires()], 'required_by': get_requiring_packages(dist.project_name) } file_list = None metadata = None if isinstance(dist, pkg_resources.DistInfoDistribution): # RECORDs should be part of .dist-info metadatas if dist.has_metadata('RECORD'): lines = dist.get_metadata_lines('RECORD') paths = [l.split(',')[0] for l in lines] paths = [os.path.join(dist.location, p) for p in paths] file_list = [os.path.relpath(p, dist.location) for p in paths] if dist.has_metadata('METADATA'): metadata = dist.get_metadata('METADATA') else: # Otherwise use pip's log for .egg-info's if dist.has_metadata('installed-files.txt'): paths = dist.get_metadata_lines('installed-files.txt') paths = [os.path.join(dist.egg_info, p) for p in paths] file_list = [os.path.relpath(p, dist.location) for p in paths] if dist.has_metadata('PKG-INFO'): metadata = dist.get_metadata('PKG-INFO') if dist.has_metadata('entry_points.txt'): entry_points = dist.get_metadata_lines('entry_points.txt') package['entry_points'] = entry_points if dist.has_metadata('INSTALLER'): for line in dist.get_metadata_lines('INSTALLER'): if line.strip(): package['installer'] = line.strip() break # @todo: Should pkg_resources.Distribution have a # `get_pkg_info` method? feed_parser = FeedParser() feed_parser.feed(metadata) pkg_info_dict = feed_parser.close() for key in ('metadata-version', 'summary', 'home-page', 'author', 'author-email', 'license'): package[key] = pkg_info_dict.get(key) # It looks like FeedParser cannot deal with repeated headers classifiers = [] for line in metadata.splitlines(): if line.startswith('Classifier: '): classifiers.append(line[len('Classifier: '):]) package['classifiers'] = classifiers if file_list: package['files'] = sorted(file_list) yield package def print_results(distributions, list_files=False, verbose=False): """ Print the informations from installed distributions found. """ results_printed = False for i, dist in enumerate(distributions): results_printed = True if i > 0: write_output("---") write_output("Name: %s", dist.get('name', '')) write_output("Version: %s", dist.get('version', '')) write_output("Summary: %s", dist.get('summary', '')) write_output("Home-page: %s", dist.get('home-page', '')) write_output("Author: %s", dist.get('author', '')) write_output("Author-email: %s", dist.get('author-email', '')) write_output("License: %s", dist.get('license', '')) write_output("Location: %s", dist.get('location', '')) write_output("Requires: %s", ', '.join(dist.get('requires', []))) write_output("Required-by: %s", ', '.join(dist.get('required_by', []))) if verbose: write_output("Metadata-Version: %s", dist.get('metadata-version', '')) write_output("Installer: %s", dist.get('installer', '')) write_output("Classifiers:") for classifier in dist.get('classifiers', []): write_output(" %s", classifier) write_output("Entry-points:") for entry in dist.get('entry_points', []): write_output(" %s", entry.strip()) if list_files: write_output("Files:") for line in dist.get('files', []): write_output(" %s", line.strip()) if "files" not in dist: write_output("Cannot locate installed-files.txt") return results_printed site-packages/pip/_internal/commands/list.py000064400000024464151464452710015200 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import json import logging from pip._vendor import six from pip._vendor.six.moves import zip_longest from pip._internal.cli import cmdoptions from pip._internal.cli.req_command import IndexGroupCommand from pip._internal.exceptions import CommandError from pip._internal.index import PackageFinder from pip._internal.models.selection_prefs import SelectionPreferences from pip._internal.self_outdated_check import make_link_collector from pip._internal.utils.misc import ( dist_is_editable, get_installed_distributions, write_output, ) from pip._internal.utils.packaging import get_installer logger = logging.getLogger(__name__) class ListCommand(IndexGroupCommand): """ List installed packages, including editables. Packages are listed in a case-insensitive sorted order. """ usage = """ %prog [options]""" def __init__(self, *args, **kw): super(ListCommand, self).__init__(*args, **kw) cmd_opts = self.cmd_opts cmd_opts.add_option( '-o', '--outdated', action='store_true', default=False, help='List outdated packages') cmd_opts.add_option( '-u', '--uptodate', action='store_true', default=False, help='List uptodate packages') cmd_opts.add_option( '-e', '--editable', action='store_true', default=False, help='List editable projects.') cmd_opts.add_option( '-l', '--local', action='store_true', default=False, help=('If in a virtualenv that has global access, do not list ' 'globally-installed packages.'), ) self.cmd_opts.add_option( '--user', dest='user', action='store_true', default=False, help='Only output packages installed in user-site.') cmd_opts.add_option(cmdoptions.list_path()) cmd_opts.add_option( '--pre', action='store_true', default=False, help=("Include pre-release and development versions. By default, " "pip only finds stable versions."), ) cmd_opts.add_option( '--format', action='store', dest='list_format', default="columns", choices=('columns', 'freeze', 'json'), help="Select the output format among: columns (default), freeze, " "or json", ) cmd_opts.add_option( '--not-required', action='store_true', dest='not_required', help="List packages that are not dependencies of " "installed packages.", ) cmd_opts.add_option( '--exclude-editable', action='store_false', dest='include_editable', help='Exclude editable package from output.', ) cmd_opts.add_option( '--include-editable', action='store_true', dest='include_editable', help='Include editable package from output.', default=True, ) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, self.parser ) self.parser.insert_option_group(0, index_opts) self.parser.insert_option_group(0, cmd_opts) def _build_package_finder(self, options, session): """ Create a package finder appropriate to this list command. """ link_collector = make_link_collector(session, options=options) # Pass allow_yanked=False to ignore yanked versions. selection_prefs = SelectionPreferences( allow_yanked=False, allow_all_prereleases=options.pre, ) return PackageFinder.create( link_collector=link_collector, selection_prefs=selection_prefs, ) def run(self, options, args): if options.outdated and options.uptodate: raise CommandError( "Options --outdated and --uptodate cannot be combined.") cmdoptions.check_list_path_option(options) packages = get_installed_distributions( local_only=options.local, user_only=options.user, editables_only=options.editable, include_editables=options.include_editable, paths=options.path, ) # get_not_required must be called firstly in order to find and # filter out all dependencies correctly. Otherwise a package # can't be identified as requirement because some parent packages # could be filtered out before. if options.not_required: packages = self.get_not_required(packages, options) if options.outdated: packages = self.get_outdated(packages, options) elif options.uptodate: packages = self.get_uptodate(packages, options) self.output_package_listing(packages, options) def get_outdated(self, packages, options): return [ dist for dist in self.iter_packages_latest_infos(packages, options) if dist.latest_version > dist.parsed_version ] def get_uptodate(self, packages, options): return [ dist for dist in self.iter_packages_latest_infos(packages, options) if dist.latest_version == dist.parsed_version ] def get_not_required(self, packages, options): dep_keys = set() for dist in packages: dep_keys.update(requirement.key for requirement in dist.requires()) return {pkg for pkg in packages if pkg.key not in dep_keys} def iter_packages_latest_infos(self, packages, options): with self._build_session(options) as session: finder = self._build_package_finder(options, session) for dist in packages: typ = 'unknown' all_candidates = finder.find_all_candidates(dist.key) if not options.pre: # Remove prereleases all_candidates = [candidate for candidate in all_candidates if not candidate.version.is_prerelease] evaluator = finder.make_candidate_evaluator( project_name=dist.project_name, ) best_candidate = evaluator.sort_best_candidate(all_candidates) if best_candidate is None: continue remote_version = best_candidate.version if best_candidate.link.is_wheel: typ = 'wheel' else: typ = 'sdist' # This is dirty but makes the rest of the code much cleaner dist.latest_version = remote_version dist.latest_filetype = typ yield dist def output_package_listing(self, packages, options): packages = sorted( packages, key=lambda dist: dist.project_name.lower(), ) if options.list_format == 'columns' and packages: data, header = format_for_columns(packages, options) self.output_package_listing_columns(data, header) elif options.list_format == 'freeze': for dist in packages: if options.verbose >= 1: write_output("%s==%s (%s)", dist.project_name, dist.version, dist.location) else: write_output("%s==%s", dist.project_name, dist.version) elif options.list_format == 'json': write_output(format_for_json(packages, options)) def output_package_listing_columns(self, data, header): # insert the header first: we need to know the size of column names if len(data) > 0: data.insert(0, header) pkg_strings, sizes = tabulate(data) # Create and add a separator. if len(data) > 0: pkg_strings.insert(1, " ".join(map(lambda x: '-' * x, sizes))) for val in pkg_strings: write_output(val) def tabulate(vals): # From pfmoore on GitHub: # https://github.com/pypa/pip/issues/3651#issuecomment-216932564 assert len(vals) > 0 sizes = [0] * max(len(x) for x in vals) for row in vals: sizes = [max(s, len(str(c))) for s, c in zip_longest(sizes, row)] result = [] for row in vals: display = " ".join([str(c).ljust(s) if c is not None else '' for s, c in zip_longest(sizes, row)]) result.append(display) return result, sizes def format_for_columns(pkgs, options): """ Convert the package data into something usable by output_package_listing_columns. """ running_outdated = options.outdated # Adjust the header for the `pip list --outdated` case. if running_outdated: header = ["Package", "Version", "Latest", "Type"] else: header = ["Package", "Version"] data = [] if options.verbose >= 1 or any(dist_is_editable(x) for x in pkgs): header.append("Location") if options.verbose >= 1: header.append("Installer") for proj in pkgs: # if we're working on the 'outdated' list, separate out the # latest_version and type row = [proj.project_name, proj.version] if running_outdated: row.append(proj.latest_version) row.append(proj.latest_filetype) if options.verbose >= 1 or dist_is_editable(proj): row.append(proj.location) if options.verbose >= 1: row.append(get_installer(proj)) data.append(row) return data, header def format_for_json(packages, options): data = [] for dist in packages: info = { 'name': dist.project_name, 'version': six.text_type(dist.version), } if options.verbose >= 1: info['location'] = dist.location info['installer'] = get_installer(dist) if options.outdated: info['latest_version'] = six.text_type(dist.latest_version) info['latest_filetype'] = dist.latest_filetype data.append(info) return json.dumps(data) site-packages/pip/_internal/commands/uninstall.py000064400000005647151464452710016240 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import from pip._vendor.packaging.utils import canonicalize_name from pip._internal.cli.base_command import Command from pip._internal.cli.req_command import SessionCommandMixin from pip._internal.exceptions import InstallationError from pip._internal.req import parse_requirements from pip._internal.req.constructors import install_req_from_line from pip._internal.utils.misc import protect_pip_from_modification_on_windows class UninstallCommand(Command, SessionCommandMixin): """ Uninstall packages. pip is able to uninstall most installed packages. Known exceptions are: - Pure distutils packages installed with ``python setup.py install``, which leave behind no metadata to determine what files were installed. - Script wrappers installed by ``python setup.py develop``. """ usage = """ %prog [options] ... %prog [options] -r ...""" def __init__(self, *args, **kw): super(UninstallCommand, self).__init__(*args, **kw) self.cmd_opts.add_option( '-r', '--requirement', dest='requirements', action='append', default=[], metavar='file', help='Uninstall all the packages listed in the given requirements ' 'file. This option can be used multiple times.', ) self.cmd_opts.add_option( '-y', '--yes', dest='yes', action='store_true', help="Don't ask for confirmation of uninstall deletions.") self.parser.insert_option_group(0, self.cmd_opts) def run(self, options, args): session = self.get_default_session(options) reqs_to_uninstall = {} for name in args: req = install_req_from_line( name, isolated=options.isolated_mode, ) if req.name: reqs_to_uninstall[canonicalize_name(req.name)] = req for filename in options.requirements: for req in parse_requirements( filename, options=options, session=session): if req.name: reqs_to_uninstall[canonicalize_name(req.name)] = req if not reqs_to_uninstall: raise InstallationError( 'You must give at least one requirement to %(name)s (see ' '"pip help %(name)s")' % dict(name=self.name) ) protect_pip_from_modification_on_windows( modifying_pip="pip" in reqs_to_uninstall ) for req in reqs_to_uninstall.values(): uninstall_pathset = req.uninstall( auto_confirm=options.yes, verbose=self.verbosity > 0, ) if uninstall_pathset: uninstall_pathset.commit() site-packages/pip/_internal/commands/install.py000064400000057757151464452710015706 0ustar00# The following comment should be removed at some point in the future. # It's included for now because without it InstallCommand.run() has a # couple errors where we have to know req.name is str rather than # Optional[str] for the InstallRequirement req. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import errno import logging import operator import os import shutil import sys from os import path from optparse import SUPPRESS_HELP from pip._vendor import pkg_resources from pip._vendor.packaging.utils import canonicalize_name from pip._internal.cache import WheelCache from pip._internal.cli import cmdoptions from pip._internal.cli.cmdoptions import make_target_python from pip._internal.cli.req_command import RequirementCommand from pip._internal.cli.status_codes import ERROR, SUCCESS from pip._internal.exceptions import ( CommandError, InstallationError, PreviousBuildDirError, ) from pip._internal.locations import distutils_scheme from pip._internal.operations.check import check_install_conflicts from pip._internal.req import RequirementSet, install_given_reqs from pip._internal.req.req_tracker import RequirementTracker from pip._internal.utils.filesystem import check_path_owner from pip._internal.utils.misc import ( ensure_dir, get_installed_version, protect_pip_from_modification_on_windows, write_output, ) from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.virtualenv import virtualenv_no_global from pip._internal.wheel import WheelBuilder if MYPY_CHECK_RUNNING: from optparse import Values from typing import Any, List, Optional from pip._internal.models.format_control import FormatControl from pip._internal.req.req_install import InstallRequirement from pip._internal.wheel import BinaryAllowedPredicate logger = logging.getLogger(__name__) def is_wheel_installed(): """ Return whether the wheel package is installed. """ try: import wheel # noqa: F401 except ImportError: return False return True def build_wheels( builder, # type: WheelBuilder pep517_requirements, # type: List[InstallRequirement] legacy_requirements, # type: List[InstallRequirement] ): # type: (...) -> List[InstallRequirement] """ Build wheels for requirements, depending on whether wheel is installed. """ # We don't build wheels for legacy requirements if wheel is not installed. should_build_legacy = is_wheel_installed() # Always build PEP 517 requirements build_failures = builder.build( pep517_requirements, should_unpack=True, ) if should_build_legacy: # We don't care about failures building legacy # requirements, as we'll fall through to a direct # install for those. builder.build( legacy_requirements, should_unpack=True, ) return build_failures def get_check_binary_allowed(format_control): # type: (FormatControl) -> BinaryAllowedPredicate def check_binary_allowed(req): # type: (InstallRequirement) -> bool if req.use_pep517: return True canonical_name = canonicalize_name(req.name) allowed_formats = format_control.get_allowed_formats(canonical_name) return "binary" in allowed_formats return check_binary_allowed class InstallCommand(RequirementCommand): """ Install packages from: - PyPI (and other indexes) using requirement specifiers. - VCS project urls. - Local project directories. - Local or remote source archives. pip also supports installing from "requirements files", which provide an easy way to specify a whole environment to be installed. """ usage = """ %prog [options] [package-index-options] ... %prog [options] -r [package-index-options] ... %prog [options] [-e] ... %prog [options] [-e] ... %prog [options] ...""" def __init__(self, *args, **kw): super(InstallCommand, self).__init__(*args, **kw) cmd_opts = self.cmd_opts cmd_opts.add_option(cmdoptions.requirements()) cmd_opts.add_option(cmdoptions.constraints()) cmd_opts.add_option(cmdoptions.no_deps()) cmd_opts.add_option(cmdoptions.pre()) cmd_opts.add_option(cmdoptions.editable()) cmd_opts.add_option( '-t', '--target', dest='target_dir', metavar='dir', default=None, help='Install packages into . ' 'By default this will not replace existing files/folders in ' '. Use --upgrade to replace existing packages in ' 'with new versions.' ) cmdoptions.add_target_python_options(cmd_opts) cmd_opts.add_option( '--user', dest='use_user_site', action='store_true', help="Install to the Python user install directory for your " "platform. Typically ~/.local/, or %APPDATA%\\Python on " "Windows. (See the Python documentation for site.USER_BASE " "for full details.)") cmd_opts.add_option( '--no-user', dest='use_user_site', action='store_false', help=SUPPRESS_HELP) cmd_opts.add_option( '--root', dest='root_path', metavar='dir', default=None, help="Install everything relative to this alternate root " "directory.") cmd_opts.add_option( '--prefix', dest='prefix_path', metavar='dir', default=None, help="Installation prefix where lib, bin and other top-level " "folders are placed") cmd_opts.add_option(cmdoptions.build_dir()) cmd_opts.add_option(cmdoptions.src()) cmd_opts.add_option( '-U', '--upgrade', dest='upgrade', action='store_true', help='Upgrade all specified packages to the newest available ' 'version. The handling of dependencies depends on the ' 'upgrade-strategy used.' ) cmd_opts.add_option( '--upgrade-strategy', dest='upgrade_strategy', default='only-if-needed', choices=['only-if-needed', 'eager'], help='Determines how dependency upgrading should be handled ' '[default: %default]. ' '"eager" - dependencies are upgraded regardless of ' 'whether the currently installed version satisfies the ' 'requirements of the upgraded package(s). ' '"only-if-needed" - are upgraded only when they do not ' 'satisfy the requirements of the upgraded package(s).' ) cmd_opts.add_option( '--force-reinstall', dest='force_reinstall', action='store_true', help='Reinstall all packages even if they are already ' 'up-to-date.') cmd_opts.add_option( '-I', '--ignore-installed', dest='ignore_installed', action='store_true', help='Ignore the installed packages, overwriting them. ' 'This can break your system if the existing package ' 'is of a different version or was installed ' 'with a different package manager!' ) cmd_opts.add_option(cmdoptions.ignore_requires_python()) cmd_opts.add_option(cmdoptions.no_build_isolation()) cmd_opts.add_option(cmdoptions.use_pep517()) cmd_opts.add_option(cmdoptions.no_use_pep517()) cmd_opts.add_option(cmdoptions.install_options()) cmd_opts.add_option(cmdoptions.global_options()) cmd_opts.add_option( "--compile", action="store_true", dest="compile", default=True, help="Compile Python source files to bytecode", ) cmd_opts.add_option( "--no-compile", action="store_false", dest="compile", help="Do not compile Python source files to bytecode", ) cmd_opts.add_option( "--no-warn-script-location", action="store_false", dest="warn_script_location", default=True, help="Do not warn when installing scripts outside PATH", ) cmd_opts.add_option( "--no-warn-conflicts", action="store_false", dest="warn_about_conflicts", default=True, help="Do not warn about broken dependencies", ) cmd_opts.add_option(cmdoptions.no_binary()) cmd_opts.add_option(cmdoptions.only_binary()) cmd_opts.add_option(cmdoptions.prefer_binary()) cmd_opts.add_option(cmdoptions.no_clean()) cmd_opts.add_option(cmdoptions.require_hashes()) cmd_opts.add_option(cmdoptions.progress_bar()) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, self.parser, ) self.parser.insert_option_group(0, index_opts) self.parser.insert_option_group(0, cmd_opts) def run(self, options, args): # type: (Values, List[Any]) -> int cmdoptions.check_install_build_global(options) def is_venv(): return (hasattr(sys, 'real_prefix') or (hasattr(sys, 'base_prefix') and sys.base_prefix != sys.prefix)) # Check whether we have root privileges and aren't in venv/virtualenv if os.getuid() == 0 and not is_venv(): command = path.basename(sys.argv[0]) if command == "__main__.py": command = path.basename(sys.executable) + " -m pip" logger.warning( "Running pip install with root privileges is " "generally not a good idea. Try `%s install --user` instead." % command ) upgrade_strategy = "to-satisfy-only" if options.upgrade: upgrade_strategy = options.upgrade_strategy if options.build_dir: options.build_dir = os.path.abspath(options.build_dir) cmdoptions.check_dist_restriction(options, check_target=True) options.src_dir = os.path.abspath(options.src_dir) install_options = options.install_options or [] if options.use_user_site: if options.prefix_path: raise CommandError( "Can not combine '--user' and '--prefix' as they imply " "different installation locations" ) if virtualenv_no_global(): raise InstallationError( "Can not perform a '--user' install. User site-packages " "are not visible in this virtualenv." ) install_options.append('--user') install_options.append('--prefix=') target_temp_dir = None # type: Optional[TempDirectory] target_temp_dir_path = None # type: Optional[str] if options.target_dir: options.ignore_installed = True options.target_dir = os.path.abspath(options.target_dir) if (os.path.exists(options.target_dir) and not os.path.isdir(options.target_dir)): raise CommandError( "Target path exists but is not a directory, will not " "continue." ) # Create a target directory for using with the target option target_temp_dir = TempDirectory(kind="target") target_temp_dir_path = target_temp_dir.path install_options.append('--home=' + target_temp_dir_path) global_options = options.global_options or [] session = self.get_default_session(options) target_python = make_target_python(options) finder = self._build_package_finder( options=options, session=session, target_python=target_python, ignore_requires_python=options.ignore_requires_python, ) build_delete = (not (options.no_clean or options.build_dir)) wheel_cache = WheelCache(options.cache_dir, options.format_control) if options.cache_dir and not check_path_owner(options.cache_dir): logger.warning( "The directory '%s' or its parent directory is not owned " "by the current user and caching wheels has been " "disabled. check the permissions and owner of that " "directory. If executing pip with sudo, you may want " "sudo's -H flag.", options.cache_dir, ) options.cache_dir = None with RequirementTracker() as req_tracker, TempDirectory( options.build_dir, delete=build_delete, kind="install" ) as directory: requirement_set = RequirementSet( require_hashes=options.require_hashes, check_supported_wheels=not options.target_dir, ) try: self.populate_requirement_set( requirement_set, args, options, finder, session, wheel_cache ) preparer = self.make_requirement_preparer( temp_build_dir=directory, options=options, req_tracker=req_tracker, ) resolver = self.make_resolver( preparer=preparer, finder=finder, session=session, options=options, wheel_cache=wheel_cache, use_user_site=options.use_user_site, ignore_installed=options.ignore_installed, ignore_requires_python=options.ignore_requires_python, force_reinstall=options.force_reinstall, upgrade_strategy=upgrade_strategy, use_pep517=options.use_pep517, ) resolver.resolve(requirement_set) try: pip_req = requirement_set.get_requirement("pip") except KeyError: modifying_pip = None else: # If we're not replacing an already installed pip, # we're not modifying it. modifying_pip = pip_req.satisfied_by is None protect_pip_from_modification_on_windows( modifying_pip=modifying_pip ) check_binary_allowed = get_check_binary_allowed( finder.format_control ) # Consider legacy and PEP517-using requirements separately legacy_requirements = [] pep517_requirements = [] for req in requirement_set.requirements.values(): if req.use_pep517: pep517_requirements.append(req) else: legacy_requirements.append(req) wheel_builder = WheelBuilder( preparer, wheel_cache, build_options=[], global_options=[], check_binary_allowed=check_binary_allowed, ) build_failures = build_wheels( builder=wheel_builder, pep517_requirements=pep517_requirements, legacy_requirements=legacy_requirements, ) # If we're using PEP 517, we cannot do a direct install # so we fail here. if build_failures: raise InstallationError( "Could not build wheels for {} which use" " PEP 517 and cannot be installed directly".format( ", ".join(r.name for r in build_failures))) to_install = resolver.get_installation_order( requirement_set ) # Consistency Checking of the package set we're installing. should_warn_about_conflicts = ( not options.ignore_dependencies and options.warn_about_conflicts ) if should_warn_about_conflicts: self._warn_about_conflicts(to_install) # Don't warn about script install locations if # --target has been specified warn_script_location = options.warn_script_location if options.target_dir: warn_script_location = False installed = install_given_reqs( to_install, install_options, global_options, root=options.root_path, home=target_temp_dir_path, prefix=options.prefix_path, pycompile=options.compile, warn_script_location=warn_script_location, use_user_site=options.use_user_site, ) lib_locations = get_lib_location_guesses( user=options.use_user_site, home=target_temp_dir_path, root=options.root_path, prefix=options.prefix_path, isolated=options.isolated_mode, ) working_set = pkg_resources.WorkingSet(lib_locations) reqs = sorted(installed, key=operator.attrgetter('name')) items = [] for req in reqs: item = req.name try: installed_version = get_installed_version( req.name, working_set=working_set ) if installed_version: item += '-' + installed_version except Exception: pass items.append(item) installed_desc = ' '.join(items) if installed_desc: write_output( 'Successfully installed %s', installed_desc, ) except EnvironmentError as error: show_traceback = (self.verbosity >= 1) message = create_env_error_message( error, show_traceback, options.use_user_site, ) logger.error(message, exc_info=show_traceback) return ERROR except PreviousBuildDirError: options.no_clean = True raise finally: # Clean up if not options.no_clean: requirement_set.cleanup_files() wheel_cache.cleanup() if options.target_dir: self._handle_target_dir( options.target_dir, target_temp_dir, options.upgrade ) return SUCCESS def _handle_target_dir(self, target_dir, target_temp_dir, upgrade): ensure_dir(target_dir) # Checking both purelib and platlib directories for installed # packages to be moved to target directory lib_dir_list = [] with target_temp_dir: # Checking both purelib and platlib directories for installed # packages to be moved to target directory scheme = distutils_scheme('', home=target_temp_dir.path) purelib_dir = scheme['purelib'] platlib_dir = scheme['platlib'] data_dir = scheme['data'] if os.path.exists(purelib_dir): lib_dir_list.append(purelib_dir) if os.path.exists(platlib_dir) and platlib_dir != purelib_dir: lib_dir_list.append(platlib_dir) if os.path.exists(data_dir): lib_dir_list.append(data_dir) for lib_dir in lib_dir_list: for item in os.listdir(lib_dir): if lib_dir == data_dir: ddir = os.path.join(data_dir, item) if any(s.startswith(ddir) for s in lib_dir_list[:-1]): continue target_item_dir = os.path.join(target_dir, item) if os.path.exists(target_item_dir): if not upgrade: logger.warning( 'Target directory %s already exists. Specify ' '--upgrade to force replacement.', target_item_dir ) continue if os.path.islink(target_item_dir): logger.warning( 'Target directory %s already exists and is ' 'a link. Pip will not automatically replace ' 'links, please remove if replacement is ' 'desired.', target_item_dir ) continue if os.path.isdir(target_item_dir): shutil.rmtree(target_item_dir) else: os.remove(target_item_dir) shutil.move( os.path.join(lib_dir, item), target_item_dir ) def _warn_about_conflicts(self, to_install): try: package_set, _dep_info = check_install_conflicts(to_install) except Exception: logger.error("Error checking for conflicts.", exc_info=True) return missing, conflicting = _dep_info # NOTE: There is some duplication here from pip check for project_name in missing: version = package_set[project_name][0] for dependency in missing[project_name]: logger.critical( "%s %s requires %s, which is not installed.", project_name, version, dependency[1], ) for project_name in conflicting: version = package_set[project_name][0] for dep_name, dep_version, req in conflicting[project_name]: logger.critical( "%s %s has requirement %s, but you'll have %s %s which is " "incompatible.", project_name, version, req, dep_name, dep_version, ) def get_lib_location_guesses(*args, **kwargs): scheme = distutils_scheme('', *args, **kwargs) return [scheme['purelib'], scheme['platlib']] def create_env_error_message(error, show_traceback, using_user_site): """Format an error message for an EnvironmentError It may occur anytime during the execution of the install command. """ parts = [] # Mention the error if we are not going to show a traceback parts.append("Could not install packages due to an EnvironmentError") if not show_traceback: parts.append(": ") parts.append(str(error)) else: parts.append(".") # Spilt the error indication from a helper message (if any) parts[-1] += "\n" # Suggest useful actions to the user: # (1) using user site-packages or (2) verifying the permissions if error.errno == errno.EACCES: user_option_part = "Consider using the `--user` option" permissions_part = "Check the permissions" if not using_user_site: parts.extend([ user_option_part, " or ", permissions_part.lower(), ]) else: parts.append(permissions_part) parts.append(".\n") return "".join(parts).strip() + "\n" site-packages/pip/_internal/commands/__pycache__/completion.cpython-38.pyc000064400000005705151464452710022661 0ustar00U .e @sXddlmZddlZddlZddlmZddlmZdZdddd Z Gd d d eZ dS) )absolute_importN)Command)get_progzJ # pip %(shell)s completion start%(script)s# pip %(shell)s completion end a _pip_completion() { COMPREPLY=( $( COMP_WORDS="${COMP_WORDS[*]}" \ COMP_CWORD=$COMP_CWORD \ PIP_AUTO_COMPLETE=1 $1 2>/dev/null ) ) } complete -o default -F _pip_completion %(prog)s aM function _pip_completion { local words cword read -Ac words read -cn cword reply=( $( COMP_WORDS="$words[*]" \ COMP_CWORD=$(( cword-1 )) \ PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null )) } compctl -K _pip_completion %(prog)s aw function __fish_complete_pip set -lx COMP_WORDS (commandline -o) "" set -lx COMP_CWORD ( \ math (contains -i -- (commandline -t) $COMP_WORDS)-1 \ ) set -lx PIP_AUTO_COMPLETE 1 string split \ -- (eval $COMP_WORDS[1]) end complete -fa "(__fish_complete_pip)" -c %(prog)s )bashzshfishcs,eZdZdZdZfddZddZZS)CompletionCommandz3A helper command to be used for command completion.Tcsltt|j|||j}|jddddddd|jdd dd dd d|jd d ddddd|jd|dS)Nz--bashz-b store_constrshellzEmit completion code for bash)actionconstdesthelpz--zshz-zrzEmit completion code for zshz--fishz-frzEmit completion code for fishr)superr__init__cmd_optsZ add_optionparserZinsert_option_group)selfargskwr __class__E/usr/lib/python3.8/site-packages/pip/_internal/commands/completion.pyr8s6zCompletionCommand.__init__cCsrt}ddt|D}|j|krXtt|jddti}tt ||jdnt j dd |dS) z-Prints the completion code of the given shellcSsg|] }d|qS)z--r).0r rrr Usz)CompletionCommand.run..prog)scriptr zERROR: You must pass %s z or N)COMPLETION_SCRIPTSkeyssortedr textwrapdedentgetrprintBASE_COMPLETIONsysstderrwritejoin)rZoptionsrZshellsZ shell_optionsrrrrrunRs   zCompletionCommand.run)__name__ __module__ __qualname____doc__Zignore_require_venvrr+ __classcell__rrrrr3s r) Z __future__rr'r"Zpip._internal.cli.base_commandrZpip._internal.utils.miscrr&rrrrrrs     #site-packages/pip/_internal/commands/__pycache__/download.cpython-38.opt-1.pyc000064400000010314151464452710023246 0ustar00U .e@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZdd lmZmZmZdd lmZeeZGd d d e ZdS) )absolute_importN) cmdoptions)make_target_python)RequirementCommand)RequirementSet)RequirementTracker)check_path_owner) ensure_dirnormalize_path write_output) TempDirectorycs,eZdZdZdZfddZddZZS)DownloadCommandaL Download packages from: - PyPI (and other indexes) using requirement specifiers. - VCS project urls. - Local project directories. - Local or remote source archives. pip also supports downloading from "requirements files", which provide an easy way to specify a whole environment to be downloaded. a %prog [options] [package-index-options] ... %prog [options] -r [package-index-options] ... %prog [options] ... %prog [options] ... %prog [options] ...c sNtt|j|||j}|t|t|t|t |t |t |t |t |t|t|t|t|t|t|t|t|jddddddtjddt|ttj|j}|jd ||jd |dS) Nz-dz--destz--destination-dirz--destination-directory download_dirdirzDownload packages into .)destmetavardefaulthelpr)superr __init__cmd_optsZ add_optionrZ constraintsZ requirements build_dirZno_depsZglobal_optionsZ no_binaryZ only_binaryZ prefer_binarysrcZpreno_cleanrequire_hashesZ progress_barZno_build_isolationZ use_pep517Z no_use_pep517oscurdirZadd_target_python_optionsZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrZ index_opts __class__C/usr/lib/python3.8/site-packages/pip/_internal/commands/download.pyr)sF zDownloadCommand.__init__c CsLd|_g|_t|tj|j|_t|j |_ t |j | |}t |}|j |||d}|jph|j }|jrt|jstd|jd|_t}t|j|dd}t|jd} || ||||d|j||||j d} |j| ||||jd} | | d d d | jD} | r$td | |js4| W5QRXW5QRX| S) NT)optionssession target_pythonzThe directory '%s' or its parent directory is not owned by the current user and caching wheels has been disabled. check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag.Zdownload)deleteZkind)r)Ztemp_build_dirr% req_trackerr)preparerfinderr&r%Zpy_version_info cSsg|] }|jqSr#)name).0Zreqr#r#r$ sz'DownloadCommand.run..zSuccessfully downloaded %s)!Zignore_installedZ editablesrZcheck_dist_restrictionrpathabspathZsrc_dirr rr Zget_default_sessionrZ_build_package_finderrr cache_dirrloggerZwarningrr rrZpopulate_requirement_setZmake_requirement_preparerZ make_resolverZpython_versionZresolvejoinZsuccessfully_downloadedr Z cleanup_files) rr%rr&r'r+Z build_deleter)Z directoryZrequirement_setr*ZresolverZ downloadedr#r#r$runQsv         zDownloadCommand.run)__name__ __module__ __qualname____doc__Zusagerr5 __classcell__r#r#r!r$r s  (r )Z __future__rZloggingrZpip._internal.clirZpip._internal.cli.cmdoptionsrZpip._internal.cli.req_commandrZpip._internal.reqrZpip._internal.req.req_trackerrZpip._internal.utils.filesystemrZpip._internal.utils.miscr r r Zpip._internal.utils.temp_dirr Z getLoggerr6r3r r#r#r#r$s         site-packages/pip/_internal/commands/__pycache__/download.cpython-38.pyc000064400000010314151464452710022307 0ustar00U .e@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZdd lmZmZmZdd lmZeeZGd d d e ZdS) )absolute_importN) cmdoptions)make_target_python)RequirementCommand)RequirementSet)RequirementTracker)check_path_owner) ensure_dirnormalize_path write_output) TempDirectorycs,eZdZdZdZfddZddZZS)DownloadCommandaL Download packages from: - PyPI (and other indexes) using requirement specifiers. - VCS project urls. - Local project directories. - Local or remote source archives. pip also supports downloading from "requirements files", which provide an easy way to specify a whole environment to be downloaded. a %prog [options] [package-index-options] ... %prog [options] -r [package-index-options] ... %prog [options] ... %prog [options] ... %prog [options] ...c sNtt|j|||j}|t|t|t|t |t |t |t |t |t|t|t|t|t|t|t|t|jddddddtjddt|ttj|j}|jd ||jd |dS) Nz-dz--destz--destination-dirz--destination-directory download_dirdirzDownload packages into .)destmetavardefaulthelpr)superr __init__cmd_optsZ add_optionrZ constraintsZ requirements build_dirZno_depsZglobal_optionsZ no_binaryZ only_binaryZ prefer_binarysrcZpreno_cleanrequire_hashesZ progress_barZno_build_isolationZ use_pep517Z no_use_pep517oscurdirZadd_target_python_optionsZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrZ index_opts __class__C/usr/lib/python3.8/site-packages/pip/_internal/commands/download.pyr)sF zDownloadCommand.__init__c CsLd|_g|_t|tj|j|_t|j |_ t |j | |}t |}|j |||d}|jph|j }|jrt|jstd|jd|_t}t|j|dd}t|jd} || ||||d|j||||j d} |j| ||||jd} | | d d d | jD} | r$td | |js4| W5QRXW5QRX| S) NT)optionssession target_pythonzThe directory '%s' or its parent directory is not owned by the current user and caching wheels has been disabled. check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag.Zdownload)deleteZkind)r)Ztemp_build_dirr% req_trackerr)preparerfinderr&r%Zpy_version_info cSsg|] }|jqSr#)name).0Zreqr#r#r$ sz'DownloadCommand.run..zSuccessfully downloaded %s)!Zignore_installedZ editablesrZcheck_dist_restrictionrpathabspathZsrc_dirr rr Zget_default_sessionrZ_build_package_finderrr cache_dirrloggerZwarningrr rrZpopulate_requirement_setZmake_requirement_preparerZ make_resolverZpython_versionZresolvejoinZsuccessfully_downloadedr Z cleanup_files) rr%rr&r'r+Z build_deleter)Z directoryZrequirement_setr*ZresolverZ downloadedr#r#r$runQsv         zDownloadCommand.run)__name__ __module__ __qualname____doc__Zusagerr5 __classcell__r#r#r!r$r s  (r )Z __future__rZloggingrZpip._internal.clirZpip._internal.cli.cmdoptionsrZpip._internal.cli.req_commandrZpip._internal.reqrZpip._internal.req.req_trackerrZpip._internal.utils.filesystemrZpip._internal.utils.miscr r r Zpip._internal.utils.temp_dirr Z getLoggerr6r3r r#r#r#r$s         site-packages/pip/_internal/commands/__pycache__/show.cpython-38.opt-1.pyc000064400000014265151464452710022430 0ustar00U .e@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z mZddlmZeeZGd d d e Zd d ZdddZdS))absolute_importN) FeedParser) pkg_resourcescanonicalize_name)Command)ERRORSUCCESS) write_outputcs0eZdZdZdZdZfddZddZZS) ShowCommandzx Show information about one or more installed packages. The output is in RFC-compliant mail header format. z$ %prog [options] ...Tcs>tt|j|||jjddddddd|jd|jdS) Nz-fz--filesfiles store_trueFz7Show the full list of installed files for each package.)destactiondefaulthelpr)superr __init__Zcmd_optsZ add_optionparserZinsert_option_group)selfargskw __class__?/usr/lib/python3.8/site-packages/pip/_internal/commands/show.pyrszShowCommand.__init__cCs8|stdtS|}t|}t||j|jds4tStS)Nz.ERROR: Please provide a package name or names.) list_filesverbose)loggerwarningrsearch_packages_info print_resultsr rr )rZoptionsrqueryresultsrrrrun*s zShowCommand.run) __name__ __module__ __qualname____doc__ZusageZignore_require_venvrr$ __classcell__rrrrr s  r c#sTitjD]}|t|j<q dd|D}tfddt||D}|r^tdd|dd}fdd|DD]ԉjj j d d D|jd }d }d }t tj rd rd }d d|D} fdd| D} fdd| D}drnd}nPdrXd} fdd| D} fdd| D}drnd}drd} | |d<drƈdD]"} | r| |d<qƐqt} | || } dD]}| |||<qg}|D](} | dr|| tdd q||d<|rHt||d<|Vqxd S)z Gather details from installed distributions. Print distribution name, version, location, and installed files. Installed files requires a pip generated 'installed-files.txt' in the distributions '.egg-info' directory. cSsg|] }t|qSrr).0namerrr Bsz(search_packages_info..csg|]\}}|kr|qSrr)r*r+pkg installedrrr,DszPackage(s) not found: %s, cst|fddtjDS)Ncs(g|] }dd|Dkr|jqS)cSsg|]}t|jqSr)rr+)r*Zrequiredrrrr,NszSsearch_packages_info..get_requiring_packages...)requires project_namer*r-Zcanonical_namerrr,KszHsearch_packages_info..get_requiring_packages..)rr working_set)Z package_namerr4rget_requiring_packagesIs z4search_packages_info..get_requiring_packagescsg|]}|kr|qSrrr3r.rrr,RscSsg|] }|jqSr)r2)r*Zdeprrrr,Ws)r+versionlocationr1 required_byNZRECORDcSsg|]}|ddqS),r)split)r*lrrrr,`scsg|]}tjj|qSr)ospathjoinr8r*pdistrrr,ascsg|]}tj|jqSrr=r>relpathr8r@rBrrr,bsZMETADATAzinstalled-files.txtcsg|]}tjj|qSr)r=r>r?Zegg_infor@rBrrr,jscsg|]}tj|jqSrrDr@rBrrr,kszPKG-INFOzentry_points.txt entry_pointsZ INSTALLER installer)metadata-versionsummary home-pageauthor author-emaillicensez Classifier: classifiersr )rr5rr2sortedziprrr?r7r8r1 isinstanceZDistInfoDistributionZ has_metadataZget_metadata_linesZ get_metadatastriprZfeedcloseget splitlines startswithappendlen)r"rAZ query_namesZmissingr6packageZ file_listZmetadatalinespathsrFlineZ feed_parserZ pkg_info_dictkeyrNr)rCr/rr 7sl                    r Fc Csd}t|D]\}}d}|dkr*tdtd|ddtd|d dtd |d dtd |d dtd|ddtd|ddtd|ddtd|ddtdd|dgtdd|dg|rdtd|ddtd|ddtd|d gD]}td!|q(td"|d#gD]}td!|qN|r td$|d%gD]}td!|q|d%|kr td&q |S)'zD Print the informations from installed distributions found. FTrz---zName: %sr+z Version: %sr7z Summary: %srIz Home-page: %srJz Author: %srKzAuthor-email: %srLz License: %srMz Location: %sr8z Requires: %sr0r1zRequired-by: %sr9zMetadata-Version: %srHz Installer: %srGz Classifiers:rNz %sz Entry-points:rFzFiles:r z!Cannot locate installed-files.txt) enumerater rTr?rR) Z distributionsrrZresults_printedirCZ classifierentryr\rrrr!sB  r!)FF)Z __future__rZloggingr=Z email.parserrZ pip._vendorrZpip._vendor.packaging.utilsrZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrr Zpip._internal.utils.miscr Z getLoggerr%rr r r!rrrrs       #Xsite-packages/pip/_internal/commands/__pycache__/help.cpython-38.pyc000064400000002242151464452710021431 0ustar00U .e@sDddlmZddlmZddlmZddlmZGdddeZdS))absolute_import)Command)SUCCESS) CommandErrorc@s eZdZdZdZdZddZdS) HelpCommandzShow help for commandsz %prog Tc Csddlm}m}m}z |d}Wntk r8tYSX||krt||}d|g}|rf|d|td|||} | j tS)Nr) commands_dictcreate_commandget_similar_commandszunknown command "%s"zmaybe you meant "%s"z - ) Zpip._internal.commandsrrr IndexErrorrappendrjoinparserZ print_help) selfZoptionsargsrrr Zcmd_nameZguessmsgZcommandr?/usr/lib/python3.8/site-packages/pip/_internal/commands/help.pyruns    zHelpCommand.runN)__name__ __module__ __qualname____doc__ZusageZignore_require_venvrrrrrr srN) Z __future__rZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrZpip._internal.exceptionsrrrrrrs    site-packages/pip/_internal/commands/__pycache__/configuration.cpython-38.opt-1.pyc000064400000014657151464452710024324 0ustar00U .e:@sddlZddlZddlZddlmZddlmZmZddlm Z m Z m Z ddl m Z ddlmZmZeeZGdddeZdS) N)Command)ERRORSUCCESS) Configurationget_configuration_fileskinds)PipError)get_prog write_outputcsxeZdZdZdZdZfddZddZdd Zd d Z d d Z ddZ ddZ ddZ ddZddZddZZS)ConfigurationCommanda9Manage local and global configuration. Subcommands: list: List the active configuration (or from the file specified) edit: Edit the configuration file in an editor get: Get the value associated with name set: Set the name=value unset: Unset the value associated with name If none of --user, --global and --site are passed, a virtual environment configuration file is used if one is active and the file exists. Otherwise, all modifications happen on the to the user file by default. Tz %prog [] list %prog [] [--editor ] edit %prog [] get name %prog [] set name value %prog [] unset name cstt|j||d|_|jjdddddd|jjdddd d d|jjd d dd d d|jjdddd dd|jd|jdS)Nz--editoreditorZstorez\Editor to use to edit the file. Uses VISUAL or EDITOR environment variables if not provided.)destactiondefaulthelpz--global global_file store_trueFz+Use the system-wide configuration file onlyz--user user_filez$Use the user configuration file onlyz--site site_filez3Use the current environment configuration file onlyr)superr __init__ configurationZcmd_optsZ add_optionparserZinsert_option_group)selfargskwargs __class__H/usr/lib/python3.8/site-packages/pip/_internal/commands/configuration.pyr0s> zConfigurationCommand.__init__c Cs|j|j|j|j|jd}|r*|d|krHtddt |t S|d}z|j ||dkd}Wn:t k r}zt|j dt WYSd}~XYnXt|j|d|_|jz||||ddWn<t k r}zt|j dt WYSd}~XYnXtS) N)listeditgetsetunsetrzNeed an action ({}) to perform.z, )r"r#r$r!) need_value)isolated load_only) list_valuesopen_in_editorget_nameset_name_value unset_nameloggererrorformatjoinsortedr_determine_filerrrZ isolated_moderloadr)roptionsrZhandlersrr'errrrunZs>    zConfigurationCommand.runcCsddtj|jftj|jftj|jffD}|s`|s8dStddttjDrXtjStjSnt |dkrt|dSt ddS)NcSsg|]\}}|r|qSrr).0keyvaluerrr sz8ConfigurationCommand._determine_file..css|]}tj|VqdS)N)ospathexists)r8Zsite_config_filerrr sz7ConfigurationCommand._determine_file..r(rzLNeed exactly one file to operate upon (--user, --site, --global) to perform.) rZUSERrZGLOBALrZSITEranyrlenr)rr5r%Z file_optionsrrrr3s$     z$ConfigurationCommand._determine_filecCs8|j|dddt|jD]\}}td||qdS)Nr rnz%s=%r) _get_n_argsr2ritemsr rr5rr9r:rrrr)sz ConfigurationCommand.list_valuescCs*|j|ddd}|j|}td|dS)Nz get [name]r(rBz%s)rDrZ get_valuer rFrrrr+s zConfigurationCommand.get_namecCs.|j|ddd\}}|j|||dS)Nzset [name] [value]rB)rDrZ set_value_save_configurationrFrrrr,sz#ConfigurationCommand.set_name_valuecCs(|j|ddd}|j||dS)Nz unset [name]r(rB)rDrZ unset_valuerH)rr5rr9rrrr-s zConfigurationCommand.unset_namec Csp||}|j}|dkr$tdzt||gWn4tjk rj}ztd|jW5d}~XYnXdS)Nz%Could not determine appropriate file.z*Editor Subprocess exited with exit code {}) _determine_editorrZget_file_to_editr subprocessZ check_callZCalledProcessErrorr0 returncode)rr5rr Zfnamer6rrrr*s  z#ConfigurationCommand.open_in_editorcCs<t||kr$d|t|}t||dkr4|dS|SdS)zJHelper to make sure the command got the right number of arguments zJGot unexpected number of arguments, expected {}. (example: "{} config {}")r(rN)rAr0r r)rrZexamplerCmsgrrrrDs z ConfigurationCommand._get_n_argscCs>z|jWn*tk r8tjdddtdYnXdS)Nz:Unable to save configuration. Please report this as a bug.r()exc_infozInternal Error.)rZsave Exceptionr.r/r)rrrrrHsz(ConfigurationCommand._save_configurationcCsD|jdk r|jSdtjkr$tjdSdtjkr8tjdStddS)NZVISUALZEDITORz"Could not determine editor to use.)r r<environr)rr5rrrrIs     z&ConfigurationCommand._determine_editor)__name__ __module__ __qualname____doc__Zignore_require_venvZusagerr7r3r)r+r,r-r*rDrHrI __classcell__rrrrr s *+ r )Zloggingr<rJZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrrZpip._internal.configurationrrrZpip._internal.exceptionsrZpip._internal.utils.miscr r Z getLoggerrPr.r rrrrs   site-packages/pip/_internal/commands/__pycache__/freeze.cpython-38.opt-1.pyc000064400000005540151464452710022724 0ustar00U .e @s|ddlmZddlZddlmZddlmZddlmZddl m Z ddl m Z ddl mZd d d d hZGd ddeZdS))absolute_importN) WheelCache) cmdoptions)Command) FormatControl)freeze) stdlib_pkgsZpipZ setuptoolsZ distributeZwheelcs0eZdZdZdZdZfddZddZZS) FreezeCommandzx Output installed packages in requirements format. packages are listed in a case-insensitive sorted order. z %prog [options])ext://sys.stderrr c stt|j|||jjddddgddd|jjdd d dgd d d|jjd dddddd|jjdddddd|jt|jjdddddtd|jjddddd|j d|jdS) Nz-rz --requirement requirementsappendfilez}Use the order in the given requirements file and its comments when generating output. This option can be used multiple times.)destactiondefaultmetavarhelpz-fz --find-links find_linksZURLzs        site-packages/pip/_internal/commands/__pycache__/check.cpython-38.pyc000064400000002440151464452710021556 0ustar00U .e@sNddlZddlmZddlmZmZddlmZee Z GdddeZ dS)N)Command)check_package_set!create_package_set_from_installed) write_outputc@seZdZdZdZddZdS) CheckCommandz7Verify installed packages have compatible dependencies.z %prog [options]c Cst\}}t|\}}|D].}||j}||D]} td||| dq0q|D]4}||j}||D]\} } } td||| | | qdqN|s|s|rdStddS)Nz*%s %s requires %s, which is not installed.rz-%s %s has requirement %s, but you have %s %s.zNo broken requirements found.)rrversionr) selfZoptionsargsZ package_setZ parsing_probsZmissingZ conflictingZ project_namerZ dependencyZdep_nameZ dep_versionZreqr @/usr/lib/python3.8/site-packages/pip/_internal/commands/check.pyruns2      zCheckCommand.runN)__name__ __module__ __qualname____doc__Zusager r r r r rsr) ZloggingZpip._internal.cli.base_commandrZpip._internal.operations.checkrrZpip._internal.utils.miscrZ getLoggerrZloggerrr r r r s    site-packages/pip/_internal/commands/__pycache__/hash.cpython-38.opt-1.pyc000064400000003706151464452710022371 0ustar00U .e@sddlmZddlZddlZddlZddlmZddlmZddl m Z m Z ddl m Z mZeeZGdddeZd d ZdS) )absolute_importN)Command)ERROR) FAVORITE_HASH STRONG_HASHES) read_chunks write_outputcs0eZdZdZdZdZfddZddZZS) HashCommandz Compute a hash of a local package archive. These can be used with --hash in a requirements file to do repeatable installs. z%prog [options] ...Tc sJtt|j|||jjdddtdtddtd|j d|jdS) Nz-az --algorithm algorithmZstorez$The hash algorithm to use: one of %sz, )destchoicesactiondefaulthelpr) superr __init__Zcmd_optsZ add_optionrrjoinparserZinsert_option_group)selfargskw __class__?/usr/lib/python3.8/site-packages/pip/_internal/commands/hash.pyrszHashCommand.__init__cCs>|s|jtjtS|j}|D]}td||t||q dS)Nz%s: --hash=%s:%s)rZ print_usagesysstderrrr r _hash_of_file)rZoptionsrr pathrrrrun)szHashCommand.run) __name__ __module__ __qualname____doc__ZusageZignore_require_venvrr __classcell__rrrrr s  r c Cs@t|d(}t|}t|D]}||qW5QRX|S)z!Return the hash digest of a file.rb)openhashlibnewrupdateZ hexdigest)rr archivehashchunkrrrr4s    r)Z __future__rr'ZloggingrZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrZpip._internal.utils.hashesrrZpip._internal.utils.miscrrZ getLoggerr Zloggerr rrrrrs    "site-packages/pip/_internal/commands/__pycache__/debug.cpython-38.opt-1.pyc000064400000006313151464452710022531 0ustar00U .eB @sddlmZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddl mZddlmZdd lmZdd lmZerdd lmZmZdd lmZeeZd dZddZddZGdddeZ dS))absolute_importN) cmdoptions)Command)make_target_python)SUCCESS) indent_log)get_pip_version)MYPY_CHECK_RUNNING) format_tag)AnyList)ValuescCstd||dS)Nz{}: {})loggerinfoformat)namevaluer@/usr/lib/python3.8/site-packages/pip/_internal/commands/debug.py show_valuesrc CsFtdttdr"tj}|j}nd}ttd|W5QRXdS)Nzsys.implementation:implementationr)rrhasattrsysrrrr)rZimplementation_namerrrshow_sys_implementations  rc Csd}t|}|}|}d}|r.d|}dt||}t||jdkrpt||krpd}|d|}nd}t8|D]}tt |q|rdj|d }t|W5QRXdS) N rz (target: {})zCompatible tags: {}{}TFz?... [First {tag_limit} tags shown. Pass --verbose to show all.]) tag_limit) rZget_tagsZ format_givenrlenrrverboserr ) optionsrZ target_pythonZtagsZformatted_targetsuffixmsgZ tags_limitedtagrrr show_tags,s,  r$cs0eZdZdZdZdZfddZddZZS) DebugCommandz$ Display debug information. z %prog Tcs4tt|j|||j}t||jd|dS)Nr)superr%__init__cmd_optsrZadd_target_python_optionsparserZinsert_option_group)selfargskwr( __class__rrr'Ws zDebugCommand.__init__cCsvtdtdttdtjtdtjtdttdttdt tdtj t t |tS) NzThis command is only meant for debugging. Do not use this with automation for parsing and getting these details, since the output and options of this command may change without notice.z pip versionz sys.versionzsys.executablezsys.getdefaultencodingzsys.getfilesystemencodingzlocale.getpreferredencodingz sys.platform)rZwarningrrrversion executablegetdefaultencodinggetfilesystemencodinglocaleZgetpreferredencodingplatformrr$r)r*r r+rrrrun^s     zDebugCommand.run) __name__ __module__ __qualname____doc__ZusageZignore_require_venvr'r5 __classcell__rrr-rr%Ns  r%)!Z __future__rr3ZloggingrZpip._internal.clirZpip._internal.cli.base_commandrZpip._internal.cli.cmdoptionsrZpip._internal.cli.status_codesrZpip._internal.utils.loggingrZpip._internal.utils.miscrZpip._internal.utils.typingr Zpip._internal.wheelr typingr r Zoptparser Z getLoggerr6rrrr$r%rrrrs&            "site-packages/pip/_internal/commands/__pycache__/uninstall.cpython-38.opt-1.pyc000064400000005206151464452710023454 0ustar00U .e @svddlmZddlmZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZGd d d eeZd S) )absolute_import)canonicalize_name)Command)SessionCommandMixin)InstallationError)parse_requirements)install_req_from_line)(protect_pip_from_modification_on_windowscs,eZdZdZdZfddZddZZS)UninstallCommandaB Uninstall packages. pip is able to uninstall most installed packages. Known exceptions are: - Pure distutils packages installed with ``python setup.py install``, which leave behind no metadata to determine what files were installed. - Script wrappers installed by ``python setup.py develop``. zU %prog [options] ... %prog [options] -r ...c sVtt|j|||jjddddgddd|jjdd d d d d |jd|jdS)Nz-rz --requirement requirementsappendfilezjUninstall all the packages listed in the given requirements file. This option can be used multiple times.)destactiondefaultmetavarhelpz-yz--yesyes store_truez2Don't ask for confirmation of uninstall deletions.)rrrr)superr __init__Zcmd_optsZ add_optionparserZinsert_option_group)selfargskw __class__D/usr/lib/python3.8/site-packages/pip/_internal/commands/uninstall.pyrs$ zUninstallCommand.__init__c Cs||}i}|D]&}t||jd}|jr||t|j<q|jD],}t|||dD]}|jrR||t|j<qRq@|stdt|jdt d|kd| D]&}|j |j |j dkd}|r|qdS) N)isolated)optionssessionzLYou must give at least one requirement to %(name)s (see "pip help %(name)s"))nameZpip)Z modifying_pipr)Z auto_confirmverbose)Zget_default_sessionrZ isolated_moder"rr rrdictr valuesZ uninstallr verbosityZcommit) rr rr!Zreqs_to_uninstallr"ZreqfilenameZuninstall_pathsetrrrrun2sB     zUninstallCommand.run)__name__ __module__ __qualname____doc__Zusagerr( __classcell__rrrrr s  r N)Z __future__rZpip._vendor.packaging.utilsrZpip._internal.cli.base_commandrZpip._internal.cli.req_commandrZpip._internal.exceptionsrZpip._internal.reqrZpip._internal.req.constructorsrZpip._internal.utils.miscr r rrrrs        site-packages/pip/_internal/commands/__pycache__/__init__.cpython-38.pyc000064400000005444151464452710022247 0ustar00U .e@s:dZddlmZddlZddlmZmZddlmZerPddl m Z ddl m Z edd Z ed e d d d fde dddfde dddfde dddfde dddfde dd d!fd"e d#d$d%fd&e d'd(d)fd*e d+d,d-fd.e d/d0d1fd2e d3d4d5fd6e d7d8d9fd:e d;de d?d@dAfgZdBdCZdDdEZdS)Fz% Package containing all pip commands )absolute_importN) OrderedDict namedtuple)MYPY_CHECK_RUNNING)Any)Command CommandInfoz module_path, class_name, summaryZinstallzpip._internal.commands.installZInstallCommandzInstall packages.Zdownloadzpip._internal.commands.downloadZDownloadCommandzDownload packages.Z uninstallz pip._internal.commands.uninstallZUninstallCommandzUninstall packages.Zfreezezpip._internal.commands.freezeZ FreezeCommandz1Output installed packages in requirements format.listzpip._internal.commands.listZ ListCommandzList installed packages.Zshowzpip._internal.commands.showZ ShowCommandz*Show information about installed packages.Zcheckzpip._internal.commands.checkZ CheckCommandz7Verify installed packages have compatible dependencies.Zconfigz$pip._internal.commands.configurationZConfigurationCommandz&Manage local and global configuration.searchzpip._internal.commands.searchZ SearchCommandzSearch PyPI for packages.Zwheelzpip._internal.commands.wheelZ WheelCommandz$Build wheels from your requirements.hashzpip._internal.commands.hashZ HashCommandz#Compute hashes of package archives.Z completionz!pip._internal.commands.completionZCompletionCommandz-A helper command used for command completion.debugzpip._internal.commands.debugZ DebugCommandz&Show information useful for debugging.helpzpip._internal.commands.helpZ HelpCommandzShow help for commands.cKs:t|\}}}t|}t||}|f||d|}|S)zF Create an instance of the Command class with the given name. )namesummary) commands_dict importlib import_modulegetattr)rkwargsZ module_path class_namermoduleZ command_classZcommandrC/usr/lib/python3.8/site-packages/pip/_internal/commands/__init__.pycreate_commandZs   rcCs6ddlm}|}||t}|r.|dSdSdS)zCommand name auto-correct.r)get_close_matchesFN)Zdifflibrlowerrkeys)rrZclose_commandsrrrget_similar_commandsgs  r)__doc__Z __future__rr collectionsrrZpip._internal.utils.typingrtypingrZpip._internal.cli.base_commandrrrrrrrrrs     < site-packages/pip/_internal/commands/__pycache__/list.cpython-38.opt-1.pyc000064400000021267151464452710022423 0ustar00U .e4)@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZdd lmZdd lmZdd lmZmZmZdd lmZeeZGd dde ZddZddZddZ dS))absolute_importN)six) zip_longest) cmdoptions)IndexGroupCommand) CommandError) PackageFinder)SelectionPreferences)make_link_collector)dist_is_editableget_installed_distributions write_output) get_installercsdeZdZdZdZfddZddZddZd d Zd d Z d dZ ddZ ddZ ddZ ZS) ListCommandzt List installed packages, including editables. Packages are listed in a case-insensitive sorted order. z %prog [options]cstt|j|||j}|jdddddd|jddddd d|jd d ddd d|jd ddddd|jjdddddd|t|jddddd|jddddddd|jddddd |jd!d"d#d$d |jd%dd#d&d'd(ttj|j }|j d)||j d)|dS)*Nz-oz --outdated store_trueFzList outdated packages)actiondefaulthelpz-uz --uptodatezList uptodate packagesz-ez --editablezList editable projects.z-lz--localzSIf in a virtualenv that has global access, do not list globally-installed packages.z--useruserz,Only output packages installed in user-site.)destrrrz--prezYInclude pre-release and development versions. By default, pip only finds stable versions.z--formatZstore list_formatcolumns)rfreezejsonzBSelect the output format among: columns (default), freeze, or json)rrrchoicesrz--not-required not_requiredz>List packages that are not dependencies of installed packages.)rrrz--exclude-editableZ store_falseinclude_editablez%Exclude editable package from output.z--include-editablez%Include editable package from output.T)rrrrr) superr__init__cmd_optsZ add_optionrZ list_pathZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrZ index_opts __class__?/usr/lib/python3.8/site-packages/pip/_internal/commands/list.pyr&s zListCommand.__init__cCs(t||d}td|jd}tj||dS)zK Create a package finder appropriate to this list command. )optionsF)Z allow_yankedZallow_all_prereleases)link_collectorselection_prefs)r r prerZcreate)r!r(sessionr)r*r&r&r'_build_package_findervs z!ListCommand._build_package_findercCs|jr|jrtdt|t|j|j|j|j |j d}|j rL| ||}|jr`| ||}n|jrr|||}|||dS)Nz5Options --outdated and --uptodate cannot be combined.)Z local_onlyZ user_onlyZeditables_onlyZinclude_editablespaths)outdatedZuptodaterrZcheck_list_path_optionr ZlocalrZeditablerpathrget_not_required get_outdated get_uptodateoutput_package_listing)r!r(r"packagesr&r&r'runs&     zListCommand.runcCsdd|||DS)NcSsg|]}|j|jkr|qSr&latest_versionZparsed_version.0distr&r&r' s z,ListCommand.get_outdated..iter_packages_latest_infosr!r5r(r&r&r'r2s zListCommand.get_outdatedcCsdd|||DS)NcSsg|]}|j|jkr|qSr&r7r9r&r&r'r<s z,ListCommand.get_uptodate..r=r?r&r&r'r3s zListCommand.get_uptodatecs:t|D]}dd|Dq fdd|DS)Ncss|] }|jVqdSNkey)r:Z requirementr&r&r' sz/ListCommand.get_not_required..csh|]}|jkr|qSr&rA)r:ZpkgZdep_keysr&r' s z/ListCommand.get_not_required..)setupdateZrequires)r!r5r(r;r&rDr'r1szListCommand.get_not_requiredc cs||}|||}|D]t}d}||j}|jsDdd|D}|j|jd}||} | dkrfq| j} | j j rzd}nd}| |_ ||_ |VqW5QRXdS)NunknowncSsg|]}|jjs|qSr&)versionZ is_prerelease)r: candidater&r&r'r<sz:ListCommand.iter_packages_latest_infos..) project_nameZwheelZsdist) Z_build_sessionr-Zfind_all_candidatesrBr+Zmake_candidate_evaluatorrKZsort_best_candidaterIlinkZis_wheelr8latest_filetype) r!r5r(r,finderr;typZall_candidatesZ evaluatorZbest_candidateZremote_versionr&r&r'r>s(    z&ListCommand.iter_packages_latest_infoscCst|ddd}|jdkr:|r:t||\}}|||n^|jdkr|D]4}|jdkrltd|j|j|jqHtd|j|jqHn|jd krtt ||dS) NcSs |jSr@)rKlower)r;r&r&r'z4ListCommand.output_package_listing..rArrz %s==%s (%s)z%s==%sr) sortedrformat_for_columnsoutput_package_listing_columnsverboser rKrIlocationformat_for_json)r!r5r(dataheaderr;r&r&r'r4s"   z"ListCommand.output_package_listingcCsbt|dkr|d|t|\}}t|dkrL|ddtdd||D] }t|qPdS)NrrS cSsd|S)N-r&)xr&r&r'rQrRz.)leninserttabulatejoinmapr )r!rZr[Z pkg_stringssizesvalr&r&r'rVs    z*ListCommand.output_package_listing_columns)__name__ __module__ __qualname____doc__Zusagerr-r6r2r3r1r>r4rV __classcell__r&r&r$r'rs PrcCspdgtdd|D}|D]}ddt||D}qg}|D](}dddt||D}||q>||fS)Nrcss|]}t|VqdSr@)r_r:r^r&r&r'rCsztabulate..cSs"g|]\}}t|tt|qSr&)maxr_strr:scr&r&r'r<sztabulate..r\cSs*g|]"\}}|dk r"t||ndqS)N)rmljustrnr&r&r'r<s)rlrrbappend)ZvalsrdrowresultZdisplayr&r&r'ras  racCs|j}|rddddg}nddg}g}|jdks@tdd|DrJ|d|jdkr^|d |D]l}|j|jg}|r||j||j|jdkst|r||j |jdkr|t |||qb||fS) z_ Convert the package data into something usable by output_package_listing_columns. ZPackageZVersionZLatestZTyperScss|]}t|VqdSr@)r rkr&r&r'rCsz%format_for_columns..ZLocationZ Installer) r/rWanyrsrKrIr8rMr rXr)Zpkgsr(Zrunning_outdatedr[rZZprojrtr&r&r'rUs(         rUcCsvg}|D]b}|jt|jd}|jdkr@|j|d<t||d<|jr`t|j|d<|j |d<| |qt |S)N)namerIrSrXZ installerr8rM) rKrZ text_typerIrWrXrr/r8rMrsrdumps)r5r(rZr;infor&r&r'rY+s      rY)!Z __future__rrZloggingZ pip._vendorrZpip._vendor.six.movesrZpip._internal.clirZpip._internal.cli.req_commandrZpip._internal.exceptionsrZpip._internal.indexrZ$pip._internal.models.selection_prefsr Z!pip._internal.self_outdated_checkr Zpip._internal.utils.miscr r r Zpip._internal.utils.packagingrZ getLoggerrfZloggerrrarUrYr&r&r&r's$           Y%site-packages/pip/_internal/commands/__pycache__/show.cpython-38.pyc000064400000014265151464452710021471 0ustar00U .e@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z mZddlmZeeZGd d d e Zd d ZdddZdS))absolute_importN) FeedParser) pkg_resourcescanonicalize_name)Command)ERRORSUCCESS) write_outputcs0eZdZdZdZdZfddZddZZS) ShowCommandzx Show information about one or more installed packages. The output is in RFC-compliant mail header format. z$ %prog [options] ...Tcs>tt|j|||jjddddddd|jd|jdS) Nz-fz--filesfiles store_trueFz7Show the full list of installed files for each package.)destactiondefaulthelpr)superr __init__Zcmd_optsZ add_optionparserZinsert_option_group)selfargskw __class__?/usr/lib/python3.8/site-packages/pip/_internal/commands/show.pyrszShowCommand.__init__cCs8|stdtS|}t|}t||j|jds4tStS)Nz.ERROR: Please provide a package name or names.) list_filesverbose)loggerwarningrsearch_packages_info print_resultsr rr )rZoptionsrqueryresultsrrrrun*s zShowCommand.run) __name__ __module__ __qualname____doc__ZusageZignore_require_venvrr$ __classcell__rrrrr s  r c#sTitjD]}|t|j<q dd|D}tfddt||D}|r^tdd|dd}fdd|DD]ԉjj j d d D|jd }d }d }t tj rd rd }d d|D} fdd| D} fdd| D}drnd}nPdrXd} fdd| D} fdd| D}drnd}drd} | |d<drƈdD]"} | r| |d<qƐqt} | || } dD]}| |||<qg}|D](} | dr|| tdd q||d<|rHt||d<|Vqxd S)z Gather details from installed distributions. Print distribution name, version, location, and installed files. Installed files requires a pip generated 'installed-files.txt' in the distributions '.egg-info' directory. cSsg|] }t|qSrr).0namerrr Bsz(search_packages_info..csg|]\}}|kr|qSrr)r*r+pkg installedrrr,DszPackage(s) not found: %s, cst|fddtjDS)Ncs(g|] }dd|Dkr|jqS)cSsg|]}t|jqSr)rr+)r*Zrequiredrrrr,NszSsearch_packages_info..get_requiring_packages...)requires project_namer*r-Zcanonical_namerrr,KszHsearch_packages_info..get_requiring_packages..)rr working_set)Z package_namerr4rget_requiring_packagesIs z4search_packages_info..get_requiring_packagescsg|]}|kr|qSrrr3r.rrr,RscSsg|] }|jqSr)r2)r*Zdeprrrr,Ws)r+versionlocationr1 required_byNZRECORDcSsg|]}|ddqS),r)split)r*lrrrr,`scsg|]}tjj|qSr)ospathjoinr8r*pdistrrr,ascsg|]}tj|jqSrr=r>relpathr8r@rBrrr,bsZMETADATAzinstalled-files.txtcsg|]}tjj|qSr)r=r>r?Zegg_infor@rBrrr,jscsg|]}tj|jqSrrDr@rBrrr,kszPKG-INFOzentry_points.txt entry_pointsZ INSTALLER installer)metadata-versionsummary home-pageauthor author-emaillicensez Classifier: classifiersr )rr5rr2sortedziprrr?r7r8r1 isinstanceZDistInfoDistributionZ has_metadataZget_metadata_linesZ get_metadatastriprZfeedcloseget splitlines startswithappendlen)r"rAZ query_namesZmissingr6packageZ file_listZmetadatalinespathsrFlineZ feed_parserZ pkg_info_dictkeyrNr)rCr/rr 7sl                    r Fc Csd}t|D]\}}d}|dkr*tdtd|ddtd|d dtd |d dtd |d dtd|ddtd|ddtd|ddtd|ddtdd|dgtdd|dg|rdtd|ddtd|ddtd|d gD]}td!|q(td"|d#gD]}td!|qN|r td$|d%gD]}td!|q|d%|kr td&q |S)'zD Print the informations from installed distributions found. FTrz---zName: %sr+z Version: %sr7z Summary: %srIz Home-page: %srJz Author: %srKzAuthor-email: %srLz License: %srMz Location: %sr8z Requires: %sr0r1zRequired-by: %sr9zMetadata-Version: %srHz Installer: %srGz Classifiers:rNz %sz Entry-points:rFzFiles:r z!Cannot locate installed-files.txt) enumerater rTr?rR) Z distributionsrrZresults_printedirCZ classifierentryr\rrrr!sB  r!)FF)Z __future__rZloggingr=Z email.parserrZ pip._vendorrZpip._vendor.packaging.utilsrZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrr Zpip._internal.utils.miscr Z getLoggerr%rr r r!rrrrs       #Xsite-packages/pip/_internal/commands/__pycache__/configuration.cpython-38.pyc000064400000014657151464452710023365 0ustar00U .e:@sddlZddlZddlZddlmZddlmZmZddlm Z m Z m Z ddl m Z ddlmZmZeeZGdddeZdS) N)Command)ERRORSUCCESS) Configurationget_configuration_fileskinds)PipError)get_prog write_outputcsxeZdZdZdZdZfddZddZdd Zd d Z d d Z ddZ ddZ ddZ ddZddZddZZS)ConfigurationCommanda9Manage local and global configuration. Subcommands: list: List the active configuration (or from the file specified) edit: Edit the configuration file in an editor get: Get the value associated with name set: Set the name=value unset: Unset the value associated with name If none of --user, --global and --site are passed, a virtual environment configuration file is used if one is active and the file exists. Otherwise, all modifications happen on the to the user file by default. Tz %prog [] list %prog [] [--editor ] edit %prog [] get name %prog [] set name value %prog [] unset name cstt|j||d|_|jjdddddd|jjdddd d d|jjd d dd d d|jjdddd dd|jd|jdS)Nz--editoreditorZstorez\Editor to use to edit the file. Uses VISUAL or EDITOR environment variables if not provided.)destactiondefaulthelpz--global global_file store_trueFz+Use the system-wide configuration file onlyz--user user_filez$Use the user configuration file onlyz--site site_filez3Use the current environment configuration file onlyr)superr __init__ configurationZcmd_optsZ add_optionparserZinsert_option_group)selfargskwargs __class__H/usr/lib/python3.8/site-packages/pip/_internal/commands/configuration.pyr0s> zConfigurationCommand.__init__c Cs|j|j|j|j|jd}|r*|d|krHtddt |t S|d}z|j ||dkd}Wn:t k r}zt|j dt WYSd}~XYnXt|j|d|_|jz||||ddWn<t k r}zt|j dt WYSd}~XYnXtS) N)listeditgetsetunsetrzNeed an action ({}) to perform.z, )r"r#r$r!) need_value)isolated load_only) list_valuesopen_in_editorget_nameset_name_value unset_nameloggererrorformatjoinsortedr_determine_filerrrZ isolated_moderloadr)roptionsrZhandlersrr'errrrunZs>    zConfigurationCommand.runcCsddtj|jftj|jftj|jffD}|s`|s8dStddttjDrXtjStjSnt |dkrt|dSt ddS)NcSsg|]\}}|r|qSrr).0keyvaluerrr sz8ConfigurationCommand._determine_file..css|]}tj|VqdS)N)ospathexists)r8Zsite_config_filerrr sz7ConfigurationCommand._determine_file..r(rzLNeed exactly one file to operate upon (--user, --site, --global) to perform.) rZUSERrZGLOBALrZSITEranyrlenr)rr5r%Z file_optionsrrrr3s$     z$ConfigurationCommand._determine_filecCs8|j|dddt|jD]\}}td||qdS)Nr rnz%s=%r) _get_n_argsr2ritemsr rr5rr9r:rrrr)sz ConfigurationCommand.list_valuescCs*|j|ddd}|j|}td|dS)Nz get [name]r(rBz%s)rDrZ get_valuer rFrrrr+s zConfigurationCommand.get_namecCs.|j|ddd\}}|j|||dS)Nzset [name] [value]rB)rDrZ set_value_save_configurationrFrrrr,sz#ConfigurationCommand.set_name_valuecCs(|j|ddd}|j||dS)Nz unset [name]r(rB)rDrZ unset_valuerH)rr5rr9rrrr-s zConfigurationCommand.unset_namec Csp||}|j}|dkr$tdzt||gWn4tjk rj}ztd|jW5d}~XYnXdS)Nz%Could not determine appropriate file.z*Editor Subprocess exited with exit code {}) _determine_editorrZget_file_to_editr subprocessZ check_callZCalledProcessErrorr0 returncode)rr5rr Zfnamer6rrrr*s  z#ConfigurationCommand.open_in_editorcCs<t||kr$d|t|}t||dkr4|dS|SdS)zJHelper to make sure the command got the right number of arguments zJGot unexpected number of arguments, expected {}. (example: "{} config {}")r(rN)rAr0r r)rrZexamplerCmsgrrrrDs z ConfigurationCommand._get_n_argscCs>z|jWn*tk r8tjdddtdYnXdS)Nz:Unable to save configuration. Please report this as a bug.r()exc_infozInternal Error.)rZsave Exceptionr.r/r)rrrrrHsz(ConfigurationCommand._save_configurationcCsD|jdk r|jSdtjkr$tjdSdtjkr8tjdStddS)NZVISUALZEDITORz"Could not determine editor to use.)r r<environr)rr5rrrrIs     z&ConfigurationCommand._determine_editor)__name__ __module__ __qualname____doc__Zignore_require_venvZusagerr7r3r)r+r,r-r*rDrHrI __classcell__rrrrr s *+ r )Zloggingr<rJZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrrZpip._internal.configurationrrrZpip._internal.exceptionsrZpip._internal.utils.miscr r Z getLoggerrPr.r rrrrs   site-packages/pip/_internal/commands/__pycache__/check.cpython-38.opt-1.pyc000064400000002440151464452710022515 0ustar00U .e@sNddlZddlmZddlmZmZddlmZee Z GdddeZ dS)N)Command)check_package_set!create_package_set_from_installed) write_outputc@seZdZdZdZddZdS) CheckCommandz7Verify installed packages have compatible dependencies.z %prog [options]c Cst\}}t|\}}|D].}||j}||D]} td||| dq0q|D]4}||j}||D]\} } } td||| | | qdqN|s|s|rdStddS)Nz*%s %s requires %s, which is not installed.rz-%s %s has requirement %s, but you have %s %s.zNo broken requirements found.)rrversionr) selfZoptionsargsZ package_setZ parsing_probsZmissingZ conflictingZ project_namerZ dependencyZdep_nameZ dep_versionZreqr @/usr/lib/python3.8/site-packages/pip/_internal/commands/check.pyruns2      zCheckCommand.runN)__name__ __module__ __qualname____doc__Zusager r r r r rsr) ZloggingZpip._internal.cli.base_commandrZpip._internal.operations.checkrrZpip._internal.utils.miscrZ getLoggerrZloggerrr r r r s    site-packages/pip/_internal/commands/__pycache__/search.cpython-38.pyc000064400000010604151464452710021747 0ustar00U .e@sddlmZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddlmZddlmZdd lmZmZdd lmZdd lmZdd lmZdd lmZddlmZddlm Z e!e"Z#GdddeeZ$ddZ%dddZ&ddZ'dS))absolute_importN) OrderedDict) pkg_resources)parse) xmlrpc_client)Command)SessionCommandMixin)NO_MATCHES_FOUNDSUCCESS) CommandError)PyPI)PipXmlrpcTransport)get_terminal_size) indent_log) write_outputcs8eZdZdZdZdZfddZddZdd ZZ S) SearchCommandz@Search for PyPI packages whose name or summary contains .z %prog [options] Tcs@tt|j|||jjddddtjdd|jd|jdS)Nz-iz--indexindexZURLz3Base URL of Python Package Index (default %default))destmetavardefaulthelpr) superr__init__Zcmd_optsZ add_optionr Zpypi_urlparserZinsert_option_group)selfargskw __class__A/usr/lib/python3.8/site-packages/pip/_internal/commands/search.pyr%szSearchCommand.__init__cCsT|s td|}|||}t|}d}tjrksz!print_results..cSsg|] }|jqSr)Z project_name)r?prrr r@psr,r-r5r<   z %-*s - %sz%s (%s)zINSTALLED: %s (latest)z INSTALLED: %sz=LATEST: %s (pre-release; install with "pip install --pre")z LATEST: %s)maxrZ working_setr8r>textwrapZwrapjoinrZget_distributionrr4 parse_versionZpreUnicodeEncodeError) r*Zname_column_widthr!Zinstalled_packagesr;r,r-ZlatestZ target_widthlineZdistrrr r'gsJ          r'cCs t|tdS)N)key)rIrL)r5rrr r8sr8)NN)(Z __future__rZloggingr$rJ collectionsrZ pip._vendorrZpip._vendor.packaging.versionrrLZpip._vendor.six.movesrZpip._internal.cli.base_commandrZpip._internal.cli.req_commandrZpip._internal.cli.status_codesr r Zpip._internal.exceptionsr Zpip._internal.models.indexr Zpip._internal.network.xmlrpcr Zpip._internal.utils.compatrZpip._internal.utils.loggingrZpip._internal.utils.miscrZ getLoggerr/Zloggerrr#r'r8rrrr s*              - )site-packages/pip/_internal/commands/__pycache__/completion.cpython-38.opt-1.pyc000064400000005705151464452710023620 0ustar00U .e @sXddlmZddlZddlZddlmZddlmZdZdddd Z Gd d d eZ dS) )absolute_importN)Command)get_progzJ # pip %(shell)s completion start%(script)s# pip %(shell)s completion end a _pip_completion() { COMPREPLY=( $( COMP_WORDS="${COMP_WORDS[*]}" \ COMP_CWORD=$COMP_CWORD \ PIP_AUTO_COMPLETE=1 $1 2>/dev/null ) ) } complete -o default -F _pip_completion %(prog)s aM function _pip_completion { local words cword read -Ac words read -cn cword reply=( $( COMP_WORDS="$words[*]" \ COMP_CWORD=$(( cword-1 )) \ PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null )) } compctl -K _pip_completion %(prog)s aw function __fish_complete_pip set -lx COMP_WORDS (commandline -o) "" set -lx COMP_CWORD ( \ math (contains -i -- (commandline -t) $COMP_WORDS)-1 \ ) set -lx PIP_AUTO_COMPLETE 1 string split \ -- (eval $COMP_WORDS[1]) end complete -fa "(__fish_complete_pip)" -c %(prog)s )bashzshfishcs,eZdZdZdZfddZddZZS)CompletionCommandz3A helper command to be used for command completion.Tcsltt|j|||j}|jddddddd|jdd dd dd d|jd d ddddd|jd|dS)Nz--bashz-b store_constrshellzEmit completion code for bash)actionconstdesthelpz--zshz-zrzEmit completion code for zshz--fishz-frzEmit completion code for fishr)superr__init__cmd_optsZ add_optionparserZinsert_option_group)selfargskwr __class__E/usr/lib/python3.8/site-packages/pip/_internal/commands/completion.pyr8s6zCompletionCommand.__init__cCsrt}ddt|D}|j|krXtt|jddti}tt ||jdnt j dd |dS) z-Prints the completion code of the given shellcSsg|] }d|qS)z--r).0r rrr Usz)CompletionCommand.run..prog)scriptr zERROR: You must pass %s z or N)COMPLETION_SCRIPTSkeyssortedr textwrapdedentgetrprintBASE_COMPLETIONsysstderrwritejoin)rZoptionsrZshellsZ shell_optionsrrrrrunRs   zCompletionCommand.run)__name__ __module__ __qualname____doc__Zignore_require_venvrr+ __classcell__rrrrr3s r) Z __future__rr'r"Zpip._internal.cli.base_commandrZpip._internal.utils.miscrr&rrrrrrs     #site-packages/pip/_internal/commands/__pycache__/list.cpython-38.pyc000064400000021336151464452710021461 0ustar00U .e4)@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZdd lmZdd lmZdd lmZmZmZdd lmZeeZGd dde ZddZddZddZ dS))absolute_importN)six) zip_longest) cmdoptions)IndexGroupCommand) CommandError) PackageFinder)SelectionPreferences)make_link_collector)dist_is_editableget_installed_distributions write_output) get_installercsdeZdZdZdZfddZddZddZd d Zd d Z d dZ ddZ ddZ ddZ ZS) ListCommandzt List installed packages, including editables. Packages are listed in a case-insensitive sorted order. z %prog [options]cstt|j|||j}|jdddddd|jddddd d|jd d ddd d|jd ddddd|jjdddddd|t|jddddd|jddddddd|jddddd |jd!d"d#d$d |jd%dd#d&d'd(ttj|j }|j d)||j d)|dS)*Nz-oz --outdated store_trueFzList outdated packages)actiondefaulthelpz-uz --uptodatezList uptodate packagesz-ez --editablezList editable projects.z-lz--localzSIf in a virtualenv that has global access, do not list globally-installed packages.z--useruserz,Only output packages installed in user-site.)destrrrz--prezYInclude pre-release and development versions. By default, pip only finds stable versions.z--formatZstore list_formatcolumns)rfreezejsonzBSelect the output format among: columns (default), freeze, or json)rrrchoicesrz--not-required not_requiredz>List packages that are not dependencies of installed packages.)rrrz--exclude-editableZ store_falseinclude_editablez%Exclude editable package from output.z--include-editablez%Include editable package from output.T)rrrrr) superr__init__cmd_optsZ add_optionrZ list_pathZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrZ index_opts __class__?/usr/lib/python3.8/site-packages/pip/_internal/commands/list.pyr&s zListCommand.__init__cCs(t||d}td|jd}tj||dS)zK Create a package finder appropriate to this list command. )optionsF)Z allow_yankedZallow_all_prereleases)link_collectorselection_prefs)r r prerZcreate)r!r(sessionr)r*r&r&r'_build_package_findervs z!ListCommand._build_package_findercCs|jr|jrtdt|t|j|j|j|j |j d}|j rL| ||}|jr`| ||}n|jrr|||}|||dS)Nz5Options --outdated and --uptodate cannot be combined.)Z local_onlyZ user_onlyZeditables_onlyZinclude_editablespaths)outdatedZuptodaterrZcheck_list_path_optionr ZlocalrZeditablerpathrget_not_required get_outdated get_uptodateoutput_package_listing)r!r(r"packagesr&r&r'runs&     zListCommand.runcCsdd|||DS)NcSsg|]}|j|jkr|qSr&latest_versionZparsed_version.0distr&r&r' s z,ListCommand.get_outdated..iter_packages_latest_infosr!r5r(r&r&r'r2s zListCommand.get_outdatedcCsdd|||DS)NcSsg|]}|j|jkr|qSr&r7r9r&r&r'r<s z,ListCommand.get_uptodate..r=r?r&r&r'r3s zListCommand.get_uptodatecs:t|D]}dd|Dq fdd|DS)Ncss|] }|jVqdSNkey)r:Z requirementr&r&r' sz/ListCommand.get_not_required..csh|]}|jkr|qSr&rA)r:ZpkgZdep_keysr&r' s z/ListCommand.get_not_required..)setupdateZrequires)r!r5r(r;r&rDr'r1szListCommand.get_not_requiredc cs||}|||}|D]t}d}||j}|jsDdd|D}|j|jd}||} | dkrfq| j} | j j rzd}nd}| |_ ||_ |VqW5QRXdS)NunknowncSsg|]}|jjs|qSr&)versionZ is_prerelease)r: candidater&r&r'r<sz:ListCommand.iter_packages_latest_infos..) project_nameZwheelZsdist) Z_build_sessionr-Zfind_all_candidatesrBr+Zmake_candidate_evaluatorrKZsort_best_candidaterIlinkZis_wheelr8latest_filetype) r!r5r(r,finderr;typZall_candidatesZ evaluatorZbest_candidateZremote_versionr&r&r'r>s(    z&ListCommand.iter_packages_latest_infoscCst|ddd}|jdkr:|r:t||\}}|||n^|jdkr|D]4}|jdkrltd|j|j|jqHtd|j|jqHn|jd krtt ||dS) NcSs |jSr@)rKlower)r;r&r&r'z4ListCommand.output_package_listing..rArrz %s==%s (%s)z%s==%sr) sortedrformat_for_columnsoutput_package_listing_columnsverboser rKrIlocationformat_for_json)r!r5r(dataheaderr;r&r&r'r4s"   z"ListCommand.output_package_listingcCsbt|dkr|d|t|\}}t|dkrL|ddtdd||D] }t|qPdS)NrrS cSsd|S)N-r&)xr&r&r'rQrRz.)leninserttabulatejoinmapr )r!rZr[Z pkg_stringssizesvalr&r&r'rVs    z*ListCommand.output_package_listing_columns)__name__ __module__ __qualname____doc__Zusagerr-r6r2r3r1r>r4rV __classcell__r&r&r$r'rs PrcCst|dkstdgtdd|D}|D]}ddt||D}q,g}|D](}dddt||D}||qN||fS)Nrcss|]}t|VqdSr@)r_r:r^r&r&r'rCsztabulate..cSs"g|]\}}t|tt|qSr&)maxr_strr:scr&r&r'r<sztabulate..r\cSs*g|]"\}}|dk r"t||ndqS)N)rmljustrnr&r&r'r<s)r_AssertionErrorrlrrbappend)ZvalsrdrowresultZdisplayr&r&r'ras  racCs|j}|rddddg}nddg}g}|jdks@tdd|DrJ|d|jdkr^|d |D]l}|j|jg}|r||j||j|jdkst|r||j |jdkr|t |||qb||fS) z_ Convert the package data into something usable by output_package_listing_columns. ZPackageZVersionZLatestZTyperScss|]}t|VqdSr@)r rkr&r&r'rCsz%format_for_columns..ZLocationZ Installer) r/rWanyrtrKrIr8rMr rXr)Zpkgsr(Zrunning_outdatedr[rZZprojrur&r&r'rUs(         rUcCsvg}|D]b}|jt|jd}|jdkr@|j|d<t||d<|jr`t|j|d<|j |d<| |qt |S)N)namerIrSrXZ installerr8rM) rKrZ text_typerIrWrXrr/r8rMrtrdumps)r5r(rZr;infor&r&r'rY+s      rY)!Z __future__rrZloggingZ pip._vendorrZpip._vendor.six.movesrZpip._internal.clirZpip._internal.cli.req_commandrZpip._internal.exceptionsrZpip._internal.indexrZ$pip._internal.models.selection_prefsr Z!pip._internal.self_outdated_checkr Zpip._internal.utils.miscr r r Zpip._internal.utils.packagingrZ getLoggerrfZloggerrrarUrYr&r&r&r's$           Y%site-packages/pip/_internal/commands/__pycache__/freeze.cpython-38.pyc000064400000005540151464452710021765 0ustar00U .e @s|ddlmZddlZddlmZddlmZddlmZddl m Z ddl m Z ddl mZd d d d hZGd ddeZdS))absolute_importN) WheelCache) cmdoptions)Command) FormatControl)freeze) stdlib_pkgsZpipZ setuptoolsZ distributeZwheelcs0eZdZdZdZdZfddZddZZS) FreezeCommandzx Output installed packages in requirements format. packages are listed in a case-insensitive sorted order. z %prog [options])ext://sys.stderrr c stt|j|||jjddddgddd|jjdd d dgd d d|jjd dddddd|jjdddddd|jt|jjdddddtd|jjddddd|j d|jdS) Nz-rz --requirement requirementsappendfilez}Use the order in the given requirements file and its comments when generating output. This option can be used multiple times.)destactiondefaultmetavarhelpz-fz --find-links find_linksZURLzs        site-packages/pip/_internal/commands/__pycache__/install.cpython-38.pyc000064400000034172151464452710022156 0ustar00U .e_@sddlmZddlZddlZddlZddlZddlZddlZddlmZddl m Z ddl m Z ddl mZddlmZddlmZdd lmZdd lmZdd lmZmZdd lmZmZmZdd lmZddl m!Z!ddl"m#Z#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*m+Z+m,Z,m-Z-ddl.m/Z/ddl0m1Z1ddl2m3Z3ddl4m5Z5e1rddl m6Z6ddl7m8Z8m9Z9m:Z:ddl;mZ>ddl4m?Z?e@eAZBddZCddZDd d!ZEGd"d#d#eZFd$d%ZGd&d'ZHdS)()absolute_importN)path) SUPPRESS_HELP) pkg_resources)canonicalize_name) WheelCache) cmdoptions)make_target_python)RequirementCommand)ERRORSUCCESS) CommandErrorInstallationErrorPreviousBuildDirErrordistutils_scheme)check_install_conflicts)RequirementSetinstall_given_reqs)RequirementTracker)check_path_owner) ensure_dirget_installed_version(protect_pip_from_modification_on_windows write_output) TempDirectory)MYPY_CHECK_RUNNING)virtualenv_no_global) WheelBuilder)Values)AnyListOptional) FormatControl)InstallRequirement)BinaryAllowedPredicatecCs(z ddl}Wntk r"YdSXdS)z8 Return whether the wheel package is installed. rNFT)wheel ImportError)r&r(B/usr/lib/python3.8/site-packages/pip/_internal/commands/install.pyis_wheel_installed<s  r*cCs*t}|j|dd}|r&|j|dd|S)zQ Build wheels for requirements, depending on whether wheel is installed. T)Z should_unpack)r*Zbuild)builderpep517_requirementslegacy_requirementsZshould_build_legacybuild_failuresr(r(r) build_wheelsHs r/csfdd}|S)Ncs&|jr dSt|j}|}d|kS)NTZbinary) use_pep517rnameZget_allowed_formats)reqZcanonical_nameZallowed_formatsformat_controlr(r)check_binary_allowedhs   z6get_check_binary_allowed..check_binary_allowedr()r4r5r(r3r)get_check_binary_allowedfs r6cs<eZdZdZdZfddZddZddZd d ZZ S) InstallCommandaI Install packages from: - PyPI (and other indexes) using requirement specifiers. - VCS project urls. - Local project directories. - Local or remote source archives. pip also supports installing from "requirements files", which provide an easy way to specify a whole environment to be installed. a% %prog [options] [package-index-options] ... %prog [options] -r [package-index-options] ... %prog [options] [-e] ... %prog [options] [-e] ... %prog [options] ...cs^tt|j|||j}|t|t|t|t |t |jdddddddt ||jddd d d |jd dd t d |jdddddd|jdddddd|t |t|jdddd dd |jdddddgdd|jddd d d |jd!d"d#d d$d |t|t|t|t|t|t|jd%d d&d'd(d)|jd*d d&d+d,|jd-d d.d'd/d)|jd0d d1d'd2d)|t|t|t|t|t|tttj|j}|jd3||jd3|dS)4Nz-tz--target target_dirdirzInstall packages into . By default this will not replace existing files/folders in . Use --upgrade to replace existing packages in with new versions.)destmetavardefaulthelp--user use_user_site store_truezInstall to the Python user install directory for your platform. Typically ~/.local/, or %APPDATA%\Python on Windows. (See the Python documentation for site.USER_BASE for full details.))r:actionr=z --no-userZ store_falsez--root root_pathz=Install everything relative to this alternate root directory.z--prefix prefix_pathzIInstallation prefix where lib, bin and other top-level folders are placedz-Uz --upgradeupgradezUpgrade all specified packages to the newest available version. The handling of dependencies depends on the upgrade-strategy used.z--upgrade-strategyupgrade_strategyzonly-if-neededZeageraGDetermines how dependency upgrading should be handled [default: %default]. "eager" - dependencies are upgraded regardless of whether the currently installed version satisfies the requirements of the upgraded package(s). "only-if-needed" - are upgraded only when they do not satisfy the requirements of the upgraded package(s).)r:r<choicesr=z--force-reinstallforce_reinstallz;Reinstall all packages even if they are already up-to-date.z-Iz--ignore-installedignore_installedzIgnore the installed packages, overwriting them. This can break your system if the existing package is of a different version or was installed with a different package manager!z --compilecompileTz'Compile Python source files to bytecode)rAr:r<r=z --no-compilez.Do not compile Python source files to bytecode)rAr:r=z--no-warn-script-locationwarn_script_locationz0Do not warn when installing scripts outside PATHz--no-warn-conflictswarn_about_conflictsz%Do not warn about broken dependenciesr)superr7__init__cmd_optsZ add_optionr requirementsZ constraintsZno_depsZpreZeditableZadd_target_python_optionsr build_dirsrcignore_requires_pythonZno_build_isolationr0Z no_use_pep517install_optionsglobal_optionsZ no_binaryZ only_binaryZ prefer_binaryno_cleanrequire_hashesZ progress_barZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrNZ index_opts __class__r(r)rMs   zInstallCommand.__init__c*Cst|dd}tdkrZ|sZttjd}|dkrLttjd}t d|d}|j rj|j }|j rtj|j |_ tj|dd tj|j|_|jpg}|jr|jrtd trtd |d |d d}d}|jrJd|_tj|j|_tj|jr,tj|js,tdtdd}|j}|d||jpTg} ||} t|} |j || | |j!d} |j"p|j } t#|j$|j%}|j$rt&|j$st d|j$d|_$t'}t|j | dd}t(|j)|j d}zz*|,|||| | ||j-|||d}|j.|| | |||j|j|j!|j/||j0d }|1|z|2d}Wnt3k rzd}Yn X|j4dk}t5|dt6| j%}g}g}|j78D]$}|j0r||n ||qt9||gg|d}t:|||d}|rtd;d o.|j?}|r@|@||jA}|jrRd!}tB||| |jC||j|jD||jd" }tE|j||jC|j|jFd#} tGH| }!tI|tJKd$d%}"g}#|"D]R}|jL}$z$tM|jL|!d&}%|%r|$d'|%7}$WntNk rYnX|#|$qd(<|#}&|&r tOd)|&WntPk r}'zN|jQd*k}(tR|'|(|j})t jS|)|(d+tTWYWHW5QRW5QRSd}'~'XYntUk rd|_"YnXW5|j"s|*|+XW5QRXW5QRX|jr|V|j||j tWS),NcSs ttdpttdotjtjkS)NZ real_prefix base_prefix)hasattrsysr]prefixr(r(r(r)is_venvs   z#InstallCommand.run..is_venvrz __main__.pyz -m pipzgRunning pip install with root privileges is generally not a good idea. Try `%s install --user` instead.zto-satisfy-onlyT)Z check_targetzVCan not combine '--user' and '--prefix' as they imply different installation locationszZCan not perform a '--user' install. User site-packages are not visible in this virtualenv.r>z --prefix=z=Target path exists but is not a directory, will not continue.target)kindz--home=)optionssession target_pythonrRzThe directory '%s' or its parent directory is not owned by the current user and caching wheels has been disabled. check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag.Zinstall)deleterc)rVZcheck_supported_wheels)Ztemp_build_dirrd req_tracker) preparerfinderrerd wheel_cacher?rHrRrGrEr0Zpip) modifying_pip)Z build_optionsrTr5)r+r,r-zPCould not build wheels for {} which use PEP 517 and cannot be installed directlyz, css|] }|jVqdSN)r1).0rr(r(r) sz%InstallCommand.run..F)roothomer`Z pycompilerJr?)userrrrqr`isolatedr1)key) working_set- zSuccessfully installed %sexc_info)XrZcheck_install_build_globalosgetuidrbasenamer_argv executableloggerwarningrDrErPabspathZcheck_dist_restrictionZsrc_dirrSr?rCr rrappendr8rHexistsisdirrrTZget_default_sessionr Z_build_package_finderrRrUr cache_dirr4rrrrVZ cleanup_filesZcleanupZpopulate_requirement_setZmake_requirement_preparerZ make_resolverrGr0ZresolveZget_requirementKeyErrorZ satisfied_byrr6rOvaluesrr/formatjoinZget_installation_orderZignore_dependenciesrK_warn_about_conflictsrJrrBrIget_lib_location_guessesZ isolated_moderZ WorkingSetsortedoperator attrgetterr1r ExceptionrEnvironmentError verbositycreate_env_error_messageerrorr r_handle_target_dirr )*rXrdrYraZcommandrErStarget_temp_dirZtarget_temp_dir_pathrTrerfrjZ build_deleterkrhZ directoryZrequirement_setriZresolverZpip_reqrlr5r-r,r2Z wheel_builderr. to_installZshould_warn_about_conflictsrJZ installedZ lib_locationsrvZreqsitemsitemZinstalled_versionZinstalled_descrshow_tracebackmessager(r(r)runs                   4 zInstallCommand.runc sft|g}|Jtd|jd}|d}|d}|d}tj|rP||tj|rn||krn||tj|r|||D]} t| D]} | |krtj|| tfdd|ddDrqtj|| } tj| r>|st d | qtj | rt d | qtj | r4t | n t| t tj| | | qqW5QRXdS) N)rrpurelibplatlibdatac3s|]}|VqdSrm) startswith)rnsddirr(r)rp+sz4InstallCommand._handle_target_dir..zKTarget directory %s already exists. Specify --upgrade to force replacement.zTarget directory %s already exists and is a link. Pip will not automatically replace links, please remove if replacement is desired.)rrrr|rrlistdirranyrrislinkrshutilZrmtreeremoveZmove) rXr8rrDZ lib_dir_listschemeZ purelib_dirZ platlib_dirZdata_dirZlib_dirrZtarget_item_dirr(rr)rsP        z!InstallCommand._handle_target_dirc Cszt|\}}Wn$tk r4tjdddYdSX|\}}|D]2}||d}||D]}td|||dqZqB|D]8}||d}||D]\} } } td||| | | qqzdS)NzError checking for conflicts.Trzrz*%s %s requires %s, which is not installed.ryzF%s %s has requirement %s, but you'll have %s %s which is incompatible.)rrrrZcritical) rXrZ package_setZ _dep_infoZmissingZ conflictingZ project_nameversionZ dependencyZdep_nameZ dep_versionr2r(r(r)rIs4   z$InstallCommand._warn_about_conflicts) __name__ __module__ __qualname____doc__ZusagerMrrr __classcell__r(r(r[r)r7ss  w8r7cOstd||}|d|dgS)Nrrr)rr)rYkwargsrr(r(r)rdsrcCsg}|d|s,|d|t|n |d|dd7<|jtjkrd}d}|st||d|gn |||d d |dS) z{Format an error message for an EnvironmentError It may occur anytime during the execution of the install command. z5Could not install packages due to an EnvironmentErrorz: .r z"Consider using the `--user` optionzCheck the permissionsz or z. r)rstrerrnoZEACCESextendlowerrstrip)rrZusing_user_sitepartsZuser_option_partZpermissions_partr(r(r)ris&      r)IZ __future__rrZloggingrr|rr_rZoptparserZ pip._vendorrZpip._vendor.packaging.utilsrZpip._internal.cacherZpip._internal.clirZpip._internal.cli.cmdoptionsr Zpip._internal.cli.req_commandr Zpip._internal.cli.status_codesr r Zpip._internal.exceptionsr rrZpip._internal.locationsrZpip._internal.operations.checkrZpip._internal.reqrrZpip._internal.req.req_trackerrZpip._internal.utils.filesystemrZpip._internal.utils.miscrrrrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrZpip._internal.wheelrrtypingr r!r"Z#pip._internal.models.format_controlr#Zpip._internal.req.req_installr$r%Z getLoggerrrr*r/r6r7rrr(r(r(r)sT                        tsite-packages/pip/_internal/commands/__pycache__/wheel.cpython-38.opt-1.pyc000064400000011015151464452710022542 0ustar00U .e@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z m Z ddl mZddlmZdd lmZdd lmZdd lmZerdd lmZdd lmZmZeeZGddde ZdS))absolute_importN) WheelCache) cmdoptions)RequirementCommand) CommandErrorPreviousBuildDirError)RequirementSet)RequirementTracker) TempDirectory)MYPY_CHECK_RUNNING) WheelBuilder)Values)AnyListcs,eZdZdZdZfddZddZZS) WheelCommanda Build Wheel archives for your requirements and dependencies. Wheel is a built-package format, and offers the advantage of not recompiling your software during every install. For more details, see the wheel docs: https://wheel.readthedocs.io/en/latest/ Requirements: setuptools>=0.8, and wheel. 'pip wheel' uses the bdist_wheel setuptools extension from the wheel package to build individual wheels. z %prog [options] ... %prog [options] -r ... %prog [options] [-e] ... %prog [options] [-e] ... %prog [options] ...csztt|j|||j}|jddddtjdd|t|t |t |jddd d d d |t |t |t |t|t|t|t|t|t|t|t|jd dd d dd|jddddd|t|tttj|j}|jd||jd|dS)Nz-wz --wheel-dir wheel_dirdirzLBuild wheels into , where the default is the current working directory.)destmetavardefaulthelpz--build-option build_optionsoptionsappendz9Extra arguments to be supplied to 'setup.py bdist_wheel'.)rractionrz--global-optionglobal_optionszZExtra global options to be supplied to the setup.py call before the 'bdist_wheel' command.)rrrrz--pre store_trueFzYInclude pre-release and development versions. By default, pip only finds stable versions.)rrrr)superr__init__cmd_optsZ add_optionoscurdirrZ no_binaryZ only_binaryZ prefer_binaryZno_build_isolation use_pep517Z no_use_pep517Z constraintsZeditable requirementssrcignore_requires_pythonZno_deps build_dirZ progress_barno_cleanrequire_hashesZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrZ index_opts __class__@/usr/lib/python3.8/site-packages/pip/_internal/commands/wheel.pyr3shzWheelCommand.__init__c Cszt||jr tj|j|_tj|j|_||}|||}|j pP|j }t |j |j }t }t|j|dd}t|jd} zz|| ||||||j||||jd} |j| |||||j|jd} | | t| ||jpg|jpg|j d} | | j} t | dkr"t!dWnt"k rBd |_ YnXW5|j s`| |XW5QRXW5QRXdS) NZwheel)deleteZkind)r()Ztemp_build_dirr req_trackerZwheel_download_dir)preparerfindersessionr wheel_cacher%r")rrr'rz"Failed to build one or more wheelsT)#rZcheck_install_build_globalr&r pathabspathZsrc_dirZget_default_sessionZ_build_package_finderr'r cache_dirZformat_controlr r rr(Z cleanup_filesZcleanupZpopulate_requirement_setZmake_requirement_preparerrZ make_resolverr%r"Zresolver rrZbuildr#valueslenrr)r*rr+r5r4Z build_deleter6r2Z directoryZrequirement_setr3ZresolverwbZbuild_failuresr/r/r0runqs|      zWheelCommand.run)__name__ __module__ __qualname____doc__Zusagerr= __classcell__r/r/r-r0rs >r) Z __future__rZloggingr Zpip._internal.cacherZpip._internal.clirZpip._internal.cli.req_commandrZpip._internal.exceptionsrrZpip._internal.reqrZpip._internal.req.req_trackerr Zpip._internal.utils.temp_dirr Zpip._internal.utils.typingr Zpip._internal.wheelr Zoptparser typingrrZ getLoggerr>Zloggerrr/r/r/r0s           site-packages/pip/_internal/commands/__pycache__/__init__.cpython-38.opt-1.pyc000064400000005444151464452710023206 0ustar00U .e@s:dZddlmZddlZddlmZmZddlmZerPddl m Z ddl m Z edd Z ed e d d d fde dddfde dddfde dddfde dddfde dd d!fd"e d#d$d%fd&e d'd(d)fd*e d+d,d-fd.e d/d0d1fd2e d3d4d5fd6e d7d8d9fd:e d;de d?d@dAfgZdBdCZdDdEZdS)Fz% Package containing all pip commands )absolute_importN) OrderedDict namedtuple)MYPY_CHECK_RUNNING)Any)Command CommandInfoz module_path, class_name, summaryZinstallzpip._internal.commands.installZInstallCommandzInstall packages.Zdownloadzpip._internal.commands.downloadZDownloadCommandzDownload packages.Z uninstallz pip._internal.commands.uninstallZUninstallCommandzUninstall packages.Zfreezezpip._internal.commands.freezeZ FreezeCommandz1Output installed packages in requirements format.listzpip._internal.commands.listZ ListCommandzList installed packages.Zshowzpip._internal.commands.showZ ShowCommandz*Show information about installed packages.Zcheckzpip._internal.commands.checkZ CheckCommandz7Verify installed packages have compatible dependencies.Zconfigz$pip._internal.commands.configurationZConfigurationCommandz&Manage local and global configuration.searchzpip._internal.commands.searchZ SearchCommandzSearch PyPI for packages.Zwheelzpip._internal.commands.wheelZ WheelCommandz$Build wheels from your requirements.hashzpip._internal.commands.hashZ HashCommandz#Compute hashes of package archives.Z completionz!pip._internal.commands.completionZCompletionCommandz-A helper command used for command completion.debugzpip._internal.commands.debugZ DebugCommandz&Show information useful for debugging.helpzpip._internal.commands.helpZ HelpCommandzShow help for commands.cKs:t|\}}}t|}t||}|f||d|}|S)zF Create an instance of the Command class with the given name. )namesummary) commands_dict importlib import_modulegetattr)rkwargsZ module_path class_namermoduleZ command_classZcommandrC/usr/lib/python3.8/site-packages/pip/_internal/commands/__init__.pycreate_commandZs   rcCs6ddlm}|}||t}|r.|dSdSdS)zCommand name auto-correct.r)get_close_matchesFN)Zdifflibrlowerrkeys)rrZclose_commandsrrrget_similar_commandsgs  r)__doc__Z __future__rr collectionsrrZpip._internal.utils.typingrtypingrZpip._internal.cli.base_commandrrrrrrrrrs     < site-packages/pip/_internal/commands/__pycache__/search.cpython-38.opt-1.pyc000064400000010604151464452710022706 0ustar00U .e@sddlmZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddlmZddlmZdd lmZmZdd lmZdd lmZdd lmZdd lmZddlmZddlm Z e!e"Z#GdddeeZ$ddZ%dddZ&ddZ'dS))absolute_importN) OrderedDict) pkg_resources)parse) xmlrpc_client)Command)SessionCommandMixin)NO_MATCHES_FOUNDSUCCESS) CommandError)PyPI)PipXmlrpcTransport)get_terminal_size) indent_log) write_outputcs8eZdZdZdZdZfddZddZdd ZZ S) SearchCommandz@Search for PyPI packages whose name or summary contains .z %prog [options] Tcs@tt|j|||jjddddtjdd|jd|jdS)Nz-iz--indexindexZURLz3Base URL of Python Package Index (default %default))destmetavardefaulthelpr) superr__init__Zcmd_optsZ add_optionr Zpypi_urlparserZinsert_option_group)selfargskw __class__A/usr/lib/python3.8/site-packages/pip/_internal/commands/search.pyr%szSearchCommand.__init__cCsT|s td|}|||}t|}d}tjrksz!print_results..cSsg|] }|jqSr)Z project_name)r?prrr r@psr,r-r5r<   z %-*s - %sz%s (%s)zINSTALLED: %s (latest)z INSTALLED: %sz=LATEST: %s (pre-release; install with "pip install --pre")z LATEST: %s)maxrZ working_setr8r>textwrapZwrapjoinrZget_distributionrr4 parse_versionZpreUnicodeEncodeError) r*Zname_column_widthr!Zinstalled_packagesr;r,r-ZlatestZ target_widthlineZdistrrr r'gsJ          r'cCs t|tdS)N)key)rIrL)r5rrr r8sr8)NN)(Z __future__rZloggingr$rJ collectionsrZ pip._vendorrZpip._vendor.packaging.versionrrLZpip._vendor.six.movesrZpip._internal.cli.base_commandrZpip._internal.cli.req_commandrZpip._internal.cli.status_codesr r Zpip._internal.exceptionsr Zpip._internal.models.indexr Zpip._internal.network.xmlrpcr Zpip._internal.utils.compatrZpip._internal.utils.loggingrZpip._internal.utils.miscrZ getLoggerr/Zloggerrr#r'r8rrrr s*              - )site-packages/pip/_internal/commands/__pycache__/uninstall.cpython-38.pyc000064400000005206151464452710022515 0ustar00U .e @svddlmZddlmZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZGd d d eeZd S) )absolute_import)canonicalize_name)Command)SessionCommandMixin)InstallationError)parse_requirements)install_req_from_line)(protect_pip_from_modification_on_windowscs,eZdZdZdZfddZddZZS)UninstallCommandaB Uninstall packages. pip is able to uninstall most installed packages. Known exceptions are: - Pure distutils packages installed with ``python setup.py install``, which leave behind no metadata to determine what files were installed. - Script wrappers installed by ``python setup.py develop``. zU %prog [options] ... %prog [options] -r ...c sVtt|j|||jjddddgddd|jjdd d d d d |jd|jdS)Nz-rz --requirement requirementsappendfilezjUninstall all the packages listed in the given requirements file. This option can be used multiple times.)destactiondefaultmetavarhelpz-yz--yesyes store_truez2Don't ask for confirmation of uninstall deletions.)rrrr)superr __init__Zcmd_optsZ add_optionparserZinsert_option_group)selfargskw __class__D/usr/lib/python3.8/site-packages/pip/_internal/commands/uninstall.pyrs$ zUninstallCommand.__init__c Cs||}i}|D]&}t||jd}|jr||t|j<q|jD],}t|||dD]}|jrR||t|j<qRq@|stdt|jdt d|kd| D]&}|j |j |j dkd}|r|qdS) N)isolated)optionssessionzLYou must give at least one requirement to %(name)s (see "pip help %(name)s"))nameZpip)Z modifying_pipr)Z auto_confirmverbose)Zget_default_sessionrZ isolated_moder"rr rrdictr valuesZ uninstallr verbosityZcommit) rr rr!Zreqs_to_uninstallr"ZreqfilenameZuninstall_pathsetrrrrun2sB     zUninstallCommand.run)__name__ __module__ __qualname____doc__Zusagerr( __classcell__rrrrr s  r N)Z __future__rZpip._vendor.packaging.utilsrZpip._internal.cli.base_commandrZpip._internal.cli.req_commandrZpip._internal.exceptionsrZpip._internal.reqrZpip._internal.req.constructorsrZpip._internal.utils.miscr r rrrrs        site-packages/pip/_internal/commands/__pycache__/debug.cpython-38.pyc000064400000006313151464452710021572 0ustar00U .eB @sddlmZddlZddlZddlZddlmZddlmZddl m Z ddl m Z ddl mZddlmZdd lmZdd lmZerdd lmZmZdd lmZeeZd dZddZddZGdddeZ dS))absolute_importN) cmdoptions)Command)make_target_python)SUCCESS) indent_log)get_pip_version)MYPY_CHECK_RUNNING) format_tag)AnyList)ValuescCstd||dS)Nz{}: {})loggerinfoformat)namevaluer@/usr/lib/python3.8/site-packages/pip/_internal/commands/debug.py show_valuesrc CsFtdttdr"tj}|j}nd}ttd|W5QRXdS)Nzsys.implementation:implementationr)rrhasattrsysrrrr)rZimplementation_namerrrshow_sys_implementations  rc Csd}t|}|}|}d}|r.d|}dt||}t||jdkrpt||krpd}|d|}nd}t8|D]}tt |q|rdj|d }t|W5QRXdS) N rz (target: {})zCompatible tags: {}{}TFz?... [First {tag_limit} tags shown. Pass --verbose to show all.]) tag_limit) rZget_tagsZ format_givenrlenrrverboserr ) optionsrZ target_pythonZtagsZformatted_targetsuffixmsgZ tags_limitedtagrrr show_tags,s,  r$cs0eZdZdZdZdZfddZddZZS) DebugCommandz$ Display debug information. z %prog Tcs4tt|j|||j}t||jd|dS)Nr)superr%__init__cmd_optsrZadd_target_python_optionsparserZinsert_option_group)selfargskwr( __class__rrr'Ws zDebugCommand.__init__cCsvtdtdttdtjtdtjtdttdttdt tdtj t t |tS) NzThis command is only meant for debugging. Do not use this with automation for parsing and getting these details, since the output and options of this command may change without notice.z pip versionz sys.versionzsys.executablezsys.getdefaultencodingzsys.getfilesystemencodingzlocale.getpreferredencodingz sys.platform)rZwarningrrrversion executablegetdefaultencodinggetfilesystemencodinglocaleZgetpreferredencodingplatformrr$r)r*r r+rrrrun^s     zDebugCommand.run) __name__ __module__ __qualname____doc__ZusageZignore_require_venvr'r5 __classcell__rrr-rr%Ns  r%)!Z __future__rr3ZloggingrZpip._internal.clirZpip._internal.cli.base_commandrZpip._internal.cli.cmdoptionsrZpip._internal.cli.status_codesrZpip._internal.utils.loggingrZpip._internal.utils.miscrZpip._internal.utils.typingr Zpip._internal.wheelr typingr r Zoptparser Z getLoggerr6rrrr$r%rrrrs&            "site-packages/pip/_internal/commands/__pycache__/help.cpython-38.opt-1.pyc000064400000002242151464452710022370 0ustar00U .e@sDddlmZddlmZddlmZddlmZGdddeZdS))absolute_import)Command)SUCCESS) CommandErrorc@s eZdZdZdZdZddZdS) HelpCommandzShow help for commandsz %prog Tc Csddlm}m}m}z |d}Wntk r8tYSX||krt||}d|g}|rf|d|td|||} | j tS)Nr) commands_dictcreate_commandget_similar_commandszunknown command "%s"zmaybe you meant "%s"z - ) Zpip._internal.commandsrrr IndexErrorrappendrjoinparserZ print_help) selfZoptionsargsrrr Zcmd_nameZguessmsgZcommandr?/usr/lib/python3.8/site-packages/pip/_internal/commands/help.pyruns    zHelpCommand.runN)__name__ __module__ __qualname____doc__ZusageZignore_require_venvrrrrrr srN) Z __future__rZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrZpip._internal.exceptionsrrrrrrs    site-packages/pip/_internal/commands/__pycache__/wheel.cpython-38.pyc000064400000011015151464452710021603 0ustar00U .e@sddlmZddlZddlZddlmZddlmZddlm Z ddl m Z m Z ddl mZddlmZdd lmZdd lmZdd lmZerdd lmZdd lmZmZeeZGddde ZdS))absolute_importN) WheelCache) cmdoptions)RequirementCommand) CommandErrorPreviousBuildDirError)RequirementSet)RequirementTracker) TempDirectory)MYPY_CHECK_RUNNING) WheelBuilder)Values)AnyListcs,eZdZdZdZfddZddZZS) WheelCommanda Build Wheel archives for your requirements and dependencies. Wheel is a built-package format, and offers the advantage of not recompiling your software during every install. For more details, see the wheel docs: https://wheel.readthedocs.io/en/latest/ Requirements: setuptools>=0.8, and wheel. 'pip wheel' uses the bdist_wheel setuptools extension from the wheel package to build individual wheels. z %prog [options] ... %prog [options] -r ... %prog [options] [-e] ... %prog [options] [-e] ... %prog [options] ...csztt|j|||j}|jddddtjdd|t|t |t |jddd d d d |t |t |t |t|t|t|t|t|t|t|t|jd dd d dd|jddddd|t|tttj|j}|jd||jd|dS)Nz-wz --wheel-dir wheel_dirdirzLBuild wheels into , where the default is the current working directory.)destmetavardefaulthelpz--build-option build_optionsoptionsappendz9Extra arguments to be supplied to 'setup.py bdist_wheel'.)rractionrz--global-optionglobal_optionszZExtra global options to be supplied to the setup.py call before the 'bdist_wheel' command.)rrrrz--pre store_trueFzYInclude pre-release and development versions. By default, pip only finds stable versions.)rrrr)superr__init__cmd_optsZ add_optionoscurdirrZ no_binaryZ only_binaryZ prefer_binaryZno_build_isolation use_pep517Z no_use_pep517Z constraintsZeditable requirementssrcignore_requires_pythonZno_deps build_dirZ progress_barno_cleanrequire_hashesZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrZ index_opts __class__@/usr/lib/python3.8/site-packages/pip/_internal/commands/wheel.pyr3shzWheelCommand.__init__c Cszt||jr tj|j|_tj|j|_||}|||}|j pP|j }t |j |j }t }t|j|dd}t|jd} zz|| ||||||j||||jd} |j| |||||j|jd} | | t| ||jpg|jpg|j d} | | j} t | dkr"t!dWnt"k rBd |_ YnXW5|j s`| |XW5QRXW5QRXdS) NZwheel)deleteZkind)r()Ztemp_build_dirr req_trackerZwheel_download_dir)preparerfindersessionr wheel_cacher%r")rrr'rz"Failed to build one or more wheelsT)#rZcheck_install_build_globalr&r pathabspathZsrc_dirZget_default_sessionZ_build_package_finderr'r cache_dirZformat_controlr r rr(Z cleanup_filesZcleanupZpopulate_requirement_setZmake_requirement_preparerrZ make_resolverr%r"Zresolver rrZbuildr#valueslenrr)r*rr+r5r4Z build_deleter6r2Z directoryZrequirement_setr3ZresolverwbZbuild_failuresr/r/r0runqs|      zWheelCommand.run)__name__ __module__ __qualname____doc__Zusagerr= __classcell__r/r/r-r0rs >r) Z __future__rZloggingr Zpip._internal.cacherZpip._internal.clirZpip._internal.cli.req_commandrZpip._internal.exceptionsrrZpip._internal.reqrZpip._internal.req.req_trackerr Zpip._internal.utils.temp_dirr Zpip._internal.utils.typingr Zpip._internal.wheelr Zoptparser typingrrZ getLoggerr>Zloggerrr/r/r/r0s           site-packages/pip/_internal/commands/__pycache__/hash.cpython-38.pyc000064400000003706151464452710021432 0ustar00U .e@sddlmZddlZddlZddlZddlmZddlmZddl m Z m Z ddl m Z mZeeZGdddeZd d ZdS) )absolute_importN)Command)ERROR) FAVORITE_HASH STRONG_HASHES) read_chunks write_outputcs0eZdZdZdZdZfddZddZZS) HashCommandz Compute a hash of a local package archive. These can be used with --hash in a requirements file to do repeatable installs. z%prog [options] ...Tc sJtt|j|||jjdddtdtddtd|j d|jdS) Nz-az --algorithm algorithmZstorez$The hash algorithm to use: one of %sz, )destchoicesactiondefaulthelpr) superr __init__Zcmd_optsZ add_optionrrjoinparserZinsert_option_group)selfargskw __class__?/usr/lib/python3.8/site-packages/pip/_internal/commands/hash.pyrszHashCommand.__init__cCs>|s|jtjtS|j}|D]}td||t||q dS)Nz%s: --hash=%s:%s)rZ print_usagesysstderrrr r _hash_of_file)rZoptionsrr pathrrrrun)szHashCommand.run) __name__ __module__ __qualname____doc__ZusageZignore_require_venvrr __classcell__rrrrr s  r c Cs@t|d(}t|}t|D]}||qW5QRX|S)z!Return the hash digest of a file.rb)openhashlibnewrupdateZ hexdigest)rr archivehashchunkrrrr4s    r)Z __future__rr'ZloggingrZpip._internal.cli.base_commandrZpip._internal.cli.status_codesrZpip._internal.utils.hashesrrZpip._internal.utils.miscrrZ getLoggerr Zloggerr rrrrrs    "site-packages/pip/_internal/commands/__pycache__/install.cpython-38.opt-1.pyc000064400000034172151464452710023115 0ustar00U .e_@sddlmZddlZddlZddlZddlZddlZddlZddlmZddl m Z ddl m Z ddl mZddlmZddlmZdd lmZdd lmZdd lmZmZdd lmZmZmZdd lmZddl m!Z!ddl"m#Z#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*m+Z+m,Z,m-Z-ddl.m/Z/ddl0m1Z1ddl2m3Z3ddl4m5Z5e1rddl m6Z6ddl7m8Z8m9Z9m:Z:ddl;mZ>ddl4m?Z?e@eAZBddZCddZDd d!ZEGd"d#d#eZFd$d%ZGd&d'ZHdS)()absolute_importN)path) SUPPRESS_HELP) pkg_resources)canonicalize_name) WheelCache) cmdoptions)make_target_python)RequirementCommand)ERRORSUCCESS) CommandErrorInstallationErrorPreviousBuildDirErrordistutils_scheme)check_install_conflicts)RequirementSetinstall_given_reqs)RequirementTracker)check_path_owner) ensure_dirget_installed_version(protect_pip_from_modification_on_windows write_output) TempDirectory)MYPY_CHECK_RUNNING)virtualenv_no_global) WheelBuilder)Values)AnyListOptional) FormatControl)InstallRequirement)BinaryAllowedPredicatecCs(z ddl}Wntk r"YdSXdS)z8 Return whether the wheel package is installed. rNFT)wheel ImportError)r&r(B/usr/lib/python3.8/site-packages/pip/_internal/commands/install.pyis_wheel_installed<s  r*cCs*t}|j|dd}|r&|j|dd|S)zQ Build wheels for requirements, depending on whether wheel is installed. T)Z should_unpack)r*Zbuild)builderpep517_requirementslegacy_requirementsZshould_build_legacybuild_failuresr(r(r) build_wheelsHs r/csfdd}|S)Ncs&|jr dSt|j}|}d|kS)NTZbinary) use_pep517rnameZget_allowed_formats)reqZcanonical_nameZallowed_formatsformat_controlr(r)check_binary_allowedhs   z6get_check_binary_allowed..check_binary_allowedr()r4r5r(r3r)get_check_binary_allowedfs r6cs<eZdZdZdZfddZddZddZd d ZZ S) InstallCommandaI Install packages from: - PyPI (and other indexes) using requirement specifiers. - VCS project urls. - Local project directories. - Local or remote source archives. pip also supports installing from "requirements files", which provide an easy way to specify a whole environment to be installed. a% %prog [options] [package-index-options] ... %prog [options] -r [package-index-options] ... %prog [options] [-e] ... %prog [options] [-e] ... %prog [options] ...cs^tt|j|||j}|t|t|t|t |t |jdddddddt ||jddd d d |jd dd t d |jdddddd|jdddddd|t |t|jdddd dd |jdddddgdd|jddd d d |jd!d"d#d d$d |t|t|t|t|t|t|jd%d d&d'd(d)|jd*d d&d+d,|jd-d d.d'd/d)|jd0d d1d'd2d)|t|t|t|t|t|tttj|j}|jd3||jd3|dS)4Nz-tz--target target_dirdirzInstall packages into . By default this will not replace existing files/folders in . Use --upgrade to replace existing packages in with new versions.)destmetavardefaulthelp--user use_user_site store_truezInstall to the Python user install directory for your platform. Typically ~/.local/, or %APPDATA%\Python on Windows. (See the Python documentation for site.USER_BASE for full details.))r:actionr=z --no-userZ store_falsez--root root_pathz=Install everything relative to this alternate root directory.z--prefix prefix_pathzIInstallation prefix where lib, bin and other top-level folders are placedz-Uz --upgradeupgradezUpgrade all specified packages to the newest available version. The handling of dependencies depends on the upgrade-strategy used.z--upgrade-strategyupgrade_strategyzonly-if-neededZeageraGDetermines how dependency upgrading should be handled [default: %default]. "eager" - dependencies are upgraded regardless of whether the currently installed version satisfies the requirements of the upgraded package(s). "only-if-needed" - are upgraded only when they do not satisfy the requirements of the upgraded package(s).)r:r<choicesr=z--force-reinstallforce_reinstallz;Reinstall all packages even if they are already up-to-date.z-Iz--ignore-installedignore_installedzIgnore the installed packages, overwriting them. This can break your system if the existing package is of a different version or was installed with a different package manager!z --compilecompileTz'Compile Python source files to bytecode)rAr:r<r=z --no-compilez.Do not compile Python source files to bytecode)rAr:r=z--no-warn-script-locationwarn_script_locationz0Do not warn when installing scripts outside PATHz--no-warn-conflictswarn_about_conflictsz%Do not warn about broken dependenciesr)superr7__init__cmd_optsZ add_optionr requirementsZ constraintsZno_depsZpreZeditableZadd_target_python_optionsr build_dirsrcignore_requires_pythonZno_build_isolationr0Z no_use_pep517install_optionsglobal_optionsZ no_binaryZ only_binaryZ prefer_binaryno_cleanrequire_hashesZ progress_barZmake_option_groupZ index_groupparserZinsert_option_group)selfargskwrNZ index_opts __class__r(r)rMs   zInstallCommand.__init__c*Cst|dd}tdkrZ|sZttjd}|dkrLttjd}t d|d}|j rj|j }|j rtj|j |_ tj|dd tj|j|_|jpg}|jr|jrtd trtd |d |d d}d}|jrJd|_tj|j|_tj|jr,tj|js,tdtdd}|j}|d||jpTg} ||} t|} |j || | |j!d} |j"p|j } t#|j$|j%}|j$rt&|j$st d|j$d|_$t'}t|j | dd}t(|j)|j d}zz*|,|||| | ||j-|||d}|j.|| | |||j|j|j!|j/||j0d }|1|z|2d}Wnt3k rzd}Yn X|j4dk}t5|dt6| j%}g}g}|j78D]$}|j0r||n ||qt9||gg|d}t:|||d}|rtd;d o.|j?}|r@|@||jA}|jrRd!}tB||| |jC||j|jD||jd" }tE|j||jC|j|jFd#} tGH| }!tI|tJKd$d%}"g}#|"D]R}|jL}$z$tM|jL|!d&}%|%r|$d'|%7}$WntNk rYnX|#|$qd(<|#}&|&r tOd)|&WntPk r}'zN|jQd*k}(tR|'|(|j})t jS|)|(d+tTWYWHW5QRW5QRSd}'~'XYntUk rd|_"YnXW5|j"s|*|+XW5QRXW5QRX|jr|V|j||j tWS),NcSs ttdpttdotjtjkS)NZ real_prefix base_prefix)hasattrsysr]prefixr(r(r(r)is_venvs   z#InstallCommand.run..is_venvrz __main__.pyz -m pipzgRunning pip install with root privileges is generally not a good idea. Try `%s install --user` instead.zto-satisfy-onlyT)Z check_targetzVCan not combine '--user' and '--prefix' as they imply different installation locationszZCan not perform a '--user' install. User site-packages are not visible in this virtualenv.r>z --prefix=z=Target path exists but is not a directory, will not continue.target)kindz--home=)optionssession target_pythonrRzThe directory '%s' or its parent directory is not owned by the current user and caching wheels has been disabled. check the permissions and owner of that directory. If executing pip with sudo, you may want sudo's -H flag.Zinstall)deleterc)rVZcheck_supported_wheels)Ztemp_build_dirrd req_tracker) preparerfinderrerd wheel_cacher?rHrRrGrEr0Zpip) modifying_pip)Z build_optionsrTr5)r+r,r-zPCould not build wheels for {} which use PEP 517 and cannot be installed directlyz, css|] }|jVqdSN)r1).0rr(r(r) sz%InstallCommand.run..F)roothomer`Z pycompilerJr?)userrrrqr`isolatedr1)key) working_set- zSuccessfully installed %sexc_info)XrZcheck_install_build_globalosgetuidrbasenamer_argv executableloggerwarningrDrErPabspathZcheck_dist_restrictionZsrc_dirrSr?rCr rrappendr8rHexistsisdirrrTZget_default_sessionr Z_build_package_finderrRrUr cache_dirr4rrrrVZ cleanup_filesZcleanupZpopulate_requirement_setZmake_requirement_preparerZ make_resolverrGr0ZresolveZget_requirementKeyErrorZ satisfied_byrr6rOvaluesrr/formatjoinZget_installation_orderZignore_dependenciesrK_warn_about_conflictsrJrrBrIget_lib_location_guessesZ isolated_moderZ WorkingSetsortedoperator attrgetterr1r ExceptionrEnvironmentError verbositycreate_env_error_messageerrorr r_handle_target_dirr )*rXrdrYraZcommandrErStarget_temp_dirZtarget_temp_dir_pathrTrerfrjZ build_deleterkrhZ directoryZrequirement_setriZresolverZpip_reqrlr5r-r,r2Z wheel_builderr. to_installZshould_warn_about_conflictsrJZ installedZ lib_locationsrvZreqsitemsitemZinstalled_versionZinstalled_descrshow_tracebackmessager(r(r)runs                   4 zInstallCommand.runc sft|g}|Jtd|jd}|d}|d}|d}tj|rP||tj|rn||krn||tj|r|||D]} t| D]} | |krtj|| tfdd|ddDrqtj|| } tj| r>|st d | qtj | rt d | qtj | r4t | n t| t tj| | | qqW5QRXdS) N)rrpurelibplatlibdatac3s|]}|VqdSrm) startswith)rnsddirr(r)rp+sz4InstallCommand._handle_target_dir..zKTarget directory %s already exists. Specify --upgrade to force replacement.zTarget directory %s already exists and is a link. Pip will not automatically replace links, please remove if replacement is desired.)rrrr|rrlistdirranyrrislinkrshutilZrmtreeremoveZmove) rXr8rrDZ lib_dir_listschemeZ purelib_dirZ platlib_dirZdata_dirZlib_dirrZtarget_item_dirr(rr)rsP        z!InstallCommand._handle_target_dirc Cszt|\}}Wn$tk r4tjdddYdSX|\}}|D]2}||d}||D]}td|||dqZqB|D]8}||d}||D]\} } } td||| | | qqzdS)NzError checking for conflicts.Trzrz*%s %s requires %s, which is not installed.ryzF%s %s has requirement %s, but you'll have %s %s which is incompatible.)rrrrZcritical) rXrZ package_setZ _dep_infoZmissingZ conflictingZ project_nameversionZ dependencyZdep_nameZ dep_versionr2r(r(r)rIs4   z$InstallCommand._warn_about_conflicts) __name__ __module__ __qualname____doc__ZusagerMrrr __classcell__r(r(r[r)r7ss  w8r7cOstd||}|d|dgS)Nrrr)rr)rYkwargsrr(r(r)rdsrcCsg}|d|s,|d|t|n |d|dd7<|jtjkrd}d}|st||d|gn |||d d |dS) z{Format an error message for an EnvironmentError It may occur anytime during the execution of the install command. z5Could not install packages due to an EnvironmentErrorz: .r z"Consider using the `--user` optionzCheck the permissionsz or z. r)rstrerrnoZEACCESextendlowerrstrip)rrZusing_user_sitepartsZuser_option_partZpermissions_partr(r(r)ris&      r)IZ __future__rrZloggingrr|rr_rZoptparserZ pip._vendorrZpip._vendor.packaging.utilsrZpip._internal.cacherZpip._internal.clirZpip._internal.cli.cmdoptionsr Zpip._internal.cli.req_commandr Zpip._internal.cli.status_codesr r Zpip._internal.exceptionsr rrZpip._internal.locationsrZpip._internal.operations.checkrZpip._internal.reqrrZpip._internal.req.req_trackerrZpip._internal.utils.filesystemrZpip._internal.utils.miscrrrrZpip._internal.utils.temp_dirrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrZpip._internal.wheelrrtypingr r!r"Z#pip._internal.models.format_controlr#Zpip._internal.req.req_installr$r%Z getLoggerrrr*r/r6r7rrr(r(r(r)sT                        tsite-packages/pip/_internal/commands/wheel.py000064400000014410151464452710015317 0ustar00# -*- coding: utf-8 -*- # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os from pip._internal.cache import WheelCache from pip._internal.cli import cmdoptions from pip._internal.cli.req_command import RequirementCommand from pip._internal.exceptions import CommandError, PreviousBuildDirError from pip._internal.req import RequirementSet from pip._internal.req.req_tracker import RequirementTracker from pip._internal.utils.temp_dir import TempDirectory from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.wheel import WheelBuilder if MYPY_CHECK_RUNNING: from optparse import Values from typing import Any, List logger = logging.getLogger(__name__) class WheelCommand(RequirementCommand): """ Build Wheel archives for your requirements and dependencies. Wheel is a built-package format, and offers the advantage of not recompiling your software during every install. For more details, see the wheel docs: https://wheel.readthedocs.io/en/latest/ Requirements: setuptools>=0.8, and wheel. 'pip wheel' uses the bdist_wheel setuptools extension from the wheel package to build individual wheels. """ usage = """ %prog [options] ... %prog [options] -r ... %prog [options] [-e] ... %prog [options] [-e] ... %prog [options] ...""" def __init__(self, *args, **kw): super(WheelCommand, self).__init__(*args, **kw) cmd_opts = self.cmd_opts cmd_opts.add_option( '-w', '--wheel-dir', dest='wheel_dir', metavar='dir', default=os.curdir, help=("Build wheels into , where the default is the " "current working directory."), ) cmd_opts.add_option(cmdoptions.no_binary()) cmd_opts.add_option(cmdoptions.only_binary()) cmd_opts.add_option(cmdoptions.prefer_binary()) cmd_opts.add_option( '--build-option', dest='build_options', metavar='options', action='append', help="Extra arguments to be supplied to 'setup.py bdist_wheel'.", ) cmd_opts.add_option(cmdoptions.no_build_isolation()) cmd_opts.add_option(cmdoptions.use_pep517()) cmd_opts.add_option(cmdoptions.no_use_pep517()) cmd_opts.add_option(cmdoptions.constraints()) cmd_opts.add_option(cmdoptions.editable()) cmd_opts.add_option(cmdoptions.requirements()) cmd_opts.add_option(cmdoptions.src()) cmd_opts.add_option(cmdoptions.ignore_requires_python()) cmd_opts.add_option(cmdoptions.no_deps()) cmd_opts.add_option(cmdoptions.build_dir()) cmd_opts.add_option(cmdoptions.progress_bar()) cmd_opts.add_option( '--global-option', dest='global_options', action='append', metavar='options', help="Extra global options to be supplied to the setup.py " "call before the 'bdist_wheel' command.") cmd_opts.add_option( '--pre', action='store_true', default=False, help=("Include pre-release and development versions. By default, " "pip only finds stable versions."), ) cmd_opts.add_option(cmdoptions.no_clean()) cmd_opts.add_option(cmdoptions.require_hashes()) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, self.parser, ) self.parser.insert_option_group(0, index_opts) self.parser.insert_option_group(0, cmd_opts) def run(self, options, args): # type: (Values, List[Any]) -> None cmdoptions.check_install_build_global(options) if options.build_dir: options.build_dir = os.path.abspath(options.build_dir) options.src_dir = os.path.abspath(options.src_dir) session = self.get_default_session(options) finder = self._build_package_finder(options, session) build_delete = (not (options.no_clean or options.build_dir)) wheel_cache = WheelCache(options.cache_dir, options.format_control) with RequirementTracker() as req_tracker, TempDirectory( options.build_dir, delete=build_delete, kind="wheel" ) as directory: requirement_set = RequirementSet( require_hashes=options.require_hashes, ) try: self.populate_requirement_set( requirement_set, args, options, finder, session, wheel_cache ) preparer = self.make_requirement_preparer( temp_build_dir=directory, options=options, req_tracker=req_tracker, wheel_download_dir=options.wheel_dir, ) resolver = self.make_resolver( preparer=preparer, finder=finder, session=session, options=options, wheel_cache=wheel_cache, ignore_requires_python=options.ignore_requires_python, use_pep517=options.use_pep517, ) resolver.resolve(requirement_set) # build wheels wb = WheelBuilder( preparer, wheel_cache, build_options=options.build_options or [], global_options=options.global_options or [], no_clean=options.no_clean, ) build_failures = wb.build( requirement_set.requirements.values(), ) if len(build_failures) != 0: raise CommandError( "Failed to build one or more wheels" ) except PreviousBuildDirError: options.no_clean = True raise finally: if not options.no_clean: requirement_set.cleanup_files() wheel_cache.cleanup() site-packages/pip/_internal/commands/download.py000064400000012711151464452710016024 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os from pip._internal.cli import cmdoptions from pip._internal.cli.cmdoptions import make_target_python from pip._internal.cli.req_command import RequirementCommand from pip._internal.req import RequirementSet from pip._internal.req.req_tracker import RequirementTracker from pip._internal.utils.filesystem import check_path_owner from pip._internal.utils.misc import ensure_dir, normalize_path, write_output from pip._internal.utils.temp_dir import TempDirectory logger = logging.getLogger(__name__) class DownloadCommand(RequirementCommand): """ Download packages from: - PyPI (and other indexes) using requirement specifiers. - VCS project urls. - Local project directories. - Local or remote source archives. pip also supports downloading from "requirements files", which provide an easy way to specify a whole environment to be downloaded. """ usage = """ %prog [options] [package-index-options] ... %prog [options] -r [package-index-options] ... %prog [options] ... %prog [options] ... %prog [options] ...""" def __init__(self, *args, **kw): super(DownloadCommand, self).__init__(*args, **kw) cmd_opts = self.cmd_opts cmd_opts.add_option(cmdoptions.constraints()) cmd_opts.add_option(cmdoptions.requirements()) cmd_opts.add_option(cmdoptions.build_dir()) cmd_opts.add_option(cmdoptions.no_deps()) cmd_opts.add_option(cmdoptions.global_options()) cmd_opts.add_option(cmdoptions.no_binary()) cmd_opts.add_option(cmdoptions.only_binary()) cmd_opts.add_option(cmdoptions.prefer_binary()) cmd_opts.add_option(cmdoptions.src()) cmd_opts.add_option(cmdoptions.pre()) cmd_opts.add_option(cmdoptions.no_clean()) cmd_opts.add_option(cmdoptions.require_hashes()) cmd_opts.add_option(cmdoptions.progress_bar()) cmd_opts.add_option(cmdoptions.no_build_isolation()) cmd_opts.add_option(cmdoptions.use_pep517()) cmd_opts.add_option(cmdoptions.no_use_pep517()) cmd_opts.add_option( '-d', '--dest', '--destination-dir', '--destination-directory', dest='download_dir', metavar='dir', default=os.curdir, help=("Download packages into ."), ) cmdoptions.add_target_python_options(cmd_opts) index_opts = cmdoptions.make_option_group( cmdoptions.index_group, self.parser, ) self.parser.insert_option_group(0, index_opts) self.parser.insert_option_group(0, cmd_opts) def run(self, options, args): options.ignore_installed = True # editable doesn't really make sense for `pip download`, but the bowels # of the RequirementSet code require that property. options.editables = [] cmdoptions.check_dist_restriction(options) options.src_dir = os.path.abspath(options.src_dir) options.download_dir = normalize_path(options.download_dir) ensure_dir(options.download_dir) session = self.get_default_session(options) target_python = make_target_python(options) finder = self._build_package_finder( options=options, session=session, target_python=target_python, ) build_delete = (not (options.no_clean or options.build_dir)) if options.cache_dir and not check_path_owner(options.cache_dir): logger.warning( "The directory '%s' or its parent directory is not owned " "by the current user and caching wheels has been " "disabled. check the permissions and owner of that " "directory. If executing pip with sudo, you may want " "sudo's -H flag.", options.cache_dir, ) options.cache_dir = None with RequirementTracker() as req_tracker, TempDirectory( options.build_dir, delete=build_delete, kind="download" ) as directory: requirement_set = RequirementSet( require_hashes=options.require_hashes, ) self.populate_requirement_set( requirement_set, args, options, finder, session, None ) preparer = self.make_requirement_preparer( temp_build_dir=directory, options=options, req_tracker=req_tracker, download_dir=options.download_dir, ) resolver = self.make_resolver( preparer=preparer, finder=finder, session=session, options=options, py_version_info=options.python_version, ) resolver.resolve(requirement_set) downloaded = ' '.join([ req.name for req in requirement_set.successfully_downloaded ]) if downloaded: write_output('Successfully downloaded %s', downloaded) # Clean up if not options.no_clean: requirement_set.cleanup_files() return requirement_set site-packages/pip/_internal/commands/search.py000064400000012034151464452710015460 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import sys import textwrap from collections import OrderedDict from pip._vendor import pkg_resources from pip._vendor.packaging.version import parse as parse_version # NOTE: XMLRPC Client is not annotated in typeshed as on 2017-07-17, which is # why we ignore the type on this import from pip._vendor.six.moves import xmlrpc_client # type: ignore from pip._internal.cli.base_command import Command from pip._internal.cli.req_command import SessionCommandMixin from pip._internal.cli.status_codes import NO_MATCHES_FOUND, SUCCESS from pip._internal.exceptions import CommandError from pip._internal.models.index import PyPI from pip._internal.network.xmlrpc import PipXmlrpcTransport from pip._internal.utils.compat import get_terminal_size from pip._internal.utils.logging import indent_log from pip._internal.utils.misc import write_output logger = logging.getLogger(__name__) class SearchCommand(Command, SessionCommandMixin): """Search for PyPI packages whose name or summary contains .""" usage = """ %prog [options] """ ignore_require_venv = True def __init__(self, *args, **kw): super(SearchCommand, self).__init__(*args, **kw) self.cmd_opts.add_option( '-i', '--index', dest='index', metavar='URL', default=PyPI.pypi_url, help='Base URL of Python Package Index (default %default)') self.parser.insert_option_group(0, self.cmd_opts) def run(self, options, args): if not args: raise CommandError('Missing required argument (search query).') query = args pypi_hits = self.search(query, options) hits = transform_hits(pypi_hits) terminal_width = None if sys.stdout.isatty(): terminal_width = get_terminal_size()[0] print_results(hits, terminal_width=terminal_width) if pypi_hits: return SUCCESS return NO_MATCHES_FOUND def search(self, query, options): index_url = options.index session = self.get_default_session(options) transport = PipXmlrpcTransport(index_url, session) pypi = xmlrpc_client.ServerProxy(index_url, transport) hits = pypi.search({'name': query, 'summary': query}, 'or') return hits def transform_hits(hits): """ The list from pypi is really a list of versions. We want a list of packages with the list of versions stored inline. This converts the list from pypi into one we can use. """ packages = OrderedDict() for hit in hits: name = hit['name'] summary = hit['summary'] version = hit['version'] if name not in packages.keys(): packages[name] = { 'name': name, 'summary': summary, 'versions': [version], } else: packages[name]['versions'].append(version) # if this is the highest version, replace summary and score if version == highest_version(packages[name]['versions']): packages[name]['summary'] = summary return list(packages.values()) def print_results(hits, name_column_width=None, terminal_width=None): if not hits: return if name_column_width is None: name_column_width = max([ len(hit['name']) + len(highest_version(hit.get('versions', ['-']))) for hit in hits ]) + 4 installed_packages = [p.project_name for p in pkg_resources.working_set] for hit in hits: name = hit['name'] summary = hit['summary'] or '' latest = highest_version(hit.get('versions', ['-'])) if terminal_width is not None: target_width = terminal_width - name_column_width - 5 if target_width > 10: # wrap and indent summary to fit terminal summary = textwrap.wrap(summary, target_width) summary = ('\n' + ' ' * (name_column_width + 3)).join(summary) line = '%-*s - %s' % (name_column_width, '%s (%s)' % (name, latest), summary) try: write_output(line) if name in installed_packages: dist = pkg_resources.get_distribution(name) with indent_log(): if dist.version == latest: write_output('INSTALLED: %s (latest)', dist.version) else: write_output('INSTALLED: %s', dist.version) if parse_version(latest).pre: write_output('LATEST: %s (pre-release; install' ' with "pip install --pre")', latest) else: write_output('LATEST: %s', latest) except UnicodeEncodeError: pass def highest_version(versions): return max(versions, key=parse_version) site-packages/pip/_internal/commands/completion.py000064400000005637151464452710016377 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import sys import textwrap from pip._internal.cli.base_command import Command from pip._internal.utils.misc import get_prog BASE_COMPLETION = """ # pip %(shell)s completion start%(script)s# pip %(shell)s completion end """ COMPLETION_SCRIPTS = { 'bash': """ _pip_completion() { COMPREPLY=( $( COMP_WORDS="${COMP_WORDS[*]}" \\ COMP_CWORD=$COMP_CWORD \\ PIP_AUTO_COMPLETE=1 $1 2>/dev/null ) ) } complete -o default -F _pip_completion %(prog)s """, 'zsh': """ function _pip_completion { local words cword read -Ac words read -cn cword reply=( $( COMP_WORDS="$words[*]" \\ COMP_CWORD=$(( cword-1 )) \\ PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null )) } compctl -K _pip_completion %(prog)s """, 'fish': """ function __fish_complete_pip set -lx COMP_WORDS (commandline -o) "" set -lx COMP_CWORD ( \\ math (contains -i -- (commandline -t) $COMP_WORDS)-1 \\ ) set -lx PIP_AUTO_COMPLETE 1 string split \\ -- (eval $COMP_WORDS[1]) end complete -fa "(__fish_complete_pip)" -c %(prog)s """, } class CompletionCommand(Command): """A helper command to be used for command completion.""" ignore_require_venv = True def __init__(self, *args, **kw): super(CompletionCommand, self).__init__(*args, **kw) cmd_opts = self.cmd_opts cmd_opts.add_option( '--bash', '-b', action='store_const', const='bash', dest='shell', help='Emit completion code for bash') cmd_opts.add_option( '--zsh', '-z', action='store_const', const='zsh', dest='shell', help='Emit completion code for zsh') cmd_opts.add_option( '--fish', '-f', action='store_const', const='fish', dest='shell', help='Emit completion code for fish') self.parser.insert_option_group(0, cmd_opts) def run(self, options, args): """Prints the completion code of the given shell""" shells = COMPLETION_SCRIPTS.keys() shell_options = ['--' + shell for shell in sorted(shells)] if options.shell in shells: script = textwrap.dedent( COMPLETION_SCRIPTS.get(options.shell, '') % { 'prog': get_prog(), } ) print(BASE_COMPLETION % {'script': script, 'shell': options.shell}) else: sys.stderr.write( 'ERROR: You must pass %s\n' % ' or '.join(shell_options) ) site-packages/pip/_internal/commands/help.py000064400000002235151464452710015145 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import from pip._internal.cli.base_command import Command from pip._internal.cli.status_codes import SUCCESS from pip._internal.exceptions import CommandError class HelpCommand(Command): """Show help for commands""" usage = """ %prog """ ignore_require_venv = True def run(self, options, args): from pip._internal.commands import ( commands_dict, create_command, get_similar_commands, ) try: # 'pip help' with no args is handled by pip.__init__.parseopt() cmd_name = args[0] # the command we need help for except IndexError: return SUCCESS if cmd_name not in commands_dict: guess = get_similar_commands(cmd_name) msg = ['unknown command "%s"' % cmd_name] if guess: msg.append('maybe you meant "%s"' % guess) raise CommandError(' - '.join(msg)) command = create_command(cmd_name) command.parser.print_help() return SUCCESS site-packages/pip/_internal/commands/debug.py000064400000006502151464452710015304 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import locale import logging import sys from pip._internal.cli import cmdoptions from pip._internal.cli.base_command import Command from pip._internal.cli.cmdoptions import make_target_python from pip._internal.cli.status_codes import SUCCESS from pip._internal.utils.logging import indent_log from pip._internal.utils.misc import get_pip_version from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.wheel import format_tag if MYPY_CHECK_RUNNING: from typing import Any, List from optparse import Values logger = logging.getLogger(__name__) def show_value(name, value): # type: (str, str) -> None logger.info('{}: {}'.format(name, value)) def show_sys_implementation(): # type: () -> None logger.info('sys.implementation:') if hasattr(sys, 'implementation'): implementation = sys.implementation # type: ignore implementation_name = implementation.name else: implementation_name = '' with indent_log(): show_value('name', implementation_name) def show_tags(options): # type: (Values) -> None tag_limit = 10 target_python = make_target_python(options) tags = target_python.get_tags() # Display the target options that were explicitly provided. formatted_target = target_python.format_given() suffix = '' if formatted_target: suffix = ' (target: {})'.format(formatted_target) msg = 'Compatible tags: {}{}'.format(len(tags), suffix) logger.info(msg) if options.verbose < 1 and len(tags) > tag_limit: tags_limited = True tags = tags[:tag_limit] else: tags_limited = False with indent_log(): for tag in tags: logger.info(format_tag(tag)) if tags_limited: msg = ( '...\n' '[First {tag_limit} tags shown. Pass --verbose to show all.]' ).format(tag_limit=tag_limit) logger.info(msg) class DebugCommand(Command): """ Display debug information. """ usage = """ %prog """ ignore_require_venv = True def __init__(self, *args, **kw): super(DebugCommand, self).__init__(*args, **kw) cmd_opts = self.cmd_opts cmdoptions.add_target_python_options(cmd_opts) self.parser.insert_option_group(0, cmd_opts) def run(self, options, args): # type: (Values, List[Any]) -> int logger.warning( "This command is only meant for debugging. " "Do not use this with automation for parsing and getting these " "details, since the output and options of this command may " "change without notice." ) show_value('pip version', get_pip_version()) show_value('sys.version', sys.version) show_value('sys.executable', sys.executable) show_value('sys.getdefaultencoding', sys.getdefaultencoding()) show_value('sys.getfilesystemencoding', sys.getfilesystemencoding()) show_value( 'locale.getpreferredencoding', locale.getpreferredencoding(), ) show_value('sys.platform', sys.platform) show_sys_implementation() show_tags(options) return SUCCESS site-packages/pip/_internal/commands/__init__.py000064400000007202151464452710015753 0ustar00""" Package containing all pip commands """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import importlib from collections import OrderedDict, namedtuple from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Any from pip._internal.cli.base_command import Command CommandInfo = namedtuple('CommandInfo', 'module_path, class_name, summary') # The ordering matters for help display. # Also, even though the module path starts with the same # "pip._internal.commands" prefix in each case, we include the full path # because it makes testing easier (specifically when modifying commands_dict # in test setup / teardown by adding info for a FakeCommand class defined # in a test-related module). # Finally, we need to pass an iterable of pairs here rather than a dict # so that the ordering won't be lost when using Python 2.7. commands_dict = OrderedDict([ ('install', CommandInfo( 'pip._internal.commands.install', 'InstallCommand', 'Install packages.', )), ('download', CommandInfo( 'pip._internal.commands.download', 'DownloadCommand', 'Download packages.', )), ('uninstall', CommandInfo( 'pip._internal.commands.uninstall', 'UninstallCommand', 'Uninstall packages.', )), ('freeze', CommandInfo( 'pip._internal.commands.freeze', 'FreezeCommand', 'Output installed packages in requirements format.', )), ('list', CommandInfo( 'pip._internal.commands.list', 'ListCommand', 'List installed packages.', )), ('show', CommandInfo( 'pip._internal.commands.show', 'ShowCommand', 'Show information about installed packages.', )), ('check', CommandInfo( 'pip._internal.commands.check', 'CheckCommand', 'Verify installed packages have compatible dependencies.', )), ('config', CommandInfo( 'pip._internal.commands.configuration', 'ConfigurationCommand', 'Manage local and global configuration.', )), ('search', CommandInfo( 'pip._internal.commands.search', 'SearchCommand', 'Search PyPI for packages.', )), ('wheel', CommandInfo( 'pip._internal.commands.wheel', 'WheelCommand', 'Build wheels from your requirements.', )), ('hash', CommandInfo( 'pip._internal.commands.hash', 'HashCommand', 'Compute hashes of package archives.', )), ('completion', CommandInfo( 'pip._internal.commands.completion', 'CompletionCommand', 'A helper command used for command completion.', )), ('debug', CommandInfo( 'pip._internal.commands.debug', 'DebugCommand', 'Show information useful for debugging.', )), ('help', CommandInfo( 'pip._internal.commands.help', 'HelpCommand', 'Show help for commands.', )), ]) # type: OrderedDict[str, CommandInfo] def create_command(name, **kwargs): # type: (str, **Any) -> Command """ Create an instance of the Command class with the given name. """ module_path, class_name, summary = commands_dict[name] module = importlib.import_module(module_path) command_class = getattr(module, class_name) command = command_class(name=name, summary=summary, **kwargs) return command def get_similar_commands(name): """Command name auto-correct.""" from difflib import get_close_matches name = name.lower() close_commands = get_close_matches(name, commands_dict.keys()) if close_commands: return close_commands[0] else: return False site-packages/pip/_internal/utils/models.py000064400000002174151464452710015041 0ustar00"""Utilities for defining models """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import operator class KeyBasedCompareMixin(object): """Provides comparison capabilities that is based on a key """ def __init__(self, key, defining_class): self._compare_key = key self._defining_class = defining_class def __hash__(self): return hash(self._compare_key) def __lt__(self, other): return self._compare(other, operator.__lt__) def __le__(self, other): return self._compare(other, operator.__le__) def __gt__(self, other): return self._compare(other, operator.__gt__) def __ge__(self, other): return self._compare(other, operator.__ge__) def __eq__(self, other): return self._compare(other, operator.__eq__) def __ne__(self, other): return self._compare(other, operator.__ne__) def _compare(self, other, method): if not isinstance(other, self._defining_class): return NotImplemented return method(self._compare_key, other._compare_key) site-packages/pip/_internal/utils/glibc.py000064400000010455151464452710014637 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False from __future__ import absolute_import import os import re import warnings from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional, Tuple def glibc_version_string(): # type: () -> Optional[str] "Returns glibc version string, or None if not using glibc." return glibc_version_string_confstr() or glibc_version_string_ctypes() def glibc_version_string_confstr(): # type: () -> Optional[str] "Primary implementation of glibc_version_string using os.confstr." # os.confstr is quite a bit faster than ctypes.DLL. It's also less likely # to be broken or missing. This strategy is used in the standard library # platform module: # https://github.com/python/cpython/blob/fcf1d003bf4f0100c9d0921ff3d70e1127ca1b71/Lib/platform.py#L175-L183 try: # os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17": _, version = os.confstr("CS_GNU_LIBC_VERSION").split() except (AttributeError, OSError, ValueError): # os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)... return None return version def glibc_version_string_ctypes(): # type: () -> Optional[str] "Fallback implementation of glibc_version_string using ctypes." try: import ctypes except ImportError: return None # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen # manpage says, "If filename is NULL, then the returned handle is for the # main program". This way we can let the linker do the work to figure out # which libc our process is actually using. process_namespace = ctypes.CDLL(None) try: gnu_get_libc_version = process_namespace.gnu_get_libc_version except AttributeError: # Symbol doesn't exist -> therefore, we are not linked to # glibc. return None # Call gnu_get_libc_version, which returns a string like "2.5" gnu_get_libc_version.restype = ctypes.c_char_p version_str = gnu_get_libc_version() # py2 / py3 compatibility: if not isinstance(version_str, str): version_str = version_str.decode("ascii") return version_str # Separated out from have_compatible_glibc for easier unit testing def check_glibc_version(version_str, required_major, minimum_minor): # type: (str, int, int) -> bool # Parse string and check against requested version. # # We use a regexp instead of str.split because we want to discard any # random junk that might come after the minor version -- this might happen # in patched/forked versions of glibc (e.g. Linaro's version of glibc # uses version strings like "2.20-2014.11"). See gh-3588. m = re.match(r"(?P[0-9]+)\.(?P[0-9]+)", version_str) if not m: warnings.warn("Expected glibc version with 2 components major.minor," " got: %s" % version_str, RuntimeWarning) return False return (int(m.group("major")) == required_major and int(m.group("minor")) >= minimum_minor) def have_compatible_glibc(required_major, minimum_minor): # type: (int, int) -> bool version_str = glibc_version_string() if version_str is None: return False return check_glibc_version(version_str, required_major, minimum_minor) # platform.libc_ver regularly returns completely nonsensical glibc # versions. E.g. on my computer, platform says: # # ~$ python2.7 -c 'import platform; print(platform.libc_ver())' # ('glibc', '2.7') # ~$ python3.5 -c 'import platform; print(platform.libc_ver())' # ('glibc', '2.9') # # But the truth is: # # ~$ ldd --version # ldd (Debian GLIBC 2.22-11) 2.22 # # This is unfortunate, because it means that the linehaul data on libc # versions that was generated by pip 8.1.2 and earlier is useless and # misleading. Solution: instead of using platform, use our code that actually # works. def libc_ver(): # type: () -> Tuple[str, str] """Try to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings in case the lookup fails. """ glibc_version = glibc_version_string() if glibc_version is None: return ("", "") else: return ("glibc", glibc_version) site-packages/pip/_internal/utils/setuptools_build.py000064400000003137151464452710017156 0ustar00import sys from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import List, Sequence # Shim to wrap setup.py invocation with setuptools # # We set sys.argv[0] to the path to the underlying setup.py file so # setuptools / distutils don't take the path to the setup.py to be "-c" when # invoking via the shim. This avoids e.g. the following manifest_maker # warning: "warning: manifest_maker: standard file '-c' not found". _SETUPTOOLS_SHIM = ( "import sys, setuptools, tokenize; sys.argv[0] = {0!r}; __file__={0!r};" "f=getattr(tokenize, 'open', open)(__file__);" "code=f.read().replace('\\r\\n', '\\n');" "f.close();" "exec(compile(code, __file__, 'exec'))" ) def make_setuptools_shim_args( setup_py_path, # type: str global_options=None, # type: Sequence[str] no_user_config=False, # type: bool unbuffered_output=False # type: bool ): # type: (...) -> List[str] """ Get setuptools command arguments with shim wrapped setup file invocation. :param setup_py_path: The path to setup.py to be wrapped. :param global_options: Additional global options. :param no_user_config: If True, disables personal user configuration. :param unbuffered_output: If True, adds the unbuffered switch to the argument list. """ args = [sys.executable] if unbuffered_output: args.append('-u') args.extend(['-c', _SETUPTOOLS_SHIM.format(setup_py_path)]) if global_options: args.extend(global_options) if no_user_config: args.append('--no-user-cfg') return args site-packages/pip/_internal/utils/deprecation.py000064400000006366151464452710016062 0ustar00""" A module that implements tooling to enable easy warnings about deprecations. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import warnings from pip._vendor.packaging.version import parse from pip import __version__ as current_version from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Any, Optional DEPRECATION_MSG_PREFIX = "DEPRECATION: " class PipDeprecationWarning(Warning): pass _original_showwarning = None # type: Any # Warnings <-> Logging Integration def _showwarning(message, category, filename, lineno, file=None, line=None): if file is not None: if _original_showwarning is not None: _original_showwarning( message, category, filename, lineno, file, line, ) elif issubclass(category, PipDeprecationWarning): # We use a specially named logger which will handle all of the # deprecation messages for pip. logger = logging.getLogger("pip._internal.deprecations") logger.warning(message) else: _original_showwarning( message, category, filename, lineno, file, line, ) def install_warning_logger(): # type: () -> None # Enable our Deprecation Warnings warnings.simplefilter("default", PipDeprecationWarning, append=True) global _original_showwarning if _original_showwarning is None: _original_showwarning = warnings.showwarning warnings.showwarning = _showwarning def deprecated(reason, replacement, gone_in, issue=None): # type: (str, Optional[str], Optional[str], Optional[int]) -> None """Helper to deprecate existing functionality. reason: Textual reason shown to the user about why this functionality has been deprecated. replacement: Textual suggestion shown to the user about what alternative functionality they can use. gone_in: The version of pip does this functionality should get removed in. Raises errors if pip's current version is greater than or equal to this. issue: Issue number on the tracker that would serve as a useful place for users to find related discussion and provide feedback. Always pass replacement, gone_in and issue as keyword arguments for clarity at the call site. """ # Construct a nice message. # This is eagerly formatted as we want it to get logged as if someone # typed this entire message out. sentences = [ (reason, DEPRECATION_MSG_PREFIX + "{}"), (gone_in, "pip {} will remove support for this functionality."), (replacement, "A possible replacement is {}."), (issue, ( "You can find discussion regarding this at " "https://github.com/pypa/pip/issues/{}." )), ] message = " ".join( template.format(val) for val, template in sentences if val is not None ) # Raise as an error if it has to be removed. if gone_in is not None and parse(current_version) >= parse(gone_in): raise PipDeprecationWarning(message) warnings.warn(message, category=PipDeprecationWarning, stacklevel=2) site-packages/pip/_internal/utils/temp_dir.py000064400000012621151464452710015357 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import errno import itertools import logging import os.path import tempfile from pip._internal.utils.misc import rmtree from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional logger = logging.getLogger(__name__) class TempDirectory(object): """Helper class that owns and cleans up a temporary directory. This class can be used as a context manager or as an OO representation of a temporary directory. Attributes: path Location to the created temporary directory delete Whether the directory should be deleted when exiting (when used as a contextmanager) Methods: cleanup() Deletes the temporary directory When used as a context manager, if the delete attribute is True, on exiting the context the temporary directory is deleted. """ def __init__( self, path=None, # type: Optional[str] delete=None, # type: Optional[bool] kind="temp" ): super(TempDirectory, self).__init__() if path is None and delete is None: # If we were not given an explicit directory, and we were not given # an explicit delete option, then we'll default to deleting. delete = True if path is None: path = self._create(kind) self._path = path self._deleted = False self.delete = delete self.kind = kind @property def path(self): # type: () -> str assert not self._deleted, ( "Attempted to access deleted path: {}".format(self._path) ) return self._path def __repr__(self): return "<{} {!r}>".format(self.__class__.__name__, self.path) def __enter__(self): return self def __exit__(self, exc, value, tb): if self.delete: self.cleanup() def _create(self, kind): """Create a temporary directory and store its path in self.path """ # We realpath here because some systems have their default tmpdir # symlinked to another directory. This tends to confuse build # scripts, so we canonicalize the path by traversing potential # symlinks here. path = os.path.realpath( tempfile.mkdtemp(prefix="pip-{}-".format(kind)) ) logger.debug("Created temporary directory: {}".format(path)) return path def cleanup(self): """Remove the temporary directory created and reset state """ self._deleted = True if os.path.exists(self._path): rmtree(self._path) class AdjacentTempDirectory(TempDirectory): """Helper class that creates a temporary directory adjacent to a real one. Attributes: original The original directory to create a temp directory for. path After calling create() or entering, contains the full path to the temporary directory. delete Whether the directory should be deleted when exiting (when used as a contextmanager) """ # The characters that may be used to name the temp directory # We always prepend a ~ and then rotate through these until # a usable name is found. # pkg_resources raises a different error for .dist-info folder # with leading '-' and invalid metadata LEADING_CHARS = "-~.=%0123456789" def __init__(self, original, delete=None): self.original = original.rstrip('/\\') super(AdjacentTempDirectory, self).__init__(delete=delete) @classmethod def _generate_names(cls, name): """Generates a series of temporary names. The algorithm replaces the leading characters in the name with ones that are valid filesystem characters, but are not valid package names (for both Python and pip definitions of package). """ for i in range(1, len(name)): for candidate in itertools.combinations_with_replacement( cls.LEADING_CHARS, i - 1): new_name = '~' + ''.join(candidate) + name[i:] if new_name != name: yield new_name # If we make it this far, we will have to make a longer name for i in range(len(cls.LEADING_CHARS)): for candidate in itertools.combinations_with_replacement( cls.LEADING_CHARS, i): new_name = '~' + ''.join(candidate) + name if new_name != name: yield new_name def _create(self, kind): root, name = os.path.split(self.original) for candidate in self._generate_names(name): path = os.path.join(root, candidate) try: os.mkdir(path) except OSError as ex: # Continue if the name exists already if ex.errno != errno.EEXIST: raise else: path = os.path.realpath(path) break else: # Final fallback on the default behavior. path = os.path.realpath( tempfile.mkdtemp(prefix="pip-{}-".format(kind)) ) logger.debug("Created temporary directory: {}".format(path)) return path site-packages/pip/_internal/utils/misc.py000064400000061753151464452710014521 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import contextlib import errno import getpass import io import logging import os import posixpath import shutil import stat import sys from collections import deque from pip._vendor import pkg_resources # NOTE: retrying is not annotated in typeshed as on 2017-07-17, which is # why we ignore the type on this import. from pip._vendor.retrying import retry # type: ignore from pip._vendor.six import PY2, text_type from pip._vendor.six.moves import input from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._vendor.six.moves.urllib.parse import unquote as urllib_unquote from pip import __version__ from pip._internal.exceptions import CommandError from pip._internal.locations import ( distutils_scheme, get_major_minor_version, site_packages, user_site, ) from pip._internal.utils.compat import ( WINDOWS, expanduser, stdlib_pkgs, str_to_display, ) from pip._internal.utils.marker_files import write_delete_marker_file from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.virtualenv import ( running_under_virtualenv, virtualenv_no_global, ) if PY2: from io import BytesIO as StringIO else: from io import StringIO if MYPY_CHECK_RUNNING: from typing import ( Any, AnyStr, Container, Iterable, List, Optional, Text, Tuple, Union, cast, ) from pip._vendor.pkg_resources import Distribution VersionInfo = Tuple[int, int, int] else: # typing's cast() is needed at runtime, but we don't want to import typing. # Thus, we use a dummy no-op version, which we tell mypy to ignore. def cast(type_, value): # type: ignore return value __all__ = ['rmtree', 'display_path', 'backup_dir', 'ask', 'splitext', 'format_size', 'is_installable_dir', 'normalize_path', 'renames', 'get_prog', 'captured_stdout', 'ensure_dir', 'get_installed_version', 'remove_auth_from_url'] logger = logging.getLogger(__name__) def get_pip_version(): # type: () -> str pip_pkg_dir = os.path.join(os.path.dirname(__file__), "..", "..") pip_pkg_dir = os.path.abspath(pip_pkg_dir) return ( 'pip {} from {} (python {})'.format( __version__, pip_pkg_dir, get_major_minor_version(), ) ) def normalize_version_info(py_version_info): # type: (Tuple[int, ...]) -> Tuple[int, int, int] """ Convert a tuple of ints representing a Python version to one of length three. :param py_version_info: a tuple of ints representing a Python version, or None to specify no version. The tuple can have any length. :return: a tuple of length three if `py_version_info` is non-None. Otherwise, return `py_version_info` unchanged (i.e. None). """ if len(py_version_info) < 3: py_version_info += (3 - len(py_version_info)) * (0,) elif len(py_version_info) > 3: py_version_info = py_version_info[:3] return cast('VersionInfo', py_version_info) def ensure_dir(path): # type: (AnyStr) -> None """os.path.makedirs without EEXIST.""" try: os.makedirs(path) except OSError as e: if e.errno != errno.EEXIST: raise def get_prog(): # type: () -> str try: prog = os.path.basename(sys.argv[0]) if prog in ('__main__.py', '-c'): return "%s -m pip" % sys.executable else: return prog except (AttributeError, TypeError, IndexError): pass return 'pip' # Retry every half second for up to 3 seconds @retry(stop_max_delay=3000, wait_fixed=500) def rmtree(dir, ignore_errors=False): # type: (str, bool) -> None shutil.rmtree(dir, ignore_errors=ignore_errors, onerror=rmtree_errorhandler) def rmtree_errorhandler(func, path, exc_info): """On Windows, the files in .svn are read-only, so when rmtree() tries to remove them, an exception is thrown. We catch that here, remove the read-only attribute, and hopefully continue without problems.""" try: has_attr_readonly = not (os.stat(path).st_mode & stat.S_IWRITE) except (IOError, OSError): # it's equivalent to os.path.exists return if has_attr_readonly: # convert to read/write os.chmod(path, stat.S_IWRITE) # use the original function to repeat the operation func(path) return else: raise def path_to_display(path): # type: (Optional[Union[str, Text]]) -> Optional[Text] """ Convert a bytes (or text) path to text (unicode in Python 2) for display and logging purposes. This function should never error out. Also, this function is mainly needed for Python 2 since in Python 3 str paths are already text. """ if path is None: return None if isinstance(path, text_type): return path # Otherwise, path is a bytes object (str in Python 2). try: display_path = path.decode(sys.getfilesystemencoding(), 'strict') except UnicodeDecodeError: # Include the full bytes to make troubleshooting easier, even though # it may not be very human readable. if PY2: # Convert the bytes to a readable str representation using # repr(), and then convert the str to unicode. # Also, we add the prefix "b" to the repr() return value both # to make the Python 2 output look like the Python 3 output, and # to signal to the user that this is a bytes representation. display_path = str_to_display('b{!r}'.format(path)) else: # Silence the "F821 undefined name 'ascii'" flake8 error since # in Python 3 ascii() is a built-in. display_path = ascii(path) # noqa: F821 return display_path def display_path(path): # type: (Union[str, Text]) -> str """Gives the display value for a given path, making it relative to cwd if possible.""" path = os.path.normcase(os.path.abspath(path)) if sys.version_info[0] == 2: path = path.decode(sys.getfilesystemencoding(), 'replace') path = path.encode(sys.getdefaultencoding(), 'replace') if path.startswith(os.getcwd() + os.path.sep): path = '.' + path[len(os.getcwd()):] return path def backup_dir(dir, ext='.bak'): # type: (str, str) -> str """Figure out the name of a directory to back up the given dir to (adding .bak, .bak2, etc)""" n = 1 extension = ext while os.path.exists(dir + extension): n += 1 extension = ext + str(n) return dir + extension def ask_path_exists(message, options): # type: (str, Iterable[str]) -> str for action in os.environ.get('PIP_EXISTS_ACTION', '').split(): if action in options: return action return ask(message, options) def _check_no_input(message): # type: (str) -> None """Raise an error if no input is allowed.""" if os.environ.get('PIP_NO_INPUT'): raise Exception( 'No input was expected ($PIP_NO_INPUT set); question: %s' % message ) def ask(message, options): # type: (str, Iterable[str]) -> str """Ask the message interactively, with the given possible responses""" while 1: _check_no_input(message) response = input(message) response = response.strip().lower() if response not in options: print( 'Your response (%r) was not one of the expected responses: ' '%s' % (response, ', '.join(options)) ) else: return response def ask_input(message): # type: (str) -> str """Ask for input interactively.""" _check_no_input(message) return input(message) def ask_password(message): # type: (str) -> str """Ask for a password interactively.""" _check_no_input(message) return getpass.getpass(message) def format_size(bytes): # type: (float) -> str if bytes > 1000 * 1000: return '%.1fMB' % (bytes / 1000.0 / 1000) elif bytes > 10 * 1000: return '%ikB' % (bytes / 1000) elif bytes > 1000: return '%.1fkB' % (bytes / 1000.0) else: return '%ibytes' % bytes def is_installable_dir(path): # type: (str) -> bool """Is path is a directory containing setup.py or pyproject.toml? """ if not os.path.isdir(path): return False setup_py = os.path.join(path, 'setup.py') if os.path.isfile(setup_py): return True pyproject_toml = os.path.join(path, 'pyproject.toml') if os.path.isfile(pyproject_toml): return True return False def read_chunks(file, size=io.DEFAULT_BUFFER_SIZE): """Yield pieces of data from a file-like object until EOF.""" while True: chunk = file.read(size) if not chunk: break yield chunk def normalize_path(path, resolve_symlinks=True): # type: (str, bool) -> str """ Convert a path to its canonical, case-normalized, absolute version. """ path = expanduser(path) if resolve_symlinks: path = os.path.realpath(path) else: path = os.path.abspath(path) return os.path.normcase(path) def splitext(path): # type: (str) -> Tuple[str, str] """Like os.path.splitext, but take off .tar too""" base, ext = posixpath.splitext(path) if base.lower().endswith('.tar'): ext = base[-4:] + ext base = base[:-4] return base, ext def renames(old, new): # type: (str, str) -> None """Like os.renames(), but handles renaming across devices.""" # Implementation borrowed from os.renames(). head, tail = os.path.split(new) if head and tail and not os.path.exists(head): os.makedirs(head) shutil.move(old, new) head, tail = os.path.split(old) if head and tail: try: os.removedirs(head) except OSError: pass def is_local(path): # type: (str) -> bool """ Return True if path is within sys.prefix, if we're running in a virtualenv. If we're not in a virtualenv, all paths are considered "local." Caution: this function assumes the head of path has been normalized with normalize_path. """ if not running_under_virtualenv(): return True return path.startswith(normalize_path(sys.prefix)) def dist_is_local(dist): # type: (Distribution) -> bool """ Return True if given Distribution object is installed locally (i.e. within current virtualenv). Always True if we're not in a virtualenv. """ return is_local(dist_location(dist)) def dist_in_usersite(dist): # type: (Distribution) -> bool """ Return True if given Distribution is installed in user site. """ return dist_location(dist).startswith(normalize_path(user_site)) def dist_in_site_packages(dist): # type: (Distribution) -> bool """ Return True if given Distribution is installed in sysconfig.get_python_lib(). """ return dist_location(dist).startswith(normalize_path(site_packages)) def dist_in_install_path(dist): """ Return True if given Distribution is installed in path matching distutils_scheme layout. """ norm_path = normalize_path(dist_location(dist)) return norm_path.startswith(normalize_path( distutils_scheme("")['purelib'].split('python')[0])) def dist_is_editable(dist): # type: (Distribution) -> bool """ Return True if given Distribution is an editable install. """ for path_item in sys.path: egg_link = os.path.join(path_item, dist.project_name + '.egg-link') if os.path.isfile(egg_link): return True return False def get_installed_distributions( local_only=True, # type: bool skip=stdlib_pkgs, # type: Container[str] include_editables=True, # type: bool editables_only=False, # type: bool user_only=False, # type: bool paths=None # type: Optional[List[str]] ): # type: (...) -> List[Distribution] """ Return a list of installed Distribution objects. If ``local_only`` is True (default), only return installations local to the current virtualenv, if in a virtualenv. ``skip`` argument is an iterable of lower-case project names to ignore; defaults to stdlib_pkgs If ``include_editables`` is False, don't report editables. If ``editables_only`` is True , only report editables. If ``user_only`` is True , only report installations in the user site directory. If ``paths`` is set, only report the distributions present at the specified list of locations. """ if paths: working_set = pkg_resources.WorkingSet(paths) else: working_set = pkg_resources.working_set if local_only: local_test = dist_is_local else: def local_test(d): return True if include_editables: def editable_test(d): return True else: def editable_test(d): return not dist_is_editable(d) if editables_only: def editables_only_test(d): return dist_is_editable(d) else: def editables_only_test(d): return True if user_only: user_test = dist_in_usersite else: def user_test(d): return True # because of pkg_resources vendoring, mypy cannot find stub in typeshed return [d for d in working_set # type: ignore if local_test(d) and d.key not in skip and editable_test(d) and editables_only_test(d) and user_test(d) ] def egg_link_path(dist): # type: (Distribution) -> Optional[str] """ Return the path for the .egg-link file if it exists, otherwise, None. There's 3 scenarios: 1) not in a virtualenv try to find in site.USER_SITE, then site_packages 2) in a no-global virtualenv try to find in site_packages 3) in a yes-global virtualenv try to find in site_packages, then site.USER_SITE (don't look in global location) For #1 and #3, there could be odd cases, where there's an egg-link in 2 locations. This method will just return the first one found. """ sites = [] if running_under_virtualenv(): sites.append(site_packages) if not virtualenv_no_global() and user_site: sites.append(user_site) else: if user_site: sites.append(user_site) sites.append(site_packages) for site in sites: egglink = os.path.join(site, dist.project_name) + '.egg-link' if os.path.isfile(egglink): return egglink return None def dist_location(dist): # type: (Distribution) -> str """ Get the site-packages location of this distribution. Generally this is dist.location, except in the case of develop-installed packages, where dist.location is the source code location, and we want to know where the egg-link file is. The returned location is normalized (in particular, with symlinks removed). """ egg_link = egg_link_path(dist) if egg_link: return normalize_path(egg_link) return normalize_path(dist.location) def write_output(msg, *args): # type: (str, str) -> None logger.info(msg, *args) def _make_build_dir(build_dir): os.makedirs(build_dir) write_delete_marker_file(build_dir) class FakeFile(object): """Wrap a list of lines in an object with readline() to make ConfigParser happy.""" def __init__(self, lines): self._gen = (l for l in lines) def readline(self): try: try: return next(self._gen) except NameError: return self._gen.next() except StopIteration: return '' def __iter__(self): return self._gen class StreamWrapper(StringIO): @classmethod def from_stream(cls, orig_stream): cls.orig_stream = orig_stream return cls() # compileall.compile_dir() needs stdout.encoding to print to stdout @property def encoding(self): return self.orig_stream.encoding @contextlib.contextmanager def captured_output(stream_name): """Return a context manager used by captured_stdout/stdin/stderr that temporarily replaces the sys stream *stream_name* with a StringIO. Taken from Lib/support/__init__.py in the CPython repo. """ orig_stdout = getattr(sys, stream_name) setattr(sys, stream_name, StreamWrapper.from_stream(orig_stdout)) try: yield getattr(sys, stream_name) finally: setattr(sys, stream_name, orig_stdout) def captured_stdout(): """Capture the output of sys.stdout: with captured_stdout() as stdout: print('hello') self.assertEqual(stdout.getvalue(), 'hello\n') Taken from Lib/support/__init__.py in the CPython repo. """ return captured_output('stdout') def captured_stderr(): """ See captured_stdout(). """ return captured_output('stderr') class cached_property(object): """A property that is only computed once per instance and then replaces itself with an ordinary attribute. Deleting the attribute resets the property. Source: https://github.com/bottlepy/bottle/blob/0.11.5/bottle.py#L175 """ def __init__(self, func): self.__doc__ = getattr(func, '__doc__') self.func = func def __get__(self, obj, cls): if obj is None: # We're being accessed from the class itself, not from an object return self value = obj.__dict__[self.func.__name__] = self.func(obj) return value def get_installed_version(dist_name, working_set=None): """Get the installed version of dist_name avoiding pkg_resources cache""" # Create a requirement that we'll look for inside of setuptools. req = pkg_resources.Requirement.parse(dist_name) if working_set is None: # We want to avoid having this cached, so we need to construct a new # working set each time. working_set = pkg_resources.WorkingSet() # Get the installed distribution from our working set dist = working_set.find(req) # Check to see if we got an installed distribution or not, if we did # we want to return it's version. return dist.version if dist else None def consume(iterator): """Consume an iterable at C speed.""" deque(iterator, maxlen=0) # Simulates an enum def enum(*sequential, **named): enums = dict(zip(sequential, range(len(sequential))), **named) reverse = {value: key for key, value in enums.items()} enums['reverse_mapping'] = reverse return type('Enum', (), enums) def build_netloc(host, port): # type: (str, Optional[int]) -> str """ Build a netloc from a host-port pair """ if port is None: return host if ':' in host: # Only wrap host with square brackets when it is IPv6 host = '[{}]'.format(host) return '{}:{}'.format(host, port) def build_url_from_netloc(netloc, scheme='https'): # type: (str, str) -> str """ Build a full URL from a netloc. """ if netloc.count(':') >= 2 and '@' not in netloc and '[' not in netloc: # It must be a bare IPv6 address, so wrap it with brackets. netloc = '[{}]'.format(netloc) return '{}://{}'.format(scheme, netloc) def parse_netloc(netloc): # type: (str) -> Tuple[str, Optional[int]] """ Return the host-port pair from a netloc. """ url = build_url_from_netloc(netloc) parsed = urllib_parse.urlparse(url) return parsed.hostname, parsed.port def split_auth_from_netloc(netloc): """ Parse out and remove the auth information from a netloc. Returns: (netloc, (username, password)). """ if '@' not in netloc: return netloc, (None, None) # Split from the right because that's how urllib.parse.urlsplit() # behaves if more than one @ is present (which can be checked using # the password attribute of urlsplit()'s return value). auth, netloc = netloc.rsplit('@', 1) if ':' in auth: # Split from the left because that's how urllib.parse.urlsplit() # behaves if more than one : is present (which again can be checked # using the password attribute of the return value) user_pass = auth.split(':', 1) else: user_pass = auth, None user_pass = tuple( None if x is None else urllib_unquote(x) for x in user_pass ) return netloc, user_pass def redact_netloc(netloc): # type: (str) -> str """ Replace the sensitive data in a netloc with "****", if it exists. For example: - "user:pass@example.com" returns "user:****@example.com" - "accesstoken@example.com" returns "****@example.com" """ netloc, (user, password) = split_auth_from_netloc(netloc) if user is None: return netloc if password is None: user = '****' password = '' else: user = urllib_parse.quote(user) password = ':****' return '{user}{password}@{netloc}'.format(user=user, password=password, netloc=netloc) def _transform_url(url, transform_netloc): """Transform and replace netloc in a url. transform_netloc is a function taking the netloc and returning a tuple. The first element of this tuple is the new netloc. The entire tuple is returned. Returns a tuple containing the transformed url as item 0 and the original tuple returned by transform_netloc as item 1. """ purl = urllib_parse.urlsplit(url) netloc_tuple = transform_netloc(purl.netloc) # stripped url url_pieces = ( purl.scheme, netloc_tuple[0], purl.path, purl.query, purl.fragment ) surl = urllib_parse.urlunsplit(url_pieces) return surl, netloc_tuple def _get_netloc(netloc): return split_auth_from_netloc(netloc) def _redact_netloc(netloc): return (redact_netloc(netloc),) def split_auth_netloc_from_url(url): # type: (str) -> Tuple[str, str, Tuple[str, str]] """ Parse a url into separate netloc, auth, and url with no auth. Returns: (url_without_auth, netloc, (username, password)) """ url_without_auth, (netloc, auth) = _transform_url(url, _get_netloc) return url_without_auth, netloc, auth def remove_auth_from_url(url): # type: (str) -> str """Return a copy of url with 'username:password@' removed.""" # username/pass params are passed to subversion through flags # and are not recognized in the url. return _transform_url(url, _get_netloc)[0] def redact_auth_from_url(url): # type: (str) -> str """Replace the password in a given url with ****.""" return _transform_url(url, _redact_netloc)[0] class HiddenText(object): def __init__( self, secret, # type: str redacted, # type: str ): # type: (...) -> None self.secret = secret self.redacted = redacted def __repr__(self): # type: (...) -> str return ''.format(str(self)) def __str__(self): # type: (...) -> str return self.redacted # This is useful for testing. def __eq__(self, other): # type: (Any) -> bool if type(self) != type(other): return False # The string being used for redaction doesn't also have to match, # just the raw, original string. return (self.secret == other.secret) # We need to provide an explicit __ne__ implementation for Python 2. # TODO: remove this when we drop PY2 support. def __ne__(self, other): # type: (Any) -> bool return not self == other def hide_value(value): # type: (str) -> HiddenText return HiddenText(value, redacted='****') def hide_url(url): # type: (str) -> HiddenText redacted = redact_auth_from_url(url) return HiddenText(url, redacted=redacted) def protect_pip_from_modification_on_windows(modifying_pip): # type: (bool) -> None """Protection of pip.exe from modification on Windows On Windows, any operation modifying pip should be run as: python -m pip ... """ pip_names = set() for ext in ('', '.exe'): pip_names.add('pip{ext}'.format(ext=ext)) pip_names.add('pip{}{ext}'.format(sys.version_info[0], ext=ext)) pip_names.add('pip{}.{}{ext}'.format(*sys.version_info[:2], ext=ext)) # See https://github.com/pypa/pip/issues/1299 for more discussion should_show_use_python_msg = ( modifying_pip and WINDOWS and os.path.basename(sys.argv[0]) in pip_names ) if should_show_use_python_msg: new_command = [ sys.executable, "-m", "pip" ] + sys.argv[1:] raise CommandError( 'To modify pip, please run the following command:\n{}' .format(" ".join(new_command)) ) def is_console_interactive(): # type: () -> bool """Is this console interactive? """ return sys.stdin is not None and sys.stdin.isatty() site-packages/pip/_internal/utils/typing.py000064400000002145151464452710015066 0ustar00"""For neatly implementing static typing in pip. `mypy` - the static type analysis tool we use - uses the `typing` module, which provides core functionality fundamental to mypy's functioning. Generally, `typing` would be imported at runtime and used in that fashion - it acts as a no-op at runtime and does not have any run-time overhead by design. As it turns out, `typing` is not vendorable - it uses separate sources for Python 2/Python 3. Thus, this codebase can not expect it to be present. To work around this, mypy allows the typing import to be behind a False-y optional to prevent it from running at runtime and type-comments can be used to remove the need for the types to be accessible directly during runtime. This module provides the False-y guard in a nicely named fashion so that a curious maintainer can reach here to read this. In pip, all static-typing related imports should be guarded as follows: from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ... Ref: https://github.com/python/mypy/issues/3216 """ MYPY_CHECK_RUNNING = False site-packages/pip/_internal/utils/inject_securetransport.py000064400000001452151464452710020353 0ustar00"""A helper module that injects SecureTransport, on import. The import should be done as early as possible, to ensure all requests and sessions (or whatever) are created after injecting SecureTransport. Note that we only do the injection on macOS, when the linked OpenSSL is too old to handle TLSv1.2. """ import sys def inject_securetransport(): # type: () -> None # Only relevant on macOS if sys.platform != "darwin": return try: import ssl except ImportError: return # Checks for OpenSSL 1.0.1 if ssl.OPENSSL_VERSION_NUMBER >= 0x1000100f: return try: from pip._vendor.urllib3.contrib import securetransport except (ImportError, OSError): return securetransport.inject_into_urllib3() inject_securetransport() site-packages/pip/_internal/utils/encoding.py000064400000002450151464452710015341 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False import codecs import locale import re import sys from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import List, Tuple, Text BOMS = [ (codecs.BOM_UTF8, 'utf-8'), (codecs.BOM_UTF16, 'utf-16'), (codecs.BOM_UTF16_BE, 'utf-16-be'), (codecs.BOM_UTF16_LE, 'utf-16-le'), (codecs.BOM_UTF32, 'utf-32'), (codecs.BOM_UTF32_BE, 'utf-32-be'), (codecs.BOM_UTF32_LE, 'utf-32-le'), ] # type: List[Tuple[bytes, Text]] ENCODING_RE = re.compile(br'coding[:=]\s*([-\w.]+)') def auto_decode(data): # type: (bytes) -> Text """Check a bytes string for a BOM to correctly detect the encoding Fallback to locale.getpreferredencoding(False) like open() on Python3""" for bom, encoding in BOMS: if data.startswith(bom): return data[len(bom):].decode(encoding) # Lets check the first two lines as in PEP263 for line in data.split(b'\n')[:2]: if line[0:1] == b'#' and ENCODING_RE.search(line): encoding = ENCODING_RE.search(line).groups()[0].decode('ascii') return data.decode(encoding) return data.decode( locale.getpreferredencoding(False) or sys.getdefaultencoding(), ) site-packages/pip/_internal/utils/filetypes.py000064400000001073151464452710015557 0ustar00"""Filetype information. """ from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Tuple WHEEL_EXTENSION = '.whl' BZ2_EXTENSIONS = ('.tar.bz2', '.tbz') # type: Tuple[str, ...] XZ_EXTENSIONS = ('.tar.xz', '.txz', '.tlz', '.tar.lz', '.tar.lzma') # type: Tuple[str, ...] ZIP_EXTENSIONS = ('.zip', WHEEL_EXTENSION) # type: Tuple[str, ...] TAR_EXTENSIONS = ('.tar.gz', '.tgz', '.tar') # type: Tuple[str, ...] ARCHIVE_EXTENSIONS = ( ZIP_EXTENSIONS + BZ2_EXTENSIONS + TAR_EXTENSIONS + XZ_EXTENSIONS ) site-packages/pip/_internal/utils/marker_files.py000064400000001467151464452710016225 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False import os.path DELETE_MARKER_MESSAGE = '''\ This file is placed here by pip to indicate the source was put here by pip. Once this package is successfully installed this source code will be deleted (unless you remove this file). ''' PIP_DELETE_MARKER_FILENAME = 'pip-delete-this-directory.txt' def has_delete_marker_file(directory): return os.path.exists(os.path.join(directory, PIP_DELETE_MARKER_FILENAME)) def write_delete_marker_file(directory): # type: (str) -> None """ Write the pip delete marker file into this directory. """ filepath = os.path.join(directory, PIP_DELETE_MARKER_FILENAME) with open(filepath, 'w') as marker_fp: marker_fp.write(DELETE_MARKER_MESSAGE) site-packages/pip/_internal/utils/virtualenv.py000064400000001573151464452710015757 0ustar00import os.path import site import sys def running_under_virtualenv(): # type: () -> bool """ Return True if we're running inside a virtualenv, False otherwise. """ if hasattr(sys, 'real_prefix'): # pypa/virtualenv case return True elif sys.prefix != getattr(sys, "base_prefix", sys.prefix): # PEP 405 venv return True return False def virtualenv_no_global(): # type: () -> bool """ Return True if in a venv and no system site packages. """ # this mirrors the logic in virtualenv.py for locating the # no-global-site-packages.txt file site_mod_dir = os.path.dirname(os.path.abspath(site.__file__)) no_global_file = os.path.join(site_mod_dir, 'no-global-site-packages.txt') if running_under_virtualenv() and os.path.isfile(no_global_file): return True else: return False site-packages/pip/_internal/utils/urls.py000064400000002711151464452710014540 0ustar00import os import sys from pip._vendor.six.moves.urllib import parse as urllib_parse from pip._vendor.six.moves.urllib import request as urllib_request from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional, Text, Union def get_url_scheme(url): # type: (Union[str, Text]) -> Optional[Text] if ':' not in url: return None return url.split(':', 1)[0].lower() def path_to_url(path): # type: (Union[str, Text]) -> str """ Convert a path to a file: URL. The path will be made absolute and have quoted path parts. """ path = os.path.normpath(os.path.abspath(path)) url = urllib_parse.urljoin('file:', urllib_request.pathname2url(path)) return url def url_to_path(url): # type: (str) -> str """ Convert a file: URL to a path. """ assert url.startswith('file:'), ( "You can only turn file: urls into filenames (not %r)" % url) _, netloc, path, _, _ = urllib_parse.urlsplit(url) if not netloc or netloc == 'localhost': # According to RFC 8089, same as empty authority. netloc = '' elif sys.platform == 'win32': # If we have a UNC path, prepend UNC share notation. netloc = '\\\\' + netloc else: raise ValueError( 'non-local file URIs are not supported on this platform: %r' % url ) path = urllib_request.url2pathname(netloc + path) return path site-packages/pip/_internal/utils/__pycache__/ui.cpython-38.pyc000064400000027042151464452710020462 0ustar00U .eR6@sxddlmZmZddlZddlZddlZddlZddlZddlm Z m Z mZddl m Z ddl mZmZddlmZmZmZddlmZddlmZdd lmZdd lmZdd lmZerdd lmZm Z m!Z!zdd l m"Z"Wne#k rdZ"YnXe$e%Z&ddZ'e'eeZ(Gddde)Z*GdddeZ+GdddeZ,Gddde)Z-Gddde)Z.Gddde.e*e-Z/Gddde/e(Z0Gddde/e+Z1Gd d!d!e/eZ2Gd"d#d#e/eZ3Gd$d%d%e/e,Z4Gd&d'd'e.e*e-eZ5e1e1fe0e5fe2e5fe3e5fe4e5fd(Z6d7d)d*Z7ej8d+d,Z9Gd-d.d.e)Z:Gd/d0d0e)Z;Gd1d2d2e;ZdS)8)absolute_importdivisionN)SIGINTdefault_int_handlersignal)six) HIDE_CURSOR SHOW_CURSOR)BarFillingCirclesBarIncrementalBar)Spinner)WINDOWS)get_indentation) format_size)MYPY_CHECK_RUNNING)AnyIteratorIO)coloramacCst|jdd}|s|St|dtt|dtg}|tt|dg7}zt||Wntk rz|YSX|SdS)NencodingZ empty_fillZfillphases)getattrfilerZ text_typelistjoinencodeUnicodeEncodeError)Z preferredZfallbackrZ charactersr:/usr/lib/python3.8/site-packages/pip/_internal/utils/ui.py_select_progress_class%s r cs4eZdZdZfddZfddZddZZS)InterruptibleMixina Helper to ensure that self.finish() gets called on keyboard interrupt. This allows downloads to be interrupted without leaving temporary state (like hidden cursors) behind. This class is similar to the progress library's existing SigIntMixin helper, but as of version 1.2, that helper has the following problems: 1. It calls sys.exit(). 2. It discards the existing SIGINT handler completely. 3. It leaves its own handler in place even after an uninterrupted finish, which will have unexpected delayed effects if the user triggers an unrelated keyboard interrupt some time after a progress-displaying download has already completed, for example. cs4tt|j||tt|j|_|jdkr0t|_dS)z= Save the original SIGINT handler for later. N)superr!__init__rr handle_sigintoriginal_handlerrselfargskwargs __class__rrr#Us zInterruptibleMixin.__init__cstt|tt|jdS)z Restore the original SIGINT handler after finishing. This should happen regardless of whether the progress display finishes normally, or gets interrupted. N)r"r!finishrrr%r'r*rrr,eszInterruptibleMixin.finishcCs||||dS)z Call self.finish() before delegating to the original SIGINT handler. This handler should only be in place while the progress display is active. N)r,r%)r'Zsignumframerrrr$osz InterruptibleMixin.handle_sigint)__name__ __module__ __qualname____doc__r#r,r$ __classcell__rrr*rr!Cs  r!c@seZdZddZdS) SilentBarcCsdSNrr-rrrupdate|szSilentBar.updateN)r/r0r1r6rrrrr4zsr4c@seZdZdZdZdZdZdS) BlueEmojiBar %(percent)d%% )u🔹u🔷u🔵N)r/r0r1suffixZ bar_prefixZ bar_suffixrrrrrr7sr7csJeZdZfddZeddZeddZeddZd d d ZZ S) DownloadProgressMixincs,tt|j||dtd|j|_dS)Nr9)r"r;r#rmessager&r*rrr#szDownloadProgressMixin.__init__cCs t|jSr5)rindexr-rrr downloadedsz DownloadProgressMixin.downloadedcCs |jdkrdStd|jdS)Ngz...z/s)Zavgrr-rrrdownload_speeds z$DownloadProgressMixin.download_speedcCs|jrd|jSdS)Nzeta %s)ZetaZeta_tdr-rrr pretty_etas z DownloadProgressMixin.pretty_etar@ccs&|D]}|V||q|dSr5)nextr,)r'itnxrrriters zDownloadProgressMixin.iter)r@) r/r0r1r#propertyr?rArCrHr3rrr*rr;s    r;cseZdZfddZZS) WindowsMixincs\trjrd_ttj||trXtrXtj_fddj_fddj_ dS)NFcs jjSr5)rwrappedisattyrr-rrz'WindowsMixin.__init__..cs jjSr5)rrKflushrr-rrrMrN) rZ hide_cursorr"rJr#rZ AnsiToWin32rrLrOr&r*r-rr#s zWindowsMixin.__init__)r/r0r1r#r3rrr*rrJsrJc@seZdZejZdZdZdS)BaseDownloadProgressBarr8z0%(downloaded)s %(download_speed)s %(pretty_eta)sN)r/r0r1sysstdoutrr=r:rrrrrPsrPc@s eZdZdS)DefaultDownloadProgressBarNr/r0r1rrrrrSsrSc@s eZdZdS)DownloadSilentBarNrTrrrrrUsrUc@s eZdZdS) DownloadBarNrTrrrrrVsrVc@s eZdZdS)DownloadFillingCirclesBarNrTrrrrrWsrWc@s eZdZdS)DownloadBlueEmojiProgressBarNrTrrrrrXsrXc@s&eZdZejZdZddZddZdS)DownloadProgressSpinnerz!%(downloaded)s %(download_speed)scCs"t|dst|j|_t|jS)N_phaser)hasattr itertoolscyclerrZrDr-rrr next_phases z"DownloadProgressSpinner.next_phasecCsN|j|}|}|j|}d||r*dnd||r6dnd|g}||dS)NrBr9)r=r^r:rZwriteln)r'r=Zphaser:linerrrr6s    zDownloadProgressSpinner.updateN) r/r0r1rQrRrr:r^r6rrrrrYsrY)ZoffZonasciiZprettyZemojicCs8|dks|dkr t|djSt|d|djSdS)Nrr@)max) BAR_TYPESrH)Z progress_barrarrrDownloadProgressProvider srcc csPtr dVn@|r"ttjkr*dVn"|tz dVW5|tXdSr5) rrLloggergetEffectiveLevelloggingINFOwriterr )rrrr hidden_cursors  ric@s$eZdZddZddZddZdS) RateLimitercCs||_d|_dS)Nr)_min_update_interval_seconds _last_update)r'min_update_interval_secondsrrrr#-szRateLimiter.__init__cCst}||j}||jkSr5)timerlrk)r'ZnowZdeltarrrready2s zRateLimiter.readycCst|_dSr5)rnrlr-rrrreset8szRateLimiter.resetN)r/r0r1r#rorprrrrrj,srjc@seZdZddZddZdS)SpinnerInterfacecCs tdSr5NotImplementedErrorr-rrrspin>szSpinnerInterface.spincCs tdSr5rrr'Z final_statusrrrr,BszSpinnerInterface.finishN)r/r0r1rtr,rrrrrq=srqc@s.eZdZd ddZddZdd Zd d ZdS) InteractiveSpinnerN-\|/?cCs\||_|dkrtj}||_t||_d|_t||_ |j dt |jdd|_ dS)NFr9z ... r) _messagerQrR_filerj _rate_limiter _finishedr\r] _spin_cyclerhr_width)r'r=rZ spin_charsrmrrrr#Hs  zInteractiveSpinner.__init__cCs\|jr td|j}|j|d|j||j|t||_|j|jdS)Nr9) r|AssertionErrorr~rzrhlenrOr{rp)r'statusZbackuprrr_writeWs     zInteractiveSpinner._writecCs,|jr dS|jsdS|t|jdSr5)r|r{rorrDr}r-rrrrtcs  zInteractiveSpinner.spincCs4|jr dS|||jd|jd|_dS)N T)r|rrzrhrOrurrrr,ks    zInteractiveSpinner.finish)Nrwrx)r/r0r1r#rrtr,rrrrrvGs   rvc@s.eZdZd ddZddZddZdd Zd S) NonInteractiveSpinner<cCs$||_d|_t||_|ddS)NFZstarted)ryr|rjr{_update)r'r=rmrrrr#zs zNonInteractiveSpinner.__init__cCs(|jr t|jtd|j|dS)Nz%s: %s)r|rr{rprdinfory)r'rrrrrs  zNonInteractiveSpinner._updatecCs&|jr dS|jsdS|ddS)Nzstill running...)r|r{rorr-rrrrts  zNonInteractiveSpinner.spincCs$|jr dS|d|fd|_dS)Nzfinished with status '%s'T)r|rrurrrr,szNonInteractiveSpinner.finishN)r)r/r0r1r#rrtr,rrrrrys rc cstjr"ttjkr"t|}nt|}z t tj |VW5QRXWn>t k rj| dYn*t k r| dYn X| ddS)NZcancelederrorZdone) rQrRrLrdrerfrgrvrriKeyboardInterruptr, Exception)r=Zspinnerrrr open_spinners    r)N)?Z __future__rr contextlibr\rfrQrnrrrZ pip._vendorrZpip._vendor.progressrr Zpip._vendor.progress.barr r r Zpip._vendor.progress.spinnerr Zpip._internal.utils.compatrZpip._internal.utils.loggingrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingrrrrrZ getLoggerr/rdr Z_BaseBarobjectr!r4r7r;rJrPrSrUrVrWrXrYrbrccontextmanagerrirjrqrvrrrrrrs~         7       2site-packages/pip/_internal/utils/__pycache__/appdirs.cpython-38.opt-1.pyc000064400000017553151464452710022454 0ustar00U .e&&@sdZddlmZddlZddlZddlmZmZddlm Z m Z ddl m Z e r\ddl mZdd Zdd d ZdddZddZddZddZe rzddlZeZWnek reZYnXddZdS)zd This code was taken from https://github.com/ActiveState/appdirs and modified to suit our purposes. )absolute_importN)PY2 text_type)WINDOWS expanduser)MYPY_CHECK_RUNNING)ListcCstr Unix: ~/.cache/ (XDG default) Windows: C:\Users\\AppData\Local\\Cache On Windows the only suggestion in the MSDN docs is that local settings go in the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming app data dir (the default returned by `user_data_dir`). Apps typically put cache data somewhere *under* the given dir here. Some examples: ...\Mozilla\Firefox\Profiles\\Cache ...\Acme\SuperApp\Cache\1.0 OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value. CSIDL_LOCAL_APPDATAZCachedarwinz~/Library/CachesZXDG_CACHE_HOMEz~/.cache)rospathnormpath_get_win_folderr isinstancer_win_path_to_bytesjoinsysplatformrgetenv)appnamer r?/usr/lib/python3.8/site-packages/pip/_internal/utils/appdirs.pyuser_cache_dirs rFcCstr,|r dpd}tjtjt||}ndtjdkrvtjtjt d|rbtjt d|ntjt d|}ntjt dt d|}|S)a Return full path to the user-specific data dir for this application. "appname" is the name of application. If None, just the system directory is returned. "roaming" (boolean, default False) can be set True to use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: macOS: ~/Library/Application Support/ if it exists, else ~/.config/ Unix: ~/.local/share/ # or in $XDG_DATA_HOME, if defined Win XP (not roaming): C:\Documents and Settings\\ ... ...Application Data\ Win XP (roaming): C:\Documents and Settings\\Local ... ...Settings\Application Data\ Win 7 (not roaming): C:\\Users\\AppData\Local\ Win 7 (roaming): C:\\Users\\AppData\Roaming\ For Unix, we follow the XDG spec and support $XDG_DATA_HOME. That means, by default "~/.local/share/". CSIDL_APPDATAr r z~/Library/Application Support/z ~/.config/Z XDG_DATA_HOMEz~/.local/share) rr r rr rrrisdirrr)rroamingconstr rrr user_data_dirHs,    rTcCsHtrt||d}n2tjdkr&t|}ntdtd}tj||}|S)arReturn full path to the user-specific config dir for this application. "appname" is the name of application. If None, just the system directory is returned. "roaming" (boolean, default True) can be set False to not use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: macOS: same as user_data_dir Unix: ~/.config/ Win *: same as user_data_dir For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME. That means, by default "~/.config/". )rr ZXDG_CONFIG_HOMEz ~/.config) rrrrr rrr r)rrr rrruser_config_dir}s  rcstr&tjtd}tj|g}nVtjdkrBtjdg}n:tdd}|rnfdd| tj D}ng}| d|S) aReturn a list of potential user-shared config dirs for this application. "appname" is the name of application. Typical user config directories are: macOS: /Library/Application Support// Unix: /etc or $XDG_CONFIG_DIRS[i]// for each value in $XDG_CONFIG_DIRS Win XP: C:\Documents and Settings\All Users\Application ... ...Data\\ Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) Win 7: Hidden, but writeable on Win 7: C:\ProgramData\\ CSIDL_COMMON_APPDATAr z/Library/Application SupportZXDG_CONFIG_DIRSz/etc/xdgcsg|]}tjt|qSr)r r rr).0xrrr sz$site_config_dirs..z/etc) rr r r rrrrrsplitpathsepappend)rr ZpathlistZxdg_config_dirsrr"rsite_config_dirss     r'cCs:ddl}dddd|}||jd}|||\}}|S)z This is a fallback technique at best. I'm not sure if using the registry for this guarantees us the correct answer for all CSIDL_* names. rNZAppDatazCommon AppDataz Local AppDatarrr z@Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders)_winregOpenKeyHKEY_CURRENT_USERZ QueryValueEx) csidl_namer)Zshell_folder_namekeyZ directoryZ_typerrr_get_win_folder_from_registrysr.cCsdddd|}td}tj}|jd|dd|d}|D]}t|dkr. If encoding using ASCII and MBCS fails, return the original Unicode path. )ASCIIZMBCS)encodeUnicodeEncodeError LookupError)r encodingrrrrs r)F)T)__doc__Z __future__rr rZpip._vendor.sixrrZpip._internal.utils.compatrrZpip._internal.utils.typingrtypingrrrrr'r.r8r3r ImportErrorrrrrrs*   1 5 ") site-packages/pip/_internal/utils/__pycache__/urls.cpython-38.opt-1.pyc000064400000002506151464452710021767 0ustar00U .e@shddlZddlZddlmZddlmZddlmZerLddl m Z m Z m Z ddZ dd Zd d ZdS) N)parse)request)MYPY_CHECK_RUNNING)OptionalTextUnioncCs d|kr dS|dddS)N:r)splitlower)urlr s    site-packages/pip/_internal/utils/__pycache__/virtualenv.cpython-38.opt-1.pyc000064400000001556151464452710023205 0ustar00U .e{@s,ddlZddlZddlZddZddZdS)NcCs*ttdrdStjttdtjkr&dSdS)zM Return True if we're running inside a virtualenv, False otherwise. Z real_prefixT base_prefixF)hasattrsysprefixgetattrrrB/usr/lib/python3.8/site-packages/pip/_internal/utils/virtualenv.pyrunning_under_virtualenvs  r cCsBtjtjtj}tj|d}tr:tj|r:dSdSdS)z? Return True if in a venv and no system site packages. zno-global-site-packages.txtTFN) ospathdirnameabspathsite__file__joinr isfile)Z site_mod_dirZno_global_filerrrvirtualenv_no_globals r)Zos.pathr rrr rrrrrssite-packages/pip/_internal/utils/__pycache__/models.cpython-38.pyc000064400000003604151464452710021326 0ustar00U .e|@s dZddlZGdddeZdS)zUtilities for defining models Nc@sXeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ dS)KeyBasedCompareMixinz/usr/lib/python3.8/site-packages/pip/_internal/utils/models.py__init__ szKeyBasedCompareMixin.__init__cCs t|jSr)hashr)rrrr __hash__szKeyBasedCompareMixin.__hash__cCs||tjSr)_compareoperator__lt__rotherrrr rszKeyBasedCompareMixin.__lt__cCs||tjSr)r r__le__rrrr rszKeyBasedCompareMixin.__le__cCs||tjSr)r r__gt__rrrr rszKeyBasedCompareMixin.__gt__cCs||tjSr)r r__ge__rrrr rszKeyBasedCompareMixin.__ge__cCs||tjSr)r r__eq__rrrr r szKeyBasedCompareMixin.__eq__cCs||tjSr)r r__ne__rrrr r#szKeyBasedCompareMixin.__ne__cCst||jstS||j|jSr) isinstancerNotImplementedr)rrmethodrrr r &s zKeyBasedCompareMixin._compareN) __name__ __module__ __qualname____doc__r r rrrrrrr rrrr r sr)rrobjectrrrrr ssite-packages/pip/_internal/utils/__pycache__/temp_dir.cpython-38.pyc000064400000011517151464452710021650 0ustar00U .e@sddlmZddlZddlZddlZddlZddlZddlm Z ddl m Z e r\ddl m Z eeZGdddeZGdd d eZdS) )absolute_importN)rmtree)MYPY_CHECK_RUNNING)OptionalcsVeZdZdZdfdd ZeddZdd Zd d Zd d Z ddZ ddZ Z S) TempDirectoryaMHelper class that owns and cleans up a temporary directory. This class can be used as a context manager or as an OO representation of a temporary directory. Attributes: path Location to the created temporary directory delete Whether the directory should be deleted when exiting (when used as a contextmanager) Methods: cleanup() Deletes the temporary directory When used as a context manager, if the delete attribute is True, on exiting the context the temporary directory is deleted. NtempcsPtt||dkr"|dkr"d}|dkr4||}||_d|_||_||_dS)NTF)superr__init___create_path_deleteddeletekind)selfpathr r __class__@/usr/lib/python3.8/site-packages/pip/_internal/utils/temp_dir.pyr +s zTempDirectory.__init__cCs|jrtd|j|jS)Nz$Attempted to access deleted path: {})r AssertionErrorformatr rrrrr@s zTempDirectory.pathcCsd|jj|jS)Nz <{} {!r}>)rr__name__rrrrr__repr__HszTempDirectory.__repr__cCs|SNrrrrr __enter__KszTempDirectory.__enter__cCs|jr|dSr)r cleanup)rexcvaluetbrrr__exit__NszTempDirectory.__exit__cCs.tjtjd|d}td||S)zECreate a temporary directory and store its path in self.path pip-{}-prefixCreated temporary directory: {})osrrealpathtempfilemkdtemprloggerdebug)rrrrrrr Rs zTempDirectory._createcCs"d|_tj|jrt|jdS)z?Remove the temporary directory created and reset state TN)r r%rexistsr rrrrrr_szTempDirectory.cleanup)NNr) r __module__ __qualname____doc__r propertyrrrr r r __classcell__rrrrrs  rcs:eZdZdZdZd fdd ZeddZdd ZZ S) AdjacentTempDirectoryaHelper class that creates a temporary directory adjacent to a real one. Attributes: original The original directory to create a temp directory for. path After calling create() or entering, contains the full path to the temporary directory. delete Whether the directory should be deleted when exiting (when used as a contextmanager) z-~.=%0123456789Ncs"|d|_tt|j|ddS)Nz/\)r )rstriporiginalrr1r )rr3r rrrr |s zAdjacentTempDirectory.__init__ccstdt|D]D}t|j|dD],}dd|||d}||kr$|Vq$qtt|jD]8}t|j|D]$}dd||}||krt|VqtqbdS)a Generates a series of temporary names. The algorithm replaces the leading characters in the name with ones that are valid filesystem characters, but are not valid package names (for both Python and pip definitions of package). ~N)rangelen itertoolscombinations_with_replacement LEADING_CHARSjoin)clsnamei candidatenew_namerrr_generate_namess  z%AdjacentTempDirectory._generate_namesc Cstj|j\}}||D]b}tj||}zt|Wn0tk rl}z|jtj kr\W5d}~XYqXtj |}qqtj t j d |d}td ||S)Nr!r"r$)r%rsplitr3rBr<mkdirOSErrorerrnoZEEXISTr&r'r(rr)r*)rrrootr>r@rZexrrrr s  zAdjacentTempDirectory._create)N) rr,r-r.r;r classmethodrBr r0rrrrr1gs  r1)Z __future__rrFr9ZloggingZos.pathr%r'Zpip._internal.utils.miscrZpip._internal.utils.typingrtypingrZ getLoggerrr)objectrr1rrrrs     Qsite-packages/pip/_internal/utils/__pycache__/filesystem.cpython-38.opt-1.pyc000064400000005325151464452710023170 0ustar00U .e @sddlZddlZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZddlmZerdd lmZmZGd d d eZd d ZddZddZeddZe dddZe reddZn eejZdS)N)contextmanager)NamedTemporaryFile)retry)PY2) get_path_uid)cast)MYPY_CHECK_RUNNING)BinaryIOIteratorc@seZdZeddZdS)NamedTemporaryFileResultcCsdSN)selfr r B/usr/lib/python3.8/site-packages/pip/_internal/utils/filesystem.pyfileszNamedTemporaryFileResult.fileN)__name__ __module__ __qualname__propertyrr r r rr sr cCsttdsdSd}||krtj|rntdkr^z t|}Wntk rTYdSX|dkSt|tjSq|tj |}}qdS)NgeteuidTrF) hasattrospathlexistsrrOSErroraccessW_OKdirname)rZpreviousZpath_uidr r rcheck_path_owners    rc Csrzt||Wn\ttfk rl||fD]8}z t|}Wntk rPYq,X|r,td|q,YnXdS)zWrap shutil.copy2() but map errors copying socket files to SpecialFileError as expected. See also https://bugs.python.org/issue37700. z`%s` is a socketN)shutilZcopy2rIOError is_socketZSpecialFileError)srcdestfZis_socket_filer r r copy2_fixed5s  r%cCstt|jSr )statS_ISSOCKrlstatst_mode)rr r rr!Msr!c csbtdtj|tj|dd8}td|}z |VW5|jt|j XW5QRXdS)zGiven a path to a file, open a temp file next to it securely and ensure it is written to disk after the context reaches its end. Fz.tmp)deletedirprefixsuffixr N) rrrrbasenamerrflushfsyncfileno)rr$resultr r radjacent_tmp_fileRs     r3i)Zstop_max_delayZ wait_fixedcCs@zt||Wn*tk r:t|t||YnXdSr )rrenamerremove)r"r#r r rreplaceis  r7)rZos.pathrr& contextlibrZtempfilerZpip._vendor.retryingrZpip._vendor.sixrZpip._internal.utils.compatrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingr r r rr%r!r3Z_replace_retryr7r r r rs.          site-packages/pip/_internal/utils/__pycache__/glibc.cpython-38.opt-1.pyc000064400000004321151464452710022057 0ustar00U .e-@sxddlmZddlZddlZddlZddlmZerDddlmZm Z ddZ ddZ d d Z d d Z d dZddZdS))absolute_importN)MYPY_CHECK_RUNNING)OptionalTuplecCs tp tS)z9Returns glibc version string, or None if not using glibc.)glibc_version_string_confstrglibc_version_string_ctypesrr=/usr/lib/python3.8/site-packages/pip/_internal/utils/glibc.pyglibc_version_stringsr c Cs8ztd\}}Wntttfk r2YdSX|S)z@Primary implementation of glibc_version_string using os.confstr.CS_GNU_LIBC_VERSIONN)osconfstrsplitAttributeErrorOSError ValueError)_versionrrr rs rcCsvz ddl}Wntk r"YdSX|d}z |j}Wntk rNYdSX|j|_|}t|tsr| d}|S)z=Fallback implementation of glibc_version_string using ctypes.rNascii) ctypes ImportErrorZCDLLgnu_get_libc_versionrZc_char_pZrestype isinstancestrdecode)rZprocess_namespacer version_strrrr r&s     rcCsHtd|}|s$td|tdSt|d|koFt|d|kS)Nz$(?P[0-9]+)\.(?P[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFmajorminor)rematchwarningswarnRuntimeWarningintgroup)rrequired_major minimum_minormrrr check_glibc_versionFs r(cCst}|dkrdSt|||S)NF)r r()r%r&rrrr have_compatible_glibcWsr)cCst}|dkrdSd|fSdS)zTry to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings in case the lookup fails. N)r*Zglibc)r )Z glibc_versionrrr libc_verpsr+)Z __future__rr rr Zpip._internal.utils.typingrtypingrrr rrr(r)r+rrrr s   site-packages/pip/_internal/utils/__pycache__/typing.cpython-38.opt-1.pyc000064400000002405151464452710022312 0ustar00U .ee@s dZdZdS)aBFor neatly implementing static typing in pip. `mypy` - the static type analysis tool we use - uses the `typing` module, which provides core functionality fundamental to mypy's functioning. Generally, `typing` would be imported at runtime and used in that fashion - it acts as a no-op at runtime and does not have any run-time overhead by design. As it turns out, `typing` is not vendorable - it uses separate sources for Python 2/Python 3. Thus, this codebase can not expect it to be present. To work around this, mypy allows the typing import to be behind a False-y optional to prevent it from running at runtime and type-comments can be used to remove the need for the types to be accessible directly during runtime. This module provides the False-y guard in a nicely named fashion so that a curious maintainer can reach here to read this. In pip, all static-typing related imports should be guarded as follows: from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ... Ref: https://github.com/python/mypy/issues/3216 FN)__doc__ZMYPY_CHECK_RUNNINGrr>/usr/lib/python3.8/site-packages/pip/_internal/utils/typing.pyssite-packages/pip/_internal/utils/__pycache__/compat.cpython-38.pyc000064400000015267151464452710021336 0ustar00U .e]% @s~dZddlmZmZddlZddlZddlZddlZddlZddl Z ddl m Z m Z ddl mZddlmZerddlmZmZmZmZz ddlZWnek rdZYnXeZz ddlZWnVek rzddlmZWn.ek rddlZeje_eje_YnXYnXd d d d d dddddg Z e!e"Z#edk pDeZ$e rddl%Z%z e%j&Z&Wne'k rxdZ&YnXe&dk Z(ndZ(ddl)m&Z&e rddZ*e+de*dZ,ndZ,d/ddZ-dd Z.e rd0dd Z/n d1dd Z/dd Z0e r dd l%m1Z1d!dZ2ndd"l3m4Z4d#dZ2d$d%Z5d&d'd(hZ6e j78d)pRe j7d*koRej9d+kZ:d,dZ;e/usr/lib/python3.8/site-packages/pip/_internal/utils/compat.py Lsz-backslashreplace_decode_fn..css|]}t|VqdSr)ord)rbrrrrNscss|]}d|VqdS)z\x%xNr)rcrrrrOs)rangestartendjoin)rZ raw_bytesrrrbackslashreplace_decode_fnKsr(backslashreplace_decodebackslashreplacecCst|tr|St}|r*t|jdkr.d}z||}WnDtk r|dkrXd}d |}t |||j|t d}YnXt t tdddd}|r|j|d d}||}|S) a For display or logging purposes, convert a bytes object (or text) to text (e.g. unicode in Python 2) safe for output. :param desc: An optional phrase describing the input data, for use in the log message if a warning is logged. Defaults to "Bytes object". This function should never error out and so can take a best effort approach. It is okay to be lossy if needed since the return value is just for display. We assume the data is in the locale preferred encoding. If it won't decode properly, we warn the user but decode as best we can. We also ensure that the output can be safely written to standard output without encoding errors. asciiutf-8Nz Bytes objectz&{} does not appear to be encoded as %s)errors __stderr__encodingr*) isinstancerlocaleZgetpreferredencodingcodecslookupnamedecodeUnicodeDecodeErrorformatloggerZwarningr)getattrsysencode)datadescr/Z decoded_dataZ msg_formatZoutput_encodingZoutput_encodedrrrstr_to_displayYs0     r>cCs t|ddS)z)r<rrrrsFcCst|tr|dS|S)Nr,)r0rr;sreplacerrrrs  cCs"t|tr|d|rdndS|S)Nr,rAstrict)r0bytesr5r?rrrrs cCs`ttdr6t|tjtjB}t|j}t|n&tj |sPt |j}n t d||S)a) Return path's uid. Does not follow symlinks: https://github.com/pypa/pip/pull/935#discussion_r5307003 Placed this function in compat due to differences on AIX and Jython, that should eventually go away. :raises OSError: When path is a symlink or can't be read. O_NOFOLLOWz1%s is a symlink; Will not return uid for symlinks) hasattrosopenO_RDONLYrDfstatst_uidclosepathislinkstatOSError)rLfdZfile_uidrrrrs     get_suffixescCsddtDS)NcSsg|] }|dqS)rr)rsuffixrrr sz*get_extension_suffixes..rQrrrrrsEXTENSION_SUFFIXEScCstSrrUrrrrrscCs0tj|}|dr,|dr,|dd}|S)zm Expand ~ and ~user constructions. Includes a workaround for https://bugs.python.org/issue14768 z~/z//N)rFrL expanduser startswith)rLZexpandedrrrrXs  rXpythonZwsgirefargparsewinZclintcCsNttjdrtj||Stjtj|}tjtj|}||kSdS)z>Provide an alternative for os.path.samefile on Windows/Python2rN)rErFrLrnormcaseabspath)Zfile1Zfile2Zpath1Zpath2rrrrs  cCs ttS) Returns a tuple (x, y) representing the width(x) and the height(y) in characters of the terminal window. )tupleshutilrrrrrrscCsdd}|dp|dp|d}|sbz(tttj}||}t|Wntk r`YnX|stjddtjdd f}t|dt|dfS) r`cSs\z4ddl}ddl}ddl}|d|||jd}Wntk rJYdSX|dkrXdS|S)NrZhhZ12345678)rr)fcntltermiosstruct unpack_fromZioctlZ TIOCGWINSZ Exception)rPrcrdrecrrrr ioctl_GWINSZsz'get_terminal_size..ioctl_GWINSZrrWZLINESZCOLUMNSP) rFrGctermidrHrKrgenvirongetint)rirhrPrrrr s)N)F)F)>__doc__Z __future__rrr2r1ZloggingrFrbr:Zpip._vendor.sixrrZpip._vendor.urllib3.utilrZpip._internal.utils.typingrtypingrr r r Z_ssl ImportErrorZsslr Z pip._vendorZipaddrZ IPAddressZ ip_addressZ IPNetworkZ ip_network__all__Z getLogger__name__r8ZHAS_TLSZimprAttributeErrorr importlib.utilr(register_errorr)r>rrrrRrZimportlib.machineryrVrXrplatformrYr4rrrErrrrrs           B        site-packages/pip/_internal/utils/__pycache__/inject_securetransport.cpython-38.pyc000064400000001644151464452710024644 0ustar00U .e*@sdZddlZddZedS)a-A helper module that injects SecureTransport, on import. The import should be done as early as possible, to ensure all requests and sessions (or whatever) are created after injecting SecureTransport. Note that we only do the injection on macOS, when the linked OpenSSL is too old to handle TLSv1.2. Nc CsxtjdkrdSz ddl}Wntk r0YdSX|jdkr@dSzddlm}Wnttfk rjYdSX|dS)Ndarwinri)securetransport) sysplatformssl ImportErrorZOPENSSL_VERSION_NUMBERZpip._vendor.urllib3.contribrOSErrorZinject_into_urllib3)rrr N/usr/lib/python3.8/site-packages/pip/_internal/utils/inject_securetransport.pyinject_securetransport s   r )__doc__rr r r r r s site-packages/pip/_internal/utils/__pycache__/marker_files.cpython-38.pyc000064400000001640151464452710022504 0ustar00U .e7@s$ddlZdZdZddZddZdS)NzThis file is placed here by pip to indicate the source was put here by pip. Once this package is successfully installed this source code will be deleted (unless you remove this file). zpip-delete-this-directory.txtcCstjtj|tS)N)ospathexistsjoinPIP_DELETE_MARKER_FILENAME) directoryrD/usr/lib/python3.8/site-packages/pip/_internal/utils/marker_files.pyhas_delete_marker_filesr c Cs2tj|t}t|d}|tW5QRXdS)z? Write the pip delete marker file into this directory. wN)rrrropenwriteDELETE_MARKER_MESSAGE)rfilepathZ marker_fprrr write_delete_marker_files r)Zos.pathrrrr rrrrr ssite-packages/pip/_internal/utils/__pycache__/deprecation.cpython-38.pyc000064400000005412151464452710022337 0ustar00U .e @sdZddlmZddlZddlZddlmZddlmZ ddl m Z e rXddl m Z mZdZGd d d eZdadd d Zd dZdddZdS)zN A module that implements tooling to enable easy warnings about deprecations. )absolute_importN)parse) __version__)MYPY_CHECK_RUNNING)AnyOptionalz DEPRECATION: c@s eZdZdS)PipDeprecationWarningN)__name__ __module__ __qualname__r r C/usr/lib/python3.8/site-packages/pip/_internal/utils/deprecation.pyrsrcCsZ|dk r$tdk rVt||||||n2t|trDtd}||nt||||||dS)Nzpip._internal.deprecations)_original_showwarning issubclassrloggingZ getLoggerZwarning)messagecategoryfilenamelinenofilelineZloggerr r r _showwarning!s*   rcCs(tjdtddtdkr$tjatt_dS)NdefaultT)append)warnings simplefilterrr showwarningrr r r r install_warning_logger2srcCsh|tdf|df|df|dfg}ddd|D}|dk rTttt|krTt|tj|td d dS) aHelper to deprecate existing functionality. reason: Textual reason shown to the user about why this functionality has been deprecated. replacement: Textual suggestion shown to the user about what alternative functionality they can use. gone_in: The version of pip does this functionality should get removed in. Raises errors if pip's current version is greater than or equal to this. issue: Issue number on the tracker that would serve as a useful place for users to find related discussion and provide feedback. Always pass replacement, gone_in and issue as keyword arguments for clarity at the call site. z{}z2pip {} will remove support for this functionality.zA possible replacement is {}.zPYou can find discussion regarding this at https://github.com/pypa/pip/issues/{}. css$|]\}}|dk r||VqdS)N)format).0valtemplater r r `szdeprecated..N)r stacklevel)DEPRECATION_MSG_PREFIXjoinrcurrent_versionrrwarn)reasonZ replacementZgone_inZissueZ sentencesrr r r deprecated>s  r+)NN)N)__doc__Z __future__rrrZpip._vendor.packaging.versionrZpiprr(Zpip._internal.utils.typingrtypingrrr&Warningrrrrr+r r r r s      site-packages/pip/_internal/utils/__pycache__/temp_dir.cpython-38.opt-1.pyc000064400000011365151464452710022610 0ustar00U .e@sddlmZddlZddlZddlZddlZddlZddlm Z ddl m Z e r\ddl m Z eeZGdddeZGdd d eZdS) )absolute_importN)rmtree)MYPY_CHECK_RUNNING)OptionalcsVeZdZdZdfdd ZeddZdd Zd d Zd d Z ddZ ddZ Z S) TempDirectoryaMHelper class that owns and cleans up a temporary directory. This class can be used as a context manager or as an OO representation of a temporary directory. Attributes: path Location to the created temporary directory delete Whether the directory should be deleted when exiting (when used as a contextmanager) Methods: cleanup() Deletes the temporary directory When used as a context manager, if the delete attribute is True, on exiting the context the temporary directory is deleted. NtempcsPtt||dkr"|dkr"d}|dkr4||}||_d|_||_||_dS)NTF)superr__init___create_path_deleteddeletekind)selfpathr r __class__@/usr/lib/python3.8/site-packages/pip/_internal/utils/temp_dir.pyr +s zTempDirectory.__init__cCs|jSN)r rrrrr@szTempDirectory.pathcCsd|jj|jS)Nz <{} {!r}>)formatr__name__rrrrr__repr__HszTempDirectory.__repr__cCs|Srrrrrr __enter__KszTempDirectory.__enter__cCs|jr|dSr)r cleanup)rexcvaluetbrrr__exit__NszTempDirectory.__exit__cCs.tjtjd|d}td||S)zECreate a temporary directory and store its path in self.path pip-{}-prefixCreated temporary directory: {})osrrealpathtempfilemkdtemprloggerdebug)rrrrrrr Rs zTempDirectory._createcCs"d|_tj|jrt|jdS)z?Remove the temporary directory created and reset state TN)r r$rexistsr rrrrrr_szTempDirectory.cleanup)NNr) r __module__ __qualname____doc__r propertyrrrrr r __classcell__rrrrrs  rcs:eZdZdZdZd fdd ZeddZdd ZZ S) AdjacentTempDirectoryaHelper class that creates a temporary directory adjacent to a real one. Attributes: original The original directory to create a temp directory for. path After calling create() or entering, contains the full path to the temporary directory. delete Whether the directory should be deleted when exiting (when used as a contextmanager) z-~.=%0123456789Ncs"|d|_tt|j|ddS)Nz/\)r )rstriporiginalrr0r )rr2r rrrr |s zAdjacentTempDirectory.__init__ccstdt|D]D}t|j|dD],}dd|||d}||kr$|Vq$qtt|jD]8}t|j|D]$}dd||}||krt|VqtqbdS)a Generates a series of temporary names. The algorithm replaces the leading characters in the name with ones that are valid filesystem characters, but are not valid package names (for both Python and pip definitions of package). ~N)rangelen itertoolscombinations_with_replacement LEADING_CHARSjoin)clsnamei candidatenew_namerrr_generate_namess  z%AdjacentTempDirectory._generate_namesc Cstj|j\}}||D]b}tj||}zt|Wn0tk rl}z|jtj kr\W5d}~XYqXtj |}qqtj t j d |d}td ||S)Nr r!r#)r$rsplitr2rAr;mkdirOSErrorerrnoZEEXISTr%r&r'rr(r))rrrootr=r?rZexrrrr s  zAdjacentTempDirectory._create)N) rr+r,r-r:r classmethodrAr r/rrrrr0gs  r0)Z __future__rrEr8ZloggingZos.pathr$r&Zpip._internal.utils.miscrZpip._internal.utils.typingrtypingrZ getLoggerrr(objectrr0rrrrs     Qsite-packages/pip/_internal/utils/__pycache__/urls.cpython-38.pyc000064400000002671151464452710021033 0ustar00U .e@shddlZddlZddlmZddlmZddlmZerLddl m Z m Z m Z ddZ dd Zd d ZdS) N)parse)request)MYPY_CHECK_RUNNING)OptionalTextUnioncCs d|kr dS|dddS)N:r)splitlower)urlr s    site-packages/pip/_internal/utils/__pycache__/subprocess.cpython-38.opt-1.pyc000064400000012736151464452710023200 0ustar00U .e& @sddlmZddlZddlZddlZddlmZddlmZddl m Z m Z ddl m Z ddlmZmZddlmZdd lmZerdd lmZmZmZmZmZmZmZmZdd lmZeeeefZ d Z!d dZ"ddZ#ddZ$ddZ%dddZ&ddZ'dS))absolute_importN) shlex_quote)InstallationError)console_to_strstr_to_display)subprocess_logger) HiddenTextpath_to_display)MYPY_CHECK_RUNNING) open_spinner)AnyCallableIterableListMappingOptionalTextUnion)SpinnerInterfacez(----------------------------------------cGs2g}|D]$}t|tr"||q||q|S)z& Create a CommandArgs object. ) isinstancelistextendappend)argsZ command_argsargrB/usr/lib/python3.8/site-packages/pip/_internal/utils/subprocess.py make_commands    rcCsddd|DS)z/ Format command arguments for display.  css,|]$}t|trtt|nt|VqdS)N)rrrstr.0rrrr ;sz&format_command_args..)joinrrrrformat_command_args1s r%cCsdd|DS)z= Return the arguments in their raw, unredacted form. cSs g|]}t|tr|jn|qSr)rrZsecretr rrr Fsz'reveal_command_args..rr$rrrreveal_command_argsAsr'c CsDt|}t|dd}t|}d|}dj|||t||td}|S)z Create and return the error message to use to log a subprocess error with command output. :param lines: A list of lines, each ending with a newline. z command bytes)ZdesczCommand errored out with exit status {exit_status}: command: {command_display} cwd: {cwd_display} Complete output ({line_count} lines): {output}{divider}) exit_statuscommand_display cwd_displayZ line_countoutputZdivider)r%rr r#formatlen LOG_DIVIDER) cmd_argscwdlinesr)Zcommandr*r+r,msgrrrmake_subprocess_output_errorKs    r4FraiseTc  Cs4|dkr g}|dkrg}|r*tj} tj} n tj} tj} t| k} | oN|dk } |dkr`t|}| d|tj }|r| ||D]}| |dqz.t jt|t jt jt j||d}|jWn6tk r}z| rtd||W5d}~XYnXg}t|j}|sqJ|}||d| || r|qz |W5|jrj|jX|jo||j|k}| r|r|dn |d|r*|dkr| s| rt||||jd }t |d !|j|}t"|n:|d krt#d ||j|n|d krnt$dt%|d&|S)a Args: show_stdout: if true, use INFO to log the subprocess's stderr and stdout streams. Otherwise, use DEBUG. Defaults to False. extra_ok_returncodes: an iterable of integer return codes that are acceptable, in addition to 0. Defaults to None, which means []. unset_environ: an iterable of environment variable names to unset prior to calling subprocess.Popen(). log_failed_cmd: if false, failed commands are not logged, only raised. NzRunning command %s)stderrstdinstdoutr1envz#Error %s while executing command %s errorZdoner5)r0r1r2r)zSCommand errored out with exit status {}: {} Check the logs for full command output.warnz$Command "%s" had error code %s in %signorezInvalid value: on_returncode=%sr()'rinfologgingINFOdebugDEBUGZgetEffectiveLevelr%osenvironcopyupdatepop subprocessPopenr'ZSTDOUTPIPEr7close ExceptionZcriticalrr8readlinerstriprZspinwait returncodeZfinishr4r;r-rZwarning ValueErrorreprr#)cmdZ show_stdoutr1Z on_returncodeZextra_ok_returncodesZ command_desc extra_environZ unset_environspinnerZlog_failed_cmdZlog_subprocessZ used_levelZshowing_subprocessZ use_spinnerr9nameprocexcZ all_outputlineZproc_had_errorr3Zexc_msgrrrcall_subprocessus               rZcsdfdd }|S)zProvide a subprocess_runner that shows a spinner message. Intended for use with for pep517's Pep517HookCaller. Thus, the runner has an API that matches what's expected by Pep517HookCaller.subprocess_runner. Nc s(t}t||||dW5QRXdS)N)r1rTrU)r rZ)rSr1rTrUmessagerrrunners z+runner_with_spinner_message..runner)NNr)r\r]rr[rrunner_with_spinner_messages r^) FNr5NNNNNT)(Z __future__rr?rCrHZpip._vendor.six.movesrZpip._internal.exceptionsrZpip._internal.utils.compatrrZpip._internal.utils.loggingrZpip._internal.utils.miscrr Zpip._internal.utils.typingr Zpip._internal.utils.uir typingr r rrrrrrrrZ CommandArgsr/rr%r'r4rZr^rrrrs>      (  ,  site-packages/pip/_internal/utils/__pycache__/virtualenv.cpython-38.pyc000064400000001556151464452710022246 0ustar00U .e{@s,ddlZddlZddlZddZddZdS)NcCs*ttdrdStjttdtjkr&dSdS)zM Return True if we're running inside a virtualenv, False otherwise. Z real_prefixT base_prefixF)hasattrsysprefixgetattrrrB/usr/lib/python3.8/site-packages/pip/_internal/utils/virtualenv.pyrunning_under_virtualenvs  r cCsBtjtjtj}tj|d}tr:tj|r:dSdSdS)z? Return True if in a venv and no system site packages. zno-global-site-packages.txtTFN) ospathdirnameabspathsite__file__joinr isfile)Z site_mod_dirZno_global_filerrrvirtualenv_no_globals r)Zos.pathr rrr rrrrrssite-packages/pip/_internal/utils/__pycache__/__init__.cpython-38.pyc000064400000000233151464452710021575 0ustar00U .e@sdS)Nrrr@/usr/lib/python3.8/site-packages/pip/_internal/utils/__init__.pysite-packages/pip/_internal/utils/__pycache__/filetypes.cpython-38.pyc000064400000001054151464452710022044 0ustar00U .e;@sLdZddlmZer ddlmZdZdZdZdefZdZ eee eZ d S) zFiletype information. )MYPY_CHECK_RUNNING)Tuplez.whl)z.tar.bz2z.tbz)z.tar.xzz.txzz.tlzz.tar.lzz .tar.lzmaz.zip)z.tar.gzz.tgzz.tarN) __doc__Zpip._internal.utils.typingrtypingrZWHEEL_EXTENSIONZBZ2_EXTENSIONSZ XZ_EXTENSIONSZZIP_EXTENSIONSZTAR_EXTENSIONSZARCHIVE_EXTENSIONSrrA/usr/lib/python3.8/site-packages/pip/_internal/utils/filetypes.pys  site-packages/pip/_internal/utils/__pycache__/packaging.cpython-38.pyc000064400000005060151464452710021765 0ustar00U .e @sddlmZddlZddlmZddlmZddlmZm Z ddl m Z ddl m Z ddlmZerdd lmZmZdd lmZdd lmZeeZd d ZddZddZddZdS))absolute_importN) FeedParser) pkg_resources) specifiersversion)NoneMetadataError) display_path)MYPY_CHECK_RUNNING)OptionalTuple)Message) DistributioncCs4|dkr dSt|}tdtt|}||kS)a Check if the given Python version matches a "Requires-Python" specifier. :param version_info: A 3-tuple of ints representing a Python major-minor-micro version to check (e.g. `sys.version_info[:3]`). :return: `True` if the given Python version satisfies the requirement. Otherwise, return `False`. :raises InvalidSpecifier: If `requires_python` has an invalid format. NT.)rZ SpecifierSetrparsejoinmapstr)requires_python version_infoZrequires_python_specifierZpython_versionrA/usr/lib/python3.8/site-packages/pip/_internal/utils/packaging.pycheck_requires_pythons  rcCsd}t|tjr&||r&||}n0|dr@d}||}ntdt|jd}|dkrht ||t }| || S)z :raises NoneMetadataError: if the distribution reports `has_metadata()` True but `get_metadata()` returns None. ZMETADATAzPKG-INFOzNo metadata found in %sN) isinstancerZDistInfoDistribution has_metadata get_metadataloggerZwarningrlocationrrZfeedclose)distZ metadata_nameZmetadataZ feed_parserrrrr,s      rcCs&t|}|d}|dk r"t|}|S)z_ Return the "Requires-Python" metadata for a distribution, or None if not present. zRequires-PythonN)rgetr)rZ pkg_info_dictrrrrget_requires_pythonGs  r!cCs2|dr.|dD]}|r|SqdS)NZ INSTALLERr)rZget_metadata_linesstrip)rlinerrr get_installerXs  r$)Z __future__rZloggingZ email.parserrZ pip._vendorrZpip._vendor.packagingrrZpip._internal.exceptionsrZpip._internal.utils.miscrZpip._internal.utils.typingr typingr r Z email.messager Zpip._vendor.pkg_resourcesr Z getLogger__name__rrrr!r$rrrrs         site-packages/pip/_internal/utils/__pycache__/unpacking.cpython-38.opt-1.pyc000064400000014055151464452710022763 0ustar00U .e%@sXdZddlmZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z m Z mZmZddlmZddlmZerddlmZmZmZmZmZeeZee ZzddlZee 7ZWnek re d YnXzddl!Z!ee7ZWn ek re d YnXd d Z"d dZ#ddZ$ddZ%dddZ&ddZ'dddZ(dS)zUtilities related archives. )absolute_importN)InstallationError)BZ2_EXTENSIONSTAR_EXTENSIONS XZ_EXTENSIONSZIP_EXTENSIONS) ensure_dir)MYPY_CHECK_RUNNING)IterableListOptionalTextUnionzbz2 module is not availablezlzma module is not availablecCstd}t||S)zBGet the current umask which involves having to set it temporarily.r)osumask)maskrA/usr/lib/python3.8/site-packages/pip/_internal/utils/unpacking.py current_umask2s  rcCsh|dd}d|krHd|kr4|d|dksd}n8|trRd}n$|drfd}ntd|d }t||}zt d d | D}| D]}|j }|rt |d }tj||}t||sd }t||||zt|j|d|Wntjk rYnX|r4t|q|rz|||Wn>tk r} ztd||j | WYqW5d} ~ XYnXqz||} WnBttfk r} ztd||j | WYqW5d} ~ XYnXttj|t|d} t | | W5QRX| |!|||j"d@rt#|dt$dBqW5| XdS)a Untar the file (with path `filename`) to the destination `location`. All files are written based on system defaults and umask (i.e. permissions are not preserved), except that regular file members with any execute permissions (user, group, or world) have "chmod +x" applied after being written. Note that for windows, any execute changes using os.chmod are no-ops per the python docs. z.gzz.tgzzr:gzzr:bz2zr:xzz.tarrz-Cannot determine compression type for file %szr:*cSsg|] }|jqSrr=).0memberrrr szuntar_file..rzQThe tar file ({}) has a file ({}) trying to install outside target directory ({})rFz/In the tar file %s the member %s is invalid: %sNr*r(r))%rlowerr2rrloggerZwarningtarfiler+r,r!Z getmembersr=rrrr/r%rr1Z data_filterreplaceZLinkOutsideDestinationErrorisdirZissymZ_extract_member ExceptionZ extractfileKeyErrorAttributeErrorr0r6r7utimerBr5r) r.r8rBZtarr;rHr>rr@excrArCrrr untar_files           rTcCstj|}|dks,|ts,t|rDt|||d dnR|dkslt |sl|t t t rxt||ntd|||td|dS)Nzapplication/zipz.whl)r9zapplication/x-gzipzZCannot unpack file %s (downloaded from %s, content-type: %s); cannot detect archive formatz%Cannot determine archive format of {})rrrealpathrJr2rr-Z is_zipfilerDrLZ is_tarfilerrrrTrKZcriticalrr1)r.r8Z content_typerrr unpack_files<     rV)T)N))__doc__Z __future__rZloggingrr6r3rLr-Zpip._internal.exceptionsrZpip._internal.utils.filetypesrrrrZpip._internal.utils.miscrZpip._internal.utils.typingr typingr r r r rZ getLogger__name__rKZSUPPORTED_EXTENSIONSbz2 ImportErrordebugZlzmarrr!r%rDrTrVrrrrsB        4^site-packages/pip/_internal/utils/__pycache__/deprecation.cpython-38.opt-1.pyc000064400000005412151464452710023276 0ustar00U .e @sdZddlmZddlZddlZddlmZddlmZ ddl m Z e rXddl m Z mZdZGd d d eZdadd d Zd dZdddZdS)zN A module that implements tooling to enable easy warnings about deprecations. )absolute_importN)parse) __version__)MYPY_CHECK_RUNNING)AnyOptionalz DEPRECATION: c@s eZdZdS)PipDeprecationWarningN)__name__ __module__ __qualname__r r C/usr/lib/python3.8/site-packages/pip/_internal/utils/deprecation.pyrsrcCsZ|dk r$tdk rVt||||||n2t|trDtd}||nt||||||dS)Nzpip._internal.deprecations)_original_showwarning issubclassrloggingZ getLoggerZwarning)messagecategoryfilenamelinenofilelineZloggerr r r _showwarning!s*   rcCs(tjdtddtdkr$tjatt_dS)NdefaultT)append)warnings simplefilterrr showwarningrr r r r install_warning_logger2srcCsh|tdf|df|df|dfg}ddd|D}|dk rTttt|krTt|tj|td d dS) aHelper to deprecate existing functionality. reason: Textual reason shown to the user about why this functionality has been deprecated. replacement: Textual suggestion shown to the user about what alternative functionality they can use. gone_in: The version of pip does this functionality should get removed in. Raises errors if pip's current version is greater than or equal to this. issue: Issue number on the tracker that would serve as a useful place for users to find related discussion and provide feedback. Always pass replacement, gone_in and issue as keyword arguments for clarity at the call site. z{}z2pip {} will remove support for this functionality.zA possible replacement is {}.zPYou can find discussion regarding this at https://github.com/pypa/pip/issues/{}. css$|]\}}|dk r||VqdS)N)format).0valtemplater r r `szdeprecated..N)r stacklevel)DEPRECATION_MSG_PREFIXjoinrcurrent_versionrrwarn)reasonZ replacementZgone_inZissueZ sentencesrr r r deprecated>s  r+)NN)N)__doc__Z __future__rrrZpip._vendor.packaging.versionrZpiprr(Zpip._internal.utils.typingrtypingrrr&Warningrrrrr+r r r r s      site-packages/pip/_internal/utils/__pycache__/setuptools_build.cpython-38.opt-1.pyc000064400000002352151464452710024401 0ustar00U .e_@s:ddlZddlmZer(ddlmZmZdZdddZdS) N)MYPY_CHECK_RUNNING)ListSequencezimport sys, setuptools, tokenize; sys.argv[0] = {0!r}; __file__={0!r};f=getattr(tokenize, 'open', open)(__file__);code=f.read().replace('\r\n', '\n');f.close();exec(compile(code, __file__, 'exec'))FcCsJtjg}|r|d|dt|g|r8|||rF|d|S)ao Get setuptools command arguments with shim wrapped setup file invocation. :param setup_py_path: The path to setup.py to be wrapped. :param global_options: Additional global options. :param no_user_config: If True, disables personal user configuration. :param unbuffered_output: If True, adds the unbuffered switch to the argument list. z-uz-cz --no-user-cfg)sys executableappendextend_SETUPTOOLS_SHIMformat)Z setup_py_pathZglobal_optionsZno_user_configZunbuffered_outputargsr H/usr/lib/python3.8/site-packages/pip/_internal/utils/setuptools_build.pymake_setuptools_shim_argss   r)NFF)rZpip._internal.utils.typingrtypingrrr rr r r r s   site-packages/pip/_internal/utils/__pycache__/misc.cpython-38.pyc000064400000055526151464452710021010 0ustar00U .ec@sddlmZddlZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl Z ddl m Z ddlmZddlmZddlmZmZddlmZddlmZdd lmZdd lmZdd lm Z dd l!m"Z"m#Z#m$Z$m%Z%dd l&m'Z'm(Z(m)Z)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0m1Z1er8ddlm2Z3n ddlm3Z3e.rddl4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;mZ>ddl?m@Z@edddddddddd d!d"d#d$gZCeDeEZFd%d&ZGd'd(ZHd)d"ZId*d ZJed+d,d-dd/dZKd0d1ZLd2d3ZMd4dZNdd6dZOd7d8ZPd9d:ZQd;dZRdd?ZTd@dZUdAdZVejWfdBdCZXddEdZYdFdZZdGdZ[dHdIZ\dJdKZ]dLdMZ^dNdOZ_dPdQZ`dRdSZadDe)dDd.d.dfdTdUZbdVdWZcdXdYZddZd[Zed\d]ZfGd^d_d_egZhGd`dadae3ZiejjdbdcZkddd!ZldedfZmGdgdhdhegZnddid#ZodjdkZpdldmZqdndoZrddqdrZsdsdtZtdudvZudwdxZvdydzZwd{d|Zxd}d~ZyddZzdd$Z{ddZ|GdddegZ}ddZ~ddZddZddZdS))absolute_importNdeque) pkg_resources)retry)PY2 text_type)input)parse)unquote) __version__) CommandError)distutils_schemeget_major_minor_version site_packages user_site)WINDOWS expanduser stdlib_pkgsstr_to_display)write_delete_marker_file)MYPY_CHECK_RUNNING)running_under_virtualenvvirtualenv_no_global)BytesIO)StringIO) AnyAnyStr ContainerIterableListOptionalTextTupleUnioncast) DistributioncCs|SN)Ztype_valuer(r(}z|jtjkr.W5d}~XYnXdS)z os.path.makedirs without EEXIST.N)r9makedirsOSErrorerrnoZEEXIST)r:er(r(r*r6qs  c CsPz0tjtjd}|dkr(dtjWS|WSWntttfk rJYnXdS)Nr)z __main__.pyz-cz %s -m pippip) r9r:basenamesysargv executableAttributeError TypeError IndexError)progr(r(r*r4{s  i i)Zstop_max_delayZ wait_fixedFcCstj||tddS)N) ignore_errorsonerror)shutilr+rmtree_errorhandler)dirrRr(r(r*r+sc CsXzt|jtj@ }Wnttfk r2YdSX|rRt|tj||dSdS)zOn Windows, the files in .svn are read-only, so when rmtree() tries to remove them, an exception is thrown. We catch that here, remove the read-only attribute, and hopefully continue without problems.N)r9statst_modeS_IWRITEIOErrorrFchmod)funcr:exc_infoZhas_attr_readonlyr(r(r*rUsrUcCsd|dkr dSt|tr|Sz|td}Wn0tk r^trRtd|}nt |}YnX|S)z Convert a bytes (or text) path to text (unicode in Python 2) for display and logging purposes. This function should never error out. Also, this function is mainly needed for Python 2 since in Python 3 str paths are already text. Nstrictzb{!r}) isinstancerdecoderKgetfilesystemencodingUnicodeDecodeErrorrrr?ascii)r:r,r(r(r*path_to_displays  rdcCsttjtj|}tjddkrB|td}|t d}| t tjj rpd|t t d}|S)zTGives the display value for a given path, making it relative to cwd if possible.rreplace.N)r9r:normcaser>rK version_infor`raencodegetdefaultencoding startswithgetcwdseprCr:r(r(r*r,s.bakcCs6d}|}tj||r.|d7}|t|}q||S)z\Figure out the name of a directory to back up the given dir to (adding .bak, .bak2, etc))r9r:existsstr)rVextn extensionr(r(r*r-s cCs2tjddD]}||kr|Sqt||S)NZPIP_EXISTS_ACTION)r9environgetsplitr.)messageoptionsactionr(r(r*ask_path_existss r~cCstjdrtd|dS)z&Raise an error if no input is allowed.Z PIP_NO_INPUTz7No input was expected ($PIP_NO_INPUT set); question: %sN)r9rxry Exceptionr{r(r(r*_check_no_inputs  rcCsFt|t|}|}||krrh)r:Zresolve_symlinksr(r(r*r23s  cCs@t|\}}|dr8|dd|}|dd}||fS)z,Like os.path.splitext, but take off .tar tooz.tarN) posixpathr/rendswith)r:basertr(r(r*r/As  cCsztj|\}}|r.|r.tj|s.t|t||tj|\}}|rv|rvzt|Wntk rtYnXdS)z7Like os.renames(), but handles renaming across devices.N) r9r:rzrrrErTZmove removedirsrF)oldnewheadtailr(r(r*r3Ks  cCsts dS|ttjS)z Return True if path is within sys.prefix, if we're running in a virtualenv. If we're not in a virtualenv, all paths are considered "local." Caution: this function assumes the head of path has been normalized with normalize_path. T)rrlr2rKprefixror(r(r*is_local]s rcCs tt|S)z Return True if given Distribution object is installed locally (i.e. within current virtualenv). Always True if we're not in a virtualenv. )r dist_locationdistr(r(r* dist_is_localls rcCst|ttS)zF Return True if given Distribution is installed in user site. )rrlr2rrr(r(r*dist_in_usersitexsrcCst|ttS)z[ Return True if given Distribution is installed in sysconfig.get_python_lib(). )rrlr2rrr(r(r*dist_in_site_packagessrcCs,tt|}|ttddddS)zf Return True if given Distribution is installed in path matching distutils_scheme layout. rwZpurelibpythonr)r2rrlrrz)rZ norm_pathr(r(r*dist_in_install_paths rcCs6tjD]*}tj||jd}tj|rdSqdS)zC Return True if given Distribution is an editable install. .egg-linkTF)rKr:r9r; project_namer)rZ path_itemegg_linkr(r(r*dist_is_editables   rcs|rt|}ntj}|r tndd|r6ddndd|rLddndd|r^tnd d fd d |DS) a^ Return a list of installed Distribution objects. If ``local_only`` is True (default), only return installations local to the current virtualenv, if in a virtualenv. ``skip`` argument is an iterable of lower-case project names to ignore; defaults to stdlib_pkgs If ``include_editables`` is False, don't report editables. If ``editables_only`` is True , only report editables. If ``user_only`` is True , only report installations in the user site directory. If ``paths`` is set, only report the distributions present at the specified list of locations. cSsdSNTr(dr(r(r* local_testsz/get_installed_distributions..local_testcSsdSrr(rr(r(r* editable_testsz2get_installed_distributions..editable_testcSs t| Sr'rrr(r(r*rscSst|Sr'rrr(r(r*editables_only_testsz8get_installed_distributions..editables_only_testcSsdSrr(rr(r(r*rscSsdSrr(rr(r(r* user_testsz.get_installed_distributions..user_testcs:g|]2}|r|jkr|r|r|r|qSr()key).0rrrrskiprr(r* s z/get_installed_distributions..)r WorkingSet working_setrr)Z local_onlyrZinclude_editablesZeditables_onlyZ user_onlypathsrr(rr*get_installed_distributionss    rcCsxg}tr*|ttsBtrB|tntr8|t|t|D],}tj||jd}tj |rF|SqFdS)a Return the path for the .egg-link file if it exists, otherwise, None. There's 3 scenarios: 1) not in a virtualenv try to find in site.USER_SITE, then site_packages 2) in a no-global virtualenv try to find in site_packages 3) in a yes-global virtualenv try to find in site_packages, then site.USER_SITE (don't look in global location) For #1 and #3, there could be odd cases, where there's an egg-link in 2 locations. This method will just return the first one found. rN) rappendrrrr9r:r;rr)rZsitesZsiteZegglinkr(r(r* egg_link_paths       rcCst|}|rt|St|jS)aO Get the site-packages location of this distribution. Generally this is dist.location, except in the case of develop-installed packages, where dist.location is the source code location, and we want to know where the egg-link file is. The returned location is normalized (in particular, with symlinks removed). )rr2location)rrr(r(r*rs rcGstj|f|dSr')loggerinfo)msgargsr(r(r* write_outputsrcCst|t|dSr')r9rEr)Z build_dirr(r(r*_make_build_dirs rc@s(eZdZdZddZddZddZdS) FakeFilezQWrap a list of lines in an object with readline() to make ConfigParser happy.cCsdd|D|_dS)Ncss|] }|VqdSr'r()rlr(r(r* &sz$FakeFile.__init__.._gen)selflinesr(r(r*__init__%szFakeFile.__init__cCsPz4zt|jWWStk r0|jYWSXWntk rJYdSXdS)Nrw)nextr NameError StopIterationrr(r(r*readline(szFakeFile.readlinecCs|jSr'rrr(r(r*__iter__1szFakeFile.__iter__N)__name__ __module__ __qualname____doc__rrrr(r(r(r*r"s rc@s$eZdZeddZeddZdS) StreamWrappercCs ||_|Sr') orig_stream)clsrr(r(r* from_stream7szStreamWrapper.from_streamcCs|jjSr')rencodingrr(r(r*r=szStreamWrapper.encodingN)rrr classmethodrpropertyrr(r(r(r*r5s rc cs@tt|}tt|t|ztt|VW5tt||XdS)zReturn a context manager used by captured_stdout/stdin/stderr that temporarily replaces the sys stream *stream_name* with a StringIO. Taken from Lib/support/__init__.py in the CPython repo. N)getattrrKsetattrrr)Z stream_nameZ orig_stdoutr(r(r*captured_outputBs  rcCstdS)zCapture the output of sys.stdout: with captured_stdout() as stdout: print('hello') self.assertEqual(stdout.getvalue(), 'hello ') Taken from Lib/support/__init__.py in the CPython repo. stdoutrr(r(r(r*r5Qs cCstdS)z See captured_stdout(). stderrrr(r(r(r*captured_stderr]src@s eZdZdZddZddZdS)cached_propertyzA property that is only computed once per instance and then replaces itself with an ordinary attribute. Deleting the attribute resets the property. Source: https://github.com/bottlepy/bottle/blob/0.11.5/bottle.py#L175 cCst|d|_||_dS)Nr)rrr\)rr\r(r(r*rls zcached_property.__init__cCs(|dkr |S||}|j|jj<|Sr')r\__dict__r)robjrr)r(r(r*__get__pszcached_property.__get__N)rrrrrrr(r(r(r*rdsrcCs4tj|}|dkrt}||}|r0|jSdS)zCGet the installed version of dist_name avoiding pkg_resources cacheN)rZ Requirementr rfindversion)Z dist_namerZreqrr(r(r*r7xs   cCst|dddS)zConsume an iterable at C speed.r)maxlenNr)iteratorr(r(r*consumesrcOs@tt|tt|f|}dd|D}||d<tdd|S)NcSsi|]\}}||qSr(r()rrr)r(r(r* szenum..Zreverse_mappingEnumr()dictziprangerCitemstype)Z sequentialZnamedZenumsreverser(r(r*enumsrcCs*|dkr |Sd|krd|}d||S)z. Build a netloc from a host-port pair N:[{}]z{}:{})r?)Zhostportr(r(r* build_netlocs  rhttpscCs4|ddkr(d|kr(d|kr(d|}d||S)z) Build a full URL from a netloc. rre@[rz{}://{})countr?)netlocschemer(r(r*build_url_from_netlocs rcCst|}t|}|j|jfS)z2 Return the host-port pair from a netloc. )r urllib_parseZurlparseZhostnamer)rurlZparsedr(r(r* parse_netlocs rcCsXd|kr|dfS|dd\}}d|kr6|dd}n|df}tdd|D}||fS)zp Parse out and remove the auth information from a netloc. Returns: (netloc, (username, password)). r)NNrqrNcss"|]}|dkrdnt|VqdSr')urllib_unquote)rxr(r(r*rsz)split_auth_from_netloc..)rsplitrztuple)rauthZ user_passr(r(r*split_auth_from_netlocsrcCsLt|\}\}}|dkr|S|dkr.d}d}nt|}d}dj|||dS)z Replace the sensitive data in a netloc with "****", if it exists. For example: - "user:pass@example.com" returns "user:****@example.com" - "accesstoken@example.com" returns "****@example.com" N****rwz:****z{user}{password}@{netloc})userpasswordr)rrZquoter?)rr r r(r(r* redact_netlocs  r cCs@t|}||j}|j|d|j|j|jf}t|}||fS)aRTransform and replace netloc in a url. transform_netloc is a function taking the netloc and returning a tuple. The first element of this tuple is the new netloc. The entire tuple is returned. Returns a tuple containing the transformed url as item 0 and the original tuple returned by transform_netloc as item 1. r)rZurlsplitrrr:ZqueryZfragmentZ urlunsplit)rZtransform_netlocZpurlZ netloc_tupleZ url_piecesZsurlr(r(r*_transform_urls   r cCst|Sr')rrr(r(r* _get_netlocsrcCs t|fSr')r r r(r(r*_redact_netlocsrcCst|t\}\}}|||fS)z Parse a url into separate netloc, auth, and url with no auth. Returns: (url_without_auth, netloc, (username, password)) r r)rZurl_without_authrrr(r(r*split_auth_netloc_from_urlsrcCst|tdS)z7Return a copy of url with 'username:password@' removed.rrrr(r(r*r8scCst|tdS)z.Replace the password in a given url with ****.r)r rrr(r(r*redact_auth_from_urlsrc@s4eZdZddZddZddZddZd d Zd S) HiddenTextcCs||_||_dSr')secretredacted)rrrr(r(r*r"szHiddenText.__init__cCsdt|S)Nz)r?rsrr(r(r*__repr__+szHiddenText.__repr__cCs|jSr'rrr(r(r*__str__/szHiddenText.__str__cCs t|t|krdS|j|jkS)NF)rrrotherr(r(r*__eq__4szHiddenText.__eq__cCs ||k Sr'r(rr(r(r*__ne__?szHiddenText.__ne__N)rrrrrrrrr(r(r(r*r!s   rcCs t|ddS)Nrr)r)r)r(r(r* hide_valueDsrcCst|}t||dS)Nr)rr)rrr(r(r*hide_urlIsrcCst}dD]P}|dj|d|djtjd|d|djtjddd |iq |oxtoxtjtj d|k}|rtj d d gtj d d}t d d |dS)zProtection of pip.exe from modification on Windows On Windows, any operation modifying pip should be run as: python -m pip ... )rwz.exezpip{ext})rtz pip{}{ext}rz pip{}.{}{ext}Nrertz-mrIrqz3To modify pip, please run the following command: {} ) setaddr?rKrirr9r:rJrLrMr r;)Z modifying_pipZ pip_namesrtZshould_show_use_python_msgZ new_commandr(r(r*(protect_pip_from_modification_on_windowsOs," r#cCstjdk otjS)z!Is this console interactive? N)rKstdinisattyr(r(r(r*is_console_interactivemsr&)F)rp)T)N)r)Z __future__r contextlibrGrioZloggingr9rrTrWrK collectionsrZ pip._vendorrZpip._vendor.retryingrZpip._vendor.sixrrZpip._vendor.six.movesr Zpip._vendor.six.moves.urllibr rZ"pip._vendor.six.moves.urllib.parser rrIr Zpip._internal.exceptionsr Zpip._internal.locationsrrrrZpip._internal.utils.compatrrrrZ pip._internal.utils.marker_filesrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrrrrtypingrrrrr r!r"r#r$r%Zpip._vendor.pkg_resourcesr&intrB__all__Z getLoggerrrr@rDr6r4r+rUrdr,r-r~rr.rrr0r1DEFAULT_BUFFER_SIZErr2r/r3rrrrrrrrrrrobjectrrcontextmanagerrr5rrr7rrrrrrr r rrrr8rrrrr#r&r(r(r(r*s            0      "          E$       #site-packages/pip/_internal/utils/__pycache__/setuptools_build.cpython-38.pyc000064400000002352151464452710023442 0ustar00U .e_@s:ddlZddlmZer(ddlmZmZdZdddZdS) N)MYPY_CHECK_RUNNING)ListSequencezimport sys, setuptools, tokenize; sys.argv[0] = {0!r}; __file__={0!r};f=getattr(tokenize, 'open', open)(__file__);code=f.read().replace('\r\n', '\n');f.close();exec(compile(code, __file__, 'exec'))FcCsJtjg}|r|d|dt|g|r8|||rF|d|S)ao Get setuptools command arguments with shim wrapped setup file invocation. :param setup_py_path: The path to setup.py to be wrapped. :param global_options: Additional global options. :param no_user_config: If True, disables personal user configuration. :param unbuffered_output: If True, adds the unbuffered switch to the argument list. z-uz-cz --no-user-cfg)sys executableappendextend_SETUPTOOLS_SHIMformat)Z setup_py_pathZglobal_optionsZno_user_configZunbuffered_outputargsr H/usr/lib/python3.8/site-packages/pip/_internal/utils/setuptools_build.pymake_setuptools_shim_argss   r)NFF)rZpip._internal.utils.typingrtypingrrr rr r r r s   site-packages/pip/_internal/utils/__pycache__/filesystem.cpython-38.pyc000064400000005325151464452710022231 0ustar00U .e @sddlZddlZddlZddlZddlmZddlmZddlm Z ddl m Z ddl m Z ddlmZddlmZerdd lmZmZGd d d eZd d ZddZddZeddZe dddZe reddZn eejZdS)N)contextmanager)NamedTemporaryFile)retry)PY2) get_path_uid)cast)MYPY_CHECK_RUNNING)BinaryIOIteratorc@seZdZeddZdS)NamedTemporaryFileResultcCsdSN)selfr r B/usr/lib/python3.8/site-packages/pip/_internal/utils/filesystem.pyfileszNamedTemporaryFileResult.fileN)__name__ __module__ __qualname__propertyrr r r rr sr cCsttdsdSd}||krtj|rntdkr^z t|}Wntk rTYdSX|dkSt|tjSq|tj |}}qdS)NgeteuidTrF) hasattrospathlexistsrrOSErroraccessW_OKdirname)rZpreviousZpath_uidr r rcheck_path_owners    rc Csrzt||Wn\ttfk rl||fD]8}z t|}Wntk rPYq,X|r,td|q,YnXdS)zWrap shutil.copy2() but map errors copying socket files to SpecialFileError as expected. See also https://bugs.python.org/issue37700. z`%s` is a socketN)shutilZcopy2rIOError is_socketZSpecialFileError)srcdestfZis_socket_filer r r copy2_fixed5s  r%cCstt|jSr )statS_ISSOCKrlstatst_mode)rr r rr!Msr!c csbtdtj|tj|dd8}td|}z |VW5|jt|j XW5QRXdS)zGiven a path to a file, open a temp file next to it securely and ensure it is written to disk after the context reaches its end. Fz.tmp)deletedirprefixsuffixr N) rrrrbasenamerrflushfsyncfileno)rr$resultr r radjacent_tmp_fileRs     r3i)Zstop_max_delayZ wait_fixedcCs@zt||Wn*tk r:t|t||YnXdSr )rrenamerremove)r"r#r r rreplaceis  r7)rZos.pathrr& contextlibrZtempfilerZpip._vendor.retryingrZpip._vendor.sixrZpip._internal.utils.compatrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingr r r rr%r!r3Z_replace_retryr7r r r rs.          site-packages/pip/_internal/utils/__pycache__/typing.cpython-38.pyc000064400000002405151464452710021353 0ustar00U .ee@s dZdZdS)aBFor neatly implementing static typing in pip. `mypy` - the static type analysis tool we use - uses the `typing` module, which provides core functionality fundamental to mypy's functioning. Generally, `typing` would be imported at runtime and used in that fashion - it acts as a no-op at runtime and does not have any run-time overhead by design. As it turns out, `typing` is not vendorable - it uses separate sources for Python 2/Python 3. Thus, this codebase can not expect it to be present. To work around this, mypy allows the typing import to be behind a False-y optional to prevent it from running at runtime and type-comments can be used to remove the need for the types to be accessible directly during runtime. This module provides the False-y guard in a nicely named fashion so that a curious maintainer can reach here to read this. In pip, all static-typing related imports should be guarded as follows: from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ... Ref: https://github.com/python/mypy/issues/3216 FN)__doc__ZMYPY_CHECK_RUNNINGrr>/usr/lib/python3.8/site-packages/pip/_internal/utils/typing.pyssite-packages/pip/_internal/utils/__pycache__/appdirs.cpython-38.pyc000064400000017553151464452710021515 0ustar00U .e&&@sdZddlmZddlZddlZddlmZmZddlm Z m Z ddl m Z e r\ddl mZdd Zdd d ZdddZddZddZddZe rzddlZeZWnek reZYnXddZdS)zd This code was taken from https://github.com/ActiveState/appdirs and modified to suit our purposes. )absolute_importN)PY2 text_type)WINDOWS expanduser)MYPY_CHECK_RUNNING)ListcCstr Unix: ~/.cache/ (XDG default) Windows: C:\Users\\AppData\Local\\Cache On Windows the only suggestion in the MSDN docs is that local settings go in the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming app data dir (the default returned by `user_data_dir`). Apps typically put cache data somewhere *under* the given dir here. Some examples: ...\Mozilla\Firefox\Profiles\\Cache ...\Acme\SuperApp\Cache\1.0 OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value. CSIDL_LOCAL_APPDATAZCachedarwinz~/Library/CachesZXDG_CACHE_HOMEz~/.cache)rospathnormpath_get_win_folderr isinstancer_win_path_to_bytesjoinsysplatformrgetenv)appnamer r?/usr/lib/python3.8/site-packages/pip/_internal/utils/appdirs.pyuser_cache_dirs rFcCstr,|r dpd}tjtjt||}ndtjdkrvtjtjt d|rbtjt d|ntjt d|}ntjt dt d|}|S)a Return full path to the user-specific data dir for this application. "appname" is the name of application. If None, just the system directory is returned. "roaming" (boolean, default False) can be set True to use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: macOS: ~/Library/Application Support/ if it exists, else ~/.config/ Unix: ~/.local/share/ # or in $XDG_DATA_HOME, if defined Win XP (not roaming): C:\Documents and Settings\\ ... ...Application Data\ Win XP (roaming): C:\Documents and Settings\\Local ... ...Settings\Application Data\ Win 7 (not roaming): C:\\Users\\AppData\Local\ Win 7 (roaming): C:\\Users\\AppData\Roaming\ For Unix, we follow the XDG spec and support $XDG_DATA_HOME. That means, by default "~/.local/share/". CSIDL_APPDATAr r z~/Library/Application Support/z ~/.config/Z XDG_DATA_HOMEz~/.local/share) rr r rr rrrisdirrr)rroamingconstr rrr user_data_dirHs,    rTcCsHtrt||d}n2tjdkr&t|}ntdtd}tj||}|S)arReturn full path to the user-specific config dir for this application. "appname" is the name of application. If None, just the system directory is returned. "roaming" (boolean, default True) can be set False to not use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: macOS: same as user_data_dir Unix: ~/.config/ Win *: same as user_data_dir For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME. That means, by default "~/.config/". )rr ZXDG_CONFIG_HOMEz ~/.config) rrrrr rrr r)rrr rrruser_config_dir}s  rcstr&tjtd}tj|g}nVtjdkrBtjdg}n:tdd}|rnfdd| tj D}ng}| d|S) aReturn a list of potential user-shared config dirs for this application. "appname" is the name of application. Typical user config directories are: macOS: /Library/Application Support// Unix: /etc or $XDG_CONFIG_DIRS[i]// for each value in $XDG_CONFIG_DIRS Win XP: C:\Documents and Settings\All Users\Application ... ...Data\\ Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) Win 7: Hidden, but writeable on Win 7: C:\ProgramData\\ CSIDL_COMMON_APPDATAr z/Library/Application SupportZXDG_CONFIG_DIRSz/etc/xdgcsg|]}tjt|qSr)r r rr).0xrrr sz$site_config_dirs..z/etc) rr r r rrrrrsplitpathsepappend)rr ZpathlistZxdg_config_dirsrr"rsite_config_dirss     r'cCs:ddl}dddd|}||jd}|||\}}|S)z This is a fallback technique at best. I'm not sure if using the registry for this guarantees us the correct answer for all CSIDL_* names. rNZAppDatazCommon AppDataz Local AppDatarrr z@Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders)_winregOpenKeyHKEY_CURRENT_USERZ QueryValueEx) csidl_namer)Zshell_folder_namekeyZ directoryZ_typerrr_get_win_folder_from_registrysr.cCsdddd|}td}tj}|jd|dd|d}|D]}t|dkr. If encoding using ASCII and MBCS fails, return the original Unicode path. )ASCIIZMBCS)encodeUnicodeEncodeError LookupError)r encodingrrrrs r)F)T)__doc__Z __future__rr rZpip._vendor.sixrrZpip._internal.utils.compatrrZpip._internal.utils.typingrtypingrrrrr'r.r8r3r ImportErrorrrrrrs*   1 5 ") site-packages/pip/_internal/utils/__pycache__/logging.cpython-38.pyc000064400000021707151464452710021475 0ustar00U .e2@sddlmZddlZddlZddlZddlZddlZddlZddlmZm Z ddl m Z ddl m Z ddlmZddlmZz ddlZWnek rddlZYnXzddlmZWnek rdZYnXdd lmZeZeZde_e d ZGd d d eZe r&e rd dZ nddZ ne r6ddZ nddZ ej!d%ddZ"ddZ#Gdddej$Z%ddZ&Gdddej'Z(Gdddej)j*Z+Gdd d eZ,Gd!d"d"eZ-d#d$Z.dS)&)absolute_importN)Filter getLogger)PY2)WINDOWS)DEPRECATION_MSG_PREFIX) ensure_dir)colorama)Forezpip.subprocessorc@seZdZdZdS)BrokenStdoutLoggingErrorzO Raised if BrokenPipeError occurs for the stdout stream while logging. N)__name__ __module__ __qualname____doc__rr?/usr/lib/python3.8/site-packages/pip/_internal/utils/logging.pyr ;sr cCs|tko|jtjtjfkSz1See the docstring for non-Windows Python 3 below.)IOErrorerrnoEINVALEPIPE exc_classexcrrr_is_broken_pipe_errorIsrcCs"|tkp |tko |jtjtjfkSr)BrokenPipeErrorOSErrorrrrrrrrrOscCs|tko|jtjkSr)rrrrrrrrUscCs|tkS)z Return whether an exception is a broken pipe error. Args: exc_class: an exception class. exc: an exception instance. )rrrrrrZsc cs.tj|7_z dVW5tj|8_XdS)zv A context manager which will cause the log output to be indented for any log messages emitted inside it. N) _log_state indentation)Znumrrr indent_loges r cCs ttddS)Nrr)getattrrrrrrget_indentationrsr"cs0eZdZfddZddZfddZZS)IndentingFormattercs$|dd|_tt|j||dS)z A logging.Formatter that obeys the indent_log() context manager. :param add_timestamp: A bool indicating output lines should be prefixed with their record's timestamp. add_timestampFN)popr$superr#__init__)selfargskwargs __class__rrr'xszIndentingFormatter.__init__cCs.|tjkrdS|trdS|tjkr*dSdS)zv Return the start of the formatted log message (not counting the prefix to add to each line). z WARNING: zERROR: )loggingWARNING startswithrERROR)r( formattedlevelnorrrget_message_starts   z$IndentingFormatter.get_message_startcsztt||}|||j}||}d|jrJ||d}d||jfdt7d fdd| dD}|S)z Calls the standard formatter, but will indent all of the log message lines by our current indentation level. r-z%Y-%m-%dT%H:%M:%Sz%s,%03d  csg|] }|qSrr).0lineprefixrr sz-IndentingFormatter.format..T) r&r#formatr4r3r$Z formatTimeZmsecsr"join splitlines)r(recordr2Z message_starttr+r8rr;s zIndentingFormatter.format)r r rr'r4r; __classcell__rrr+rr#vs r#csfdd}|S)Ncsdt|tjjgS)Nr-)r<listr ZStyleZ RESET_ALL)Zinpcolorsrrwrappedsz_color_wrap..wrappedr)rCrDrrBr _color_wraps rEcsheZdZer.ejeejfej eej fgZ ngZ d ddZ ddZ ddZdd Zfd d ZZS) ColorizedStreamHandlerNcCs.tj||||_tr*tr*t|j|_dSN)r. StreamHandlerr' _no_colorrr AnsiToWin32stream)r(rKno_colorrrrr'szColorizedStreamHandler.__init__cCs"trtr|jjtjkS|jtjkS)zA Return whether the handler is using sys.stdout. )rr rKrDsysstdoutr(rrr _using_stdoutsz$ColorizedStreamHandler._using_stdoutcCsXtr |jrdSt|jtjs"|jn|jj}t|dr@|r@dStj ddkrTdSdS)NFisattyTZTERMZANSI) r rI isinstancerKrJrDhasattrrQosenvironget)r(Z real_streamrrr should_colors z#ColorizedStreamHandler.should_colorcCs@tj||}|r<|jD]\}}|j|kr||}qmsglevelZcolorrrrr;s zColorizedStreamHandler.formatcs@tdd\}}|r0|r0t||r0ttt||S)Nr)rMexc_inforPrr r&rF handleError)r(r>rrr+rrr\s  z"ColorizedStreamHandler.handleError)NN)r r rr r.r1rEr ZREDr/ZYELLOWrXr'rPrWr;r\r@rrr+rrFs   rFc@seZdZddZdS)BetterRotatingFileHandlercCs ttj|jtjj|SrG) rrTpathdirnameZ baseFilenamer.handlersRotatingFileHandler_openrOrrrrbszBetterRotatingFileHandler._openN)r r rrbrrrrr]sr]c@seZdZddZddZdS)MaxLevelFiltercCs ||_dSrG)rZ)r(rZrrrr'szMaxLevelFilter.__init__cCs |j|jkSrG)r3rZr(r>rrrfilterszMaxLevelFilter.filterN)r r rr'rerrrrrcsrccs eZdZdZfddZZS)ExcludeLoggerFilterzQ A logging Filter that excludes records from a logger (or its children). cstt|| SrG)r&rfrerdr+rrreszExcludeLoggerFilter.filter)r r rrrer@rrr+rrf srfc Csf|dkrd}n.|dkrd}n |dkr*d}n|dkr8d}nd }tt|}|d k }|r\|}d}nd }|}|d krpdnd}d dd} ddd} dddg|rdgng} tjdddtjddtjddtjddtddtdd d!d"|| d#|| d$d%d&gd'd(d| d#|| d)d%gd'd(|| d#|| d)d*gd'd(d| d+|d d,d-d.|| d/d0d1|iid2|S)3znConfigures and sets up all of the logging Returns the requested logging level, as its integer value. DEBUGr/r1ZCRITICALINFONz /dev/null)rlr1zext://sys.stdoutzext://sys.stderr)rNstderrz2pip._internal.utils.logging.ColorizedStreamHandlerz5pip._internal.utils.logging.BetterRotatingFileHandler)rKfileconsoleconsole_errorsconsole_subprocessuser_logFz*pip._internal.utils.logging.MaxLevelFilter)()rZzlogging.Filter)rsnamez/pip._internal.utils.logging.ExcludeLoggerFilter)exclude_warningsrestrict_to_subprocessexclude_subprocessz %(message)s)rsr;T)rsr;r$)indentindent_with_timestamprKrNrwrurx)rZclassrLrKfilters formatterrmrvrnry)rZrzfilenameZdelayr|)rorprqrr)rZr`z pip._vendorrZ)versionZdisable_existing_loggersr{Z formattersr`rootZloggers)r!r.ZconfigZ dictConfigr/subprocess_loggerrtr#) verbosityrLZ user_log_filerZZ level_numberZinclude_user_logZadditional_log_fileZ root_levelZvendored_log_levelZ log_streamsZhandler_classesr`rrr setup_loggings      $Jr)r)/Z __future__r contextlibrr.Zlogging.handlersrTrMrrZpip._vendor.sixrZpip._internal.utils.compatrZpip._internal.utils.deprecationrZpip._internal.utils.miscrZ threading ImportErrorZdummy_threadingZ pip._vendorr Z _colorama ExceptionZpip._vendor.coloramar Zlocalrrrr rcontextmanagerr r"Z Formatterr#rErHrFr`rar]rcrfrrrrrsT              2K  site-packages/pip/_internal/utils/__pycache__/unpacking.cpython-38.pyc000064400000014055151464452710022024 0ustar00U .e%@sXdZddlmZddlZddlZddlZddlZddlZddlZddl m Z ddl m Z m Z mZmZddlmZddlmZerddlmZmZmZmZmZeeZee ZzddlZee 7ZWnek re d YnXzddl!Z!ee7ZWn ek re d YnXd d Z"d dZ#ddZ$ddZ%dddZ&ddZ'dddZ(dS)zUtilities related archives. )absolute_importN)InstallationError)BZ2_EXTENSIONSTAR_EXTENSIONS XZ_EXTENSIONSZIP_EXTENSIONS) ensure_dir)MYPY_CHECK_RUNNING)IterableListOptionalTextUnionzbz2 module is not availablezlzma module is not availablecCstd}t||S)zBGet the current umask which involves having to set it temporarily.r)osumask)maskrA/usr/lib/python3.8/site-packages/pip/_internal/utils/unpacking.py current_umask2s  rcCsh|dd}d|krHd|kr4|d|dksd}n8|trRd}n$|drfd}ntd|d }t||}zt d d | D}| D]}|j }|rt |d }tj||}t||sd }t||||zt|j|d|Wntjk rYnX|r4t|q|rz|||Wn>tk r} ztd||j | WYqW5d} ~ XYnXqz||} WnBttfk r} ztd||j | WYqW5d} ~ XYnXttj|t|d} t | | W5QRX| |!|||j"d@rt#|dt$dBqW5| XdS)a Untar the file (with path `filename`) to the destination `location`. All files are written based on system defaults and umask (i.e. permissions are not preserved), except that regular file members with any execute permissions (user, group, or world) have "chmod +x" applied after being written. Note that for windows, any execute changes using os.chmod are no-ops per the python docs. z.gzz.tgzzr:gzzr:bz2zr:xzz.tarrz-Cannot determine compression type for file %szr:*cSsg|] }|jqSrr=).0memberrrr szuntar_file..rzQThe tar file ({}) has a file ({}) trying to install outside target directory ({})rFz/In the tar file %s the member %s is invalid: %sNr*r(r))%rlowerr2rrloggerZwarningtarfiler+r,r!Z getmembersr=rrrr/r%rr1Z data_filterreplaceZLinkOutsideDestinationErrorisdirZissymZ_extract_member ExceptionZ extractfileKeyErrorAttributeErrorr0r6r7utimerBr5r) r.r8rBZtarr;rHr>rr@excrArCrrr untar_files           rTcCstj|}|dks,|ts,t|rDt|||d dnR|dkslt |sl|t t t rxt||ntd|||td|dS)Nzapplication/zipz.whl)r9zapplication/x-gzipzZCannot unpack file %s (downloaded from %s, content-type: %s); cannot detect archive formatz%Cannot determine archive format of {})rrrealpathrJr2rr-Z is_zipfilerDrLZ is_tarfilerrrrTrKZcriticalrr1)r.r8Z content_typerrr unpack_files<     rV)T)N))__doc__Z __future__rZloggingrr6r3rLr-Zpip._internal.exceptionsrZpip._internal.utils.filetypesrrrrZpip._internal.utils.miscrZpip._internal.utils.typingr typingr r r r rZ getLogger__name__rKZSUPPORTED_EXTENSIONSbz2 ImportErrordebugZlzmarrr!r%rDrTrVrrrrsB        4^site-packages/pip/_internal/utils/__pycache__/inject_securetransport.cpython-38.opt-1.pyc000064400000001644151464452710025603 0ustar00U .e*@sdZddlZddZedS)a-A helper module that injects SecureTransport, on import. The import should be done as early as possible, to ensure all requests and sessions (or whatever) are created after injecting SecureTransport. Note that we only do the injection on macOS, when the linked OpenSSL is too old to handle TLSv1.2. Nc CsxtjdkrdSz ddl}Wntk r0YdSX|jdkr@dSzddlm}Wnttfk rjYdSX|dS)Ndarwinri)securetransport) sysplatformssl ImportErrorZOPENSSL_VERSION_NUMBERZpip._vendor.urllib3.contribrOSErrorZinject_into_urllib3)rrr N/usr/lib/python3.8/site-packages/pip/_internal/utils/inject_securetransport.pyinject_securetransport s   r )__doc__rr r r r r s site-packages/pip/_internal/utils/__pycache__/hashes.cpython-38.opt-1.pyc000064400000010060151464452710022247 0ustar00U .e@sddlmZddlZddlmZmZmZddlmZm Z m Z ddl m Z ddl mZerddlmZmZmZmZmZddlmZerdd lmZn dd lmZd Zd d d gZGdddeZGdddeZdS))absolute_importN) iteritemsiterkeys itervalues) HashMismatch HashMissingInstallationError) read_chunks)MYPY_CHECK_RUNNING)DictListBinaryIONoReturnIterator)PY3)_Hash)_hashZsha256Zsha384Zsha512c@s^eZdZdZdddZeddZddZd d Zd d Z d dZ ddZ ddZ ddZ dS)HasheszaA wrapper that builds multiple hashes at once and checks them against known-good values NcCs|dkr in||_dS)zo :param hashes: A dict of algorithm names pointing to lists of allowed hex digests N)_allowed)selfhashesr>/usr/lib/python3.8/site-packages/pip/_internal/utils/hashes.py__init__,szHashes.__init__cCstdd|jDS)Ncss|]}t|VqdSN)len).0Zdigestsrrr 7sz&Hashes.digest_count..)sumrvaluesrrrr digest_count4szHashes.digest_countcCs||j|gkS)z/Return whether the given hex digest is allowed.)rget)r hash_nameZ hex_digestrrris_hash_allowed9szHashes.is_hash_allowedc Csi}t|jD]<}zt|||<Wqttfk rHtd|YqXq|D]}t|D]}||q\qPt |D] \}}| |j|krvdSqv| |dS)zCheck good hashes against ones built from iterable of chunks of data. Raise HashMismatch if none match. zUnknown hash name: %sN) rrhashlibnew ValueError TypeErrorrrupdater hexdigest_raise)rZchunksgotsr#chunkhashZgotrrrcheck_against_chunksAs zHashes.check_against_chunkscCst|j|dSr)rrrr,rrrr+Ysz Hashes._raisecCs|t|S)zaCheck good hashes against a file-like object Raise HashMismatch if none match. )r/r )rfilerrrcheck_against_file]szHashes.check_against_filec Cs,t|d}||W5QRSQRXdS)Nrb)openr2)rpathr1rrrcheck_against_pathfs zHashes.check_against_pathcCs t|jS)z,Return whether I know any known-good hashes.)boolrr rrr __nonzero__kszHashes.__nonzero__cCs|Sr)r8r rrr__bool__pszHashes.__bool__)N)__name__ __module__ __qualname____doc__rpropertyr!r$r/r+r2r6r8r9rrrrr's   rcs(eZdZdZfddZddZZS) MissingHasheszA workalike for Hashes used when we're missing a hash for a requirement It computes the actual hash of the requirement and raises a HashMissing exception showing it to the user. cstt|jtgiddS)z!Don't offer the ``hashes`` kwarg.)rN)superr?r FAVORITE_HASHr  __class__rrr|szMissingHashes.__init__cCst|tdSr)rrAr*r0rrrr+szMissingHashes._raise)r:r;r<r=rr+ __classcell__rrrBrr?us r?)Z __future__rr%Zpip._vendor.sixrrrZpip._internal.exceptionsrrrZpip._internal.utils.miscr Zpip._internal.utils.typingr typingr r r rrrrrrAZ STRONG_HASHESobjectrr?rrrrs      Nsite-packages/pip/_internal/utils/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000233151464452710022534 0ustar00U .e@sdS)Nrrr@/usr/lib/python3.8/site-packages/pip/_internal/utils/__init__.pysite-packages/pip/_internal/utils/__pycache__/marker_files.cpython-38.opt-1.pyc000064400000001640151464452710023443 0ustar00U .e7@s$ddlZdZdZddZddZdS)NzThis file is placed here by pip to indicate the source was put here by pip. Once this package is successfully installed this source code will be deleted (unless you remove this file). zpip-delete-this-directory.txtcCstjtj|tS)N)ospathexistsjoinPIP_DELETE_MARKER_FILENAME) directoryrD/usr/lib/python3.8/site-packages/pip/_internal/utils/marker_files.pyhas_delete_marker_filesr c Cs2tj|t}t|d}|tW5QRXdS)z? Write the pip delete marker file into this directory. wN)rrrropenwriteDELETE_MARKER_MESSAGE)rfilepathZ marker_fprrr write_delete_marker_files r)Zos.pathrrrr rrrrr ssite-packages/pip/_internal/utils/__pycache__/encoding.cpython-38.pyc000064400000002337151464452710021633 0ustar00U .e(@sddlZddlZddlZddlZddlmZerDddlmZmZm Z ej dfej dfej dfej dfejdfejd fejd fgZed Zd d ZdS)N)MYPY_CHECK_RUNNING)ListTupleTextzutf-8zutf-16z utf-16-bez utf-16-lezutf-32z utf-32-bez utf-32-lescoding[:=]\s*([-\w.]+)cCstD],\}}||r|t|d|Sq|dddD]D}|dddkrDt|rDt|dd}||SqD|t dpt S) zCheck a bytes string for a BOM to correctly detect the encoding Fallback to locale.getpreferredencoding(False) like open() on Python3N r#asciiF) BOMS startswithlendecodesplit ENCODING_REsearchgroupslocaleZgetpreferredencodingsysgetdefaultencoding)dataZbomencodingliner@/usr/lib/python3.8/site-packages/pip/_internal/utils/encoding.py auto_decodes  r)codecsrrerZpip._internal.utils.typingrtypingrrrBOM_UTF8 BOM_UTF16 BOM_UTF16_BE BOM_UTF16_LE BOM_UTF32 BOM_UTF32_BE BOM_UTF32_LEr compilerrrrrrs   site-packages/pip/_internal/utils/__pycache__/subprocess.cpython-38.pyc000064400000012736151464452710022241 0ustar00U .e& @sddlmZddlZddlZddlZddlmZddlmZddl m Z m Z ddl m Z ddlmZmZddlmZdd lmZerdd lmZmZmZmZmZmZmZmZdd lmZeeeefZ d Z!d dZ"ddZ#ddZ$ddZ%dddZ&ddZ'dS))absolute_importN) shlex_quote)InstallationError)console_to_strstr_to_display)subprocess_logger) HiddenTextpath_to_display)MYPY_CHECK_RUNNING) open_spinner)AnyCallableIterableListMappingOptionalTextUnion)SpinnerInterfacez(----------------------------------------cGs2g}|D]$}t|tr"||q||q|S)z& Create a CommandArgs object. ) isinstancelistextendappend)argsZ command_argsargrB/usr/lib/python3.8/site-packages/pip/_internal/utils/subprocess.py make_commands    rcCsddd|DS)z/ Format command arguments for display.  css,|]$}t|trtt|nt|VqdS)N)rrrstr.0rrrr ;sz&format_command_args..)joinrrrrformat_command_args1s r%cCsdd|DS)z= Return the arguments in their raw, unredacted form. cSs g|]}t|tr|jn|qSr)rrZsecretr rrr Fsz'reveal_command_args..rr$rrrreveal_command_argsAsr'c CsDt|}t|dd}t|}d|}dj|||t||td}|S)z Create and return the error message to use to log a subprocess error with command output. :param lines: A list of lines, each ending with a newline. z command bytes)ZdesczCommand errored out with exit status {exit_status}: command: {command_display} cwd: {cwd_display} Complete output ({line_count} lines): {output}{divider}) exit_statuscommand_display cwd_displayZ line_countoutputZdivider)r%rr r#formatlen LOG_DIVIDER) cmd_argscwdlinesr)Zcommandr*r+r,msgrrrmake_subprocess_output_errorKs    r4FraiseTc  Cs4|dkr g}|dkrg}|r*tj} tj} n tj} tj} t| k} | oN|dk } |dkr`t|}| d|tj }|r| ||D]}| |dqz.t jt|t jt jt j||d}|jWn6tk r}z| rtd||W5d}~XYnXg}t|j}|sqJ|}||d| || r|qz |W5|jrj|jX|jo||j|k}| r|r|dn |d|r*|dkr| s| rt||||jd }t |d !|j|}t"|n:|d krt#d ||j|n|d krnt$dt%|d&|S)a Args: show_stdout: if true, use INFO to log the subprocess's stderr and stdout streams. Otherwise, use DEBUG. Defaults to False. extra_ok_returncodes: an iterable of integer return codes that are acceptable, in addition to 0. Defaults to None, which means []. unset_environ: an iterable of environment variable names to unset prior to calling subprocess.Popen(). log_failed_cmd: if false, failed commands are not logged, only raised. NzRunning command %s)stderrstdinstdoutr1envz#Error %s while executing command %s errorZdoner5)r0r1r2r)zSCommand errored out with exit status {}: {} Check the logs for full command output.warnz$Command "%s" had error code %s in %signorezInvalid value: on_returncode=%sr()'rinfologgingINFOdebugDEBUGZgetEffectiveLevelr%osenvironcopyupdatepop subprocessPopenr'ZSTDOUTPIPEr7close ExceptionZcriticalrr8readlinerstriprZspinwait returncodeZfinishr4r;r-rZwarning ValueErrorreprr#)cmdZ show_stdoutr1Z on_returncodeZextra_ok_returncodesZ command_desc extra_environZ unset_environspinnerZlog_failed_cmdZlog_subprocessZ used_levelZshowing_subprocessZ use_spinnerr9nameprocexcZ all_outputlineZproc_had_errorr3Zexc_msgrrrcall_subprocessus               rZcsdfdd }|S)zProvide a subprocess_runner that shows a spinner message. Intended for use with for pep517's Pep517HookCaller. Thus, the runner has an API that matches what's expected by Pep517HookCaller.subprocess_runner. Nc s(t}t||||dW5QRXdS)N)r1rTrU)r rZ)rSr1rTrUmessagerrrunners z+runner_with_spinner_message..runner)NNr)r\r]rr[rrunner_with_spinner_messages r^) FNr5NNNNNT)(Z __future__rr?rCrHZpip._vendor.six.movesrZpip._internal.exceptionsrZpip._internal.utils.compatrrZpip._internal.utils.loggingrZpip._internal.utils.miscrr Zpip._internal.utils.typingr Zpip._internal.utils.uir typingr r rrrrrrrrZ CommandArgsr/rr%r'r4rZr^rrrrs>      (  ,  site-packages/pip/_internal/utils/__pycache__/hashes.cpython-38.pyc000064400000010060151464452710021310 0ustar00U .e@sddlmZddlZddlmZmZmZddlmZm Z m Z ddl m Z ddl mZerddlmZmZmZmZmZddlmZerdd lmZn dd lmZd Zd d d gZGdddeZGdddeZdS))absolute_importN) iteritemsiterkeys itervalues) HashMismatch HashMissingInstallationError) read_chunks)MYPY_CHECK_RUNNING)DictListBinaryIONoReturnIterator)PY3)_Hash)_hashZsha256Zsha384Zsha512c@s^eZdZdZdddZeddZddZd d Zd d Z d dZ ddZ ddZ ddZ dS)HasheszaA wrapper that builds multiple hashes at once and checks them against known-good values NcCs|dkr in||_dS)zo :param hashes: A dict of algorithm names pointing to lists of allowed hex digests N)_allowed)selfhashesr>/usr/lib/python3.8/site-packages/pip/_internal/utils/hashes.py__init__,szHashes.__init__cCstdd|jDS)Ncss|]}t|VqdSN)len).0Zdigestsrrr 7sz&Hashes.digest_count..)sumrvaluesrrrr digest_count4szHashes.digest_countcCs||j|gkS)z/Return whether the given hex digest is allowed.)rget)r hash_nameZ hex_digestrrris_hash_allowed9szHashes.is_hash_allowedc Csi}t|jD]<}zt|||<Wqttfk rHtd|YqXq|D]}t|D]}||q\qPt |D] \}}| |j|krvdSqv| |dS)zCheck good hashes against ones built from iterable of chunks of data. Raise HashMismatch if none match. zUnknown hash name: %sN) rrhashlibnew ValueError TypeErrorrrupdater hexdigest_raise)rZchunksgotsr#chunkhashZgotrrrcheck_against_chunksAs zHashes.check_against_chunkscCst|j|dSr)rrrr,rrrr+Ysz Hashes._raisecCs|t|S)zaCheck good hashes against a file-like object Raise HashMismatch if none match. )r/r )rfilerrrcheck_against_file]szHashes.check_against_filec Cs,t|d}||W5QRSQRXdS)Nrb)openr2)rpathr1rrrcheck_against_pathfs zHashes.check_against_pathcCs t|jS)z,Return whether I know any known-good hashes.)boolrr rrr __nonzero__kszHashes.__nonzero__cCs|Sr)r8r rrr__bool__pszHashes.__bool__)N)__name__ __module__ __qualname____doc__rpropertyr!r$r/r+r2r6r8r9rrrrr's   rcs(eZdZdZfddZddZZS) MissingHasheszA workalike for Hashes used when we're missing a hash for a requirement It computes the actual hash of the requirement and raises a HashMissing exception showing it to the user. cstt|jtgiddS)z!Don't offer the ``hashes`` kwarg.)rN)superr?r FAVORITE_HASHr  __class__rrr|szMissingHashes.__init__cCst|tdSr)rrAr*r0rrrr+szMissingHashes._raise)r:r;r<r=rr+ __classcell__rrrBrr?us r?)Z __future__rr%Zpip._vendor.sixrrrZpip._internal.exceptionsrrrZpip._internal.utils.miscr Zpip._internal.utils.typingr typingr r r rrrrrrAZ STRONG_HASHESobjectrr?rrrrs      Nsite-packages/pip/_internal/utils/__pycache__/encoding.cpython-38.opt-1.pyc000064400000002337151464452710022572 0ustar00U .e(@sddlZddlZddlZddlZddlmZerDddlmZmZm Z ej dfej dfej dfej dfejdfejd fejd fgZed Zd d ZdS)N)MYPY_CHECK_RUNNING)ListTupleTextzutf-8zutf-16z utf-16-bez utf-16-lezutf-32z utf-32-bez utf-32-lescoding[:=]\s*([-\w.]+)cCstD],\}}||r|t|d|Sq|dddD]D}|dddkrDt|rDt|dd}||SqD|t dpt S) zCheck a bytes string for a BOM to correctly detect the encoding Fallback to locale.getpreferredencoding(False) like open() on Python3N r#asciiF) BOMS startswithlendecodesplit ENCODING_REsearchgroupslocaleZgetpreferredencodingsysgetdefaultencoding)dataZbomencodingliner@/usr/lib/python3.8/site-packages/pip/_internal/utils/encoding.py auto_decodes  r)codecsrrerZpip._internal.utils.typingrtypingrrrBOM_UTF8 BOM_UTF16 BOM_UTF16_BE BOM_UTF16_LE BOM_UTF32 BOM_UTF32_BE BOM_UTF32_LEr compilerrrrrrs   site-packages/pip/_internal/utils/__pycache__/compat.cpython-38.opt-1.pyc000064400000015267151464452710022275 0ustar00U .e]% @s~dZddlmZmZddlZddlZddlZddlZddlZddl Z ddl m Z m Z ddl mZddlmZerddlmZmZmZmZz ddlZWnek rdZYnXeZz ddlZWnVek rzddlmZWn.ek rddlZeje_eje_YnXYnXd d d d d dddddg Z e!e"Z#edk pDeZ$e rddl%Z%z e%j&Z&Wne'k rxdZ&YnXe&dk Z(ndZ(ddl)m&Z&e rddZ*e+de*dZ,ndZ,d/ddZ-dd Z.e rd0dd Z/n d1dd Z/dd Z0e r dd l%m1Z1d!dZ2ndd"l3m4Z4d#dZ2d$d%Z5d&d'd(hZ6e j78d)pRe j7d*koRej9d+kZ:d,dZ;e/usr/lib/python3.8/site-packages/pip/_internal/utils/compat.py Lsz-backslashreplace_decode_fn..css|]}t|VqdSr)ord)rbrrrrNscss|]}d|VqdS)z\x%xNr)rcrrrrOs)rangestartendjoin)rZ raw_bytesrrrbackslashreplace_decode_fnKsr(backslashreplace_decodebackslashreplacecCst|tr|St}|r*t|jdkr.d}z||}WnDtk r|dkrXd}d |}t |||j|t d}YnXt t tdddd}|r|j|d d}||}|S) a For display or logging purposes, convert a bytes object (or text) to text (e.g. unicode in Python 2) safe for output. :param desc: An optional phrase describing the input data, for use in the log message if a warning is logged. Defaults to "Bytes object". This function should never error out and so can take a best effort approach. It is okay to be lossy if needed since the return value is just for display. We assume the data is in the locale preferred encoding. If it won't decode properly, we warn the user but decode as best we can. We also ensure that the output can be safely written to standard output without encoding errors. asciiutf-8Nz Bytes objectz&{} does not appear to be encoded as %s)errors __stderr__encodingr*) isinstancerlocaleZgetpreferredencodingcodecslookupnamedecodeUnicodeDecodeErrorformatloggerZwarningr)getattrsysencode)datadescr/Z decoded_dataZ msg_formatZoutput_encodingZoutput_encodedrrrstr_to_displayYs0     r>cCs t|ddS)z)r<rrrrsFcCst|tr|dS|S)Nr,)r0rr;sreplacerrrrs  cCs"t|tr|d|rdndS|S)Nr,rAstrict)r0bytesr5r?rrrrs cCs`ttdr6t|tjtjB}t|j}t|n&tj |sPt |j}n t d||S)a) Return path's uid. Does not follow symlinks: https://github.com/pypa/pip/pull/935#discussion_r5307003 Placed this function in compat due to differences on AIX and Jython, that should eventually go away. :raises OSError: When path is a symlink or can't be read. O_NOFOLLOWz1%s is a symlink; Will not return uid for symlinks) hasattrosopenO_RDONLYrDfstatst_uidclosepathislinkstatOSError)rLfdZfile_uidrrrrs     get_suffixescCsddtDS)NcSsg|] }|dqS)rr)rsuffixrrr sz*get_extension_suffixes..rQrrrrrsEXTENSION_SUFFIXEScCstSrrUrrrrrscCs0tj|}|dr,|dr,|dd}|S)zm Expand ~ and ~user constructions. Includes a workaround for https://bugs.python.org/issue14768 z~/z//N)rFrL expanduser startswith)rLZexpandedrrrrXs  rXpythonZwsgirefargparsewinZclintcCsNttjdrtj||Stjtj|}tjtj|}||kSdS)z>Provide an alternative for os.path.samefile on Windows/Python2rN)rErFrLrnormcaseabspath)Zfile1Zfile2Zpath1Zpath2rrrrs  cCs ttS) Returns a tuple (x, y) representing the width(x) and the height(y) in characters of the terminal window. )tupleshutilrrrrrrscCsdd}|dp|dp|d}|sbz(tttj}||}t|Wntk r`YnX|stjddtjdd f}t|dt|dfS) r`cSs\z4ddl}ddl}ddl}|d|||jd}Wntk rJYdSX|dkrXdS|S)NrZhhZ12345678)rr)fcntltermiosstruct unpack_fromZioctlZ TIOCGWINSZ Exception)rPrcrdrecrrrr ioctl_GWINSZsz'get_terminal_size..ioctl_GWINSZrrWZLINESZCOLUMNSP) rFrGctermidrHrKrgenvirongetint)rirhrPrrrr s)N)F)F)>__doc__Z __future__rrr2r1ZloggingrFrbr:Zpip._vendor.sixrrZpip._vendor.urllib3.utilrZpip._internal.utils.typingrtypingrr r r Z_ssl ImportErrorZsslr Z pip._vendorZipaddrZ IPAddressZ ip_addressZ IPNetworkZ ip_network__all__Z getLogger__name__r8ZHAS_TLSZimprAttributeErrorr importlib.utilr(register_errorr)r>rrrrRrZimportlib.machineryrVrXrplatformrYr4rrrErrrrrs           B        site-packages/pip/_internal/utils/__pycache__/misc.cpython-38.opt-1.pyc000064400000055526151464452710021747 0ustar00U .ec@sddlmZddlZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl Z ddl m Z ddlmZddlmZddlmZmZddlmZddlmZdd lmZdd lmZdd lm Z dd l!m"Z"m#Z#m$Z$m%Z%dd l&m'Z'm(Z(m)Z)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0m1Z1er8ddlm2Z3n ddlm3Z3e.rddl4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;mZ>ddl?m@Z@edddddddddd d!d"d#d$gZCeDeEZFd%d&ZGd'd(ZHd)d"ZId*d ZJed+d,d-dd/dZKd0d1ZLd2d3ZMd4dZNdd6dZOd7d8ZPd9d:ZQd;dZRdd?ZTd@dZUdAdZVejWfdBdCZXddEdZYdFdZZdGdZ[dHdIZ\dJdKZ]dLdMZ^dNdOZ_dPdQZ`dRdSZadDe)dDd.d.dfdTdUZbdVdWZcdXdYZddZd[Zed\d]ZfGd^d_d_egZhGd`dadae3ZiejjdbdcZkddd!ZldedfZmGdgdhdhegZnddid#ZodjdkZpdldmZqdndoZrddqdrZsdsdtZtdudvZudwdxZvdydzZwd{d|Zxd}d~ZyddZzdd$Z{ddZ|GdddegZ}ddZ~ddZddZddZdS))absolute_importNdeque) pkg_resources)retry)PY2 text_type)input)parse)unquote) __version__) CommandError)distutils_schemeget_major_minor_version site_packages user_site)WINDOWS expanduser stdlib_pkgsstr_to_display)write_delete_marker_file)MYPY_CHECK_RUNNING)running_under_virtualenvvirtualenv_no_global)BytesIO)StringIO) AnyAnyStr ContainerIterableListOptionalTextTupleUnioncast) DistributioncCs|SN)Ztype_valuer(r(}z|jtjkr.W5d}~XYnXdS)z os.path.makedirs without EEXIST.N)r9makedirsOSErrorerrnoZEEXIST)r:er(r(r*r6qs  c CsPz0tjtjd}|dkr(dtjWS|WSWntttfk rJYnXdS)Nr)z __main__.pyz-cz %s -m pippip) r9r:basenamesysargv executableAttributeError TypeError IndexError)progr(r(r*r4{s  i i)Zstop_max_delayZ wait_fixedFcCstj||tddS)N) ignore_errorsonerror)shutilr+rmtree_errorhandler)dirrRr(r(r*r+sc CsXzt|jtj@ }Wnttfk r2YdSX|rRt|tj||dSdS)zOn Windows, the files in .svn are read-only, so when rmtree() tries to remove them, an exception is thrown. We catch that here, remove the read-only attribute, and hopefully continue without problems.N)r9statst_modeS_IWRITEIOErrorrFchmod)funcr:exc_infoZhas_attr_readonlyr(r(r*rUsrUcCsd|dkr dSt|tr|Sz|td}Wn0tk r^trRtd|}nt |}YnX|S)z Convert a bytes (or text) path to text (unicode in Python 2) for display and logging purposes. This function should never error out. Also, this function is mainly needed for Python 2 since in Python 3 str paths are already text. Nstrictzb{!r}) isinstancerdecoderKgetfilesystemencodingUnicodeDecodeErrorrrr?ascii)r:r,r(r(r*path_to_displays  rdcCsttjtj|}tjddkrB|td}|t d}| t tjj rpd|t t d}|S)zTGives the display value for a given path, making it relative to cwd if possible.rreplace.N)r9r:normcaser>rK version_infor`raencodegetdefaultencoding startswithgetcwdseprCr:r(r(r*r,s.bakcCs6d}|}tj||r.|d7}|t|}q||S)z\Figure out the name of a directory to back up the given dir to (adding .bak, .bak2, etc))r9r:existsstr)rVextn extensionr(r(r*r-s cCs2tjddD]}||kr|Sqt||S)NZPIP_EXISTS_ACTION)r9environgetsplitr.)messageoptionsactionr(r(r*ask_path_existss r~cCstjdrtd|dS)z&Raise an error if no input is allowed.Z PIP_NO_INPUTz7No input was expected ($PIP_NO_INPUT set); question: %sN)r9rxry Exceptionr{r(r(r*_check_no_inputs  rcCsFt|t|}|}||krrh)r:Zresolve_symlinksr(r(r*r23s  cCs@t|\}}|dr8|dd|}|dd}||fS)z,Like os.path.splitext, but take off .tar tooz.tarN) posixpathr/rendswith)r:basertr(r(r*r/As  cCsztj|\}}|r.|r.tj|s.t|t||tj|\}}|rv|rvzt|Wntk rtYnXdS)z7Like os.renames(), but handles renaming across devices.N) r9r:rzrrrErTZmove removedirsrF)oldnewheadtailr(r(r*r3Ks  cCsts dS|ttjS)z Return True if path is within sys.prefix, if we're running in a virtualenv. If we're not in a virtualenv, all paths are considered "local." Caution: this function assumes the head of path has been normalized with normalize_path. T)rrlr2rKprefixror(r(r*is_local]s rcCs tt|S)z Return True if given Distribution object is installed locally (i.e. within current virtualenv). Always True if we're not in a virtualenv. )r dist_locationdistr(r(r* dist_is_localls rcCst|ttS)zF Return True if given Distribution is installed in user site. )rrlr2rrr(r(r*dist_in_usersitexsrcCst|ttS)z[ Return True if given Distribution is installed in sysconfig.get_python_lib(). )rrlr2rrr(r(r*dist_in_site_packagessrcCs,tt|}|ttddddS)zf Return True if given Distribution is installed in path matching distutils_scheme layout. rwZpurelibpythonr)r2rrlrrz)rZ norm_pathr(r(r*dist_in_install_paths rcCs6tjD]*}tj||jd}tj|rdSqdS)zC Return True if given Distribution is an editable install. .egg-linkTF)rKr:r9r; project_namer)rZ path_itemegg_linkr(r(r*dist_is_editables   rcs|rt|}ntj}|r tndd|r6ddndd|rLddndd|r^tnd d fd d |DS) a^ Return a list of installed Distribution objects. If ``local_only`` is True (default), only return installations local to the current virtualenv, if in a virtualenv. ``skip`` argument is an iterable of lower-case project names to ignore; defaults to stdlib_pkgs If ``include_editables`` is False, don't report editables. If ``editables_only`` is True , only report editables. If ``user_only`` is True , only report installations in the user site directory. If ``paths`` is set, only report the distributions present at the specified list of locations. cSsdSNTr(dr(r(r* local_testsz/get_installed_distributions..local_testcSsdSrr(rr(r(r* editable_testsz2get_installed_distributions..editable_testcSs t| Sr'rrr(r(r*rscSst|Sr'rrr(r(r*editables_only_testsz8get_installed_distributions..editables_only_testcSsdSrr(rr(r(r*rscSsdSrr(rr(r(r* user_testsz.get_installed_distributions..user_testcs:g|]2}|r|jkr|r|r|r|qSr()key).0rrrrskiprr(r* s z/get_installed_distributions..)r WorkingSet working_setrr)Z local_onlyrZinclude_editablesZeditables_onlyZ user_onlypathsrr(rr*get_installed_distributionss    rcCsxg}tr*|ttsBtrB|tntr8|t|t|D],}tj||jd}tj |rF|SqFdS)a Return the path for the .egg-link file if it exists, otherwise, None. There's 3 scenarios: 1) not in a virtualenv try to find in site.USER_SITE, then site_packages 2) in a no-global virtualenv try to find in site_packages 3) in a yes-global virtualenv try to find in site_packages, then site.USER_SITE (don't look in global location) For #1 and #3, there could be odd cases, where there's an egg-link in 2 locations. This method will just return the first one found. rN) rappendrrrr9r:r;rr)rZsitesZsiteZegglinkr(r(r* egg_link_paths       rcCst|}|rt|St|jS)aO Get the site-packages location of this distribution. Generally this is dist.location, except in the case of develop-installed packages, where dist.location is the source code location, and we want to know where the egg-link file is. The returned location is normalized (in particular, with symlinks removed). )rr2location)rrr(r(r*rs rcGstj|f|dSr')loggerinfo)msgargsr(r(r* write_outputsrcCst|t|dSr')r9rEr)Z build_dirr(r(r*_make_build_dirs rc@s(eZdZdZddZddZddZdS) FakeFilezQWrap a list of lines in an object with readline() to make ConfigParser happy.cCsdd|D|_dS)Ncss|] }|VqdSr'r()rlr(r(r* &sz$FakeFile.__init__.._gen)selflinesr(r(r*__init__%szFakeFile.__init__cCsPz4zt|jWWStk r0|jYWSXWntk rJYdSXdS)Nrw)nextr NameError StopIterationrr(r(r*readline(szFakeFile.readlinecCs|jSr'rrr(r(r*__iter__1szFakeFile.__iter__N)__name__ __module__ __qualname____doc__rrrr(r(r(r*r"s rc@s$eZdZeddZeddZdS) StreamWrappercCs ||_|Sr') orig_stream)clsrr(r(r* from_stream7szStreamWrapper.from_streamcCs|jjSr')rencodingrr(r(r*r=szStreamWrapper.encodingN)rrr classmethodrpropertyrr(r(r(r*r5s rc cs@tt|}tt|t|ztt|VW5tt||XdS)zReturn a context manager used by captured_stdout/stdin/stderr that temporarily replaces the sys stream *stream_name* with a StringIO. Taken from Lib/support/__init__.py in the CPython repo. N)getattrrKsetattrrr)Z stream_nameZ orig_stdoutr(r(r*captured_outputBs  rcCstdS)zCapture the output of sys.stdout: with captured_stdout() as stdout: print('hello') self.assertEqual(stdout.getvalue(), 'hello ') Taken from Lib/support/__init__.py in the CPython repo. stdoutrr(r(r(r*r5Qs cCstdS)z See captured_stdout(). stderrrr(r(r(r*captured_stderr]src@s eZdZdZddZddZdS)cached_propertyzA property that is only computed once per instance and then replaces itself with an ordinary attribute. Deleting the attribute resets the property. Source: https://github.com/bottlepy/bottle/blob/0.11.5/bottle.py#L175 cCst|d|_||_dS)Nr)rrr\)rr\r(r(r*rls zcached_property.__init__cCs(|dkr |S||}|j|jj<|Sr')r\__dict__r)robjrr)r(r(r*__get__pszcached_property.__get__N)rrrrrrr(r(r(r*rdsrcCs4tj|}|dkrt}||}|r0|jSdS)zCGet the installed version of dist_name avoiding pkg_resources cacheN)rZ Requirementr rfindversion)Z dist_namerZreqrr(r(r*r7xs   cCst|dddS)zConsume an iterable at C speed.r)maxlenNr)iteratorr(r(r*consumesrcOs@tt|tt|f|}dd|D}||d<tdd|S)NcSsi|]\}}||qSr(r()rrr)r(r(r* szenum..Zreverse_mappingEnumr()dictziprangerCitemstype)Z sequentialZnamedZenumsreverser(r(r*enumsrcCs*|dkr |Sd|krd|}d||S)z. Build a netloc from a host-port pair N:[{}]z{}:{})r?)Zhostportr(r(r* build_netlocs  rhttpscCs4|ddkr(d|kr(d|kr(d|}d||S)z) Build a full URL from a netloc. rre@[rz{}://{})countr?)netlocschemer(r(r*build_url_from_netlocs rcCst|}t|}|j|jfS)z2 Return the host-port pair from a netloc. )r urllib_parseZurlparseZhostnamer)rurlZparsedr(r(r* parse_netlocs rcCsXd|kr|dfS|dd\}}d|kr6|dd}n|df}tdd|D}||fS)zp Parse out and remove the auth information from a netloc. Returns: (netloc, (username, password)). r)NNrqrNcss"|]}|dkrdnt|VqdSr')urllib_unquote)rxr(r(r*rsz)split_auth_from_netloc..)rsplitrztuple)rauthZ user_passr(r(r*split_auth_from_netlocsrcCsLt|\}\}}|dkr|S|dkr.d}d}nt|}d}dj|||dS)z Replace the sensitive data in a netloc with "****", if it exists. For example: - "user:pass@example.com" returns "user:****@example.com" - "accesstoken@example.com" returns "****@example.com" N****rwz:****z{user}{password}@{netloc})userpasswordr)rrZquoter?)rr r r(r(r* redact_netlocs  r cCs@t|}||j}|j|d|j|j|jf}t|}||fS)aRTransform and replace netloc in a url. transform_netloc is a function taking the netloc and returning a tuple. The first element of this tuple is the new netloc. The entire tuple is returned. Returns a tuple containing the transformed url as item 0 and the original tuple returned by transform_netloc as item 1. r)rZurlsplitrrr:ZqueryZfragmentZ urlunsplit)rZtransform_netlocZpurlZ netloc_tupleZ url_piecesZsurlr(r(r*_transform_urls   r cCst|Sr')rrr(r(r* _get_netlocsrcCs t|fSr')r r r(r(r*_redact_netlocsrcCst|t\}\}}|||fS)z Parse a url into separate netloc, auth, and url with no auth. Returns: (url_without_auth, netloc, (username, password)) r r)rZurl_without_authrrr(r(r*split_auth_netloc_from_urlsrcCst|tdS)z7Return a copy of url with 'username:password@' removed.rrrr(r(r*r8scCst|tdS)z.Replace the password in a given url with ****.r)r rrr(r(r*redact_auth_from_urlsrc@s4eZdZddZddZddZddZd d Zd S) HiddenTextcCs||_||_dSr')secretredacted)rrrr(r(r*r"szHiddenText.__init__cCsdt|S)Nz)r?rsrr(r(r*__repr__+szHiddenText.__repr__cCs|jSr'rrr(r(r*__str__/szHiddenText.__str__cCs t|t|krdS|j|jkS)NF)rrrotherr(r(r*__eq__4szHiddenText.__eq__cCs ||k Sr'r(rr(r(r*__ne__?szHiddenText.__ne__N)rrrrrrrrr(r(r(r*r!s   rcCs t|ddS)Nrr)r)r)r(r(r* hide_valueDsrcCst|}t||dS)Nr)rr)rrr(r(r*hide_urlIsrcCst}dD]P}|dj|d|djtjd|d|djtjddd |iq |oxtoxtjtj d|k}|rtj d d gtj d d}t d d |dS)zProtection of pip.exe from modification on Windows On Windows, any operation modifying pip should be run as: python -m pip ... )rwz.exezpip{ext})rtz pip{}{ext}rz pip{}.{}{ext}Nrertz-mrIrqz3To modify pip, please run the following command: {} ) setaddr?rKrirr9r:rJrLrMr r;)Z modifying_pipZ pip_namesrtZshould_show_use_python_msgZ new_commandr(r(r*(protect_pip_from_modification_on_windowsOs," r#cCstjdk otjS)z!Is this console interactive? N)rKstdinisattyr(r(r(r*is_console_interactivemsr&)F)rp)T)N)r)Z __future__r contextlibrGrioZloggingr9rrTrWrK collectionsrZ pip._vendorrZpip._vendor.retryingrZpip._vendor.sixrrZpip._vendor.six.movesr Zpip._vendor.six.moves.urllibr rZ"pip._vendor.six.moves.urllib.parser rrIr Zpip._internal.exceptionsr Zpip._internal.locationsrrrrZpip._internal.utils.compatrrrrZ pip._internal.utils.marker_filesrZpip._internal.utils.typingrZpip._internal.utils.virtualenvrrrrtypingrrrrr r!r"r#r$r%Zpip._vendor.pkg_resourcesr&intrB__all__Z getLoggerrrr@rDr6r4r+rUrdr,r-r~rr.rrr0r1DEFAULT_BUFFER_SIZErr2r/r3rrrrrrrrrrrobjectrrcontextmanagerrr5rrr7rrrrrrr r rrrr8rrrrr#r&r(r(r(r*s            0      "          E$       #site-packages/pip/_internal/utils/__pycache__/packaging.cpython-38.opt-1.pyc000064400000005060151464452710022724 0ustar00U .e @sddlmZddlZddlmZddlmZddlmZm Z ddl m Z ddl m Z ddlmZerdd lmZmZdd lmZdd lmZeeZd d ZddZddZddZdS))absolute_importN) FeedParser) pkg_resources) specifiersversion)NoneMetadataError) display_path)MYPY_CHECK_RUNNING)OptionalTuple)Message) DistributioncCs4|dkr dSt|}tdtt|}||kS)a Check if the given Python version matches a "Requires-Python" specifier. :param version_info: A 3-tuple of ints representing a Python major-minor-micro version to check (e.g. `sys.version_info[:3]`). :return: `True` if the given Python version satisfies the requirement. Otherwise, return `False`. :raises InvalidSpecifier: If `requires_python` has an invalid format. NT.)rZ SpecifierSetrparsejoinmapstr)requires_python version_infoZrequires_python_specifierZpython_versionrA/usr/lib/python3.8/site-packages/pip/_internal/utils/packaging.pycheck_requires_pythons  rcCsd}t|tjr&||r&||}n0|dr@d}||}ntdt|jd}|dkrht ||t }| || S)z :raises NoneMetadataError: if the distribution reports `has_metadata()` True but `get_metadata()` returns None. ZMETADATAzPKG-INFOzNo metadata found in %sN) isinstancerZDistInfoDistribution has_metadata get_metadataloggerZwarningrlocationrrZfeedclose)distZ metadata_nameZmetadataZ feed_parserrrrr,s      rcCs&t|}|d}|dk r"t|}|S)z_ Return the "Requires-Python" metadata for a distribution, or None if not present. zRequires-PythonN)rgetr)rZ pkg_info_dictrrrrget_requires_pythonGs  r!cCs2|dr.|dD]}|r|SqdS)NZ INSTALLERr)rZget_metadata_linesstrip)rlinerrr get_installerXs  r$)Z __future__rZloggingZ email.parserrZ pip._vendorrZpip._vendor.packagingrrZpip._internal.exceptionsrZpip._internal.utils.miscrZpip._internal.utils.typingr typingr r Z email.messager Zpip._vendor.pkg_resourcesr Z getLogger__name__rrrr!r$rrrrs         site-packages/pip/_internal/utils/__pycache__/glibc.cpython-38.pyc000064400000004321151464452710021120 0ustar00U .e-@sxddlmZddlZddlZddlZddlmZerDddlmZm Z ddZ ddZ d d Z d d Z d dZddZdS))absolute_importN)MYPY_CHECK_RUNNING)OptionalTuplecCs tp tS)z9Returns glibc version string, or None if not using glibc.)glibc_version_string_confstrglibc_version_string_ctypesrr=/usr/lib/python3.8/site-packages/pip/_internal/utils/glibc.pyglibc_version_stringsr c Cs8ztd\}}Wntttfk r2YdSX|S)z@Primary implementation of glibc_version_string using os.confstr.CS_GNU_LIBC_VERSIONN)osconfstrsplitAttributeErrorOSError ValueError)_versionrrr rs rcCsvz ddl}Wntk r"YdSX|d}z |j}Wntk rNYdSX|j|_|}t|tsr| d}|S)z=Fallback implementation of glibc_version_string using ctypes.rNascii) ctypes ImportErrorZCDLLgnu_get_libc_versionrZc_char_pZrestype isinstancestrdecode)rZprocess_namespacer version_strrrr r&s     rcCsHtd|}|s$td|tdSt|d|koFt|d|kS)Nz$(?P[0-9]+)\.(?P[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFmajorminor)rematchwarningswarnRuntimeWarningintgroup)rrequired_major minimum_minormrrr check_glibc_versionFs r(cCst}|dkrdSt|||S)NF)r r()r%r&rrrr have_compatible_glibcWsr)cCst}|dkrdSd|fSdS)zTry to determine the glibc version Returns a tuple of strings (lib, version) which default to empty strings in case the lookup fails. N)r*Zglibc)r )Z glibc_versionrrr libc_verpsr+)Z __future__rr rr Zpip._internal.utils.typingrtypingrrr rrr(r)r+rrrr s   site-packages/pip/_internal/utils/__pycache__/ui.cpython-38.opt-1.pyc000064400000026753151464452710021431 0ustar00U .eR6@sxddlmZmZddlZddlZddlZddlZddlZddlm Z m Z mZddl m Z ddl mZmZddlmZmZmZddlmZddlmZdd lmZdd lmZdd lmZerdd lmZm Z m!Z!zdd l m"Z"Wne#k rdZ"YnXe$e%Z&ddZ'e'eeZ(Gddde)Z*GdddeZ+GdddeZ,Gddde)Z-Gddde)Z.Gddde.e*e-Z/Gddde/e(Z0Gddde/e+Z1Gd d!d!e/eZ2Gd"d#d#e/eZ3Gd$d%d%e/e,Z4Gd&d'd'e.e*e-eZ5e1e1fe0e5fe2e5fe3e5fe4e5fd(Z6d7d)d*Z7ej8d+d,Z9Gd-d.d.e)Z:Gd/d0d0e)Z;Gd1d2d2e;ZdS)8)absolute_importdivisionN)SIGINTdefault_int_handlersignal)six) HIDE_CURSOR SHOW_CURSOR)BarFillingCirclesBarIncrementalBar)Spinner)WINDOWS)get_indentation) format_size)MYPY_CHECK_RUNNING)AnyIteratorIO)coloramacCst|jdd}|s|St|dtt|dtg}|tt|dg7}zt||Wntk rz|YSX|SdS)NencodingZ empty_fillZfillphases)getattrfilerZ text_typelistjoinencodeUnicodeEncodeError)Z preferredZfallbackrZ charactersr:/usr/lib/python3.8/site-packages/pip/_internal/utils/ui.py_select_progress_class%s r cs4eZdZdZfddZfddZddZZS)InterruptibleMixina Helper to ensure that self.finish() gets called on keyboard interrupt. This allows downloads to be interrupted without leaving temporary state (like hidden cursors) behind. This class is similar to the progress library's existing SigIntMixin helper, but as of version 1.2, that helper has the following problems: 1. It calls sys.exit(). 2. It discards the existing SIGINT handler completely. 3. It leaves its own handler in place even after an uninterrupted finish, which will have unexpected delayed effects if the user triggers an unrelated keyboard interrupt some time after a progress-displaying download has already completed, for example. cs4tt|j||tt|j|_|jdkr0t|_dS)z= Save the original SIGINT handler for later. N)superr!__init__rr handle_sigintoriginal_handlerrselfargskwargs __class__rrr#Us zInterruptibleMixin.__init__cstt|tt|jdS)z Restore the original SIGINT handler after finishing. This should happen regardless of whether the progress display finishes normally, or gets interrupted. N)r"r!finishrrr%r'r*rrr,eszInterruptibleMixin.finishcCs||||dS)z Call self.finish() before delegating to the original SIGINT handler. This handler should only be in place while the progress display is active. N)r,r%)r'Zsignumframerrrr$osz InterruptibleMixin.handle_sigint)__name__ __module__ __qualname____doc__r#r,r$ __classcell__rrr*rr!Cs  r!c@seZdZddZdS) SilentBarcCsdSNrr-rrrupdate|szSilentBar.updateN)r/r0r1r6rrrrr4zsr4c@seZdZdZdZdZdZdS) BlueEmojiBar %(percent)d%% )u🔹u🔷u🔵N)r/r0r1suffixZ bar_prefixZ bar_suffixrrrrrr7sr7csJeZdZfddZeddZeddZeddZd d d ZZ S) DownloadProgressMixincs,tt|j||dtd|j|_dS)Nr9)r"r;r#rmessager&r*rrr#szDownloadProgressMixin.__init__cCs t|jSr5)rindexr-rrr downloadedsz DownloadProgressMixin.downloadedcCs |jdkrdStd|jdS)Ngz...z/s)Zavgrr-rrrdownload_speeds z$DownloadProgressMixin.download_speedcCs|jrd|jSdS)Nzeta %s)ZetaZeta_tdr-rrr pretty_etas z DownloadProgressMixin.pretty_etar@ccs&|D]}|V||q|dSr5)nextr,)r'itnxrrriters zDownloadProgressMixin.iter)r@) r/r0r1r#propertyr?rArCrHr3rrr*rr;s    r;cseZdZfddZZS) WindowsMixincs\trjrd_ttj||trXtrXtj_fddj_fddj_ dS)NFcs jjSr5)rwrappedisattyrr-rrz'WindowsMixin.__init__..cs jjSr5)rrKflushrr-rrrMrN) rZ hide_cursorr"rJr#rZ AnsiToWin32rrLrOr&r*r-rr#s zWindowsMixin.__init__)r/r0r1r#r3rrr*rrJsrJc@seZdZejZdZdZdS)BaseDownloadProgressBarr8z0%(downloaded)s %(download_speed)s %(pretty_eta)sN)r/r0r1sysstdoutrr=r:rrrrrPsrPc@s eZdZdS)DefaultDownloadProgressBarNr/r0r1rrrrrSsrSc@s eZdZdS)DownloadSilentBarNrTrrrrrUsrUc@s eZdZdS) DownloadBarNrTrrrrrVsrVc@s eZdZdS)DownloadFillingCirclesBarNrTrrrrrWsrWc@s eZdZdS)DownloadBlueEmojiProgressBarNrTrrrrrXsrXc@s&eZdZejZdZddZddZdS)DownloadProgressSpinnerz!%(downloaded)s %(download_speed)scCs"t|dst|j|_t|jS)N_phaser)hasattr itertoolscyclerrZrDr-rrr next_phases z"DownloadProgressSpinner.next_phasecCsN|j|}|}|j|}d||r*dnd||r6dnd|g}||dS)NrBr9)r=r^r:rZwriteln)r'r=Zphaser:linerrrr6s    zDownloadProgressSpinner.updateN) r/r0r1rQrRrr:r^r6rrrrrYsrY)ZoffZonasciiZprettyZemojicCs8|dks|dkr t|djSt|d|djSdS)Nrr@)max) BAR_TYPESrH)Z progress_barrarrrDownloadProgressProvider srcc csPtr dVn@|r"ttjkr*dVn"|tz dVW5|tXdSr5) rrLloggergetEffectiveLevelloggingINFOwriterr )rrrr hidden_cursors  ric@s$eZdZddZddZddZdS) RateLimitercCs||_d|_dS)Nr)_min_update_interval_seconds _last_update)r'min_update_interval_secondsrrrr#-szRateLimiter.__init__cCst}||j}||jkSr5)timerlrk)r'ZnowZdeltarrrready2s zRateLimiter.readycCst|_dSr5)rnrlr-rrrreset8szRateLimiter.resetN)r/r0r1r#rorprrrrrj,srjc@seZdZddZddZdS)SpinnerInterfacecCs tdSr5NotImplementedErrorr-rrrspin>szSpinnerInterface.spincCs tdSr5rrr'Z final_statusrrrr,BszSpinnerInterface.finishN)r/r0r1rtr,rrrrrq=srqc@s.eZdZd ddZddZdd Zd d ZdS) InteractiveSpinnerN-\|/?cCs\||_|dkrtj}||_t||_d|_t||_ |j dt |jdd|_ dS)NFr9z ... r) _messagerQrR_filerj _rate_limiter _finishedr\r] _spin_cyclerhr_width)r'r=rZ spin_charsrmrrrr#Hs  zInteractiveSpinner.__init__cCsRd|j}|j|d|j||j|t||_|j|jdS)Nr9)r~rzrhlenrOr{rp)r'statusZbackuprrr_writeWs     zInteractiveSpinner._writecCs,|jr dS|jsdS|t|jdSr5)r|r{rorrDr}r-rrrrtcs  zInteractiveSpinner.spincCs4|jr dS|||jd|jd|_dS)N T)r|rrzrhrOrurrrr,ks    zInteractiveSpinner.finish)Nrwrx)r/r0r1r#rrtr,rrrrrvGs   rvc@s.eZdZd ddZddZddZdd Zd S) NonInteractiveSpinner<cCs$||_d|_t||_|ddS)NFZstarted)ryr|rjr{_update)r'r=rmrrrr#zs zNonInteractiveSpinner.__init__cCs|jtd|j|dS)Nz%s: %s)r{rprdinfory)r'rrrrrs zNonInteractiveSpinner._updatecCs&|jr dS|jsdS|ddS)Nzstill running...)r|r{rorr-rrrrts  zNonInteractiveSpinner.spincCs$|jr dS|d|fd|_dS)Nzfinished with status '%s'T)r|rrurrrr,szNonInteractiveSpinner.finishN)r)r/r0r1r#rrtr,rrrrrys rc cstjr"ttjkr"t|}nt|}z t tj |VW5QRXWn>t k rj| dYn*t k r| dYn X| ddS)NZcancelederrorZdone) rQrRrLrdrerfrgrvrriKeyboardInterruptr, Exception)r=Zspinnerrrr open_spinners    r)N)?Z __future__rr contextlibr\rfrQrnrrrZ pip._vendorrZpip._vendor.progressrr Zpip._vendor.progress.barr r r Zpip._vendor.progress.spinnerr Zpip._internal.utils.compatrZpip._internal.utils.loggingrZpip._internal.utils.miscrZpip._internal.utils.typingrtypingrrrrrZ getLoggerr/rdr Z_BaseBarobjectr!r4r7r;rJrPrSrUrVrWrXrYrbrccontextmanagerrirjrqrvrrrrrrs~         7       2site-packages/pip/_internal/utils/__pycache__/logging.cpython-38.opt-1.pyc000064400000021707151464452710022434 0ustar00U .e2@sddlmZddlZddlZddlZddlZddlZddlZddlmZm Z ddl m Z ddl m Z ddlmZddlmZz ddlZWnek rddlZYnXzddlmZWnek rdZYnXdd lmZeZeZde_e d ZGd d d eZe r&e rd dZ nddZ ne r6ddZ nddZ ej!d%ddZ"ddZ#Gdddej$Z%ddZ&Gdddej'Z(Gdddej)j*Z+Gdd d eZ,Gd!d"d"eZ-d#d$Z.dS)&)absolute_importN)Filter getLogger)PY2)WINDOWS)DEPRECATION_MSG_PREFIX) ensure_dir)colorama)Forezpip.subprocessorc@seZdZdZdS)BrokenStdoutLoggingErrorzO Raised if BrokenPipeError occurs for the stdout stream while logging. N)__name__ __module__ __qualname____doc__rr?/usr/lib/python3.8/site-packages/pip/_internal/utils/logging.pyr ;sr cCs|tko|jtjtjfkSz1See the docstring for non-Windows Python 3 below.)IOErrorerrnoEINVALEPIPE exc_classexcrrr_is_broken_pipe_errorIsrcCs"|tkp |tko |jtjtjfkSr)BrokenPipeErrorOSErrorrrrrrrrrOscCs|tko|jtjkSr)rrrrrrrrUscCs|tkS)z Return whether an exception is a broken pipe error. Args: exc_class: an exception class. exc: an exception instance. )rrrrrrZsc cs.tj|7_z dVW5tj|8_XdS)zv A context manager which will cause the log output to be indented for any log messages emitted inside it. N) _log_state indentation)Znumrrr indent_loges r cCs ttddS)Nrr)getattrrrrrrget_indentationrsr"cs0eZdZfddZddZfddZZS)IndentingFormattercs$|dd|_tt|j||dS)z A logging.Formatter that obeys the indent_log() context manager. :param add_timestamp: A bool indicating output lines should be prefixed with their record's timestamp. add_timestampFN)popr$superr#__init__)selfargskwargs __class__rrr'xszIndentingFormatter.__init__cCs.|tjkrdS|trdS|tjkr*dSdS)zv Return the start of the formatted log message (not counting the prefix to add to each line). z WARNING: zERROR: )loggingWARNING startswithrERROR)r( formattedlevelnorrrget_message_starts   z$IndentingFormatter.get_message_startcsztt||}|||j}||}d|jrJ||d}d||jfdt7d fdd| dD}|S)z Calls the standard formatter, but will indent all of the log message lines by our current indentation level. r-z%Y-%m-%dT%H:%M:%Sz%s,%03d  csg|] }|qSrr).0lineprefixrr sz-IndentingFormatter.format..T) r&r#formatr4r3r$Z formatTimeZmsecsr"join splitlines)r(recordr2Z message_starttr+r8rr;s zIndentingFormatter.format)r r rr'r4r; __classcell__rrr+rr#vs r#csfdd}|S)Ncsdt|tjjgS)Nr-)r<listr ZStyleZ RESET_ALL)Zinpcolorsrrwrappedsz_color_wrap..wrappedr)rCrDrrBr _color_wraps rEcsheZdZer.ejeejfej eej fgZ ngZ d ddZ ddZ ddZdd Zfd d ZZS) ColorizedStreamHandlerNcCs.tj||||_tr*tr*t|j|_dSN)r. StreamHandlerr' _no_colorrr AnsiToWin32stream)r(rKno_colorrrrr'szColorizedStreamHandler.__init__cCs"trtr|jjtjkS|jtjkS)zA Return whether the handler is using sys.stdout. )rr rKrDsysstdoutr(rrr _using_stdoutsz$ColorizedStreamHandler._using_stdoutcCsXtr |jrdSt|jtjs"|jn|jj}t|dr@|r@dStj ddkrTdSdS)NFisattyTZTERMZANSI) r rI isinstancerKrJrDhasattrrQosenvironget)r(Z real_streamrrr should_colors z#ColorizedStreamHandler.should_colorcCs@tj||}|r<|jD]\}}|j|kr||}qmsglevelZcolorrrrr;s zColorizedStreamHandler.formatcs@tdd\}}|r0|r0t||r0ttt||S)Nr)rMexc_inforPrr r&rF handleError)r(r>rrr+rrr\s  z"ColorizedStreamHandler.handleError)NN)r r rr r.r1rEr ZREDr/ZYELLOWrXr'rPrWr;r\r@rrr+rrFs   rFc@seZdZddZdS)BetterRotatingFileHandlercCs ttj|jtjj|SrG) rrTpathdirnameZ baseFilenamer.handlersRotatingFileHandler_openrOrrrrbszBetterRotatingFileHandler._openN)r r rrbrrrrr]sr]c@seZdZddZddZdS)MaxLevelFiltercCs ||_dSrG)rZ)r(rZrrrr'szMaxLevelFilter.__init__cCs |j|jkSrG)r3rZr(r>rrrfilterszMaxLevelFilter.filterN)r r rr'rerrrrrcsrccs eZdZdZfddZZS)ExcludeLoggerFilterzQ A logging Filter that excludes records from a logger (or its children). cstt|| SrG)r&rfrerdr+rrreszExcludeLoggerFilter.filter)r r rrrer@rrr+rrf srfc Csf|dkrd}n.|dkrd}n |dkr*d}n|dkr8d}nd }tt|}|d k }|r\|}d}nd }|}|d krpdnd}d dd} ddd} dddg|rdgng} tjdddtjddtjddtjddtddtdd d!d"|| d#|| d$d%d&gd'd(d| d#|| d)d%gd'd(|| d#|| d)d*gd'd(d| d+|d d,d-d.|| d/d0d1|iid2|S)3znConfigures and sets up all of the logging Returns the requested logging level, as its integer value. DEBUGr/r1ZCRITICALINFONz /dev/null)rlr1zext://sys.stdoutzext://sys.stderr)rNstderrz2pip._internal.utils.logging.ColorizedStreamHandlerz5pip._internal.utils.logging.BetterRotatingFileHandler)rKfileconsoleconsole_errorsconsole_subprocessuser_logFz*pip._internal.utils.logging.MaxLevelFilter)()rZzlogging.Filter)rsnamez/pip._internal.utils.logging.ExcludeLoggerFilter)exclude_warningsrestrict_to_subprocessexclude_subprocessz %(message)s)rsr;T)rsr;r$)indentindent_with_timestamprKrNrwrurx)rZclassrLrKfilters formatterrmrvrnry)rZrzfilenameZdelayr|)rorprqrr)rZr`z pip._vendorrZ)versionZdisable_existing_loggersr{Z formattersr`rootZloggers)r!r.ZconfigZ dictConfigr/subprocess_loggerrtr#) verbosityrLZ user_log_filerZZ level_numberZinclude_user_logZadditional_log_fileZ root_levelZvendored_log_levelZ log_streamsZhandler_classesr`rrr setup_loggings      $Jr)r)/Z __future__r contextlibrr.Zlogging.handlersrTrMrrZpip._vendor.sixrZpip._internal.utils.compatrZpip._internal.utils.deprecationrZpip._internal.utils.miscrZ threading ImportErrorZdummy_threadingZ pip._vendorr Z _colorama ExceptionZpip._vendor.coloramar Zlocalrrrr rcontextmanagerr r"Z Formatterr#rErHrFr`rar]rcrfrrrrrsT              2K  site-packages/pip/_internal/utils/__pycache__/filetypes.cpython-38.opt-1.pyc000064400000001054151464452710023003 0ustar00U .e;@sLdZddlmZer ddlmZdZdZdZdefZdZ eee eZ d S) zFiletype information. )MYPY_CHECK_RUNNING)Tuplez.whl)z.tar.bz2z.tbz)z.tar.xzz.txzz.tlzz.tar.lzz .tar.lzmaz.zip)z.tar.gzz.tgzz.tarN) __doc__Zpip._internal.utils.typingrtypingrZWHEEL_EXTENSIONZBZ2_EXTENSIONSZ XZ_EXTENSIONSZZIP_EXTENSIONSZTAR_EXTENSIONSZARCHIVE_EXTENSIONSrrA/usr/lib/python3.8/site-packages/pip/_internal/utils/filetypes.pys  site-packages/pip/_internal/utils/__pycache__/models.cpython-38.opt-1.pyc000064400000003604151464452710022265 0ustar00U .e|@s dZddlZGdddeZdS)zUtilities for defining models Nc@sXeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ dS)KeyBasedCompareMixinz/usr/lib/python3.8/site-packages/pip/_internal/utils/models.py__init__ szKeyBasedCompareMixin.__init__cCs t|jSr)hashr)rrrr __hash__szKeyBasedCompareMixin.__hash__cCs||tjSr)_compareoperator__lt__rotherrrr rszKeyBasedCompareMixin.__lt__cCs||tjSr)r r__le__rrrr rszKeyBasedCompareMixin.__le__cCs||tjSr)r r__gt__rrrr rszKeyBasedCompareMixin.__gt__cCs||tjSr)r r__ge__rrrr rszKeyBasedCompareMixin.__ge__cCs||tjSr)r r__eq__rrrr r szKeyBasedCompareMixin.__eq__cCs||tjSr)r r__ne__rrrr r#szKeyBasedCompareMixin.__ne__cCst||jstS||j|jSr) isinstancerNotImplementedr)rrmethodrrr r &s zKeyBasedCompareMixin._compareN) __name__ __module__ __qualname____doc__r r rrrrrrr rrrr r sr)rrobjectrrrrr ssite-packages/pip/_internal/utils/hashes.py000064400000007664151464452710015042 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import hashlib from pip._vendor.six import iteritems, iterkeys, itervalues from pip._internal.exceptions import ( HashMismatch, HashMissing, InstallationError, ) from pip._internal.utils.misc import read_chunks from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import ( Dict, List, BinaryIO, NoReturn, Iterator ) from pip._vendor.six import PY3 if PY3: from hashlib import _Hash else: from hashlib import _hash as _Hash # The recommended hash algo of the moment. Change this whenever the state of # the art changes; it won't hurt backward compatibility. FAVORITE_HASH = 'sha256' # Names of hashlib algorithms allowed by the --hash option and ``pip hash`` # Currently, those are the ones at least as collision-resistant as sha256. STRONG_HASHES = ['sha256', 'sha384', 'sha512'] class Hashes(object): """A wrapper that builds multiple hashes at once and checks them against known-good values """ def __init__(self, hashes=None): # type: (Dict[str, List[str]]) -> None """ :param hashes: A dict of algorithm names pointing to lists of allowed hex digests """ self._allowed = {} if hashes is None else hashes @property def digest_count(self): # type: () -> int return sum(len(digests) for digests in self._allowed.values()) def is_hash_allowed( self, hash_name, # type: str hex_digest, # type: str ): """Return whether the given hex digest is allowed.""" return hex_digest in self._allowed.get(hash_name, []) def check_against_chunks(self, chunks): # type: (Iterator[bytes]) -> None """Check good hashes against ones built from iterable of chunks of data. Raise HashMismatch if none match. """ gots = {} for hash_name in iterkeys(self._allowed): try: gots[hash_name] = hashlib.new(hash_name) except (ValueError, TypeError): raise InstallationError('Unknown hash name: %s' % hash_name) for chunk in chunks: for hash in itervalues(gots): hash.update(chunk) for hash_name, got in iteritems(gots): if got.hexdigest() in self._allowed[hash_name]: return self._raise(gots) def _raise(self, gots): # type: (Dict[str, _Hash]) -> NoReturn raise HashMismatch(self._allowed, gots) def check_against_file(self, file): # type: (BinaryIO) -> None """Check good hashes against a file-like object Raise HashMismatch if none match. """ return self.check_against_chunks(read_chunks(file)) def check_against_path(self, path): # type: (str) -> None with open(path, 'rb') as file: return self.check_against_file(file) def __nonzero__(self): # type: () -> bool """Return whether I know any known-good hashes.""" return bool(self._allowed) def __bool__(self): # type: () -> bool return self.__nonzero__() class MissingHashes(Hashes): """A workalike for Hashes used when we're missing a hash for a requirement It computes the actual hash of the requirement and raises a HashMissing exception showing it to the user. """ def __init__(self): # type: () -> None """Don't offer the ``hashes`` kwarg.""" # Pass our favorite hash in to generate a "gotten hash". With the # empty list, it will never match, so an error will always raise. super(MissingHashes, self).__init__(hashes={FAVORITE_HASH: []}) def _raise(self, gots): # type: (Dict[str, _Hash]) -> NoReturn raise HashMissing(gots[FAVORITE_HASH].hexdigest()) site-packages/pip/_internal/utils/logging.py000064400000031351151464452710015203 0ustar00# The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import contextlib import errno import logging import logging.handlers import os import sys from logging import Filter, getLogger from pip._vendor.six import PY2 from pip._internal.utils.compat import WINDOWS from pip._internal.utils.deprecation import DEPRECATION_MSG_PREFIX from pip._internal.utils.misc import ensure_dir try: import threading except ImportError: import dummy_threading as threading # type: ignore try: # Use "import as" and set colorama in the else clause to avoid mypy # errors and get the following correct revealed type for colorama: # `Union[_importlib_modulespec.ModuleType, None]` # Otherwise, we get an error like the following in the except block: # > Incompatible types in assignment (expression has type "None", # variable has type Module) # TODO: eliminate the need to use "import as" once mypy addresses some # of its issues with conditional imports. Here is an umbrella issue: # https://github.com/python/mypy/issues/1297 from pip._vendor import colorama as _colorama # Lots of different errors can come from this, including SystemError and # ImportError. except Exception: colorama = None else: # Import Fore explicitly rather than accessing below as colorama.Fore # to avoid the following error running mypy: # > Module has no attribute "Fore" # TODO: eliminate the need to import Fore once mypy addresses some of its # issues with conditional imports. This particular case could be an # instance of the following issue (but also see the umbrella issue above): # https://github.com/python/mypy/issues/3500 from pip._vendor.colorama import Fore colorama = _colorama _log_state = threading.local() _log_state.indentation = 0 subprocess_logger = getLogger('pip.subprocessor') class BrokenStdoutLoggingError(Exception): """ Raised if BrokenPipeError occurs for the stdout stream while logging. """ pass # BrokenPipeError does not exist in Python 2 and, in addition, manifests # differently in Windows and non-Windows. if WINDOWS: # In Windows, a broken pipe can show up as EINVAL rather than EPIPE: # https://bugs.python.org/issue19612 # https://bugs.python.org/issue30418 if PY2: def _is_broken_pipe_error(exc_class, exc): """See the docstring for non-Windows Python 3 below.""" return (exc_class is IOError and exc.errno in (errno.EINVAL, errno.EPIPE)) else: # In Windows, a broken pipe IOError became OSError in Python 3. def _is_broken_pipe_error(exc_class, exc): """See the docstring for non-Windows Python 3 below.""" return ((exc_class is BrokenPipeError) or # noqa: F821 (exc_class is OSError and exc.errno in (errno.EINVAL, errno.EPIPE))) elif PY2: def _is_broken_pipe_error(exc_class, exc): """See the docstring for non-Windows Python 3 below.""" return (exc_class is IOError and exc.errno == errno.EPIPE) else: # Then we are in the non-Windows Python 3 case. def _is_broken_pipe_error(exc_class, exc): """ Return whether an exception is a broken pipe error. Args: exc_class: an exception class. exc: an exception instance. """ return (exc_class is BrokenPipeError) # noqa: F821 @contextlib.contextmanager def indent_log(num=2): """ A context manager which will cause the log output to be indented for any log messages emitted inside it. """ _log_state.indentation += num try: yield finally: _log_state.indentation -= num def get_indentation(): return getattr(_log_state, 'indentation', 0) class IndentingFormatter(logging.Formatter): def __init__(self, *args, **kwargs): """ A logging.Formatter that obeys the indent_log() context manager. :param add_timestamp: A bool indicating output lines should be prefixed with their record's timestamp. """ self.add_timestamp = kwargs.pop("add_timestamp", False) super(IndentingFormatter, self).__init__(*args, **kwargs) def get_message_start(self, formatted, levelno): """ Return the start of the formatted log message (not counting the prefix to add to each line). """ if levelno < logging.WARNING: return '' if formatted.startswith(DEPRECATION_MSG_PREFIX): # Then the message already has a prefix. We don't want it to # look like "WARNING: DEPRECATION: ...." return '' if levelno < logging.ERROR: return 'WARNING: ' return 'ERROR: ' def format(self, record): """ Calls the standard formatter, but will indent all of the log message lines by our current indentation level. """ formatted = super(IndentingFormatter, self).format(record) message_start = self.get_message_start(formatted, record.levelno) formatted = message_start + formatted prefix = '' if self.add_timestamp: # TODO: Use Formatter.default_time_format after dropping PY2. t = self.formatTime(record, "%Y-%m-%dT%H:%M:%S") prefix = '%s,%03d ' % (t, record.msecs) prefix += " " * get_indentation() formatted = "".join([ prefix + line for line in formatted.splitlines(True) ]) return formatted def _color_wrap(*colors): def wrapped(inp): return "".join(list(colors) + [inp, colorama.Style.RESET_ALL]) return wrapped class ColorizedStreamHandler(logging.StreamHandler): # Don't build up a list of colors if we don't have colorama if colorama: COLORS = [ # This needs to be in order from highest logging level to lowest. (logging.ERROR, _color_wrap(Fore.RED)), (logging.WARNING, _color_wrap(Fore.YELLOW)), ] else: COLORS = [] def __init__(self, stream=None, no_color=None): logging.StreamHandler.__init__(self, stream) self._no_color = no_color if WINDOWS and colorama: self.stream = colorama.AnsiToWin32(self.stream) def _using_stdout(self): """ Return whether the handler is using sys.stdout. """ if WINDOWS and colorama: # Then self.stream is an AnsiToWin32 object. return self.stream.wrapped is sys.stdout return self.stream is sys.stdout def should_color(self): # Don't colorize things if we do not have colorama or if told not to if not colorama or self._no_color: return False real_stream = ( self.stream if not isinstance(self.stream, colorama.AnsiToWin32) else self.stream.wrapped ) # If the stream is a tty we should color it if hasattr(real_stream, "isatty") and real_stream.isatty(): return True # If we have an ANSI term we should color it if os.environ.get("TERM") == "ANSI": return True # If anything else we should not color it return False def format(self, record): msg = logging.StreamHandler.format(self, record) if self.should_color(): for level, color in self.COLORS: if record.levelno >= level: msg = color(msg) break return msg # The logging module says handleError() can be customized. def handleError(self, record): exc_class, exc = sys.exc_info()[:2] # If a broken pipe occurred while calling write() or flush() on the # stdout stream in logging's Handler.emit(), then raise our special # exception so we can handle it in main() instead of logging the # broken pipe error and continuing. if (exc_class and self._using_stdout() and _is_broken_pipe_error(exc_class, exc)): raise BrokenStdoutLoggingError() return super(ColorizedStreamHandler, self).handleError(record) class BetterRotatingFileHandler(logging.handlers.RotatingFileHandler): def _open(self): ensure_dir(os.path.dirname(self.baseFilename)) return logging.handlers.RotatingFileHandler._open(self) class MaxLevelFilter(Filter): def __init__(self, level): self.level = level def filter(self, record): return record.levelno < self.level class ExcludeLoggerFilter(Filter): """ A logging Filter that excludes records from a logger (or its children). """ def filter(self, record): # The base Filter class allows only records from a logger (or its # children). return not super(ExcludeLoggerFilter, self).filter(record) def setup_logging(verbosity, no_color, user_log_file): """Configures and sets up all of the logging Returns the requested logging level, as its integer value. """ # Determine the level to be logging at. if verbosity >= 1: level = "DEBUG" elif verbosity == -1: level = "WARNING" elif verbosity == -2: level = "ERROR" elif verbosity <= -3: level = "CRITICAL" else: level = "INFO" level_number = getattr(logging, level) # The "root" logger should match the "console" level *unless* we also need # to log to a user log file. include_user_log = user_log_file is not None if include_user_log: additional_log_file = user_log_file root_level = "DEBUG" else: additional_log_file = "/dev/null" root_level = level # Disable any logging besides WARNING unless we have DEBUG level logging # enabled for vendored libraries. vendored_log_level = "WARNING" if level in ["INFO", "ERROR"] else "DEBUG" # Shorthands for clarity log_streams = { "stdout": "ext://sys.stdout", "stderr": "ext://sys.stderr", } handler_classes = { "stream": "pip._internal.utils.logging.ColorizedStreamHandler", "file": "pip._internal.utils.logging.BetterRotatingFileHandler", } handlers = ["console", "console_errors", "console_subprocess"] + ( ["user_log"] if include_user_log else [] ) logging.config.dictConfig({ "version": 1, "disable_existing_loggers": False, "filters": { "exclude_warnings": { "()": "pip._internal.utils.logging.MaxLevelFilter", "level": logging.WARNING, }, "restrict_to_subprocess": { "()": "logging.Filter", "name": subprocess_logger.name, }, "exclude_subprocess": { "()": "pip._internal.utils.logging.ExcludeLoggerFilter", "name": subprocess_logger.name, }, }, "formatters": { "indent": { "()": IndentingFormatter, "format": "%(message)s", }, "indent_with_timestamp": { "()": IndentingFormatter, "format": "%(message)s", "add_timestamp": True, }, }, "handlers": { "console": { "level": level, "class": handler_classes["stream"], "no_color": no_color, "stream": log_streams["stdout"], "filters": ["exclude_subprocess", "exclude_warnings"], "formatter": "indent", }, "console_errors": { "level": "WARNING", "class": handler_classes["stream"], "no_color": no_color, "stream": log_streams["stderr"], "filters": ["exclude_subprocess"], "formatter": "indent", }, # A handler responsible for logging to the console messages # from the "subprocessor" logger. "console_subprocess": { "level": level, "class": handler_classes["stream"], "no_color": no_color, "stream": log_streams["stderr"], "filters": ["restrict_to_subprocess"], "formatter": "indent", }, "user_log": { "level": "DEBUG", "class": handler_classes["file"], "filename": additional_log_file, "delay": True, "formatter": "indent_with_timestamp", }, }, "root": { "level": root_level, "handlers": handlers, }, "loggers": { "pip._vendor": { "level": vendored_log_level } }, }) return level_number site-packages/pip/_internal/utils/filesystem.py000064400000006406151464452710015744 0ustar00import os import os.path import shutil import stat from contextlib import contextmanager from tempfile import NamedTemporaryFile # NOTE: retrying is not annotated in typeshed as on 2017-07-17, which is # why we ignore the type on this import. from pip._vendor.retrying import retry # type: ignore from pip._vendor.six import PY2 from pip._internal.utils.compat import get_path_uid from pip._internal.utils.misc import cast from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import BinaryIO, Iterator class NamedTemporaryFileResult(BinaryIO): @property def file(self): # type: () -> BinaryIO pass def check_path_owner(path): # type: (str) -> bool # If we don't have a way to check the effective uid of this process, then # we'll just assume that we own the directory. if not hasattr(os, "geteuid"): return True previous = None while path != previous: if os.path.lexists(path): # Check if path is writable by current user. if os.geteuid() == 0: # Special handling for root user in order to handle properly # cases where users use sudo without -H flag. try: path_uid = get_path_uid(path) except OSError: return False return path_uid == 0 else: return os.access(path, os.W_OK) else: previous, path = path, os.path.dirname(path) return False # assume we don't own the path def copy2_fixed(src, dest): # type: (str, str) -> None """Wrap shutil.copy2() but map errors copying socket files to SpecialFileError as expected. See also https://bugs.python.org/issue37700. """ try: shutil.copy2(src, dest) except (OSError, IOError): for f in [src, dest]: try: is_socket_file = is_socket(f) except OSError: # An error has already occurred. Another error here is not # a problem and we can ignore it. pass else: if is_socket_file: raise shutil.SpecialFileError("`%s` is a socket" % f) raise def is_socket(path): # type: (str) -> bool return stat.S_ISSOCK(os.lstat(path).st_mode) @contextmanager def adjacent_tmp_file(path): # type: (str) -> Iterator[NamedTemporaryFileResult] """Given a path to a file, open a temp file next to it securely and ensure it is written to disk after the context reaches its end. """ with NamedTemporaryFile( delete=False, dir=os.path.dirname(path), prefix=os.path.basename(path), suffix='.tmp', ) as f: result = cast('NamedTemporaryFileResult', f) try: yield result finally: result.file.flush() os.fsync(result.file.fileno()) _replace_retry = retry(stop_max_delay=1000, wait_fixed=250) if PY2: @_replace_retry def replace(src, dest): # type: (str, str) -> None try: os.rename(src, dest) except OSError: os.remove(dest) os.rename(src, dest) else: replace = _replace_retry(os.replace) site-packages/pip/_internal/utils/appdirs.py000064400000023046151464452710015221 0ustar00""" This code was taken from https://github.com/ActiveState/appdirs and modified to suit our purposes. """ # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import import os import sys from pip._vendor.six import PY2, text_type from pip._internal.utils.compat import WINDOWS, expanduser from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import List def user_cache_dir(appname): # type: (str) -> str r""" Return full path to the user-specific cache dir for this application. "appname" is the name of application. Typical user cache directories are: macOS: ~/Library/Caches/ Unix: ~/.cache/ (XDG default) Windows: C:\Users\\AppData\Local\\Cache On Windows the only suggestion in the MSDN docs is that local settings go in the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming app data dir (the default returned by `user_data_dir`). Apps typically put cache data somewhere *under* the given dir here. Some examples: ...\Mozilla\Firefox\Profiles\\Cache ...\Acme\SuperApp\Cache\1.0 OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value. """ if WINDOWS: # Get the base path path = os.path.normpath(_get_win_folder("CSIDL_LOCAL_APPDATA")) # When using Python 2, return paths as bytes on Windows like we do on # other operating systems. See helper function docs for more details. if PY2 and isinstance(path, text_type): path = _win_path_to_bytes(path) # Add our app name and Cache directory to it path = os.path.join(path, appname, "Cache") elif sys.platform == "darwin": # Get the base path path = expanduser("~/Library/Caches") # Add our app name to it path = os.path.join(path, appname) else: # Get the base path path = os.getenv("XDG_CACHE_HOME", expanduser("~/.cache")) # Add our app name to it path = os.path.join(path, appname) return path def user_data_dir(appname, roaming=False): # type: (str, bool) -> str r""" Return full path to the user-specific data dir for this application. "appname" is the name of application. If None, just the system directory is returned. "roaming" (boolean, default False) can be set True to use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: macOS: ~/Library/Application Support/ if it exists, else ~/.config/ Unix: ~/.local/share/ # or in $XDG_DATA_HOME, if defined Win XP (not roaming): C:\Documents and Settings\\ ... ...Application Data\ Win XP (roaming): C:\Documents and Settings\\Local ... ...Settings\Application Data\ Win 7 (not roaming): C:\\Users\\AppData\Local\ Win 7 (roaming): C:\\Users\\AppData\Roaming\ For Unix, we follow the XDG spec and support $XDG_DATA_HOME. That means, by default "~/.local/share/". """ if WINDOWS: const = roaming and "CSIDL_APPDATA" or "CSIDL_LOCAL_APPDATA" path = os.path.join(os.path.normpath(_get_win_folder(const)), appname) elif sys.platform == "darwin": path = os.path.join( expanduser('~/Library/Application Support/'), appname, ) if os.path.isdir(os.path.join( expanduser('~/Library/Application Support/'), appname, ) ) else os.path.join( expanduser('~/.config/'), appname, ) else: path = os.path.join( os.getenv('XDG_DATA_HOME', expanduser("~/.local/share")), appname, ) return path def user_config_dir(appname, roaming=True): # type: (str, bool) -> str """Return full path to the user-specific config dir for this application. "appname" is the name of application. If None, just the system directory is returned. "roaming" (boolean, default True) can be set False to not use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: macOS: same as user_data_dir Unix: ~/.config/ Win *: same as user_data_dir For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME. That means, by default "~/.config/". """ if WINDOWS: path = user_data_dir(appname, roaming=roaming) elif sys.platform == "darwin": path = user_data_dir(appname) else: path = os.getenv('XDG_CONFIG_HOME', expanduser("~/.config")) path = os.path.join(path, appname) return path # for the discussion regarding site_config_dirs locations # see def site_config_dirs(appname): # type: (str) -> List[str] r"""Return a list of potential user-shared config dirs for this application. "appname" is the name of application. Typical user config directories are: macOS: /Library/Application Support// Unix: /etc or $XDG_CONFIG_DIRS[i]// for each value in $XDG_CONFIG_DIRS Win XP: C:\Documents and Settings\All Users\Application ... ...Data\\ Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) Win 7: Hidden, but writeable on Win 7: C:\ProgramData\\ """ if WINDOWS: path = os.path.normpath(_get_win_folder("CSIDL_COMMON_APPDATA")) pathlist = [os.path.join(path, appname)] elif sys.platform == 'darwin': pathlist = [os.path.join('/Library/Application Support', appname)] else: # try looking in $XDG_CONFIG_DIRS xdg_config_dirs = os.getenv('XDG_CONFIG_DIRS', '/etc/xdg') if xdg_config_dirs: pathlist = [ os.path.join(expanduser(x), appname) for x in xdg_config_dirs.split(os.pathsep) ] else: pathlist = [] # always look in /etc directly as well pathlist.append('/etc') return pathlist # -- Windows support functions -- def _get_win_folder_from_registry(csidl_name): # type: (str) -> str """ This is a fallback technique at best. I'm not sure if using the registry for this guarantees us the correct answer for all CSIDL_* names. """ import _winreg shell_folder_name = { "CSIDL_APPDATA": "AppData", "CSIDL_COMMON_APPDATA": "Common AppData", "CSIDL_LOCAL_APPDATA": "Local AppData", }[csidl_name] key = _winreg.OpenKey( _winreg.HKEY_CURRENT_USER, r"Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders" ) directory, _type = _winreg.QueryValueEx(key, shell_folder_name) return directory def _get_win_folder_with_ctypes(csidl_name): # type: (str) -> str # On Python 2, ctypes.create_unicode_buffer().value returns "unicode", # which isn't the same as str in the annotation above. csidl_const = { "CSIDL_APPDATA": 26, "CSIDL_COMMON_APPDATA": 35, "CSIDL_LOCAL_APPDATA": 28, }[csidl_name] buf = ctypes.create_unicode_buffer(1024) windll = ctypes.windll # type: ignore windll.shell32.SHGetFolderPathW(None, csidl_const, None, 0, buf) # Downgrade to short path name if have highbit chars. See # . has_high_char = False for c in buf: if ord(c) > 255: has_high_char = True break if has_high_char: buf2 = ctypes.create_unicode_buffer(1024) if windll.kernel32.GetShortPathNameW(buf.value, buf2, 1024): buf = buf2 # The type: ignore is explained under the type annotation for this function return buf.value # type: ignore if WINDOWS: try: import ctypes _get_win_folder = _get_win_folder_with_ctypes except ImportError: _get_win_folder = _get_win_folder_from_registry def _win_path_to_bytes(path): """Encode Windows paths to bytes. Only used on Python 2. Motivation is to be consistent with other operating systems where paths are also returned as bytes. This avoids problems mixing bytes and Unicode elsewhere in the codebase. For more details and discussion see . If encoding using ASCII and MBCS fails, return the original Unicode path. """ for encoding in ('ASCII', 'MBCS'): try: return path.encode(encoding) except (UnicodeEncodeError, LookupError): pass return path site-packages/pip/_internal/utils/compat.py000064400000022535151464452710015044 0ustar00"""Stuff that differs in different Python versions and platform distributions.""" # The following comment should be removed at some point in the future. # mypy: disallow-untyped-defs=False from __future__ import absolute_import, division import codecs import locale import logging import os import shutil import sys from pip._vendor.six import PY2, text_type from pip._vendor.urllib3.util import IS_PYOPENSSL from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional, Text, Tuple, Union try: import _ssl # noqa except ImportError: ssl = None else: # This additional assignment was needed to prevent a mypy error. ssl = _ssl try: import ipaddress except ImportError: try: from pip._vendor import ipaddress # type: ignore except ImportError: import ipaddr as ipaddress # type: ignore ipaddress.ip_address = ipaddress.IPAddress # type: ignore ipaddress.ip_network = ipaddress.IPNetwork # type: ignore __all__ = [ "ipaddress", "uses_pycache", "console_to_str", "native_str", "get_path_uid", "stdlib_pkgs", "WINDOWS", "samefile", "get_terminal_size", "get_extension_suffixes", ] logger = logging.getLogger(__name__) HAS_TLS = (ssl is not None) or IS_PYOPENSSL if PY2: import imp try: cache_from_source = imp.cache_from_source # type: ignore except AttributeError: # does not use __pycache__ cache_from_source = None uses_pycache = cache_from_source is not None else: uses_pycache = True from importlib.util import cache_from_source if PY2: # In Python 2.7, backslashreplace exists # but does not support use for decoding. # We implement our own replace handler for this # situation, so that we can consistently use # backslash replacement for all versions. def backslashreplace_decode_fn(err): raw_bytes = (err.object[i] for i in range(err.start, err.end)) # Python 2 gave us characters - convert to numeric bytes raw_bytes = (ord(b) for b in raw_bytes) return u"".join(u"\\x%x" % c for c in raw_bytes), err.end codecs.register_error( "backslashreplace_decode", backslashreplace_decode_fn, ) backslashreplace_decode = "backslashreplace_decode" else: backslashreplace_decode = "backslashreplace" def str_to_display(data, desc=None): # type: (Union[bytes, Text], Optional[str]) -> Text """ For display or logging purposes, convert a bytes object (or text) to text (e.g. unicode in Python 2) safe for output. :param desc: An optional phrase describing the input data, for use in the log message if a warning is logged. Defaults to "Bytes object". This function should never error out and so can take a best effort approach. It is okay to be lossy if needed since the return value is just for display. We assume the data is in the locale preferred encoding. If it won't decode properly, we warn the user but decode as best we can. We also ensure that the output can be safely written to standard output without encoding errors. """ if isinstance(data, text_type): return data # Otherwise, data is a bytes object (str in Python 2). # First, get the encoding we assume. This is the preferred # encoding for the locale, unless that is not found, or # it is ASCII, in which case assume UTF-8 encoding = locale.getpreferredencoding() if (not encoding) or codecs.lookup(encoding).name == "ascii": encoding = "utf-8" # Now try to decode the data - if we fail, warn the user and # decode with replacement. try: decoded_data = data.decode(encoding) except UnicodeDecodeError: if desc is None: desc = 'Bytes object' msg_format = '{} does not appear to be encoded as %s'.format(desc) logger.warning(msg_format, encoding) decoded_data = data.decode(encoding, errors=backslashreplace_decode) # Make sure we can print the output, by encoding it to the output # encoding with replacement of unencodable characters, and then # decoding again. # We use stderr's encoding because it's less likely to be # redirected and if we don't find an encoding we skip this # step (on the assumption that output is wrapped by something # that won't fail). # The double getattr is to deal with the possibility that we're # being called in a situation where sys.__stderr__ doesn't exist, # or doesn't have an encoding attribute. Neither of these cases # should occur in normal pip use, but there's no harm in checking # in case people use pip in (unsupported) unusual situations. output_encoding = getattr(getattr(sys, "__stderr__", None), "encoding", None) if output_encoding: output_encoded = decoded_data.encode( output_encoding, errors="backslashreplace" ) decoded_data = output_encoded.decode(output_encoding) return decoded_data def console_to_str(data): # type: (bytes) -> Text """Return a string, safe for output, of subprocess output. """ return str_to_display(data, desc='Subprocess output') if PY2: def native_str(s, replace=False): # type: (str, bool) -> str # Replace is ignored -- unicode to UTF-8 can't fail if isinstance(s, text_type): return s.encode('utf-8') return s else: def native_str(s, replace=False): # type: (str, bool) -> str if isinstance(s, bytes): return s.decode('utf-8', 'replace' if replace else 'strict') return s def get_path_uid(path): # type: (str) -> int """ Return path's uid. Does not follow symlinks: https://github.com/pypa/pip/pull/935#discussion_r5307003 Placed this function in compat due to differences on AIX and Jython, that should eventually go away. :raises OSError: When path is a symlink or can't be read. """ if hasattr(os, 'O_NOFOLLOW'): fd = os.open(path, os.O_RDONLY | os.O_NOFOLLOW) file_uid = os.fstat(fd).st_uid os.close(fd) else: # AIX and Jython # WARNING: time of check vulnerability, but best we can do w/o NOFOLLOW if not os.path.islink(path): # older versions of Jython don't have `os.fstat` file_uid = os.stat(path).st_uid else: # raise OSError for parity with os.O_NOFOLLOW above raise OSError( "%s is a symlink; Will not return uid for symlinks" % path ) return file_uid if PY2: from imp import get_suffixes def get_extension_suffixes(): return [suffix[0] for suffix in get_suffixes()] else: from importlib.machinery import EXTENSION_SUFFIXES def get_extension_suffixes(): return EXTENSION_SUFFIXES def expanduser(path): # type: (str) -> str """ Expand ~ and ~user constructions. Includes a workaround for https://bugs.python.org/issue14768 """ expanded = os.path.expanduser(path) if path.startswith('~/') and expanded.startswith('//'): expanded = expanded[1:] return expanded # packages in the stdlib that may have installation metadata, but should not be # considered 'installed'. this theoretically could be determined based on # dist.location (py27:`sysconfig.get_paths()['stdlib']`, # py26:sysconfig.get_config_vars('LIBDEST')), but fear platform variation may # make this ineffective, so hard-coding stdlib_pkgs = {"python", "wsgiref", "argparse"} # windows detection, covers cpython and ironpython WINDOWS = (sys.platform.startswith("win") or (sys.platform == 'cli' and os.name == 'nt')) def samefile(file1, file2): # type: (str, str) -> bool """Provide an alternative for os.path.samefile on Windows/Python2""" if hasattr(os.path, 'samefile'): return os.path.samefile(file1, file2) else: path1 = os.path.normcase(os.path.abspath(file1)) path2 = os.path.normcase(os.path.abspath(file2)) return path1 == path2 if hasattr(shutil, 'get_terminal_size'): def get_terminal_size(): # type: () -> Tuple[int, int] """ Returns a tuple (x, y) representing the width(x) and the height(y) in characters of the terminal window. """ return tuple(shutil.get_terminal_size()) # type: ignore else: def get_terminal_size(): # type: () -> Tuple[int, int] """ Returns a tuple (x, y) representing the width(x) and the height(y) in characters of the terminal window. """ def ioctl_GWINSZ(fd): try: import fcntl import termios import struct cr = struct.unpack_from( 'hh', fcntl.ioctl(fd, termios.TIOCGWINSZ, '12345678') ) except Exception: return None if cr == (0, 0): return None return cr cr = ioctl_GWINSZ(0) or ioctl_GWINSZ(1) or ioctl_GWINSZ(2) if not cr: try: fd = os.open(os.ctermid(), os.O_RDONLY) cr = ioctl_GWINSZ(fd) os.close(fd) except Exception: pass if not cr: cr = (os.environ.get('LINES', 25), os.environ.get('COLUMNS', 80)) return int(cr[1]), int(cr[0]) site-packages/pip/_internal/utils/packaging.py000064400000005733151464452710015506 0ustar00from __future__ import absolute_import import logging from email.parser import FeedParser from pip._vendor import pkg_resources from pip._vendor.packaging import specifiers, version from pip._internal.exceptions import NoneMetadataError from pip._internal.utils.misc import display_path from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Optional, Tuple from email.message import Message from pip._vendor.pkg_resources import Distribution logger = logging.getLogger(__name__) def check_requires_python(requires_python, version_info): # type: (Optional[str], Tuple[int, ...]) -> bool """ Check if the given Python version matches a "Requires-Python" specifier. :param version_info: A 3-tuple of ints representing a Python major-minor-micro version to check (e.g. `sys.version_info[:3]`). :return: `True` if the given Python version satisfies the requirement. Otherwise, return `False`. :raises InvalidSpecifier: If `requires_python` has an invalid format. """ if requires_python is None: # The package provides no information return True requires_python_specifier = specifiers.SpecifierSet(requires_python) python_version = version.parse('.'.join(map(str, version_info))) return python_version in requires_python_specifier def get_metadata(dist): # type: (Distribution) -> Message """ :raises NoneMetadataError: if the distribution reports `has_metadata()` True but `get_metadata()` returns None. """ metadata_name = 'METADATA' if (isinstance(dist, pkg_resources.DistInfoDistribution) and dist.has_metadata(metadata_name)): metadata = dist.get_metadata(metadata_name) elif dist.has_metadata('PKG-INFO'): metadata_name = 'PKG-INFO' metadata = dist.get_metadata(metadata_name) else: logger.warning("No metadata found in %s", display_path(dist.location)) metadata = '' if metadata is None: raise NoneMetadataError(dist, metadata_name) feed_parser = FeedParser() # The following line errors out if with a "NoneType" TypeError if # passed metadata=None. feed_parser.feed(metadata) return feed_parser.close() def get_requires_python(dist): # type: (pkg_resources.Distribution) -> Optional[str] """ Return the "Requires-Python" metadata for a distribution, or None if not present. """ pkg_info_dict = get_metadata(dist) requires_python = pkg_info_dict.get('Requires-Python') if requires_python is not None: # Convert to a str to satisfy the type checker, since requires_python # can be a Header object. requires_python = str(requires_python) return requires_python def get_installer(dist): # type: (Distribution) -> str if dist.has_metadata('INSTALLER'): for line in dist.get_metadata_lines('INSTALLER'): if line.strip(): return line.strip() return '' site-packages/pip/_internal/utils/ui.py000064400000033122151464452710014170 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import, division import contextlib import itertools import logging import sys import time from signal import SIGINT, default_int_handler, signal from pip._vendor import six from pip._vendor.progress import HIDE_CURSOR, SHOW_CURSOR from pip._vendor.progress.bar import Bar, FillingCirclesBar, IncrementalBar from pip._vendor.progress.spinner import Spinner from pip._internal.utils.compat import WINDOWS from pip._internal.utils.logging import get_indentation from pip._internal.utils.misc import format_size from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Any, Iterator, IO try: from pip._vendor import colorama # Lots of different errors can come from this, including SystemError and # ImportError. except Exception: colorama = None logger = logging.getLogger(__name__) def _select_progress_class(preferred, fallback): encoding = getattr(preferred.file, "encoding", None) # If we don't know what encoding this file is in, then we'll just assume # that it doesn't support unicode and use the ASCII bar. if not encoding: return fallback # Collect all of the possible characters we want to use with the preferred # bar. characters = [ getattr(preferred, "empty_fill", six.text_type()), getattr(preferred, "fill", six.text_type()), ] characters += list(getattr(preferred, "phases", [])) # Try to decode the characters we're using for the bar using the encoding # of the given file, if this works then we'll assume that we can use the # fancier bar and if not we'll fall back to the plaintext bar. try: six.text_type().join(characters).encode(encoding) except UnicodeEncodeError: return fallback else: return preferred _BaseBar = _select_progress_class(IncrementalBar, Bar) # type: Any class InterruptibleMixin(object): """ Helper to ensure that self.finish() gets called on keyboard interrupt. This allows downloads to be interrupted without leaving temporary state (like hidden cursors) behind. This class is similar to the progress library's existing SigIntMixin helper, but as of version 1.2, that helper has the following problems: 1. It calls sys.exit(). 2. It discards the existing SIGINT handler completely. 3. It leaves its own handler in place even after an uninterrupted finish, which will have unexpected delayed effects if the user triggers an unrelated keyboard interrupt some time after a progress-displaying download has already completed, for example. """ def __init__(self, *args, **kwargs): """ Save the original SIGINT handler for later. """ super(InterruptibleMixin, self).__init__(*args, **kwargs) self.original_handler = signal(SIGINT, self.handle_sigint) # If signal() returns None, the previous handler was not installed from # Python, and we cannot restore it. This probably should not happen, # but if it does, we must restore something sensible instead, at least. # The least bad option should be Python's default SIGINT handler, which # just raises KeyboardInterrupt. if self.original_handler is None: self.original_handler = default_int_handler def finish(self): """ Restore the original SIGINT handler after finishing. This should happen regardless of whether the progress display finishes normally, or gets interrupted. """ super(InterruptibleMixin, self).finish() signal(SIGINT, self.original_handler) def handle_sigint(self, signum, frame): """ Call self.finish() before delegating to the original SIGINT handler. This handler should only be in place while the progress display is active. """ self.finish() self.original_handler(signum, frame) class SilentBar(Bar): def update(self): pass class BlueEmojiBar(IncrementalBar): suffix = "%(percent)d%%" bar_prefix = " " bar_suffix = " " phases = (u"\U0001F539", u"\U0001F537", u"\U0001F535") # type: Any class DownloadProgressMixin(object): def __init__(self, *args, **kwargs): super(DownloadProgressMixin, self).__init__(*args, **kwargs) self.message = (" " * (get_indentation() + 2)) + self.message @property def downloaded(self): return format_size(self.index) @property def download_speed(self): # Avoid zero division errors... if self.avg == 0.0: return "..." return format_size(1 / self.avg) + "/s" @property def pretty_eta(self): if self.eta: return "eta %s" % self.eta_td return "" def iter(self, it, n=1): for x in it: yield x self.next(n) self.finish() class WindowsMixin(object): def __init__(self, *args, **kwargs): # The Windows terminal does not support the hide/show cursor ANSI codes # even with colorama. So we'll ensure that hide_cursor is False on # Windows. # This call needs to go before the super() call, so that hide_cursor # is set in time. The base progress bar class writes the "hide cursor" # code to the terminal in its init, so if we don't set this soon # enough, we get a "hide" with no corresponding "show"... if WINDOWS and self.hide_cursor: self.hide_cursor = False super(WindowsMixin, self).__init__(*args, **kwargs) # Check if we are running on Windows and we have the colorama module, # if we do then wrap our file with it. if WINDOWS and colorama: self.file = colorama.AnsiToWin32(self.file) # The progress code expects to be able to call self.file.isatty() # but the colorama.AnsiToWin32() object doesn't have that, so we'll # add it. self.file.isatty = lambda: self.file.wrapped.isatty() # The progress code expects to be able to call self.file.flush() # but the colorama.AnsiToWin32() object doesn't have that, so we'll # add it. self.file.flush = lambda: self.file.wrapped.flush() class BaseDownloadProgressBar(WindowsMixin, InterruptibleMixin, DownloadProgressMixin): file = sys.stdout message = "%(percent)d%%" suffix = "%(downloaded)s %(download_speed)s %(pretty_eta)s" # NOTE: The "type: ignore" comments on the following classes are there to # work around https://github.com/python/typing/issues/241 class DefaultDownloadProgressBar(BaseDownloadProgressBar, _BaseBar): pass class DownloadSilentBar(BaseDownloadProgressBar, SilentBar): # type: ignore pass class DownloadBar(BaseDownloadProgressBar, # type: ignore Bar): pass class DownloadFillingCirclesBar(BaseDownloadProgressBar, # type: ignore FillingCirclesBar): pass class DownloadBlueEmojiProgressBar(BaseDownloadProgressBar, # type: ignore BlueEmojiBar): pass class DownloadProgressSpinner(WindowsMixin, InterruptibleMixin, DownloadProgressMixin, Spinner): file = sys.stdout suffix = "%(downloaded)s %(download_speed)s" def next_phase(self): if not hasattr(self, "_phaser"): self._phaser = itertools.cycle(self.phases) return next(self._phaser) def update(self): message = self.message % self phase = self.next_phase() suffix = self.suffix % self line = ''.join([ message, " " if message else "", phase, " " if suffix else "", suffix, ]) self.writeln(line) BAR_TYPES = { "off": (DownloadSilentBar, DownloadSilentBar), "on": (DefaultDownloadProgressBar, DownloadProgressSpinner), "ascii": (DownloadBar, DownloadProgressSpinner), "pretty": (DownloadFillingCirclesBar, DownloadProgressSpinner), "emoji": (DownloadBlueEmojiProgressBar, DownloadProgressSpinner) } def DownloadProgressProvider(progress_bar, max=None): if max is None or max == 0: return BAR_TYPES[progress_bar][1]().iter else: return BAR_TYPES[progress_bar][0](max=max).iter ################################################################ # Generic "something is happening" spinners # # We don't even try using progress.spinner.Spinner here because it's actually # simpler to reimplement from scratch than to coerce their code into doing # what we need. ################################################################ @contextlib.contextmanager def hidden_cursor(file): # type: (IO) -> Iterator[None] # The Windows terminal does not support the hide/show cursor ANSI codes, # even via colorama. So don't even try. if WINDOWS: yield # We don't want to clutter the output with control characters if we're # writing to a file, or if the user is running with --quiet. # See https://github.com/pypa/pip/issues/3418 elif not file.isatty() or logger.getEffectiveLevel() > logging.INFO: yield else: file.write(HIDE_CURSOR) try: yield finally: file.write(SHOW_CURSOR) class RateLimiter(object): def __init__(self, min_update_interval_seconds): # type: (float) -> None self._min_update_interval_seconds = min_update_interval_seconds self._last_update = 0 # type: float def ready(self): # type: () -> bool now = time.time() delta = now - self._last_update return delta >= self._min_update_interval_seconds def reset(self): # type: () -> None self._last_update = time.time() class SpinnerInterface(object): def spin(self): # type: () -> None raise NotImplementedError() def finish(self, final_status): # type: (str) -> None raise NotImplementedError() class InteractiveSpinner(SpinnerInterface): def __init__(self, message, file=None, spin_chars="-\\|/", # Empirically, 8 updates/second looks nice min_update_interval_seconds=0.125): self._message = message if file is None: file = sys.stdout self._file = file self._rate_limiter = RateLimiter(min_update_interval_seconds) self._finished = False self._spin_cycle = itertools.cycle(spin_chars) self._file.write(" " * get_indentation() + self._message + " ... ") self._width = 0 def _write(self, status): assert not self._finished # Erase what we wrote before by backspacing to the beginning, writing # spaces to overwrite the old text, and then backspacing again backup = "\b" * self._width self._file.write(backup + " " * self._width + backup) # Now we have a blank slate to add our status self._file.write(status) self._width = len(status) self._file.flush() self._rate_limiter.reset() def spin(self): # type: () -> None if self._finished: return if not self._rate_limiter.ready(): return self._write(next(self._spin_cycle)) def finish(self, final_status): # type: (str) -> None if self._finished: return self._write(final_status) self._file.write("\n") self._file.flush() self._finished = True # Used for dumb terminals, non-interactive installs (no tty), etc. # We still print updates occasionally (once every 60 seconds by default) to # act as a keep-alive for systems like Travis-CI that take lack-of-output as # an indication that a task has frozen. class NonInteractiveSpinner(SpinnerInterface): def __init__(self, message, min_update_interval_seconds=60): # type: (str, float) -> None self._message = message self._finished = False self._rate_limiter = RateLimiter(min_update_interval_seconds) self._update("started") def _update(self, status): assert not self._finished self._rate_limiter.reset() logger.info("%s: %s", self._message, status) def spin(self): # type: () -> None if self._finished: return if not self._rate_limiter.ready(): return self._update("still running...") def finish(self, final_status): # type: (str) -> None if self._finished: return self._update("finished with status '%s'" % (final_status,)) self._finished = True @contextlib.contextmanager def open_spinner(message): # type: (str) -> Iterator[SpinnerInterface] # Interactive spinner goes directly to sys.stdout rather than being routed # through the logging system, but it acts like it has level INFO, # i.e. it's only displayed if we're at level INFO or better. # Non-interactive spinner goes through the logging system, so it is always # in sync with logging configuration. if sys.stdout.isatty() and logger.getEffectiveLevel() <= logging.INFO: spinner = InteractiveSpinner(message) # type: SpinnerInterface else: spinner = NonInteractiveSpinner(message) try: with hidden_cursor(sys.stdout): yield spinner except KeyboardInterrupt: spinner.finish("canceled") raise except Exception: spinner.finish("error") raise else: spinner.finish("done") site-packages/pip/_internal/utils/subprocess.py000064400000023267151464452710015754 0ustar00# The following comment should be removed at some point in the future. # mypy: strict-optional=False from __future__ import absolute_import import logging import os import subprocess from pip._vendor.six.moves import shlex_quote from pip._internal.exceptions import InstallationError from pip._internal.utils.compat import console_to_str, str_to_display from pip._internal.utils.logging import subprocess_logger from pip._internal.utils.misc import HiddenText, path_to_display from pip._internal.utils.typing import MYPY_CHECK_RUNNING from pip._internal.utils.ui import open_spinner if MYPY_CHECK_RUNNING: from typing import ( Any, Callable, Iterable, List, Mapping, Optional, Text, Union, ) from pip._internal.utils.ui import SpinnerInterface CommandArgs = List[Union[str, HiddenText]] LOG_DIVIDER = '----------------------------------------' def make_command(*args): # type: (Union[str, HiddenText, CommandArgs]) -> CommandArgs """ Create a CommandArgs object. """ command_args = [] # type: CommandArgs for arg in args: # Check for list instead of CommandArgs since CommandArgs is # only known during type-checking. if isinstance(arg, list): command_args.extend(arg) else: # Otherwise, arg is str or HiddenText. command_args.append(arg) return command_args def format_command_args(args): # type: (Union[List[str], CommandArgs]) -> str """ Format command arguments for display. """ # For HiddenText arguments, display the redacted form by calling str(). # Also, we don't apply str() to arguments that aren't HiddenText since # this can trigger a UnicodeDecodeError in Python 2 if the argument # has type unicode and includes a non-ascii character. (The type # checker doesn't ensure the annotations are correct in all cases.) return ' '.join( shlex_quote(str(arg)) if isinstance(arg, HiddenText) else shlex_quote(arg) for arg in args ) def reveal_command_args(args): # type: (Union[List[str], CommandArgs]) -> List[str] """ Return the arguments in their raw, unredacted form. """ return [ arg.secret if isinstance(arg, HiddenText) else arg for arg in args ] def make_subprocess_output_error( cmd_args, # type: Union[List[str], CommandArgs] cwd, # type: Optional[str] lines, # type: List[Text] exit_status, # type: int ): # type: (...) -> Text """ Create and return the error message to use to log a subprocess error with command output. :param lines: A list of lines, each ending with a newline. """ command = format_command_args(cmd_args) # Convert `command` and `cwd` to text (unicode in Python 2) so we can use # them as arguments in the unicode format string below. This avoids # "UnicodeDecodeError: 'ascii' codec can't decode byte ..." in Python 2 # if either contains a non-ascii character. command_display = str_to_display(command, desc='command bytes') cwd_display = path_to_display(cwd) # We know the joined output value ends in a newline. output = ''.join(lines) msg = ( # Use a unicode string to avoid "UnicodeEncodeError: 'ascii' # codec can't encode character ..." in Python 2 when a format # argument (e.g. `output`) has a non-ascii character. u'Command errored out with exit status {exit_status}:\n' ' command: {command_display}\n' ' cwd: {cwd_display}\n' 'Complete output ({line_count} lines):\n{output}{divider}' ).format( exit_status=exit_status, command_display=command_display, cwd_display=cwd_display, line_count=len(lines), output=output, divider=LOG_DIVIDER, ) return msg def call_subprocess( cmd, # type: Union[List[str], CommandArgs] show_stdout=False, # type: bool cwd=None, # type: Optional[str] on_returncode='raise', # type: str extra_ok_returncodes=None, # type: Optional[Iterable[int]] command_desc=None, # type: Optional[str] extra_environ=None, # type: Optional[Mapping[str, Any]] unset_environ=None, # type: Optional[Iterable[str]] spinner=None, # type: Optional[SpinnerInterface] log_failed_cmd=True # type: Optional[bool] ): # type: (...) -> Text """ Args: show_stdout: if true, use INFO to log the subprocess's stderr and stdout streams. Otherwise, use DEBUG. Defaults to False. extra_ok_returncodes: an iterable of integer return codes that are acceptable, in addition to 0. Defaults to None, which means []. unset_environ: an iterable of environment variable names to unset prior to calling subprocess.Popen(). log_failed_cmd: if false, failed commands are not logged, only raised. """ if extra_ok_returncodes is None: extra_ok_returncodes = [] if unset_environ is None: unset_environ = [] # Most places in pip use show_stdout=False. What this means is-- # # - We connect the child's output (combined stderr and stdout) to a # single pipe, which we read. # - We log this output to stderr at DEBUG level as it is received. # - If DEBUG logging isn't enabled (e.g. if --verbose logging wasn't # requested), then we show a spinner so the user can still see the # subprocess is in progress. # - If the subprocess exits with an error, we log the output to stderr # at ERROR level if it hasn't already been displayed to the console # (e.g. if --verbose logging wasn't enabled). This way we don't log # the output to the console twice. # # If show_stdout=True, then the above is still done, but with DEBUG # replaced by INFO. if show_stdout: # Then log the subprocess output at INFO level. log_subprocess = subprocess_logger.info used_level = logging.INFO else: # Then log the subprocess output using DEBUG. This also ensures # it will be logged to the log file (aka user_log), if enabled. log_subprocess = subprocess_logger.debug used_level = logging.DEBUG # Whether the subprocess will be visible in the console. showing_subprocess = subprocess_logger.getEffectiveLevel() <= used_level # Only use the spinner if we're not showing the subprocess output # and we have a spinner. use_spinner = not showing_subprocess and spinner is not None if command_desc is None: command_desc = format_command_args(cmd) log_subprocess("Running command %s", command_desc) env = os.environ.copy() if extra_environ: env.update(extra_environ) for name in unset_environ: env.pop(name, None) try: proc = subprocess.Popen( # Convert HiddenText objects to the underlying str. reveal_command_args(cmd), stderr=subprocess.STDOUT, stdin=subprocess.PIPE, stdout=subprocess.PIPE, cwd=cwd, env=env, ) proc.stdin.close() except Exception as exc: if log_failed_cmd: subprocess_logger.critical( "Error %s while executing command %s", exc, command_desc, ) raise all_output = [] while True: # The "line" value is a unicode string in Python 2. line = console_to_str(proc.stdout.readline()) if not line: break line = line.rstrip() all_output.append(line + '\n') # Show the line immediately. log_subprocess(line) # Update the spinner. if use_spinner: spinner.spin() try: proc.wait() finally: if proc.stdout: proc.stdout.close() proc_had_error = ( proc.returncode and proc.returncode not in extra_ok_returncodes ) if use_spinner: if proc_had_error: spinner.finish("error") else: spinner.finish("done") if proc_had_error: if on_returncode == 'raise': if not showing_subprocess and log_failed_cmd: # Then the subprocess streams haven't been logged to the # console yet. msg = make_subprocess_output_error( cmd_args=cmd, cwd=cwd, lines=all_output, exit_status=proc.returncode, ) subprocess_logger.error(msg) exc_msg = ( 'Command errored out with exit status {}: {} ' 'Check the logs for full command output.' ).format(proc.returncode, command_desc) raise InstallationError(exc_msg) elif on_returncode == 'warn': subprocess_logger.warning( 'Command "%s" had error code %s in %s', command_desc, proc.returncode, cwd, ) elif on_returncode == 'ignore': pass else: raise ValueError('Invalid value: on_returncode=%s' % repr(on_returncode)) return ''.join(all_output) def runner_with_spinner_message(message): # type: (str) -> Callable """Provide a subprocess_runner that shows a spinner message. Intended for use with for pep517's Pep517HookCaller. Thus, the runner has an API that matches what's expected by Pep517HookCaller.subprocess_runner. """ def runner( cmd, # type: List[str] cwd=None, # type: Optional[str] extra_environ=None # type: Optional[Mapping[str, Any]] ): # type: (...) -> None with open_spinner(message) as spinner: call_subprocess( cmd, cwd=cwd, extra_environ=extra_environ, spinner=spinner, ) return runner site-packages/pip/_internal/utils/__init__.py000064400000000000151464452710015277 0ustar00site-packages/pip/_internal/utils/unpacking.py000064400000022722151464452710015536 0ustar00"""Utilities related archives. """ # The following comment should be removed at some point in the future. # mypy: strict-optional=False # mypy: disallow-untyped-defs=False from __future__ import absolute_import import logging import os import shutil import stat import tarfile import zipfile from pip._internal.exceptions import InstallationError from pip._internal.utils.filetypes import ( BZ2_EXTENSIONS, TAR_EXTENSIONS, XZ_EXTENSIONS, ZIP_EXTENSIONS, ) from pip._internal.utils.misc import ensure_dir from pip._internal.utils.typing import MYPY_CHECK_RUNNING if MYPY_CHECK_RUNNING: from typing import Iterable, List, Optional, Text, Union logger = logging.getLogger(__name__) SUPPORTED_EXTENSIONS = ZIP_EXTENSIONS + TAR_EXTENSIONS try: import bz2 # noqa SUPPORTED_EXTENSIONS += BZ2_EXTENSIONS except ImportError: logger.debug('bz2 module is not available') try: # Only for Python 3.3+ import lzma # noqa SUPPORTED_EXTENSIONS += XZ_EXTENSIONS except ImportError: logger.debug('lzma module is not available') def current_umask(): """Get the current umask which involves having to set it temporarily.""" mask = os.umask(0) os.umask(mask) return mask def split_leading_dir(path): # type: (Union[str, Text]) -> List[Union[str, Text]] path = path.lstrip('/').lstrip('\\') if ( '/' in path and ( ('\\' in path and path.find('/') < path.find('\\')) or '\\' not in path ) ): return path.split('/', 1) elif '\\' in path: return path.split('\\', 1) else: return [path, ''] def has_leading_dir(paths): # type: (Iterable[Union[str, Text]]) -> bool """Returns true if all the paths have the same leading path name (i.e., everything is in one subdirectory in an archive)""" common_prefix = None for path in paths: prefix, rest = split_leading_dir(path) if not prefix: return False elif common_prefix is None: common_prefix = prefix elif prefix != common_prefix: return False return True def is_within_directory(directory, target): # type: ((Union[str, Text]), (Union[str, Text])) -> bool """ Return true if the absolute path of target is within the directory """ abs_directory = os.path.abspath(directory) abs_target = os.path.abspath(target) prefix = os.path.commonprefix([abs_directory, abs_target]) return prefix == abs_directory def unzip_file(filename, location, flatten=True): # type: (str, str, bool) -> None """ Unzip the file (with path `filename`) to the destination `location`. All files are written based on system defaults and umask (i.e. permissions are not preserved), except that regular file members with any execute permissions (user, group, or world) have "chmod +x" applied after being written. Note that for windows, any execute changes using os.chmod are no-ops per the python docs. """ ensure_dir(location) zipfp = open(filename, 'rb') try: zip = zipfile.ZipFile(zipfp, allowZip64=True) leading = has_leading_dir(zip.namelist()) and flatten for info in zip.infolist(): name = info.filename fn = name if leading: fn = split_leading_dir(name)[1] fn = os.path.join(location, fn) dir = os.path.dirname(fn) if not is_within_directory(location, fn): message = ( 'The zip file ({}) has a file ({}) trying to install ' 'outside target directory ({})' ) raise InstallationError(message.format(filename, fn, location)) if fn.endswith('/') or fn.endswith('\\'): # A directory ensure_dir(fn) else: ensure_dir(dir) # Don't use read() to avoid allocating an arbitrarily large # chunk of memory for the file's content fp = zip.open(name) try: with open(fn, 'wb') as destfp: shutil.copyfileobj(fp, destfp) finally: fp.close() mode = info.external_attr >> 16 # if mode and regular file and any execute permissions for # user/group/world? if mode and stat.S_ISREG(mode) and mode & 0o111: # make dest file have execute for user/group/world # (chmod +x) no-op on windows per python docs os.chmod(fn, (0o777 - current_umask() | 0o111)) finally: zipfp.close() def untar_file(filename, location): # type: (str, str) -> None """ Untar the file (with path `filename`) to the destination `location`. All files are written based on system defaults and umask (i.e. permissions are not preserved), except that regular file members with any execute permissions (user, group, or world) have "chmod +x" applied after being written. Note that for windows, any execute changes using os.chmod are no-ops per the python docs. """ ensure_dir(location) if filename.lower().endswith('.gz') or filename.lower().endswith('.tgz'): mode = 'r:gz' elif filename.lower().endswith(BZ2_EXTENSIONS): mode = 'r:bz2' elif filename.lower().endswith(XZ_EXTENSIONS): mode = 'r:xz' elif filename.lower().endswith('.tar'): mode = 'r' else: logger.warning( 'Cannot determine compression type for file %s', filename, ) mode = 'r:*' tar = tarfile.open(filename, mode) try: leading = has_leading_dir([ member.name for member in tar.getmembers() ]) for member in tar.getmembers(): fn = member.name if leading: # https://github.com/python/mypy/issues/1174 fn = split_leading_dir(fn)[1] # type: ignore path = os.path.join(location, fn) if not is_within_directory(location, path): message = ( 'The tar file ({}) has a file ({}) trying to install ' 'outside target directory ({})' ) raise InstallationError( message.format(filename, path, location) ) # Call the `data` filter for its side effect (raising exception) try: tarfile.data_filter(member.replace(name=fn), location) except tarfile.LinkOutsideDestinationError: pass if member.isdir(): ensure_dir(path) elif member.issym(): try: # https://github.com/python/typeshed/issues/2673 tar._extract_member(member, path) # type: ignore except Exception as exc: # Some corrupt tar files seem to produce this # (specifically bad symlinks) logger.warning( 'In the tar file %s the member %s is invalid: %s', filename, member.name, exc, ) continue else: try: fp = tar.extractfile(member) except (KeyError, AttributeError) as exc: # Some corrupt tar files seem to produce this # (specifically bad symlinks) logger.warning( 'In the tar file %s the member %s is invalid: %s', filename, member.name, exc, ) continue ensure_dir(os.path.dirname(path)) with open(path, 'wb') as destfp: shutil.copyfileobj(fp, destfp) fp.close() # Update the timestamp (useful for cython compiled files) # https://github.com/python/typeshed/issues/2673 tar.utime(member, path) # type: ignore # member have any execute permissions for user/group/world? if member.mode & 0o111: # make dest file have execute for user/group/world # no-op on windows per python docs os.chmod(path, (0o777 - current_umask() | 0o111)) finally: tar.close() def unpack_file( filename, # type: str location, # type: str content_type=None, # type: Optional[str] ): # type: (...) -> None filename = os.path.realpath(filename) if ( content_type == 'application/zip' or filename.lower().endswith(ZIP_EXTENSIONS) or zipfile.is_zipfile(filename) ): unzip_file( filename, location, flatten=not filename.endswith('.whl') ) elif ( content_type == 'application/x-gzip' or tarfile.is_tarfile(filename) or filename.lower().endswith( TAR_EXTENSIONS + BZ2_EXTENSIONS + XZ_EXTENSIONS ) ): untar_file(filename, location) else: # FIXME: handle? # FIXME: magic signatures? logger.critical( 'Cannot unpack file %s (downloaded from %s, content-type: %s); ' 'cannot detect archive format', filename, location, content_type, ) raise InstallationError( 'Cannot determine archive format of {}'.format(location) ) site-packages/pip/__main__.py000064400000001164151464452710012161 0ustar00from __future__ import absolute_import import os import sys # If we are running from a wheel, add the wheel to sys.path # This allows the usage python pip-*.whl/pip install pip-*.whl if __package__ == '': # __file__ is pip-*.whl/pip/__main__.py # first dirname call strips of '/__main__.py', second strips off '/pip' # Resulting path is the name of the wheel itself # Add that to sys.path so we can import pip path = os.path.dirname(os.path.dirname(__file__)) sys.path.insert(0, path) from pip._internal.main import main as _main # isort:skip # noqa if __name__ == '__main__': sys.exit(_main()) site-packages/pip/__pycache__/__init__.cpython-38.pyc000064400000000241151464452710016461 0ustar00U .e@sdZdS)z19.3.1N) __version__rr0/usr/lib/python3.8/site-packages/pip/__init__.pysite-packages/pip/__pycache__/__main__.cpython-38.opt-1.pyc000064400000000642151464452710017406 0ustar00U .et@sjddlmZddlZddlZedkrFejejeZejdeddl m Z e dkrfe e dS))absolute_importN)main__main__)Z __future__rossys __package__pathdirname__file__insertZpip._internal.mainrZ_main__name__exitrr0/usr/lib/python3.8/site-packages/pip/__main__.pys  site-packages/pip/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000241151464452710017420 0ustar00U .e@sdZdS)z19.3.1N) __version__rr0/usr/lib/python3.8/site-packages/pip/__init__.pysite-packages/pip/__pycache__/__main__.cpython-38.pyc000064400000000642151464452710016447 0ustar00U .et@sjddlmZddlZddlZedkrFejejeZejdeddl m Z e dkrfe e dS))absolute_importN)main__main__)Z __future__rossys __package__pathdirname__file__insertZpip._internal.mainrZ_main__name__exitrr0/usr/lib/python3.8/site-packages/pip/__main__.pys  site-packages/pip/_vendor/msgpack/fallback.py000064400000111116151464452710015260 0ustar00"""Fallback pure Python implementation of msgpack""" import sys import struct import warnings if sys.version_info[0] == 2: PY2 = True int_types = (int, long) def dict_iteritems(d): return d.iteritems() else: PY2 = False int_types = int unicode = str xrange = range def dict_iteritems(d): return d.items() if sys.version_info < (3, 5): # Ugly hack... RecursionError = RuntimeError def _is_recursionerror(e): return len(e.args) == 1 and isinstance(e.args[0], str) and \ e.args[0].startswith('maximum recursion depth exceeded') else: def _is_recursionerror(e): return True if hasattr(sys, 'pypy_version_info'): # cStringIO is slow on PyPy, StringIO is faster. However: PyPy's own # StringBuilder is fastest. from __pypy__ import newlist_hint try: from __pypy__.builders import BytesBuilder as StringBuilder except ImportError: from __pypy__.builders import StringBuilder USING_STRINGBUILDER = True class StringIO(object): def __init__(self, s=b''): if s: self.builder = StringBuilder(len(s)) self.builder.append(s) else: self.builder = StringBuilder() def write(self, s): if isinstance(s, memoryview): s = s.tobytes() elif isinstance(s, bytearray): s = bytes(s) self.builder.append(s) def getvalue(self): return self.builder.build() else: USING_STRINGBUILDER = False from io import BytesIO as StringIO newlist_hint = lambda size: [] from .exceptions import ( BufferFull, OutOfData, ExtraData, FormatError, StackError, ) from . import ExtType EX_SKIP = 0 EX_CONSTRUCT = 1 EX_READ_ARRAY_HEADER = 2 EX_READ_MAP_HEADER = 3 TYPE_IMMEDIATE = 0 TYPE_ARRAY = 1 TYPE_MAP = 2 TYPE_RAW = 3 TYPE_BIN = 4 TYPE_EXT = 5 DEFAULT_RECURSE_LIMIT = 511 def _check_type_strict(obj, t, type=type, tuple=tuple): if type(t) is tuple: return type(obj) in t else: return type(obj) is t def _get_data_from_buffer(obj): try: view = memoryview(obj) except TypeError: # try to use legacy buffer protocol if 2.7, otherwise re-raise if PY2: view = memoryview(buffer(obj)) warnings.warn("using old buffer interface to unpack %s; " "this leads to unpacking errors if slicing is used and " "will be removed in a future version" % type(obj), RuntimeWarning, stacklevel=3) else: raise if view.itemsize != 1: raise ValueError("cannot unpack from multi-byte object") return view def unpack(stream, **kwargs): warnings.warn( "Direct calling implementation's unpack() is deprecated, Use msgpack.unpack() or unpackb() instead.", DeprecationWarning, stacklevel=2) data = stream.read() return unpackb(data, **kwargs) def unpackb(packed, **kwargs): """ Unpack an object from `packed`. Raises ``ExtraData`` when *packed* contains extra bytes. Raises ``ValueError`` when *packed* is incomplete. Raises ``FormatError`` when *packed* is not valid msgpack. Raises ``StackError`` when *packed* contains too nested. Other exceptions can be raised during unpacking. See :class:`Unpacker` for options. """ unpacker = Unpacker(None, max_buffer_size=len(packed), **kwargs) unpacker.feed(packed) try: ret = unpacker._unpack() except OutOfData: raise ValueError("Unpack failed: incomplete input") except RecursionError as e: if _is_recursionerror(e): raise StackError raise if unpacker._got_extradata(): raise ExtraData(ret, unpacker._get_extradata()) return ret if sys.version_info < (2, 7, 6): def _unpack_from(f, b, o=0): """Explicit typcast for legacy struct.unpack_from""" return struct.unpack_from(f, bytes(b), o) else: _unpack_from = struct.unpack_from class Unpacker(object): """Streaming unpacker. arguments: :param file_like: File-like object having `.read(n)` method. If specified, unpacker reads serialized data from it and :meth:`feed()` is not usable. :param int read_size: Used as `file_like.read(read_size)`. (default: `min(16*1024, max_buffer_size)`) :param bool use_list: If true, unpack msgpack array to Python list. Otherwise, unpack to Python tuple. (default: True) :param bool raw: If true, unpack msgpack raw to Python bytes (default). Otherwise, unpack to Python str (or unicode on Python 2) by decoding with UTF-8 encoding (recommended). Currently, the default is true, but it will be changed to false in near future. So you must specify it explicitly for keeping backward compatibility. *encoding* option which is deprecated overrides this option. :param bool strict_map_key: If true, only str or bytes are accepted for map (dict) keys. It's False by default for backward-compatibility. But it will be True from msgpack 1.0. :param callable object_hook: When specified, it should be callable. Unpacker calls it with a dict argument after unpacking msgpack map. (See also simplejson) :param callable object_pairs_hook: When specified, it should be callable. Unpacker calls it with a list of key-value pairs after unpacking msgpack map. (See also simplejson) :param str encoding: Encoding used for decoding msgpack raw. If it is None (default), msgpack raw is deserialized to Python bytes. :param str unicode_errors: (deprecated) Used for decoding msgpack raw with *encoding*. (default: `'strict'`) :param int max_buffer_size: Limits size of data waiting unpacked. 0 means system's INT_MAX (default). Raises `BufferFull` exception when it is insufficient. You should set this parameter when unpacking data from untrusted source. :param int max_str_len: Deprecated, use *max_buffer_size* instead. Limits max length of str. (default: max_buffer_size or 1024*1024) :param int max_bin_len: Deprecated, use *max_buffer_size* instead. Limits max length of bin. (default: max_buffer_size or 1024*1024) :param int max_array_len: Limits max length of array. (default: max_buffer_size or 128*1024) :param int max_map_len: Limits max length of map. (default: max_buffer_size//2 or 32*1024) :param int max_ext_len: Deprecated, use *max_buffer_size* instead. Limits max size of ext type. (default: max_buffer_size or 1024*1024) Example of streaming deserialize from file-like object:: unpacker = Unpacker(file_like, raw=False, max_buffer_size=10*1024*1024) for o in unpacker: process(o) Example of streaming deserialize from socket:: unpacker = Unpacker(raw=False, max_buffer_size=10*1024*1024) while True: buf = sock.recv(1024**2) if not buf: break unpacker.feed(buf) for o in unpacker: process(o) Raises ``ExtraData`` when *packed* contains extra bytes. Raises ``OutOfData`` when *packed* is incomplete. Raises ``FormatError`` when *packed* is not valid msgpack. Raises ``StackError`` when *packed* contains too nested. Other exceptions can be raised during unpacking. """ def __init__(self, file_like=None, read_size=0, use_list=True, raw=True, strict_map_key=False, object_hook=None, object_pairs_hook=None, list_hook=None, encoding=None, unicode_errors=None, max_buffer_size=0, ext_hook=ExtType, max_str_len=-1, max_bin_len=-1, max_array_len=-1, max_map_len=-1, max_ext_len=-1): if encoding is not None: warnings.warn( "encoding is deprecated, Use raw=False instead.", DeprecationWarning, stacklevel=2) if unicode_errors is None: unicode_errors = 'strict' if file_like is None: self._feeding = True else: if not callable(file_like.read): raise TypeError("`file_like.read` must be callable") self.file_like = file_like self._feeding = False #: array of bytes fed. self._buffer = bytearray() #: Which position we currently reads self._buff_i = 0 # When Unpacker is used as an iterable, between the calls to next(), # the buffer is not "consumed" completely, for efficiency sake. # Instead, it is done sloppily. To make sure we raise BufferFull at # the correct moments, we have to keep track of how sloppy we were. # Furthermore, when the buffer is incomplete (that is: in the case # we raise an OutOfData) we need to rollback the buffer to the correct # state, which _buf_checkpoint records. self._buf_checkpoint = 0 if max_str_len == -1: max_str_len = max_buffer_size or 1024*1024 if max_bin_len == -1: max_bin_len = max_buffer_size or 1024*1024 if max_array_len == -1: max_array_len = max_buffer_size or 128*1024 if max_map_len == -1: max_map_len = max_buffer_size//2 or 32*1024 if max_ext_len == -1: max_ext_len = max_buffer_size or 1024*1024 self._max_buffer_size = max_buffer_size or 2**31-1 if read_size > self._max_buffer_size: raise ValueError("read_size must be smaller than max_buffer_size") self._read_size = read_size or min(self._max_buffer_size, 16*1024) self._raw = bool(raw) self._strict_map_key = bool(strict_map_key) self._encoding = encoding self._unicode_errors = unicode_errors self._use_list = use_list self._list_hook = list_hook self._object_hook = object_hook self._object_pairs_hook = object_pairs_hook self._ext_hook = ext_hook self._max_str_len = max_str_len self._max_bin_len = max_bin_len self._max_array_len = max_array_len self._max_map_len = max_map_len self._max_ext_len = max_ext_len self._stream_offset = 0 if list_hook is not None and not callable(list_hook): raise TypeError('`list_hook` is not callable') if object_hook is not None and not callable(object_hook): raise TypeError('`object_hook` is not callable') if object_pairs_hook is not None and not callable(object_pairs_hook): raise TypeError('`object_pairs_hook` is not callable') if object_hook is not None and object_pairs_hook is not None: raise TypeError("object_pairs_hook and object_hook are mutually " "exclusive") if not callable(ext_hook): raise TypeError("`ext_hook` is not callable") def feed(self, next_bytes): assert self._feeding view = _get_data_from_buffer(next_bytes) if (len(self._buffer) - self._buff_i + len(view) > self._max_buffer_size): raise BufferFull # Strip buffer before checkpoint before reading file. if self._buf_checkpoint > 0: del self._buffer[:self._buf_checkpoint] self._buff_i -= self._buf_checkpoint self._buf_checkpoint = 0 # Use extend here: INPLACE_ADD += doesn't reliably typecast memoryview in jython self._buffer.extend(view) def _consume(self): """ Gets rid of the used parts of the buffer. """ self._stream_offset += self._buff_i - self._buf_checkpoint self._buf_checkpoint = self._buff_i def _got_extradata(self): return self._buff_i < len(self._buffer) def _get_extradata(self): return self._buffer[self._buff_i:] def read_bytes(self, n): return self._read(n) def _read(self, n): # (int) -> bytearray self._reserve(n) i = self._buff_i self._buff_i = i+n return self._buffer[i:i+n] def _reserve(self, n): remain_bytes = len(self._buffer) - self._buff_i - n # Fast path: buffer has n bytes already if remain_bytes >= 0: return if self._feeding: self._buff_i = self._buf_checkpoint raise OutOfData # Strip buffer before checkpoint before reading file. if self._buf_checkpoint > 0: del self._buffer[:self._buf_checkpoint] self._buff_i -= self._buf_checkpoint self._buf_checkpoint = 0 # Read from file remain_bytes = -remain_bytes while remain_bytes > 0: to_read_bytes = max(self._read_size, remain_bytes) read_data = self.file_like.read(to_read_bytes) if not read_data: break assert isinstance(read_data, bytes) self._buffer += read_data remain_bytes -= len(read_data) if len(self._buffer) < n + self._buff_i: self._buff_i = 0 # rollback raise OutOfData def _read_header(self, execute=EX_CONSTRUCT): typ = TYPE_IMMEDIATE n = 0 obj = None self._reserve(1) b = self._buffer[self._buff_i] self._buff_i += 1 if b & 0b10000000 == 0: obj = b elif b & 0b11100000 == 0b11100000: obj = -1 - (b ^ 0xff) elif b & 0b11100000 == 0b10100000: n = b & 0b00011111 typ = TYPE_RAW if n > self._max_str_len: raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len) obj = self._read(n) elif b & 0b11110000 == 0b10010000: n = b & 0b00001111 typ = TYPE_ARRAY if n > self._max_array_len: raise ValueError("%s exceeds max_array_len(%s)", n, self._max_array_len) elif b & 0b11110000 == 0b10000000: n = b & 0b00001111 typ = TYPE_MAP if n > self._max_map_len: raise ValueError("%s exceeds max_map_len(%s)", n, self._max_map_len) elif b == 0xc0: obj = None elif b == 0xc2: obj = False elif b == 0xc3: obj = True elif b == 0xc4: typ = TYPE_BIN self._reserve(1) n = self._buffer[self._buff_i] self._buff_i += 1 if n > self._max_bin_len: raise ValueError("%s exceeds max_bin_len(%s)" % (n, self._max_bin_len)) obj = self._read(n) elif b == 0xc5: typ = TYPE_BIN self._reserve(2) n = _unpack_from(">H", self._buffer, self._buff_i)[0] self._buff_i += 2 if n > self._max_bin_len: raise ValueError("%s exceeds max_bin_len(%s)" % (n, self._max_bin_len)) obj = self._read(n) elif b == 0xc6: typ = TYPE_BIN self._reserve(4) n = _unpack_from(">I", self._buffer, self._buff_i)[0] self._buff_i += 4 if n > self._max_bin_len: raise ValueError("%s exceeds max_bin_len(%s)" % (n, self._max_bin_len)) obj = self._read(n) elif b == 0xc7: # ext 8 typ = TYPE_EXT self._reserve(2) L, n = _unpack_from('Bb', self._buffer, self._buff_i) self._buff_i += 2 if L > self._max_ext_len: raise ValueError("%s exceeds max_ext_len(%s)" % (L, self._max_ext_len)) obj = self._read(L) elif b == 0xc8: # ext 16 typ = TYPE_EXT self._reserve(3) L, n = _unpack_from('>Hb', self._buffer, self._buff_i) self._buff_i += 3 if L > self._max_ext_len: raise ValueError("%s exceeds max_ext_len(%s)" % (L, self._max_ext_len)) obj = self._read(L) elif b == 0xc9: # ext 32 typ = TYPE_EXT self._reserve(5) L, n = _unpack_from('>Ib', self._buffer, self._buff_i) self._buff_i += 5 if L > self._max_ext_len: raise ValueError("%s exceeds max_ext_len(%s)" % (L, self._max_ext_len)) obj = self._read(L) elif b == 0xca: self._reserve(4) obj = _unpack_from(">f", self._buffer, self._buff_i)[0] self._buff_i += 4 elif b == 0xcb: self._reserve(8) obj = _unpack_from(">d", self._buffer, self._buff_i)[0] self._buff_i += 8 elif b == 0xcc: self._reserve(1) obj = self._buffer[self._buff_i] self._buff_i += 1 elif b == 0xcd: self._reserve(2) obj = _unpack_from(">H", self._buffer, self._buff_i)[0] self._buff_i += 2 elif b == 0xce: self._reserve(4) obj = _unpack_from(">I", self._buffer, self._buff_i)[0] self._buff_i += 4 elif b == 0xcf: self._reserve(8) obj = _unpack_from(">Q", self._buffer, self._buff_i)[0] self._buff_i += 8 elif b == 0xd0: self._reserve(1) obj = _unpack_from("b", self._buffer, self._buff_i)[0] self._buff_i += 1 elif b == 0xd1: self._reserve(2) obj = _unpack_from(">h", self._buffer, self._buff_i)[0] self._buff_i += 2 elif b == 0xd2: self._reserve(4) obj = _unpack_from(">i", self._buffer, self._buff_i)[0] self._buff_i += 4 elif b == 0xd3: self._reserve(8) obj = _unpack_from(">q", self._buffer, self._buff_i)[0] self._buff_i += 8 elif b == 0xd4: # fixext 1 typ = TYPE_EXT if self._max_ext_len < 1: raise ValueError("%s exceeds max_ext_len(%s)" % (1, self._max_ext_len)) self._reserve(2) n, obj = _unpack_from("b1s", self._buffer, self._buff_i) self._buff_i += 2 elif b == 0xd5: # fixext 2 typ = TYPE_EXT if self._max_ext_len < 2: raise ValueError("%s exceeds max_ext_len(%s)" % (2, self._max_ext_len)) self._reserve(3) n, obj = _unpack_from("b2s", self._buffer, self._buff_i) self._buff_i += 3 elif b == 0xd6: # fixext 4 typ = TYPE_EXT if self._max_ext_len < 4: raise ValueError("%s exceeds max_ext_len(%s)" % (4, self._max_ext_len)) self._reserve(5) n, obj = _unpack_from("b4s", self._buffer, self._buff_i) self._buff_i += 5 elif b == 0xd7: # fixext 8 typ = TYPE_EXT if self._max_ext_len < 8: raise ValueError("%s exceeds max_ext_len(%s)" % (8, self._max_ext_len)) self._reserve(9) n, obj = _unpack_from("b8s", self._buffer, self._buff_i) self._buff_i += 9 elif b == 0xd8: # fixext 16 typ = TYPE_EXT if self._max_ext_len < 16: raise ValueError("%s exceeds max_ext_len(%s)" % (16, self._max_ext_len)) self._reserve(17) n, obj = _unpack_from("b16s", self._buffer, self._buff_i) self._buff_i += 17 elif b == 0xd9: typ = TYPE_RAW self._reserve(1) n = self._buffer[self._buff_i] self._buff_i += 1 if n > self._max_str_len: raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len) obj = self._read(n) elif b == 0xda: typ = TYPE_RAW self._reserve(2) n, = _unpack_from(">H", self._buffer, self._buff_i) self._buff_i += 2 if n > self._max_str_len: raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len) obj = self._read(n) elif b == 0xdb: typ = TYPE_RAW self._reserve(4) n, = _unpack_from(">I", self._buffer, self._buff_i) self._buff_i += 4 if n > self._max_str_len: raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len) obj = self._read(n) elif b == 0xdc: typ = TYPE_ARRAY self._reserve(2) n, = _unpack_from(">H", self._buffer, self._buff_i) self._buff_i += 2 if n > self._max_array_len: raise ValueError("%s exceeds max_array_len(%s)", n, self._max_array_len) elif b == 0xdd: typ = TYPE_ARRAY self._reserve(4) n, = _unpack_from(">I", self._buffer, self._buff_i) self._buff_i += 4 if n > self._max_array_len: raise ValueError("%s exceeds max_array_len(%s)", n, self._max_array_len) elif b == 0xde: self._reserve(2) n, = _unpack_from(">H", self._buffer, self._buff_i) self._buff_i += 2 if n > self._max_map_len: raise ValueError("%s exceeds max_map_len(%s)", n, self._max_map_len) typ = TYPE_MAP elif b == 0xdf: self._reserve(4) n, = _unpack_from(">I", self._buffer, self._buff_i) self._buff_i += 4 if n > self._max_map_len: raise ValueError("%s exceeds max_map_len(%s)", n, self._max_map_len) typ = TYPE_MAP else: raise FormatError("Unknown header: 0x%x" % b) return typ, n, obj def _unpack(self, execute=EX_CONSTRUCT): typ, n, obj = self._read_header(execute) if execute == EX_READ_ARRAY_HEADER: if typ != TYPE_ARRAY: raise ValueError("Expected array") return n if execute == EX_READ_MAP_HEADER: if typ != TYPE_MAP: raise ValueError("Expected map") return n # TODO should we eliminate the recursion? if typ == TYPE_ARRAY: if execute == EX_SKIP: for i in xrange(n): # TODO check whether we need to call `list_hook` self._unpack(EX_SKIP) return ret = newlist_hint(n) for i in xrange(n): ret.append(self._unpack(EX_CONSTRUCT)) if self._list_hook is not None: ret = self._list_hook(ret) # TODO is the interaction between `list_hook` and `use_list` ok? return ret if self._use_list else tuple(ret) if typ == TYPE_MAP: if execute == EX_SKIP: for i in xrange(n): # TODO check whether we need to call hooks self._unpack(EX_SKIP) self._unpack(EX_SKIP) return if self._object_pairs_hook is not None: ret = self._object_pairs_hook( (self._unpack(EX_CONSTRUCT), self._unpack(EX_CONSTRUCT)) for _ in xrange(n)) else: ret = {} for _ in xrange(n): key = self._unpack(EX_CONSTRUCT) if self._strict_map_key and type(key) not in (unicode, bytes): raise ValueError("%s is not allowed for map key" % str(type(key))) ret[key] = self._unpack(EX_CONSTRUCT) if self._object_hook is not None: ret = self._object_hook(ret) return ret if execute == EX_SKIP: return if typ == TYPE_RAW: if self._encoding is not None: obj = obj.decode(self._encoding, self._unicode_errors) elif self._raw: obj = bytes(obj) else: obj = obj.decode('utf_8') return obj if typ == TYPE_EXT: return self._ext_hook(n, bytes(obj)) if typ == TYPE_BIN: return bytes(obj) assert typ == TYPE_IMMEDIATE return obj def __iter__(self): return self def __next__(self): try: ret = self._unpack(EX_CONSTRUCT) self._consume() return ret except OutOfData: self._consume() raise StopIteration except RecursionError: raise StackError next = __next__ def skip(self): self._unpack(EX_SKIP) self._consume() def unpack(self): try: ret = self._unpack(EX_CONSTRUCT) except RecursionError: raise StackError self._consume() return ret def read_array_header(self): ret = self._unpack(EX_READ_ARRAY_HEADER) self._consume() return ret def read_map_header(self): ret = self._unpack(EX_READ_MAP_HEADER) self._consume() return ret def tell(self): return self._stream_offset class Packer(object): """ MessagePack Packer usage: packer = Packer() astream.write(packer.pack(a)) astream.write(packer.pack(b)) Packer's constructor has some keyword arguments: :param callable default: Convert user type to builtin type that Packer supports. See also simplejson's document. :param bool use_single_float: Use single precision float type for float. (default: False) :param bool autoreset: Reset buffer after each pack and return its content as `bytes`. (default: True). If set this to false, use `bytes()` to get content and `.reset()` to clear buffer. :param bool use_bin_type: Use bin type introduced in msgpack spec 2.0 for bytes. It also enables str8 type for unicode. :param bool strict_types: If set to true, types will be checked to be exact. Derived classes from serializeable types will not be serialized and will be treated as unsupported type and forwarded to default. Additionally tuples will not be serialized as lists. This is useful when trying to implement accurate serialization for python types. :param str encoding: (deprecated) Convert unicode to bytes with this encoding. (default: 'utf-8') :param str unicode_errors: Error handler for encoding unicode. (default: 'strict') """ def __init__(self, default=None, encoding=None, unicode_errors=None, use_single_float=False, autoreset=True, use_bin_type=False, strict_types=False): if encoding is None: encoding = 'utf_8' else: warnings.warn( "encoding is deprecated, Use raw=False instead.", DeprecationWarning, stacklevel=2) if unicode_errors is None: unicode_errors = 'strict' self._strict_types = strict_types self._use_float = use_single_float self._autoreset = autoreset self._use_bin_type = use_bin_type self._encoding = encoding self._unicode_errors = unicode_errors self._buffer = StringIO() if default is not None: if not callable(default): raise TypeError("default must be callable") self._default = default def _pack(self, obj, nest_limit=DEFAULT_RECURSE_LIMIT, check=isinstance, check_type_strict=_check_type_strict): default_used = False if self._strict_types: check = check_type_strict list_types = list else: list_types = (list, tuple) while True: if nest_limit < 0: raise ValueError("recursion limit exceeded") if obj is None: return self._buffer.write(b"\xc0") if check(obj, bool): if obj: return self._buffer.write(b"\xc3") return self._buffer.write(b"\xc2") if check(obj, int_types): if 0 <= obj < 0x80: return self._buffer.write(struct.pack("B", obj)) if -0x20 <= obj < 0: return self._buffer.write(struct.pack("b", obj)) if 0x80 <= obj <= 0xff: return self._buffer.write(struct.pack("BB", 0xcc, obj)) if -0x80 <= obj < 0: return self._buffer.write(struct.pack(">Bb", 0xd0, obj)) if 0xff < obj <= 0xffff: return self._buffer.write(struct.pack(">BH", 0xcd, obj)) if -0x8000 <= obj < -0x80: return self._buffer.write(struct.pack(">Bh", 0xd1, obj)) if 0xffff < obj <= 0xffffffff: return self._buffer.write(struct.pack(">BI", 0xce, obj)) if -0x80000000 <= obj < -0x8000: return self._buffer.write(struct.pack(">Bi", 0xd2, obj)) if 0xffffffff < obj <= 0xffffffffffffffff: return self._buffer.write(struct.pack(">BQ", 0xcf, obj)) if -0x8000000000000000 <= obj < -0x80000000: return self._buffer.write(struct.pack(">Bq", 0xd3, obj)) if not default_used and self._default is not None: obj = self._default(obj) default_used = True continue raise OverflowError("Integer value out of range") if check(obj, (bytes, bytearray)): n = len(obj) if n >= 2**32: raise ValueError("%s is too large" % type(obj).__name__) self._pack_bin_header(n) return self._buffer.write(obj) if check(obj, unicode): if self._encoding is None: raise TypeError( "Can't encode unicode string: " "no encoding is specified") obj = obj.encode(self._encoding, self._unicode_errors) n = len(obj) if n >= 2**32: raise ValueError("String is too large") self._pack_raw_header(n) return self._buffer.write(obj) if check(obj, memoryview): n = len(obj) * obj.itemsize if n >= 2**32: raise ValueError("Memoryview is too large") self._pack_bin_header(n) return self._buffer.write(obj) if check(obj, float): if self._use_float: return self._buffer.write(struct.pack(">Bf", 0xca, obj)) return self._buffer.write(struct.pack(">Bd", 0xcb, obj)) if check(obj, ExtType): code = obj.code data = obj.data assert isinstance(code, int) assert isinstance(data, bytes) L = len(data) if L == 1: self._buffer.write(b'\xd4') elif L == 2: self._buffer.write(b'\xd5') elif L == 4: self._buffer.write(b'\xd6') elif L == 8: self._buffer.write(b'\xd7') elif L == 16: self._buffer.write(b'\xd8') elif L <= 0xff: self._buffer.write(struct.pack(">BB", 0xc7, L)) elif L <= 0xffff: self._buffer.write(struct.pack(">BH", 0xc8, L)) else: self._buffer.write(struct.pack(">BI", 0xc9, L)) self._buffer.write(struct.pack("b", code)) self._buffer.write(data) return if check(obj, list_types): n = len(obj) self._pack_array_header(n) for i in xrange(n): self._pack(obj[i], nest_limit - 1) return if check(obj, dict): return self._pack_map_pairs(len(obj), dict_iteritems(obj), nest_limit - 1) if not default_used and self._default is not None: obj = self._default(obj) default_used = 1 continue raise TypeError("Cannot serialize %r" % (obj, )) def pack(self, obj): try: self._pack(obj) except: self._buffer = StringIO() # force reset raise if self._autoreset: ret = self._buffer.getvalue() self._buffer = StringIO() return ret def pack_map_pairs(self, pairs): self._pack_map_pairs(len(pairs), pairs) if self._autoreset: ret = self._buffer.getvalue() self._buffer = StringIO() return ret def pack_array_header(self, n): if n >= 2**32: raise ValueError self._pack_array_header(n) if self._autoreset: ret = self._buffer.getvalue() self._buffer = StringIO() return ret def pack_map_header(self, n): if n >= 2**32: raise ValueError self._pack_map_header(n) if self._autoreset: ret = self._buffer.getvalue() self._buffer = StringIO() return ret def pack_ext_type(self, typecode, data): if not isinstance(typecode, int): raise TypeError("typecode must have int type.") if not 0 <= typecode <= 127: raise ValueError("typecode should be 0-127") if not isinstance(data, bytes): raise TypeError("data must have bytes type") L = len(data) if L > 0xffffffff: raise ValueError("Too large data") if L == 1: self._buffer.write(b'\xd4') elif L == 2: self._buffer.write(b'\xd5') elif L == 4: self._buffer.write(b'\xd6') elif L == 8: self._buffer.write(b'\xd7') elif L == 16: self._buffer.write(b'\xd8') elif L <= 0xff: self._buffer.write(b'\xc7' + struct.pack('B', L)) elif L <= 0xffff: self._buffer.write(b'\xc8' + struct.pack('>H', L)) else: self._buffer.write(b'\xc9' + struct.pack('>I', L)) self._buffer.write(struct.pack('B', typecode)) self._buffer.write(data) def _pack_array_header(self, n): if n <= 0x0f: return self._buffer.write(struct.pack('B', 0x90 + n)) if n <= 0xffff: return self._buffer.write(struct.pack(">BH", 0xdc, n)) if n <= 0xffffffff: return self._buffer.write(struct.pack(">BI", 0xdd, n)) raise ValueError("Array is too large") def _pack_map_header(self, n): if n <= 0x0f: return self._buffer.write(struct.pack('B', 0x80 + n)) if n <= 0xffff: return self._buffer.write(struct.pack(">BH", 0xde, n)) if n <= 0xffffffff: return self._buffer.write(struct.pack(">BI", 0xdf, n)) raise ValueError("Dict is too large") def _pack_map_pairs(self, n, pairs, nest_limit=DEFAULT_RECURSE_LIMIT): self._pack_map_header(n) for (k, v) in pairs: self._pack(k, nest_limit - 1) self._pack(v, nest_limit - 1) def _pack_raw_header(self, n): if n <= 0x1f: self._buffer.write(struct.pack('B', 0xa0 + n)) elif self._use_bin_type and n <= 0xff: self._buffer.write(struct.pack('>BB', 0xd9, n)) elif n <= 0xffff: self._buffer.write(struct.pack(">BH", 0xda, n)) elif n <= 0xffffffff: self._buffer.write(struct.pack(">BI", 0xdb, n)) else: raise ValueError('Raw is too large') def _pack_bin_header(self, n): if not self._use_bin_type: return self._pack_raw_header(n) elif n <= 0xff: return self._buffer.write(struct.pack('>BB', 0xc4, n)) elif n <= 0xffff: return self._buffer.write(struct.pack(">BH", 0xc5, n)) elif n <= 0xffffffff: return self._buffer.write(struct.pack(">BI", 0xc6, n)) else: raise ValueError('Bin is too large') def bytes(self): """Return internal buffer contents as bytes object""" return self._buffer.getvalue() def reset(self): """Reset internal buffer. This method is usaful only when autoreset=False. """ self._buffer = StringIO() def getbuffer(self): """Return view of internal buffer.""" if USING_STRINGBUILDER or PY2: return memoryview(self.bytes()) else: return self._buffer.getbuffer() site-packages/pip/_vendor/msgpack/_version.py000064400000000024151464452710015340 0ustar00version = (0, 6, 2) site-packages/pip/_vendor/msgpack/__pycache__/_version.cpython-38.pyc000064400000000266151464452710021636 0ustar00U .e@sdZdS))N)versionrr@/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/_version.pysite-packages/pip/_vendor/msgpack/__pycache__/fallback.cpython-38.pyc000064400000063241151464452710021553 0ustar00U .eN@sdZddlZddlZddlZejddkr@dZeefZddZ ndZeZe Z e Z ddZ ejd krpeZd d Znd d Zeed rddlmZzddlmZWn ek rddlmZYnXdZGdddeZndZddlmZddZddlmZmZm Z m!Z!m"Z"ddl#m$Z$dZ%dZ&dZ'dZ(dZ)dZ*dZ+dZ,dZ-dZ.dZ/e0e1fddZ2dd Z3d!d"Z4d#d$Z5ejd%krd,d&d'Z6nej7Z6Gd(d)d)eZ8Gd*d+d+eZ9dS)-z.Fallback pure Python implementation of msgpackNTcCs|SN)Z iteritemsdr@/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/fallback.pydict_iteritems srFcCs|Sr)itemsrrrrrs)cCs.t|jdko,t|jdto,|jddS)Nrz maximum recursion depth exceeded)lenargs isinstancestr startswitherrr_is_recursionerrorsrcCsdS)NTrrrrrrsZpypy_version_info) newlist_hint) BytesBuilder) StringBuilderc@s&eZdZd ddZddZddZdS) StringIOcCs,|r tt||_|j|nt|_dSr)rr builderappendselfsrrr__init__*szStringIO.__init__cCs6t|tr|}nt|tr&t|}|j|dSr)r memoryviewtobytes bytearraybytesrrrrrrwrite0s    zStringIO.writecCs |jSr)rZbuildrrrrgetvalue6szStringIO.getvalueN)r)__name__ __module__ __qualname__rr$r&rrrrr)s r)BytesIOcCsgSrr)sizerrr;rr,r ) BufferFull OutOfData ExtraData FormatError StackError)ExtTyper r icCs(|||kr|||kS|||kSdSrr)objttypetuplerrr_check_type_strictXs  r8cCsdz t|}Wn@tk rLtrFtt|}tjdt|tddnYnX|jdkr`t d|S)Nzusing old buffer interface to unpack %s; this leads to unpacking errors if slicing is used and will be removed in a future versionr  stacklevelr z$cannot unpack from multi-byte object) r TypeErrorPY2bufferwarningswarnr6RuntimeWarningitemsize ValueError)r4viewrrr_get_data_from_buffer_s   rDcKs$tjdtdd|}t|f|S)NzbDirect calling implementation's unpack() is deprecated, Use msgpack.unpack() or unpackb() instead.rr9)r>r?DeprecationWarningreadunpackb)streamkwargsdatarrrunpackqsrKc Kstddt|i|}||z |}WnJtk rHtdYn0tk rv}zt|rdtW5d}~XYnX| rt || |S)aw Unpack an object from `packed`. Raises ``ExtraData`` when *packed* contains extra bytes. Raises ``ValueError`` when *packed* is incomplete. Raises ``FormatError`` when *packed* is not valid msgpack. Raises ``StackError`` when *packed* contains too nested. Other exceptions can be raised during unpacking. See :class:`Unpacker` for options. Nmax_buffer_sizezUnpack failed: incomplete input)N) Unpackerr feed_unpackr.rBRecursionErrorrr1_got_extradatar/_get_extradata)ZpackedrIZunpackerretrrrrrGys    rG)rcCst|t||S)z.Explicit typcast for legacy struct.unpack_from)struct unpack_fromr#)fborrr _unpack_fromsr[c@seZdZdZdddddddddddedddddfddZd d Zd d Zd dZddZ ddZ ddZ ddZ e fddZe fddZddZddZeZdd Zd!d"Zd#d$Zd%d&Zd'd(ZdS))rMa Streaming unpacker. arguments: :param file_like: File-like object having `.read(n)` method. If specified, unpacker reads serialized data from it and :meth:`feed()` is not usable. :param int read_size: Used as `file_like.read(read_size)`. (default: `min(16*1024, max_buffer_size)`) :param bool use_list: If true, unpack msgpack array to Python list. Otherwise, unpack to Python tuple. (default: True) :param bool raw: If true, unpack msgpack raw to Python bytes (default). Otherwise, unpack to Python str (or unicode on Python 2) by decoding with UTF-8 encoding (recommended). Currently, the default is true, but it will be changed to false in near future. So you must specify it explicitly for keeping backward compatibility. *encoding* option which is deprecated overrides this option. :param bool strict_map_key: If true, only str or bytes are accepted for map (dict) keys. It's False by default for backward-compatibility. But it will be True from msgpack 1.0. :param callable object_hook: When specified, it should be callable. Unpacker calls it with a dict argument after unpacking msgpack map. (See also simplejson) :param callable object_pairs_hook: When specified, it should be callable. Unpacker calls it with a list of key-value pairs after unpacking msgpack map. (See also simplejson) :param str encoding: Encoding used for decoding msgpack raw. If it is None (default), msgpack raw is deserialized to Python bytes. :param str unicode_errors: (deprecated) Used for decoding msgpack raw with *encoding*. (default: `'strict'`) :param int max_buffer_size: Limits size of data waiting unpacked. 0 means system's INT_MAX (default). Raises `BufferFull` exception when it is insufficient. You should set this parameter when unpacking data from untrusted source. :param int max_str_len: Deprecated, use *max_buffer_size* instead. Limits max length of str. (default: max_buffer_size or 1024*1024) :param int max_bin_len: Deprecated, use *max_buffer_size* instead. Limits max length of bin. (default: max_buffer_size or 1024*1024) :param int max_array_len: Limits max length of array. (default: max_buffer_size or 128*1024) :param int max_map_len: Limits max length of map. (default: max_buffer_size//2 or 32*1024) :param int max_ext_len: Deprecated, use *max_buffer_size* instead. Limits max size of ext type. (default: max_buffer_size or 1024*1024) Example of streaming deserialize from file-like object:: unpacker = Unpacker(file_like, raw=False, max_buffer_size=10*1024*1024) for o in unpacker: process(o) Example of streaming deserialize from socket:: unpacker = Unpacker(raw=False, max_buffer_size=10*1024*1024) while True: buf = sock.recv(1024**2) if not buf: break unpacker.feed(buf) for o in unpacker: process(o) Raises ``ExtraData`` when *packed* contains extra bytes. Raises ``OutOfData`` when *packed* is incomplete. Raises ``FormatError`` when *packed* is not valid msgpack. Raises ``StackError`` when *packed* contains too nested. Other exceptions can be raised during unpacking. NrTFcCs| dk rtjdtdd| dkr$d} |dkr4d|_nt|jsFtd||_d|_t|_ d|_ d|_ | d krv| ptd } |d kr| pd }|d kr| pd }|d kr| dpd }|d kr| pd }| pd |_ ||j krt d|pt|j d|_t||_t||_| |_| |_||_||_||_||_| |_| |_||_||_||_||_d|_|dk rft|sftd|dk rt|std|dk rt|std|dk r|dk rtdt| stddS)N.encoding is deprecated, Use raw=False instead.rr9strictTz!`file_like.read` must be callableFrr\iiiiz.read_size must be smaller than max_buffer_sizei@z`list_hook` is not callablez`object_hook` is not callablez#`object_pairs_hook` is not callablez8object_pairs_hook and object_hook are mutually exclusivez`ext_hook` is not callable) r>r?rE_feedingcallablerFr; file_liker"_buffer_buff_i_buf_checkpoint_max_buffer_sizerBmin _read_sizebool_raw_strict_map_key _encoding_unicode_errors _use_list _list_hook _object_hook_object_pairs_hook _ext_hook _max_str_len _max_bin_len_max_array_len _max_map_len _max_ext_len_stream_offset)rraZ read_sizeZuse_listrawZstrict_map_keyZ object_hookZobject_pairs_hookZ list_hookencodingunicode_errorsrLZext_hookZ max_str_lenZ max_bin_lenZ max_array_lenZ max_map_lenZ max_ext_lenrrrrsp         zUnpacker.__init__cCsr|js tt|}t|j|jt||jkr4t|jdkrb|jd|j=|j|j8_d|_|j |dSNr) r_AssertionErrorrDr rbrcrer-rdextend)rZ next_bytesrCrrrrNPs  z Unpacker.feedcCs"|j|j|j7_|j|_dS)z+ Gets rid of the used parts of the buffer. N)rwrcrdr%rrr_consume_szUnpacker._consumecCs|jt|jkSr)rcr rbr%rrrrQdszUnpacker._got_extradatacCs|j|jdSr)rbrcr%rrrrRgszUnpacker._get_extradatacCs ||Sr)_readrnrrr read_bytesjszUnpacker.read_bytescCs,|||j}|||_|j|||Sr)_reservercrb)rrirrrrms  zUnpacker._readcCst|j|j|}|dkr dS|jr2|j|_t|jdkr`|jd|j=|j|j8_d|_| }|dkrt|j|}|j |}|sqt |t st |j|7_|t|8}qft|j||jkrd|_tdSr{) r rbrcr_rdr.maxrgrarFrr#r|)rrZ remain_bytesZ to_read_bytesZ read_datarrrrts,   zUnpacker._reservecCs& t}d}d}|d|j|j}|jd7_|d@dkrD|} n|d@dkr`d|dA} n|d@dkr|d@}t}||jkrtd ||j||} n~|d @d kr|d @}t}||j krtd ||j  nJ|d @dkr |d @}t }||j k rtd||j  n|dkrd} n|dkr.d}n|dkr@d}n|dkrt }|d|j|j}|jd7_||j krtd||j f||}n~|dkrt }|dtd|j|jd}|jd7_||j krtd||j f||}n|dkrjt }|dtd|j|jd}|jd7_||j kr\td||j f||}n|dkrt}|dtd|j|j\}}|jd7_||jkrtd||jf||}nL|dkr6t}|d td!|j|j\}}|jd 7_||jkr(td||jf||}n|d"krt}|d#td$|j|j\}}|jd#7_||jkrtd||jf||}n|d%kr|dtd&|j|jd}|jd7_nF|d'kr|d(td)|j|jd}|jd(7_n |d*krB|d|j|j}|jd7_n|d+kr||dtd|j|jd}|jd7_n|d,kr|dtd|j|jd}|jd7_nf|d-kr|d(td.|j|jd}|jd(7_n,|d/kr*|dtd0|j|jd}|jd7_n|d1krd|dtd2|j|jd}|jd7_n|d3kr|dtd4|j|jd}|jd7_n~|d5kr|d(td6|j|jd}|jd(7_nD|d7kr4t}|jdkrtdd|jf|dtd8|j|j\}}|jd7_n|d9krt}|jdkr`tdd|jf|d td:|j|j\}}|jd 7_n|d;krt}|jdkrtdd|jf|d#td<|j|j\}}|jd#7_n0|d=krHt}|jd(krtdd(|jf|d>td?|j|j\}}|jd>7_n|d@krt}|jdAkrttddA|jf|dBtdC|j|j\}}|jdB7_nx|dDkrt}|d|j|j}|jd7_||jkrtd ||j||}n|dEkr^t}|dtd|j|j\}|jd7_||jkrPtd ||j||}n|dFkrt}|dtd|j|j\}|jd7_||jkrtd ||j||}n^|dGk rt}|dtd|j|j\}|jd7_||j k rtd ||j n|dHk rht}|dtd|j|j\}|jd7_||j k rtd ||j n|dIk r|dtd|j|j\}|jd7_||j k rtd||j t }n`|dJk r|dtd|j|j\}|jd7_||j k r td||j t }n tdK||||fS)LNrr r\z%s exceeds max_str_len(%s)z%s exceeds max_array_len(%s)z%s exceeds max_map_len(%s)FTz%s exceeds max_bin_len(%s)r>Hr3>IZBbz%s exceeds max_ext_len(%s)r z>Hbr z>Ibz>fz>dz>QrYz>hz>iz>qZb1sZb2sZb4s Zb8sZb16szUnknown header: 0x%x)TYPE_IMMEDIATErrbrcTYPE_RAWrrrBr TYPE_ARRAYrtTYPE_MAPruTYPE_BINrsr[TYPE_EXTrvr0)rexecutetyprr4rYLrrr _read_headers                                                                                          zUnpacker._read_headerc s*|\}}}|tkr,|tkr(td|S|tkrH|tkrDtd|S|tkr|tkrtt|D]}tq`dSt |}t|D]}| t qj dk r |}j r|St|S|tkr|tkrt|D]}ttqdSjdk r"fddt|D}npi}t|D]L}t }jrjt|ttfkrjtdtt|t ||<q.jdk r|}|S|tkrdS|tkrjdk r|jj}njrt|}n |d}|S|tkr|t|S|tkrt|S|tks&t|S)NzExpected arrayz Expected mapc3s"|]}ttfVqdSr)rO EX_CONSTRUCT).0_r%rr sz#Unpacker._unpack..z%s is not allowed for map keyutf_8) rEX_READ_ARRAY_HEADERrrBEX_READ_MAP_HEADERrEX_SKIPxrangerOrrrrnrmr7rprjr6unicoder#rrorrkdecoderlrirrqrrr|) rrrrr4rrSrkeyrr%rrOhsl                      zUnpacker._unpackcCs|Srrr%rrr__iter__szUnpacker.__iter__cCsTz|t}||WStk r8|tYntk rNtYnXdSr)rOrr~r. StopIterationrPr1rrSrrr__next__s zUnpacker.__next__cCs|t|dSr)rOrr~r%rrrskips z Unpacker.skipcCs4z|t}Wntk r&tYnX||Sr)rOrrPr1r~rrrrrKs  zUnpacker.unpackcCs|t}||Sr)rOrr~rrrrread_array_headers zUnpacker.read_array_headercCs|t}||Sr)rOrr~rrrrread_map_headers zUnpacker.read_map_headercCs|jSr)rwr%rrrtellsz Unpacker.tell)r'r(r)__doc__r2rrNr~rQrRrrrrrrOrrnextrrKrrrrrrrrMsBa  R U @ rMc@seZdZdZd#ddZeeefddZd d Z d d Z d dZ ddZ ddZ ddZddZefddZddZddZddZdd Zd!d"ZdS)$Packerah MessagePack Packer usage: packer = Packer() astream.write(packer.pack(a)) astream.write(packer.pack(b)) Packer's constructor has some keyword arguments: :param callable default: Convert user type to builtin type that Packer supports. See also simplejson's document. :param bool use_single_float: Use single precision float type for float. (default: False) :param bool autoreset: Reset buffer after each pack and return its content as `bytes`. (default: True). If set this to false, use `bytes()` to get content and `.reset()` to clear buffer. :param bool use_bin_type: Use bin type introduced in msgpack spec 2.0 for bytes. It also enables str8 type for unicode. :param bool strict_types: If set to true, types will be checked to be exact. Derived classes from serializeable types will not be serialized and will be treated as unsupported type and forwarded to default. Additionally tuples will not be serialized as lists. This is useful when trying to implement accurate serialization for python types. :param str encoding: (deprecated) Convert unicode to bytes with this encoding. (default: 'utf-8') :param str unicode_errors: Error handler for encoding unicode. (default: 'strict') NFTcCsx|dkrd}ntjdtdd|dkr*d}||_||_||_||_||_||_t |_ |dk rnt |snt d||_ dS)Nrr]rr9r^zdefault must be callable)r>r?rE _strict_types _use_float _autoreset _use_bin_typerkrlrrbr`r;_default)rdefaultryrzZuse_single_floatZ autoresetZ use_bin_typeZ strict_typesrrrrs(zPacker.__init__c Csd}|jr|}t}nttf}|dkr,td|dkr@|jdS||trf|rZ|jdS|jdS||trd|krdkrnn|jt d|Sd |krdkrnn|jt d |Sd|krd krnn|jt d d |Sd|krdkr*nn|jt dd|Sd |krBdkr\nn|jt dd|Sd|krtdkrnn|jt dd|Sd|krdkrnn|jt dd|Sd|krdkrnn|jt dd|Sd|kr dkr$nn|jt dd|Sd |krBbr>BHriz>Bhr>BIriz>Birlz>BQrlz>BqrTzInteger value out of rangez%s is too largez5Can't encode unicode string: no encoding is specifiedzString is too largezMemoryview is too largez>Bfrz>Bdrr rr3rr>BBrrrzCannot serialize %r)(rlistr7rBrbr$rh int_typesrVpackr OverflowErrorr#r"r r6r'_pack_bin_headerrrkr;encoderl_pack_raw_headerr rAfloatrr2coderJrintr|_pack_array_headerr_packdict_pack_map_pairsr) rr4 nest_limitZcheckZcheck_type_strictZ default_usedZ list_typesrrrJrrrrrrs                                 z Packer._packcCsFz||Wnt|_YnX|jrB|j}t|_|SdSr)rrrbrr&)rr4rSrrrrs z Packer.packcCs0|t|||jr,|j}t|_|SdSr)rr rrbr&r)rpairsrSrrrpack_map_pairss  zPacker.pack_map_pairscCs6|dkr t|||jr2|j}t|_|SdSNr)rBrrrbr&rrrrSrrrpack_array_headers  zPacker.pack_array_headercCs6|dkr t|||jr2|j}t|_|SdSr)rB_pack_map_headerrrbr&rrrrrpack_map_headers  zPacker.pack_map_headercCsJt|tstdd|kr&dks0ntdt|tsBtdt|}|dkrZtd|dkrp|jd n|d kr|jd n|d kr|jd n|dkr|jdnt|dkr|jdn^|dkr|jdt d|n<|dkr|jdt d|n|jdt d||jt d||j|dS)Nztypecode must have int type.rztypecode should be 0-127zdata must have bytes typerzToo large datar rrrr3rrrrrrrrrr) rrr;rBr#r rbr$rVr)rtypecoderJrrrr pack_ext_types4   zPacker.pack_ext_typecCsh|dkr |jtdd|S|dkr>|jtdd|S|dkr\|jtdd |Std dS) NrrrrrrrrrzArray is too largerbr$rVrrBrrrrrszPacker._pack_array_headercCsh|dkr |jtdd|S|dkr>|jtdd|S|dkr\|jtdd |Std dS) NrrrrrrrrrzDict is too largerrrrrrszPacker._pack_map_headercCs<|||D](\}}|||d|||dqdS)Nr )rr)rrrrkvrrrrs  zPacker._pack_map_pairscCs|dkr"|jtdd|nn|jrH|dkrH|jtdd|nH|dkrh|jtdd |n(|d kr|jtd d |ntd dS)NrrrrrrrrrrrrzRaw is too large)rbr$rVrrrBrrrrrszPacker._pack_raw_headercCsv|js||S|dkr.|jtdd|S|dkrL|jtdd|S|dkrj|jtdd |Std dS) NrrrrrrrrrzBin is too large)rrrbr$rVrrBrrrrrs zPacker._pack_bin_headercCs |jS)z/Return internal buffer contents as bytes object)rbr&r%rrrr#sz Packer.bytescCs t|_dS)zYReset internal buffer. This method is usaful only when autoreset=False. N)rrbr%rrrresetsz Packer.resetcCs"tstrt|S|jSdS)zReturn view of internal buffer.N)USING_STRINGBUILDERr<r r#rb getbufferr%rrrr s zPacker.getbuffer)NNNFTFF)r'r(r)rrDEFAULT_RECURSE_LIMITrr8rrrrrrrrrrrr#rr rrrrrs0(  n       r)r):rsysrVr> version_infor<rZlongrrrrranger RuntimeErrorrPrhasattrZ__pypy__rZ__pypy__.buildersrr ImportErrorr objectrior* exceptionsr-r.r/r0r1r2rrrrrrrrrrr r6r7r8rDrKrGr[rWrMrrrrrsj         :site-packages/pip/_vendor/msgpack/__pycache__/__init__.cpython-38.pyc000064400000003576151464452710021560 0ustar00U .e@sddlmZddlTddlmZGdddeddZddlZejd r`dd l m Z m Z m Z nBzdd l m Z m Z m Z Wn(ek rdd l m Z m Z m Z YnXd d Zd dZddZeZe ZeZeZdS))version)*) namedtuplecs eZdZdZfddZZS)ExtTypez'ExtType represents ext type in msgpack.csVt|tstdt|ts$tdd|kr8dksBntdtt||||S)Nzcode must be intzdata must be bytesrzcode must be 0~127) isinstanceint TypeErrorbytes ValueErrorsuperr__new__)clscodedata __class__@/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/__init__.pyr s  zExtType.__new__)__name__ __module__ __qualname____doc__r __classcell__rrrrrsrz code dataNZMSGPACK_PUREPYTHON)PackerunpackbUnpackercKstf|}|||dS)zX Pack object `o` and write it to `stream` See :class:`Packer` for options. N)rwritepack)ostreamkwargsZpackerrrrrs rcKstf||S)zW Pack object `o` and return packed bytes See :class:`Packer` for options. )rr)r r"rrrpackb(sr#cKs|}t|f|S)z Unpack an object from `stream`. Raises `ExtraData` when `stream` contains extra bytes. See :class:`Unpacker` for options. )readr)r!r"rrrrunpack1sr%)Z_versionr exceptions collectionsrrosenvirongetZfallbackrrrZ _cmsgpack ImportErrorrr#r%loadloadsdumpdumpsrrrrs"       site-packages/pip/_vendor/msgpack/__pycache__/fallback.cpython-38.opt-1.pyc000064400000063024151464452710022511 0ustar00U .eN@sdZddlZddlZddlZejddkr@dZeefZddZ ndZeZe Z e Z ddZ ejd krpeZd d Znd d Zeed rddlmZzddlmZWn ek rddlmZYnXdZGdddeZndZddlmZddZddlmZmZm Z m!Z!m"Z"ddl#m$Z$dZ%dZ&dZ'dZ(dZ)dZ*dZ+dZ,dZ-dZ.dZ/e0e1fddZ2dd Z3d!d"Z4d#d$Z5ejd%krd,d&d'Z6nej7Z6Gd(d)d)eZ8Gd*d+d+eZ9dS)-z.Fallback pure Python implementation of msgpackNTcCs|SN)Z iteritemsdr@/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/fallback.pydict_iteritems srFcCs|Sr)itemsrrrrrs)cCs.t|jdko,t|jdto,|jddS)Nrz maximum recursion depth exceeded)lenargs isinstancestr startswitherrr_is_recursionerrorsrcCsdS)NTrrrrrrsZpypy_version_info) newlist_hint) BytesBuilder) StringBuilderc@s&eZdZd ddZddZddZdS) StringIOcCs,|r tt||_|j|nt|_dSr)rr builderappendselfsrrr__init__*szStringIO.__init__cCs6t|tr|}nt|tr&t|}|j|dSr)r memoryviewtobytes bytearraybytesrrrrrrwrite0s    zStringIO.writecCs |jSr)rZbuildrrrrgetvalue6szStringIO.getvalueN)r)__name__ __module__ __qualname__rr$r&rrrrr)s r)BytesIOcCsgSrr)sizerrr;rr,r ) BufferFull OutOfData ExtraData FormatError StackError)ExtTyper r icCs(|||kr|||kS|||kSdSrr)objttypetuplerrr_check_type_strictXs  r8cCsdz t|}Wn@tk rLtrFtt|}tjdt|tddnYnX|jdkr`t d|S)Nzusing old buffer interface to unpack %s; this leads to unpacking errors if slicing is used and will be removed in a future versionr  stacklevelr z$cannot unpack from multi-byte object) r TypeErrorPY2bufferwarningswarnr6RuntimeWarningitemsize ValueError)r4viewrrr_get_data_from_buffer_s   rDcKs$tjdtdd|}t|f|S)NzbDirect calling implementation's unpack() is deprecated, Use msgpack.unpack() or unpackb() instead.rr9)r>r?DeprecationWarningreadunpackb)streamkwargsdatarrrunpackqsrKc Kstddt|i|}||z |}WnJtk rHtdYn0tk rv}zt|rdtW5d}~XYnX| rt || |S)aw Unpack an object from `packed`. Raises ``ExtraData`` when *packed* contains extra bytes. Raises ``ValueError`` when *packed* is incomplete. Raises ``FormatError`` when *packed* is not valid msgpack. Raises ``StackError`` when *packed* contains too nested. Other exceptions can be raised during unpacking. See :class:`Unpacker` for options. Nmax_buffer_sizezUnpack failed: incomplete input)N) Unpackerr feed_unpackr.rBRecursionErrorrr1_got_extradatar/_get_extradata)ZpackedrIZunpackerretrrrrrGys    rG)rcCst|t||S)z.Explicit typcast for legacy struct.unpack_from)struct unpack_fromr#)fborrr _unpack_fromsr[c@seZdZdZdddddddddddedddddfddZd d Zd d Zd dZddZ ddZ ddZ ddZ e fddZe fddZddZddZeZdd Zd!d"Zd#d$Zd%d&Zd'd(ZdS))rMa Streaming unpacker. arguments: :param file_like: File-like object having `.read(n)` method. If specified, unpacker reads serialized data from it and :meth:`feed()` is not usable. :param int read_size: Used as `file_like.read(read_size)`. (default: `min(16*1024, max_buffer_size)`) :param bool use_list: If true, unpack msgpack array to Python list. Otherwise, unpack to Python tuple. (default: True) :param bool raw: If true, unpack msgpack raw to Python bytes (default). Otherwise, unpack to Python str (or unicode on Python 2) by decoding with UTF-8 encoding (recommended). Currently, the default is true, but it will be changed to false in near future. So you must specify it explicitly for keeping backward compatibility. *encoding* option which is deprecated overrides this option. :param bool strict_map_key: If true, only str or bytes are accepted for map (dict) keys. It's False by default for backward-compatibility. But it will be True from msgpack 1.0. :param callable object_hook: When specified, it should be callable. Unpacker calls it with a dict argument after unpacking msgpack map. (See also simplejson) :param callable object_pairs_hook: When specified, it should be callable. Unpacker calls it with a list of key-value pairs after unpacking msgpack map. (See also simplejson) :param str encoding: Encoding used for decoding msgpack raw. If it is None (default), msgpack raw is deserialized to Python bytes. :param str unicode_errors: (deprecated) Used for decoding msgpack raw with *encoding*. (default: `'strict'`) :param int max_buffer_size: Limits size of data waiting unpacked. 0 means system's INT_MAX (default). Raises `BufferFull` exception when it is insufficient. You should set this parameter when unpacking data from untrusted source. :param int max_str_len: Deprecated, use *max_buffer_size* instead. Limits max length of str. (default: max_buffer_size or 1024*1024) :param int max_bin_len: Deprecated, use *max_buffer_size* instead. Limits max length of bin. (default: max_buffer_size or 1024*1024) :param int max_array_len: Limits max length of array. (default: max_buffer_size or 128*1024) :param int max_map_len: Limits max length of map. (default: max_buffer_size//2 or 32*1024) :param int max_ext_len: Deprecated, use *max_buffer_size* instead. Limits max size of ext type. (default: max_buffer_size or 1024*1024) Example of streaming deserialize from file-like object:: unpacker = Unpacker(file_like, raw=False, max_buffer_size=10*1024*1024) for o in unpacker: process(o) Example of streaming deserialize from socket:: unpacker = Unpacker(raw=False, max_buffer_size=10*1024*1024) while True: buf = sock.recv(1024**2) if not buf: break unpacker.feed(buf) for o in unpacker: process(o) Raises ``ExtraData`` when *packed* contains extra bytes. Raises ``OutOfData`` when *packed* is incomplete. Raises ``FormatError`` when *packed* is not valid msgpack. Raises ``StackError`` when *packed* contains too nested. Other exceptions can be raised during unpacking. NrTFcCs| dk rtjdtdd| dkr$d} |dkr4d|_nt|jsFtd||_d|_t|_ d|_ d|_ | d krv| ptd } |d kr| pd }|d kr| pd }|d kr| dpd }|d kr| pd }| pd |_ ||j krt d|pt|j d|_t||_t||_| |_| |_||_||_||_||_| |_| |_||_||_||_||_d|_|dk rft|sftd|dk rt|std|dk rt|std|dk r|dk rtdt| stddS)N.encoding is deprecated, Use raw=False instead.rr9strictTz!`file_like.read` must be callableFrr\iiiiz.read_size must be smaller than max_buffer_sizei@z`list_hook` is not callablez`object_hook` is not callablez#`object_pairs_hook` is not callablez8object_pairs_hook and object_hook are mutually exclusivez`ext_hook` is not callable) r>r?rE_feedingcallablerFr; file_liker"_buffer_buff_i_buf_checkpoint_max_buffer_sizerBmin _read_sizebool_raw_strict_map_key _encoding_unicode_errors _use_list _list_hook _object_hook_object_pairs_hook _ext_hook _max_str_len _max_bin_len_max_array_len _max_map_len _max_ext_len_stream_offset)rraZ read_sizeZuse_listrawZstrict_map_keyZ object_hookZobject_pairs_hookZ list_hookencodingunicode_errorsrLZext_hookZ max_str_lenZ max_bin_lenZ max_array_lenZ max_map_lenZ max_ext_lenrrrrsp         zUnpacker.__init__cCsht|}t|j|jt||jkr*t|jdkrX|jd|j=|j|j8_d|_|j|dSNr)rDr rbrcrer-rdextend)rZ next_bytesrCrrrrNPs z Unpacker.feedcCs"|j|j|j7_|j|_dS)z+ Gets rid of the used parts of the buffer. N)rwrcrdr%rrr_consume_szUnpacker._consumecCs|jt|jkSr)rcr rbr%rrrrQdszUnpacker._got_extradatacCs|j|jdSr)rbrcr%rrrrRgszUnpacker._get_extradatacCs ||Sr)_readrnrrr read_bytesjszUnpacker.read_bytescCs,|||j}|||_|j|||Sr)_reservercrb)rrirrrr~ms  zUnpacker._readcCst|j|j|}|dkr dS|jr2|j|_t|jdkr`|jd|j=|j|j8_d|_| }|dkrt|j|}|j |}|sq|j|7_|t|8}qft|j||jkrd|_tdSr{) r rbrcr_rdr.maxrgrarF)rrZ remain_bytesZ to_read_bytesZ read_datarrrrts*   zUnpacker._reservecCs& t}d}d}|d|j|j}|jd7_|d@dkrD|} n|d@dkr`d|dA} n|d@dkr|d@}t}||jkrtd ||j||} n~|d @d kr|d @}t}||j krtd ||j  nJ|d @dkr |d @}t }||j k rtd||j  n|dkrd} n|dkr.d}n|dkr@d}n|dkrt }|d|j|j}|jd7_||j krtd||j f||}n~|dkrt }|dtd|j|jd}|jd7_||j krtd||j f||}n|dkrjt }|dtd|j|jd}|jd7_||j kr\td||j f||}n|dkrt}|dtd|j|j\}}|jd7_||jkrtd||jf||}nL|dkr6t}|d td!|j|j\}}|jd 7_||jkr(td||jf||}n|d"krt}|d#td$|j|j\}}|jd#7_||jkrtd||jf||}n|d%kr|dtd&|j|jd}|jd7_nF|d'kr|d(td)|j|jd}|jd(7_n |d*krB|d|j|j}|jd7_n|d+kr||dtd|j|jd}|jd7_n|d,kr|dtd|j|jd}|jd7_nf|d-kr|d(td.|j|jd}|jd(7_n,|d/kr*|dtd0|j|jd}|jd7_n|d1krd|dtd2|j|jd}|jd7_n|d3kr|dtd4|j|jd}|jd7_n~|d5kr|d(td6|j|jd}|jd(7_nD|d7kr4t}|jdkrtdd|jf|dtd8|j|j\}}|jd7_n|d9krt}|jdkr`tdd|jf|d td:|j|j\}}|jd 7_n|d;krt}|jdkrtdd|jf|d#td<|j|j\}}|jd#7_n0|d=krHt}|jd(krtdd(|jf|d>td?|j|j\}}|jd>7_n|d@krt}|jdAkrttddA|jf|dBtdC|j|j\}}|jdB7_nx|dDkrt}|d|j|j}|jd7_||jkrtd ||j||}n|dEkr^t}|dtd|j|j\}|jd7_||jkrPtd ||j||}n|dFkrt}|dtd|j|j\}|jd7_||jkrtd ||j||}n^|dGk rt}|dtd|j|j\}|jd7_||j k rtd ||j n|dHk rht}|dtd|j|j\}|jd7_||j k rtd ||j n|dIk r|dtd|j|j\}|jd7_||j k rtd||j t }n`|dJk r|dtd|j|j\}|jd7_||j k r td||j t }n tdK||||fS)LNrr r\z%s exceeds max_str_len(%s)z%s exceeds max_array_len(%s)z%s exceeds max_map_len(%s)FTz%s exceeds max_bin_len(%s)r>Hr3>IZBbz%s exceeds max_ext_len(%s)r z>Hbr z>Ibz>fz>dz>QrYz>hz>iz>qZb1sZb2sZb4s Zb8sZb16szUnknown header: 0x%x)TYPE_IMMEDIATErrbrcTYPE_RAWrrrBr~ TYPE_ARRAYrtTYPE_MAPruTYPE_BINrsr[TYPE_EXTrvr0)rexecutetyprr4rYLrrr _read_headers                                                                                          zUnpacker._read_headerc s|\}}}|tkr,|tkr(td|S|tkrH|tkrDtd|S|tkr|tkrtt|D]}tq`dSt |}t|D]}| t qj dk r |}j r|St|S|tkr|tkrt|D]}ttqdSjdk r"fddt|D}npi}t|D]L}t }jrjt|ttfkrjtdtt|t ||<q.jdk r|}|S|tkrdS|tkrjdk r|jj}njrt|}n |d}|S|tkr|t|S|tkrt|S|S)NzExpected arrayz Expected mapc3s"|]}ttfVqdSr)rO EX_CONSTRUCT).0_r%rr sz#Unpacker._unpack..z%s is not allowed for map keyutf_8)rEX_READ_ARRAY_HEADERrrBEX_READ_MAP_HEADERrEX_SKIPxrangerOrrrrnrmr7rprjr6unicoder#rrorrkdecoderlrirrqr) rrrrr4rrSrkeyrr%rrOhsj                      zUnpacker._unpackcCs|Srrr%rrr__iter__szUnpacker.__iter__cCsTz|t}||WStk r8|tYntk rNtYnXdSr)rOrr}r. StopIterationrPr1rrSrrr__next__s zUnpacker.__next__cCs|t|dSr)rOrr}r%rrrskips z Unpacker.skipcCs4z|t}Wntk r&tYnX||Sr)rOrrPr1r}rrrrrKs  zUnpacker.unpackcCs|t}||Sr)rOrr}rrrrread_array_headers zUnpacker.read_array_headercCs|t}||Sr)rOrr}rrrrread_map_headers zUnpacker.read_map_headercCs|jSr)rwr%rrrtellsz Unpacker.tell)r'r(r)__doc__r2rrNr}rQrRrr~rrrrOrrnextrrKrrrrrrrrMsBa  R U @ rMc@seZdZdZd#ddZeeefddZd d Z d d Z d dZ ddZ ddZ ddZddZefddZddZddZddZdd Zd!d"ZdS)$Packerah MessagePack Packer usage: packer = Packer() astream.write(packer.pack(a)) astream.write(packer.pack(b)) Packer's constructor has some keyword arguments: :param callable default: Convert user type to builtin type that Packer supports. See also simplejson's document. :param bool use_single_float: Use single precision float type for float. (default: False) :param bool autoreset: Reset buffer after each pack and return its content as `bytes`. (default: True). If set this to false, use `bytes()` to get content and `.reset()` to clear buffer. :param bool use_bin_type: Use bin type introduced in msgpack spec 2.0 for bytes. It also enables str8 type for unicode. :param bool strict_types: If set to true, types will be checked to be exact. Derived classes from serializeable types will not be serialized and will be treated as unsupported type and forwarded to default. Additionally tuples will not be serialized as lists. This is useful when trying to implement accurate serialization for python types. :param str encoding: (deprecated) Convert unicode to bytes with this encoding. (default: 'utf-8') :param str unicode_errors: Error handler for encoding unicode. (default: 'strict') NFTcCsx|dkrd}ntjdtdd|dkr*d}||_||_||_||_||_||_t |_ |dk rnt |snt d||_ dS)Nrr]rr9r^zdefault must be callable)r>r?rE _strict_types _use_float _autoreset _use_bin_typerkrlrrbr`r;_default)rdefaultryrzZuse_single_floatZ autoresetZ use_bin_typeZ strict_typesrrrrs(zPacker.__init__c Csdd}|jr|}t}nttf}|dkr,td|dkr@|jdS||trf|rZ|jdS|jdS||trd|krdkrnn|jt d|Sd |krdkrnn|jt d |Sd|krd krnn|jt d d |Sd|krdkr*nn|jt dd|Sd |krBdkr\nn|jt dd|Sd|krtdkrnn|jt dd|Sd|krdkrnn|jt dd|Sd|krdkrnn|jt dd|Sd|kr dkr$nn|jt dd|Sd |krBbr>BHriz>Bhr>BIriz>Birlz>BQrlz>BqrTzInteger value out of rangez%s is too largez5Can't encode unicode string: no encoding is specifiedzString is too largezMemoryview is too largez>Bfrz>Bdrr rr3rr>BBrrrzCannot serialize %r)%rlistr7rBrbr$rh int_typesrVpackr OverflowErrorr#r"r r6r'_pack_bin_headerrrkr;encoderl_pack_raw_headerr rAfloatrr2coderJ_pack_array_headerr_packdict_pack_map_pairsr) rr4 nest_limitZcheckZcheck_type_strictZ default_usedZ list_typesrrrJrrrrrrs                                 z Packer._packcCsFz||Wnt|_YnX|jrB|j}t|_|SdSr)rrrbrr&)rr4rSrrrrs z Packer.packcCs0|t|||jr,|j}t|_|SdSr)rr rrbr&r)rpairsrSrrrpack_map_pairss  zPacker.pack_map_pairscCs6|dkr t|||jr2|j}t|_|SdSNr)rBrrrbr&rrrrSrrrpack_array_headers  zPacker.pack_array_headercCs6|dkr t|||jr2|j}t|_|SdSr)rB_pack_map_headerrrbr&rrrrrpack_map_headers  zPacker.pack_map_headercCsJt|tstdd|kr&dks0ntdt|tsBtdt|}|dkrZtd|dkrp|jd n|d kr|jd n|d kr|jd n|dkr|jdnt|dkr|jdn^|dkr|jdt d|n<|dkr|jdt d|n|jdt d||jt d||j|dS)Nztypecode must have int type.rztypecode should be 0-127zdata must have bytes typerzToo large datar rrrr3rrrrrrrrrr) rintr;rBr#r rbr$rVr)rtypecoderJrrrr pack_ext_types4   zPacker.pack_ext_typecCsh|dkr |jtdd|S|dkr>|jtdd|S|dkr\|jtdd |Std dS) NrrrrrrrrrzArray is too largerbr$rVrrBrrrrrszPacker._pack_array_headercCsh|dkr |jtdd|S|dkr>|jtdd|S|dkr\|jtdd |Std dS) NrrrrrrrrrzDict is too largerrrrrrszPacker._pack_map_headercCs<|||D](\}}|||d|||dqdS)Nr )rr)rrrrkvrrrrs  zPacker._pack_map_pairscCs|dkr"|jtdd|nn|jrH|dkrH|jtdd|nH|dkrh|jtdd |n(|d kr|jtd d |ntd dS)NrrrrrrrrrrrrzRaw is too large)rbr$rVrrrBrrrrrszPacker._pack_raw_headercCsv|js||S|dkr.|jtdd|S|dkrL|jtdd|S|dkrj|jtdd |Std dS) NrrrrrrrrrzBin is too large)rrrbr$rVrrBrrrrrs zPacker._pack_bin_headercCs |jS)z/Return internal buffer contents as bytes object)rbr&r%rrrr#sz Packer.bytescCs t|_dS)zYReset internal buffer. This method is usaful only when autoreset=False. N)rrbr%rrrresetsz Packer.resetcCs"tstrt|S|jSdS)zReturn view of internal buffer.N)USING_STRINGBUILDERr<r r#rb getbufferr%rrrr s zPacker.getbuffer)NNNFTFF)r'r(r)rrDEFAULT_RECURSE_LIMITrr8rrrrrrrrrrrr#rr rrrrrs0(  n       r)r):rsysrVr> version_infor<rZlongrrrrranger RuntimeErrorrPrhasattrZ__pypy__rZ__pypy__.buildersrr ImportErrorrobjectrior* exceptionsr-r.r/r0r1r2rrrrrrrrrrr r6r7r8rDrKrGr[rWrMrrrrrsj         :site-packages/pip/_vendor/msgpack/__pycache__/exceptions.cpython-38.pyc000064400000003430151464452710022167 0ustar00U .e9@sxGdddeZGdddeZGdddeZGdddeeZGdd d eeZeZGd d d eZeZ eZ e Z d S) c@seZdZdZdS)UnpackExceptionzBase class for some exceptions raised while unpacking. NOTE: unpack may raise exception other than subclass of UnpackException. If you want to catch all error, catch Exception instead. N__name__ __module__ __qualname____doc__rrB/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/exceptions.pyrsrc@s eZdZdS) BufferFullNrrrrrrrr sr c@s eZdZdS) OutOfDataNr rrrrr sr c@seZdZdZdS) FormatErrorzInvalid msgpack formatNrrrrrr sr c@seZdZdZdS) StackErrorz Too nestedNrrrrrr sr c@s eZdZdZddZddZdS) ExtraDatazExtraData is raised when there is trailing data. This exception is raised while only one-shot (not streaming) unpack. cCs||_||_dS)N)unpackedextra)selfrrrrr__init__%szExtraData.__init__cCsdS)Nzunpack(b) received extra data.r)rrrr__str__)szExtraData.__str__N)rrrrrrrrrrrsrN) Exceptionrr r ValueErrorr r ZUnpackValueErrorrZ PackExceptionZPackValueError OverflowErrorZPackOverflowErrorrrrrs site-packages/pip/_vendor/msgpack/__pycache__/__init__.cpython-38.opt-1.pyc000064400000003576151464452710022517 0ustar00U .e@sddlmZddlTddlmZGdddeddZddlZejd r`dd l m Z m Z m Z nBzdd l m Z m Z m Z Wn(ek rdd l m Z m Z m Z YnXd d Zd dZddZeZe ZeZeZdS))version)*) namedtuplecs eZdZdZfddZZS)ExtTypez'ExtType represents ext type in msgpack.csVt|tstdt|ts$tdd|kr8dksBntdtt||||S)Nzcode must be intzdata must be bytesrzcode must be 0~127) isinstanceint TypeErrorbytes ValueErrorsuperr__new__)clscodedata __class__@/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/__init__.pyr s  zExtType.__new__)__name__ __module__ __qualname____doc__r __classcell__rrrrrsrz code dataNZMSGPACK_PUREPYTHON)PackerunpackbUnpackercKstf|}|||dS)zX Pack object `o` and write it to `stream` See :class:`Packer` for options. N)rwritepack)ostreamkwargsZpackerrrrrs rcKstf||S)zW Pack object `o` and return packed bytes See :class:`Packer` for options. )rr)r r"rrrpackb(sr#cKs|}t|f|S)z Unpack an object from `stream`. Raises `ExtraData` when `stream` contains extra bytes. See :class:`Unpacker` for options. )readr)r!r"rrrrunpack1sr%)Z_versionr exceptions collectionsrrosenvirongetZfallbackrrrZ _cmsgpack ImportErrorrr#r%loadloadsdumpdumpsrrrrs"       site-packages/pip/_vendor/msgpack/__pycache__/exceptions.cpython-38.opt-1.pyc000064400000003430151464452710023126 0ustar00U .e9@sxGdddeZGdddeZGdddeZGdddeeZGdd d eeZeZGd d d eZeZ eZ e Z d S) c@seZdZdZdS)UnpackExceptionzBase class for some exceptions raised while unpacking. NOTE: unpack may raise exception other than subclass of UnpackException. If you want to catch all error, catch Exception instead. N__name__ __module__ __qualname____doc__rrB/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/exceptions.pyrsrc@s eZdZdS) BufferFullNrrrrrrrr sr c@s eZdZdS) OutOfDataNr rrrrr sr c@seZdZdZdS) FormatErrorzInvalid msgpack formatNrrrrrr sr c@seZdZdZdS) StackErrorz Too nestedNrrrrrr sr c@s eZdZdZddZddZdS) ExtraDatazExtraData is raised when there is trailing data. This exception is raised while only one-shot (not streaming) unpack. cCs||_||_dS)N)unpackedextra)selfrrrrr__init__%szExtraData.__init__cCsdS)Nzunpack(b) received extra data.r)rrrr__str__)szExtraData.__str__N)rrrrrrrrrrrsrN) Exceptionrr r ValueErrorr r ZUnpackValueErrorrZ PackExceptionZPackValueError OverflowErrorZPackOverflowErrorrrrrs site-packages/pip/_vendor/msgpack/__pycache__/_version.cpython-38.opt-1.pyc000064400000000266151464452710022575 0ustar00U .e@sdZdS))N)versionrr@/usr/lib/python3.8/site-packages/pip/_vendor/msgpack/_version.pysite-packages/pip/_vendor/msgpack/exceptions.py000064400000002071151464452710015701 0ustar00class UnpackException(Exception): """Base class for some exceptions raised while unpacking. NOTE: unpack may raise exception other than subclass of UnpackException. If you want to catch all error, catch Exception instead. """ class BufferFull(UnpackException): pass class OutOfData(UnpackException): pass class FormatError(ValueError, UnpackException): """Invalid msgpack format""" class StackError(ValueError, UnpackException): """Too nested""" # Deprecated. Use ValueError instead UnpackValueError = ValueError class ExtraData(UnpackValueError): """ExtraData is raised when there is trailing data. This exception is raised while only one-shot (not streaming) unpack. """ def __init__(self, unpacked, extra): self.unpacked = unpacked self.extra = extra def __str__(self): return "unpack(b) received extra data." # Deprecated. Use Exception instead to catch all exception during packing. PackException = Exception PackValueError = ValueError PackOverflowError = OverflowError site-packages/pip/_vendor/msgpack/__init__.py000064400000002777151464452710015274 0ustar00# coding: utf-8 from ._version import version from .exceptions import * from collections import namedtuple class ExtType(namedtuple('ExtType', 'code data')): """ExtType represents ext type in msgpack.""" def __new__(cls, code, data): if not isinstance(code, int): raise TypeError("code must be int") if not isinstance(data, bytes): raise TypeError("data must be bytes") if not 0 <= code <= 127: raise ValueError("code must be 0~127") return super(ExtType, cls).__new__(cls, code, data) import os if os.environ.get('MSGPACK_PUREPYTHON'): from .fallback import Packer, unpackb, Unpacker else: try: from ._cmsgpack import Packer, unpackb, Unpacker except ImportError: from .fallback import Packer, unpackb, Unpacker def pack(o, stream, **kwargs): """ Pack object `o` and write it to `stream` See :class:`Packer` for options. """ packer = Packer(**kwargs) stream.write(packer.pack(o)) def packb(o, **kwargs): """ Pack object `o` and return packed bytes See :class:`Packer` for options. """ return Packer(**kwargs).pack(o) def unpack(stream, **kwargs): """ Unpack an object from `stream`. Raises `ExtraData` when `stream` contains extra bytes. See :class:`Unpacker` for options. """ data = stream.read() return unpackb(data, **kwargs) # alias for compatibility to simplejson/marshal/pickle. load = unpack loads = unpackb dump = pack dumps = packb site-packages/pip/_vendor/packaging/requirements.py000064400000011134151464452710016542 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import string import re from pip._vendor.pyparsing import stringStart, stringEnd, originalTextFor, ParseException from pip._vendor.pyparsing import ZeroOrMore, Word, Optional, Regex, Combine from pip._vendor.pyparsing import Literal as L # noqa from pip._vendor.six.moves.urllib import parse as urlparse from .markers import MARKER_EXPR, Marker from .specifiers import LegacySpecifier, Specifier, SpecifierSet class InvalidRequirement(ValueError): """ An invalid requirement was found, users should refer to PEP 508. """ ALPHANUM = Word(string.ascii_letters + string.digits) LBRACKET = L("[").suppress() RBRACKET = L("]").suppress() LPAREN = L("(").suppress() RPAREN = L(")").suppress() COMMA = L(",").suppress() SEMICOLON = L(";").suppress() AT = L("@").suppress() PUNCTUATION = Word("-_.") IDENTIFIER_END = ALPHANUM | (ZeroOrMore(PUNCTUATION) + ALPHANUM) IDENTIFIER = Combine(ALPHANUM + ZeroOrMore(IDENTIFIER_END)) NAME = IDENTIFIER("name") EXTRA = IDENTIFIER URI = Regex(r"[^ ]+")("url") URL = AT + URI EXTRAS_LIST = EXTRA + ZeroOrMore(COMMA + EXTRA) EXTRAS = (LBRACKET + Optional(EXTRAS_LIST) + RBRACKET)("extras") VERSION_PEP440 = Regex(Specifier._regex_str, re.VERBOSE | re.IGNORECASE) VERSION_LEGACY = Regex(LegacySpecifier._regex_str, re.VERBOSE | re.IGNORECASE) VERSION_ONE = VERSION_PEP440 ^ VERSION_LEGACY VERSION_MANY = Combine( VERSION_ONE + ZeroOrMore(COMMA + VERSION_ONE), joinString=",", adjacent=False )("_raw_spec") _VERSION_SPEC = Optional(((LPAREN + VERSION_MANY + RPAREN) | VERSION_MANY)) _VERSION_SPEC.setParseAction(lambda s, l, t: t._raw_spec or "") VERSION_SPEC = originalTextFor(_VERSION_SPEC)("specifier") VERSION_SPEC.setParseAction(lambda s, l, t: t[1]) MARKER_EXPR = originalTextFor(MARKER_EXPR())("marker") MARKER_EXPR.setParseAction( lambda s, l, t: Marker(s[t._original_start : t._original_end]) ) MARKER_SEPARATOR = SEMICOLON MARKER = MARKER_SEPARATOR + MARKER_EXPR VERSION_AND_MARKER = VERSION_SPEC + Optional(MARKER) URL_AND_MARKER = URL + Optional(MARKER) NAMED_REQUIREMENT = NAME + Optional(EXTRAS) + (URL_AND_MARKER | VERSION_AND_MARKER) REQUIREMENT = stringStart + NAMED_REQUIREMENT + stringEnd # pyparsing isn't thread safe during initialization, so we do it eagerly, see # issue #104 REQUIREMENT.parseString("x[]") class Requirement(object): """Parse a requirement. Parse a given requirement string into its parts, such as name, specifier, URL, and extras. Raises InvalidRequirement on a badly-formed requirement string. """ # TODO: Can we test whether something is contained within a requirement? # If so how do we do that? Do we need to test against the _name_ of # the thing as well as the version? What about the markers? # TODO: Can we normalize the name and extra name? def __init__(self, requirement_string): try: req = REQUIREMENT.parseString(requirement_string) except ParseException as e: raise InvalidRequirement( 'Parse error at "{0!r}": {1}'.format( requirement_string[e.loc : e.loc + 8], e.msg ) ) self.name = req.name if req.url: parsed_url = urlparse.urlparse(req.url) if parsed_url.scheme == "file": if urlparse.urlunparse(parsed_url) != req.url: raise InvalidRequirement("Invalid URL given") elif not (parsed_url.scheme and parsed_url.netloc) or ( not parsed_url.scheme and not parsed_url.netloc ): raise InvalidRequirement("Invalid URL: {0}".format(req.url)) self.url = req.url else: self.url = None self.extras = set(req.extras.asList() if req.extras else []) self.specifier = SpecifierSet(req.specifier) self.marker = req.marker if req.marker else None def __str__(self): parts = [self.name] if self.extras: parts.append("[{0}]".format(",".join(sorted(self.extras)))) if self.specifier: parts.append(str(self.specifier)) if self.url: parts.append("@ {0}".format(self.url)) if self.marker: parts.append(" ") if self.marker: parts.append("; {0}".format(self.marker)) return "".join(parts) def __repr__(self): return "".format(str(self)) site-packages/pip/_vendor/packaging/_structures.py000064400000002610151464452710016400 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function class Infinity(object): def __repr__(self): return "Infinity" def __hash__(self): return hash(repr(self)) def __lt__(self, other): return False def __le__(self, other): return False def __eq__(self, other): return isinstance(other, self.__class__) def __ne__(self, other): return not isinstance(other, self.__class__) def __gt__(self, other): return True def __ge__(self, other): return True def __neg__(self): return NegativeInfinity Infinity = Infinity() class NegativeInfinity(object): def __repr__(self): return "-Infinity" def __hash__(self): return hash(repr(self)) def __lt__(self, other): return True def __le__(self, other): return True def __eq__(self, other): return isinstance(other, self.__class__) def __ne__(self, other): return not isinstance(other, self.__class__) def __gt__(self, other): return False def __ge__(self, other): return False def __neg__(self): return Infinity NegativeInfinity = NegativeInfinity() site-packages/pip/_vendor/packaging/utils.py000064400000002760151464452710015164 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import re from .version import InvalidVersion, Version _canonicalize_regex = re.compile(r"[-_.]+") def canonicalize_name(name): # This is taken from PEP 503. return _canonicalize_regex.sub("-", name).lower() def canonicalize_version(version): """ This is very similar to Version.__str__, but has one subtle differences with the way it handles the release segment. """ try: version = Version(version) except InvalidVersion: # Legacy versions cannot be normalized return version parts = [] # Epoch if version.epoch != 0: parts.append("{0}!".format(version.epoch)) # Release segment # NB: This strips trailing '.0's to normalize parts.append(re.sub(r"(\.0)+$", "", ".".join(str(x) for x in version.release))) # Pre-release if version.pre is not None: parts.append("".join(str(x) for x in version.pre)) # Post-release if version.post is not None: parts.append(".post{0}".format(version.post)) # Development release if version.dev is not None: parts.append(".dev{0}".format(version.dev)) # Local version segment if version.local is not None: parts.append("+{0}".format(version.local)) return "".join(parts) site-packages/pip/_vendor/packaging/tags.py000064400000031205151464452710014756 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import import distutils.util try: from importlib.machinery import EXTENSION_SUFFIXES except ImportError: # pragma: no cover import imp EXTENSION_SUFFIXES = [x[0] for x in imp.get_suffixes()] del imp import platform import re import sys import sysconfig import warnings INTERPRETER_SHORT_NAMES = { "python": "py", # Generic. "cpython": "cp", "pypy": "pp", "ironpython": "ip", "jython": "jy", } _32_BIT_INTERPRETER = sys.maxsize <= 2 ** 32 class Tag(object): __slots__ = ["_interpreter", "_abi", "_platform"] def __init__(self, interpreter, abi, platform): self._interpreter = interpreter.lower() self._abi = abi.lower() self._platform = platform.lower() @property def interpreter(self): return self._interpreter @property def abi(self): return self._abi @property def platform(self): return self._platform def __eq__(self, other): return ( (self.platform == other.platform) and (self.abi == other.abi) and (self.interpreter == other.interpreter) ) def __hash__(self): return hash((self._interpreter, self._abi, self._platform)) def __str__(self): return "{}-{}-{}".format(self._interpreter, self._abi, self._platform) def __repr__(self): return "<{self} @ {self_id}>".format(self=self, self_id=id(self)) def parse_tag(tag): tags = set() interpreters, abis, platforms = tag.split("-") for interpreter in interpreters.split("."): for abi in abis.split("."): for platform_ in platforms.split("."): tags.add(Tag(interpreter, abi, platform_)) return frozenset(tags) def _normalize_string(string): return string.replace(".", "_").replace("-", "_") def _cpython_interpreter(py_version): # TODO: Is using py_version_nodot for interpreter version critical? return "cp{major}{minor}".format(major=py_version[0], minor=py_version[1]) def _cpython_abis(py_version): abis = [] version = "{}{}".format(*py_version[:2]) debug = pymalloc = ucs4 = "" with_debug = sysconfig.get_config_var("Py_DEBUG") has_refcount = hasattr(sys, "gettotalrefcount") # Windows doesn't set Py_DEBUG, so checking for support of debug-compiled # extension modules is the best option. # https://github.com/pypa/pip/issues/3383#issuecomment-173267692 has_ext = "_d.pyd" in EXTENSION_SUFFIXES if with_debug or (with_debug is None and (has_refcount or has_ext)): debug = "d" if py_version < (3, 8): with_pymalloc = sysconfig.get_config_var("WITH_PYMALLOC") if with_pymalloc or with_pymalloc is None: pymalloc = "m" if py_version < (3, 3): unicode_size = sysconfig.get_config_var("Py_UNICODE_SIZE") if unicode_size == 4 or ( unicode_size is None and sys.maxunicode == 0x10FFFF ): ucs4 = "u" elif debug: # Debug builds can also load "normal" extension modules. # We can also assume no UCS-4 or pymalloc requirement. abis.append("cp{version}".format(version=version)) abis.insert( 0, "cp{version}{debug}{pymalloc}{ucs4}".format( version=version, debug=debug, pymalloc=pymalloc, ucs4=ucs4 ), ) return abis def _cpython_tags(py_version, interpreter, abis, platforms): for abi in abis: for platform_ in platforms: yield Tag(interpreter, abi, platform_) for tag in (Tag(interpreter, "abi3", platform_) for platform_ in platforms): yield tag for tag in (Tag(interpreter, "none", platform_) for platform_ in platforms): yield tag # PEP 384 was first implemented in Python 3.2. for minor_version in range(py_version[1] - 1, 1, -1): for platform_ in platforms: interpreter = "cp{major}{minor}".format( major=py_version[0], minor=minor_version ) yield Tag(interpreter, "abi3", platform_) def _pypy_interpreter(): return "pp{py_major}{pypy_major}{pypy_minor}".format( py_major=sys.version_info[0], pypy_major=sys.pypy_version_info.major, pypy_minor=sys.pypy_version_info.minor, ) def _generic_abi(): abi = sysconfig.get_config_var("SOABI") if abi: return _normalize_string(abi) else: return "none" def _pypy_tags(py_version, interpreter, abi, platforms): for tag in (Tag(interpreter, abi, platform) for platform in platforms): yield tag for tag in (Tag(interpreter, "none", platform) for platform in platforms): yield tag def _generic_tags(interpreter, py_version, abi, platforms): for tag in (Tag(interpreter, abi, platform) for platform in platforms): yield tag if abi != "none": tags = (Tag(interpreter, "none", platform_) for platform_ in platforms) for tag in tags: yield tag def _py_interpreter_range(py_version): """ Yield Python versions in descending order. After the latest version, the major-only version will be yielded, and then all following versions up to 'end'. """ yield "py{major}{minor}".format(major=py_version[0], minor=py_version[1]) yield "py{major}".format(major=py_version[0]) for minor in range(py_version[1] - 1, -1, -1): yield "py{major}{minor}".format(major=py_version[0], minor=minor) def _independent_tags(interpreter, py_version, platforms): """ Return the sequence of tags that are consistent across implementations. The tags consist of: - py*-none- - -none-any - py*-none-any """ for version in _py_interpreter_range(py_version): for platform_ in platforms: yield Tag(version, "none", platform_) yield Tag(interpreter, "none", "any") for version in _py_interpreter_range(py_version): yield Tag(version, "none", "any") def _mac_arch(arch, is_32bit=_32_BIT_INTERPRETER): if not is_32bit: return arch if arch.startswith("ppc"): return "ppc" return "i386" def _mac_binary_formats(version, cpu_arch): formats = [cpu_arch] if cpu_arch == "x86_64": if version < (10, 4): return [] formats.extend(["intel", "fat64", "fat32"]) elif cpu_arch == "i386": if version < (10, 4): return [] formats.extend(["intel", "fat32", "fat"]) elif cpu_arch == "ppc64": # TODO: Need to care about 32-bit PPC for ppc64 through 10.2? if version > (10, 5) or version < (10, 4): return [] formats.append("fat64") elif cpu_arch == "ppc": if version > (10, 6): return [] formats.extend(["fat32", "fat"]) formats.append("universal") return formats def _mac_platforms(version=None, arch=None): version_str, _, cpu_arch = platform.mac_ver() if version is None: version = tuple(map(int, version_str.split(".")[:2])) if arch is None: arch = _mac_arch(cpu_arch) platforms = [] for minor_version in range(version[1], -1, -1): compat_version = version[0], minor_version binary_formats = _mac_binary_formats(compat_version, arch) for binary_format in binary_formats: platforms.append( "macosx_{major}_{minor}_{binary_format}".format( major=compat_version[0], minor=compat_version[1], binary_format=binary_format, ) ) return platforms # From PEP 513. def _is_manylinux_compatible(name, glibc_version): # Check for presence of _manylinux module. try: import _manylinux return bool(getattr(_manylinux, name + "_compatible")) except (ImportError, AttributeError): # Fall through to heuristic check below. pass return _have_compatible_glibc(*glibc_version) def _glibc_version_string(): # Returns glibc version string, or None if not using glibc. import ctypes # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen # manpage says, "If filename is NULL, then the returned handle is for the # main program". This way we can let the linker do the work to figure out # which libc our process is actually using. process_namespace = ctypes.CDLL(None) try: gnu_get_libc_version = process_namespace.gnu_get_libc_version except AttributeError: # Symbol doesn't exist -> therefore, we are not linked to # glibc. return None # Call gnu_get_libc_version, which returns a string like "2.5" gnu_get_libc_version.restype = ctypes.c_char_p version_str = gnu_get_libc_version() # py2 / py3 compatibility: if not isinstance(version_str, str): version_str = version_str.decode("ascii") return version_str # Separated out from have_compatible_glibc for easier unit testing. def _check_glibc_version(version_str, required_major, minimum_minor): # Parse string and check against requested version. # # We use a regexp instead of str.split because we want to discard any # random junk that might come after the minor version -- this might happen # in patched/forked versions of glibc (e.g. Linaro's version of glibc # uses version strings like "2.20-2014.11"). See gh-3588. m = re.match(r"(?P[0-9]+)\.(?P[0-9]+)", version_str) if not m: warnings.warn( "Expected glibc version with 2 components major.minor," " got: %s" % version_str, RuntimeWarning, ) return False return ( int(m.group("major")) == required_major and int(m.group("minor")) >= minimum_minor ) def _have_compatible_glibc(required_major, minimum_minor): version_str = _glibc_version_string() if version_str is None: return False return _check_glibc_version(version_str, required_major, minimum_minor) def _linux_platforms(is_32bit=_32_BIT_INTERPRETER): linux = _normalize_string(distutils.util.get_platform()) if linux == "linux_x86_64" and is_32bit: linux = "linux_i686" manylinux_support = ( ("manylinux2014", (2, 17)), # CentOS 7 w/ glibc 2.17 (PEP 599) ("manylinux2010", (2, 12)), # CentOS 6 w/ glibc 2.12 (PEP 571) ("manylinux1", (2, 5)), # CentOS 5 w/ glibc 2.5 (PEP 513) ) manylinux_support_iter = iter(manylinux_support) for name, glibc_version in manylinux_support_iter: if _is_manylinux_compatible(name, glibc_version): platforms = [linux.replace("linux", name)] break else: platforms = [] # Support for a later manylinux implies support for an earlier version. platforms += [linux.replace("linux", name) for name, _ in manylinux_support_iter] platforms.append(linux) return platforms def _generic_platforms(): platform = _normalize_string(distutils.util.get_platform()) return [platform] def _interpreter_name(): name = platform.python_implementation().lower() return INTERPRETER_SHORT_NAMES.get(name) or name def _generic_interpreter(name, py_version): version = sysconfig.get_config_var("py_version_nodot") if not version: version = "".join(map(str, py_version[:2])) return "{name}{version}".format(name=name, version=version) def sys_tags(): """ Returns the sequence of tag triples for the running interpreter. The order of the sequence corresponds to priority order for the interpreter, from most to least important. """ py_version = sys.version_info[:2] interpreter_name = _interpreter_name() if platform.system() == "Darwin": platforms = _mac_platforms() elif platform.system() == "Linux": platforms = _linux_platforms() else: platforms = _generic_platforms() if interpreter_name == "cp": interpreter = _cpython_interpreter(py_version) abis = _cpython_abis(py_version) for tag in _cpython_tags(py_version, interpreter, abis, platforms): yield tag elif interpreter_name == "pp": interpreter = _pypy_interpreter() abi = _generic_abi() for tag in _pypy_tags(py_version, interpreter, abi, platforms): yield tag else: interpreter = _generic_interpreter(interpreter_name, py_version) abi = _generic_abi() for tag in _generic_tags(interpreter, py_version, abi, platforms): yield tag for tag in _independent_tags(interpreter, py_version, platforms): yield tag site-packages/pip/_vendor/packaging/version.py000064400000027312151464452710015511 0ustar00# This file is dual licensed under the terms of the Apache License, Version # 2.0, and the BSD License. See the LICENSE file in the root of this repository # for complete details. from __future__ import absolute_import, division, print_function import collections import itertools import re from ._structures import Infinity __all__ = ["parse", "Version", "LegacyVersion", "InvalidVersion", "VERSION_PATTERN"] _Version = collections.namedtuple( "_Version", ["epoch", "release", "dev", "pre", "post", "local"] ) def parse(version): """ Parse the given version string and return either a :class:`Version` object or a :class:`LegacyVersion` object depending on if the given version is a valid PEP 440 version or a legacy version. """ try: return Version(version) except InvalidVersion: return LegacyVersion(version) class InvalidVersion(ValueError): """ An invalid version was found, users should refer to PEP 440. """ class _BaseVersion(object): def __hash__(self): return hash(self._key) def __lt__(self, other): return self._compare(other, lambda s, o: s < o) def __le__(self, other): return self._compare(other, lambda s, o: s <= o) def __eq__(self, other): return self._compare(other, lambda s, o: s == o) def __ge__(self, other): return self._compare(other, lambda s, o: s >= o) def __gt__(self, other): return self._compare(other, lambda s, o: s > o) def __ne__(self, other): return self._compare(other, lambda s, o: s != o) def _compare(self, other, method): if not isinstance(other, _BaseVersion): return NotImplemented return method(self._key, other._key) class LegacyVersion(_BaseVersion): def __init__(self, version): self._version = str(version) self._key = _legacy_cmpkey(self._version) def __str__(self): return self._version def __repr__(self): return "".format(repr(str(self))) @property def public(self): return self._version @property def base_version(self): return self._version @property def epoch(self): return -1 @property def release(self): return None @property def pre(self): return None @property def post(self): return None @property def dev(self): return None @property def local(self): return None @property def is_prerelease(self): return False @property def is_postrelease(self): return False @property def is_devrelease(self): return False _legacy_version_component_re = re.compile(r"(\d+ | [a-z]+ | \.| -)", re.VERBOSE) _legacy_version_replacement_map = { "pre": "c", "preview": "c", "-": "final-", "rc": "c", "dev": "@", } def _parse_version_parts(s): for part in _legacy_version_component_re.split(s): part = _legacy_version_replacement_map.get(part, part) if not part or part == ".": continue if part[:1] in "0123456789": # pad for numeric comparison yield part.zfill(8) else: yield "*" + part # ensure that alpha/beta/candidate are before final yield "*final" def _legacy_cmpkey(version): # We hardcode an epoch of -1 here. A PEP 440 version can only have a epoch # greater than or equal to 0. This will effectively put the LegacyVersion, # which uses the defacto standard originally implemented by setuptools, # as before all PEP 440 versions. epoch = -1 # This scheme is taken from pkg_resources.parse_version setuptools prior to # it's adoption of the packaging library. parts = [] for part in _parse_version_parts(version.lower()): if part.startswith("*"): # remove "-" before a prerelease tag if part < "*final": while parts and parts[-1] == "*final-": parts.pop() # remove trailing zeros from each series of numeric parts while parts and parts[-1] == "00000000": parts.pop() parts.append(part) parts = tuple(parts) return epoch, parts # Deliberately not anchored to the start and end of the string, to make it # easier for 3rd party code to reuse VERSION_PATTERN = r""" v? (?: (?:(?P[0-9]+)!)? # epoch (?P[0-9]+(?:\.[0-9]+)*) # release segment (?P
                                          # pre-release
            [-_\.]?
            (?P(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P[0-9]+)?
        )?
        (?P                                         # post release
            (?:-(?P[0-9]+))
            |
            (?:
                [-_\.]?
                (?Ppost|rev|r)
                [-_\.]?
                (?P[0-9]+)?
            )
        )?
        (?P                                          # dev release
            [-_\.]?
            (?Pdev)
            [-_\.]?
            (?P[0-9]+)?
        )?
    )
    (?:\+(?P[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
"""


class Version(_BaseVersion):

    _regex = re.compile(r"^\s*" + VERSION_PATTERN + r"\s*$", re.VERBOSE | re.IGNORECASE)

    def __init__(self, version):
        # Validate the version and parse it into pieces
        match = self._regex.search(version)
        if not match:
            raise InvalidVersion("Invalid version: '{0}'".format(version))

        # Store the parsed out pieces of the version
        self._version = _Version(
            epoch=int(match.group("epoch")) if match.group("epoch") else 0,
            release=tuple(int(i) for i in match.group("release").split(".")),
            pre=_parse_letter_version(match.group("pre_l"), match.group("pre_n")),
            post=_parse_letter_version(
                match.group("post_l"), match.group("post_n1") or match.group("post_n2")
            ),
            dev=_parse_letter_version(match.group("dev_l"), match.group("dev_n")),
            local=_parse_local_version(match.group("local")),
        )

        # Generate a key which will be used for sorting
        self._key = _cmpkey(
            self._version.epoch,
            self._version.release,
            self._version.pre,
            self._version.post,
            self._version.dev,
            self._version.local,
        )

    def __repr__(self):
        return "".format(repr(str(self)))

    def __str__(self):
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append("{0}!".format(self.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        # Pre-release
        if self.pre is not None:
            parts.append("".join(str(x) for x in self.pre))

        # Post-release
        if self.post is not None:
            parts.append(".post{0}".format(self.post))

        # Development release
        if self.dev is not None:
            parts.append(".dev{0}".format(self.dev))

        # Local version segment
        if self.local is not None:
            parts.append("+{0}".format(self.local))

        return "".join(parts)

    @property
    def epoch(self):
        return self._version.epoch

    @property
    def release(self):
        return self._version.release

    @property
    def pre(self):
        return self._version.pre

    @property
    def post(self):
        return self._version.post[1] if self._version.post else None

    @property
    def dev(self):
        return self._version.dev[1] if self._version.dev else None

    @property
    def local(self):
        if self._version.local:
            return ".".join(str(x) for x in self._version.local)
        else:
            return None

    @property
    def public(self):
        return str(self).split("+", 1)[0]

    @property
    def base_version(self):
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append("{0}!".format(self.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        return "".join(parts)

    @property
    def is_prerelease(self):
        return self.dev is not None or self.pre is not None

    @property
    def is_postrelease(self):
        return self.post is not None

    @property
    def is_devrelease(self):
        return self.dev is not None


def _parse_letter_version(letter, number):
    if letter:
        # We consider there to be an implicit 0 in a pre-release if there is
        # not a numeral associated with it.
        if number is None:
            number = 0

        # We normalize any letters to their lower case form
        letter = letter.lower()

        # We consider some words to be alternate spellings of other words and
        # in those cases we want to normalize the spellings to our preferred
        # spelling.
        if letter == "alpha":
            letter = "a"
        elif letter == "beta":
            letter = "b"
        elif letter in ["c", "pre", "preview"]:
            letter = "rc"
        elif letter in ["rev", "r"]:
            letter = "post"

        return letter, int(number)
    if not letter and number:
        # We assume if we are given a number, but we are not given a letter
        # then this is using the implicit post release syntax (e.g. 1.0-1)
        letter = "post"

        return letter, int(number)


_local_version_separators = re.compile(r"[\._-]")


def _parse_local_version(local):
    """
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    """
    if local is not None:
        return tuple(
            part.lower() if not part.isdigit() else int(part)
            for part in _local_version_separators.split(local)
        )


def _cmpkey(epoch, release, pre, post, dev, local):
    # When we compare a release version, we want to compare it with all of the
    # trailing zeros removed. So we'll use a reverse the list, drop all the now
    # leading zeros until we come to something non zero, then take the rest
    # re-reverse it back into the correct order and make it a tuple and use
    # that for our sorting key.
    release = tuple(
        reversed(list(itertools.dropwhile(lambda x: x == 0, reversed(release))))
    )

    # We need to "trick" the sorting algorithm to put 1.0.dev0 before 1.0a0.
    # We'll do this by abusing the pre segment, but we _only_ want to do this
    # if there is not a pre or a post segment. If we have one of those then
    # the normal sorting rules will handle this case correctly.
    if pre is None and post is None and dev is not None:
        pre = -Infinity
    # Versions without a pre-release (except as noted above) should sort after
    # those with one.
    elif pre is None:
        pre = Infinity

    # Versions without a post segment should sort before those with one.
    if post is None:
        post = -Infinity

    # Versions without a development segment should sort after those with one.
    if dev is None:
        dev = Infinity

    if local is None:
        # Versions without a local segment should sort before those with one.
        local = -Infinity
    else:
        # Versions with a local segment need that segment parsed to implement
        # the sorting rules in PEP440.
        # - Alpha numeric segments sort before numeric segments
        # - Alpha numeric segments sort lexicographically
        # - Numeric segments sort numerically
        # - Shorter versions sort before longer versions when the prefixes
        #   match exactly
        local = tuple((i, "") if isinstance(i, int) else (-Infinity, i) for i in local)

    return epoch, release, pre, post, dev, local
site-packages/pip/_vendor/packaging/__pycache__/version.cpython-38.opt-1.pyc000064400000027432151464452710022741 0ustar00U

.e.	@sddlmZmZmZddlZddlZddlZddlmZddddd	gZ	e
d
ddd
dddgZddZGddde
ZGdddeZGdddeZedejZddddddZddZddZdZGd ddeZd!d"Zed#Zd$d%Zd&d'ZdS)()absolute_importdivisionprint_functionN)InfinityparseVersion
LegacyVersionInvalidVersionVERSION_PATTERN_VersionepochreleasedevprepostlocalcCs,z
t|WStk
r&t|YSXdS)z
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rr
r	)versionrA/usr/lib/python3.8/site-packages/pip/_vendor/packaging/version.pyrs
c@seZdZdZdS)r
zF
    An invalid version was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrrrr
!sc@sLeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
dS)_BaseVersioncCs
t|jSN)hash_keyselfrrr__hash__(sz_BaseVersion.__hash__cCs||ddS)NcSs||kSrrsorrr,z%_BaseVersion.__lt__.._comparerotherrrr__lt__+sz_BaseVersion.__lt__cCs||ddS)NcSs||kSrrr!rrrr$/r%z%_BaseVersion.__le__..r&r(rrr__le__.sz_BaseVersion.__le__cCs||ddS)NcSs||kSrrr!rrrr$2r%z%_BaseVersion.__eq__..r&r(rrr__eq__1sz_BaseVersion.__eq__cCs||ddS)NcSs||kSrrr!rrrr$5r%z%_BaseVersion.__ge__..r&r(rrr__ge__4sz_BaseVersion.__ge__cCs||ddS)NcSs||kSrrr!rrrr$8r%z%_BaseVersion.__gt__..r&r(rrr__gt__7sz_BaseVersion.__gt__cCs||ddS)NcSs||kSrrr!rrrr$;r%z%_BaseVersion.__ne__..r&r(rrr__ne__:sz_BaseVersion.__ne__cCst|tstS||j|jSr)
isinstancerNotImplementedr)rr)methodrrrr'=s
z_BaseVersion._compareN)rrrr r*r+r,r-r.r/r'rrrrr'src@seZdZddZddZddZeddZed	d
ZeddZ	ed
dZ
eddZeddZeddZ
eddZeddZeddZeddZdS)r	cCst||_t|j|_dSr)str_version_legacy_cmpkeyr)rrrrr__init__Es
zLegacyVersion.__init__cCs|jSrr4rrrr__str__IszLegacyVersion.__str__cCsdtt|S)Nzformatreprr3rrrr__repr__LszLegacyVersion.__repr__cCs|jSrr7rrrrpublicOszLegacyVersion.publiccCs|jSrr7rrrrbase_versionSszLegacyVersion.base_versioncCsdS)Nrrrrrr
WszLegacyVersion.epochcCsdSrrrrrrr[szLegacyVersion.releasecCsdSrrrrrrr_szLegacyVersion.precCsdSrrrrrrrcszLegacyVersion.postcCsdSrrrrrrrgszLegacyVersion.devcCsdSrrrrrrrkszLegacyVersion.localcCsdSNFrrrrr
is_prereleaseoszLegacyVersion.is_prereleasecCsdSr@rrrrris_postreleasesszLegacyVersion.is_postreleasecCsdSr@rrrrr
is_devreleasewszLegacyVersion.is_devreleaseN)rrrr6r8r<propertyr=r>r
rrrrrrArBrCrrrrr	Ds2









z(\d+ | [a-z]+ | \.| -)czfinal-@)rpreview-rcrccs\t|D]F}t||}|r
|dkr(q
|dddkrF|dVq
d|Vq
dVdS)N.r
0123456789**final)_legacy_version_component_resplit_legacy_version_replacement_mapgetzfill)r"partrrr_parse_version_partssrUcCszd}g}t|D]T}|dr^|dkrD|rD|ddkrD|q*|r^|ddkr^|qD||qt|}||fS)Nr?rMrNz*final-Z00000000)rUlower
startswithpopappendtuple)rr
partsrTrrrr5s


r5a
    v?
    (?:
        (?:(?P[0-9]+)!)?                           # epoch
        (?P[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P
                                          # pre-release
            [-_\.]?
            (?P(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P[0-9]+)?
        )?
        (?P                                         # post release
            (?:-(?P[0-9]+))
            |
            (?:
                [-_\.]?
                (?Ppost|rev|r)
                [-_\.]?
                (?P[0-9]+)?
            )
        )?
        (?P                                          # dev release
            [-_\.]?
            (?Pdev)
            [-_\.]?
            (?P[0-9]+)?
        )?
    )
    (?:\+(?P[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@seZdZededejejBZddZ	ddZ
ddZed	d
Z
eddZed
dZeddZeddZeddZeddZeddZeddZeddZeddZdS) rz^\s*z\s*$c
Cs|j|}|std|t|dr8t|dndtdd|ddDt	|d|d	t	|d
|dp|dt	|d
|dt
|dd|_t|jj
|jj|jj|jj|jj|jj|_dS)NzInvalid version: '{0}'r
rcss|]}t|VqdSr)int.0irrr	sz#Version.__init__..rrJZpre_lZpre_nZpost_lZpost_n1Zpost_n2Zdev_lZdev_nrr
rrrrr)_regexsearchr
r:rgroupr\rZrP_parse_letter_version_parse_local_versionr4_cmpkeyr
rrrrrr)rrmatchrrrr6s,zVersion.__init__cCsdtt|S)Nzr9rrrrr<szVersion.__repr__cCsg}|jdkr |d|j|ddd|jD|jdk	rb|ddd|jD|jdk	r~|d|j|jdk	r|d	|j|jdk	r|d
|jd|S)Nr{0}!rJcss|]}t|VqdSrr3r^xrrrr`sz"Version.__str__..css|]}t|VqdSrrjrkrrrr`sz.post{0}z.dev{0}z+{0})	r
rYr:joinrrrrrrr[rrrr8s




zVersion.__str__cCs|jjSr)r4r
rrrrr
sz
Version.epochcCs|jjSr)r4rrrrrrszVersion.releasecCs|jjSr)r4rrrrrrszVersion.precCs|jjr|jjdSdSNr)r4rrrrrr szVersion.postcCs|jjr|jjdSdSrp)r4rrrrrr$szVersion.devcCs(|jjr ddd|jjDSdSdS)NrJcss|]}t|VqdSrrjrkrrrr`+sz Version.local..)r4rrnrrrrr(sz
Version.localcCst|dddS)N+rr)r3rPrrrrr=/szVersion.publiccCsFg}|jdkr |d|j|ddd|jDd|S)NrrirJcss|]}t|VqdSrrjrkrrrr`<sz'Version.base_version..rm)r
rYr:rnrrorrrr>3s

zVersion.base_versioncCs|jdk	p|jdk	Sr)rrrrrrrA@szVersion.is_prereleasecCs
|jdk	Sr)rrrrrrBDszVersion.is_postreleasecCs
|jdk	Sr)rrrrrrCHszVersion.is_devreleaseN)rrrrecompilerVERBOSE
IGNORECASErbr6r<r8rDr
rrrrrr=r>rArBrCrrrrrs4









cCsv|rZ|dkrd}|}|dkr&d}n(|dkr4d}n|dkrBd}n|dkrNd	}|t|fS|sr|rrd	}|t|fSdS)
NrZalphaaZbetab)rErrGrI)Zrevrr)rVr\)ZletterZnumberrrrreMs rez[\._-]cCs$|dk	r tddt|DSdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|s|nt|VqdSr)isdigitrVr\)r^rTrrrr`tsz'_parse_local_version..)rZ_local_version_separatorsrP)rrrrrfosrfcCsttttddt|}|dkr@|dkr@|dk	r@t}n|dkrLt}|dkrZt}|dkrft}|dkrvt}ntdd|D}||||||fS)NcSs|dkS)Nrr)rlrrrr$r%z_cmpkey..css*|]"}t|tr|dfnt|fVqdS)rmN)r0r\rr]rrrr`sz_cmpkey..)rZreversedlist	itertools	dropwhilerrarrrrgzs	rg)Z
__future__rrrcollectionsr}rrZ_structuresr__all__
namedtuplerr
ValueErrorr
objectrr	rsrtrOrQrUr5rrrerzrfrgrrrrs88	 x
site-packages/pip/_vendor/packaging/__pycache__/version.cpython-38.pyc000064400000027432151464452710022002 0ustar00U

.e.	@sddlmZmZmZddlZddlZddlZddlmZddddd	gZ	e
d
ddd
dddgZddZGddde
ZGdddeZGdddeZedejZddddddZddZddZdZGd ddeZd!d"Zed#Zd$d%Zd&d'ZdS)()absolute_importdivisionprint_functionN)InfinityparseVersion
LegacyVersionInvalidVersionVERSION_PATTERN_VersionepochreleasedevprepostlocalcCs,z
t|WStk
r&t|YSXdS)z
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rr
r	)versionrA/usr/lib/python3.8/site-packages/pip/_vendor/packaging/version.pyrs
c@seZdZdZdS)r
zF
    An invalid version was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrrrr
!sc@sLeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
dS)_BaseVersioncCs
t|jSN)hash_keyselfrrr__hash__(sz_BaseVersion.__hash__cCs||ddS)NcSs||kSrrsorrr,z%_BaseVersion.__lt__.._comparerotherrrr__lt__+sz_BaseVersion.__lt__cCs||ddS)NcSs||kSrrr!rrrr$/r%z%_BaseVersion.__le__..r&r(rrr__le__.sz_BaseVersion.__le__cCs||ddS)NcSs||kSrrr!rrrr$2r%z%_BaseVersion.__eq__..r&r(rrr__eq__1sz_BaseVersion.__eq__cCs||ddS)NcSs||kSrrr!rrrr$5r%z%_BaseVersion.__ge__..r&r(rrr__ge__4sz_BaseVersion.__ge__cCs||ddS)NcSs||kSrrr!rrrr$8r%z%_BaseVersion.__gt__..r&r(rrr__gt__7sz_BaseVersion.__gt__cCs||ddS)NcSs||kSrrr!rrrr$;r%z%_BaseVersion.__ne__..r&r(rrr__ne__:sz_BaseVersion.__ne__cCst|tstS||j|jSr)
isinstancerNotImplementedr)rr)methodrrrr'=s
z_BaseVersion._compareN)rrrr r*r+r,r-r.r/r'rrrrr'src@seZdZddZddZddZeddZed	d
ZeddZ	ed
dZ
eddZeddZeddZ
eddZeddZeddZeddZdS)r	cCst||_t|j|_dSr)str_version_legacy_cmpkeyr)rrrrr__init__Es
zLegacyVersion.__init__cCs|jSrr4rrrr__str__IszLegacyVersion.__str__cCsdtt|S)Nzformatreprr3rrrr__repr__LszLegacyVersion.__repr__cCs|jSrr7rrrrpublicOszLegacyVersion.publiccCs|jSrr7rrrrbase_versionSszLegacyVersion.base_versioncCsdS)Nrrrrrr
WszLegacyVersion.epochcCsdSrrrrrrr[szLegacyVersion.releasecCsdSrrrrrrr_szLegacyVersion.precCsdSrrrrrrrcszLegacyVersion.postcCsdSrrrrrrrgszLegacyVersion.devcCsdSrrrrrrrkszLegacyVersion.localcCsdSNFrrrrr
is_prereleaseoszLegacyVersion.is_prereleasecCsdSr@rrrrris_postreleasesszLegacyVersion.is_postreleasecCsdSr@rrrrr
is_devreleasewszLegacyVersion.is_devreleaseN)rrrr6r8r<propertyr=r>r
rrrrrrArBrCrrrrr	Ds2









z(\d+ | [a-z]+ | \.| -)czfinal-@)rpreview-rcrccs\t|D]F}t||}|r
|dkr(q
|dddkrF|dVq
d|Vq
dVdS)N.r
0123456789**final)_legacy_version_component_resplit_legacy_version_replacement_mapgetzfill)r"partrrr_parse_version_partssrUcCszd}g}t|D]T}|dr^|dkrD|rD|ddkrD|q*|r^|ddkr^|qD||qt|}||fS)Nr?rMrNz*final-Z00000000)rUlower
startswithpopappendtuple)rr
partsrTrrrr5s


r5a
    v?
    (?:
        (?:(?P[0-9]+)!)?                           # epoch
        (?P[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P
                                          # pre-release
            [-_\.]?
            (?P(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P[0-9]+)?
        )?
        (?P                                         # post release
            (?:-(?P[0-9]+))
            |
            (?:
                [-_\.]?
                (?Ppost|rev|r)
                [-_\.]?
                (?P[0-9]+)?
            )
        )?
        (?P                                          # dev release
            [-_\.]?
            (?Pdev)
            [-_\.]?
            (?P[0-9]+)?
        )?
    )
    (?:\+(?P[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@seZdZededejejBZddZ	ddZ
ddZed	d
Z
eddZed
dZeddZeddZeddZeddZeddZeddZeddZeddZdS) rz^\s*z\s*$c
Cs|j|}|std|t|dr8t|dndtdd|ddDt	|d|d	t	|d
|dp|dt	|d
|dt
|dd|_t|jj
|jj|jj|jj|jj|jj|_dS)NzInvalid version: '{0}'r
rcss|]}t|VqdSr)int.0irrr	sz#Version.__init__..rrJZpre_lZpre_nZpost_lZpost_n1Zpost_n2Zdev_lZdev_nrr
rrrrr)_regexsearchr
r:rgroupr\rZrP_parse_letter_version_parse_local_versionr4_cmpkeyr
rrrrrr)rrmatchrrrr6s,zVersion.__init__cCsdtt|S)Nzr9rrrrr<szVersion.__repr__cCsg}|jdkr |d|j|ddd|jD|jdk	rb|ddd|jD|jdk	r~|d|j|jdk	r|d	|j|jdk	r|d
|jd|S)Nr{0}!rJcss|]}t|VqdSrr3r^xrrrr`sz"Version.__str__..css|]}t|VqdSrrjrkrrrr`sz.post{0}z.dev{0}z+{0})	r
rYr:joinrrrrrrr[rrrr8s




zVersion.__str__cCs|jjSr)r4r
rrrrr
sz
Version.epochcCs|jjSr)r4rrrrrrszVersion.releasecCs|jjSr)r4rrrrrrszVersion.precCs|jjr|jjdSdSNr)r4rrrrrr szVersion.postcCs|jjr|jjdSdSrp)r4rrrrrr$szVersion.devcCs(|jjr ddd|jjDSdSdS)NrJcss|]}t|VqdSrrjrkrrrr`+sz Version.local..)r4rrnrrrrr(sz
Version.localcCst|dddS)N+rr)r3rPrrrrr=/szVersion.publiccCsFg}|jdkr |d|j|ddd|jDd|S)NrrirJcss|]}t|VqdSrrjrkrrrr`<sz'Version.base_version..rm)r
rYr:rnrrorrrr>3s

zVersion.base_versioncCs|jdk	p|jdk	Sr)rrrrrrrA@szVersion.is_prereleasecCs
|jdk	Sr)rrrrrrBDszVersion.is_postreleasecCs
|jdk	Sr)rrrrrrCHszVersion.is_devreleaseN)rrrrecompilerVERBOSE
IGNORECASErbr6r<r8rDr
rrrrrr=r>rArBrCrrrrrs4









cCsv|rZ|dkrd}|}|dkr&d}n(|dkr4d}n|dkrBd}n|dkrNd	}|t|fS|sr|rrd	}|t|fSdS)
NrZalphaaZbetab)rErrGrI)Zrevrr)rVr\)ZletterZnumberrrrreMs rez[\._-]cCs$|dk	r tddt|DSdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|s|nt|VqdSr)isdigitrVr\)r^rTrrrr`tsz'_parse_local_version..)rZ_local_version_separatorsrP)rrrrrfosrfcCsttttddt|}|dkr@|dkr@|dk	r@t}n|dkrLt}|dkrZt}|dkrft}|dkrvt}ntdd|D}||||||fS)NcSs|dkS)Nrr)rlrrrr$r%z_cmpkey..css*|]"}t|tr|dfnt|fVqdS)rmN)r0r\rr]rrrr`sz_cmpkey..)rZreversedlist	itertools	dropwhilerrarrrrgzs	rg)Z
__future__rrrcollectionsr}rrZ_structuresr__all__
namedtuplerr
ValueErrorr
objectrr	rsrtrOrQrUr5rrrerzrfrgrrrrs88	 x
site-packages/pip/_vendor/packaging/__pycache__/requirements.cpython-38.opt-1.pyc000064400000007610151464452710023773 0ustar00U

.e\@s|ddlmZmZmZddlZddlZddlmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZddlmZddlmZddlmZmZdd	lmZmZmZGd
ddeZeejejZ ed!Z"ed
!Z#ed!Z$ed!Z%ed!Z&ed!Z'ed!Z(edZ)e ee)e BZ*ee ee*Z+e+dZ,e+Z-eddZ.e(e.Z/e-ee&e-Z0e"e
e0e#dZ1eej2ej3ej4BZ5eej2ej3ej4BZ6e5e6AZ7ee7ee&e7ddddZ8e
e$e8e%e8BZ9e9:dde	e9dZ;e;:dde	edZe:d de'Ze/e
e=Z?e,e
e1e?e>BZ@ee@eZAeABd!Gd"d#d#eCZDdS)$)absolute_importdivisionprint_functionN)stringStart	stringEndoriginalTextForParseException)
ZeroOrMoreWordOptionalRegexCombine)Literal)parse)MARKER_EXPRMarker)LegacySpecifier	SpecifierSpecifierSetc@seZdZdZdS)InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)__name__
__module____qualname____doc__rrF/usr/lib/python3.8/site-packages/pip/_vendor/packaging/requirements.pyrsr[](),;@z-_.namez[^ ]+urlextrasF)Z
joinStringZadjacent	_raw_speccCs
|jpdS)N)r'sltrrr7r-	specifiercCs|dS)Nrrr)rrrr-:r.markercCst||j|jS)N)rZ_original_startZ
_original_endr)rrrr->r.zx[]c@s(eZdZdZddZddZddZdS)	RequirementzParse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    c
Cszt|}WnDtk
rR}z&td||j|jd|jW5d}~XYnX|j|_|jrt		|j}|j
dkrt	||jkrtdn(|j
r|jr|j
s|jstd|j|j|_nd|_t
|jr|jng|_t|j|_|jr|jnd|_dS)NzParse error at "{0!r}": {1}filezInvalid URL givenzInvalid URL: {0})REQUIREMENTparseStringrrformatZlocmsgr$r%urlparseZschemeZ
urlunparseZnetlocsetr&ZasListrr/r0)selfZrequirement_stringZreqeZ
parsed_urlrrr__init__[s4


zRequirement.__init__cCs|jg}|jr*|ddt|j|jr@|t|j|jrh|d|j|j	rh|d|j	r|d|j	d|S)Nz[{0}]r!z@ {0} z; {0}r()
r$r&appendr6joinsortedr/strr%r0)r:partsrrr__str__vs
zRequirement.__str__cCsdt|S)Nz)r6rA)r:rrr__repr__szRequirement.__repr__N)rrrrr<rCrDrrrrr1Nsr1)EZ
__future__rrrstringreZpip._vendor.pyparsingrrrrr	r
rrr
rLZpip._vendor.six.moves.urllibrr8ZmarkersrrZ
specifiersrrr
ValueErrorrZ
ascii_lettersZdigitsZALPHANUMsuppressZLBRACKETZRBRACKETZLPARENZRPARENCOMMAZ	SEMICOLONATZPUNCTUATIONZIDENTIFIER_ENDZ
IDENTIFIERNAMEZEXTRAZURIZURLZEXTRAS_LISTZEXTRASZ
_regex_strVERBOSE
IGNORECASEZVERSION_PEP440ZVERSION_LEGACYZVERSION_ONEZVERSION_MANYZ
_VERSION_SPECZsetParseActionZVERSION_SPECZMARKER_SEPARATORZMARKERZVERSION_AND_MARKERZURL_AND_MARKERZNAMED_REQUIREMENTr4r5objectr1rrrrsh
site-packages/pip/_vendor/packaging/__pycache__/__init__.cpython-38.pyc000064400000001027151464452710022044 0ustar00U

.e2@sTddlmZmZmZddlmZmZmZmZm	Z	m
Z
mZmZdddddd	d
dgZ
dS)
)absolute_importdivisionprint_function)
__author__
__copyright__	__email____license____summary__	__title____uri____version__rr
rr
rrr	rN)Z
__future__rrr	__about__rrrr	r
rrr
__all__rrB/usr/lib/python3.8/site-packages/pip/_vendor/packaging/__init__.pys(site-packages/pip/_vendor/packaging/__pycache__/_compat.cpython-38.opt-1.pyc000064400000001713151464452710022670 0ustar00U

.ea@sVddlmZmZmZddlZejddkZejddkZerDefZ	ne
fZ	ddZdS))absolute_importdivisionprint_functionNcs&Gfddd}t|ddiS)z/
    Create a base class with a metaclass.
    cseZdZfddZdS)z!with_metaclass..metaclasscs||S)N)clsnameZ
this_basesdbasesmetarA/usr/lib/python3.8/site-packages/pip/_vendor/packaging/_compat.py__new__sz)with_metaclass..metaclass.__new__N)__name__
__module____qualname__rrrrr	metaclasssrZtemporary_classr)typer)r
rrrrrwith_metaclasssr)Z
__future__rrrsysversion_infoZPY2ZPY3strZstring_typesZ
basestringrrrrrssite-packages/pip/_vendor/packaging/__pycache__/tags.cpython-38.opt-1.pyc000064400000025064151464452710022211 0ustar00U

.e2@snddlmZddlZzddlmZWn0ek
rTddlZddeDZ[YnXddl	Z	ddl
Z
ddlZddlZddl
Z
dddd	d
dZejdkZGd
ddeZddZddZddZddZddZddZddZddZdd Zd!d"Zd#d$Zefd%d&Zd'd(Zd=d)d*Z d+d,Z!d-d.Z"d/d0Z#d1d2Z$efd3d4Z%d5d6Z&d7d8Z'd9d:Z(d;d<Z)dS)>)absolute_importN)EXTENSION_SUFFIXEScCsg|]}|dqS)r).0xrr>/usr/lib/python3.8/site-packages/pip/_vendor/packaging/tags.py
srpycpppZipZjy)pythonZcpythonZpypyZ
ironpythonZjythonlc@sbeZdZdddgZddZeddZedd	Zed
dZdd
Z	ddZ
ddZddZdS)Tag_interpreter_abi	_platformcCs"||_||_||_dSN)lowerrrr)selfinterpreterabiplatformrrr__init__'s

zTag.__init__cCs|jSr)rrrrrr,szTag.interpretercCs|jSr)rrrrrr0szTag.abicCs|jSr)rrrrrr4szTag.platformcCs$|j|jko"|j|jko"|j|jkSr)rrr)rotherrrr__eq__8s


z
Tag.__eq__cCst|j|j|jfSr)hashrrrrrrr__hash__?szTag.__hash__cCsd|j|j|jS)Nz{}-{}-{})formatrrrrrrr__str__BszTag.__str__cCsdj|t|dS)Nz<{self} @ {self_id}>)rZself_id)ridrrrr__repr__EszTag.__repr__N)
__name__
__module____qualname__	__slots__rpropertyrrrrrrr rrrrr
#s



r
c	Cs`t}|d\}}}|dD]6}|dD]&}|dD]}|t|||qsz _cpython_tags..c3s|]}td|VqdSnoneNrSrTrUrrrVsr8r7rr9rR)r
ranger)r<rr.r/rr0r,
minor_versionrrUr
_cpython_tagssr\cCsdjtjdtjjtjjdS)Nz$pp{py_major}{pypy_major}{pypy_minor}r)Zpy_majorZ
pypy_majorZ
pypy_minor)rrMversion_infoZpypy_version_infor:r;rrrr_pypy_interpreters
r^cCstd}|rt|SdSdS)NZSOABIrX)rJrKr6)rrrr_generic_abis
r_c#sBfdd|DD]
}|Vqfdd|DD]
}|Vq2dS)Nc3s|]}t|VqdSrrSrrrrrrrVsz_pypy_tags..c3s|]}td|VqdSrWrSr`rUrrrVsr)r<rrr/r,rrar
_pypy_tagssrbc#sNfdd|DD]
}|VqdkrJfdd|D}|D]
}|Vq>dS)Nc3s|]}t|VqdSrrSr`rarrrVsz _generic_tags..rXc3s|]}td|VqdSrWrSrTrUrrrVsr)rr<rr/r,r-rrar
_generic_tagssrcccs\dj|d|ddVdj|ddVt|ddddD]}dj|d|dVq>dS)	z
    Yield Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all following versions up to 'end'.
    zpy{major}{minor}rr8r9z	py{major})r:rYN)rrZ)r<r;rrr_py_interpreter_rangesrdccsTt|D]}|D]}t|d|Vqqt|ddVt|D]}t|ddVq
    - -none-any
    - py*-none-any
    rXanyN)rdr
)rr<r/rFr0rrr_independent_tagss	rfcCs|s|S|drdSdS)Nppci386)
startswith)archis_32bitrrr	_mac_archs

rlcCs|g}|dkr,|dkrgS|dddgnp|dkrR|dkr@gS|dddgnJ|dkrz|d	ksj|dkrngS|dn"|d
kr|dkrgS|ddg|d|S)
NZx86_64)
rDZintelZfat64Zfat32rhZfatZppc64)rmrg)rmZ	universal)extendrO)rFcpu_archZformatsrrr_mac_binary_formatss&
rrc
	Cst\}}}|dkr2ttt|ddd}|dkrBt|}g}t|dddD]B}|d|f}t||}|D]"}	|	dj
|d|d|	dqtqV|S)Nr'r>r8rYrz&macosx_{major}_{minor}_{binary_format})r:r;
binary_format)rZmac_vertuplemapintr)rlrZrrrOr)
rFrjversion_strr2rqr/r[Zcompat_versionZbinary_formatsrsrrr_mac_platformss$
rxc	Cs>zddl}tt||dWSttfk
r4YnXt|S)NrZ_compatible)
_manylinuxboolgetattrImportErrorAttributeError_have_compatible_glibc)name
glibc_versionryrrr_is_manylinux_compatible	srcCsZddl}|d}z
|j}Wntk
r2YdSX|j|_|}t|tsV|d}|S)Nrascii)	ctypesZCDLLgnu_get_libc_versionr}Zc_char_pZrestype
isinstancestrdecode)rZprocess_namespacerrwrrr_glibc_version_strings



rcCsHtd|}|s$td|tdSt|d|koFt|d|kS)Nz$(?P[0-9]+)\.(?P[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFr:r;)rematchwarningswarnRuntimeWarningrvgroup)rwrequired_major
minimum_minorrCrrr_check_glibc_version1srcCst}|dkrdSt|||S)NF)rr)rrrwrrrr~Fsr~cs|ttjdkr|rdd}t|}|D]$\}}t||r.d|g}qXq.g}|fdd|D7}||S)NZlinux_x86_64Z
linux_i686))Z
manylinux2014)r>)Z
manylinux2010)r>)Z
manylinux1)r>rnlinuxcsg|]\}}d|qSrr3)rrr2rrrr^sz$_linux_platforms..)r6	distutilsutilget_platformiterrr4rO)rkZmanylinux_supportZmanylinux_support_iterrrr/rrr_linux_platformsMs

rcCsttj}|gSr)r6rrr)rrrr_generic_platformscsrcCst}t|p|Sr)rZpython_implementationrINTERPRETER_SHORT_NAMESget)rrrr_interpreter_namehsrcCs4td}|s&dtt|dd}dj||dS)NZpy_version_nodotr?r>z{name}{version})rrF)rJrKjoinrurr)rr<rFrrr_generic_interpreterms
rccstjdd}t}tdkr(t}ntdkrZDarwinZLinuxr
r)rMr]rrsystemrxrrr=rQr\r^r_rbrrcrf)r<Zinterpreter_namer/rr.r,rrrrsys_tagsts.


r)NN)*Z
__future__rZdistutils.utilrZimportlib.machineryrr|ZimpZget_suffixesrrrMrJrrmaxsizeZ_32_BIT_INTERPRETERobjectr
r1r6r=rQr\r^r_rbrcrdrfrlrrrxrrrr~rrrrrrrrrsV	
&
#	



site-packages/pip/_vendor/packaging/__pycache__/__about__.cpython-38.pyc000064400000001271151464452710022214 0ustar00U

.e@sPddlmZmZmZdddddddd	gZd
ZdZdZd
ZdZ	dZ
dZde	ZdS))absolute_importdivisionprint_function	__title____summary____uri____version__
__author__	__email____license__
__copyright__Z	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz19.2z)Donald Stufft and individual contributorszdonald@stufft.ioz"BSD or Apache License, Version 2.0zCopyright 2014-2019 %sN)
Z
__future__rrr__all__rrrrr	r
rrrrC/usr/lib/python3.8/site-packages/pip/_vendor/packaging/__about__.pys"site-packages/pip/_vendor/packaging/__pycache__/_compat.cpython-38.pyc000064400000001713151464452710021731 0ustar00U

.ea@sVddlmZmZmZddlZejddkZejddkZerDefZ	ne
fZ	ddZdS))absolute_importdivisionprint_functionNcs&Gfddd}t|ddiS)z/
    Create a base class with a metaclass.
    cseZdZfddZdS)z!with_metaclass..metaclasscs||S)N)clsnameZ
this_basesdbasesmetarA/usr/lib/python3.8/site-packages/pip/_vendor/packaging/_compat.py__new__sz)with_metaclass..metaclass.__new__N)__name__
__module____qualname__rrrrr	metaclasssrZtemporary_classr)typer)r
rrrrrwith_metaclasssr)Z
__future__rrrsysversion_infoZPY2ZPY3strZstring_typesZ
basestringrrrrrssite-packages/pip/_vendor/packaging/__pycache__/__about__.cpython-38.opt-1.pyc000064400000001271151464452710023153 0ustar00U

.e@sPddlmZmZmZdddddddd	gZd
ZdZdZd
ZdZ	dZ
dZde	ZdS))absolute_importdivisionprint_function	__title____summary____uri____version__
__author__	__email____license__
__copyright__Z	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz19.2z)Donald Stufft and individual contributorszdonald@stufft.ioz"BSD or Apache License, Version 2.0zCopyright 2014-2019 %sN)
Z
__future__rrr__all__rrrrr	r
rrrrC/usr/lib/python3.8/site-packages/pip/_vendor/packaging/__about__.pys"site-packages/pip/_vendor/packaging/__pycache__/_structures.cpython-38.pyc000064400000005301151464452710022666 0ustar00U

.e@sDddlmZmZmZGdddeZeZGdddeZeZdS))absolute_importdivisionprint_functionc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)InfinitycCsdS)NrselfrrE/usr/lib/python3.8/site-packages/pip/_vendor/packaging/_structures.py__repr__szInfinity.__repr__cCstt|SNhashreprrrrr	__hash__szInfinity.__hash__cCsdSNFrrotherrrr	__lt__szInfinity.__lt__cCsdSrrrrrr	__le__szInfinity.__le__cCst||jSr
isinstance	__class__rrrr	__eq__szInfinity.__eq__cCst||jSrrrrrr	__ne__szInfinity.__ne__cCsdSNTrrrrr	__gt__szInfinity.__gt__cCsdSrrrrrr	__ge__szInfinity.__ge__cCstSr)NegativeInfinityrrrr	__neg__ szInfinity.__neg__N__name__
__module____qualname__r
rrrrrrrrrrrr	rsrc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)rcCsdS)Nz	-Infinityrrrrr	r
(szNegativeInfinity.__repr__cCstt|Srrrrrr	r+szNegativeInfinity.__hash__cCsdSrrrrrr	r.szNegativeInfinity.__lt__cCsdSrrrrrr	r1szNegativeInfinity.__le__cCst||jSrrrrrr	r4szNegativeInfinity.__eq__cCst||jSrrrrrr	r7szNegativeInfinity.__ne__cCsdSrrrrrr	r:szNegativeInfinity.__gt__cCsdSrrrrrr	r=szNegativeInfinity.__ge__cCstSr)rrrrr	r@szNegativeInfinity.__neg__Nrrrrr	r'srN)Z
__future__rrrobjectrrrrrr	ssite-packages/pip/_vendor/packaging/__pycache__/requirements.cpython-38.pyc000064400000007610151464452710023034 0ustar00U

.e\@s|ddlmZmZmZddlZddlZddlmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZddlmZddlmZddlmZmZdd	lmZmZmZGd
ddeZeejejZ ed!Z"ed
!Z#ed!Z$ed!Z%ed!Z&ed!Z'ed!Z(edZ)e ee)e BZ*ee ee*Z+e+dZ,e+Z-eddZ.e(e.Z/e-ee&e-Z0e"e
e0e#dZ1eej2ej3ej4BZ5eej2ej3ej4BZ6e5e6AZ7ee7ee&e7ddddZ8e
e$e8e%e8BZ9e9:dde	e9dZ;e;:dde	edZe:d de'Ze/e
e=Z?e,e
e1e?e>BZ@ee@eZAeABd!Gd"d#d#eCZDdS)$)absolute_importdivisionprint_functionN)stringStart	stringEndoriginalTextForParseException)
ZeroOrMoreWordOptionalRegexCombine)Literal)parse)MARKER_EXPRMarker)LegacySpecifier	SpecifierSpecifierSetc@seZdZdZdS)InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)__name__
__module____qualname____doc__rrF/usr/lib/python3.8/site-packages/pip/_vendor/packaging/requirements.pyrsr[](),;@z-_.namez[^ ]+urlextrasF)Z
joinStringZadjacent	_raw_speccCs
|jpdS)N)r'sltrrr7r-	specifiercCs|dS)Nrrr)rrrr-:r.markercCst||j|jS)N)rZ_original_startZ
_original_endr)rrrr->r.zx[]c@s(eZdZdZddZddZddZdS)	RequirementzParse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    c
Cszt|}WnDtk
rR}z&td||j|jd|jW5d}~XYnX|j|_|jrt		|j}|j
dkrt	||jkrtdn(|j
r|jr|j
s|jstd|j|j|_nd|_t
|jr|jng|_t|j|_|jr|jnd|_dS)NzParse error at "{0!r}": {1}filezInvalid URL givenzInvalid URL: {0})REQUIREMENTparseStringrrformatZlocmsgr$r%urlparseZschemeZ
urlunparseZnetlocsetr&ZasListrr/r0)selfZrequirement_stringZreqeZ
parsed_urlrrr__init__[s4


zRequirement.__init__cCs|jg}|jr*|ddt|j|jr@|t|j|jrh|d|j|j	rh|d|j	r|d|j	d|S)Nz[{0}]r!z@ {0} z; {0}r()
r$r&appendr6joinsortedr/strr%r0)r:partsrrr__str__vs
zRequirement.__str__cCsdt|S)Nz)r6rA)r:rrr__repr__szRequirement.__repr__N)rrrrr<rCrDrrrrr1Nsr1)EZ
__future__rrrstringreZpip._vendor.pyparsingrrrrr	r
rrr
rLZpip._vendor.six.moves.urllibrr8ZmarkersrrZ
specifiersrrr
ValueErrorrZ
ascii_lettersZdigitsZALPHANUMsuppressZLBRACKETZRBRACKETZLPARENZRPARENCOMMAZ	SEMICOLONATZPUNCTUATIONZIDENTIFIER_ENDZ
IDENTIFIERNAMEZEXTRAZURIZURLZEXTRAS_LISTZEXTRASZ
_regex_strVERBOSE
IGNORECASEZVERSION_PEP440ZVERSION_LEGACYZVERSION_ONEZVERSION_MANYZ
_VERSION_SPECZsetParseActionZVERSION_SPECZMARKER_SEPARATORZMARKERZVERSION_AND_MARKERZURL_AND_MARKERZNAMED_REQUIREMENTr4r5objectr1rrrrsh
site-packages/pip/_vendor/packaging/__pycache__/_structures.cpython-38.opt-1.pyc000064400000005301151464452710023625 0ustar00U

.e@sDddlmZmZmZGdddeZeZGdddeZeZdS))absolute_importdivisionprint_functionc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)InfinitycCsdS)NrselfrrE/usr/lib/python3.8/site-packages/pip/_vendor/packaging/_structures.py__repr__szInfinity.__repr__cCstt|SNhashreprrrrr	__hash__szInfinity.__hash__cCsdSNFrrotherrrr	__lt__szInfinity.__lt__cCsdSrrrrrr	__le__szInfinity.__le__cCst||jSr
isinstance	__class__rrrr	__eq__szInfinity.__eq__cCst||jSrrrrrr	__ne__szInfinity.__ne__cCsdSNTrrrrr	__gt__szInfinity.__gt__cCsdSrrrrrr	__ge__szInfinity.__ge__cCstSr)NegativeInfinityrrrr	__neg__ szInfinity.__neg__N__name__
__module____qualname__r
rrrrrrrrrrrr	rsrc@sTeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZdS)rcCsdS)Nz	-Infinityrrrrr	r
(szNegativeInfinity.__repr__cCstt|Srrrrrr	r+szNegativeInfinity.__hash__cCsdSrrrrrr	r.szNegativeInfinity.__lt__cCsdSrrrrrr	r1szNegativeInfinity.__le__cCst||jSrrrrrr	r4szNegativeInfinity.__eq__cCst||jSrrrrrr	r7szNegativeInfinity.__ne__cCsdSrrrrrr	r:szNegativeInfinity.__gt__cCsdSrrrrrr	r=szNegativeInfinity.__ge__cCstSr)rrrrr	r@szNegativeInfinity.__neg__Nrrrrr	r'srN)Z
__future__rrrobjectrrrrrr	ssite-packages/pip/_vendor/packaging/__pycache__/utils.cpython-38.pyc000064400000002624151464452710021451 0ustar00U

.e@sJddlmZmZmZddlZddlmZmZedZ	ddZ
dd	ZdS)
)absolute_importdivisionprint_functionN)InvalidVersionVersionz[-_.]+cCstd|S)N-)_canonicalize_regexsublower)namer
?/usr/lib/python3.8/site-packages/pip/_vendor/packaging/utils.pycanonicalize_namesrc
Cszt|}Wntk
r$|YSXg}|jdkrF|d|j|tddddd|jD|j	dk	r|dd	d|j	D|j
dk	r|d
|j
|jdk	r|d|j|jdk	r|d|jd|S)
z
    This is very similar to Version.__str__, but has one subtle differences
    with the way it handles the release segment.
    rz{0}!z(\.0)+$.css|]}t|VqdSNstr.0xr
r
r	'sz'canonicalize_version..Ncss|]}t|VqdSrrrr
r
rr+sz.post{0}z.dev{0}z+{0})
rrZepochappendformatrer
joinreleaseZpreZpostZdevZlocal)versionpartsr
r
rcanonicalize_versions"

&



r )Z
__future__rrrrrrrcompiler	rr r
r
r
rs

site-packages/pip/_vendor/packaging/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001027151464452710023003 0ustar00U

.e2@sTddlmZmZmZddlmZmZmZmZm	Z	m
Z
mZmZdddddd	d
dgZ
dS)
)absolute_importdivisionprint_function)
__author__
__copyright__	__email____license____summary__	__title____uri____version__rr
rr
rrr	rN)Z
__future__rrr	__about__rrrr	r
rrr
__all__rrB/usr/lib/python3.8/site-packages/pip/_vendor/packaging/__init__.pys(site-packages/pip/_vendor/packaging/__pycache__/tags.cpython-38.pyc000064400000025064151464452710021252 0ustar00U

.e2@snddlmZddlZzddlmZWn0ek
rTddlZddeDZ[YnXddl	Z	ddl
Z
ddlZddlZddl
Z
dddd	d
dZejdkZGd
ddeZddZddZddZddZddZddZddZddZdd Zd!d"Zd#d$Zefd%d&Zd'd(Zd=d)d*Z d+d,Z!d-d.Z"d/d0Z#d1d2Z$efd3d4Z%d5d6Z&d7d8Z'd9d:Z(d;d<Z)dS)>)absolute_importN)EXTENSION_SUFFIXEScCsg|]}|dqS)r).0xrr>/usr/lib/python3.8/site-packages/pip/_vendor/packaging/tags.py
srpycpppZipZjy)pythonZcpythonZpypyZ
ironpythonZjythonlc@sbeZdZdddgZddZeddZedd	Zed
dZdd
Z	ddZ
ddZddZdS)Tag_interpreter_abi	_platformcCs"||_||_||_dSN)lowerrrr)selfinterpreterabiplatformrrr__init__'s

zTag.__init__cCs|jSr)rrrrrr,szTag.interpretercCs|jSr)rrrrrr0szTag.abicCs|jSr)rrrrrr4szTag.platformcCs$|j|jko"|j|jko"|j|jkSr)rrr)rotherrrr__eq__8s


z
Tag.__eq__cCst|j|j|jfSr)hashrrrrrrr__hash__?szTag.__hash__cCsd|j|j|jS)Nz{}-{}-{})formatrrrrrrr__str__BszTag.__str__cCsdj|t|dS)Nz<{self} @ {self_id}>)rZself_id)ridrrrr__repr__EszTag.__repr__N)
__name__
__module____qualname__	__slots__rpropertyrrrrrrr rrrrr
#s



r
c	Cs`t}|d\}}}|dD]6}|dD]&}|dD]}|t|||qsz _cpython_tags..c3s|]}td|VqdSnoneNrSrTrUrrrVsr8r7rr9rR)r
ranger)r<rr.r/rr0r,
minor_versionrrUr
_cpython_tagssr\cCsdjtjdtjjtjjdS)Nz$pp{py_major}{pypy_major}{pypy_minor}r)Zpy_majorZ
pypy_majorZ
pypy_minor)rrMversion_infoZpypy_version_infor:r;rrrr_pypy_interpreters
r^cCstd}|rt|SdSdS)NZSOABIrX)rJrKr6)rrrr_generic_abis
r_c#sBfdd|DD]
}|Vqfdd|DD]
}|Vq2dS)Nc3s|]}t|VqdSrrSrrrrrrrVsz_pypy_tags..c3s|]}td|VqdSrWrSr`rUrrrVsr)r<rrr/r,rrar
_pypy_tagssrbc#sNfdd|DD]
}|VqdkrJfdd|D}|D]
}|Vq>dS)Nc3s|]}t|VqdSrrSr`rarrrVsz _generic_tags..rXc3s|]}td|VqdSrWrSrTrUrrrVsr)rr<rr/r,r-rrar
_generic_tagssrcccs\dj|d|ddVdj|ddVt|ddddD]}dj|d|dVq>dS)	z
    Yield Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all following versions up to 'end'.
    zpy{major}{minor}rr8r9z	py{major})r:rYN)rrZ)r<r;rrr_py_interpreter_rangesrdccsTt|D]}|D]}t|d|Vqqt|ddVt|D]}t|ddVq
    - -none-any
    - py*-none-any
    rXanyN)rdr
)rr<r/rFr0rrr_independent_tagss	rfcCs|s|S|drdSdS)Nppci386)
startswith)archis_32bitrrr	_mac_archs

rlcCs|g}|dkr,|dkrgS|dddgnp|dkrR|dkr@gS|dddgnJ|dkrz|d	ksj|dkrngS|dn"|d
kr|dkrgS|ddg|d|S)
NZx86_64)
rDZintelZfat64Zfat32rhZfatZppc64)rmrg)rmZ	universal)extendrO)rFcpu_archZformatsrrr_mac_binary_formatss&
rrc
	Cst\}}}|dkr2ttt|ddd}|dkrBt|}g}t|dddD]B}|d|f}t||}|D]"}	|	dj
|d|d|	dqtqV|S)Nr'r>r8rYrz&macosx_{major}_{minor}_{binary_format})r:r;
binary_format)rZmac_vertuplemapintr)rlrZrrrOr)
rFrjversion_strr2rqr/r[Zcompat_versionZbinary_formatsrsrrr_mac_platformss$
rxc	Cs>zddl}tt||dWSttfk
r4YnXt|S)NrZ_compatible)
_manylinuxboolgetattrImportErrorAttributeError_have_compatible_glibc)name
glibc_versionryrrr_is_manylinux_compatible	srcCsZddl}|d}z
|j}Wntk
r2YdSX|j|_|}t|tsV|d}|S)Nrascii)	ctypesZCDLLgnu_get_libc_versionr}Zc_char_pZrestype
isinstancestrdecode)rZprocess_namespacerrwrrr_glibc_version_strings



rcCsHtd|}|s$td|tdSt|d|koFt|d|kS)Nz$(?P[0-9]+)\.(?P[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFr:r;)rematchwarningswarnRuntimeWarningrvgroup)rwrequired_major
minimum_minorrCrrr_check_glibc_version1srcCst}|dkrdSt|||S)NF)rr)rrrwrrrr~Fsr~cs|ttjdkr|rdd}t|}|D]$\}}t||r.d|g}qXq.g}|fdd|D7}||S)NZlinux_x86_64Z
linux_i686))Z
manylinux2014)r>)Z
manylinux2010)r>)Z
manylinux1)r>rnlinuxcsg|]\}}d|qSrr3)rrr2rrrr^sz$_linux_platforms..)r6	distutilsutilget_platformiterrr4rO)rkZmanylinux_supportZmanylinux_support_iterrrr/rrr_linux_platformsMs

rcCsttj}|gSr)r6rrr)rrrr_generic_platformscsrcCst}t|p|Sr)rZpython_implementationrINTERPRETER_SHORT_NAMESget)rrrr_interpreter_namehsrcCs4td}|s&dtt|dd}dj||dS)NZpy_version_nodotr?r>z{name}{version})rrF)rJrKjoinrurr)rr<rFrrr_generic_interpreterms
rccstjdd}t}tdkr(t}ntdkrZDarwinZLinuxr
r)rMr]rrsystemrxrrr=rQr\r^r_rbrrcrf)r<Zinterpreter_namer/rr.r,rrrrsys_tagsts.


r)NN)*Z
__future__rZdistutils.utilrZimportlib.machineryrr|ZimpZget_suffixesrrrMrJrrmaxsizeZ_32_BIT_INTERPRETERobjectr
r1r6r=rQr\r^r_rbrcrdrfrlrrrxrrrr~rrrrrrrrrsV	
&
#	



site-packages/pip/_vendor/packaging/__pycache__/markers.cpython-38.opt-1.pyc000064400000021155151464452710022714 0ustar00U

.e: 	@s@ddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
mZmZddlm
Z
mZmZmZddlmZddlmZddlmZmZd	d
ddd
gZGdd	d	eZGdd
d
eZGdddeZGdddeZGdddeZGdddeZ GdddeZ!ededBedBedBedBedBedBed Bed!Bed"Bed#Bed$Bed%Bed&Bed'Bed(Bed)Bed*BZ"d#d"ddddd+Z#e"$d,d-ed.ed/Bed0Bed1Bed2Bed3Bed4Bed5BZ%e%ed6Bed7BZ&e&$d8d-ed9ed:BZ'e'$d;d-ed<ed=BZ(e"e'BZ)ee)e&e)Z*e*$d>d-ed?+Z,ed@+Z-eZ.e*ee,e.e-BZ/e.e/e
e(e.>ee.eZ0dAdBZ1dSdDdEZ2dFd-dGd-ej3ej4ej5ej6ej7ej8dHZ9dIdJZ:eZ;dKdLZdQd
Z?GdRddeZ@dS)T)absolute_importdivisionprint_functionN)ParseExceptionParseResultsstringStart	stringEnd)
ZeroOrMoreGroupForwardQuotedString)Literal)string_types)	SpecifierInvalidSpecifier
InvalidMarkerUndefinedComparisonUndefinedEnvironmentNameMarkerdefault_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N__name__
__module____qualname____doc__rrA/usr/lib/python3.8/site-packages/pip/_vendor/packaging/markers.pyrsc@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrrrrrr"sc@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrrrrrr(sc@s,eZdZddZddZddZddZd	S)
NodecCs
||_dSN)value)selfr rrr__init__0sz
Node.__init__cCs
t|jSr)strr r!rrr__str__3szNode.__str__cCsd|jjt|S)Nz<{0}({1!r})>)format	__class__rr#r$rrr__repr__6sz
Node.__repr__cCstdSr)NotImplementedErrorr$rrr	serialize9szNode.serializeN)rrrr"r%r(r*rrrrr/src@seZdZddZdS)VariablecCst|Srr#r$rrrr*>szVariable.serializeNrrrr*rrrrr+=sr+c@seZdZddZdS)ValuecCs
d|S)Nz"{0}")r&r$rrrr*CszValue.serializeNr-rrrrr.Bsr.c@seZdZddZdS)OpcCst|Srr,r$rrrr*HszOp.serializeNr-rrrrr/Gsr/implementation_versionplatform_python_implementationimplementation_namepython_full_versionplatform_releaseplatform_versionplatform_machineplatform_systempython_versionsys_platformos_nameos.namesys.platformplatform.versionplatform.machineplatform.python_implementationpython_implementationZextra)r;r<r=r>r?r@cCstt|d|dSNr)r+ALIASESgetsltrrrhrHz=====>=<=!=z~=><not inincCst|dSrA)r/rDrrrrHorI'"cCst|dSrA)r.rDrrrrHrrIandorcCst|dSrA)tuplerDrrrrHyrI()cCs t|trdd|DS|SdS)NcSsg|]}t|qSr)_coerce_parse_result).0irrr
sz(_coerce_parse_result..)
isinstancer)resultsrrrrYs
rYTcCst|tr4t|dkr4t|dttfr4t|dSt|trndd|D}|rZd|Sdd|dSn"t|trddd	|DS|SdS)
Nrrcss|]}t|ddVqdS)F)firstN)_format_markerrZmrrr	sz!_format_marker.. rWrXcSsg|]}|qSr)r*rarrrr\sz"_format_marker..)r]listlenrVr`join)markerr_innerrrrr`s



r`cCs||kSrrlhsrhsrrrrHrIcCs||kSrrrjrrrrHrI)rQrPrOrLrJrMrKrNcCslztd||g}Wntk
r.YnX||St|}|dkrbtd||||||S)Nz#Undefined {0!r} on {1!r} and {2!r}.)	rrgr*rcontains
_operatorsrCrr&)rkoprlspecZoperrrr_eval_ops
rrcCs&||t}|tkr"td||S)Nz/{0!r} does not exist in evaluation environment.)rC
_undefinedrr&)environmentnamer rrr_get_envsrvc	Csgg}|D]}t|tr.|dt||q
t|tr|\}}}t|tr`t||j}|j}n|j}t||j}|dt|||q
|dkr
|gq
t	dd|DS)NrUcss|]}t|VqdSr)all)rZitemrrrrcsz$_evaluate_markers..)
r]reappend_evaluate_markersrVr+rvr rrany)	ZmarkersrtgroupsrhrkrprlZ	lhs_valueZ	rhs_valuerrrr{s



r{cCs2d|}|j}|dkr.||dt|j7}|S)Nz{0.major}.{0.minor}.{0.micro}finalr)r&releaselevelr#serial)infoversionZkindrrrformat_full_versions

rcCsrttdr ttjj}tjj}nd}d}||tjtt	t
tttd
tddtjdS)Nimplementation0rm.)r2r0r:r6r4r7r5r3r1r8r9)hasattrsysrrrruosplatformmachinereleasesystemr8r@rgZpython_version_tuple)Ziverr2rrrrs"

c@s.eZdZddZddZddZd
dd	ZdS)rc
Cs`ztt||_WnFtk
rZ}z(d|||j|jd}t|W5d}~XYnXdS)Nz+Invalid marker: {0!r}, parse error at {1!r})rYMARKERZparseString_markersrr&Zlocr)r!rheZerr_strrrrr"szMarker.__init__cCs
t|jSr)r`rr$rrrr%szMarker.__str__cCsdt|S)Nz)r&r#r$rrrr(szMarker.__repr__NcCs$t}|dk	r||t|j|S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)rupdater{r)r!rtZcurrent_environmentrrrevaluates	
zMarker.evaluate)N)rrrr"r%r(rrrrrrs	)T)AZ
__future__rrroperatorrrrZpip._vendor.pyparsingrrrrr	r
rrr
LZ_compatrZ
specifiersrr__all__
ValueErrorrrrobjectrr+r.r/ZVARIABLErBZsetParseActionZVERSION_CMPZ	MARKER_OPZMARKER_VALUEZBOOLOPZ
MARKER_VARZMARKER_ITEMsuppressZLPARENZRPARENZMARKER_EXPRZMARKER_ATOMrrYr`ltleeqnegegtrorrrsrvr{rrrrrrrs		

>
site-packages/pip/_vendor/packaging/__pycache__/specifiers.cpython-38.opt-1.pyc000064400000046417151464452710023414 0ustar00U

.el@sddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZm
Z
mZGdddeZGdd	d	e
ejeZGd
ddeZGdd
d
eZddZGdddeZedZddZddZGdddeZdS))absolute_importdivisionprint_functionN)string_typeswith_metaclass)Version
LegacyVersionparsec@seZdZdZdS)InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrD/usr/lib/python3.8/site-packages/pip/_vendor/packaging/specifiers.pyrsrc@seZdZejddZejddZejddZejddZej	d	d
Z
e
jdd
Z
ejdd
dZejdddZ
dS)
BaseSpecifiercCsdS)z
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nrselfrrr__str__szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr__hash__szBaseSpecifier.__hash__cCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nrrotherrrr__eq__#szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrrrrr__ne__*szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrrprereleases1szBaseSpecifier.prereleasescCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrvaluerrrr8sNcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nrritemrrrrcontains?szBaseSpecifier.containscCsdS)z
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)riterablerrrrfilterEszBaseSpecifier.filter)N)N)rr
rabcabstractmethodrrrrabstractpropertyrsetterr r"rrrrrs 





rc@seZdZiZd ddZddZddZd	d
ZddZd
dZ	ddZ
ddZeddZ
eddZeddZejddZddZd!ddZd"ddZdS)#_IndividualSpecifierNcCsF|j|}|std||d|df|_||_dS)NzInvalid specifier: '{0}'operatorversion)_regexsearchrformatgroupstrip_spec_prereleases)rspecrmatchrrr__init__Qs
z_IndividualSpecifier.__init__cCs0|jdk	rd|jnd}d|jjt||S)N, prereleases={0!r}r(z<{0}({1!r}{2})>)r1r-r	__class__rstrrZprerrr__repr__[s
z_IndividualSpecifier.__repr__cCsdj|jS)Nz{0}{1})r-r0rrrrrdsz_IndividualSpecifier.__str__cCs
t|jSN)hashr0rrrrrgsz_IndividualSpecifier.__hash__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:
isinstancerr6rNotImplementedr0rrrrrjs
z_IndividualSpecifier.__eq__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:r<rrrrrus
z_IndividualSpecifier.__ne__cCst|d|j|S)Nz_compare_{0})getattrr-
_operators)roprrr
_get_operatorsz"_IndividualSpecifier._get_operatorcCst|ttfst|}|Sr:)r=r	rr
rr*rrr_coerce_versionsz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nrr0rrrrr)sz_IndividualSpecifier.operatorcCs
|jdS)NrrErrrrr*sz_IndividualSpecifier.versioncCs|jSr:r1rrrrrsz _IndividualSpecifier.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:r rrrrr__contains__sz!_IndividualSpecifier.__contains__cCs:|dkr|j}||}|jr&|s&dS||j||jSNF)rrD
is_prereleaserBr)r*rrrrr s

z_IndividualSpecifier.containsccsd}g}d|dk	r|ndi}|D]B}||}|j|f|r |jrX|sX|jsX||q d}|Vq |s||r||D]
}|VqpdS)NFrT)rDr rKrappend)rr!rZyieldedfound_prereleaseskwr*parsed_versionrrrr"s"
z_IndividualSpecifier.filter)r(N)N)N)rr
rr@r4r9rrrrrBrDpropertyr)r*rr&rIr r"rrrrr'Ms(

	




r'c@sveZdZdZededejejBZdddddd	d
Z	ddZ
d
dZddZddZ
ddZddZddZdS)LegacySpecifiera
        (?P(==|!=|<=|>=|<|>))
        \s*
        (?P
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        ^\s*\s*$equal	not_equalless_than_equalgreater_than_equal	less_thangreater_than)==!=<=>=<>cCst|tstt|}|Sr:)r=r	r7rCrrrrDs
zLegacySpecifier._coerce_versioncCs|||kSr:rDrprospectiver2rrr_compare_equalszLegacySpecifier._compare_equalcCs|||kSr:r`rarrr_compare_not_equalsz"LegacySpecifier._compare_not_equalcCs|||kSr:r`rarrr_compare_less_than_equalsz(LegacySpecifier._compare_less_than_equalcCs|||kSr:r`rarrr_compare_greater_than_equalsz+LegacySpecifier._compare_greater_than_equalcCs|||kSr:r`rarrr_compare_less_thansz"LegacySpecifier._compare_less_thancCs|||kSr:r`rarrr_compare_greater_thansz%LegacySpecifier._compare_greater_thanN)rr
r
_regex_strrecompileVERBOSE
IGNORECASEr+r@rDrcrdrerfrgrhrrrrrQs 	rQcstfdd}|S)Ncst|tsdS|||SrJ)r=rrafnrrwrappeds
z)_require_version_compare..wrapped)	functoolswraps)rorprrnr_require_version_comparesrsc	@seZdZdZededejejBZdddddd	d
ddZ	e
d
dZe
ddZe
ddZ
e
ddZe
ddZe
ddZe
ddZddZeddZejddZd S)!	Specifiera
        (?P(~=|==|!=|<=|>=|<|>|===))
        (?P
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?z/Specifier._compare_compatible...*r]rZ)joinlist	itertools	takewhile_version_splitrB)rrbr2prefixrrr_compare_compatiblevs
zSpecifier._compare_compatiblecCsp|drPt|j}t|dd}tt|}|dt|}t||\}}nt|}|jsht|j}||kS)Nr~)endswithrZpublicrr7len_pad_versionlocalrarrrrcs


zSpecifier._compare_equalcCs|||Sr:)rcrarrrrdszSpecifier._compare_not_equalcCs|t|kSr:rrarrrresz"Specifier._compare_less_than_equalcCs|t|kSr:rrarrrrfsz%Specifier._compare_greater_than_equalcCs<t|}||ksdS|js8|jr8t|jt|jkr8dSdSNFT)rrKbase_versionrarrrrgszSpecifier._compare_less_thancCs^t|}||ksdS|js8|jr8t|jt|jkr8dS|jdk	rZt|jt|jkrZdSdSr)rZis_postreleaserrrarrrrhs
zSpecifier._compare_greater_thancCst|t|kSr:)r7lowerrarrr_compare_arbitraryszSpecifier._compare_arbitrarycCsR|jdk	r|jS|j\}}|dkrN|dkr@|dr@|dd}t|jrNdSdS)N)rZr]r\rurvrZr~rTF)r1r0rr
rK)rr)r*rrrrs


zSpecifier.prereleasescCs
||_dSr:rFrrrrrsN)rr
rrirjrkrlrmr+r@rsrrcrdrerfrgrhrrPrr&rrrrrt
s<]

"





rtz^([0-9]+)((?:a|b|c|rc)[0-9]+)$cCs@g}|dD],}t|}|r0||q||q|S)Nrw)split
_prefix_regexr,extendgroupsrL)r*resultrr3rrrrs
rc
Csgg}}|ttdd||ttdd|||t|dd||t|dd|ddgtdt|dt|d|ddgtdt|dt|dttj|ttj|fS)NcSs|Sr:isdigitryrrrr{*r|z_pad_version..cSs|Sr:rryrrrr{+r|rr0)rLrrrrinsertmaxchain)leftrightZ
left_splitZright_splitrrrr&s
,,rc@seZdZdddZddZddZd	d
ZddZd
dZddZ	ddZ
ddZeddZ
e
jddZ
ddZdddZd ddZdS)!SpecifierSetr(Nc	Csndd|dD}t}|D]:}z|t|Wqtk
rV|t|YqXqt||_||_dS)NcSsg|]}|r|qSr)r/.0srrr
<sz)SpecifierSet.__init__..,)	rsetaddrtrrQ	frozenset_specsr1)rZ
specifiersrZparsed	specifierrrrr49s
zSpecifierSet.__init__cCs*|jdk	rd|jnd}dt||S)Nr5r(z)r1r-rr7r8rrrr9Ns
zSpecifierSet.__repr__cCsdtdd|jDS)Nrcss|]}t|VqdSr:)r7rrrr	Xsz'SpecifierSet.__str__..)rsortedrrrrrrWszSpecifierSet.__str__cCs
t|jSr:)r;rrrrrrZszSpecifierSet.__hash__cCst|trt|}nt|ts"tSt}t|j|jB|_|jdkrX|jdk	rX|j|_n<|jdk	rv|jdkrv|j|_n|j|jkr|j|_ntd|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)r=rrr>rrr1
ValueError)rrrrrr__and__]s 





zSpecifierSet.__and__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:r=rrr'r7r>rrrrrrts



zSpecifierSet.__eq__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:rrrrrr~s



zSpecifierSet.__ne__cCs
t|jSr:)rrrrrr__len__szSpecifierSet.__len__cCs
t|jSr:)iterrrrrr__iter__szSpecifierSet.__iter__cCs.|jdk	r|jS|jsdStdd|jDS)Ncss|]}|jVqdSr:rrrrrrsz+SpecifierSet.prereleases..)r1ranyrrrrrs

zSpecifierSet.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:rGrHrrrrIszSpecifierSet.__contains__csLtttfstdkr$|js2jr2dStfdd|jDS)NFc3s|]}|jdVqdS)rNrGrrrrrrsz(SpecifierSet.contains..)r=r	rr
rrKallrrrrrr s
zSpecifierSet.containscCs|dkr|j}|jr6|jD]}|j|t|d}q|Sg}g}|D]P}t|ttfs^t|}n|}t|trnqB|jr|s|s|	|qB|	|qB|s|r|dkr|S|SdS)Nr)
rrr"boolr=r	rr
rKrL)rr!rr2ZfilteredrMrrOrrrr"s*



zSpecifierSet.filter)r(N)N)N)rr
rr4r9rrrrrrrrPrr&rIr r"rrrrr8s 
	




r)Z
__future__rrrr#rqrrjZ_compatrrr*rr	r
rrABCMetaobjectrr'rQrsrtrkrrrrrrrrs&81

site-packages/pip/_vendor/packaging/__pycache__/utils.cpython-38.opt-1.pyc000064400000002624151464452710022410 0ustar00U

.e@sJddlmZmZmZddlZddlmZmZedZ	ddZ
dd	ZdS)
)absolute_importdivisionprint_functionN)InvalidVersionVersionz[-_.]+cCstd|S)N-)_canonicalize_regexsublower)namer
?/usr/lib/python3.8/site-packages/pip/_vendor/packaging/utils.pycanonicalize_namesrc
Cszt|}Wntk
r$|YSXg}|jdkrF|d|j|tddddd|jD|j	dk	r|dd	d|j	D|j
dk	r|d
|j
|jdk	r|d|j|jdk	r|d|jd|S)
z
    This is very similar to Version.__str__, but has one subtle differences
    with the way it handles the release segment.
    rz{0}!z(\.0)+$.css|]}t|VqdSNstr.0xr
r
r	'sz'canonicalize_version..Ncss|]}t|VqdSrrrr
r
rr+sz.post{0}z.dev{0}z+{0})
rrZepochappendformatrer
joinreleaseZpreZpostZdevZlocal)versionpartsr
r
rcanonicalize_versions"

&



r )Z
__future__rrrrrrrcompiler	rr r
r
r
rs

site-packages/pip/_vendor/packaging/__pycache__/specifiers.cpython-38.pyc000064400000046417151464452710022455 0ustar00U

.el@sddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZm
Z
mZGdddeZGdd	d	e
ejeZGd
ddeZGdd
d
eZddZGdddeZedZddZddZGdddeZdS))absolute_importdivisionprint_functionN)string_typeswith_metaclass)Version
LegacyVersionparsec@seZdZdZdS)InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)__name__
__module____qualname____doc__rrD/usr/lib/python3.8/site-packages/pip/_vendor/packaging/specifiers.pyrsrc@seZdZejddZejddZejddZejddZej	d	d
Z
e
jdd
Z
ejdd
dZejdddZ
dS)
BaseSpecifiercCsdS)z
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nrselfrrr__str__szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr__hash__szBaseSpecifier.__hash__cCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nrrotherrrr__eq__#szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrrrrr__ne__*szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrrprereleases1szBaseSpecifier.prereleasescCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrvaluerrrr8sNcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nrritemrrrrcontains?szBaseSpecifier.containscCsdS)z
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)riterablerrrrfilterEszBaseSpecifier.filter)N)N)rr
rabcabstractmethodrrrrabstractpropertyrsetterr r"rrrrrs 





rc@seZdZiZd ddZddZddZd	d
ZddZd
dZ	ddZ
ddZeddZ
eddZeddZejddZddZd!ddZd"ddZdS)#_IndividualSpecifierNcCsF|j|}|std||d|df|_||_dS)NzInvalid specifier: '{0}'operatorversion)_regexsearchrformatgroupstrip_spec_prereleases)rspecrmatchrrr__init__Qs
z_IndividualSpecifier.__init__cCs0|jdk	rd|jnd}d|jjt||S)N, prereleases={0!r}r(z<{0}({1!r}{2})>)r1r-r	__class__rstrrZprerrr__repr__[s
z_IndividualSpecifier.__repr__cCsdj|jS)Nz{0}{1})r-r0rrrrrdsz_IndividualSpecifier.__str__cCs
t|jSN)hashr0rrrrrgsz_IndividualSpecifier.__hash__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:
isinstancerr6rNotImplementedr0rrrrrjs
z_IndividualSpecifier.__eq__cCsPt|tr4z||}WqDtk
r0tYSXnt||jsDtS|j|jkSr:r<rrrrrus
z_IndividualSpecifier.__ne__cCst|d|j|S)Nz_compare_{0})getattrr-
_operators)roprrr
_get_operatorsz"_IndividualSpecifier._get_operatorcCst|ttfst|}|Sr:)r=r	rr
rr*rrr_coerce_versionsz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nrr0rrrrr)sz_IndividualSpecifier.operatorcCs
|jdS)NrrErrrrr*sz_IndividualSpecifier.versioncCs|jSr:r1rrrrrsz _IndividualSpecifier.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:r rrrrr__contains__sz!_IndividualSpecifier.__contains__cCs:|dkr|j}||}|jr&|s&dS||j||jSNF)rrD
is_prereleaserBr)r*rrrrr s

z_IndividualSpecifier.containsccsd}g}d|dk	r|ndi}|D]B}||}|j|f|r |jrX|sX|jsX||q d}|Vq |s||r||D]
}|VqpdS)NFrT)rDr rKrappend)rr!rZyieldedfound_prereleaseskwr*parsed_versionrrrr"s"
z_IndividualSpecifier.filter)r(N)N)N)rr
rr@r4r9rrrrrBrDpropertyr)r*rr&rIr r"rrrrr'Ms(

	




r'c@sveZdZdZededejejBZdddddd	d
Z	ddZ
d
dZddZddZ
ddZddZddZdS)LegacySpecifiera
        (?P(==|!=|<=|>=|<|>))
        \s*
        (?P
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        ^\s*\s*$equal	not_equalless_than_equalgreater_than_equal	less_thangreater_than)==!=<=>=<>cCst|tstt|}|Sr:)r=r	r7rCrrrrDs
zLegacySpecifier._coerce_versioncCs|||kSr:rDrprospectiver2rrr_compare_equalszLegacySpecifier._compare_equalcCs|||kSr:r`rarrr_compare_not_equalsz"LegacySpecifier._compare_not_equalcCs|||kSr:r`rarrr_compare_less_than_equalsz(LegacySpecifier._compare_less_than_equalcCs|||kSr:r`rarrr_compare_greater_than_equalsz+LegacySpecifier._compare_greater_than_equalcCs|||kSr:r`rarrr_compare_less_thansz"LegacySpecifier._compare_less_thancCs|||kSr:r`rarrr_compare_greater_thansz%LegacySpecifier._compare_greater_thanN)rr
r
_regex_strrecompileVERBOSE
IGNORECASEr+r@rDrcrdrerfrgrhrrrrrQs 	rQcstfdd}|S)Ncst|tsdS|||SrJ)r=rrafnrrwrappeds
z)_require_version_compare..wrapped)	functoolswraps)rorprrnr_require_version_comparesrsc	@seZdZdZededejejBZdddddd	d
ddZ	e
d
dZe
ddZe
ddZ
e
ddZe
ddZe
ddZe
ddZddZeddZejddZd S)!	Specifiera
        (?P(~=|==|!=|<=|>=|<|>|===))
        (?P
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?z/Specifier._compare_compatible...*r]rZ)joinlist	itertools	takewhile_version_splitrB)rrbr2prefixrrr_compare_compatiblevs
zSpecifier._compare_compatiblecCsp|drPt|j}t|dd}tt|}|dt|}t||\}}nt|}|jsht|j}||kS)Nr~)endswithrZpublicrr7len_pad_versionlocalrarrrrcs


zSpecifier._compare_equalcCs|||Sr:)rcrarrrrdszSpecifier._compare_not_equalcCs|t|kSr:rrarrrresz"Specifier._compare_less_than_equalcCs|t|kSr:rrarrrrfsz%Specifier._compare_greater_than_equalcCs<t|}||ksdS|js8|jr8t|jt|jkr8dSdSNFT)rrKbase_versionrarrrrgszSpecifier._compare_less_thancCs^t|}||ksdS|js8|jr8t|jt|jkr8dS|jdk	rZt|jt|jkrZdSdSr)rZis_postreleaserrrarrrrhs
zSpecifier._compare_greater_thancCst|t|kSr:)r7lowerrarrr_compare_arbitraryszSpecifier._compare_arbitrarycCsR|jdk	r|jS|j\}}|dkrN|dkr@|dr@|dd}t|jrNdSdS)N)rZr]r\rurvrZr~rTF)r1r0rr
rK)rr)r*rrrrs


zSpecifier.prereleasescCs
||_dSr:rFrrrrrsN)rr
rrirjrkrlrmr+r@rsrrcrdrerfrgrhrrPrr&rrrrrt
s<]

"





rtz^([0-9]+)((?:a|b|c|rc)[0-9]+)$cCs@g}|dD],}t|}|r0||q||q|S)Nrw)split
_prefix_regexr,extendgroupsrL)r*resultrr3rrrrs
rc
Csgg}}|ttdd||ttdd|||t|dd||t|dd|ddgtdt|dt|d|ddgtdt|dt|dttj|ttj|fS)NcSs|Sr:isdigitryrrrr{*r|z_pad_version..cSs|Sr:rryrrrr{+r|rr0)rLrrrrinsertmaxchain)leftrightZ
left_splitZright_splitrrrr&s
,,rc@seZdZdddZddZddZd	d
ZddZd
dZddZ	ddZ
ddZeddZ
e
jddZ
ddZdddZd ddZdS)!SpecifierSetr(Nc	Csndd|dD}t}|D]:}z|t|Wqtk
rV|t|YqXqt||_||_dS)NcSsg|]}|r|qSr)r/.0srrr
<sz)SpecifierSet.__init__..,)	rsetaddrtrrQ	frozenset_specsr1)rZ
specifiersrZparsed	specifierrrrr49s
zSpecifierSet.__init__cCs*|jdk	rd|jnd}dt||S)Nr5r(z)r1r-rr7r8rrrr9Ns
zSpecifierSet.__repr__cCsdtdd|jDS)Nrcss|]}t|VqdSr:)r7rrrr	Xsz'SpecifierSet.__str__..)rsortedrrrrrrWszSpecifierSet.__str__cCs
t|jSr:)r;rrrrrrZszSpecifierSet.__hash__cCst|trt|}nt|ts"tSt}t|j|jB|_|jdkrX|jdk	rX|j|_n<|jdk	rv|jdkrv|j|_n|j|jkr|j|_ntd|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)r=rrr>rrr1
ValueError)rrrrrr__and__]s 





zSpecifierSet.__and__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:r=rrr'r7r>rrrrrrts



zSpecifierSet.__eq__cCsFt|trt|}n&t|tr,tt|}nt|ts:tS|j|jkSr:rrrrrr~s



zSpecifierSet.__ne__cCs
t|jSr:)rrrrrr__len__szSpecifierSet.__len__cCs
t|jSr:)iterrrrrr__iter__szSpecifierSet.__iter__cCs.|jdk	r|jS|jsdStdd|jDS)Ncss|]}|jVqdSr:rrrrrrsz+SpecifierSet.prereleases..)r1ranyrrrrrs

zSpecifierSet.prereleasescCs
||_dSr:rFrrrrrscCs
||Sr:rGrHrrrrIszSpecifierSet.__contains__csLtttfstdkr$|js2jr2dStfdd|jDS)NFc3s|]}|jdVqdS)rNrGrrrrrrsz(SpecifierSet.contains..)r=r	rr
rrKallrrrrrr s
zSpecifierSet.containscCs|dkr|j}|jr6|jD]}|j|t|d}q|Sg}g}|D]P}t|ttfs^t|}n|}t|trnqB|jr|s|s|	|qB|	|qB|s|r|dkr|S|SdS)Nr)
rrr"boolr=r	rr
rKrL)rr!rr2ZfilteredrMrrOrrrr"s*



zSpecifierSet.filter)r(N)N)N)rr
rr4r9rrrrrrrrPrr&rIr r"rrrrr8s 
	




r)Z
__future__rrrr#rqrrjZ_compatrrr*rr	r
rrABCMetaobjectrr'rQrsrtrkrrrrrrrrs&81

site-packages/pip/_vendor/packaging/__pycache__/markers.cpython-38.pyc000064400000021322151464452710021751 0ustar00U

.e: 	@s@ddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
mZmZddlm
Z
mZmZmZddlmZddlmZddlmZmZd	d
ddd
gZGdd	d	eZGdd
d
eZGdddeZGdddeZGdddeZGdddeZ GdddeZ!ededBedBedBedBedBedBed Bed!Bed"Bed#Bed$Bed%Bed&Bed'Bed(Bed)Bed*BZ"d#d"ddddd+Z#e"$d,d-ed.ed/Bed0Bed1Bed2Bed3Bed4Bed5BZ%e%ed6Bed7BZ&e&$d8d-ed9ed:BZ'e'$d;d-ed<ed=BZ(e"e'BZ)ee)e&e)Z*e*$d>d-ed?+Z,ed@+Z-eZ.e*ee,e.e-BZ/e.e/e
e(e.>ee.eZ0dAdBZ1dSdDdEZ2dFd-dGd-ej3ej4ej5ej6ej7ej8dHZ9dIdJZ:eZ;dKdLZdQd
Z?GdRddeZ@dS)T)absolute_importdivisionprint_functionN)ParseExceptionParseResultsstringStart	stringEnd)
ZeroOrMoreGroupForwardQuotedString)Literal)string_types)	SpecifierInvalidSpecifier
InvalidMarkerUndefinedComparisonUndefinedEnvironmentNameMarkerdefault_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N__name__
__module____qualname____doc__rrA/usr/lib/python3.8/site-packages/pip/_vendor/packaging/markers.pyrsc@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrrrrrr"sc@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrrrrrr(sc@s,eZdZddZddZddZddZd	S)
NodecCs
||_dSN)value)selfr rrr__init__0sz
Node.__init__cCs
t|jSr)strr r!rrr__str__3szNode.__str__cCsd|jjt|S)Nz<{0}({1!r})>)format	__class__rr#r$rrr__repr__6sz
Node.__repr__cCstdSr)NotImplementedErrorr$rrr	serialize9szNode.serializeN)rrrr"r%r(r*rrrrr/src@seZdZddZdS)VariablecCst|Srr#r$rrrr*>szVariable.serializeNrrrr*rrrrr+=sr+c@seZdZddZdS)ValuecCs
d|S)Nz"{0}")r&r$rrrr*CszValue.serializeNr-rrrrr.Bsr.c@seZdZddZdS)OpcCst|Srr,r$rrrr*HszOp.serializeNr-rrrrr/Gsr/implementation_versionplatform_python_implementationimplementation_namepython_full_versionplatform_releaseplatform_versionplatform_machineplatform_systempython_versionsys_platformos_nameos.namesys.platformplatform.versionplatform.machineplatform.python_implementationpython_implementationZextra)r;r<r=r>r?r@cCstt|d|dSNr)r+ALIASESgetsltrrrhrHz=====>=<=!=z~=><not inincCst|dSrA)r/rDrrrrHorI'"cCst|dSrA)r.rDrrrrHrrIandorcCst|dSrA)tuplerDrrrrHyrI()cCs t|trdd|DS|SdS)NcSsg|]}t|qSr)_coerce_parse_result).0irrr
sz(_coerce_parse_result..)
isinstancer)resultsrrrrYs
rYTcCst|tttfstt|trHt|dkrHt|dttfrHt|dSt|trdd|D}|rnd|Sdd|dSn"t|trddd	|DS|SdS)
Nrrcss|]}t|ddVqdS)F)firstN)_format_markerrZmrrr	sz!_format_marker.. rWrXcSsg|]}|qSr)r*rarrrr\sz"_format_marker..)r]listrVrAssertionErrorlenr`join)markerr_innerrrrr`s 



r`cCs||kSrrlhsrhsrrrrHrIcCs||kSrrrkrrrrHrI)rQrPrOrLrJrMrKrNcCslztd||g}Wntk
r.YnX||St|}|dkrbtd||||||S)Nz#Undefined {0!r} on {1!r} and {2!r}.)	rrhr*rcontains
_operatorsrCrr&)rloprmspecZoperrrr_eval_ops
rscCs&||t}|tkr"td||S)Nz/{0!r} does not exist in evaluation environment.)rC
_undefinedrr&)environmentnamer rrr_get_envsrwc	Csgg}|D]}t|tttfs"tt|trB|dt||q
t|tr|\}}}t|trtt||j	}|j	}n|j	}t||j	}|dt
|||q
|dkst|dkr
|gq
tdd|DS)N)rTrUrUcss|]}t|VqdSr)all)rZitemrrrrcsz$_evaluate_markers..)r]rerVrrfappend_evaluate_markersr+rwr rsany)	ZmarkersrugroupsrirlrqrmZ	lhs_valueZ	rhs_valuerrrr|s"



r|cCs2d|}|j}|dkr.||dt|j7}|S)Nz{0.major}.{0.minor}.{0.micro}finalr)r&releaselevelr#serial)infoversionZkindrrrformat_full_versions

rcCsrttdr ttjj}tjj}nd}d}||tjtt	t
tttd
tddtjdS)Nimplementation0rn.)r2r0r:r6r4r7r5r3r1r8r9)hasattrsysrrrrvosplatformmachinereleasesystemr8r@rhZpython_version_tuple)Ziverr2rrrrs"

c@s.eZdZddZddZddZd
dd	ZdS)rc
Cs`ztt||_WnFtk
rZ}z(d|||j|jd}t|W5d}~XYnXdS)Nz+Invalid marker: {0!r}, parse error at {1!r})rYMARKERZparseString_markersrr&Zlocr)r!rieZerr_strrrrr"szMarker.__init__cCs
t|jSr)r`rr$rrrr%szMarker.__str__cCsdt|S)Nz)r&r#r$rrrr(szMarker.__repr__NcCs$t}|dk	r||t|j|S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)rupdater|r)r!ruZcurrent_environmentrrrevaluates	
zMarker.evaluate)N)rrrr"r%r(rrrrrrs	)T)AZ
__future__rrroperatorrrrZpip._vendor.pyparsingrrrrr	r
rrr
LZ_compatrZ
specifiersrr__all__
ValueErrorrrrobjectrr+r.r/ZVARIABLErBZsetParseActionZVERSION_CMPZ	MARKER_OPZMARKER_VALUEZBOOLOPZ
MARKER_VARZMARKER_ITEMsuppressZLPARENZRPARENZMARKER_EXPRZMARKER_ATOMrrYr`ltleeqnegegtrprsrtrwr|rrrrrrrs		

>
site-packages/pip/_vendor/packaging/__about__.py000064400000001350151464452710015724 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]

__title__ = "packaging"
__summary__ = "Core utilities for Python packages"
__uri__ = "https://github.com/pypa/packaging"

__version__ = "19.2"

__author__ = "Donald Stufft and individual contributors"
__email__ = "donald@stufft.io"

__license__ = "BSD or Apache License, Version 2.0"
__copyright__ = "Copyright 2014-2019 %s" % __author__
site-packages/pip/_vendor/packaging/_compat.py000064400000001541151464452710015442 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import sys


PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3

# flake8: noqa

if PY3:
    string_types = (str,)
else:
    string_types = (basestring,)


def with_metaclass(meta, *bases):
    """
    Create a base class with a metaclass.
    """
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(meta):
        def __new__(cls, name, this_bases, d):
            return meta(name, bases, d)

    return type.__new__(metaclass, "temporary_class", (), {})
site-packages/pip/_vendor/packaging/specifiers.py000064400000066202151464452710016161 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import abc
import functools
import itertools
import re

from ._compat import string_types, with_metaclass
from .version import Version, LegacyVersion, parse


class InvalidSpecifier(ValueError):
    """
    An invalid specifier was found, users should refer to PEP 440.
    """


class BaseSpecifier(with_metaclass(abc.ABCMeta, object)):
    @abc.abstractmethod
    def __str__(self):
        """
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        """

    @abc.abstractmethod
    def __hash__(self):
        """
        Returns a hash value for this Specifier like object.
        """

    @abc.abstractmethod
    def __eq__(self, other):
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        """

    @abc.abstractmethod
    def __ne__(self, other):
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        """

    @abc.abstractproperty
    def prereleases(self):
        """
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @prereleases.setter
    def prereleases(self, value):
        """
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @abc.abstractmethod
    def contains(self, item, prereleases=None):
        """
        Determines if the given item is contained within this specifier.
        """

    @abc.abstractmethod
    def filter(self, iterable, prereleases=None):
        """
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        """


class _IndividualSpecifier(BaseSpecifier):

    _operators = {}

    def __init__(self, spec="", prereleases=None):
        match = self._regex.search(spec)
        if not match:
            raise InvalidSpecifier("Invalid specifier: '{0}'".format(spec))

        self._spec = (match.group("operator").strip(), match.group("version").strip())

        # Store whether or not this Specifier should accept prereleases
        self._prereleases = prereleases

    def __repr__(self):
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "<{0}({1!r}{2})>".format(self.__class__.__name__, str(self), pre)

    def __str__(self):
        return "{0}{1}".format(*self._spec)

    def __hash__(self):
        return hash(self._spec)

    def __eq__(self, other):
        if isinstance(other, string_types):
            try:
                other = self.__class__(other)
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec == other._spec

    def __ne__(self, other):
        if isinstance(other, string_types):
            try:
                other = self.__class__(other)
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec != other._spec

    def _get_operator(self, op):
        return getattr(self, "_compare_{0}".format(self._operators[op]))

    def _coerce_version(self, version):
        if not isinstance(version, (LegacyVersion, Version)):
            version = parse(version)
        return version

    @property
    def operator(self):
        return self._spec[0]

    @property
    def version(self):
        return self._spec[1]

    @property
    def prereleases(self):
        return self._prereleases

    @prereleases.setter
    def prereleases(self, value):
        self._prereleases = value

    def __contains__(self, item):
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # Determine if prereleases are to be allowed or not.
        if prereleases is None:
            prereleases = self.prereleases

        # Normalize item to a Version or LegacyVersion, this allows us to have
        # a shortcut for ``"2.0" in Specifier(">=2")
        item = self._coerce_version(item)

        # Determine if we should be supporting prereleases in this specifier
        # or not, if we do not support prereleases than we can short circuit
        # logic if this version is a prereleases.
        if item.is_prerelease and not prereleases:
            return False

        # Actually do the comparison to determine if this item is contained
        # within this Specifier or not.
        return self._get_operator(self.operator)(item, self.version)

    def filter(self, iterable, prereleases=None):
        yielded = False
        found_prereleases = []

        kw = {"prereleases": prereleases if prereleases is not None else True}

        # Attempt to iterate over all the values in the iterable and if any of
        # them match, yield them.
        for version in iterable:
            parsed_version = self._coerce_version(version)

            if self.contains(parsed_version, **kw):
                # If our version is a prerelease, and we were not set to allow
                # prereleases, then we'll store it for later incase nothing
                # else matches this specifier.
                if parsed_version.is_prerelease and not (
                    prereleases or self.prereleases
                ):
                    found_prereleases.append(version)
                # Either this is not a prerelease, or we should have been
                # accepting prereleases from the beginning.
                else:
                    yielded = True
                    yield version

        # Now that we've iterated over everything, determine if we've yielded
        # any values, and if we have not and we have any prereleases stored up
        # then we will go ahead and yield the prereleases.
        if not yielded and found_prereleases:
            for version in found_prereleases:
                yield version


class LegacySpecifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P(==|!=|<=|>=|<|>))
        \s*
        (?P
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
    }

    def _coerce_version(self, version):
        if not isinstance(version, LegacyVersion):
            version = LegacyVersion(str(version))
        return version

    def _compare_equal(self, prospective, spec):
        return prospective == self._coerce_version(spec)

    def _compare_not_equal(self, prospective, spec):
        return prospective != self._coerce_version(spec)

    def _compare_less_than_equal(self, prospective, spec):
        return prospective <= self._coerce_version(spec)

    def _compare_greater_than_equal(self, prospective, spec):
        return prospective >= self._coerce_version(spec)

    def _compare_less_than(self, prospective, spec):
        return prospective < self._coerce_version(spec)

    def _compare_greater_than(self, prospective, spec):
        return prospective > self._coerce_version(spec)


def _require_version_compare(fn):
    @functools.wraps(fn)
    def wrapped(self, prospective, spec):
        if not isinstance(prospective, Version):
            return False
        return fn(self, prospective, spec)

    return wrapped


class Specifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P(~=|==|!=|<=|>=|<|>|===))
        (?P
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
        "===": "arbitrary",
    }

    @_require_version_compare
    def _compare_compatible(self, prospective, spec):
        # Compatible releases have an equivalent combination of >= and ==. That
        # is that ~=2.2 is equivalent to >=2.2,==2.*. This allows us to
        # implement this in terms of the other specifiers instead of
        # implementing it ourselves. The only thing we need to do is construct
        # the other specifiers.

        # We want everything but the last item in the version, but we want to
        # ignore post and dev releases and we want to treat the pre-release as
        # it's own separate segment.
        prefix = ".".join(
            list(
                itertools.takewhile(
                    lambda x: (not x.startswith("post") and not x.startswith("dev")),
                    _version_split(spec),
                )
            )[:-1]
        )

        # Add the prefix notation to the end of our string
        prefix += ".*"

        return self._get_operator(">=")(prospective, spec) and self._get_operator("==")(
            prospective, prefix
        )

    @_require_version_compare
    def _compare_equal(self, prospective, spec):
        # We need special logic to handle prefix matching
        if spec.endswith(".*"):
            # In the case of prefix matching we want to ignore local segment.
            prospective = Version(prospective.public)
            # Split the spec out by dots, and pretend that there is an implicit
            # dot in between a release segment and a pre-release segment.
            spec = _version_split(spec[:-2])  # Remove the trailing .*

            # Split the prospective version out by dots, and pretend that there
            # is an implicit dot in between a release segment and a pre-release
            # segment.
            prospective = _version_split(str(prospective))

            # Shorten the prospective version to be the same length as the spec
            # so that we can determine if the specifier is a prefix of the
            # prospective version or not.
            prospective = prospective[: len(spec)]

            # Pad out our two sides with zeros so that they both equal the same
            # length.
            spec, prospective = _pad_version(spec, prospective)
        else:
            # Convert our spec string into a Version
            spec = Version(spec)

            # If the specifier does not have a local segment, then we want to
            # act as if the prospective version also does not have a local
            # segment.
            if not spec.local:
                prospective = Version(prospective.public)

        return prospective == spec

    @_require_version_compare
    def _compare_not_equal(self, prospective, spec):
        return not self._compare_equal(prospective, spec)

    @_require_version_compare
    def _compare_less_than_equal(self, prospective, spec):
        return prospective <= Version(spec)

    @_require_version_compare
    def _compare_greater_than_equal(self, prospective, spec):
        return prospective >= Version(spec)

    @_require_version_compare
    def _compare_less_than(self, prospective, spec):
        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec)

        # Check to see if the prospective version is less than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective < spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a pre-release version, that we do not accept pre-release
        # versions for the version mentioned in the specifier (e.g. <3.1 should
        # not match 3.1.dev0, but should match 3.0.dev0).
        if not spec.is_prerelease and prospective.is_prerelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # less than the spec version *and* it's not a pre-release of the same
        # version in the spec.
        return True

    @_require_version_compare
    def _compare_greater_than(self, prospective, spec):
        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec)

        # Check to see if the prospective version is greater than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective > spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a post-release version, that we do not accept
        # post-release versions for the version mentioned in the specifier
        # (e.g. >3.1 should not match 3.0.post0, but should match 3.2.post0).
        if not spec.is_postrelease and prospective.is_postrelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # Ensure that we do not allow a local version of the version mentioned
        # in the specifier, which is technically greater than, to match.
        if prospective.local is not None:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # greater than the spec version *and* it's not a pre-release of the
        # same version in the spec.
        return True

    def _compare_arbitrary(self, prospective, spec):
        return str(prospective).lower() == str(spec).lower()

    @property
    def prereleases(self):
        # If there is an explicit prereleases set for this, then we'll just
        # blindly use that.
        if self._prereleases is not None:
            return self._prereleases

        # Look at all of our specifiers and determine if they are inclusive
        # operators, and if they are if they are including an explicit
        # prerelease.
        operator, version = self._spec
        if operator in ["==", ">=", "<=", "~=", "==="]:
            # The == specifier can include a trailing .*, if it does we
            # want to remove before parsing.
            if operator == "==" and version.endswith(".*"):
                version = version[:-2]

            # Parse the version, and if it is a pre-release than this
            # specifier allows pre-releases.
            if parse(version).is_prerelease:
                return True

        return False

    @prereleases.setter
    def prereleases(self, value):
        self._prereleases = value


_prefix_regex = re.compile(r"^([0-9]+)((?:a|b|c|rc)[0-9]+)$")


def _version_split(version):
    result = []
    for item in version.split("."):
        match = _prefix_regex.search(item)
        if match:
            result.extend(match.groups())
        else:
            result.append(item)
    return result


def _pad_version(left, right):
    left_split, right_split = [], []

    # Get the release segment of our versions
    left_split.append(list(itertools.takewhile(lambda x: x.isdigit(), left)))
    right_split.append(list(itertools.takewhile(lambda x: x.isdigit(), right)))

    # Get the rest of our versions
    left_split.append(left[len(left_split[0]) :])
    right_split.append(right[len(right_split[0]) :])

    # Insert our padding
    left_split.insert(1, ["0"] * max(0, len(right_split[0]) - len(left_split[0])))
    right_split.insert(1, ["0"] * max(0, len(left_split[0]) - len(right_split[0])))

    return (list(itertools.chain(*left_split)), list(itertools.chain(*right_split)))


class SpecifierSet(BaseSpecifier):
    def __init__(self, specifiers="", prereleases=None):
        # Split on , to break each indidivual specifier into it's own item, and
        # strip each item to remove leading/trailing whitespace.
        specifiers = [s.strip() for s in specifiers.split(",") if s.strip()]

        # Parsed each individual specifier, attempting first to make it a
        # Specifier and falling back to a LegacySpecifier.
        parsed = set()
        for specifier in specifiers:
            try:
                parsed.add(Specifier(specifier))
            except InvalidSpecifier:
                parsed.add(LegacySpecifier(specifier))

        # Turn our parsed specifiers into a frozen set and save them for later.
        self._specs = frozenset(parsed)

        # Store our prereleases value so we can use it later to determine if
        # we accept prereleases or not.
        self._prereleases = prereleases

    def __repr__(self):
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "".format(str(self), pre)

    def __str__(self):
        return ",".join(sorted(str(s) for s in self._specs))

    def __hash__(self):
        return hash(self._specs)

    def __and__(self, other):
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        specifier = SpecifierSet()
        specifier._specs = frozenset(self._specs | other._specs)

        if self._prereleases is None and other._prereleases is not None:
            specifier._prereleases = other._prereleases
        elif self._prereleases is not None and other._prereleases is None:
            specifier._prereleases = self._prereleases
        elif self._prereleases == other._prereleases:
            specifier._prereleases = self._prereleases
        else:
            raise ValueError(
                "Cannot combine SpecifierSets with True and False prerelease "
                "overrides."
            )

        return specifier

    def __eq__(self, other):
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif isinstance(other, _IndividualSpecifier):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs == other._specs

    def __ne__(self, other):
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif isinstance(other, _IndividualSpecifier):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs != other._specs

    def __len__(self):
        return len(self._specs)

    def __iter__(self):
        return iter(self._specs)

    @property
    def prereleases(self):
        # If we have been given an explicit prerelease modifier, then we'll
        # pass that through here.
        if self._prereleases is not None:
            return self._prereleases

        # If we don't have any specifiers, and we don't have a forced value,
        # then we'll just return None since we don't know if this should have
        # pre-releases or not.
        if not self._specs:
            return None

        # Otherwise we'll see if any of the given specifiers accept
        # prereleases, if any of them do we'll return True, otherwise False.
        return any(s.prereleases for s in self._specs)

    @prereleases.setter
    def prereleases(self, value):
        self._prereleases = value

    def __contains__(self, item):
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # Ensure that our item is a Version or LegacyVersion instance.
        if not isinstance(item, (LegacyVersion, Version)):
            item = parse(item)

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # We can determine if we're going to allow pre-releases by looking to
        # see if any of the underlying items supports them. If none of them do
        # and this item is a pre-release then we do not allow it and we can
        # short circuit that here.
        # Note: This means that 1.0.dev1 would not be contained in something
        #       like >=1.0.devabc however it would be in >=1.0.debabc,>0.0.dev0
        if not prereleases and item.is_prerelease:
            return False

        # We simply dispatch to the underlying specs here to make sure that the
        # given version is contained within all of them.
        # Note: This use of all() here means that an empty set of specifiers
        #       will always return True, this is an explicit design decision.
        return all(s.contains(item, prereleases=prereleases) for s in self._specs)

    def filter(self, iterable, prereleases=None):
        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # If we have any specifiers, then we want to wrap our iterable in the
        # filter method for each one, this will act as a logical AND amongst
        # each specifier.
        if self._specs:
            for spec in self._specs:
                iterable = spec.filter(iterable, prereleases=bool(prereleases))
            return iterable
        # If we do not have any specifiers, then we need to have a rough filter
        # which will filter out any pre-releases, unless there are no final
        # releases, and which will filter out LegacyVersion in general.
        else:
            filtered = []
            found_prereleases = []

            for item in iterable:
                # Ensure that we some kind of Version class for this item.
                if not isinstance(item, (LegacyVersion, Version)):
                    parsed_version = parse(item)
                else:
                    parsed_version = item

                # Filter out any item which is parsed as a LegacyVersion
                if isinstance(parsed_version, LegacyVersion):
                    continue

                # Store any item which is a pre-release for later unless we've
                # already found a final version or we are accepting prereleases
                if parsed_version.is_prerelease and not prereleases:
                    if not filtered:
                        found_prereleases.append(item)
                else:
                    filtered.append(item)

            # If we've found no items except for pre-releases, then we'll go
            # ahead and use the pre-releases
            if not filtered and found_prereleases and prereleases is None:
                return found_prereleases

            return filtered
site-packages/pip/_vendor/packaging/markers.py000064400000020072151464452710015464 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import operator
import os
import platform
import sys

from pip._vendor.pyparsing import ParseException, ParseResults, stringStart, stringEnd
from pip._vendor.pyparsing import ZeroOrMore, Group, Forward, QuotedString
from pip._vendor.pyparsing import Literal as L  # noqa

from ._compat import string_types
from .specifiers import Specifier, InvalidSpecifier


__all__ = [
    "InvalidMarker",
    "UndefinedComparison",
    "UndefinedEnvironmentName",
    "Marker",
    "default_environment",
]


class InvalidMarker(ValueError):
    """
    An invalid marker was found, users should refer to PEP 508.
    """


class UndefinedComparison(ValueError):
    """
    An invalid operation was attempted on a value that doesn't support it.
    """


class UndefinedEnvironmentName(ValueError):
    """
    A name was attempted to be used that does not exist inside of the
    environment.
    """


class Node(object):
    def __init__(self, value):
        self.value = value

    def __str__(self):
        return str(self.value)

    def __repr__(self):
        return "<{0}({1!r})>".format(self.__class__.__name__, str(self))

    def serialize(self):
        raise NotImplementedError


class Variable(Node):
    def serialize(self):
        return str(self)


class Value(Node):
    def serialize(self):
        return '"{0}"'.format(self)


class Op(Node):
    def serialize(self):
        return str(self)


VARIABLE = (
    L("implementation_version")
    | L("platform_python_implementation")
    | L("implementation_name")
    | L("python_full_version")
    | L("platform_release")
    | L("platform_version")
    | L("platform_machine")
    | L("platform_system")
    | L("python_version")
    | L("sys_platform")
    | L("os_name")
    | L("os.name")
    | L("sys.platform")  # PEP-345
    | L("platform.version")  # PEP-345
    | L("platform.machine")  # PEP-345
    | L("platform.python_implementation")  # PEP-345
    | L("python_implementation")  # PEP-345
    | L("extra")  # undocumented setuptools legacy
)
ALIASES = {
    "os.name": "os_name",
    "sys.platform": "sys_platform",
    "platform.version": "platform_version",
    "platform.machine": "platform_machine",
    "platform.python_implementation": "platform_python_implementation",
    "python_implementation": "platform_python_implementation",
}
VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0])))

VERSION_CMP = (
    L("===") | L("==") | L(">=") | L("<=") | L("!=") | L("~=") | L(">") | L("<")
)

MARKER_OP = VERSION_CMP | L("not in") | L("in")
MARKER_OP.setParseAction(lambda s, l, t: Op(t[0]))

MARKER_VALUE = QuotedString("'") | QuotedString('"')
MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0]))

BOOLOP = L("and") | L("or")

MARKER_VAR = VARIABLE | MARKER_VALUE

MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR)
MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0]))

LPAREN = L("(").suppress()
RPAREN = L(")").suppress()

MARKER_EXPR = Forward()
MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN)
MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR)

MARKER = stringStart + MARKER_EXPR + stringEnd


def _coerce_parse_result(results):
    if isinstance(results, ParseResults):
        return [_coerce_parse_result(i) for i in results]
    else:
        return results


def _format_marker(marker, first=True):
    assert isinstance(marker, (list, tuple, string_types))

    # Sometimes we have a structure like [[...]] which is a single item list
    # where the single item is itself it's own list. In that case we want skip
    # the rest of this function so that we don't get extraneous () on the
    # outside.
    if (
        isinstance(marker, list)
        and len(marker) == 1
        and isinstance(marker[0], (list, tuple))
    ):
        return _format_marker(marker[0])

    if isinstance(marker, list):
        inner = (_format_marker(m, first=False) for m in marker)
        if first:
            return " ".join(inner)
        else:
            return "(" + " ".join(inner) + ")"
    elif isinstance(marker, tuple):
        return " ".join([m.serialize() for m in marker])
    else:
        return marker


_operators = {
    "in": lambda lhs, rhs: lhs in rhs,
    "not in": lambda lhs, rhs: lhs not in rhs,
    "<": operator.lt,
    "<=": operator.le,
    "==": operator.eq,
    "!=": operator.ne,
    ">=": operator.ge,
    ">": operator.gt,
}


def _eval_op(lhs, op, rhs):
    try:
        spec = Specifier("".join([op.serialize(), rhs]))
    except InvalidSpecifier:
        pass
    else:
        return spec.contains(lhs)

    oper = _operators.get(op.serialize())
    if oper is None:
        raise UndefinedComparison(
            "Undefined {0!r} on {1!r} and {2!r}.".format(op, lhs, rhs)
        )

    return oper(lhs, rhs)


_undefined = object()


def _get_env(environment, name):
    value = environment.get(name, _undefined)

    if value is _undefined:
        raise UndefinedEnvironmentName(
            "{0!r} does not exist in evaluation environment.".format(name)
        )

    return value


def _evaluate_markers(markers, environment):
    groups = [[]]

    for marker in markers:
        assert isinstance(marker, (list, tuple, string_types))

        if isinstance(marker, list):
            groups[-1].append(_evaluate_markers(marker, environment))
        elif isinstance(marker, tuple):
            lhs, op, rhs = marker

            if isinstance(lhs, Variable):
                lhs_value = _get_env(environment, lhs.value)
                rhs_value = rhs.value
            else:
                lhs_value = lhs.value
                rhs_value = _get_env(environment, rhs.value)

            groups[-1].append(_eval_op(lhs_value, op, rhs_value))
        else:
            assert marker in ["and", "or"]
            if marker == "or":
                groups.append([])

    return any(all(item) for item in groups)


def format_full_version(info):
    version = "{0.major}.{0.minor}.{0.micro}".format(info)
    kind = info.releaselevel
    if kind != "final":
        version += kind[0] + str(info.serial)
    return version


def default_environment():
    if hasattr(sys, "implementation"):
        iver = format_full_version(sys.implementation.version)
        implementation_name = sys.implementation.name
    else:
        iver = "0"
        implementation_name = ""

    return {
        "implementation_name": implementation_name,
        "implementation_version": iver,
        "os_name": os.name,
        "platform_machine": platform.machine(),
        "platform_release": platform.release(),
        "platform_system": platform.system(),
        "platform_version": platform.version(),
        "python_full_version": platform.python_version(),
        "platform_python_implementation": platform.python_implementation(),
        "python_version": ".".join(platform.python_version_tuple()[:2]),
        "sys_platform": sys.platform,
    }


class Marker(object):
    def __init__(self, marker):
        try:
            self._markers = _coerce_parse_result(MARKER.parseString(marker))
        except ParseException as e:
            err_str = "Invalid marker: {0!r}, parse error at {1!r}".format(
                marker, marker[e.loc : e.loc + 8]
            )
            raise InvalidMarker(err_str)

    def __str__(self):
        return _format_marker(self._markers)

    def __repr__(self):
        return "".format(str(self))

    def evaluate(self, environment=None):
        """Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        """
        current_environment = default_environment()
        if environment is not None:
            current_environment.update(environment)

        return _evaluate_markers(self._markers, current_environment)
site-packages/pip/_vendor/packaging/__init__.py000064400000001062151464452710015555 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

from .__about__ import (
    __author__,
    __copyright__,
    __email__,
    __license__,
    __summary__,
    __title__,
    __uri__,
    __version__,
)

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]
site-packages/pip/_vendor/pytoml/utils.py000064400000003201151464452710014553 0ustar00import datetime
import re

rfc3339_re = re.compile(r'(\d{4})-(\d{2})-(\d{2})T(\d{2}):(\d{2}):(\d{2})(\.\d+)?(?:Z|([+-]\d{2}):(\d{2}))')

def parse_rfc3339(v):
    m = rfc3339_re.match(v)
    if not m or m.group(0) != v:
        return None
    return parse_rfc3339_re(m)

def parse_rfc3339_re(m):
    r = map(int, m.groups()[:6])
    if m.group(7):
        micro = float(m.group(7))
    else:
        micro = 0

    if m.group(8):
        g = int(m.group(8), 10) * 60 + int(m.group(9), 10)
        tz = _TimeZone(datetime.timedelta(0, g * 60))
    else:
        tz = _TimeZone(datetime.timedelta(0, 0))

    y, m, d, H, M, S = r
    return datetime.datetime(y, m, d, H, M, S, int(micro * 1000000), tz)


def format_rfc3339(v):
    offs = v.utcoffset()
    offs = int(offs.total_seconds()) // 60 if offs is not None else 0

    if offs == 0:
        suffix = 'Z'
    else:
        if offs > 0:
            suffix = '+'
        else:
            suffix = '-'
            offs = -offs
        suffix = '{0}{1:02}:{2:02}'.format(suffix, offs // 60, offs % 60)

    if v.microsecond:
        return v.strftime('%Y-%m-%dT%H:%M:%S.%f') + suffix
    else:
        return v.strftime('%Y-%m-%dT%H:%M:%S') + suffix

class _TimeZone(datetime.tzinfo):
    def __init__(self, offset):
        self._offset = offset

    def utcoffset(self, dt):
        return self._offset

    def dst(self, dt):
        return None

    def tzname(self, dt):
        m = self._offset.total_seconds() // 60
        if m < 0:
            res = '-'
            m = -m
        else:
            res = '+'
        h = m // 60
        m = m - h * 60
        return '{}{:.02}{:.02}'.format(res, h, m)
site-packages/pip/_vendor/pytoml/parser.py000064400000024105151464452710014715 0ustar00import re, sys
from .core import TomlError
from .utils import rfc3339_re, parse_rfc3339_re

if sys.version_info[0] == 2:
    _chr = unichr
else:
    _chr = chr

def load(fin, translate=lambda t, x, v: v, object_pairs_hook=dict):
    return loads(fin.read(), translate=translate, object_pairs_hook=object_pairs_hook, filename=getattr(fin, 'name', repr(fin)))

def loads(s, filename='', translate=lambda t, x, v: v, object_pairs_hook=dict):
    if isinstance(s, bytes):
        s = s.decode('utf-8')

    s = s.replace('\r\n', '\n')

    root = object_pairs_hook()
    tables = object_pairs_hook()
    scope = root

    src = _Source(s, filename=filename)
    ast = _p_toml(src, object_pairs_hook=object_pairs_hook)

    def error(msg):
        raise TomlError(msg, pos[0], pos[1], filename)

    def process_value(v, object_pairs_hook):
        kind, text, value, pos = v
        if kind == 'array':
            if value and any(k != value[0][0] for k, t, v, p in value[1:]):
                error('array-type-mismatch')
            value = [process_value(item, object_pairs_hook=object_pairs_hook) for item in value]
        elif kind == 'table':
            value = object_pairs_hook([(k, process_value(value[k], object_pairs_hook=object_pairs_hook)) for k in value])
        return translate(kind, text, value)

    for kind, value, pos in ast:
        if kind == 'kv':
            k, v = value
            if k in scope:
                error('duplicate_keys. Key "{0}" was used more than once.'.format(k))
            scope[k] = process_value(v, object_pairs_hook=object_pairs_hook)
        else:
            is_table_array = (kind == 'table_array')
            cur = tables
            for name in value[:-1]:
                if isinstance(cur.get(name), list):
                    d, cur = cur[name][-1]
                else:
                    d, cur = cur.setdefault(name, (None, object_pairs_hook()))

            scope = object_pairs_hook()
            name = value[-1]
            if name not in cur:
                if is_table_array:
                    cur[name] = [(scope, object_pairs_hook())]
                else:
                    cur[name] = (scope, object_pairs_hook())
            elif isinstance(cur[name], list):
                if not is_table_array:
                    error('table_type_mismatch')
                cur[name].append((scope, object_pairs_hook()))
            else:
                if is_table_array:
                    error('table_type_mismatch')
                old_scope, next_table = cur[name]
                if old_scope is not None:
                    error('duplicate_tables')
                cur[name] = (scope, next_table)

    def merge_tables(scope, tables):
        if scope is None:
            scope = object_pairs_hook()
        for k in tables:
            if k in scope:
                error('key_table_conflict')
            v = tables[k]
            if isinstance(v, list):
                scope[k] = [merge_tables(sc, tbl) for sc, tbl in v]
            else:
                scope[k] = merge_tables(v[0], v[1])
        return scope

    return merge_tables(root, tables)

class _Source:
    def __init__(self, s, filename=None):
        self.s = s
        self._pos = (1, 1)
        self._last = None
        self._filename = filename
        self.backtrack_stack = []

    def last(self):
        return self._last

    def pos(self):
        return self._pos

    def fail(self):
        return self._expect(None)

    def consume_dot(self):
        if self.s:
            self._last = self.s[0]
            self.s = self[1:]
            self._advance(self._last)
            return self._last
        return None

    def expect_dot(self):
        return self._expect(self.consume_dot())

    def consume_eof(self):
        if not self.s:
            self._last = ''
            return True
        return False

    def expect_eof(self):
        return self._expect(self.consume_eof())

    def consume(self, s):
        if self.s.startswith(s):
            self.s = self.s[len(s):]
            self._last = s
            self._advance(s)
            return True
        return False

    def expect(self, s):
        return self._expect(self.consume(s))

    def consume_re(self, re):
        m = re.match(self.s)
        if m:
            self.s = self.s[len(m.group(0)):]
            self._last = m
            self._advance(m.group(0))
            return m
        return None

    def expect_re(self, re):
        return self._expect(self.consume_re(re))

    def __enter__(self):
        self.backtrack_stack.append((self.s, self._pos))

    def __exit__(self, type, value, traceback):
        if type is None:
            self.backtrack_stack.pop()
        else:
            self.s, self._pos = self.backtrack_stack.pop()
        return type == TomlError

    def commit(self):
        self.backtrack_stack[-1] = (self.s, self._pos)

    def _expect(self, r):
        if not r:
            raise TomlError('msg', self._pos[0], self._pos[1], self._filename)
        return r

    def _advance(self, s):
        suffix_pos = s.rfind('\n')
        if suffix_pos == -1:
            self._pos = (self._pos[0], self._pos[1] + len(s))
        else:
            self._pos = (self._pos[0] + s.count('\n'), len(s) - suffix_pos)

_ews_re = re.compile(r'(?:[ \t]|#[^\n]*\n|#[^\n]*\Z|\n)*')
def _p_ews(s):
    s.expect_re(_ews_re)

_ws_re = re.compile(r'[ \t]*')
def _p_ws(s):
    s.expect_re(_ws_re)

_escapes = { 'b': '\b', 'n': '\n', 'r': '\r', 't': '\t', '"': '"',
    '\\': '\\', 'f': '\f' }

_basicstr_re = re.compile(r'[^"\\\000-\037]*')
_short_uni_re = re.compile(r'u([0-9a-fA-F]{4})')
_long_uni_re = re.compile(r'U([0-9a-fA-F]{8})')
_escapes_re = re.compile(r'[btnfr\"\\]')
_newline_esc_re = re.compile('\n[ \t\n]*')
def _p_basicstr_content(s, content=_basicstr_re):
    res = []
    while True:
        res.append(s.expect_re(content).group(0))
        if not s.consume('\\'):
            break
        if s.consume_re(_newline_esc_re):
            pass
        elif s.consume_re(_short_uni_re) or s.consume_re(_long_uni_re):
            v = int(s.last().group(1), 16)
            if 0xd800 <= v < 0xe000:
                s.fail()
            res.append(_chr(v))
        else:
            s.expect_re(_escapes_re)
            res.append(_escapes[s.last().group(0)])
    return ''.join(res)

_key_re = re.compile(r'[0-9a-zA-Z-_]+')
def _p_key(s):
    with s:
        s.expect('"')
        r = _p_basicstr_content(s, _basicstr_re)
        s.expect('"')
        return r
    if s.consume('\''):
        if s.consume('\'\''):
            s.consume('\n')
            r = s.expect_re(_litstr_ml_re).group(0)
            s.expect('\'\'\'')
        else:
            r = s.expect_re(_litstr_re).group(0)
            s.expect('\'')
        return r
    return s.expect_re(_key_re).group(0)

_float_re = re.compile(r'[+-]?(?:0|[1-9](?:_?\d)*)(?:\.\d(?:_?\d)*)?(?:[eE][+-]?(?:\d(?:_?\d)*))?')

_basicstr_ml_re = re.compile(r'(?:""?(?!")|[^"\\\000-\011\013-\037])*')
_litstr_re = re.compile(r"[^'\000\010\012-\037]*")
_litstr_ml_re = re.compile(r"(?:(?:|'|'')(?:[^'\000-\010\013-\037]))*")
def _p_value(s, object_pairs_hook):
    pos = s.pos()

    if s.consume('true'):
        return 'bool', s.last(), True, pos
    if s.consume('false'):
        return 'bool', s.last(), False, pos

    if s.consume('"'):
        if s.consume('""'):
            s.consume('\n')
            r = _p_basicstr_content(s, _basicstr_ml_re)
            s.expect('"""')
        else:
            r = _p_basicstr_content(s, _basicstr_re)
            s.expect('"')
        return 'str', r, r, pos

    if s.consume('\''):
        if s.consume('\'\''):
            s.consume('\n')
            r = s.expect_re(_litstr_ml_re).group(0)
            s.expect('\'\'\'')
        else:
            r = s.expect_re(_litstr_re).group(0)
            s.expect('\'')
        return 'str', r, r, pos

    if s.consume_re(rfc3339_re):
        m = s.last()
        return 'datetime', m.group(0), parse_rfc3339_re(m), pos

    if s.consume_re(_float_re):
        m = s.last().group(0)
        r = m.replace('_','')
        if '.' in m or 'e' in m or 'E' in m:
            return 'float', m, float(r), pos
        else:
            return 'int', m, int(r, 10), pos

    if s.consume('['):
        items = []
        with s:
            while True:
                _p_ews(s)
                items.append(_p_value(s, object_pairs_hook=object_pairs_hook))
                s.commit()
                _p_ews(s)
                s.expect(',')
                s.commit()
        _p_ews(s)
        s.expect(']')
        return 'array', None, items, pos

    if s.consume('{'):
        _p_ws(s)
        items = object_pairs_hook()
        if not s.consume('}'):
            k = _p_key(s)
            _p_ws(s)
            s.expect('=')
            _p_ws(s)
            items[k] = _p_value(s, object_pairs_hook=object_pairs_hook)
            _p_ws(s)
            while s.consume(','):
                _p_ws(s)
                k = _p_key(s)
                _p_ws(s)
                s.expect('=')
                _p_ws(s)
                items[k] = _p_value(s, object_pairs_hook=object_pairs_hook)
                _p_ws(s)
            s.expect('}')
        return 'table', None, items, pos

    s.fail()

def _p_stmt(s, object_pairs_hook):
    pos = s.pos()
    if s.consume(   '['):
        is_array = s.consume('[')
        _p_ws(s)
        keys = [_p_key(s)]
        _p_ws(s)
        while s.consume('.'):
            _p_ws(s)
            keys.append(_p_key(s))
            _p_ws(s)
        s.expect(']')
        if is_array:
            s.expect(']')
        return 'table_array' if is_array else 'table', keys, pos

    key = _p_key(s)
    _p_ws(s)
    s.expect('=')
    _p_ws(s)
    value = _p_value(s, object_pairs_hook=object_pairs_hook)
    return 'kv', (key, value), pos

_stmtsep_re = re.compile(r'(?:[ \t]*(?:#[^\n]*)?\n)+[ \t]*')
def _p_toml(s, object_pairs_hook):
    stmts = []
    _p_ews(s)
    with s:
        stmts.append(_p_stmt(s, object_pairs_hook=object_pairs_hook))
        while True:
            s.commit()
            s.expect_re(_stmtsep_re)
            stmts.append(_p_stmt(s, object_pairs_hook=object_pairs_hook))
    _p_ews(s)
    s.expect_eof()
    return stmts
site-packages/pip/_vendor/pytoml/test.py000064400000001775151464452710014410 0ustar00import datetime
from .utils import format_rfc3339

try:
    _string_types = (str, unicode)
    _int_types = (int, long)
except NameError:
    _string_types = str
    _int_types = int

def translate_to_test(v):
    if isinstance(v, dict):
        return { k: translate_to_test(v) for k, v in v.items() }
    if isinstance(v, list):
        a = [translate_to_test(x) for x in v]
        if v and isinstance(v[0], dict):
            return a
        else:
            return {'type': 'array', 'value': a}
    if isinstance(v, datetime.datetime):
        return {'type': 'datetime', 'value': format_rfc3339(v)}
    if isinstance(v, bool):
        return {'type': 'bool', 'value': 'true' if v else 'false'}
    if isinstance(v, _int_types):
        return {'type': 'integer', 'value': str(v)}
    if isinstance(v, float):
        return {'type': 'float', 'value': '{:.17}'.format(v)}
    if isinstance(v, _string_types):
        return {'type': 'string', 'value': v}
    raise RuntimeError('unexpected value: {!r}'.format(v))
site-packages/pip/_vendor/pytoml/__pycache__/test.cpython-38.opt-1.pyc000064400000002303151464452710021621 0ustar00U

.e@sRddlZddlmZzeefZeefZWne	k
rDeZeZYnXddZ
dS)N)format_rfc3339cCst|trdd|DSt|trTdd|D}|rJt|dtrJ|Sd|dSt|tjrndt|dSt|trd	|rd
nddSt|trdt|dSt|t	rd
d
|dSt|trd|dStd
|dS)NcSsi|]\}}|t|qStranslate_to_test).0kvrr;/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/test.py

sz%translate_to_test..cSsg|]}t|qSrr)rxrrr

sz%translate_to_test..rZarray)typevaluedatetimebooltrueZfalseZintegerfloatz{:.17}stringzunexpected value: {!r})
isinstancedictitemslistrrr
_int_typesstrrformat
_string_typesRuntimeError)r	arrr
rs$







r)rZutilsrrZunicoderintZlongr	NameErrorrrrrr
s
site-packages/pip/_vendor/pytoml/__pycache__/core.cpython-38.pyc000064400000001625151464452710020641 0ustar00U

.e@sGdddeZdS)c@s$eZdZddZddZddZdS)	TomlErrorcCs.t|||||||_||_||_||_dS)N)RuntimeError__init__messagelinecolfilename)selfrrrrr	;/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/core.pyrs
zTomlError.__init__cCsd|j|j|j|jS)Nz{}({}, {}): {})formatrrrrrr	r	r
__str__	szTomlError.__str__cCsd|j|j|j|jS)Nz!TomlError({!r}, {!r}, {!r}, {!r}))rrrrrrr	r	r
__repr__szTomlError.__repr__N)__name__
__module____qualname__rr
rr	r	r	r
rsrN)rrr	r	r	r
site-packages/pip/_vendor/pytoml/__pycache__/writer.cpython-38.opt-1.pyc000064400000007171151464452710022166 0ustar00U

.e)
@sddlmZddlZddlZddlZddlZddlZddlmZzddl	m
ZWnek
rhdZYnXej
ddkreZeZdd	d
Zddd
dddddZddZejejdZddZddZdddZdS))unicode_literalsN)format_rfc3339)PurePathFcCst}t|||d|S)N)	sort_keys)ioStringIOdumpgetvalue)objrfoutrr=/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/writer.pydumpssrnr\tbf")

r	rcsgdfdd}dtkr}|dkrT|dt|n$t|dkrx|dt|d7q|d	d
d	S)Nrcs"krdS)Nr)appendriressstartrrflushsz_escape_string..flushz"\

	r z\u%04xrr)lenr_escapesordjoin)r!r#crrr_escape_strings
r+z-_cCstdd|Drt|S|S)Ncss|]}|tkVqdSN)
_key_chars.0r*rrr	4sz_escape_id..)anyr+)r!rrr
_escape_id3sr2cCs
t|tr|rdSdSt|ts*t|tr2t|St|trjt|sPt|r`t	d
|nt|Snt|ts~t|trt
|St|tjrt|St|trd
ddd|DSt|trd
dd	d|DSt|trt
t|St|dS)
NtrueZfalsez{0} is not a valid TOML valuez[{0}]z, css|]}t|VqdSr,)
_format_value)r/r
rrrr0Hsz _format_value..z{{{0}}}css&|]\}}dt|t|VqdS)z{} = {}N)formatr2r4)r/kr
rrrr0Js)
isinstanceboolintlongunicodefloatmathZisnanZisinf
ValueErrorr5reprbytesr+datetimerlistr)dictitems_path_typesstrRuntimeError)vrrrr49s&




r4cs\d|dfg}|rX|\}}r^dddD}|rN|d|n|d||rnt|n|}g}d}	|D]|}
t|
tr|f|
dfqt|
t	r|
rt
dd|
Dr|fd	d|
Dq|
dkr|d
td}	q|dtt
|
d}	q|t|sH|	r|r|d
qdS)NrF.css|]}t|VqdSr,)r2r.rrrr0Wszdump..z[[{0}]]
z[{0}]
css|]}t|tVqdSr,)r7rC)r/orrrr0dsc3s|]}f|dfVqdS)TNr)r/dr6namerrr0esz<#{} = null  # To use: uncomment and replace null with value
Tz
{0} = {1}
r)popr)writer5sortedkeysr7rCrrBallextendr2r4reversed)r
rrZtablestableZis_arrayZsection_nameZ
table_keysZ
new_tablesZhas_kvrHrrLrrQs6
 
r)F)F)Z
__future__rr	rAr=stringsysZutilsrpathlibrrEImportErrorversion_infor9r:rFr;rr'r+ZdigitsZ
ascii_lettersr-r2r4rrrrrs (

site-packages/pip/_vendor/pytoml/__pycache__/parser.cpython-38.pyc000064400000023542151464452710021207 0ustar00U

.eE(@sNddlZddlZddlmZddlmZmZejddkr@eZ	ne
Z	ddefdd	Zd
ddefdd
Z
GdddZedZddZedZddZddddddddZedZedZed Zed!Zed"Zefd#d$Zed%Zd&d'Zed(Zed)Zed*Zed+Z d,d-Z!d.d/Z"ed0Z#d1d2Z$dS)3N	TomlError)
rfc3339_reparse_rfc3339_recCs|SNtxvr	r	=/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/parser.py
rc	Cs t|||t|dt|dS)Nname)	translateobject_pairs_hookfilename)loadsreadgetattrrepr)Zfinrrr	r	rload
srzcCs|Srr	r
r	r	rr
rcst|tr|d}|dd}}}|}t|d}t|d}fddfdd	|D]D\}	}
|	d
kr|
\}}||krd||d||<qj|	dk}
|}|
dd
D]<}t||tr||d
\}}q|	|df\}}qƈ}|
d
}||krD|
r4|fg||<n|f||<qjt||trx|
sbd||
|fqj|
rd||\}}|dk	rd||f||<qjfdd||S)Nzutf-8z

)rrcst|dddSNrrr)msg)rposr	rerrorszloads..errorcs|\}}}|dkrTr>tfddddDr>dfddDn"|dkrvfd	dD||S)
Narrayc3s&|]\}}}}|ddkVqdS)rNr	).0krr
p)valuer	r	 s
z/loads..process_value..rzarray-type-mismatchcsg|]}|dqSrr	)r!item)r
process_valuer	r
"sz0loads..process_value..tablecs g|]}||dfqSr&r	)r!r")rr(r$r	rr)$s)any)r
rkindtextr)rr(r)rr$rr(s"zloads..process_valuekvz2duplicate_keys. Key "{0}" was used more than once.table_arrayZtable_type_mismatchZduplicate_tablescsl|dkr}|D]T}||kr&d||}t|trPfdd|D||<q|d|d||<q|S)NZkey_table_conflictcsg|]\}}||qSr	r	)r!ZscZtbl)merge_tablesr	rr)Qsz/loads..merge_tables..rr)
isinstancelist)scopetablesr"r
)rr1rr	rr1Is
zloads..merge_tables)r2bytesdecodereplace_Source_p_tomlformatgetr3
setdefaultappend)srrrrootr5r4srcZastr,r$r"r
Zis_table_arrayZcurrdZ	old_scopeZ
next_tabler	)rrr1rrr(rrr
sP





rc@seZdZd$ddZddZddZdd	Zd
dZdd
ZddZ	ddZ
ddZddZddZ
ddZddZddZddZd d!Zd"d#ZdS)%r9NcCs"||_d|_d|_||_g|_dS)N)rr)r?_pos_last	_filenamebacktrack_stack)selfr?rr	r	r__init__Ys
z_Source.__init__cCs|jSr)rDrGr	r	rlast`sz_Source.lastcCs|jSr)rCrIr	r	rrcsz_Source.poscCs
|dSr)_expectrIr	r	rfailfsz_Source.failcCs6|jr2|jd|_|dd|_||j|jSdSr)r?rD_advancerIr	r	rconsume_dotisz_Source.consume_dotcCs||Sr)rKrNrIr	r	r
expect_dotqsz_Source.expect_dotcCs|jsd|_dSdS)NTF)r?rDrIr	r	rconsume_eoftsz_Source.consume_eofcCs||Sr)rKrQrIr	r	r
expect_eofzsz_Source.expect_eofcCs8|j|r4|jt|d|_||_||dSdS)NTF)r?
startswithlenrDrMrGr?r	r	rconsume}s
z_Source.consumecCs|||Sr)rKrVrUr	r	rexpectsz_Source.expectcCsH||j}|rD|jt|dd|_||_||d|SdS)Nr)matchr?rTgrouprDrM)rGremr	r	r
consume_resz_Source.consume_recCs|||Sr)rKr\)rGrZr	r	r	expect_resz_Source.expect_recCs|j|j|jfdSr)rFr>r?rCrIr	r	r	__enter__sz_Source.__enter__cCs.|dkr|jn|j\|_|_|tkSr)rFpopr?rCr)rGtyper$	tracebackr	r	r__exit__sz_Source.__exit__cCs|j|jf|jd<dS)Nr0)r?rCrFrIr	r	rcommitsz_Source.commitcCs$|s td|jd|jd|j|S)Nrrr)rrCrE)rGrr	r	rrKsz_Source._expectcCsX|d}|dkr2|jd|jdt|f|_n"|jd|dt||f|_dS)Nrr0rr)rfindrCrTcount)rGr?Z
suffix_posr	r	rrMs
 z_Source._advance)N)__name__
__module____qualname__rHrJrrLrNrOrQrRrVrWr\r]r^rbrcrKrMr	r	r	rr9Xs"
	r9z!(?:[ \t]|#[^\n]*\n|#[^\n]*\Z|\n)*cCs|tdSr)r]_ews_rer?r	r	r_p_ewssrlz[ \t]*cCs|tdSr)r]_ws_rerkr	r	r_p_wssrnr
	"\)bnrdrrrrsfz[^"\\\000-\037]*zu([0-9a-fA-F]{4})zU([0-9a-fA-F]{8})z[btnfr\"\\]z
[ 	
]*cCsg}|||d|ds&q|tr2q|tsF|trt|	dd}d|krndkrznn|
|t|q|t|t
|	dqd|S)NrrsriirP)r>r]rYrVr\_newline_esc_re
_short_uni_re_long_uni_reintrJrL_chr_escapes_re_escapesjoin)r?Zcontentresr
r	r	r_p_basicstr_contents


rz[0-9a-zA-Z-_]+c
Cs|0|dt|t}|d|W5QRSQRX|dr|drt|d|td}|dn|td}|d|S|tdS)Nrr'''rr''')	rWr_basicstr_rerVr]
_litstr_ml_rerY
_litstr_re_key_re)r?rdr	r	r_p_keys






rzH[+-]?(?:0|[1-9](?:_?\d)*)(?:\.\d(?:_?\d)*)?(?:[eE][+-]?(?:\d(?:_?\d)*))?z&(?:""?(?!")|[^"\\\000-\011\013-\037])*z[^'\000\010\012-\037]*z((?:(?:|'|'')(?:[^'\000-\010\013-\037]))*c	Cs|}|dr"d|d|fS|dr_p_valuercrnrrL)r?rrrdr[itemsr"r	r	rrs~
















rcCs|}|dr|d}t|t|g}t||dr`t||t|t|q6|d|rx|d|rdnd||fSt|}t||dt|t||d}d||f|fS)	Nrrrr/r*rrr.)rrVrnrr>rWr)r?rrZis_arraykeyskeyr$r	r	r_p_stmt3s(







rz(?:[ \t]*(?:#[^\n]*)?\n)+[ \t]*c	Cshg}t||>|t||d||t|t||dq$W5QRXt|||S)Nr)rlr>rrcr]_stmtsep_rerR)r?rZstmtsr	r	rr:Ks
r:)%rZsysZcorerZutilsrrversion_infoZunichrr}chrdictrrr9compilerjrlrmrnrrrzr{r~ryrrrrrrrrrrr:r	r	r	rs@KU












M
site-packages/pip/_vendor/pytoml/__pycache__/__init__.cpython-38.pyc000064400000000524151464452710021445 0ustar00U

.e@s<ddlmZddlmZmZddlmZddlmZm	Z	dS))	TomlError)loadloads)translate_to_test)dumpdumpsN)
ZcorerparserrrZtestrwriterrrr
r
?/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/__init__.pyssite-packages/pip/_vendor/pytoml/__pycache__/core.cpython-38.opt-1.pyc000064400000001625151464452710021600 0ustar00U

.e@sGdddeZdS)c@s$eZdZddZddZddZdS)	TomlErrorcCs.t|||||||_||_||_||_dS)N)RuntimeError__init__messagelinecolfilename)selfrrrrr	;/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/core.pyrs
zTomlError.__init__cCsd|j|j|j|jS)Nz{}({}, {}): {})formatrrrrrr	r	r
__str__	szTomlError.__str__cCsd|j|j|j|jS)Nz!TomlError({!r}, {!r}, {!r}, {!r}))rrrrrrr	r	r
__repr__szTomlError.__repr__N)__name__
__module____qualname__rr
rr	r	r	r
rsrN)rrr	r	r	r
site-packages/pip/_vendor/pytoml/__pycache__/writer.cpython-38.pyc000064400000007171151464452710021227 0ustar00U

.e)
@sddlmZddlZddlZddlZddlZddlZddlmZzddl	m
ZWnek
rhdZYnXej
ddkreZeZdd	d
Zddd
dddddZddZejejdZddZddZdddZdS))unicode_literalsN)format_rfc3339)PurePathFcCst}t|||d|S)N)	sort_keys)ioStringIOdumpgetvalue)objrfoutrr=/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/writer.pydumpssrnr\tbf")

r	rcsgdfdd}dtkr}|dkrT|dt|n$t|dkrx|dt|d7q|d	d
d	S)Nrcs"krdS)Nr)appendriressstartrrflushsz_escape_string..flushz"\

	r z\u%04xrr)lenr_escapesordjoin)r!r#crrr_escape_strings
r+z-_cCstdd|Drt|S|S)Ncss|]}|tkVqdSN)
_key_chars.0r*rrr	4sz_escape_id..)anyr+)r!rrr
_escape_id3sr2cCs
t|tr|rdSdSt|ts*t|tr2t|St|trjt|sPt|r`t	d
|nt|Snt|ts~t|trt
|St|tjrt|St|trd
ddd|DSt|trd
dd	d|DSt|trt
t|St|dS)
NtrueZfalsez{0} is not a valid TOML valuez[{0}]z, css|]}t|VqdSr,)
_format_value)r/r
rrrr0Hsz _format_value..z{{{0}}}css&|]\}}dt|t|VqdS)z{} = {}N)formatr2r4)r/kr
rrrr0Js)
isinstanceboolintlongunicodefloatmathZisnanZisinf
ValueErrorr5reprbytesr+datetimerlistr)dictitems_path_typesstrRuntimeError)vrrrr49s&




r4cs\d|dfg}|rX|\}}r^dddD}|rN|d|n|d||rnt|n|}g}d}	|D]|}
t|
tr|f|
dfqt|
t	r|
rt
dd|
Dr|fd	d|
Dq|
dkr|d
td}	q|dtt
|
d}	q|t|sH|	r|r|d
qdS)NrF.css|]}t|VqdSr,)r2r.rrrr0Wszdump..z[[{0}]]
z[{0}]
css|]}t|tVqdSr,)r7rC)r/orrrr0dsc3s|]}f|dfVqdS)TNr)r/dr6namerrr0esz<#{} = null  # To use: uncomment and replace null with value
Tz
{0} = {1}
r)popr)writer5sortedkeysr7rCrrBallextendr2r4reversed)r
rrZtablestableZis_arrayZsection_nameZ
table_keysZ
new_tablesZhas_kvrHrrLrrQs6
 
r)F)F)Z
__future__rr	rAr=stringsysZutilsrpathlibrrEImportErrorversion_infor9r:rFr;rr'r+ZdigitsZ
ascii_lettersr-r2r4rrrrrs (

site-packages/pip/_vendor/pytoml/__pycache__/test.cpython-38.pyc000064400000002303151464452710020662 0ustar00U

.e@sRddlZddlmZzeefZeefZWne	k
rDeZeZYnXddZ
dS)N)format_rfc3339cCst|trdd|DSt|trTdd|D}|rJt|dtrJ|Sd|dSt|tjrndt|dSt|trd	|rd
nddSt|trdt|dSt|t	rd
d
|dSt|trd|dStd
|dS)NcSsi|]\}}|t|qStranslate_to_test).0kvrr;/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/test.py

sz%translate_to_test..cSsg|]}t|qSrr)rxrrr

sz%translate_to_test..rZarray)typevaluedatetimebooltrueZfalseZintegerfloatz{:.17}stringzunexpected value: {!r})
isinstancedictitemslistrrr
_int_typesstrrformat
_string_typesRuntimeError)r	arrr
rs$







r)rZutilsrrZunicoderintZlongr	NameErrorrrrrr
s
site-packages/pip/_vendor/pytoml/__pycache__/parser.cpython-38.opt-1.pyc000064400000023542151464452710022146 0ustar00U

.eE(@sNddlZddlZddlmZddlmZmZejddkr@eZ	ne
Z	ddefdd	Zd
ddefdd
Z
GdddZedZddZedZddZddddddddZedZedZed Zed!Zed"Zefd#d$Zed%Zd&d'Zed(Zed)Zed*Zed+Z d,d-Z!d.d/Z"ed0Z#d1d2Z$dS)3N	TomlError)
rfc3339_reparse_rfc3339_recCs|SNtxvr	r	=/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/parser.py
rc	Cs t|||t|dt|dS)Nname)	translateobject_pairs_hookfilename)loadsreadgetattrrepr)Zfinrrr	r	rload
srzcCs|Srr	r
r	r	rr
rcst|tr|d}|dd}}}|}t|d}t|d}fddfdd	|D]D\}	}
|	d
kr|
\}}||krd||d||<qj|	dk}
|}|
dd
D]<}t||tr||d
\}}q|	|df\}}qƈ}|
d
}||krD|
r4|fg||<n|f||<qjt||trx|
sbd||
|fqj|
rd||\}}|dk	rd||f||<qjfdd||S)Nzutf-8z

)rrcst|dddSNrrr)msg)rposr	rerrorszloads..errorcs|\}}}|dkrTr>tfddddDr>dfddDn"|dkrvfd	dD||S)
Narrayc3s&|]\}}}}|ddkVqdS)rNr	).0krr
p)valuer	r	 s
z/loads..process_value..rzarray-type-mismatchcsg|]}|dqSrr	)r!item)r
process_valuer	r
"sz0loads..process_value..tablecs g|]}||dfqSr&r	)r!r")rr(r$r	rr)$s)any)r
rkindtextr)rr(r)rr$rr(s"zloads..process_valuekvz2duplicate_keys. Key "{0}" was used more than once.table_arrayZtable_type_mismatchZduplicate_tablescsl|dkr}|D]T}||kr&d||}t|trPfdd|D||<q|d|d||<q|S)NZkey_table_conflictcsg|]\}}||qSr	r	)r!ZscZtbl)merge_tablesr	rr)Qsz/loads..merge_tables..rr)
isinstancelist)scopetablesr"r
)rr1rr	rr1Is
zloads..merge_tables)r2bytesdecodereplace_Source_p_tomlformatgetr3
setdefaultappend)srrrrootr5r4srcZastr,r$r"r
Zis_table_arrayZcurrdZ	old_scopeZ
next_tabler	)rrr1rrr(rrr
sP





rc@seZdZd$ddZddZddZdd	Zd
dZdd
ZddZ	ddZ
ddZddZddZ
ddZddZddZddZd d!Zd"d#ZdS)%r9NcCs"||_d|_d|_||_g|_dS)N)rr)r?_pos_last	_filenamebacktrack_stack)selfr?rr	r	r__init__Ys
z_Source.__init__cCs|jSr)rDrGr	r	rlast`sz_Source.lastcCs|jSr)rCrIr	r	rrcsz_Source.poscCs
|dSr)_expectrIr	r	rfailfsz_Source.failcCs6|jr2|jd|_|dd|_||j|jSdSr)r?rD_advancerIr	r	rconsume_dotisz_Source.consume_dotcCs||Sr)rKrNrIr	r	r
expect_dotqsz_Source.expect_dotcCs|jsd|_dSdS)NTF)r?rDrIr	r	rconsume_eoftsz_Source.consume_eofcCs||Sr)rKrQrIr	r	r
expect_eofzsz_Source.expect_eofcCs8|j|r4|jt|d|_||_||dSdS)NTF)r?
startswithlenrDrMrGr?r	r	rconsume}s
z_Source.consumecCs|||Sr)rKrVrUr	r	rexpectsz_Source.expectcCsH||j}|rD|jt|dd|_||_||d|SdS)Nr)matchr?rTgrouprDrM)rGremr	r	r
consume_resz_Source.consume_recCs|||Sr)rKr\)rGrZr	r	r	expect_resz_Source.expect_recCs|j|j|jfdSr)rFr>r?rCrIr	r	r	__enter__sz_Source.__enter__cCs.|dkr|jn|j\|_|_|tkSr)rFpopr?rCr)rGtyper$	tracebackr	r	r__exit__sz_Source.__exit__cCs|j|jf|jd<dS)Nr0)r?rCrFrIr	r	rcommitsz_Source.commitcCs$|s td|jd|jd|j|S)Nrrr)rrCrE)rGrr	r	rrKsz_Source._expectcCsX|d}|dkr2|jd|jdt|f|_n"|jd|dt||f|_dS)Nrr0rr)rfindrCrTcount)rGr?Z
suffix_posr	r	rrMs
 z_Source._advance)N)__name__
__module____qualname__rHrJrrLrNrOrQrRrVrWr\r]r^rbrcrKrMr	r	r	rr9Xs"
	r9z!(?:[ \t]|#[^\n]*\n|#[^\n]*\Z|\n)*cCs|tdSr)r]_ews_rer?r	r	r_p_ewssrlz[ \t]*cCs|tdSr)r]_ws_rerkr	r	r_p_wssrnr
	"\)bnrdrrrrsfz[^"\\\000-\037]*zu([0-9a-fA-F]{4})zU([0-9a-fA-F]{8})z[btnfr\"\\]z
[ 	
]*cCsg}|||d|ds&q|tr2q|tsF|trt|	dd}d|krndkrznn|
|t|q|t|t
|	dqd|S)NrrsriirP)r>r]rYrVr\_newline_esc_re
_short_uni_re_long_uni_reintrJrL_chr_escapes_re_escapesjoin)r?Zcontentresr
r	r	r_p_basicstr_contents


rz[0-9a-zA-Z-_]+c
Cs|0|dt|t}|d|W5QRSQRX|dr|drt|d|td}|dn|td}|d|S|tdS)Nrr'''rr''')	rWr_basicstr_rerVr]
_litstr_ml_rerY
_litstr_re_key_re)r?rdr	r	r_p_keys






rzH[+-]?(?:0|[1-9](?:_?\d)*)(?:\.\d(?:_?\d)*)?(?:[eE][+-]?(?:\d(?:_?\d)*))?z&(?:""?(?!")|[^"\\\000-\011\013-\037])*z[^'\000\010\012-\037]*z((?:(?:|'|'')(?:[^'\000-\010\013-\037]))*c	Cs|}|dr"d|d|fS|dr_p_valuercrnrrL)r?rrrdr[itemsr"r	r	rrs~
















rcCs|}|dr|d}t|t|g}t||dr`t||t|t|q6|d|rx|d|rdnd||fSt|}t||dt|t||d}d||f|fS)	Nrrrr/r*rrr.)rrVrnrr>rWr)r?rrZis_arraykeyskeyr$r	r	r_p_stmt3s(







rz(?:[ \t]*(?:#[^\n]*)?\n)+[ \t]*c	Cshg}t||>|t||d||t|t||dq$W5QRXt|||S)Nr)rlr>rrcr]_stmtsep_rerR)r?rZstmtsr	r	rr:Ks
r:)%rZsysZcorerZutilsrrversion_infoZunichrr}chrdictrrr9compilerjrlrmrnrrrzr{r~ryrrrrrrrrrrr:r	r	r	rs@KU












M
site-packages/pip/_vendor/pytoml/__pycache__/utils.cpython-38.pyc000064400000004121151464452710021043 0ustar00U

.e@sHddlZddlZedZddZddZddZGd	d
d
ejZdS)NzP(\d{4})-(\d{2})-(\d{2})T(\d{2}):(\d{2}):(\d{2})(\.\d+)?(?:Z|([+-]\d{2}):(\d{2}))cCs(t|}|r|d|kr dSt|S)Nr)
rfc3339_rematchgroupparse_rfc3339_re)vmrs
site-packages/pip/_vendor/pytoml/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000524151464452710022404 0ustar00U

.e@s<ddlmZddlmZmZddlmZddlmZm	Z	dS))	TomlError)loadloads)translate_to_test)dumpdumpsN)
ZcorerparserrrZtestrwriterrrr
r
?/usr/lib/python3.8/site-packages/pip/_vendor/pytoml/__init__.pyssite-packages/pip/_vendor/pytoml/__pycache__/utils.cpython-38.opt-1.pyc000064400000004121151464452710022002 0ustar00U

.e@sHddlZddlZedZddZddZddZGd	d
d
ejZdS)NzP(\d{4})-(\d{2})-(\d{2})T(\d{2}):(\d{2}):(\d{2})(\.\d+)?(?:Z|([+-]\d{2}):(\d{2}))cCs(t|}|r|d|kr dSt|S)Nr)
rfc3339_rematchgroupparse_rfc3339_re)vmrs
site-packages/pip/_vendor/pytoml/writer.py000064400000006451151464452710014741 0ustar00from __future__ import unicode_literals
import io, datetime, math, string, sys

from .utils import format_rfc3339

try:
    from pathlib import PurePath as _path_types
except ImportError:
    _path_types = ()


if sys.version_info[0] == 3:
    long = int
    unicode = str


def dumps(obj, sort_keys=False):
    fout = io.StringIO()
    dump(obj, fout, sort_keys=sort_keys)
    return fout.getvalue()


_escapes = {'\n': 'n', '\r': 'r', '\\': '\\', '\t': 't', '\b': 'b', '\f': 'f', '"': '"'}


def _escape_string(s):
    res = []
    start = 0

    def flush():
        if start != i:
            res.append(s[start:i])
        return i + 1

    i = 0
    while i < len(s):
        c = s[i]
        if c in '"\\\n\r\t\b\f':
            start = flush()
            res.append('\\' + _escapes[c])
        elif ord(c) < 0x20:
            start = flush()
            res.append('\\u%04x' % ord(c))
        i += 1

    flush()
    return '"' + ''.join(res) + '"'


_key_chars = string.digits + string.ascii_letters + '-_'
def _escape_id(s):
    if any(c not in _key_chars for c in s):
        return _escape_string(s)
    return s


def _format_value(v):
    if isinstance(v, bool):
        return 'true' if v else 'false'
    if isinstance(v, int) or isinstance(v, long):
        return unicode(v)
    if isinstance(v, float):
        if math.isnan(v) or math.isinf(v):
            raise ValueError("{0} is not a valid TOML value".format(v))
        else:
            return repr(v)
    elif isinstance(v, unicode) or isinstance(v, bytes):
        return _escape_string(v)
    elif isinstance(v, datetime.datetime):
        return format_rfc3339(v)
    elif isinstance(v, list):
        return '[{0}]'.format(', '.join(_format_value(obj) for obj in v))
    elif isinstance(v, dict):
        return '{{{0}}}'.format(', '.join('{} = {}'.format(_escape_id(k), _format_value(obj)) for k, obj in v.items()))
    elif isinstance(v, _path_types):
        return _escape_string(str(v))
    else:
        raise RuntimeError(v)


def dump(obj, fout, sort_keys=False):
    tables = [((), obj, False)]

    while tables:
        name, table, is_array = tables.pop()
        if name:
            section_name = '.'.join(_escape_id(c) for c in name)
            if is_array:
                fout.write('[[{0}]]\n'.format(section_name))
            else:
                fout.write('[{0}]\n'.format(section_name))

        table_keys = sorted(table.keys()) if sort_keys else table.keys()
        new_tables = []
        has_kv = False
        for k in table_keys:
            v = table[k]
            if isinstance(v, dict):
                new_tables.append((name + (k,), v, False))
            elif isinstance(v, list) and v and all(isinstance(o, dict) for o in v):
                new_tables.extend((name + (k,), d, True) for d in v)
            elif v is None:
                # based on mojombo's comment: https://github.com/toml-lang/toml/issues/146#issuecomment-25019344
                fout.write(
                    '#{} = null  # To use: uncomment and replace null with value\n'.format(_escape_id(k)))
                has_kv = True
            else:
                fout.write('{0} = {1}\n'.format(_escape_id(k), _format_value(v)))
                has_kv = True

        tables.extend(reversed(new_tables))

        if (name or has_kv) and tables:
            fout.write('\n')
site-packages/pip/_vendor/pytoml/__init__.py000064400000000177151464452710015163 0ustar00from .core import TomlError
from .parser import load, loads
from .test import translate_to_test
from .writer import dump, dumpssite-packages/pip/_vendor/pytoml/core.py000064400000000775151464452710014360 0ustar00class TomlError(RuntimeError):
    def __init__(self, message, line, col, filename):
        RuntimeError.__init__(self, message, line, col, filename)
        self.message = message
        self.line = line
        self.col = col
        self.filename = filename

    def __str__(self):
        return '{}({}, {}): {}'.format(self.filename, self.line, self.col, self.message)

    def __repr__(self):
        return 'TomlError({!r}, {!r}, {!r}, {!r})'.format(self.message, self.line, self.col, self.filename)
site-packages/pip/_vendor/ipaddress.py000064400000233754151464452710014067 0ustar00# Copyright 2007 Google Inc.
#  Licensed to PSF under a Contributor Agreement.

"""A fast, lightweight IPv4/IPv6 manipulation library in Python.

This library is used to create/poke/manipulate IPv4 and IPv6 addresses
and networks.

"""

from __future__ import unicode_literals


import itertools
import struct

__version__ = '1.0.22'

# Compatibility functions
_compat_int_types = (int,)
try:
    _compat_int_types = (int, long)
except NameError:
    pass
try:
    _compat_str = unicode
except NameError:
    _compat_str = str
    assert bytes != str
if b'\0'[0] == 0:  # Python 3 semantics
    def _compat_bytes_to_byte_vals(byt):
        return byt
else:
    def _compat_bytes_to_byte_vals(byt):
        return [struct.unpack(b'!B', b)[0] for b in byt]
try:
    _compat_int_from_byte_vals = int.from_bytes
except AttributeError:
    def _compat_int_from_byte_vals(bytvals, endianess):
        assert endianess == 'big'
        res = 0
        for bv in bytvals:
            assert isinstance(bv, _compat_int_types)
            res = (res << 8) + bv
        return res


def _compat_to_bytes(intval, length, endianess):
    assert isinstance(intval, _compat_int_types)
    assert endianess == 'big'
    if length == 4:
        if intval < 0 or intval >= 2 ** 32:
            raise struct.error("integer out of range for 'I' format code")
        return struct.pack(b'!I', intval)
    elif length == 16:
        if intval < 0 or intval >= 2 ** 128:
            raise struct.error("integer out of range for 'QQ' format code")
        return struct.pack(b'!QQ', intval >> 64, intval & 0xffffffffffffffff)
    else:
        raise NotImplementedError()


if hasattr(int, 'bit_length'):
    # Not int.bit_length , since that won't work in 2.7 where long exists
    def _compat_bit_length(i):
        return i.bit_length()
else:
    def _compat_bit_length(i):
        for res in itertools.count():
            if i >> res == 0:
                return res


def _compat_range(start, end, step=1):
    assert step > 0
    i = start
    while i < end:
        yield i
        i += step


class _TotalOrderingMixin(object):
    __slots__ = ()

    # Helper that derives the other comparison operations from
    # __lt__ and __eq__
    # We avoid functools.total_ordering because it doesn't handle
    # NotImplemented correctly yet (http://bugs.python.org/issue10042)
    def __eq__(self, other):
        raise NotImplementedError

    def __ne__(self, other):
        equal = self.__eq__(other)
        if equal is NotImplemented:
            return NotImplemented
        return not equal

    def __lt__(self, other):
        raise NotImplementedError

    def __le__(self, other):
        less = self.__lt__(other)
        if less is NotImplemented or not less:
            return self.__eq__(other)
        return less

    def __gt__(self, other):
        less = self.__lt__(other)
        if less is NotImplemented:
            return NotImplemented
        equal = self.__eq__(other)
        if equal is NotImplemented:
            return NotImplemented
        return not (less or equal)

    def __ge__(self, other):
        less = self.__lt__(other)
        if less is NotImplemented:
            return NotImplemented
        return not less


IPV4LENGTH = 32
IPV6LENGTH = 128


class AddressValueError(ValueError):
    """A Value Error related to the address."""


class NetmaskValueError(ValueError):
    """A Value Error related to the netmask."""


def ip_address(address):
    """Take an IP string/int and return an object of the correct type.

    Args:
        address: A string or integer, the IP address.  Either IPv4 or
          IPv6 addresses may be supplied; integers less than 2**32 will
          be considered to be IPv4 by default.

    Returns:
        An IPv4Address or IPv6Address object.

    Raises:
        ValueError: if the *address* passed isn't either a v4 or a v6
          address

    """
    try:
        return IPv4Address(address)
    except (AddressValueError, NetmaskValueError):
        pass

    try:
        return IPv6Address(address)
    except (AddressValueError, NetmaskValueError):
        pass

    if isinstance(address, bytes):
        raise AddressValueError(
            '%r does not appear to be an IPv4 or IPv6 address. '
            'Did you pass in a bytes (str in Python 2) instead of'
            ' a unicode object?' % address)

    raise ValueError('%r does not appear to be an IPv4 or IPv6 address' %
                     address)


def ip_network(address, strict=True):
    """Take an IP string/int and return an object of the correct type.

    Args:
        address: A string or integer, the IP network.  Either IPv4 or
          IPv6 networks may be supplied; integers less than 2**32 will
          be considered to be IPv4 by default.

    Returns:
        An IPv4Network or IPv6Network object.

    Raises:
        ValueError: if the string passed isn't either a v4 or a v6
          address. Or if the network has host bits set.

    """
    try:
        return IPv4Network(address, strict)
    except (AddressValueError, NetmaskValueError):
        pass

    try:
        return IPv6Network(address, strict)
    except (AddressValueError, NetmaskValueError):
        pass

    if isinstance(address, bytes):
        raise AddressValueError(
            '%r does not appear to be an IPv4 or IPv6 network. '
            'Did you pass in a bytes (str in Python 2) instead of'
            ' a unicode object?' % address)

    raise ValueError('%r does not appear to be an IPv4 or IPv6 network' %
                     address)


def ip_interface(address):
    """Take an IP string/int and return an object of the correct type.

    Args:
        address: A string or integer, the IP address.  Either IPv4 or
          IPv6 addresses may be supplied; integers less than 2**32 will
          be considered to be IPv4 by default.

    Returns:
        An IPv4Interface or IPv6Interface object.

    Raises:
        ValueError: if the string passed isn't either a v4 or a v6
          address.

    Notes:
        The IPv?Interface classes describe an Address on a particular
        Network, so they're basically a combination of both the Address
        and Network classes.

    """
    try:
        return IPv4Interface(address)
    except (AddressValueError, NetmaskValueError):
        pass

    try:
        return IPv6Interface(address)
    except (AddressValueError, NetmaskValueError):
        pass

    raise ValueError('%r does not appear to be an IPv4 or IPv6 interface' %
                     address)


def v4_int_to_packed(address):
    """Represent an address as 4 packed bytes in network (big-endian) order.

    Args:
        address: An integer representation of an IPv4 IP address.

    Returns:
        The integer address packed as 4 bytes in network (big-endian) order.

    Raises:
        ValueError: If the integer is negative or too large to be an
          IPv4 IP address.

    """
    try:
        return _compat_to_bytes(address, 4, 'big')
    except (struct.error, OverflowError):
        raise ValueError("Address negative or too large for IPv4")


def v6_int_to_packed(address):
    """Represent an address as 16 packed bytes in network (big-endian) order.

    Args:
        address: An integer representation of an IPv6 IP address.

    Returns:
        The integer address packed as 16 bytes in network (big-endian) order.

    """
    try:
        return _compat_to_bytes(address, 16, 'big')
    except (struct.error, OverflowError):
        raise ValueError("Address negative or too large for IPv6")


def _split_optional_netmask(address):
    """Helper to split the netmask and raise AddressValueError if needed"""
    addr = _compat_str(address).split('/')
    if len(addr) > 2:
        raise AddressValueError("Only one '/' permitted in %r" % address)
    return addr


def _find_address_range(addresses):
    """Find a sequence of sorted deduplicated IPv#Address.

    Args:
        addresses: a list of IPv#Address objects.

    Yields:
        A tuple containing the first and last IP addresses in the sequence.

    """
    it = iter(addresses)
    first = last = next(it)
    for ip in it:
        if ip._ip != last._ip + 1:
            yield first, last
            first = ip
        last = ip
    yield first, last


def _count_righthand_zero_bits(number, bits):
    """Count the number of zero bits on the right hand side.

    Args:
        number: an integer.
        bits: maximum number of bits to count.

    Returns:
        The number of zero bits on the right hand side of the number.

    """
    if number == 0:
        return bits
    return min(bits, _compat_bit_length(~number & (number - 1)))


def summarize_address_range(first, last):
    """Summarize a network range given the first and last IP addresses.

    Example:
        >>> list(summarize_address_range(IPv4Address('192.0.2.0'),
        ...                              IPv4Address('192.0.2.130')))
        ...                                #doctest: +NORMALIZE_WHITESPACE
        [IPv4Network('192.0.2.0/25'), IPv4Network('192.0.2.128/31'),
         IPv4Network('192.0.2.130/32')]

    Args:
        first: the first IPv4Address or IPv6Address in the range.
        last: the last IPv4Address or IPv6Address in the range.

    Returns:
        An iterator of the summarized IPv(4|6) network objects.

    Raise:
        TypeError:
            If the first and last objects are not IP addresses.
            If the first and last objects are not the same version.
        ValueError:
            If the last object is not greater than the first.
            If the version of the first address is not 4 or 6.

    """
    if (not (isinstance(first, _BaseAddress) and
             isinstance(last, _BaseAddress))):
        raise TypeError('first and last must be IP addresses, not networks')
    if first.version != last.version:
        raise TypeError("%s and %s are not of the same version" % (
                        first, last))
    if first > last:
        raise ValueError('last IP address must be greater than first')

    if first.version == 4:
        ip = IPv4Network
    elif first.version == 6:
        ip = IPv6Network
    else:
        raise ValueError('unknown IP version')

    ip_bits = first._max_prefixlen
    first_int = first._ip
    last_int = last._ip
    while first_int <= last_int:
        nbits = min(_count_righthand_zero_bits(first_int, ip_bits),
                    _compat_bit_length(last_int - first_int + 1) - 1)
        net = ip((first_int, ip_bits - nbits))
        yield net
        first_int += 1 << nbits
        if first_int - 1 == ip._ALL_ONES:
            break


def _collapse_addresses_internal(addresses):
    """Loops through the addresses, collapsing concurrent netblocks.

    Example:

        ip1 = IPv4Network('192.0.2.0/26')
        ip2 = IPv4Network('192.0.2.64/26')
        ip3 = IPv4Network('192.0.2.128/26')
        ip4 = IPv4Network('192.0.2.192/26')

        _collapse_addresses_internal([ip1, ip2, ip3, ip4]) ->
          [IPv4Network('192.0.2.0/24')]

        This shouldn't be called directly; it is called via
          collapse_addresses([]).

    Args:
        addresses: A list of IPv4Network's or IPv6Network's

    Returns:
        A list of IPv4Network's or IPv6Network's depending on what we were
        passed.

    """
    # First merge
    to_merge = list(addresses)
    subnets = {}
    while to_merge:
        net = to_merge.pop()
        supernet = net.supernet()
        existing = subnets.get(supernet)
        if existing is None:
            subnets[supernet] = net
        elif existing != net:
            # Merge consecutive subnets
            del subnets[supernet]
            to_merge.append(supernet)
    # Then iterate over resulting networks, skipping subsumed subnets
    last = None
    for net in sorted(subnets.values()):
        if last is not None:
            # Since they are sorted,
            # last.network_address <= net.network_address is a given.
            if last.broadcast_address >= net.broadcast_address:
                continue
        yield net
        last = net


def collapse_addresses(addresses):
    """Collapse a list of IP objects.

    Example:
        collapse_addresses([IPv4Network('192.0.2.0/25'),
                            IPv4Network('192.0.2.128/25')]) ->
                           [IPv4Network('192.0.2.0/24')]

    Args:
        addresses: An iterator of IPv4Network or IPv6Network objects.

    Returns:
        An iterator of the collapsed IPv(4|6)Network objects.

    Raises:
        TypeError: If passed a list of mixed version objects.

    """
    addrs = []
    ips = []
    nets = []

    # split IP addresses and networks
    for ip in addresses:
        if isinstance(ip, _BaseAddress):
            if ips and ips[-1]._version != ip._version:
                raise TypeError("%s and %s are not of the same version" % (
                                ip, ips[-1]))
            ips.append(ip)
        elif ip._prefixlen == ip._max_prefixlen:
            if ips and ips[-1]._version != ip._version:
                raise TypeError("%s and %s are not of the same version" % (
                                ip, ips[-1]))
            try:
                ips.append(ip.ip)
            except AttributeError:
                ips.append(ip.network_address)
        else:
            if nets and nets[-1]._version != ip._version:
                raise TypeError("%s and %s are not of the same version" % (
                                ip, nets[-1]))
            nets.append(ip)

    # sort and dedup
    ips = sorted(set(ips))

    # find consecutive address ranges in the sorted sequence and summarize them
    if ips:
        for first, last in _find_address_range(ips):
            addrs.extend(summarize_address_range(first, last))

    return _collapse_addresses_internal(addrs + nets)


def get_mixed_type_key(obj):
    """Return a key suitable for sorting between networks and addresses.

    Address and Network objects are not sortable by default; they're
    fundamentally different so the expression

        IPv4Address('192.0.2.0') <= IPv4Network('192.0.2.0/24')

    doesn't make any sense.  There are some times however, where you may wish
    to have ipaddress sort these for you anyway. If you need to do this, you
    can use this function as the key= argument to sorted().

    Args:
      obj: either a Network or Address object.
    Returns:
      appropriate key.

    """
    if isinstance(obj, _BaseNetwork):
        return obj._get_networks_key()
    elif isinstance(obj, _BaseAddress):
        return obj._get_address_key()
    return NotImplemented


class _IPAddressBase(_TotalOrderingMixin):

    """The mother class."""

    __slots__ = ()

    @property
    def exploded(self):
        """Return the longhand version of the IP address as a string."""
        return self._explode_shorthand_ip_string()

    @property
    def compressed(self):
        """Return the shorthand version of the IP address as a string."""
        return _compat_str(self)

    @property
    def reverse_pointer(self):
        """The name of the reverse DNS pointer for the IP address, e.g.:
            >>> ipaddress.ip_address("127.0.0.1").reverse_pointer
            '1.0.0.127.in-addr.arpa'
            >>> ipaddress.ip_address("2001:db8::1").reverse_pointer
            '1.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.8.b.d.0.1.0.0.2.ip6.arpa'

        """
        return self._reverse_pointer()

    @property
    def version(self):
        msg = '%200s has no version specified' % (type(self),)
        raise NotImplementedError(msg)

    def _check_int_address(self, address):
        if address < 0:
            msg = "%d (< 0) is not permitted as an IPv%d address"
            raise AddressValueError(msg % (address, self._version))
        if address > self._ALL_ONES:
            msg = "%d (>= 2**%d) is not permitted as an IPv%d address"
            raise AddressValueError(msg % (address, self._max_prefixlen,
                                           self._version))

    def _check_packed_address(self, address, expected_len):
        address_len = len(address)
        if address_len != expected_len:
            msg = (
                '%r (len %d != %d) is not permitted as an IPv%d address. '
                'Did you pass in a bytes (str in Python 2) instead of'
                ' a unicode object?')
            raise AddressValueError(msg % (address, address_len,
                                           expected_len, self._version))

    @classmethod
    def _ip_int_from_prefix(cls, prefixlen):
        """Turn the prefix length into a bitwise netmask

        Args:
            prefixlen: An integer, the prefix length.

        Returns:
            An integer.

        """
        return cls._ALL_ONES ^ (cls._ALL_ONES >> prefixlen)

    @classmethod
    def _prefix_from_ip_int(cls, ip_int):
        """Return prefix length from the bitwise netmask.

        Args:
            ip_int: An integer, the netmask in expanded bitwise format

        Returns:
            An integer, the prefix length.

        Raises:
            ValueError: If the input intermingles zeroes & ones
        """
        trailing_zeroes = _count_righthand_zero_bits(ip_int,
                                                     cls._max_prefixlen)
        prefixlen = cls._max_prefixlen - trailing_zeroes
        leading_ones = ip_int >> trailing_zeroes
        all_ones = (1 << prefixlen) - 1
        if leading_ones != all_ones:
            byteslen = cls._max_prefixlen // 8
            details = _compat_to_bytes(ip_int, byteslen, 'big')
            msg = 'Netmask pattern %r mixes zeroes & ones'
            raise ValueError(msg % details)
        return prefixlen

    @classmethod
    def _report_invalid_netmask(cls, netmask_str):
        msg = '%r is not a valid netmask' % netmask_str
        raise NetmaskValueError(msg)

    @classmethod
    def _prefix_from_prefix_string(cls, prefixlen_str):
        """Return prefix length from a numeric string

        Args:
            prefixlen_str: The string to be converted

        Returns:
            An integer, the prefix length.

        Raises:
            NetmaskValueError: If the input is not a valid netmask
        """
        # int allows a leading +/- as well as surrounding whitespace,
        # so we ensure that isn't the case
        if not _BaseV4._DECIMAL_DIGITS.issuperset(prefixlen_str):
            cls._report_invalid_netmask(prefixlen_str)
        try:
            prefixlen = int(prefixlen_str)
        except ValueError:
            cls._report_invalid_netmask(prefixlen_str)
        if not (0 <= prefixlen <= cls._max_prefixlen):
            cls._report_invalid_netmask(prefixlen_str)
        return prefixlen

    @classmethod
    def _prefix_from_ip_string(cls, ip_str):
        """Turn a netmask/hostmask string into a prefix length

        Args:
            ip_str: The netmask/hostmask to be converted

        Returns:
            An integer, the prefix length.

        Raises:
            NetmaskValueError: If the input is not a valid netmask/hostmask
        """
        # Parse the netmask/hostmask like an IP address.
        try:
            ip_int = cls._ip_int_from_string(ip_str)
        except AddressValueError:
            cls._report_invalid_netmask(ip_str)

        # Try matching a netmask (this would be /1*0*/ as a bitwise regexp).
        # Note that the two ambiguous cases (all-ones and all-zeroes) are
        # treated as netmasks.
        try:
            return cls._prefix_from_ip_int(ip_int)
        except ValueError:
            pass

        # Invert the bits, and try matching a /0+1+/ hostmask instead.
        ip_int ^= cls._ALL_ONES
        try:
            return cls._prefix_from_ip_int(ip_int)
        except ValueError:
            cls._report_invalid_netmask(ip_str)

    def __reduce__(self):
        return self.__class__, (_compat_str(self),)


class _BaseAddress(_IPAddressBase):

    """A generic IP object.

    This IP class contains the version independent methods which are
    used by single IP addresses.
    """

    __slots__ = ()

    def __int__(self):
        return self._ip

    def __eq__(self, other):
        try:
            return (self._ip == other._ip and
                    self._version == other._version)
        except AttributeError:
            return NotImplemented

    def __lt__(self, other):
        if not isinstance(other, _IPAddressBase):
            return NotImplemented
        if not isinstance(other, _BaseAddress):
            raise TypeError('%s and %s are not of the same type' % (
                self, other))
        if self._version != other._version:
            raise TypeError('%s and %s are not of the same version' % (
                self, other))
        if self._ip != other._ip:
            return self._ip < other._ip
        return False

    # Shorthand for Integer addition and subtraction. This is not
    # meant to ever support addition/subtraction of addresses.
    def __add__(self, other):
        if not isinstance(other, _compat_int_types):
            return NotImplemented
        return self.__class__(int(self) + other)

    def __sub__(self, other):
        if not isinstance(other, _compat_int_types):
            return NotImplemented
        return self.__class__(int(self) - other)

    def __repr__(self):
        return '%s(%r)' % (self.__class__.__name__, _compat_str(self))

    def __str__(self):
        return _compat_str(self._string_from_ip_int(self._ip))

    def __hash__(self):
        return hash(hex(int(self._ip)))

    def _get_address_key(self):
        return (self._version, self)

    def __reduce__(self):
        return self.__class__, (self._ip,)


class _BaseNetwork(_IPAddressBase):

    """A generic IP network object.

    This IP class contains the version independent methods which are
    used by networks.

    """
    def __init__(self, address):
        self._cache = {}

    def __repr__(self):
        return '%s(%r)' % (self.__class__.__name__, _compat_str(self))

    def __str__(self):
        return '%s/%d' % (self.network_address, self.prefixlen)

    def hosts(self):
        """Generate Iterator over usable hosts in a network.

        This is like __iter__ except it doesn't return the network
        or broadcast addresses.

        """
        network = int(self.network_address)
        broadcast = int(self.broadcast_address)
        for x in _compat_range(network + 1, broadcast):
            yield self._address_class(x)

    def __iter__(self):
        network = int(self.network_address)
        broadcast = int(self.broadcast_address)
        for x in _compat_range(network, broadcast + 1):
            yield self._address_class(x)

    def __getitem__(self, n):
        network = int(self.network_address)
        broadcast = int(self.broadcast_address)
        if n >= 0:
            if network + n > broadcast:
                raise IndexError('address out of range')
            return self._address_class(network + n)
        else:
            n += 1
            if broadcast + n < network:
                raise IndexError('address out of range')
            return self._address_class(broadcast + n)

    def __lt__(self, other):
        if not isinstance(other, _IPAddressBase):
            return NotImplemented
        if not isinstance(other, _BaseNetwork):
            raise TypeError('%s and %s are not of the same type' % (
                            self, other))
        if self._version != other._version:
            raise TypeError('%s and %s are not of the same version' % (
                            self, other))
        if self.network_address != other.network_address:
            return self.network_address < other.network_address
        if self.netmask != other.netmask:
            return self.netmask < other.netmask
        return False

    def __eq__(self, other):
        try:
            return (self._version == other._version and
                    self.network_address == other.network_address and
                    int(self.netmask) == int(other.netmask))
        except AttributeError:
            return NotImplemented

    def __hash__(self):
        return hash(int(self.network_address) ^ int(self.netmask))

    def __contains__(self, other):
        # always false if one is v4 and the other is v6.
        if self._version != other._version:
            return False
        # dealing with another network.
        if isinstance(other, _BaseNetwork):
            return False
        # dealing with another address
        else:
            # address
            return (int(self.network_address) <= int(other._ip) <=
                    int(self.broadcast_address))

    def overlaps(self, other):
        """Tell if self is partly contained in other."""
        return self.network_address in other or (
            self.broadcast_address in other or (
                other.network_address in self or (
                    other.broadcast_address in self)))

    @property
    def broadcast_address(self):
        x = self._cache.get('broadcast_address')
        if x is None:
            x = self._address_class(int(self.network_address) |
                                    int(self.hostmask))
            self._cache['broadcast_address'] = x
        return x

    @property
    def hostmask(self):
        x = self._cache.get('hostmask')
        if x is None:
            x = self._address_class(int(self.netmask) ^ self._ALL_ONES)
            self._cache['hostmask'] = x
        return x

    @property
    def with_prefixlen(self):
        return '%s/%d' % (self.network_address, self._prefixlen)

    @property
    def with_netmask(self):
        return '%s/%s' % (self.network_address, self.netmask)

    @property
    def with_hostmask(self):
        return '%s/%s' % (self.network_address, self.hostmask)

    @property
    def num_addresses(self):
        """Number of hosts in the current subnet."""
        return int(self.broadcast_address) - int(self.network_address) + 1

    @property
    def _address_class(self):
        # Returning bare address objects (rather than interfaces) allows for
        # more consistent behaviour across the network address, broadcast
        # address and individual host addresses.
        msg = '%200s has no associated address class' % (type(self),)
        raise NotImplementedError(msg)

    @property
    def prefixlen(self):
        return self._prefixlen

    def address_exclude(self, other):
        """Remove an address from a larger block.

        For example:

            addr1 = ip_network('192.0.2.0/28')
            addr2 = ip_network('192.0.2.1/32')
            list(addr1.address_exclude(addr2)) =
                [IPv4Network('192.0.2.0/32'), IPv4Network('192.0.2.2/31'),
                 IPv4Network('192.0.2.4/30'), IPv4Network('192.0.2.8/29')]

        or IPv6:

            addr1 = ip_network('2001:db8::1/32')
            addr2 = ip_network('2001:db8::1/128')
            list(addr1.address_exclude(addr2)) =
                [ip_network('2001:db8::1/128'),
                 ip_network('2001:db8::2/127'),
                 ip_network('2001:db8::4/126'),
                 ip_network('2001:db8::8/125'),
                 ...
                 ip_network('2001:db8:8000::/33')]

        Args:
            other: An IPv4Network or IPv6Network object of the same type.

        Returns:
            An iterator of the IPv(4|6)Network objects which is self
            minus other.

        Raises:
            TypeError: If self and other are of differing address
              versions, or if other is not a network object.
            ValueError: If other is not completely contained by self.

        """
        if not self._version == other._version:
            raise TypeError("%s and %s are not of the same version" % (
                            self, other))

        if not isinstance(other, _BaseNetwork):
            raise TypeError("%s is not a network object" % other)

        if not other.subnet_of(self):
            raise ValueError('%s not contained in %s' % (other, self))
        if other == self:
            return

        # Make sure we're comparing the network of other.
        other = other.__class__('%s/%s' % (other.network_address,
                                           other.prefixlen))

        s1, s2 = self.subnets()
        while s1 != other and s2 != other:
            if other.subnet_of(s1):
                yield s2
                s1, s2 = s1.subnets()
            elif other.subnet_of(s2):
                yield s1
                s1, s2 = s2.subnets()
            else:
                # If we got here, there's a bug somewhere.
                raise AssertionError('Error performing exclusion: '
                                     's1: %s s2: %s other: %s' %
                                     (s1, s2, other))
        if s1 == other:
            yield s2
        elif s2 == other:
            yield s1
        else:
            # If we got here, there's a bug somewhere.
            raise AssertionError('Error performing exclusion: '
                                 's1: %s s2: %s other: %s' %
                                 (s1, s2, other))

    def compare_networks(self, other):
        """Compare two IP objects.

        This is only concerned about the comparison of the integer
        representation of the network addresses.  This means that the
        host bits aren't considered at all in this method.  If you want
        to compare host bits, you can easily enough do a
        'HostA._ip < HostB._ip'

        Args:
            other: An IP object.

        Returns:
            If the IP versions of self and other are the same, returns:

            -1 if self < other:
              eg: IPv4Network('192.0.2.0/25') < IPv4Network('192.0.2.128/25')
              IPv6Network('2001:db8::1000/124') <
                  IPv6Network('2001:db8::2000/124')
            0 if self == other
              eg: IPv4Network('192.0.2.0/24') == IPv4Network('192.0.2.0/24')
              IPv6Network('2001:db8::1000/124') ==
                  IPv6Network('2001:db8::1000/124')
            1 if self > other
              eg: IPv4Network('192.0.2.128/25') > IPv4Network('192.0.2.0/25')
                  IPv6Network('2001:db8::2000/124') >
                      IPv6Network('2001:db8::1000/124')

          Raises:
              TypeError if the IP versions are different.

        """
        # does this need to raise a ValueError?
        if self._version != other._version:
            raise TypeError('%s and %s are not of the same type' % (
                            self, other))
        # self._version == other._version below here:
        if self.network_address < other.network_address:
            return -1
        if self.network_address > other.network_address:
            return 1
        # self.network_address == other.network_address below here:
        if self.netmask < other.netmask:
            return -1
        if self.netmask > other.netmask:
            return 1
        return 0

    def _get_networks_key(self):
        """Network-only key function.

        Returns an object that identifies this address' network and
        netmask. This function is a suitable "key" argument for sorted()
        and list.sort().

        """
        return (self._version, self.network_address, self.netmask)

    def subnets(self, prefixlen_diff=1, new_prefix=None):
        """The subnets which join to make the current subnet.

        In the case that self contains only one IP
        (self._prefixlen == 32 for IPv4 or self._prefixlen == 128
        for IPv6), yield an iterator with just ourself.

        Args:
            prefixlen_diff: An integer, the amount the prefix length
              should be increased by. This should not be set if
              new_prefix is also set.
            new_prefix: The desired new prefix length. This must be a
              larger number (smaller prefix) than the existing prefix.
              This should not be set if prefixlen_diff is also set.

        Returns:
            An iterator of IPv(4|6) objects.

        Raises:
            ValueError: The prefixlen_diff is too small or too large.
                OR
            prefixlen_diff and new_prefix are both set or new_prefix
              is a smaller number than the current prefix (smaller
              number means a larger network)

        """
        if self._prefixlen == self._max_prefixlen:
            yield self
            return

        if new_prefix is not None:
            if new_prefix < self._prefixlen:
                raise ValueError('new prefix must be longer')
            if prefixlen_diff != 1:
                raise ValueError('cannot set prefixlen_diff and new_prefix')
            prefixlen_diff = new_prefix - self._prefixlen

        if prefixlen_diff < 0:
            raise ValueError('prefix length diff must be > 0')
        new_prefixlen = self._prefixlen + prefixlen_diff

        if new_prefixlen > self._max_prefixlen:
            raise ValueError(
                'prefix length diff %d is invalid for netblock %s' % (
                    new_prefixlen, self))

        start = int(self.network_address)
        end = int(self.broadcast_address) + 1
        step = (int(self.hostmask) + 1) >> prefixlen_diff
        for new_addr in _compat_range(start, end, step):
            current = self.__class__((new_addr, new_prefixlen))
            yield current

    def supernet(self, prefixlen_diff=1, new_prefix=None):
        """The supernet containing the current network.

        Args:
            prefixlen_diff: An integer, the amount the prefix length of
              the network should be decreased by.  For example, given a
              /24 network and a prefixlen_diff of 3, a supernet with a
              /21 netmask is returned.

        Returns:
            An IPv4 network object.

        Raises:
            ValueError: If self.prefixlen - prefixlen_diff < 0. I.e., you have
              a negative prefix length.
                OR
            If prefixlen_diff and new_prefix are both set or new_prefix is a
              larger number than the current prefix (larger number means a
              smaller network)

        """
        if self._prefixlen == 0:
            return self

        if new_prefix is not None:
            if new_prefix > self._prefixlen:
                raise ValueError('new prefix must be shorter')
            if prefixlen_diff != 1:
                raise ValueError('cannot set prefixlen_diff and new_prefix')
            prefixlen_diff = self._prefixlen - new_prefix

        new_prefixlen = self.prefixlen - prefixlen_diff
        if new_prefixlen < 0:
            raise ValueError(
                'current prefixlen is %d, cannot have a prefixlen_diff of %d' %
                (self.prefixlen, prefixlen_diff))
        return self.__class__((
            int(self.network_address) & (int(self.netmask) << prefixlen_diff),
            new_prefixlen))

    @property
    def is_multicast(self):
        """Test if the address is reserved for multicast use.

        Returns:
            A boolean, True if the address is a multicast address.
            See RFC 2373 2.7 for details.

        """
        return (self.network_address.is_multicast and
                self.broadcast_address.is_multicast)

    @staticmethod
    def _is_subnet_of(a, b):
        try:
            # Always false if one is v4 and the other is v6.
            if a._version != b._version:
                raise TypeError("%s and %s are not of the same version" (a, b))
            return (b.network_address <= a.network_address and
                    b.broadcast_address >= a.broadcast_address)
        except AttributeError:
            raise TypeError("Unable to test subnet containment "
                            "between %s and %s" % (a, b))

    def subnet_of(self, other):
        """Return True if this network is a subnet of other."""
        return self._is_subnet_of(self, other)

    def supernet_of(self, other):
        """Return True if this network is a supernet of other."""
        return self._is_subnet_of(other, self)

    @property
    def is_reserved(self):
        """Test if the address is otherwise IETF reserved.

        Returns:
            A boolean, True if the address is within one of the
            reserved IPv6 Network ranges.

        """
        return (self.network_address.is_reserved and
                self.broadcast_address.is_reserved)

    @property
    def is_link_local(self):
        """Test if the address is reserved for link-local.

        Returns:
            A boolean, True if the address is reserved per RFC 4291.

        """
        return (self.network_address.is_link_local and
                self.broadcast_address.is_link_local)

    @property
    def is_private(self):
        """Test if this address is allocated for private networks.

        Returns:
            A boolean, True if the address is reserved per
            iana-ipv4-special-registry or iana-ipv6-special-registry.

        """
        return (self.network_address.is_private and
                self.broadcast_address.is_private)

    @property
    def is_global(self):
        """Test if this address is allocated for public networks.

        Returns:
            A boolean, True if the address is not reserved per
            iana-ipv4-special-registry or iana-ipv6-special-registry.

        """
        return not self.is_private

    @property
    def is_unspecified(self):
        """Test if the address is unspecified.

        Returns:
            A boolean, True if this is the unspecified address as defined in
            RFC 2373 2.5.2.

        """
        return (self.network_address.is_unspecified and
                self.broadcast_address.is_unspecified)

    @property
    def is_loopback(self):
        """Test if the address is a loopback address.

        Returns:
            A boolean, True if the address is a loopback address as defined in
            RFC 2373 2.5.3.

        """
        return (self.network_address.is_loopback and
                self.broadcast_address.is_loopback)


class _BaseV4(object):

    """Base IPv4 object.

    The following methods are used by IPv4 objects in both single IP
    addresses and networks.

    """

    __slots__ = ()
    _version = 4
    # Equivalent to 255.255.255.255 or 32 bits of 1's.
    _ALL_ONES = (2 ** IPV4LENGTH) - 1
    _DECIMAL_DIGITS = frozenset('0123456789')

    # the valid octets for host and netmasks. only useful for IPv4.
    _valid_mask_octets = frozenset([255, 254, 252, 248, 240, 224, 192, 128, 0])

    _max_prefixlen = IPV4LENGTH
    # There are only a handful of valid v4 netmasks, so we cache them all
    # when constructed (see _make_netmask()).
    _netmask_cache = {}

    def _explode_shorthand_ip_string(self):
        return _compat_str(self)

    @classmethod
    def _make_netmask(cls, arg):
        """Make a (netmask, prefix_len) tuple from the given argument.

        Argument can be:
        - an integer (the prefix length)
        - a string representing the prefix length (e.g. "24")
        - a string representing the prefix netmask (e.g. "255.255.255.0")
        """
        if arg not in cls._netmask_cache:
            if isinstance(arg, _compat_int_types):
                prefixlen = arg
            else:
                try:
                    # Check for a netmask in prefix length form
                    prefixlen = cls._prefix_from_prefix_string(arg)
                except NetmaskValueError:
                    # Check for a netmask or hostmask in dotted-quad form.
                    # This may raise NetmaskValueError.
                    prefixlen = cls._prefix_from_ip_string(arg)
            netmask = IPv4Address(cls._ip_int_from_prefix(prefixlen))
            cls._netmask_cache[arg] = netmask, prefixlen
        return cls._netmask_cache[arg]

    @classmethod
    def _ip_int_from_string(cls, ip_str):
        """Turn the given IP string into an integer for comparison.

        Args:
            ip_str: A string, the IP ip_str.

        Returns:
            The IP ip_str as an integer.

        Raises:
            AddressValueError: if ip_str isn't a valid IPv4 Address.

        """
        if not ip_str:
            raise AddressValueError('Address cannot be empty')

        octets = ip_str.split('.')
        if len(octets) != 4:
            raise AddressValueError("Expected 4 octets in %r" % ip_str)

        try:
            return _compat_int_from_byte_vals(
                map(cls._parse_octet, octets), 'big')
        except ValueError as exc:
            raise AddressValueError("%s in %r" % (exc, ip_str))

    @classmethod
    def _parse_octet(cls, octet_str):
        """Convert a decimal octet into an integer.

        Args:
            octet_str: A string, the number to parse.

        Returns:
            The octet as an integer.

        Raises:
            ValueError: if the octet isn't strictly a decimal from [0..255].

        """
        if not octet_str:
            raise ValueError("Empty octet not permitted")
        # Whitelist the characters, since int() allows a lot of bizarre stuff.
        if not cls._DECIMAL_DIGITS.issuperset(octet_str):
            msg = "Only decimal digits permitted in %r"
            raise ValueError(msg % octet_str)
        # We do the length check second, since the invalid character error
        # is likely to be more informative for the user
        if len(octet_str) > 3:
            msg = "At most 3 characters permitted in %r"
            raise ValueError(msg % octet_str)
        # Convert to integer (we know digits are legal)
        octet_int = int(octet_str, 10)
        # Any octets that look like they *might* be written in octal,
        # and which don't look exactly the same in both octal and
        # decimal are rejected as ambiguous
        if octet_int > 7 and octet_str[0] == '0':
            msg = "Ambiguous (octal/decimal) value in %r not permitted"
            raise ValueError(msg % octet_str)
        if octet_int > 255:
            raise ValueError("Octet %d (> 255) not permitted" % octet_int)
        return octet_int

    @classmethod
    def _string_from_ip_int(cls, ip_int):
        """Turns a 32-bit integer into dotted decimal notation.

        Args:
            ip_int: An integer, the IP address.

        Returns:
            The IP address as a string in dotted decimal notation.

        """
        return '.'.join(_compat_str(struct.unpack(b'!B', b)[0]
                                    if isinstance(b, bytes)
                                    else b)
                        for b in _compat_to_bytes(ip_int, 4, 'big'))

    def _is_hostmask(self, ip_str):
        """Test if the IP string is a hostmask (rather than a netmask).

        Args:
            ip_str: A string, the potential hostmask.

        Returns:
            A boolean, True if the IP string is a hostmask.

        """
        bits = ip_str.split('.')
        try:
            parts = [x for x in map(int, bits) if x in self._valid_mask_octets]
        except ValueError:
            return False
        if len(parts) != len(bits):
            return False
        if parts[0] < parts[-1]:
            return True
        return False

    def _reverse_pointer(self):
        """Return the reverse DNS pointer name for the IPv4 address.

        This implements the method described in RFC1035 3.5.

        """
        reverse_octets = _compat_str(self).split('.')[::-1]
        return '.'.join(reverse_octets) + '.in-addr.arpa'

    @property
    def max_prefixlen(self):
        return self._max_prefixlen

    @property
    def version(self):
        return self._version


class IPv4Address(_BaseV4, _BaseAddress):

    """Represent and manipulate single IPv4 Addresses."""

    __slots__ = ('_ip', '__weakref__')

    def __init__(self, address):

        """
        Args:
            address: A string or integer representing the IP

              Additionally, an integer can be passed, so
              IPv4Address('192.0.2.1') == IPv4Address(3221225985).
              or, more generally
              IPv4Address(int(IPv4Address('192.0.2.1'))) ==
                IPv4Address('192.0.2.1')

        Raises:
            AddressValueError: If ipaddress isn't a valid IPv4 address.

        """
        # Efficient constructor from integer.
        if isinstance(address, _compat_int_types):
            self._check_int_address(address)
            self._ip = address
            return

        # Constructing from a packed address
        if isinstance(address, bytes):
            self._check_packed_address(address, 4)
            bvs = _compat_bytes_to_byte_vals(address)
            self._ip = _compat_int_from_byte_vals(bvs, 'big')
            return

        # Assume input argument to be string or any object representation
        # which converts into a formatted IP string.
        addr_str = _compat_str(address)
        if '/' in addr_str:
            raise AddressValueError("Unexpected '/' in %r" % address)
        self._ip = self._ip_int_from_string(addr_str)

    @property
    def packed(self):
        """The binary representation of this address."""
        return v4_int_to_packed(self._ip)

    @property
    def is_reserved(self):
        """Test if the address is otherwise IETF reserved.

         Returns:
             A boolean, True if the address is within the
             reserved IPv4 Network range.

        """
        return self in self._constants._reserved_network

    @property
    def is_private(self):
        """Test if this address is allocated for private networks.

        Returns:
            A boolean, True if the address is reserved per
            iana-ipv4-special-registry.

        """
        return any(self in net for net in self._constants._private_networks)

    @property
    def is_global(self):
        return (
            self not in self._constants._public_network and
            not self.is_private)

    @property
    def is_multicast(self):
        """Test if the address is reserved for multicast use.

        Returns:
            A boolean, True if the address is multicast.
            See RFC 3171 for details.

        """
        return self in self._constants._multicast_network

    @property
    def is_unspecified(self):
        """Test if the address is unspecified.

        Returns:
            A boolean, True if this is the unspecified address as defined in
            RFC 5735 3.

        """
        return self == self._constants._unspecified_address

    @property
    def is_loopback(self):
        """Test if the address is a loopback address.

        Returns:
            A boolean, True if the address is a loopback per RFC 3330.

        """
        return self in self._constants._loopback_network

    @property
    def is_link_local(self):
        """Test if the address is reserved for link-local.

        Returns:
            A boolean, True if the address is link-local per RFC 3927.

        """
        return self in self._constants._linklocal_network


class IPv4Interface(IPv4Address):

    def __init__(self, address):
        if isinstance(address, (bytes, _compat_int_types)):
            IPv4Address.__init__(self, address)
            self.network = IPv4Network(self._ip)
            self._prefixlen = self._max_prefixlen
            return

        if isinstance(address, tuple):
            IPv4Address.__init__(self, address[0])
            if len(address) > 1:
                self._prefixlen = int(address[1])
            else:
                self._prefixlen = self._max_prefixlen

            self.network = IPv4Network(address, strict=False)
            self.netmask = self.network.netmask
            self.hostmask = self.network.hostmask
            return

        addr = _split_optional_netmask(address)
        IPv4Address.__init__(self, addr[0])

        self.network = IPv4Network(address, strict=False)
        self._prefixlen = self.network._prefixlen

        self.netmask = self.network.netmask
        self.hostmask = self.network.hostmask

    def __str__(self):
        return '%s/%d' % (self._string_from_ip_int(self._ip),
                          self.network.prefixlen)

    def __eq__(self, other):
        address_equal = IPv4Address.__eq__(self, other)
        if not address_equal or address_equal is NotImplemented:
            return address_equal
        try:
            return self.network == other.network
        except AttributeError:
            # An interface with an associated network is NOT the
            # same as an unassociated address. That's why the hash
            # takes the extra info into account.
            return False

    def __lt__(self, other):
        address_less = IPv4Address.__lt__(self, other)
        if address_less is NotImplemented:
            return NotImplemented
        try:
            return (self.network < other.network or
                    self.network == other.network and address_less)
        except AttributeError:
            # We *do* allow addresses and interfaces to be sorted. The
            # unassociated address is considered less than all interfaces.
            return False

    def __hash__(self):
        return self._ip ^ self._prefixlen ^ int(self.network.network_address)

    __reduce__ = _IPAddressBase.__reduce__

    @property
    def ip(self):
        return IPv4Address(self._ip)

    @property
    def with_prefixlen(self):
        return '%s/%s' % (self._string_from_ip_int(self._ip),
                          self._prefixlen)

    @property
    def with_netmask(self):
        return '%s/%s' % (self._string_from_ip_int(self._ip),
                          self.netmask)

    @property
    def with_hostmask(self):
        return '%s/%s' % (self._string_from_ip_int(self._ip),
                          self.hostmask)


class IPv4Network(_BaseV4, _BaseNetwork):

    """This class represents and manipulates 32-bit IPv4 network + addresses..

    Attributes: [examples for IPv4Network('192.0.2.0/27')]
        .network_address: IPv4Address('192.0.2.0')
        .hostmask: IPv4Address('0.0.0.31')
        .broadcast_address: IPv4Address('192.0.2.32')
        .netmask: IPv4Address('255.255.255.224')
        .prefixlen: 27

    """
    # Class to use when creating address objects
    _address_class = IPv4Address

    def __init__(self, address, strict=True):

        """Instantiate a new IPv4 network object.

        Args:
            address: A string or integer representing the IP [& network].
              '192.0.2.0/24'
              '192.0.2.0/255.255.255.0'
              '192.0.0.2/0.0.0.255'
              are all functionally the same in IPv4. Similarly,
              '192.0.2.1'
              '192.0.2.1/255.255.255.255'
              '192.0.2.1/32'
              are also functionally equivalent. That is to say, failing to
              provide a subnetmask will create an object with a mask of /32.

              If the mask (portion after the / in the argument) is given in
              dotted quad form, it is treated as a netmask if it starts with a
              non-zero field (e.g. /255.0.0.0 == /8) and as a hostmask if it
              starts with a zero field (e.g. 0.255.255.255 == /8), with the
              single exception of an all-zero mask which is treated as a
              netmask == /0. If no mask is given, a default of /32 is used.

              Additionally, an integer can be passed, so
              IPv4Network('192.0.2.1') == IPv4Network(3221225985)
              or, more generally
              IPv4Interface(int(IPv4Interface('192.0.2.1'))) ==
                IPv4Interface('192.0.2.1')

        Raises:
            AddressValueError: If ipaddress isn't a valid IPv4 address.
            NetmaskValueError: If the netmask isn't valid for
              an IPv4 address.
            ValueError: If strict is True and a network address is not
              supplied.

        """
        _BaseNetwork.__init__(self, address)

        # Constructing from a packed address or integer
        if isinstance(address, (_compat_int_types, bytes)):
            self.network_address = IPv4Address(address)
            self.netmask, self._prefixlen = self._make_netmask(
                self._max_prefixlen)
            # fixme: address/network test here.
            return

        if isinstance(address, tuple):
            if len(address) > 1:
                arg = address[1]
            else:
                # We weren't given an address[1]
                arg = self._max_prefixlen
            self.network_address = IPv4Address(address[0])
            self.netmask, self._prefixlen = self._make_netmask(arg)
            packed = int(self.network_address)
            if packed & int(self.netmask) != packed:
                if strict:
                    raise ValueError('%s has host bits set' % self)
                else:
                    self.network_address = IPv4Address(packed &
                                                       int(self.netmask))
            return

        # Assume input argument to be string or any object representation
        # which converts into a formatted IP prefix string.
        addr = _split_optional_netmask(address)
        self.network_address = IPv4Address(self._ip_int_from_string(addr[0]))

        if len(addr) == 2:
            arg = addr[1]
        else:
            arg = self._max_prefixlen
        self.netmask, self._prefixlen = self._make_netmask(arg)

        if strict:
            if (IPv4Address(int(self.network_address) & int(self.netmask)) !=
                    self.network_address):
                raise ValueError('%s has host bits set' % self)
        self.network_address = IPv4Address(int(self.network_address) &
                                           int(self.netmask))

        if self._prefixlen == (self._max_prefixlen - 1):
            self.hosts = self.__iter__

    @property
    def is_global(self):
        """Test if this address is allocated for public networks.

        Returns:
            A boolean, True if the address is not reserved per
            iana-ipv4-special-registry.

        """
        return (not (self.network_address in IPv4Network('100.64.0.0/10') and
                self.broadcast_address in IPv4Network('100.64.0.0/10')) and
                not self.is_private)


class _IPv4Constants(object):

    _linklocal_network = IPv4Network('169.254.0.0/16')

    _loopback_network = IPv4Network('127.0.0.0/8')

    _multicast_network = IPv4Network('224.0.0.0/4')

    _public_network = IPv4Network('100.64.0.0/10')

    _private_networks = [
        IPv4Network('0.0.0.0/8'),
        IPv4Network('10.0.0.0/8'),
        IPv4Network('127.0.0.0/8'),
        IPv4Network('169.254.0.0/16'),
        IPv4Network('172.16.0.0/12'),
        IPv4Network('192.0.0.0/29'),
        IPv4Network('192.0.0.170/31'),
        IPv4Network('192.0.2.0/24'),
        IPv4Network('192.168.0.0/16'),
        IPv4Network('198.18.0.0/15'),
        IPv4Network('198.51.100.0/24'),
        IPv4Network('203.0.113.0/24'),
        IPv4Network('240.0.0.0/4'),
        IPv4Network('255.255.255.255/32'),
    ]

    _reserved_network = IPv4Network('240.0.0.0/4')

    _unspecified_address = IPv4Address('0.0.0.0')


IPv4Address._constants = _IPv4Constants


class _BaseV6(object):

    """Base IPv6 object.

    The following methods are used by IPv6 objects in both single IP
    addresses and networks.

    """

    __slots__ = ()
    _version = 6
    _ALL_ONES = (2 ** IPV6LENGTH) - 1
    _HEXTET_COUNT = 8
    _HEX_DIGITS = frozenset('0123456789ABCDEFabcdef')
    _max_prefixlen = IPV6LENGTH

    # There are only a bunch of valid v6 netmasks, so we cache them all
    # when constructed (see _make_netmask()).
    _netmask_cache = {}

    @classmethod
    def _make_netmask(cls, arg):
        """Make a (netmask, prefix_len) tuple from the given argument.

        Argument can be:
        - an integer (the prefix length)
        - a string representing the prefix length (e.g. "24")
        - a string representing the prefix netmask (e.g. "255.255.255.0")
        """
        if arg not in cls._netmask_cache:
            if isinstance(arg, _compat_int_types):
                prefixlen = arg
            else:
                prefixlen = cls._prefix_from_prefix_string(arg)
            netmask = IPv6Address(cls._ip_int_from_prefix(prefixlen))
            cls._netmask_cache[arg] = netmask, prefixlen
        return cls._netmask_cache[arg]

    @classmethod
    def _ip_int_from_string(cls, ip_str):
        """Turn an IPv6 ip_str into an integer.

        Args:
            ip_str: A string, the IPv6 ip_str.

        Returns:
            An int, the IPv6 address

        Raises:
            AddressValueError: if ip_str isn't a valid IPv6 Address.

        """
        if not ip_str:
            raise AddressValueError('Address cannot be empty')

        parts = ip_str.split(':')

        # An IPv6 address needs at least 2 colons (3 parts).
        _min_parts = 3
        if len(parts) < _min_parts:
            msg = "At least %d parts expected in %r" % (_min_parts, ip_str)
            raise AddressValueError(msg)

        # If the address has an IPv4-style suffix, convert it to hexadecimal.
        if '.' in parts[-1]:
            try:
                ipv4_int = IPv4Address(parts.pop())._ip
            except AddressValueError as exc:
                raise AddressValueError("%s in %r" % (exc, ip_str))
            parts.append('%x' % ((ipv4_int >> 16) & 0xFFFF))
            parts.append('%x' % (ipv4_int & 0xFFFF))

        # An IPv6 address can't have more than 8 colons (9 parts).
        # The extra colon comes from using the "::" notation for a single
        # leading or trailing zero part.
        _max_parts = cls._HEXTET_COUNT + 1
        if len(parts) > _max_parts:
            msg = "At most %d colons permitted in %r" % (
                _max_parts - 1, ip_str)
            raise AddressValueError(msg)

        # Disregarding the endpoints, find '::' with nothing in between.
        # This indicates that a run of zeroes has been skipped.
        skip_index = None
        for i in _compat_range(1, len(parts) - 1):
            if not parts[i]:
                if skip_index is not None:
                    # Can't have more than one '::'
                    msg = "At most one '::' permitted in %r" % ip_str
                    raise AddressValueError(msg)
                skip_index = i

        # parts_hi is the number of parts to copy from above/before the '::'
        # parts_lo is the number of parts to copy from below/after the '::'
        if skip_index is not None:
            # If we found a '::', then check if it also covers the endpoints.
            parts_hi = skip_index
            parts_lo = len(parts) - skip_index - 1
            if not parts[0]:
                parts_hi -= 1
                if parts_hi:
                    msg = "Leading ':' only permitted as part of '::' in %r"
                    raise AddressValueError(msg % ip_str)  # ^: requires ^::
            if not parts[-1]:
                parts_lo -= 1
                if parts_lo:
                    msg = "Trailing ':' only permitted as part of '::' in %r"
                    raise AddressValueError(msg % ip_str)  # :$ requires ::$
            parts_skipped = cls._HEXTET_COUNT - (parts_hi + parts_lo)
            if parts_skipped < 1:
                msg = "Expected at most %d other parts with '::' in %r"
                raise AddressValueError(msg % (cls._HEXTET_COUNT - 1, ip_str))
        else:
            # Otherwise, allocate the entire address to parts_hi.  The
            # endpoints could still be empty, but _parse_hextet() will check
            # for that.
            if len(parts) != cls._HEXTET_COUNT:
                msg = "Exactly %d parts expected without '::' in %r"
                raise AddressValueError(msg % (cls._HEXTET_COUNT, ip_str))
            if not parts[0]:
                msg = "Leading ':' only permitted as part of '::' in %r"
                raise AddressValueError(msg % ip_str)  # ^: requires ^::
            if not parts[-1]:
                msg = "Trailing ':' only permitted as part of '::' in %r"
                raise AddressValueError(msg % ip_str)  # :$ requires ::$
            parts_hi = len(parts)
            parts_lo = 0
            parts_skipped = 0

        try:
            # Now, parse the hextets into a 128-bit integer.
            ip_int = 0
            for i in range(parts_hi):
                ip_int <<= 16
                ip_int |= cls._parse_hextet(parts[i])
            ip_int <<= 16 * parts_skipped
            for i in range(-parts_lo, 0):
                ip_int <<= 16
                ip_int |= cls._parse_hextet(parts[i])
            return ip_int
        except ValueError as exc:
            raise AddressValueError("%s in %r" % (exc, ip_str))

    @classmethod
    def _parse_hextet(cls, hextet_str):
        """Convert an IPv6 hextet string into an integer.

        Args:
            hextet_str: A string, the number to parse.

        Returns:
            The hextet as an integer.

        Raises:
            ValueError: if the input isn't strictly a hex number from
              [0..FFFF].

        """
        # Whitelist the characters, since int() allows a lot of bizarre stuff.
        if not cls._HEX_DIGITS.issuperset(hextet_str):
            raise ValueError("Only hex digits permitted in %r" % hextet_str)
        # We do the length check second, since the invalid character error
        # is likely to be more informative for the user
        if len(hextet_str) > 4:
            msg = "At most 4 characters permitted in %r"
            raise ValueError(msg % hextet_str)
        # Length check means we can skip checking the integer value
        return int(hextet_str, 16)

    @classmethod
    def _compress_hextets(cls, hextets):
        """Compresses a list of hextets.

        Compresses a list of strings, replacing the longest continuous
        sequence of "0" in the list with "" and adding empty strings at
        the beginning or at the end of the string such that subsequently
        calling ":".join(hextets) will produce the compressed version of
        the IPv6 address.

        Args:
            hextets: A list of strings, the hextets to compress.

        Returns:
            A list of strings.

        """
        best_doublecolon_start = -1
        best_doublecolon_len = 0
        doublecolon_start = -1
        doublecolon_len = 0
        for index, hextet in enumerate(hextets):
            if hextet == '0':
                doublecolon_len += 1
                if doublecolon_start == -1:
                    # Start of a sequence of zeros.
                    doublecolon_start = index
                if doublecolon_len > best_doublecolon_len:
                    # This is the longest sequence of zeros so far.
                    best_doublecolon_len = doublecolon_len
                    best_doublecolon_start = doublecolon_start
            else:
                doublecolon_len = 0
                doublecolon_start = -1

        if best_doublecolon_len > 1:
            best_doublecolon_end = (best_doublecolon_start +
                                    best_doublecolon_len)
            # For zeros at the end of the address.
            if best_doublecolon_end == len(hextets):
                hextets += ['']
            hextets[best_doublecolon_start:best_doublecolon_end] = ['']
            # For zeros at the beginning of the address.
            if best_doublecolon_start == 0:
                hextets = [''] + hextets

        return hextets

    @classmethod
    def _string_from_ip_int(cls, ip_int=None):
        """Turns a 128-bit integer into hexadecimal notation.

        Args:
            ip_int: An integer, the IP address.

        Returns:
            A string, the hexadecimal representation of the address.

        Raises:
            ValueError: The address is bigger than 128 bits of all ones.

        """
        if ip_int is None:
            ip_int = int(cls._ip)

        if ip_int > cls._ALL_ONES:
            raise ValueError('IPv6 address is too large')

        hex_str = '%032x' % ip_int
        hextets = ['%x' % int(hex_str[x:x + 4], 16) for x in range(0, 32, 4)]

        hextets = cls._compress_hextets(hextets)
        return ':'.join(hextets)

    def _explode_shorthand_ip_string(self):
        """Expand a shortened IPv6 address.

        Args:
            ip_str: A string, the IPv6 address.

        Returns:
            A string, the expanded IPv6 address.

        """
        if isinstance(self, IPv6Network):
            ip_str = _compat_str(self.network_address)
        elif isinstance(self, IPv6Interface):
            ip_str = _compat_str(self.ip)
        else:
            ip_str = _compat_str(self)

        ip_int = self._ip_int_from_string(ip_str)
        hex_str = '%032x' % ip_int
        parts = [hex_str[x:x + 4] for x in range(0, 32, 4)]
        if isinstance(self, (_BaseNetwork, IPv6Interface)):
            return '%s/%d' % (':'.join(parts), self._prefixlen)
        return ':'.join(parts)

    def _reverse_pointer(self):
        """Return the reverse DNS pointer name for the IPv6 address.

        This implements the method described in RFC3596 2.5.

        """
        reverse_chars = self.exploded[::-1].replace(':', '')
        return '.'.join(reverse_chars) + '.ip6.arpa'

    @property
    def max_prefixlen(self):
        return self._max_prefixlen

    @property
    def version(self):
        return self._version


class IPv6Address(_BaseV6, _BaseAddress):

    """Represent and manipulate single IPv6 Addresses."""

    __slots__ = ('_ip', '__weakref__')

    def __init__(self, address):
        """Instantiate a new IPv6 address object.

        Args:
            address: A string or integer representing the IP

              Additionally, an integer can be passed, so
              IPv6Address('2001:db8::') ==
                IPv6Address(42540766411282592856903984951653826560)
              or, more generally
              IPv6Address(int(IPv6Address('2001:db8::'))) ==
                IPv6Address('2001:db8::')

        Raises:
            AddressValueError: If address isn't a valid IPv6 address.

        """
        # Efficient constructor from integer.
        if isinstance(address, _compat_int_types):
            self._check_int_address(address)
            self._ip = address
            return

        # Constructing from a packed address
        if isinstance(address, bytes):
            self._check_packed_address(address, 16)
            bvs = _compat_bytes_to_byte_vals(address)
            self._ip = _compat_int_from_byte_vals(bvs, 'big')
            return

        # Assume input argument to be string or any object representation
        # which converts into a formatted IP string.
        addr_str = _compat_str(address)
        if '/' in addr_str:
            raise AddressValueError("Unexpected '/' in %r" % address)
        self._ip = self._ip_int_from_string(addr_str)

    @property
    def packed(self):
        """The binary representation of this address."""
        return v6_int_to_packed(self._ip)

    @property
    def is_multicast(self):
        """Test if the address is reserved for multicast use.

        Returns:
            A boolean, True if the address is a multicast address.
            See RFC 2373 2.7 for details.

        """
        return self in self._constants._multicast_network

    @property
    def is_reserved(self):
        """Test if the address is otherwise IETF reserved.

        Returns:
            A boolean, True if the address is within one of the
            reserved IPv6 Network ranges.

        """
        return any(self in x for x in self._constants._reserved_networks)

    @property
    def is_link_local(self):
        """Test if the address is reserved for link-local.

        Returns:
            A boolean, True if the address is reserved per RFC 4291.

        """
        return self in self._constants._linklocal_network

    @property
    def is_site_local(self):
        """Test if the address is reserved for site-local.

        Note that the site-local address space has been deprecated by RFC 3879.
        Use is_private to test if this address is in the space of unique local
        addresses as defined by RFC 4193.

        Returns:
            A boolean, True if the address is reserved per RFC 3513 2.5.6.

        """
        return self in self._constants._sitelocal_network

    @property
    def is_private(self):
        """Test if this address is allocated for private networks.

        Returns:
            A boolean, True if the address is reserved per
            iana-ipv6-special-registry.

        """
        return any(self in net for net in self._constants._private_networks)

    @property
    def is_global(self):
        """Test if this address is allocated for public networks.

        Returns:
            A boolean, true if the address is not reserved per
            iana-ipv6-special-registry.

        """
        return not self.is_private

    @property
    def is_unspecified(self):
        """Test if the address is unspecified.

        Returns:
            A boolean, True if this is the unspecified address as defined in
            RFC 2373 2.5.2.

        """
        return self._ip == 0

    @property
    def is_loopback(self):
        """Test if the address is a loopback address.

        Returns:
            A boolean, True if the address is a loopback address as defined in
            RFC 2373 2.5.3.

        """
        return self._ip == 1

    @property
    def ipv4_mapped(self):
        """Return the IPv4 mapped address.

        Returns:
            If the IPv6 address is a v4 mapped address, return the
            IPv4 mapped address. Return None otherwise.

        """
        if (self._ip >> 32) != 0xFFFF:
            return None
        return IPv4Address(self._ip & 0xFFFFFFFF)

    @property
    def teredo(self):
        """Tuple of embedded teredo IPs.

        Returns:
            Tuple of the (server, client) IPs or None if the address
            doesn't appear to be a teredo address (doesn't start with
            2001::/32)

        """
        if (self._ip >> 96) != 0x20010000:
            return None
        return (IPv4Address((self._ip >> 64) & 0xFFFFFFFF),
                IPv4Address(~self._ip & 0xFFFFFFFF))

    @property
    def sixtofour(self):
        """Return the IPv4 6to4 embedded address.

        Returns:
            The IPv4 6to4-embedded address if present or None if the
            address doesn't appear to contain a 6to4 embedded address.

        """
        if (self._ip >> 112) != 0x2002:
            return None
        return IPv4Address((self._ip >> 80) & 0xFFFFFFFF)


class IPv6Interface(IPv6Address):

    def __init__(self, address):
        if isinstance(address, (bytes, _compat_int_types)):
            IPv6Address.__init__(self, address)
            self.network = IPv6Network(self._ip)
            self._prefixlen = self._max_prefixlen
            return
        if isinstance(address, tuple):
            IPv6Address.__init__(self, address[0])
            if len(address) > 1:
                self._prefixlen = int(address[1])
            else:
                self._prefixlen = self._max_prefixlen
            self.network = IPv6Network(address, strict=False)
            self.netmask = self.network.netmask
            self.hostmask = self.network.hostmask
            return

        addr = _split_optional_netmask(address)
        IPv6Address.__init__(self, addr[0])
        self.network = IPv6Network(address, strict=False)
        self.netmask = self.network.netmask
        self._prefixlen = self.network._prefixlen
        self.hostmask = self.network.hostmask

    def __str__(self):
        return '%s/%d' % (self._string_from_ip_int(self._ip),
                          self.network.prefixlen)

    def __eq__(self, other):
        address_equal = IPv6Address.__eq__(self, other)
        if not address_equal or address_equal is NotImplemented:
            return address_equal
        try:
            return self.network == other.network
        except AttributeError:
            # An interface with an associated network is NOT the
            # same as an unassociated address. That's why the hash
            # takes the extra info into account.
            return False

    def __lt__(self, other):
        address_less = IPv6Address.__lt__(self, other)
        if address_less is NotImplemented:
            return NotImplemented
        try:
            return (self.network < other.network or
                    self.network == other.network and address_less)
        except AttributeError:
            # We *do* allow addresses and interfaces to be sorted. The
            # unassociated address is considered less than all interfaces.
            return False

    def __hash__(self):
        return self._ip ^ self._prefixlen ^ int(self.network.network_address)

    __reduce__ = _IPAddressBase.__reduce__

    @property
    def ip(self):
        return IPv6Address(self._ip)

    @property
    def with_prefixlen(self):
        return '%s/%s' % (self._string_from_ip_int(self._ip),
                          self._prefixlen)

    @property
    def with_netmask(self):
        return '%s/%s' % (self._string_from_ip_int(self._ip),
                          self.netmask)

    @property
    def with_hostmask(self):
        return '%s/%s' % (self._string_from_ip_int(self._ip),
                          self.hostmask)

    @property
    def is_unspecified(self):
        return self._ip == 0 and self.network.is_unspecified

    @property
    def is_loopback(self):
        return self._ip == 1 and self.network.is_loopback


class IPv6Network(_BaseV6, _BaseNetwork):

    """This class represents and manipulates 128-bit IPv6 networks.

    Attributes: [examples for IPv6('2001:db8::1000/124')]
        .network_address: IPv6Address('2001:db8::1000')
        .hostmask: IPv6Address('::f')
        .broadcast_address: IPv6Address('2001:db8::100f')
        .netmask: IPv6Address('ffff:ffff:ffff:ffff:ffff:ffff:ffff:fff0')
        .prefixlen: 124

    """

    # Class to use when creating address objects
    _address_class = IPv6Address

    def __init__(self, address, strict=True):
        """Instantiate a new IPv6 Network object.

        Args:
            address: A string or integer representing the IPv6 network or the
              IP and prefix/netmask.
              '2001:db8::/128'
              '2001:db8:0000:0000:0000:0000:0000:0000/128'
              '2001:db8::'
              are all functionally the same in IPv6.  That is to say,
              failing to provide a subnetmask will create an object with
              a mask of /128.

              Additionally, an integer can be passed, so
              IPv6Network('2001:db8::') ==
                IPv6Network(42540766411282592856903984951653826560)
              or, more generally
              IPv6Network(int(IPv6Network('2001:db8::'))) ==
                IPv6Network('2001:db8::')

            strict: A boolean. If true, ensure that we have been passed
              A true network address, eg, 2001:db8::1000/124 and not an
              IP address on a network, eg, 2001:db8::1/124.

        Raises:
            AddressValueError: If address isn't a valid IPv6 address.
            NetmaskValueError: If the netmask isn't valid for
              an IPv6 address.
            ValueError: If strict was True and a network address was not
              supplied.

        """
        _BaseNetwork.__init__(self, address)

        # Efficient constructor from integer or packed address
        if isinstance(address, (bytes, _compat_int_types)):
            self.network_address = IPv6Address(address)
            self.netmask, self._prefixlen = self._make_netmask(
                self._max_prefixlen)
            return

        if isinstance(address, tuple):
            if len(address) > 1:
                arg = address[1]
            else:
                arg = self._max_prefixlen
            self.netmask, self._prefixlen = self._make_netmask(arg)
            self.network_address = IPv6Address(address[0])
            packed = int(self.network_address)
            if packed & int(self.netmask) != packed:
                if strict:
                    raise ValueError('%s has host bits set' % self)
                else:
                    self.network_address = IPv6Address(packed &
                                                       int(self.netmask))
            return

        # Assume input argument to be string or any object representation
        # which converts into a formatted IP prefix string.
        addr = _split_optional_netmask(address)

        self.network_address = IPv6Address(self._ip_int_from_string(addr[0]))

        if len(addr) == 2:
            arg = addr[1]
        else:
            arg = self._max_prefixlen
        self.netmask, self._prefixlen = self._make_netmask(arg)

        if strict:
            if (IPv6Address(int(self.network_address) & int(self.netmask)) !=
                    self.network_address):
                raise ValueError('%s has host bits set' % self)
        self.network_address = IPv6Address(int(self.network_address) &
                                           int(self.netmask))

        if self._prefixlen == (self._max_prefixlen - 1):
            self.hosts = self.__iter__

    def hosts(self):
        """Generate Iterator over usable hosts in a network.

          This is like __iter__ except it doesn't return the
          Subnet-Router anycast address.

        """
        network = int(self.network_address)
        broadcast = int(self.broadcast_address)
        for x in _compat_range(network + 1, broadcast + 1):
            yield self._address_class(x)

    @property
    def is_site_local(self):
        """Test if the address is reserved for site-local.

        Note that the site-local address space has been deprecated by RFC 3879.
        Use is_private to test if this address is in the space of unique local
        addresses as defined by RFC 4193.

        Returns:
            A boolean, True if the address is reserved per RFC 3513 2.5.6.

        """
        return (self.network_address.is_site_local and
                self.broadcast_address.is_site_local)


class _IPv6Constants(object):

    _linklocal_network = IPv6Network('fe80::/10')

    _multicast_network = IPv6Network('ff00::/8')

    _private_networks = [
        IPv6Network('::1/128'),
        IPv6Network('::/128'),
        IPv6Network('::ffff:0:0/96'),
        IPv6Network('100::/64'),
        IPv6Network('2001::/23'),
        IPv6Network('2001:2::/48'),
        IPv6Network('2001:db8::/32'),
        IPv6Network('2001:10::/28'),
        IPv6Network('fc00::/7'),
        IPv6Network('fe80::/10'),
    ]

    _reserved_networks = [
        IPv6Network('::/8'), IPv6Network('100::/8'),
        IPv6Network('200::/7'), IPv6Network('400::/6'),
        IPv6Network('800::/5'), IPv6Network('1000::/4'),
        IPv6Network('4000::/3'), IPv6Network('6000::/3'),
        IPv6Network('8000::/3'), IPv6Network('A000::/3'),
        IPv6Network('C000::/3'), IPv6Network('E000::/4'),
        IPv6Network('F000::/5'), IPv6Network('F800::/6'),
        IPv6Network('FE00::/9'),
    ]

    _sitelocal_network = IPv6Network('fec0::/10')


IPv6Address._constants = _IPv6Constants
site-packages/pip/_vendor/idna/codec.py000064400000006343151464452710014071 0ustar00from .core import encode, decode, alabel, ulabel, IDNAError
import codecs
import re

_unicode_dots_re = re.compile(u'[\u002e\u3002\uff0e\uff61]')

class Codec(codecs.Codec):

    def encode(self, data, errors='strict'):

        if errors != 'strict':
            raise IDNAError("Unsupported error handling \"{0}\"".format(errors))

        if not data:
            return "", 0

        return encode(data), len(data)

    def decode(self, data, errors='strict'):

        if errors != 'strict':
            raise IDNAError("Unsupported error handling \"{0}\"".format(errors))

        if not data:
            return u"", 0

        return decode(data), len(data)

class IncrementalEncoder(codecs.BufferedIncrementalEncoder):
    def _buffer_encode(self, data, errors, final):
        if errors != 'strict':
            raise IDNAError("Unsupported error handling \"{0}\"".format(errors))

        if not data:
            return ("", 0)

        labels = _unicode_dots_re.split(data)
        trailing_dot = u''
        if labels:
            if not labels[-1]:
                trailing_dot = '.'
                del labels[-1]
            elif not final:
                # Keep potentially unfinished label until the next call
                del labels[-1]
                if labels:
                    trailing_dot = '.'

        result = []
        size = 0
        for label in labels:
            result.append(alabel(label))
            if size:
                size += 1
            size += len(label)

        # Join with U+002E
        result = ".".join(result) + trailing_dot
        size += len(trailing_dot)
        return (result, size)

class IncrementalDecoder(codecs.BufferedIncrementalDecoder):
    def _buffer_decode(self, data, errors, final):
        if errors != 'strict':
            raise IDNAError("Unsupported error handling \"{0}\"".format(errors))

        if not data:
            return (u"", 0)

        # IDNA allows decoding to operate on Unicode strings, too.
        if isinstance(data, unicode):
            labels = _unicode_dots_re.split(data)
        else:
            # Must be ASCII string
            data = str(data)
            unicode(data, "ascii")
            labels = data.split(".")

        trailing_dot = u''
        if labels:
            if not labels[-1]:
                trailing_dot = u'.'
                del labels[-1]
            elif not final:
                # Keep potentially unfinished label until the next call
                del labels[-1]
                if labels:
                    trailing_dot = u'.'

        result = []
        size = 0
        for label in labels:
            result.append(ulabel(label))
            if size:
                size += 1
            size += len(label)

        result = u".".join(result) + trailing_dot
        size += len(trailing_dot)
        return (result, size)


class StreamWriter(Codec, codecs.StreamWriter):
    pass

class StreamReader(Codec, codecs.StreamReader):
    pass

def getregentry():
    return codecs.CodecInfo(
        name='idna',
        encode=Codec().encode,
        decode=Codec().decode,
        incrementalencoder=IncrementalEncoder,
        incrementaldecoder=IncrementalDecoder,
        streamwriter=StreamWriter,
        streamreader=StreamReader,
    )
site-packages/pip/_vendor/idna/package_data.py000064400000000025151464452710015367 0ustar00__version__ = '2.8'

site-packages/pip/_vendor/idna/uts46data.py000064400000603224151464452710014634 0ustar00# This file is automatically generated by tools/idna-data
# vim: set fileencoding=utf-8 :

"""IDNA Mapping Table from UTS46."""


__version__ = "11.0.0"
def _seg_0():
    return [
    (0x0, '3'),
    (0x1, '3'),
    (0x2, '3'),
    (0x3, '3'),
    (0x4, '3'),
    (0x5, '3'),
    (0x6, '3'),
    (0x7, '3'),
    (0x8, '3'),
    (0x9, '3'),
    (0xA, '3'),
    (0xB, '3'),
    (0xC, '3'),
    (0xD, '3'),
    (0xE, '3'),
    (0xF, '3'),
    (0x10, '3'),
    (0x11, '3'),
    (0x12, '3'),
    (0x13, '3'),
    (0x14, '3'),
    (0x15, '3'),
    (0x16, '3'),
    (0x17, '3'),
    (0x18, '3'),
    (0x19, '3'),
    (0x1A, '3'),
    (0x1B, '3'),
    (0x1C, '3'),
    (0x1D, '3'),
    (0x1E, '3'),
    (0x1F, '3'),
    (0x20, '3'),
    (0x21, '3'),
    (0x22, '3'),
    (0x23, '3'),
    (0x24, '3'),
    (0x25, '3'),
    (0x26, '3'),
    (0x27, '3'),
    (0x28, '3'),
    (0x29, '3'),
    (0x2A, '3'),
    (0x2B, '3'),
    (0x2C, '3'),
    (0x2D, 'V'),
    (0x2E, 'V'),
    (0x2F, '3'),
    (0x30, 'V'),
    (0x31, 'V'),
    (0x32, 'V'),
    (0x33, 'V'),
    (0x34, 'V'),
    (0x35, 'V'),
    (0x36, 'V'),
    (0x37, 'V'),
    (0x38, 'V'),
    (0x39, 'V'),
    (0x3A, '3'),
    (0x3B, '3'),
    (0x3C, '3'),
    (0x3D, '3'),
    (0x3E, '3'),
    (0x3F, '3'),
    (0x40, '3'),
    (0x41, 'M', u'a'),
    (0x42, 'M', u'b'),
    (0x43, 'M', u'c'),
    (0x44, 'M', u'd'),
    (0x45, 'M', u'e'),
    (0x46, 'M', u'f'),
    (0x47, 'M', u'g'),
    (0x48, 'M', u'h'),
    (0x49, 'M', u'i'),
    (0x4A, 'M', u'j'),
    (0x4B, 'M', u'k'),
    (0x4C, 'M', u'l'),
    (0x4D, 'M', u'm'),
    (0x4E, 'M', u'n'),
    (0x4F, 'M', u'o'),
    (0x50, 'M', u'p'),
    (0x51, 'M', u'q'),
    (0x52, 'M', u'r'),
    (0x53, 'M', u's'),
    (0x54, 'M', u't'),
    (0x55, 'M', u'u'),
    (0x56, 'M', u'v'),
    (0x57, 'M', u'w'),
    (0x58, 'M', u'x'),
    (0x59, 'M', u'y'),
    (0x5A, 'M', u'z'),
    (0x5B, '3'),
    (0x5C, '3'),
    (0x5D, '3'),
    (0x5E, '3'),
    (0x5F, '3'),
    (0x60, '3'),
    (0x61, 'V'),
    (0x62, 'V'),
    (0x63, 'V'),
    ]

def _seg_1():
    return [
    (0x64, 'V'),
    (0x65, 'V'),
    (0x66, 'V'),
    (0x67, 'V'),
    (0x68, 'V'),
    (0x69, 'V'),
    (0x6A, 'V'),
    (0x6B, 'V'),
    (0x6C, 'V'),
    (0x6D, 'V'),
    (0x6E, 'V'),
    (0x6F, 'V'),
    (0x70, 'V'),
    (0x71, 'V'),
    (0x72, 'V'),
    (0x73, 'V'),
    (0x74, 'V'),
    (0x75, 'V'),
    (0x76, 'V'),
    (0x77, 'V'),
    (0x78, 'V'),
    (0x79, 'V'),
    (0x7A, 'V'),
    (0x7B, '3'),
    (0x7C, '3'),
    (0x7D, '3'),
    (0x7E, '3'),
    (0x7F, '3'),
    (0x80, 'X'),
    (0x81, 'X'),
    (0x82, 'X'),
    (0x83, 'X'),
    (0x84, 'X'),
    (0x85, 'X'),
    (0x86, 'X'),
    (0x87, 'X'),
    (0x88, 'X'),
    (0x89, 'X'),
    (0x8A, 'X'),
    (0x8B, 'X'),
    (0x8C, 'X'),
    (0x8D, 'X'),
    (0x8E, 'X'),
    (0x8F, 'X'),
    (0x90, 'X'),
    (0x91, 'X'),
    (0x92, 'X'),
    (0x93, 'X'),
    (0x94, 'X'),
    (0x95, 'X'),
    (0x96, 'X'),
    (0x97, 'X'),
    (0x98, 'X'),
    (0x99, 'X'),
    (0x9A, 'X'),
    (0x9B, 'X'),
    (0x9C, 'X'),
    (0x9D, 'X'),
    (0x9E, 'X'),
    (0x9F, 'X'),
    (0xA0, '3', u' '),
    (0xA1, 'V'),
    (0xA2, 'V'),
    (0xA3, 'V'),
    (0xA4, 'V'),
    (0xA5, 'V'),
    (0xA6, 'V'),
    (0xA7, 'V'),
    (0xA8, '3', u' ̈'),
    (0xA9, 'V'),
    (0xAA, 'M', u'a'),
    (0xAB, 'V'),
    (0xAC, 'V'),
    (0xAD, 'I'),
    (0xAE, 'V'),
    (0xAF, '3', u' ̄'),
    (0xB0, 'V'),
    (0xB1, 'V'),
    (0xB2, 'M', u'2'),
    (0xB3, 'M', u'3'),
    (0xB4, '3', u' ́'),
    (0xB5, 'M', u'μ'),
    (0xB6, 'V'),
    (0xB7, 'V'),
    (0xB8, '3', u' ̧'),
    (0xB9, 'M', u'1'),
    (0xBA, 'M', u'o'),
    (0xBB, 'V'),
    (0xBC, 'M', u'1⁄4'),
    (0xBD, 'M', u'1⁄2'),
    (0xBE, 'M', u'3⁄4'),
    (0xBF, 'V'),
    (0xC0, 'M', u'à'),
    (0xC1, 'M', u'á'),
    (0xC2, 'M', u'â'),
    (0xC3, 'M', u'ã'),
    (0xC4, 'M', u'ä'),
    (0xC5, 'M', u'å'),
    (0xC6, 'M', u'æ'),
    (0xC7, 'M', u'ç'),
    ]

def _seg_2():
    return [
    (0xC8, 'M', u'è'),
    (0xC9, 'M', u'é'),
    (0xCA, 'M', u'ê'),
    (0xCB, 'M', u'ë'),
    (0xCC, 'M', u'ì'),
    (0xCD, 'M', u'í'),
    (0xCE, 'M', u'î'),
    (0xCF, 'M', u'ï'),
    (0xD0, 'M', u'ð'),
    (0xD1, 'M', u'ñ'),
    (0xD2, 'M', u'ò'),
    (0xD3, 'M', u'ó'),
    (0xD4, 'M', u'ô'),
    (0xD5, 'M', u'õ'),
    (0xD6, 'M', u'ö'),
    (0xD7, 'V'),
    (0xD8, 'M', u'ø'),
    (0xD9, 'M', u'ù'),
    (0xDA, 'M', u'ú'),
    (0xDB, 'M', u'û'),
    (0xDC, 'M', u'ü'),
    (0xDD, 'M', u'ý'),
    (0xDE, 'M', u'þ'),
    (0xDF, 'D', u'ss'),
    (0xE0, 'V'),
    (0xE1, 'V'),
    (0xE2, 'V'),
    (0xE3, 'V'),
    (0xE4, 'V'),
    (0xE5, 'V'),
    (0xE6, 'V'),
    (0xE7, 'V'),
    (0xE8, 'V'),
    (0xE9, 'V'),
    (0xEA, 'V'),
    (0xEB, 'V'),
    (0xEC, 'V'),
    (0xED, 'V'),
    (0xEE, 'V'),
    (0xEF, 'V'),
    (0xF0, 'V'),
    (0xF1, 'V'),
    (0xF2, 'V'),
    (0xF3, 'V'),
    (0xF4, 'V'),
    (0xF5, 'V'),
    (0xF6, 'V'),
    (0xF7, 'V'),
    (0xF8, 'V'),
    (0xF9, 'V'),
    (0xFA, 'V'),
    (0xFB, 'V'),
    (0xFC, 'V'),
    (0xFD, 'V'),
    (0xFE, 'V'),
    (0xFF, 'V'),
    (0x100, 'M', u'ā'),
    (0x101, 'V'),
    (0x102, 'M', u'ă'),
    (0x103, 'V'),
    (0x104, 'M', u'ą'),
    (0x105, 'V'),
    (0x106, 'M', u'ć'),
    (0x107, 'V'),
    (0x108, 'M', u'ĉ'),
    (0x109, 'V'),
    (0x10A, 'M', u'ċ'),
    (0x10B, 'V'),
    (0x10C, 'M', u'č'),
    (0x10D, 'V'),
    (0x10E, 'M', u'ď'),
    (0x10F, 'V'),
    (0x110, 'M', u'đ'),
    (0x111, 'V'),
    (0x112, 'M', u'ē'),
    (0x113, 'V'),
    (0x114, 'M', u'ĕ'),
    (0x115, 'V'),
    (0x116, 'M', u'ė'),
    (0x117, 'V'),
    (0x118, 'M', u'ę'),
    (0x119, 'V'),
    (0x11A, 'M', u'ě'),
    (0x11B, 'V'),
    (0x11C, 'M', u'ĝ'),
    (0x11D, 'V'),
    (0x11E, 'M', u'ğ'),
    (0x11F, 'V'),
    (0x120, 'M', u'ġ'),
    (0x121, 'V'),
    (0x122, 'M', u'ģ'),
    (0x123, 'V'),
    (0x124, 'M', u'ĥ'),
    (0x125, 'V'),
    (0x126, 'M', u'ħ'),
    (0x127, 'V'),
    (0x128, 'M', u'ĩ'),
    (0x129, 'V'),
    (0x12A, 'M', u'ī'),
    (0x12B, 'V'),
    ]

def _seg_3():
    return [
    (0x12C, 'M', u'ĭ'),
    (0x12D, 'V'),
    (0x12E, 'M', u'į'),
    (0x12F, 'V'),
    (0x130, 'M', u'i̇'),
    (0x131, 'V'),
    (0x132, 'M', u'ij'),
    (0x134, 'M', u'ĵ'),
    (0x135, 'V'),
    (0x136, 'M', u'ķ'),
    (0x137, 'V'),
    (0x139, 'M', u'ĺ'),
    (0x13A, 'V'),
    (0x13B, 'M', u'ļ'),
    (0x13C, 'V'),
    (0x13D, 'M', u'ľ'),
    (0x13E, 'V'),
    (0x13F, 'M', u'l·'),
    (0x141, 'M', u'ł'),
    (0x142, 'V'),
    (0x143, 'M', u'ń'),
    (0x144, 'V'),
    (0x145, 'M', u'ņ'),
    (0x146, 'V'),
    (0x147, 'M', u'ň'),
    (0x148, 'V'),
    (0x149, 'M', u'ʼn'),
    (0x14A, 'M', u'ŋ'),
    (0x14B, 'V'),
    (0x14C, 'M', u'ō'),
    (0x14D, 'V'),
    (0x14E, 'M', u'ŏ'),
    (0x14F, 'V'),
    (0x150, 'M', u'ő'),
    (0x151, 'V'),
    (0x152, 'M', u'œ'),
    (0x153, 'V'),
    (0x154, 'M', u'ŕ'),
    (0x155, 'V'),
    (0x156, 'M', u'ŗ'),
    (0x157, 'V'),
    (0x158, 'M', u'ř'),
    (0x159, 'V'),
    (0x15A, 'M', u'ś'),
    (0x15B, 'V'),
    (0x15C, 'M', u'ŝ'),
    (0x15D, 'V'),
    (0x15E, 'M', u'ş'),
    (0x15F, 'V'),
    (0x160, 'M', u'š'),
    (0x161, 'V'),
    (0x162, 'M', u'ţ'),
    (0x163, 'V'),
    (0x164, 'M', u'ť'),
    (0x165, 'V'),
    (0x166, 'M', u'ŧ'),
    (0x167, 'V'),
    (0x168, 'M', u'ũ'),
    (0x169, 'V'),
    (0x16A, 'M', u'ū'),
    (0x16B, 'V'),
    (0x16C, 'M', u'ŭ'),
    (0x16D, 'V'),
    (0x16E, 'M', u'ů'),
    (0x16F, 'V'),
    (0x170, 'M', u'ű'),
    (0x171, 'V'),
    (0x172, 'M', u'ų'),
    (0x173, 'V'),
    (0x174, 'M', u'ŵ'),
    (0x175, 'V'),
    (0x176, 'M', u'ŷ'),
    (0x177, 'V'),
    (0x178, 'M', u'ÿ'),
    (0x179, 'M', u'ź'),
    (0x17A, 'V'),
    (0x17B, 'M', u'ż'),
    (0x17C, 'V'),
    (0x17D, 'M', u'ž'),
    (0x17E, 'V'),
    (0x17F, 'M', u's'),
    (0x180, 'V'),
    (0x181, 'M', u'ɓ'),
    (0x182, 'M', u'ƃ'),
    (0x183, 'V'),
    (0x184, 'M', u'ƅ'),
    (0x185, 'V'),
    (0x186, 'M', u'ɔ'),
    (0x187, 'M', u'ƈ'),
    (0x188, 'V'),
    (0x189, 'M', u'ɖ'),
    (0x18A, 'M', u'ɗ'),
    (0x18B, 'M', u'ƌ'),
    (0x18C, 'V'),
    (0x18E, 'M', u'ǝ'),
    (0x18F, 'M', u'ə'),
    (0x190, 'M', u'ɛ'),
    (0x191, 'M', u'ƒ'),
    (0x192, 'V'),
    (0x193, 'M', u'ɠ'),
    ]

def _seg_4():
    return [
    (0x194, 'M', u'ɣ'),
    (0x195, 'V'),
    (0x196, 'M', u'ɩ'),
    (0x197, 'M', u'ɨ'),
    (0x198, 'M', u'ƙ'),
    (0x199, 'V'),
    (0x19C, 'M', u'ɯ'),
    (0x19D, 'M', u'ɲ'),
    (0x19E, 'V'),
    (0x19F, 'M', u'ɵ'),
    (0x1A0, 'M', u'ơ'),
    (0x1A1, 'V'),
    (0x1A2, 'M', u'ƣ'),
    (0x1A3, 'V'),
    (0x1A4, 'M', u'ƥ'),
    (0x1A5, 'V'),
    (0x1A6, 'M', u'ʀ'),
    (0x1A7, 'M', u'ƨ'),
    (0x1A8, 'V'),
    (0x1A9, 'M', u'ʃ'),
    (0x1AA, 'V'),
    (0x1AC, 'M', u'ƭ'),
    (0x1AD, 'V'),
    (0x1AE, 'M', u'ʈ'),
    (0x1AF, 'M', u'ư'),
    (0x1B0, 'V'),
    (0x1B1, 'M', u'ʊ'),
    (0x1B2, 'M', u'ʋ'),
    (0x1B3, 'M', u'ƴ'),
    (0x1B4, 'V'),
    (0x1B5, 'M', u'ƶ'),
    (0x1B6, 'V'),
    (0x1B7, 'M', u'ʒ'),
    (0x1B8, 'M', u'ƹ'),
    (0x1B9, 'V'),
    (0x1BC, 'M', u'ƽ'),
    (0x1BD, 'V'),
    (0x1C4, 'M', u'dž'),
    (0x1C7, 'M', u'lj'),
    (0x1CA, 'M', u'nj'),
    (0x1CD, 'M', u'ǎ'),
    (0x1CE, 'V'),
    (0x1CF, 'M', u'ǐ'),
    (0x1D0, 'V'),
    (0x1D1, 'M', u'ǒ'),
    (0x1D2, 'V'),
    (0x1D3, 'M', u'ǔ'),
    (0x1D4, 'V'),
    (0x1D5, 'M', u'ǖ'),
    (0x1D6, 'V'),
    (0x1D7, 'M', u'ǘ'),
    (0x1D8, 'V'),
    (0x1D9, 'M', u'ǚ'),
    (0x1DA, 'V'),
    (0x1DB, 'M', u'ǜ'),
    (0x1DC, 'V'),
    (0x1DE, 'M', u'ǟ'),
    (0x1DF, 'V'),
    (0x1E0, 'M', u'ǡ'),
    (0x1E1, 'V'),
    (0x1E2, 'M', u'ǣ'),
    (0x1E3, 'V'),
    (0x1E4, 'M', u'ǥ'),
    (0x1E5, 'V'),
    (0x1E6, 'M', u'ǧ'),
    (0x1E7, 'V'),
    (0x1E8, 'M', u'ǩ'),
    (0x1E9, 'V'),
    (0x1EA, 'M', u'ǫ'),
    (0x1EB, 'V'),
    (0x1EC, 'M', u'ǭ'),
    (0x1ED, 'V'),
    (0x1EE, 'M', u'ǯ'),
    (0x1EF, 'V'),
    (0x1F1, 'M', u'dz'),
    (0x1F4, 'M', u'ǵ'),
    (0x1F5, 'V'),
    (0x1F6, 'M', u'ƕ'),
    (0x1F7, 'M', u'ƿ'),
    (0x1F8, 'M', u'ǹ'),
    (0x1F9, 'V'),
    (0x1FA, 'M', u'ǻ'),
    (0x1FB, 'V'),
    (0x1FC, 'M', u'ǽ'),
    (0x1FD, 'V'),
    (0x1FE, 'M', u'ǿ'),
    (0x1FF, 'V'),
    (0x200, 'M', u'ȁ'),
    (0x201, 'V'),
    (0x202, 'M', u'ȃ'),
    (0x203, 'V'),
    (0x204, 'M', u'ȅ'),
    (0x205, 'V'),
    (0x206, 'M', u'ȇ'),
    (0x207, 'V'),
    (0x208, 'M', u'ȉ'),
    (0x209, 'V'),
    (0x20A, 'M', u'ȋ'),
    (0x20B, 'V'),
    (0x20C, 'M', u'ȍ'),
    ]

def _seg_5():
    return [
    (0x20D, 'V'),
    (0x20E, 'M', u'ȏ'),
    (0x20F, 'V'),
    (0x210, 'M', u'ȑ'),
    (0x211, 'V'),
    (0x212, 'M', u'ȓ'),
    (0x213, 'V'),
    (0x214, 'M', u'ȕ'),
    (0x215, 'V'),
    (0x216, 'M', u'ȗ'),
    (0x217, 'V'),
    (0x218, 'M', u'ș'),
    (0x219, 'V'),
    (0x21A, 'M', u'ț'),
    (0x21B, 'V'),
    (0x21C, 'M', u'ȝ'),
    (0x21D, 'V'),
    (0x21E, 'M', u'ȟ'),
    (0x21F, 'V'),
    (0x220, 'M', u'ƞ'),
    (0x221, 'V'),
    (0x222, 'M', u'ȣ'),
    (0x223, 'V'),
    (0x224, 'M', u'ȥ'),
    (0x225, 'V'),
    (0x226, 'M', u'ȧ'),
    (0x227, 'V'),
    (0x228, 'M', u'ȩ'),
    (0x229, 'V'),
    (0x22A, 'M', u'ȫ'),
    (0x22B, 'V'),
    (0x22C, 'M', u'ȭ'),
    (0x22D, 'V'),
    (0x22E, 'M', u'ȯ'),
    (0x22F, 'V'),
    (0x230, 'M', u'ȱ'),
    (0x231, 'V'),
    (0x232, 'M', u'ȳ'),
    (0x233, 'V'),
    (0x23A, 'M', u'ⱥ'),
    (0x23B, 'M', u'ȼ'),
    (0x23C, 'V'),
    (0x23D, 'M', u'ƚ'),
    (0x23E, 'M', u'ⱦ'),
    (0x23F, 'V'),
    (0x241, 'M', u'ɂ'),
    (0x242, 'V'),
    (0x243, 'M', u'ƀ'),
    (0x244, 'M', u'ʉ'),
    (0x245, 'M', u'ʌ'),
    (0x246, 'M', u'ɇ'),
    (0x247, 'V'),
    (0x248, 'M', u'ɉ'),
    (0x249, 'V'),
    (0x24A, 'M', u'ɋ'),
    (0x24B, 'V'),
    (0x24C, 'M', u'ɍ'),
    (0x24D, 'V'),
    (0x24E, 'M', u'ɏ'),
    (0x24F, 'V'),
    (0x2B0, 'M', u'h'),
    (0x2B1, 'M', u'ɦ'),
    (0x2B2, 'M', u'j'),
    (0x2B3, 'M', u'r'),
    (0x2B4, 'M', u'ɹ'),
    (0x2B5, 'M', u'ɻ'),
    (0x2B6, 'M', u'ʁ'),
    (0x2B7, 'M', u'w'),
    (0x2B8, 'M', u'y'),
    (0x2B9, 'V'),
    (0x2D8, '3', u' ̆'),
    (0x2D9, '3', u' ̇'),
    (0x2DA, '3', u' ̊'),
    (0x2DB, '3', u' ̨'),
    (0x2DC, '3', u' ̃'),
    (0x2DD, '3', u' ̋'),
    (0x2DE, 'V'),
    (0x2E0, 'M', u'ɣ'),
    (0x2E1, 'M', u'l'),
    (0x2E2, 'M', u's'),
    (0x2E3, 'M', u'x'),
    (0x2E4, 'M', u'ʕ'),
    (0x2E5, 'V'),
    (0x340, 'M', u'̀'),
    (0x341, 'M', u'́'),
    (0x342, 'V'),
    (0x343, 'M', u'̓'),
    (0x344, 'M', u'̈́'),
    (0x345, 'M', u'ι'),
    (0x346, 'V'),
    (0x34F, 'I'),
    (0x350, 'V'),
    (0x370, 'M', u'ͱ'),
    (0x371, 'V'),
    (0x372, 'M', u'ͳ'),
    (0x373, 'V'),
    (0x374, 'M', u'ʹ'),
    (0x375, 'V'),
    (0x376, 'M', u'ͷ'),
    (0x377, 'V'),
    ]

def _seg_6():
    return [
    (0x378, 'X'),
    (0x37A, '3', u' ι'),
    (0x37B, 'V'),
    (0x37E, '3', u';'),
    (0x37F, 'M', u'ϳ'),
    (0x380, 'X'),
    (0x384, '3', u' ́'),
    (0x385, '3', u' ̈́'),
    (0x386, 'M', u'ά'),
    (0x387, 'M', u'·'),
    (0x388, 'M', u'έ'),
    (0x389, 'M', u'ή'),
    (0x38A, 'M', u'ί'),
    (0x38B, 'X'),
    (0x38C, 'M', u'ό'),
    (0x38D, 'X'),
    (0x38E, 'M', u'ύ'),
    (0x38F, 'M', u'ώ'),
    (0x390, 'V'),
    (0x391, 'M', u'α'),
    (0x392, 'M', u'β'),
    (0x393, 'M', u'γ'),
    (0x394, 'M', u'δ'),
    (0x395, 'M', u'ε'),
    (0x396, 'M', u'ζ'),
    (0x397, 'M', u'η'),
    (0x398, 'M', u'θ'),
    (0x399, 'M', u'ι'),
    (0x39A, 'M', u'κ'),
    (0x39B, 'M', u'λ'),
    (0x39C, 'M', u'μ'),
    (0x39D, 'M', u'ν'),
    (0x39E, 'M', u'ξ'),
    (0x39F, 'M', u'ο'),
    (0x3A0, 'M', u'π'),
    (0x3A1, 'M', u'ρ'),
    (0x3A2, 'X'),
    (0x3A3, 'M', u'σ'),
    (0x3A4, 'M', u'τ'),
    (0x3A5, 'M', u'υ'),
    (0x3A6, 'M', u'φ'),
    (0x3A7, 'M', u'χ'),
    (0x3A8, 'M', u'ψ'),
    (0x3A9, 'M', u'ω'),
    (0x3AA, 'M', u'ϊ'),
    (0x3AB, 'M', u'ϋ'),
    (0x3AC, 'V'),
    (0x3C2, 'D', u'σ'),
    (0x3C3, 'V'),
    (0x3CF, 'M', u'ϗ'),
    (0x3D0, 'M', u'β'),
    (0x3D1, 'M', u'θ'),
    (0x3D2, 'M', u'υ'),
    (0x3D3, 'M', u'ύ'),
    (0x3D4, 'M', u'ϋ'),
    (0x3D5, 'M', u'φ'),
    (0x3D6, 'M', u'π'),
    (0x3D7, 'V'),
    (0x3D8, 'M', u'ϙ'),
    (0x3D9, 'V'),
    (0x3DA, 'M', u'ϛ'),
    (0x3DB, 'V'),
    (0x3DC, 'M', u'ϝ'),
    (0x3DD, 'V'),
    (0x3DE, 'M', u'ϟ'),
    (0x3DF, 'V'),
    (0x3E0, 'M', u'ϡ'),
    (0x3E1, 'V'),
    (0x3E2, 'M', u'ϣ'),
    (0x3E3, 'V'),
    (0x3E4, 'M', u'ϥ'),
    (0x3E5, 'V'),
    (0x3E6, 'M', u'ϧ'),
    (0x3E7, 'V'),
    (0x3E8, 'M', u'ϩ'),
    (0x3E9, 'V'),
    (0x3EA, 'M', u'ϫ'),
    (0x3EB, 'V'),
    (0x3EC, 'M', u'ϭ'),
    (0x3ED, 'V'),
    (0x3EE, 'M', u'ϯ'),
    (0x3EF, 'V'),
    (0x3F0, 'M', u'κ'),
    (0x3F1, 'M', u'ρ'),
    (0x3F2, 'M', u'σ'),
    (0x3F3, 'V'),
    (0x3F4, 'M', u'θ'),
    (0x3F5, 'M', u'ε'),
    (0x3F6, 'V'),
    (0x3F7, 'M', u'ϸ'),
    (0x3F8, 'V'),
    (0x3F9, 'M', u'σ'),
    (0x3FA, 'M', u'ϻ'),
    (0x3FB, 'V'),
    (0x3FD, 'M', u'ͻ'),
    (0x3FE, 'M', u'ͼ'),
    (0x3FF, 'M', u'ͽ'),
    (0x400, 'M', u'ѐ'),
    (0x401, 'M', u'ё'),
    (0x402, 'M', u'ђ'),
    ]

def _seg_7():
    return [
    (0x403, 'M', u'ѓ'),
    (0x404, 'M', u'є'),
    (0x405, 'M', u'ѕ'),
    (0x406, 'M', u'і'),
    (0x407, 'M', u'ї'),
    (0x408, 'M', u'ј'),
    (0x409, 'M', u'љ'),
    (0x40A, 'M', u'њ'),
    (0x40B, 'M', u'ћ'),
    (0x40C, 'M', u'ќ'),
    (0x40D, 'M', u'ѝ'),
    (0x40E, 'M', u'ў'),
    (0x40F, 'M', u'џ'),
    (0x410, 'M', u'а'),
    (0x411, 'M', u'б'),
    (0x412, 'M', u'в'),
    (0x413, 'M', u'г'),
    (0x414, 'M', u'д'),
    (0x415, 'M', u'е'),
    (0x416, 'M', u'ж'),
    (0x417, 'M', u'з'),
    (0x418, 'M', u'и'),
    (0x419, 'M', u'й'),
    (0x41A, 'M', u'к'),
    (0x41B, 'M', u'л'),
    (0x41C, 'M', u'м'),
    (0x41D, 'M', u'н'),
    (0x41E, 'M', u'о'),
    (0x41F, 'M', u'п'),
    (0x420, 'M', u'р'),
    (0x421, 'M', u'с'),
    (0x422, 'M', u'т'),
    (0x423, 'M', u'у'),
    (0x424, 'M', u'ф'),
    (0x425, 'M', u'х'),
    (0x426, 'M', u'ц'),
    (0x427, 'M', u'ч'),
    (0x428, 'M', u'ш'),
    (0x429, 'M', u'щ'),
    (0x42A, 'M', u'ъ'),
    (0x42B, 'M', u'ы'),
    (0x42C, 'M', u'ь'),
    (0x42D, 'M', u'э'),
    (0x42E, 'M', u'ю'),
    (0x42F, 'M', u'я'),
    (0x430, 'V'),
    (0x460, 'M', u'ѡ'),
    (0x461, 'V'),
    (0x462, 'M', u'ѣ'),
    (0x463, 'V'),
    (0x464, 'M', u'ѥ'),
    (0x465, 'V'),
    (0x466, 'M', u'ѧ'),
    (0x467, 'V'),
    (0x468, 'M', u'ѩ'),
    (0x469, 'V'),
    (0x46A, 'M', u'ѫ'),
    (0x46B, 'V'),
    (0x46C, 'M', u'ѭ'),
    (0x46D, 'V'),
    (0x46E, 'M', u'ѯ'),
    (0x46F, 'V'),
    (0x470, 'M', u'ѱ'),
    (0x471, 'V'),
    (0x472, 'M', u'ѳ'),
    (0x473, 'V'),
    (0x474, 'M', u'ѵ'),
    (0x475, 'V'),
    (0x476, 'M', u'ѷ'),
    (0x477, 'V'),
    (0x478, 'M', u'ѹ'),
    (0x479, 'V'),
    (0x47A, 'M', u'ѻ'),
    (0x47B, 'V'),
    (0x47C, 'M', u'ѽ'),
    (0x47D, 'V'),
    (0x47E, 'M', u'ѿ'),
    (0x47F, 'V'),
    (0x480, 'M', u'ҁ'),
    (0x481, 'V'),
    (0x48A, 'M', u'ҋ'),
    (0x48B, 'V'),
    (0x48C, 'M', u'ҍ'),
    (0x48D, 'V'),
    (0x48E, 'M', u'ҏ'),
    (0x48F, 'V'),
    (0x490, 'M', u'ґ'),
    (0x491, 'V'),
    (0x492, 'M', u'ғ'),
    (0x493, 'V'),
    (0x494, 'M', u'ҕ'),
    (0x495, 'V'),
    (0x496, 'M', u'җ'),
    (0x497, 'V'),
    (0x498, 'M', u'ҙ'),
    (0x499, 'V'),
    (0x49A, 'M', u'қ'),
    (0x49B, 'V'),
    (0x49C, 'M', u'ҝ'),
    (0x49D, 'V'),
    ]

def _seg_8():
    return [
    (0x49E, 'M', u'ҟ'),
    (0x49F, 'V'),
    (0x4A0, 'M', u'ҡ'),
    (0x4A1, 'V'),
    (0x4A2, 'M', u'ң'),
    (0x4A3, 'V'),
    (0x4A4, 'M', u'ҥ'),
    (0x4A5, 'V'),
    (0x4A6, 'M', u'ҧ'),
    (0x4A7, 'V'),
    (0x4A8, 'M', u'ҩ'),
    (0x4A9, 'V'),
    (0x4AA, 'M', u'ҫ'),
    (0x4AB, 'V'),
    (0x4AC, 'M', u'ҭ'),
    (0x4AD, 'V'),
    (0x4AE, 'M', u'ү'),
    (0x4AF, 'V'),
    (0x4B0, 'M', u'ұ'),
    (0x4B1, 'V'),
    (0x4B2, 'M', u'ҳ'),
    (0x4B3, 'V'),
    (0x4B4, 'M', u'ҵ'),
    (0x4B5, 'V'),
    (0x4B6, 'M', u'ҷ'),
    (0x4B7, 'V'),
    (0x4B8, 'M', u'ҹ'),
    (0x4B9, 'V'),
    (0x4BA, 'M', u'һ'),
    (0x4BB, 'V'),
    (0x4BC, 'M', u'ҽ'),
    (0x4BD, 'V'),
    (0x4BE, 'M', u'ҿ'),
    (0x4BF, 'V'),
    (0x4C0, 'X'),
    (0x4C1, 'M', u'ӂ'),
    (0x4C2, 'V'),
    (0x4C3, 'M', u'ӄ'),
    (0x4C4, 'V'),
    (0x4C5, 'M', u'ӆ'),
    (0x4C6, 'V'),
    (0x4C7, 'M', u'ӈ'),
    (0x4C8, 'V'),
    (0x4C9, 'M', u'ӊ'),
    (0x4CA, 'V'),
    (0x4CB, 'M', u'ӌ'),
    (0x4CC, 'V'),
    (0x4CD, 'M', u'ӎ'),
    (0x4CE, 'V'),
    (0x4D0, 'M', u'ӑ'),
    (0x4D1, 'V'),
    (0x4D2, 'M', u'ӓ'),
    (0x4D3, 'V'),
    (0x4D4, 'M', u'ӕ'),
    (0x4D5, 'V'),
    (0x4D6, 'M', u'ӗ'),
    (0x4D7, 'V'),
    (0x4D8, 'M', u'ә'),
    (0x4D9, 'V'),
    (0x4DA, 'M', u'ӛ'),
    (0x4DB, 'V'),
    (0x4DC, 'M', u'ӝ'),
    (0x4DD, 'V'),
    (0x4DE, 'M', u'ӟ'),
    (0x4DF, 'V'),
    (0x4E0, 'M', u'ӡ'),
    (0x4E1, 'V'),
    (0x4E2, 'M', u'ӣ'),
    (0x4E3, 'V'),
    (0x4E4, 'M', u'ӥ'),
    (0x4E5, 'V'),
    (0x4E6, 'M', u'ӧ'),
    (0x4E7, 'V'),
    (0x4E8, 'M', u'ө'),
    (0x4E9, 'V'),
    (0x4EA, 'M', u'ӫ'),
    (0x4EB, 'V'),
    (0x4EC, 'M', u'ӭ'),
    (0x4ED, 'V'),
    (0x4EE, 'M', u'ӯ'),
    (0x4EF, 'V'),
    (0x4F0, 'M', u'ӱ'),
    (0x4F1, 'V'),
    (0x4F2, 'M', u'ӳ'),
    (0x4F3, 'V'),
    (0x4F4, 'M', u'ӵ'),
    (0x4F5, 'V'),
    (0x4F6, 'M', u'ӷ'),
    (0x4F7, 'V'),
    (0x4F8, 'M', u'ӹ'),
    (0x4F9, 'V'),
    (0x4FA, 'M', u'ӻ'),
    (0x4FB, 'V'),
    (0x4FC, 'M', u'ӽ'),
    (0x4FD, 'V'),
    (0x4FE, 'M', u'ӿ'),
    (0x4FF, 'V'),
    (0x500, 'M', u'ԁ'),
    (0x501, 'V'),
    (0x502, 'M', u'ԃ'),
    ]

def _seg_9():
    return [
    (0x503, 'V'),
    (0x504, 'M', u'ԅ'),
    (0x505, 'V'),
    (0x506, 'M', u'ԇ'),
    (0x507, 'V'),
    (0x508, 'M', u'ԉ'),
    (0x509, 'V'),
    (0x50A, 'M', u'ԋ'),
    (0x50B, 'V'),
    (0x50C, 'M', u'ԍ'),
    (0x50D, 'V'),
    (0x50E, 'M', u'ԏ'),
    (0x50F, 'V'),
    (0x510, 'M', u'ԑ'),
    (0x511, 'V'),
    (0x512, 'M', u'ԓ'),
    (0x513, 'V'),
    (0x514, 'M', u'ԕ'),
    (0x515, 'V'),
    (0x516, 'M', u'ԗ'),
    (0x517, 'V'),
    (0x518, 'M', u'ԙ'),
    (0x519, 'V'),
    (0x51A, 'M', u'ԛ'),
    (0x51B, 'V'),
    (0x51C, 'M', u'ԝ'),
    (0x51D, 'V'),
    (0x51E, 'M', u'ԟ'),
    (0x51F, 'V'),
    (0x520, 'M', u'ԡ'),
    (0x521, 'V'),
    (0x522, 'M', u'ԣ'),
    (0x523, 'V'),
    (0x524, 'M', u'ԥ'),
    (0x525, 'V'),
    (0x526, 'M', u'ԧ'),
    (0x527, 'V'),
    (0x528, 'M', u'ԩ'),
    (0x529, 'V'),
    (0x52A, 'M', u'ԫ'),
    (0x52B, 'V'),
    (0x52C, 'M', u'ԭ'),
    (0x52D, 'V'),
    (0x52E, 'M', u'ԯ'),
    (0x52F, 'V'),
    (0x530, 'X'),
    (0x531, 'M', u'ա'),
    (0x532, 'M', u'բ'),
    (0x533, 'M', u'գ'),
    (0x534, 'M', u'դ'),
    (0x535, 'M', u'ե'),
    (0x536, 'M', u'զ'),
    (0x537, 'M', u'է'),
    (0x538, 'M', u'ը'),
    (0x539, 'M', u'թ'),
    (0x53A, 'M', u'ժ'),
    (0x53B, 'M', u'ի'),
    (0x53C, 'M', u'լ'),
    (0x53D, 'M', u'խ'),
    (0x53E, 'M', u'ծ'),
    (0x53F, 'M', u'կ'),
    (0x540, 'M', u'հ'),
    (0x541, 'M', u'ձ'),
    (0x542, 'M', u'ղ'),
    (0x543, 'M', u'ճ'),
    (0x544, 'M', u'մ'),
    (0x545, 'M', u'յ'),
    (0x546, 'M', u'ն'),
    (0x547, 'M', u'շ'),
    (0x548, 'M', u'ո'),
    (0x549, 'M', u'չ'),
    (0x54A, 'M', u'պ'),
    (0x54B, 'M', u'ջ'),
    (0x54C, 'M', u'ռ'),
    (0x54D, 'M', u'ս'),
    (0x54E, 'M', u'վ'),
    (0x54F, 'M', u'տ'),
    (0x550, 'M', u'ր'),
    (0x551, 'M', u'ց'),
    (0x552, 'M', u'ւ'),
    (0x553, 'M', u'փ'),
    (0x554, 'M', u'ք'),
    (0x555, 'M', u'օ'),
    (0x556, 'M', u'ֆ'),
    (0x557, 'X'),
    (0x559, 'V'),
    (0x587, 'M', u'եւ'),
    (0x588, 'V'),
    (0x58B, 'X'),
    (0x58D, 'V'),
    (0x590, 'X'),
    (0x591, 'V'),
    (0x5C8, 'X'),
    (0x5D0, 'V'),
    (0x5EB, 'X'),
    (0x5EF, 'V'),
    (0x5F5, 'X'),
    (0x606, 'V'),
    (0x61C, 'X'),
    (0x61E, 'V'),
    ]

def _seg_10():
    return [
    (0x675, 'M', u'اٴ'),
    (0x676, 'M', u'وٴ'),
    (0x677, 'M', u'ۇٴ'),
    (0x678, 'M', u'يٴ'),
    (0x679, 'V'),
    (0x6DD, 'X'),
    (0x6DE, 'V'),
    (0x70E, 'X'),
    (0x710, 'V'),
    (0x74B, 'X'),
    (0x74D, 'V'),
    (0x7B2, 'X'),
    (0x7C0, 'V'),
    (0x7FB, 'X'),
    (0x7FD, 'V'),
    (0x82E, 'X'),
    (0x830, 'V'),
    (0x83F, 'X'),
    (0x840, 'V'),
    (0x85C, 'X'),
    (0x85E, 'V'),
    (0x85F, 'X'),
    (0x860, 'V'),
    (0x86B, 'X'),
    (0x8A0, 'V'),
    (0x8B5, 'X'),
    (0x8B6, 'V'),
    (0x8BE, 'X'),
    (0x8D3, 'V'),
    (0x8E2, 'X'),
    (0x8E3, 'V'),
    (0x958, 'M', u'क़'),
    (0x959, 'M', u'ख़'),
    (0x95A, 'M', u'ग़'),
    (0x95B, 'M', u'ज़'),
    (0x95C, 'M', u'ड़'),
    (0x95D, 'M', u'ढ़'),
    (0x95E, 'M', u'फ़'),
    (0x95F, 'M', u'य़'),
    (0x960, 'V'),
    (0x984, 'X'),
    (0x985, 'V'),
    (0x98D, 'X'),
    (0x98F, 'V'),
    (0x991, 'X'),
    (0x993, 'V'),
    (0x9A9, 'X'),
    (0x9AA, 'V'),
    (0x9B1, 'X'),
    (0x9B2, 'V'),
    (0x9B3, 'X'),
    (0x9B6, 'V'),
    (0x9BA, 'X'),
    (0x9BC, 'V'),
    (0x9C5, 'X'),
    (0x9C7, 'V'),
    (0x9C9, 'X'),
    (0x9CB, 'V'),
    (0x9CF, 'X'),
    (0x9D7, 'V'),
    (0x9D8, 'X'),
    (0x9DC, 'M', u'ড়'),
    (0x9DD, 'M', u'ঢ়'),
    (0x9DE, 'X'),
    (0x9DF, 'M', u'য়'),
    (0x9E0, 'V'),
    (0x9E4, 'X'),
    (0x9E6, 'V'),
    (0x9FF, 'X'),
    (0xA01, 'V'),
    (0xA04, 'X'),
    (0xA05, 'V'),
    (0xA0B, 'X'),
    (0xA0F, 'V'),
    (0xA11, 'X'),
    (0xA13, 'V'),
    (0xA29, 'X'),
    (0xA2A, 'V'),
    (0xA31, 'X'),
    (0xA32, 'V'),
    (0xA33, 'M', u'ਲ਼'),
    (0xA34, 'X'),
    (0xA35, 'V'),
    (0xA36, 'M', u'ਸ਼'),
    (0xA37, 'X'),
    (0xA38, 'V'),
    (0xA3A, 'X'),
    (0xA3C, 'V'),
    (0xA3D, 'X'),
    (0xA3E, 'V'),
    (0xA43, 'X'),
    (0xA47, 'V'),
    (0xA49, 'X'),
    (0xA4B, 'V'),
    (0xA4E, 'X'),
    (0xA51, 'V'),
    (0xA52, 'X'),
    (0xA59, 'M', u'ਖ਼'),
    (0xA5A, 'M', u'ਗ਼'),
    (0xA5B, 'M', u'ਜ਼'),
    ]

def _seg_11():
    return [
    (0xA5C, 'V'),
    (0xA5D, 'X'),
    (0xA5E, 'M', u'ਫ਼'),
    (0xA5F, 'X'),
    (0xA66, 'V'),
    (0xA77, 'X'),
    (0xA81, 'V'),
    (0xA84, 'X'),
    (0xA85, 'V'),
    (0xA8E, 'X'),
    (0xA8F, 'V'),
    (0xA92, 'X'),
    (0xA93, 'V'),
    (0xAA9, 'X'),
    (0xAAA, 'V'),
    (0xAB1, 'X'),
    (0xAB2, 'V'),
    (0xAB4, 'X'),
    (0xAB5, 'V'),
    (0xABA, 'X'),
    (0xABC, 'V'),
    (0xAC6, 'X'),
    (0xAC7, 'V'),
    (0xACA, 'X'),
    (0xACB, 'V'),
    (0xACE, 'X'),
    (0xAD0, 'V'),
    (0xAD1, 'X'),
    (0xAE0, 'V'),
    (0xAE4, 'X'),
    (0xAE6, 'V'),
    (0xAF2, 'X'),
    (0xAF9, 'V'),
    (0xB00, 'X'),
    (0xB01, 'V'),
    (0xB04, 'X'),
    (0xB05, 'V'),
    (0xB0D, 'X'),
    (0xB0F, 'V'),
    (0xB11, 'X'),
    (0xB13, 'V'),
    (0xB29, 'X'),
    (0xB2A, 'V'),
    (0xB31, 'X'),
    (0xB32, 'V'),
    (0xB34, 'X'),
    (0xB35, 'V'),
    (0xB3A, 'X'),
    (0xB3C, 'V'),
    (0xB45, 'X'),
    (0xB47, 'V'),
    (0xB49, 'X'),
    (0xB4B, 'V'),
    (0xB4E, 'X'),
    (0xB56, 'V'),
    (0xB58, 'X'),
    (0xB5C, 'M', u'ଡ଼'),
    (0xB5D, 'M', u'ଢ଼'),
    (0xB5E, 'X'),
    (0xB5F, 'V'),
    (0xB64, 'X'),
    (0xB66, 'V'),
    (0xB78, 'X'),
    (0xB82, 'V'),
    (0xB84, 'X'),
    (0xB85, 'V'),
    (0xB8B, 'X'),
    (0xB8E, 'V'),
    (0xB91, 'X'),
    (0xB92, 'V'),
    (0xB96, 'X'),
    (0xB99, 'V'),
    (0xB9B, 'X'),
    (0xB9C, 'V'),
    (0xB9D, 'X'),
    (0xB9E, 'V'),
    (0xBA0, 'X'),
    (0xBA3, 'V'),
    (0xBA5, 'X'),
    (0xBA8, 'V'),
    (0xBAB, 'X'),
    (0xBAE, 'V'),
    (0xBBA, 'X'),
    (0xBBE, 'V'),
    (0xBC3, 'X'),
    (0xBC6, 'V'),
    (0xBC9, 'X'),
    (0xBCA, 'V'),
    (0xBCE, 'X'),
    (0xBD0, 'V'),
    (0xBD1, 'X'),
    (0xBD7, 'V'),
    (0xBD8, 'X'),
    (0xBE6, 'V'),
    (0xBFB, 'X'),
    (0xC00, 'V'),
    (0xC0D, 'X'),
    (0xC0E, 'V'),
    (0xC11, 'X'),
    (0xC12, 'V'),
    ]

def _seg_12():
    return [
    (0xC29, 'X'),
    (0xC2A, 'V'),
    (0xC3A, 'X'),
    (0xC3D, 'V'),
    (0xC45, 'X'),
    (0xC46, 'V'),
    (0xC49, 'X'),
    (0xC4A, 'V'),
    (0xC4E, 'X'),
    (0xC55, 'V'),
    (0xC57, 'X'),
    (0xC58, 'V'),
    (0xC5B, 'X'),
    (0xC60, 'V'),
    (0xC64, 'X'),
    (0xC66, 'V'),
    (0xC70, 'X'),
    (0xC78, 'V'),
    (0xC8D, 'X'),
    (0xC8E, 'V'),
    (0xC91, 'X'),
    (0xC92, 'V'),
    (0xCA9, 'X'),
    (0xCAA, 'V'),
    (0xCB4, 'X'),
    (0xCB5, 'V'),
    (0xCBA, 'X'),
    (0xCBC, 'V'),
    (0xCC5, 'X'),
    (0xCC6, 'V'),
    (0xCC9, 'X'),
    (0xCCA, 'V'),
    (0xCCE, 'X'),
    (0xCD5, 'V'),
    (0xCD7, 'X'),
    (0xCDE, 'V'),
    (0xCDF, 'X'),
    (0xCE0, 'V'),
    (0xCE4, 'X'),
    (0xCE6, 'V'),
    (0xCF0, 'X'),
    (0xCF1, 'V'),
    (0xCF3, 'X'),
    (0xD00, 'V'),
    (0xD04, 'X'),
    (0xD05, 'V'),
    (0xD0D, 'X'),
    (0xD0E, 'V'),
    (0xD11, 'X'),
    (0xD12, 'V'),
    (0xD45, 'X'),
    (0xD46, 'V'),
    (0xD49, 'X'),
    (0xD4A, 'V'),
    (0xD50, 'X'),
    (0xD54, 'V'),
    (0xD64, 'X'),
    (0xD66, 'V'),
    (0xD80, 'X'),
    (0xD82, 'V'),
    (0xD84, 'X'),
    (0xD85, 'V'),
    (0xD97, 'X'),
    (0xD9A, 'V'),
    (0xDB2, 'X'),
    (0xDB3, 'V'),
    (0xDBC, 'X'),
    (0xDBD, 'V'),
    (0xDBE, 'X'),
    (0xDC0, 'V'),
    (0xDC7, 'X'),
    (0xDCA, 'V'),
    (0xDCB, 'X'),
    (0xDCF, 'V'),
    (0xDD5, 'X'),
    (0xDD6, 'V'),
    (0xDD7, 'X'),
    (0xDD8, 'V'),
    (0xDE0, 'X'),
    (0xDE6, 'V'),
    (0xDF0, 'X'),
    (0xDF2, 'V'),
    (0xDF5, 'X'),
    (0xE01, 'V'),
    (0xE33, 'M', u'ํา'),
    (0xE34, 'V'),
    (0xE3B, 'X'),
    (0xE3F, 'V'),
    (0xE5C, 'X'),
    (0xE81, 'V'),
    (0xE83, 'X'),
    (0xE84, 'V'),
    (0xE85, 'X'),
    (0xE87, 'V'),
    (0xE89, 'X'),
    (0xE8A, 'V'),
    (0xE8B, 'X'),
    (0xE8D, 'V'),
    (0xE8E, 'X'),
    (0xE94, 'V'),
    ]

def _seg_13():
    return [
    (0xE98, 'X'),
    (0xE99, 'V'),
    (0xEA0, 'X'),
    (0xEA1, 'V'),
    (0xEA4, 'X'),
    (0xEA5, 'V'),
    (0xEA6, 'X'),
    (0xEA7, 'V'),
    (0xEA8, 'X'),
    (0xEAA, 'V'),
    (0xEAC, 'X'),
    (0xEAD, 'V'),
    (0xEB3, 'M', u'ໍາ'),
    (0xEB4, 'V'),
    (0xEBA, 'X'),
    (0xEBB, 'V'),
    (0xEBE, 'X'),
    (0xEC0, 'V'),
    (0xEC5, 'X'),
    (0xEC6, 'V'),
    (0xEC7, 'X'),
    (0xEC8, 'V'),
    (0xECE, 'X'),
    (0xED0, 'V'),
    (0xEDA, 'X'),
    (0xEDC, 'M', u'ຫນ'),
    (0xEDD, 'M', u'ຫມ'),
    (0xEDE, 'V'),
    (0xEE0, 'X'),
    (0xF00, 'V'),
    (0xF0C, 'M', u'་'),
    (0xF0D, 'V'),
    (0xF43, 'M', u'གྷ'),
    (0xF44, 'V'),
    (0xF48, 'X'),
    (0xF49, 'V'),
    (0xF4D, 'M', u'ཌྷ'),
    (0xF4E, 'V'),
    (0xF52, 'M', u'དྷ'),
    (0xF53, 'V'),
    (0xF57, 'M', u'བྷ'),
    (0xF58, 'V'),
    (0xF5C, 'M', u'ཛྷ'),
    (0xF5D, 'V'),
    (0xF69, 'M', u'ཀྵ'),
    (0xF6A, 'V'),
    (0xF6D, 'X'),
    (0xF71, 'V'),
    (0xF73, 'M', u'ཱི'),
    (0xF74, 'V'),
    (0xF75, 'M', u'ཱུ'),
    (0xF76, 'M', u'ྲྀ'),
    (0xF77, 'M', u'ྲཱྀ'),
    (0xF78, 'M', u'ླྀ'),
    (0xF79, 'M', u'ླཱྀ'),
    (0xF7A, 'V'),
    (0xF81, 'M', u'ཱྀ'),
    (0xF82, 'V'),
    (0xF93, 'M', u'ྒྷ'),
    (0xF94, 'V'),
    (0xF98, 'X'),
    (0xF99, 'V'),
    (0xF9D, 'M', u'ྜྷ'),
    (0xF9E, 'V'),
    (0xFA2, 'M', u'ྡྷ'),
    (0xFA3, 'V'),
    (0xFA7, 'M', u'ྦྷ'),
    (0xFA8, 'V'),
    (0xFAC, 'M', u'ྫྷ'),
    (0xFAD, 'V'),
    (0xFB9, 'M', u'ྐྵ'),
    (0xFBA, 'V'),
    (0xFBD, 'X'),
    (0xFBE, 'V'),
    (0xFCD, 'X'),
    (0xFCE, 'V'),
    (0xFDB, 'X'),
    (0x1000, 'V'),
    (0x10A0, 'X'),
    (0x10C7, 'M', u'ⴧ'),
    (0x10C8, 'X'),
    (0x10CD, 'M', u'ⴭ'),
    (0x10CE, 'X'),
    (0x10D0, 'V'),
    (0x10FC, 'M', u'ნ'),
    (0x10FD, 'V'),
    (0x115F, 'X'),
    (0x1161, 'V'),
    (0x1249, 'X'),
    (0x124A, 'V'),
    (0x124E, 'X'),
    (0x1250, 'V'),
    (0x1257, 'X'),
    (0x1258, 'V'),
    (0x1259, 'X'),
    (0x125A, 'V'),
    (0x125E, 'X'),
    (0x1260, 'V'),
    (0x1289, 'X'),
    (0x128A, 'V'),
    ]

def _seg_14():
    return [
    (0x128E, 'X'),
    (0x1290, 'V'),
    (0x12B1, 'X'),
    (0x12B2, 'V'),
    (0x12B6, 'X'),
    (0x12B8, 'V'),
    (0x12BF, 'X'),
    (0x12C0, 'V'),
    (0x12C1, 'X'),
    (0x12C2, 'V'),
    (0x12C6, 'X'),
    (0x12C8, 'V'),
    (0x12D7, 'X'),
    (0x12D8, 'V'),
    (0x1311, 'X'),
    (0x1312, 'V'),
    (0x1316, 'X'),
    (0x1318, 'V'),
    (0x135B, 'X'),
    (0x135D, 'V'),
    (0x137D, 'X'),
    (0x1380, 'V'),
    (0x139A, 'X'),
    (0x13A0, 'V'),
    (0x13F6, 'X'),
    (0x13F8, 'M', u'Ᏸ'),
    (0x13F9, 'M', u'Ᏹ'),
    (0x13FA, 'M', u'Ᏺ'),
    (0x13FB, 'M', u'Ᏻ'),
    (0x13FC, 'M', u'Ᏼ'),
    (0x13FD, 'M', u'Ᏽ'),
    (0x13FE, 'X'),
    (0x1400, 'V'),
    (0x1680, 'X'),
    (0x1681, 'V'),
    (0x169D, 'X'),
    (0x16A0, 'V'),
    (0x16F9, 'X'),
    (0x1700, 'V'),
    (0x170D, 'X'),
    (0x170E, 'V'),
    (0x1715, 'X'),
    (0x1720, 'V'),
    (0x1737, 'X'),
    (0x1740, 'V'),
    (0x1754, 'X'),
    (0x1760, 'V'),
    (0x176D, 'X'),
    (0x176E, 'V'),
    (0x1771, 'X'),
    (0x1772, 'V'),
    (0x1774, 'X'),
    (0x1780, 'V'),
    (0x17B4, 'X'),
    (0x17B6, 'V'),
    (0x17DE, 'X'),
    (0x17E0, 'V'),
    (0x17EA, 'X'),
    (0x17F0, 'V'),
    (0x17FA, 'X'),
    (0x1800, 'V'),
    (0x1806, 'X'),
    (0x1807, 'V'),
    (0x180B, 'I'),
    (0x180E, 'X'),
    (0x1810, 'V'),
    (0x181A, 'X'),
    (0x1820, 'V'),
    (0x1879, 'X'),
    (0x1880, 'V'),
    (0x18AB, 'X'),
    (0x18B0, 'V'),
    (0x18F6, 'X'),
    (0x1900, 'V'),
    (0x191F, 'X'),
    (0x1920, 'V'),
    (0x192C, 'X'),
    (0x1930, 'V'),
    (0x193C, 'X'),
    (0x1940, 'V'),
    (0x1941, 'X'),
    (0x1944, 'V'),
    (0x196E, 'X'),
    (0x1970, 'V'),
    (0x1975, 'X'),
    (0x1980, 'V'),
    (0x19AC, 'X'),
    (0x19B0, 'V'),
    (0x19CA, 'X'),
    (0x19D0, 'V'),
    (0x19DB, 'X'),
    (0x19DE, 'V'),
    (0x1A1C, 'X'),
    (0x1A1E, 'V'),
    (0x1A5F, 'X'),
    (0x1A60, 'V'),
    (0x1A7D, 'X'),
    (0x1A7F, 'V'),
    (0x1A8A, 'X'),
    (0x1A90, 'V'),
    ]

def _seg_15():
    return [
    (0x1A9A, 'X'),
    (0x1AA0, 'V'),
    (0x1AAE, 'X'),
    (0x1AB0, 'V'),
    (0x1ABF, 'X'),
    (0x1B00, 'V'),
    (0x1B4C, 'X'),
    (0x1B50, 'V'),
    (0x1B7D, 'X'),
    (0x1B80, 'V'),
    (0x1BF4, 'X'),
    (0x1BFC, 'V'),
    (0x1C38, 'X'),
    (0x1C3B, 'V'),
    (0x1C4A, 'X'),
    (0x1C4D, 'V'),
    (0x1C80, 'M', u'в'),
    (0x1C81, 'M', u'д'),
    (0x1C82, 'M', u'о'),
    (0x1C83, 'M', u'с'),
    (0x1C84, 'M', u'т'),
    (0x1C86, 'M', u'ъ'),
    (0x1C87, 'M', u'ѣ'),
    (0x1C88, 'M', u'ꙋ'),
    (0x1C89, 'X'),
    (0x1CC0, 'V'),
    (0x1CC8, 'X'),
    (0x1CD0, 'V'),
    (0x1CFA, 'X'),
    (0x1D00, 'V'),
    (0x1D2C, 'M', u'a'),
    (0x1D2D, 'M', u'æ'),
    (0x1D2E, 'M', u'b'),
    (0x1D2F, 'V'),
    (0x1D30, 'M', u'd'),
    (0x1D31, 'M', u'e'),
    (0x1D32, 'M', u'ǝ'),
    (0x1D33, 'M', u'g'),
    (0x1D34, 'M', u'h'),
    (0x1D35, 'M', u'i'),
    (0x1D36, 'M', u'j'),
    (0x1D37, 'M', u'k'),
    (0x1D38, 'M', u'l'),
    (0x1D39, 'M', u'm'),
    (0x1D3A, 'M', u'n'),
    (0x1D3B, 'V'),
    (0x1D3C, 'M', u'o'),
    (0x1D3D, 'M', u'ȣ'),
    (0x1D3E, 'M', u'p'),
    (0x1D3F, 'M', u'r'),
    (0x1D40, 'M', u't'),
    (0x1D41, 'M', u'u'),
    (0x1D42, 'M', u'w'),
    (0x1D43, 'M', u'a'),
    (0x1D44, 'M', u'ɐ'),
    (0x1D45, 'M', u'ɑ'),
    (0x1D46, 'M', u'ᴂ'),
    (0x1D47, 'M', u'b'),
    (0x1D48, 'M', u'd'),
    (0x1D49, 'M', u'e'),
    (0x1D4A, 'M', u'ə'),
    (0x1D4B, 'M', u'ɛ'),
    (0x1D4C, 'M', u'ɜ'),
    (0x1D4D, 'M', u'g'),
    (0x1D4E, 'V'),
    (0x1D4F, 'M', u'k'),
    (0x1D50, 'M', u'm'),
    (0x1D51, 'M', u'ŋ'),
    (0x1D52, 'M', u'o'),
    (0x1D53, 'M', u'ɔ'),
    (0x1D54, 'M', u'ᴖ'),
    (0x1D55, 'M', u'ᴗ'),
    (0x1D56, 'M', u'p'),
    (0x1D57, 'M', u't'),
    (0x1D58, 'M', u'u'),
    (0x1D59, 'M', u'ᴝ'),
    (0x1D5A, 'M', u'ɯ'),
    (0x1D5B, 'M', u'v'),
    (0x1D5C, 'M', u'ᴥ'),
    (0x1D5D, 'M', u'β'),
    (0x1D5E, 'M', u'γ'),
    (0x1D5F, 'M', u'δ'),
    (0x1D60, 'M', u'φ'),
    (0x1D61, 'M', u'χ'),
    (0x1D62, 'M', u'i'),
    (0x1D63, 'M', u'r'),
    (0x1D64, 'M', u'u'),
    (0x1D65, 'M', u'v'),
    (0x1D66, 'M', u'β'),
    (0x1D67, 'M', u'γ'),
    (0x1D68, 'M', u'ρ'),
    (0x1D69, 'M', u'φ'),
    (0x1D6A, 'M', u'χ'),
    (0x1D6B, 'V'),
    (0x1D78, 'M', u'н'),
    (0x1D79, 'V'),
    (0x1D9B, 'M', u'ɒ'),
    (0x1D9C, 'M', u'c'),
    (0x1D9D, 'M', u'ɕ'),
    (0x1D9E, 'M', u'ð'),
    ]

def _seg_16():
    return [
    (0x1D9F, 'M', u'ɜ'),
    (0x1DA0, 'M', u'f'),
    (0x1DA1, 'M', u'ɟ'),
    (0x1DA2, 'M', u'ɡ'),
    (0x1DA3, 'M', u'ɥ'),
    (0x1DA4, 'M', u'ɨ'),
    (0x1DA5, 'M', u'ɩ'),
    (0x1DA6, 'M', u'ɪ'),
    (0x1DA7, 'M', u'ᵻ'),
    (0x1DA8, 'M', u'ʝ'),
    (0x1DA9, 'M', u'ɭ'),
    (0x1DAA, 'M', u'ᶅ'),
    (0x1DAB, 'M', u'ʟ'),
    (0x1DAC, 'M', u'ɱ'),
    (0x1DAD, 'M', u'ɰ'),
    (0x1DAE, 'M', u'ɲ'),
    (0x1DAF, 'M', u'ɳ'),
    (0x1DB0, 'M', u'ɴ'),
    (0x1DB1, 'M', u'ɵ'),
    (0x1DB2, 'M', u'ɸ'),
    (0x1DB3, 'M', u'ʂ'),
    (0x1DB4, 'M', u'ʃ'),
    (0x1DB5, 'M', u'ƫ'),
    (0x1DB6, 'M', u'ʉ'),
    (0x1DB7, 'M', u'ʊ'),
    (0x1DB8, 'M', u'ᴜ'),
    (0x1DB9, 'M', u'ʋ'),
    (0x1DBA, 'M', u'ʌ'),
    (0x1DBB, 'M', u'z'),
    (0x1DBC, 'M', u'ʐ'),
    (0x1DBD, 'M', u'ʑ'),
    (0x1DBE, 'M', u'ʒ'),
    (0x1DBF, 'M', u'θ'),
    (0x1DC0, 'V'),
    (0x1DFA, 'X'),
    (0x1DFB, 'V'),
    (0x1E00, 'M', u'ḁ'),
    (0x1E01, 'V'),
    (0x1E02, 'M', u'ḃ'),
    (0x1E03, 'V'),
    (0x1E04, 'M', u'ḅ'),
    (0x1E05, 'V'),
    (0x1E06, 'M', u'ḇ'),
    (0x1E07, 'V'),
    (0x1E08, 'M', u'ḉ'),
    (0x1E09, 'V'),
    (0x1E0A, 'M', u'ḋ'),
    (0x1E0B, 'V'),
    (0x1E0C, 'M', u'ḍ'),
    (0x1E0D, 'V'),
    (0x1E0E, 'M', u'ḏ'),
    (0x1E0F, 'V'),
    (0x1E10, 'M', u'ḑ'),
    (0x1E11, 'V'),
    (0x1E12, 'M', u'ḓ'),
    (0x1E13, 'V'),
    (0x1E14, 'M', u'ḕ'),
    (0x1E15, 'V'),
    (0x1E16, 'M', u'ḗ'),
    (0x1E17, 'V'),
    (0x1E18, 'M', u'ḙ'),
    (0x1E19, 'V'),
    (0x1E1A, 'M', u'ḛ'),
    (0x1E1B, 'V'),
    (0x1E1C, 'M', u'ḝ'),
    (0x1E1D, 'V'),
    (0x1E1E, 'M', u'ḟ'),
    (0x1E1F, 'V'),
    (0x1E20, 'M', u'ḡ'),
    (0x1E21, 'V'),
    (0x1E22, 'M', u'ḣ'),
    (0x1E23, 'V'),
    (0x1E24, 'M', u'ḥ'),
    (0x1E25, 'V'),
    (0x1E26, 'M', u'ḧ'),
    (0x1E27, 'V'),
    (0x1E28, 'M', u'ḩ'),
    (0x1E29, 'V'),
    (0x1E2A, 'M', u'ḫ'),
    (0x1E2B, 'V'),
    (0x1E2C, 'M', u'ḭ'),
    (0x1E2D, 'V'),
    (0x1E2E, 'M', u'ḯ'),
    (0x1E2F, 'V'),
    (0x1E30, 'M', u'ḱ'),
    (0x1E31, 'V'),
    (0x1E32, 'M', u'ḳ'),
    (0x1E33, 'V'),
    (0x1E34, 'M', u'ḵ'),
    (0x1E35, 'V'),
    (0x1E36, 'M', u'ḷ'),
    (0x1E37, 'V'),
    (0x1E38, 'M', u'ḹ'),
    (0x1E39, 'V'),
    (0x1E3A, 'M', u'ḻ'),
    (0x1E3B, 'V'),
    (0x1E3C, 'M', u'ḽ'),
    (0x1E3D, 'V'),
    (0x1E3E, 'M', u'ḿ'),
    (0x1E3F, 'V'),
    ]

def _seg_17():
    return [
    (0x1E40, 'M', u'ṁ'),
    (0x1E41, 'V'),
    (0x1E42, 'M', u'ṃ'),
    (0x1E43, 'V'),
    (0x1E44, 'M', u'ṅ'),
    (0x1E45, 'V'),
    (0x1E46, 'M', u'ṇ'),
    (0x1E47, 'V'),
    (0x1E48, 'M', u'ṉ'),
    (0x1E49, 'V'),
    (0x1E4A, 'M', u'ṋ'),
    (0x1E4B, 'V'),
    (0x1E4C, 'M', u'ṍ'),
    (0x1E4D, 'V'),
    (0x1E4E, 'M', u'ṏ'),
    (0x1E4F, 'V'),
    (0x1E50, 'M', u'ṑ'),
    (0x1E51, 'V'),
    (0x1E52, 'M', u'ṓ'),
    (0x1E53, 'V'),
    (0x1E54, 'M', u'ṕ'),
    (0x1E55, 'V'),
    (0x1E56, 'M', u'ṗ'),
    (0x1E57, 'V'),
    (0x1E58, 'M', u'ṙ'),
    (0x1E59, 'V'),
    (0x1E5A, 'M', u'ṛ'),
    (0x1E5B, 'V'),
    (0x1E5C, 'M', u'ṝ'),
    (0x1E5D, 'V'),
    (0x1E5E, 'M', u'ṟ'),
    (0x1E5F, 'V'),
    (0x1E60, 'M', u'ṡ'),
    (0x1E61, 'V'),
    (0x1E62, 'M', u'ṣ'),
    (0x1E63, 'V'),
    (0x1E64, 'M', u'ṥ'),
    (0x1E65, 'V'),
    (0x1E66, 'M', u'ṧ'),
    (0x1E67, 'V'),
    (0x1E68, 'M', u'ṩ'),
    (0x1E69, 'V'),
    (0x1E6A, 'M', u'ṫ'),
    (0x1E6B, 'V'),
    (0x1E6C, 'M', u'ṭ'),
    (0x1E6D, 'V'),
    (0x1E6E, 'M', u'ṯ'),
    (0x1E6F, 'V'),
    (0x1E70, 'M', u'ṱ'),
    (0x1E71, 'V'),
    (0x1E72, 'M', u'ṳ'),
    (0x1E73, 'V'),
    (0x1E74, 'M', u'ṵ'),
    (0x1E75, 'V'),
    (0x1E76, 'M', u'ṷ'),
    (0x1E77, 'V'),
    (0x1E78, 'M', u'ṹ'),
    (0x1E79, 'V'),
    (0x1E7A, 'M', u'ṻ'),
    (0x1E7B, 'V'),
    (0x1E7C, 'M', u'ṽ'),
    (0x1E7D, 'V'),
    (0x1E7E, 'M', u'ṿ'),
    (0x1E7F, 'V'),
    (0x1E80, 'M', u'ẁ'),
    (0x1E81, 'V'),
    (0x1E82, 'M', u'ẃ'),
    (0x1E83, 'V'),
    (0x1E84, 'M', u'ẅ'),
    (0x1E85, 'V'),
    (0x1E86, 'M', u'ẇ'),
    (0x1E87, 'V'),
    (0x1E88, 'M', u'ẉ'),
    (0x1E89, 'V'),
    (0x1E8A, 'M', u'ẋ'),
    (0x1E8B, 'V'),
    (0x1E8C, 'M', u'ẍ'),
    (0x1E8D, 'V'),
    (0x1E8E, 'M', u'ẏ'),
    (0x1E8F, 'V'),
    (0x1E90, 'M', u'ẑ'),
    (0x1E91, 'V'),
    (0x1E92, 'M', u'ẓ'),
    (0x1E93, 'V'),
    (0x1E94, 'M', u'ẕ'),
    (0x1E95, 'V'),
    (0x1E9A, 'M', u'aʾ'),
    (0x1E9B, 'M', u'ṡ'),
    (0x1E9C, 'V'),
    (0x1E9E, 'M', u'ss'),
    (0x1E9F, 'V'),
    (0x1EA0, 'M', u'ạ'),
    (0x1EA1, 'V'),
    (0x1EA2, 'M', u'ả'),
    (0x1EA3, 'V'),
    (0x1EA4, 'M', u'ấ'),
    (0x1EA5, 'V'),
    (0x1EA6, 'M', u'ầ'),
    (0x1EA7, 'V'),
    (0x1EA8, 'M', u'ẩ'),
    ]

def _seg_18():
    return [
    (0x1EA9, 'V'),
    (0x1EAA, 'M', u'ẫ'),
    (0x1EAB, 'V'),
    (0x1EAC, 'M', u'ậ'),
    (0x1EAD, 'V'),
    (0x1EAE, 'M', u'ắ'),
    (0x1EAF, 'V'),
    (0x1EB0, 'M', u'ằ'),
    (0x1EB1, 'V'),
    (0x1EB2, 'M', u'ẳ'),
    (0x1EB3, 'V'),
    (0x1EB4, 'M', u'ẵ'),
    (0x1EB5, 'V'),
    (0x1EB6, 'M', u'ặ'),
    (0x1EB7, 'V'),
    (0x1EB8, 'M', u'ẹ'),
    (0x1EB9, 'V'),
    (0x1EBA, 'M', u'ẻ'),
    (0x1EBB, 'V'),
    (0x1EBC, 'M', u'ẽ'),
    (0x1EBD, 'V'),
    (0x1EBE, 'M', u'ế'),
    (0x1EBF, 'V'),
    (0x1EC0, 'M', u'ề'),
    (0x1EC1, 'V'),
    (0x1EC2, 'M', u'ể'),
    (0x1EC3, 'V'),
    (0x1EC4, 'M', u'ễ'),
    (0x1EC5, 'V'),
    (0x1EC6, 'M', u'ệ'),
    (0x1EC7, 'V'),
    (0x1EC8, 'M', u'ỉ'),
    (0x1EC9, 'V'),
    (0x1ECA, 'M', u'ị'),
    (0x1ECB, 'V'),
    (0x1ECC, 'M', u'ọ'),
    (0x1ECD, 'V'),
    (0x1ECE, 'M', u'ỏ'),
    (0x1ECF, 'V'),
    (0x1ED0, 'M', u'ố'),
    (0x1ED1, 'V'),
    (0x1ED2, 'M', u'ồ'),
    (0x1ED3, 'V'),
    (0x1ED4, 'M', u'ổ'),
    (0x1ED5, 'V'),
    (0x1ED6, 'M', u'ỗ'),
    (0x1ED7, 'V'),
    (0x1ED8, 'M', u'ộ'),
    (0x1ED9, 'V'),
    (0x1EDA, 'M', u'ớ'),
    (0x1EDB, 'V'),
    (0x1EDC, 'M', u'ờ'),
    (0x1EDD, 'V'),
    (0x1EDE, 'M', u'ở'),
    (0x1EDF, 'V'),
    (0x1EE0, 'M', u'ỡ'),
    (0x1EE1, 'V'),
    (0x1EE2, 'M', u'ợ'),
    (0x1EE3, 'V'),
    (0x1EE4, 'M', u'ụ'),
    (0x1EE5, 'V'),
    (0x1EE6, 'M', u'ủ'),
    (0x1EE7, 'V'),
    (0x1EE8, 'M', u'ứ'),
    (0x1EE9, 'V'),
    (0x1EEA, 'M', u'ừ'),
    (0x1EEB, 'V'),
    (0x1EEC, 'M', u'ử'),
    (0x1EED, 'V'),
    (0x1EEE, 'M', u'ữ'),
    (0x1EEF, 'V'),
    (0x1EF0, 'M', u'ự'),
    (0x1EF1, 'V'),
    (0x1EF2, 'M', u'ỳ'),
    (0x1EF3, 'V'),
    (0x1EF4, 'M', u'ỵ'),
    (0x1EF5, 'V'),
    (0x1EF6, 'M', u'ỷ'),
    (0x1EF7, 'V'),
    (0x1EF8, 'M', u'ỹ'),
    (0x1EF9, 'V'),
    (0x1EFA, 'M', u'ỻ'),
    (0x1EFB, 'V'),
    (0x1EFC, 'M', u'ỽ'),
    (0x1EFD, 'V'),
    (0x1EFE, 'M', u'ỿ'),
    (0x1EFF, 'V'),
    (0x1F08, 'M', u'ἀ'),
    (0x1F09, 'M', u'ἁ'),
    (0x1F0A, 'M', u'ἂ'),
    (0x1F0B, 'M', u'ἃ'),
    (0x1F0C, 'M', u'ἄ'),
    (0x1F0D, 'M', u'ἅ'),
    (0x1F0E, 'M', u'ἆ'),
    (0x1F0F, 'M', u'ἇ'),
    (0x1F10, 'V'),
    (0x1F16, 'X'),
    (0x1F18, 'M', u'ἐ'),
    (0x1F19, 'M', u'ἑ'),
    (0x1F1A, 'M', u'ἒ'),
    ]

def _seg_19():
    return [
    (0x1F1B, 'M', u'ἓ'),
    (0x1F1C, 'M', u'ἔ'),
    (0x1F1D, 'M', u'ἕ'),
    (0x1F1E, 'X'),
    (0x1F20, 'V'),
    (0x1F28, 'M', u'ἠ'),
    (0x1F29, 'M', u'ἡ'),
    (0x1F2A, 'M', u'ἢ'),
    (0x1F2B, 'M', u'ἣ'),
    (0x1F2C, 'M', u'ἤ'),
    (0x1F2D, 'M', u'ἥ'),
    (0x1F2E, 'M', u'ἦ'),
    (0x1F2F, 'M', u'ἧ'),
    (0x1F30, 'V'),
    (0x1F38, 'M', u'ἰ'),
    (0x1F39, 'M', u'ἱ'),
    (0x1F3A, 'M', u'ἲ'),
    (0x1F3B, 'M', u'ἳ'),
    (0x1F3C, 'M', u'ἴ'),
    (0x1F3D, 'M', u'ἵ'),
    (0x1F3E, 'M', u'ἶ'),
    (0x1F3F, 'M', u'ἷ'),
    (0x1F40, 'V'),
    (0x1F46, 'X'),
    (0x1F48, 'M', u'ὀ'),
    (0x1F49, 'M', u'ὁ'),
    (0x1F4A, 'M', u'ὂ'),
    (0x1F4B, 'M', u'ὃ'),
    (0x1F4C, 'M', u'ὄ'),
    (0x1F4D, 'M', u'ὅ'),
    (0x1F4E, 'X'),
    (0x1F50, 'V'),
    (0x1F58, 'X'),
    (0x1F59, 'M', u'ὑ'),
    (0x1F5A, 'X'),
    (0x1F5B, 'M', u'ὓ'),
    (0x1F5C, 'X'),
    (0x1F5D, 'M', u'ὕ'),
    (0x1F5E, 'X'),
    (0x1F5F, 'M', u'ὗ'),
    (0x1F60, 'V'),
    (0x1F68, 'M', u'ὠ'),
    (0x1F69, 'M', u'ὡ'),
    (0x1F6A, 'M', u'ὢ'),
    (0x1F6B, 'M', u'ὣ'),
    (0x1F6C, 'M', u'ὤ'),
    (0x1F6D, 'M', u'ὥ'),
    (0x1F6E, 'M', u'ὦ'),
    (0x1F6F, 'M', u'ὧ'),
    (0x1F70, 'V'),
    (0x1F71, 'M', u'ά'),
    (0x1F72, 'V'),
    (0x1F73, 'M', u'έ'),
    (0x1F74, 'V'),
    (0x1F75, 'M', u'ή'),
    (0x1F76, 'V'),
    (0x1F77, 'M', u'ί'),
    (0x1F78, 'V'),
    (0x1F79, 'M', u'ό'),
    (0x1F7A, 'V'),
    (0x1F7B, 'M', u'ύ'),
    (0x1F7C, 'V'),
    (0x1F7D, 'M', u'ώ'),
    (0x1F7E, 'X'),
    (0x1F80, 'M', u'ἀι'),
    (0x1F81, 'M', u'ἁι'),
    (0x1F82, 'M', u'ἂι'),
    (0x1F83, 'M', u'ἃι'),
    (0x1F84, 'M', u'ἄι'),
    (0x1F85, 'M', u'ἅι'),
    (0x1F86, 'M', u'ἆι'),
    (0x1F87, 'M', u'ἇι'),
    (0x1F88, 'M', u'ἀι'),
    (0x1F89, 'M', u'ἁι'),
    (0x1F8A, 'M', u'ἂι'),
    (0x1F8B, 'M', u'ἃι'),
    (0x1F8C, 'M', u'ἄι'),
    (0x1F8D, 'M', u'ἅι'),
    (0x1F8E, 'M', u'ἆι'),
    (0x1F8F, 'M', u'ἇι'),
    (0x1F90, 'M', u'ἠι'),
    (0x1F91, 'M', u'ἡι'),
    (0x1F92, 'M', u'ἢι'),
    (0x1F93, 'M', u'ἣι'),
    (0x1F94, 'M', u'ἤι'),
    (0x1F95, 'M', u'ἥι'),
    (0x1F96, 'M', u'ἦι'),
    (0x1F97, 'M', u'ἧι'),
    (0x1F98, 'M', u'ἠι'),
    (0x1F99, 'M', u'ἡι'),
    (0x1F9A, 'M', u'ἢι'),
    (0x1F9B, 'M', u'ἣι'),
    (0x1F9C, 'M', u'ἤι'),
    (0x1F9D, 'M', u'ἥι'),
    (0x1F9E, 'M', u'ἦι'),
    (0x1F9F, 'M', u'ἧι'),
    (0x1FA0, 'M', u'ὠι'),
    (0x1FA1, 'M', u'ὡι'),
    (0x1FA2, 'M', u'ὢι'),
    (0x1FA3, 'M', u'ὣι'),
    ]

def _seg_20():
    return [
    (0x1FA4, 'M', u'ὤι'),
    (0x1FA5, 'M', u'ὥι'),
    (0x1FA6, 'M', u'ὦι'),
    (0x1FA7, 'M', u'ὧι'),
    (0x1FA8, 'M', u'ὠι'),
    (0x1FA9, 'M', u'ὡι'),
    (0x1FAA, 'M', u'ὢι'),
    (0x1FAB, 'M', u'ὣι'),
    (0x1FAC, 'M', u'ὤι'),
    (0x1FAD, 'M', u'ὥι'),
    (0x1FAE, 'M', u'ὦι'),
    (0x1FAF, 'M', u'ὧι'),
    (0x1FB0, 'V'),
    (0x1FB2, 'M', u'ὰι'),
    (0x1FB3, 'M', u'αι'),
    (0x1FB4, 'M', u'άι'),
    (0x1FB5, 'X'),
    (0x1FB6, 'V'),
    (0x1FB7, 'M', u'ᾶι'),
    (0x1FB8, 'M', u'ᾰ'),
    (0x1FB9, 'M', u'ᾱ'),
    (0x1FBA, 'M', u'ὰ'),
    (0x1FBB, 'M', u'ά'),
    (0x1FBC, 'M', u'αι'),
    (0x1FBD, '3', u' ̓'),
    (0x1FBE, 'M', u'ι'),
    (0x1FBF, '3', u' ̓'),
    (0x1FC0, '3', u' ͂'),
    (0x1FC1, '3', u' ̈͂'),
    (0x1FC2, 'M', u'ὴι'),
    (0x1FC3, 'M', u'ηι'),
    (0x1FC4, 'M', u'ήι'),
    (0x1FC5, 'X'),
    (0x1FC6, 'V'),
    (0x1FC7, 'M', u'ῆι'),
    (0x1FC8, 'M', u'ὲ'),
    (0x1FC9, 'M', u'έ'),
    (0x1FCA, 'M', u'ὴ'),
    (0x1FCB, 'M', u'ή'),
    (0x1FCC, 'M', u'ηι'),
    (0x1FCD, '3', u' ̓̀'),
    (0x1FCE, '3', u' ̓́'),
    (0x1FCF, '3', u' ̓͂'),
    (0x1FD0, 'V'),
    (0x1FD3, 'M', u'ΐ'),
    (0x1FD4, 'X'),
    (0x1FD6, 'V'),
    (0x1FD8, 'M', u'ῐ'),
    (0x1FD9, 'M', u'ῑ'),
    (0x1FDA, 'M', u'ὶ'),
    (0x1FDB, 'M', u'ί'),
    (0x1FDC, 'X'),
    (0x1FDD, '3', u' ̔̀'),
    (0x1FDE, '3', u' ̔́'),
    (0x1FDF, '3', u' ̔͂'),
    (0x1FE0, 'V'),
    (0x1FE3, 'M', u'ΰ'),
    (0x1FE4, 'V'),
    (0x1FE8, 'M', u'ῠ'),
    (0x1FE9, 'M', u'ῡ'),
    (0x1FEA, 'M', u'ὺ'),
    (0x1FEB, 'M', u'ύ'),
    (0x1FEC, 'M', u'ῥ'),
    (0x1FED, '3', u' ̈̀'),
    (0x1FEE, '3', u' ̈́'),
    (0x1FEF, '3', u'`'),
    (0x1FF0, 'X'),
    (0x1FF2, 'M', u'ὼι'),
    (0x1FF3, 'M', u'ωι'),
    (0x1FF4, 'M', u'ώι'),
    (0x1FF5, 'X'),
    (0x1FF6, 'V'),
    (0x1FF7, 'M', u'ῶι'),
    (0x1FF8, 'M', u'ὸ'),
    (0x1FF9, 'M', u'ό'),
    (0x1FFA, 'M', u'ὼ'),
    (0x1FFB, 'M', u'ώ'),
    (0x1FFC, 'M', u'ωι'),
    (0x1FFD, '3', u' ́'),
    (0x1FFE, '3', u' ̔'),
    (0x1FFF, 'X'),
    (0x2000, '3', u' '),
    (0x200B, 'I'),
    (0x200C, 'D', u''),
    (0x200E, 'X'),
    (0x2010, 'V'),
    (0x2011, 'M', u'‐'),
    (0x2012, 'V'),
    (0x2017, '3', u' ̳'),
    (0x2018, 'V'),
    (0x2024, 'X'),
    (0x2027, 'V'),
    (0x2028, 'X'),
    (0x202F, '3', u' '),
    (0x2030, 'V'),
    (0x2033, 'M', u'′′'),
    (0x2034, 'M', u'′′′'),
    (0x2035, 'V'),
    (0x2036, 'M', u'‵‵'),
    (0x2037, 'M', u'‵‵‵'),
    ]

def _seg_21():
    return [
    (0x2038, 'V'),
    (0x203C, '3', u'!!'),
    (0x203D, 'V'),
    (0x203E, '3', u' ̅'),
    (0x203F, 'V'),
    (0x2047, '3', u'??'),
    (0x2048, '3', u'?!'),
    (0x2049, '3', u'!?'),
    (0x204A, 'V'),
    (0x2057, 'M', u'′′′′'),
    (0x2058, 'V'),
    (0x205F, '3', u' '),
    (0x2060, 'I'),
    (0x2061, 'X'),
    (0x2064, 'I'),
    (0x2065, 'X'),
    (0x2070, 'M', u'0'),
    (0x2071, 'M', u'i'),
    (0x2072, 'X'),
    (0x2074, 'M', u'4'),
    (0x2075, 'M', u'5'),
    (0x2076, 'M', u'6'),
    (0x2077, 'M', u'7'),
    (0x2078, 'M', u'8'),
    (0x2079, 'M', u'9'),
    (0x207A, '3', u'+'),
    (0x207B, 'M', u'−'),
    (0x207C, '3', u'='),
    (0x207D, '3', u'('),
    (0x207E, '3', u')'),
    (0x207F, 'M', u'n'),
    (0x2080, 'M', u'0'),
    (0x2081, 'M', u'1'),
    (0x2082, 'M', u'2'),
    (0x2083, 'M', u'3'),
    (0x2084, 'M', u'4'),
    (0x2085, 'M', u'5'),
    (0x2086, 'M', u'6'),
    (0x2087, 'M', u'7'),
    (0x2088, 'M', u'8'),
    (0x2089, 'M', u'9'),
    (0x208A, '3', u'+'),
    (0x208B, 'M', u'−'),
    (0x208C, '3', u'='),
    (0x208D, '3', u'('),
    (0x208E, '3', u')'),
    (0x208F, 'X'),
    (0x2090, 'M', u'a'),
    (0x2091, 'M', u'e'),
    (0x2092, 'M', u'o'),
    (0x2093, 'M', u'x'),
    (0x2094, 'M', u'ə'),
    (0x2095, 'M', u'h'),
    (0x2096, 'M', u'k'),
    (0x2097, 'M', u'l'),
    (0x2098, 'M', u'm'),
    (0x2099, 'M', u'n'),
    (0x209A, 'M', u'p'),
    (0x209B, 'M', u's'),
    (0x209C, 'M', u't'),
    (0x209D, 'X'),
    (0x20A0, 'V'),
    (0x20A8, 'M', u'rs'),
    (0x20A9, 'V'),
    (0x20C0, 'X'),
    (0x20D0, 'V'),
    (0x20F1, 'X'),
    (0x2100, '3', u'a/c'),
    (0x2101, '3', u'a/s'),
    (0x2102, 'M', u'c'),
    (0x2103, 'M', u'°c'),
    (0x2104, 'V'),
    (0x2105, '3', u'c/o'),
    (0x2106, '3', u'c/u'),
    (0x2107, 'M', u'ɛ'),
    (0x2108, 'V'),
    (0x2109, 'M', u'°f'),
    (0x210A, 'M', u'g'),
    (0x210B, 'M', u'h'),
    (0x210F, 'M', u'ħ'),
    (0x2110, 'M', u'i'),
    (0x2112, 'M', u'l'),
    (0x2114, 'V'),
    (0x2115, 'M', u'n'),
    (0x2116, 'M', u'no'),
    (0x2117, 'V'),
    (0x2119, 'M', u'p'),
    (0x211A, 'M', u'q'),
    (0x211B, 'M', u'r'),
    (0x211E, 'V'),
    (0x2120, 'M', u'sm'),
    (0x2121, 'M', u'tel'),
    (0x2122, 'M', u'tm'),
    (0x2123, 'V'),
    (0x2124, 'M', u'z'),
    (0x2125, 'V'),
    (0x2126, 'M', u'ω'),
    (0x2127, 'V'),
    (0x2128, 'M', u'z'),
    (0x2129, 'V'),
    ]

def _seg_22():
    return [
    (0x212A, 'M', u'k'),
    (0x212B, 'M', u'å'),
    (0x212C, 'M', u'b'),
    (0x212D, 'M', u'c'),
    (0x212E, 'V'),
    (0x212F, 'M', u'e'),
    (0x2131, 'M', u'f'),
    (0x2132, 'X'),
    (0x2133, 'M', u'm'),
    (0x2134, 'M', u'o'),
    (0x2135, 'M', u'א'),
    (0x2136, 'M', u'ב'),
    (0x2137, 'M', u'ג'),
    (0x2138, 'M', u'ד'),
    (0x2139, 'M', u'i'),
    (0x213A, 'V'),
    (0x213B, 'M', u'fax'),
    (0x213C, 'M', u'π'),
    (0x213D, 'M', u'γ'),
    (0x213F, 'M', u'π'),
    (0x2140, 'M', u'∑'),
    (0x2141, 'V'),
    (0x2145, 'M', u'd'),
    (0x2147, 'M', u'e'),
    (0x2148, 'M', u'i'),
    (0x2149, 'M', u'j'),
    (0x214A, 'V'),
    (0x2150, 'M', u'1⁄7'),
    (0x2151, 'M', u'1⁄9'),
    (0x2152, 'M', u'1⁄10'),
    (0x2153, 'M', u'1⁄3'),
    (0x2154, 'M', u'2⁄3'),
    (0x2155, 'M', u'1⁄5'),
    (0x2156, 'M', u'2⁄5'),
    (0x2157, 'M', u'3⁄5'),
    (0x2158, 'M', u'4⁄5'),
    (0x2159, 'M', u'1⁄6'),
    (0x215A, 'M', u'5⁄6'),
    (0x215B, 'M', u'1⁄8'),
    (0x215C, 'M', u'3⁄8'),
    (0x215D, 'M', u'5⁄8'),
    (0x215E, 'M', u'7⁄8'),
    (0x215F, 'M', u'1⁄'),
    (0x2160, 'M', u'i'),
    (0x2161, 'M', u'ii'),
    (0x2162, 'M', u'iii'),
    (0x2163, 'M', u'iv'),
    (0x2164, 'M', u'v'),
    (0x2165, 'M', u'vi'),
    (0x2166, 'M', u'vii'),
    (0x2167, 'M', u'viii'),
    (0x2168, 'M', u'ix'),
    (0x2169, 'M', u'x'),
    (0x216A, 'M', u'xi'),
    (0x216B, 'M', u'xii'),
    (0x216C, 'M', u'l'),
    (0x216D, 'M', u'c'),
    (0x216E, 'M', u'd'),
    (0x216F, 'M', u'm'),
    (0x2170, 'M', u'i'),
    (0x2171, 'M', u'ii'),
    (0x2172, 'M', u'iii'),
    (0x2173, 'M', u'iv'),
    (0x2174, 'M', u'v'),
    (0x2175, 'M', u'vi'),
    (0x2176, 'M', u'vii'),
    (0x2177, 'M', u'viii'),
    (0x2178, 'M', u'ix'),
    (0x2179, 'M', u'x'),
    (0x217A, 'M', u'xi'),
    (0x217B, 'M', u'xii'),
    (0x217C, 'M', u'l'),
    (0x217D, 'M', u'c'),
    (0x217E, 'M', u'd'),
    (0x217F, 'M', u'm'),
    (0x2180, 'V'),
    (0x2183, 'X'),
    (0x2184, 'V'),
    (0x2189, 'M', u'0⁄3'),
    (0x218A, 'V'),
    (0x218C, 'X'),
    (0x2190, 'V'),
    (0x222C, 'M', u'∫∫'),
    (0x222D, 'M', u'∫∫∫'),
    (0x222E, 'V'),
    (0x222F, 'M', u'∮∮'),
    (0x2230, 'M', u'∮∮∮'),
    (0x2231, 'V'),
    (0x2260, '3'),
    (0x2261, 'V'),
    (0x226E, '3'),
    (0x2270, 'V'),
    (0x2329, 'M', u'〈'),
    (0x232A, 'M', u'〉'),
    (0x232B, 'V'),
    (0x2427, 'X'),
    (0x2440, 'V'),
    (0x244B, 'X'),
    (0x2460, 'M', u'1'),
    (0x2461, 'M', u'2'),
    ]

def _seg_23():
    return [
    (0x2462, 'M', u'3'),
    (0x2463, 'M', u'4'),
    (0x2464, 'M', u'5'),
    (0x2465, 'M', u'6'),
    (0x2466, 'M', u'7'),
    (0x2467, 'M', u'8'),
    (0x2468, 'M', u'9'),
    (0x2469, 'M', u'10'),
    (0x246A, 'M', u'11'),
    (0x246B, 'M', u'12'),
    (0x246C, 'M', u'13'),
    (0x246D, 'M', u'14'),
    (0x246E, 'M', u'15'),
    (0x246F, 'M', u'16'),
    (0x2470, 'M', u'17'),
    (0x2471, 'M', u'18'),
    (0x2472, 'M', u'19'),
    (0x2473, 'M', u'20'),
    (0x2474, '3', u'(1)'),
    (0x2475, '3', u'(2)'),
    (0x2476, '3', u'(3)'),
    (0x2477, '3', u'(4)'),
    (0x2478, '3', u'(5)'),
    (0x2479, '3', u'(6)'),
    (0x247A, '3', u'(7)'),
    (0x247B, '3', u'(8)'),
    (0x247C, '3', u'(9)'),
    (0x247D, '3', u'(10)'),
    (0x247E, '3', u'(11)'),
    (0x247F, '3', u'(12)'),
    (0x2480, '3', u'(13)'),
    (0x2481, '3', u'(14)'),
    (0x2482, '3', u'(15)'),
    (0x2483, '3', u'(16)'),
    (0x2484, '3', u'(17)'),
    (0x2485, '3', u'(18)'),
    (0x2486, '3', u'(19)'),
    (0x2487, '3', u'(20)'),
    (0x2488, 'X'),
    (0x249C, '3', u'(a)'),
    (0x249D, '3', u'(b)'),
    (0x249E, '3', u'(c)'),
    (0x249F, '3', u'(d)'),
    (0x24A0, '3', u'(e)'),
    (0x24A1, '3', u'(f)'),
    (0x24A2, '3', u'(g)'),
    (0x24A3, '3', u'(h)'),
    (0x24A4, '3', u'(i)'),
    (0x24A5, '3', u'(j)'),
    (0x24A6, '3', u'(k)'),
    (0x24A7, '3', u'(l)'),
    (0x24A8, '3', u'(m)'),
    (0x24A9, '3', u'(n)'),
    (0x24AA, '3', u'(o)'),
    (0x24AB, '3', u'(p)'),
    (0x24AC, '3', u'(q)'),
    (0x24AD, '3', u'(r)'),
    (0x24AE, '3', u'(s)'),
    (0x24AF, '3', u'(t)'),
    (0x24B0, '3', u'(u)'),
    (0x24B1, '3', u'(v)'),
    (0x24B2, '3', u'(w)'),
    (0x24B3, '3', u'(x)'),
    (0x24B4, '3', u'(y)'),
    (0x24B5, '3', u'(z)'),
    (0x24B6, 'M', u'a'),
    (0x24B7, 'M', u'b'),
    (0x24B8, 'M', u'c'),
    (0x24B9, 'M', u'd'),
    (0x24BA, 'M', u'e'),
    (0x24BB, 'M', u'f'),
    (0x24BC, 'M', u'g'),
    (0x24BD, 'M', u'h'),
    (0x24BE, 'M', u'i'),
    (0x24BF, 'M', u'j'),
    (0x24C0, 'M', u'k'),
    (0x24C1, 'M', u'l'),
    (0x24C2, 'M', u'm'),
    (0x24C3, 'M', u'n'),
    (0x24C4, 'M', u'o'),
    (0x24C5, 'M', u'p'),
    (0x24C6, 'M', u'q'),
    (0x24C7, 'M', u'r'),
    (0x24C8, 'M', u's'),
    (0x24C9, 'M', u't'),
    (0x24CA, 'M', u'u'),
    (0x24CB, 'M', u'v'),
    (0x24CC, 'M', u'w'),
    (0x24CD, 'M', u'x'),
    (0x24CE, 'M', u'y'),
    (0x24CF, 'M', u'z'),
    (0x24D0, 'M', u'a'),
    (0x24D1, 'M', u'b'),
    (0x24D2, 'M', u'c'),
    (0x24D3, 'M', u'd'),
    (0x24D4, 'M', u'e'),
    (0x24D5, 'M', u'f'),
    (0x24D6, 'M', u'g'),
    (0x24D7, 'M', u'h'),
    (0x24D8, 'M', u'i'),
    ]

def _seg_24():
    return [
    (0x24D9, 'M', u'j'),
    (0x24DA, 'M', u'k'),
    (0x24DB, 'M', u'l'),
    (0x24DC, 'M', u'm'),
    (0x24DD, 'M', u'n'),
    (0x24DE, 'M', u'o'),
    (0x24DF, 'M', u'p'),
    (0x24E0, 'M', u'q'),
    (0x24E1, 'M', u'r'),
    (0x24E2, 'M', u's'),
    (0x24E3, 'M', u't'),
    (0x24E4, 'M', u'u'),
    (0x24E5, 'M', u'v'),
    (0x24E6, 'M', u'w'),
    (0x24E7, 'M', u'x'),
    (0x24E8, 'M', u'y'),
    (0x24E9, 'M', u'z'),
    (0x24EA, 'M', u'0'),
    (0x24EB, 'V'),
    (0x2A0C, 'M', u'∫∫∫∫'),
    (0x2A0D, 'V'),
    (0x2A74, '3', u'::='),
    (0x2A75, '3', u'=='),
    (0x2A76, '3', u'==='),
    (0x2A77, 'V'),
    (0x2ADC, 'M', u'⫝̸'),
    (0x2ADD, 'V'),
    (0x2B74, 'X'),
    (0x2B76, 'V'),
    (0x2B96, 'X'),
    (0x2B98, 'V'),
    (0x2BC9, 'X'),
    (0x2BCA, 'V'),
    (0x2BFF, 'X'),
    (0x2C00, 'M', u'ⰰ'),
    (0x2C01, 'M', u'ⰱ'),
    (0x2C02, 'M', u'ⰲ'),
    (0x2C03, 'M', u'ⰳ'),
    (0x2C04, 'M', u'ⰴ'),
    (0x2C05, 'M', u'ⰵ'),
    (0x2C06, 'M', u'ⰶ'),
    (0x2C07, 'M', u'ⰷ'),
    (0x2C08, 'M', u'ⰸ'),
    (0x2C09, 'M', u'ⰹ'),
    (0x2C0A, 'M', u'ⰺ'),
    (0x2C0B, 'M', u'ⰻ'),
    (0x2C0C, 'M', u'ⰼ'),
    (0x2C0D, 'M', u'ⰽ'),
    (0x2C0E, 'M', u'ⰾ'),
    (0x2C0F, 'M', u'ⰿ'),
    (0x2C10, 'M', u'ⱀ'),
    (0x2C11, 'M', u'ⱁ'),
    (0x2C12, 'M', u'ⱂ'),
    (0x2C13, 'M', u'ⱃ'),
    (0x2C14, 'M', u'ⱄ'),
    (0x2C15, 'M', u'ⱅ'),
    (0x2C16, 'M', u'ⱆ'),
    (0x2C17, 'M', u'ⱇ'),
    (0x2C18, 'M', u'ⱈ'),
    (0x2C19, 'M', u'ⱉ'),
    (0x2C1A, 'M', u'ⱊ'),
    (0x2C1B, 'M', u'ⱋ'),
    (0x2C1C, 'M', u'ⱌ'),
    (0x2C1D, 'M', u'ⱍ'),
    (0x2C1E, 'M', u'ⱎ'),
    (0x2C1F, 'M', u'ⱏ'),
    (0x2C20, 'M', u'ⱐ'),
    (0x2C21, 'M', u'ⱑ'),
    (0x2C22, 'M', u'ⱒ'),
    (0x2C23, 'M', u'ⱓ'),
    (0x2C24, 'M', u'ⱔ'),
    (0x2C25, 'M', u'ⱕ'),
    (0x2C26, 'M', u'ⱖ'),
    (0x2C27, 'M', u'ⱗ'),
    (0x2C28, 'M', u'ⱘ'),
    (0x2C29, 'M', u'ⱙ'),
    (0x2C2A, 'M', u'ⱚ'),
    (0x2C2B, 'M', u'ⱛ'),
    (0x2C2C, 'M', u'ⱜ'),
    (0x2C2D, 'M', u'ⱝ'),
    (0x2C2E, 'M', u'ⱞ'),
    (0x2C2F, 'X'),
    (0x2C30, 'V'),
    (0x2C5F, 'X'),
    (0x2C60, 'M', u'ⱡ'),
    (0x2C61, 'V'),
    (0x2C62, 'M', u'ɫ'),
    (0x2C63, 'M', u'ᵽ'),
    (0x2C64, 'M', u'ɽ'),
    (0x2C65, 'V'),
    (0x2C67, 'M', u'ⱨ'),
    (0x2C68, 'V'),
    (0x2C69, 'M', u'ⱪ'),
    (0x2C6A, 'V'),
    (0x2C6B, 'M', u'ⱬ'),
    (0x2C6C, 'V'),
    (0x2C6D, 'M', u'ɑ'),
    (0x2C6E, 'M', u'ɱ'),
    (0x2C6F, 'M', u'ɐ'),
    (0x2C70, 'M', u'ɒ'),
    ]

def _seg_25():
    return [
    (0x2C71, 'V'),
    (0x2C72, 'M', u'ⱳ'),
    (0x2C73, 'V'),
    (0x2C75, 'M', u'ⱶ'),
    (0x2C76, 'V'),
    (0x2C7C, 'M', u'j'),
    (0x2C7D, 'M', u'v'),
    (0x2C7E, 'M', u'ȿ'),
    (0x2C7F, 'M', u'ɀ'),
    (0x2C80, 'M', u'ⲁ'),
    (0x2C81, 'V'),
    (0x2C82, 'M', u'ⲃ'),
    (0x2C83, 'V'),
    (0x2C84, 'M', u'ⲅ'),
    (0x2C85, 'V'),
    (0x2C86, 'M', u'ⲇ'),
    (0x2C87, 'V'),
    (0x2C88, 'M', u'ⲉ'),
    (0x2C89, 'V'),
    (0x2C8A, 'M', u'ⲋ'),
    (0x2C8B, 'V'),
    (0x2C8C, 'M', u'ⲍ'),
    (0x2C8D, 'V'),
    (0x2C8E, 'M', u'ⲏ'),
    (0x2C8F, 'V'),
    (0x2C90, 'M', u'ⲑ'),
    (0x2C91, 'V'),
    (0x2C92, 'M', u'ⲓ'),
    (0x2C93, 'V'),
    (0x2C94, 'M', u'ⲕ'),
    (0x2C95, 'V'),
    (0x2C96, 'M', u'ⲗ'),
    (0x2C97, 'V'),
    (0x2C98, 'M', u'ⲙ'),
    (0x2C99, 'V'),
    (0x2C9A, 'M', u'ⲛ'),
    (0x2C9B, 'V'),
    (0x2C9C, 'M', u'ⲝ'),
    (0x2C9D, 'V'),
    (0x2C9E, 'M', u'ⲟ'),
    (0x2C9F, 'V'),
    (0x2CA0, 'M', u'ⲡ'),
    (0x2CA1, 'V'),
    (0x2CA2, 'M', u'ⲣ'),
    (0x2CA3, 'V'),
    (0x2CA4, 'M', u'ⲥ'),
    (0x2CA5, 'V'),
    (0x2CA6, 'M', u'ⲧ'),
    (0x2CA7, 'V'),
    (0x2CA8, 'M', u'ⲩ'),
    (0x2CA9, 'V'),
    (0x2CAA, 'M', u'ⲫ'),
    (0x2CAB, 'V'),
    (0x2CAC, 'M', u'ⲭ'),
    (0x2CAD, 'V'),
    (0x2CAE, 'M', u'ⲯ'),
    (0x2CAF, 'V'),
    (0x2CB0, 'M', u'ⲱ'),
    (0x2CB1, 'V'),
    (0x2CB2, 'M', u'ⲳ'),
    (0x2CB3, 'V'),
    (0x2CB4, 'M', u'ⲵ'),
    (0x2CB5, 'V'),
    (0x2CB6, 'M', u'ⲷ'),
    (0x2CB7, 'V'),
    (0x2CB8, 'M', u'ⲹ'),
    (0x2CB9, 'V'),
    (0x2CBA, 'M', u'ⲻ'),
    (0x2CBB, 'V'),
    (0x2CBC, 'M', u'ⲽ'),
    (0x2CBD, 'V'),
    (0x2CBE, 'M', u'ⲿ'),
    (0x2CBF, 'V'),
    (0x2CC0, 'M', u'ⳁ'),
    (0x2CC1, 'V'),
    (0x2CC2, 'M', u'ⳃ'),
    (0x2CC3, 'V'),
    (0x2CC4, 'M', u'ⳅ'),
    (0x2CC5, 'V'),
    (0x2CC6, 'M', u'ⳇ'),
    (0x2CC7, 'V'),
    (0x2CC8, 'M', u'ⳉ'),
    (0x2CC9, 'V'),
    (0x2CCA, 'M', u'ⳋ'),
    (0x2CCB, 'V'),
    (0x2CCC, 'M', u'ⳍ'),
    (0x2CCD, 'V'),
    (0x2CCE, 'M', u'ⳏ'),
    (0x2CCF, 'V'),
    (0x2CD0, 'M', u'ⳑ'),
    (0x2CD1, 'V'),
    (0x2CD2, 'M', u'ⳓ'),
    (0x2CD3, 'V'),
    (0x2CD4, 'M', u'ⳕ'),
    (0x2CD5, 'V'),
    (0x2CD6, 'M', u'ⳗ'),
    (0x2CD7, 'V'),
    (0x2CD8, 'M', u'ⳙ'),
    (0x2CD9, 'V'),
    (0x2CDA, 'M', u'ⳛ'),
    ]

def _seg_26():
    return [
    (0x2CDB, 'V'),
    (0x2CDC, 'M', u'ⳝ'),
    (0x2CDD, 'V'),
    (0x2CDE, 'M', u'ⳟ'),
    (0x2CDF, 'V'),
    (0x2CE0, 'M', u'ⳡ'),
    (0x2CE1, 'V'),
    (0x2CE2, 'M', u'ⳣ'),
    (0x2CE3, 'V'),
    (0x2CEB, 'M', u'ⳬ'),
    (0x2CEC, 'V'),
    (0x2CED, 'M', u'ⳮ'),
    (0x2CEE, 'V'),
    (0x2CF2, 'M', u'ⳳ'),
    (0x2CF3, 'V'),
    (0x2CF4, 'X'),
    (0x2CF9, 'V'),
    (0x2D26, 'X'),
    (0x2D27, 'V'),
    (0x2D28, 'X'),
    (0x2D2D, 'V'),
    (0x2D2E, 'X'),
    (0x2D30, 'V'),
    (0x2D68, 'X'),
    (0x2D6F, 'M', u'ⵡ'),
    (0x2D70, 'V'),
    (0x2D71, 'X'),
    (0x2D7F, 'V'),
    (0x2D97, 'X'),
    (0x2DA0, 'V'),
    (0x2DA7, 'X'),
    (0x2DA8, 'V'),
    (0x2DAF, 'X'),
    (0x2DB0, 'V'),
    (0x2DB7, 'X'),
    (0x2DB8, 'V'),
    (0x2DBF, 'X'),
    (0x2DC0, 'V'),
    (0x2DC7, 'X'),
    (0x2DC8, 'V'),
    (0x2DCF, 'X'),
    (0x2DD0, 'V'),
    (0x2DD7, 'X'),
    (0x2DD8, 'V'),
    (0x2DDF, 'X'),
    (0x2DE0, 'V'),
    (0x2E4F, 'X'),
    (0x2E80, 'V'),
    (0x2E9A, 'X'),
    (0x2E9B, 'V'),
    (0x2E9F, 'M', u'母'),
    (0x2EA0, 'V'),
    (0x2EF3, 'M', u'龟'),
    (0x2EF4, 'X'),
    (0x2F00, 'M', u'一'),
    (0x2F01, 'M', u'丨'),
    (0x2F02, 'M', u'丶'),
    (0x2F03, 'M', u'丿'),
    (0x2F04, 'M', u'乙'),
    (0x2F05, 'M', u'亅'),
    (0x2F06, 'M', u'二'),
    (0x2F07, 'M', u'亠'),
    (0x2F08, 'M', u'人'),
    (0x2F09, 'M', u'儿'),
    (0x2F0A, 'M', u'入'),
    (0x2F0B, 'M', u'八'),
    (0x2F0C, 'M', u'冂'),
    (0x2F0D, 'M', u'冖'),
    (0x2F0E, 'M', u'冫'),
    (0x2F0F, 'M', u'几'),
    (0x2F10, 'M', u'凵'),
    (0x2F11, 'M', u'刀'),
    (0x2F12, 'M', u'力'),
    (0x2F13, 'M', u'勹'),
    (0x2F14, 'M', u'匕'),
    (0x2F15, 'M', u'匚'),
    (0x2F16, 'M', u'匸'),
    (0x2F17, 'M', u'十'),
    (0x2F18, 'M', u'卜'),
    (0x2F19, 'M', u'卩'),
    (0x2F1A, 'M', u'厂'),
    (0x2F1B, 'M', u'厶'),
    (0x2F1C, 'M', u'又'),
    (0x2F1D, 'M', u'口'),
    (0x2F1E, 'M', u'囗'),
    (0x2F1F, 'M', u'土'),
    (0x2F20, 'M', u'士'),
    (0x2F21, 'M', u'夂'),
    (0x2F22, 'M', u'夊'),
    (0x2F23, 'M', u'夕'),
    (0x2F24, 'M', u'大'),
    (0x2F25, 'M', u'女'),
    (0x2F26, 'M', u'子'),
    (0x2F27, 'M', u'宀'),
    (0x2F28, 'M', u'寸'),
    (0x2F29, 'M', u'小'),
    (0x2F2A, 'M', u'尢'),
    (0x2F2B, 'M', u'尸'),
    (0x2F2C, 'M', u'屮'),
    (0x2F2D, 'M', u'山'),
    ]

def _seg_27():
    return [
    (0x2F2E, 'M', u'巛'),
    (0x2F2F, 'M', u'工'),
    (0x2F30, 'M', u'己'),
    (0x2F31, 'M', u'巾'),
    (0x2F32, 'M', u'干'),
    (0x2F33, 'M', u'幺'),
    (0x2F34, 'M', u'广'),
    (0x2F35, 'M', u'廴'),
    (0x2F36, 'M', u'廾'),
    (0x2F37, 'M', u'弋'),
    (0x2F38, 'M', u'弓'),
    (0x2F39, 'M', u'彐'),
    (0x2F3A, 'M', u'彡'),
    (0x2F3B, 'M', u'彳'),
    (0x2F3C, 'M', u'心'),
    (0x2F3D, 'M', u'戈'),
    (0x2F3E, 'M', u'戶'),
    (0x2F3F, 'M', u'手'),
    (0x2F40, 'M', u'支'),
    (0x2F41, 'M', u'攴'),
    (0x2F42, 'M', u'文'),
    (0x2F43, 'M', u'斗'),
    (0x2F44, 'M', u'斤'),
    (0x2F45, 'M', u'方'),
    (0x2F46, 'M', u'无'),
    (0x2F47, 'M', u'日'),
    (0x2F48, 'M', u'曰'),
    (0x2F49, 'M', u'月'),
    (0x2F4A, 'M', u'木'),
    (0x2F4B, 'M', u'欠'),
    (0x2F4C, 'M', u'止'),
    (0x2F4D, 'M', u'歹'),
    (0x2F4E, 'M', u'殳'),
    (0x2F4F, 'M', u'毋'),
    (0x2F50, 'M', u'比'),
    (0x2F51, 'M', u'毛'),
    (0x2F52, 'M', u'氏'),
    (0x2F53, 'M', u'气'),
    (0x2F54, 'M', u'水'),
    (0x2F55, 'M', u'火'),
    (0x2F56, 'M', u'爪'),
    (0x2F57, 'M', u'父'),
    (0x2F58, 'M', u'爻'),
    (0x2F59, 'M', u'爿'),
    (0x2F5A, 'M', u'片'),
    (0x2F5B, 'M', u'牙'),
    (0x2F5C, 'M', u'牛'),
    (0x2F5D, 'M', u'犬'),
    (0x2F5E, 'M', u'玄'),
    (0x2F5F, 'M', u'玉'),
    (0x2F60, 'M', u'瓜'),
    (0x2F61, 'M', u'瓦'),
    (0x2F62, 'M', u'甘'),
    (0x2F63, 'M', u'生'),
    (0x2F64, 'M', u'用'),
    (0x2F65, 'M', u'田'),
    (0x2F66, 'M', u'疋'),
    (0x2F67, 'M', u'疒'),
    (0x2F68, 'M', u'癶'),
    (0x2F69, 'M', u'白'),
    (0x2F6A, 'M', u'皮'),
    (0x2F6B, 'M', u'皿'),
    (0x2F6C, 'M', u'目'),
    (0x2F6D, 'M', u'矛'),
    (0x2F6E, 'M', u'矢'),
    (0x2F6F, 'M', u'石'),
    (0x2F70, 'M', u'示'),
    (0x2F71, 'M', u'禸'),
    (0x2F72, 'M', u'禾'),
    (0x2F73, 'M', u'穴'),
    (0x2F74, 'M', u'立'),
    (0x2F75, 'M', u'竹'),
    (0x2F76, 'M', u'米'),
    (0x2F77, 'M', u'糸'),
    (0x2F78, 'M', u'缶'),
    (0x2F79, 'M', u'网'),
    (0x2F7A, 'M', u'羊'),
    (0x2F7B, 'M', u'羽'),
    (0x2F7C, 'M', u'老'),
    (0x2F7D, 'M', u'而'),
    (0x2F7E, 'M', u'耒'),
    (0x2F7F, 'M', u'耳'),
    (0x2F80, 'M', u'聿'),
    (0x2F81, 'M', u'肉'),
    (0x2F82, 'M', u'臣'),
    (0x2F83, 'M', u'自'),
    (0x2F84, 'M', u'至'),
    (0x2F85, 'M', u'臼'),
    (0x2F86, 'M', u'舌'),
    (0x2F87, 'M', u'舛'),
    (0x2F88, 'M', u'舟'),
    (0x2F89, 'M', u'艮'),
    (0x2F8A, 'M', u'色'),
    (0x2F8B, 'M', u'艸'),
    (0x2F8C, 'M', u'虍'),
    (0x2F8D, 'M', u'虫'),
    (0x2F8E, 'M', u'血'),
    (0x2F8F, 'M', u'行'),
    (0x2F90, 'M', u'衣'),
    (0x2F91, 'M', u'襾'),
    ]

def _seg_28():
    return [
    (0x2F92, 'M', u'見'),
    (0x2F93, 'M', u'角'),
    (0x2F94, 'M', u'言'),
    (0x2F95, 'M', u'谷'),
    (0x2F96, 'M', u'豆'),
    (0x2F97, 'M', u'豕'),
    (0x2F98, 'M', u'豸'),
    (0x2F99, 'M', u'貝'),
    (0x2F9A, 'M', u'赤'),
    (0x2F9B, 'M', u'走'),
    (0x2F9C, 'M', u'足'),
    (0x2F9D, 'M', u'身'),
    (0x2F9E, 'M', u'車'),
    (0x2F9F, 'M', u'辛'),
    (0x2FA0, 'M', u'辰'),
    (0x2FA1, 'M', u'辵'),
    (0x2FA2, 'M', u'邑'),
    (0x2FA3, 'M', u'酉'),
    (0x2FA4, 'M', u'釆'),
    (0x2FA5, 'M', u'里'),
    (0x2FA6, 'M', u'金'),
    (0x2FA7, 'M', u'長'),
    (0x2FA8, 'M', u'門'),
    (0x2FA9, 'M', u'阜'),
    (0x2FAA, 'M', u'隶'),
    (0x2FAB, 'M', u'隹'),
    (0x2FAC, 'M', u'雨'),
    (0x2FAD, 'M', u'靑'),
    (0x2FAE, 'M', u'非'),
    (0x2FAF, 'M', u'面'),
    (0x2FB0, 'M', u'革'),
    (0x2FB1, 'M', u'韋'),
    (0x2FB2, 'M', u'韭'),
    (0x2FB3, 'M', u'音'),
    (0x2FB4, 'M', u'頁'),
    (0x2FB5, 'M', u'風'),
    (0x2FB6, 'M', u'飛'),
    (0x2FB7, 'M', u'食'),
    (0x2FB8, 'M', u'首'),
    (0x2FB9, 'M', u'香'),
    (0x2FBA, 'M', u'馬'),
    (0x2FBB, 'M', u'骨'),
    (0x2FBC, 'M', u'高'),
    (0x2FBD, 'M', u'髟'),
    (0x2FBE, 'M', u'鬥'),
    (0x2FBF, 'M', u'鬯'),
    (0x2FC0, 'M', u'鬲'),
    (0x2FC1, 'M', u'鬼'),
    (0x2FC2, 'M', u'魚'),
    (0x2FC3, 'M', u'鳥'),
    (0x2FC4, 'M', u'鹵'),
    (0x2FC5, 'M', u'鹿'),
    (0x2FC6, 'M', u'麥'),
    (0x2FC7, 'M', u'麻'),
    (0x2FC8, 'M', u'黃'),
    (0x2FC9, 'M', u'黍'),
    (0x2FCA, 'M', u'黑'),
    (0x2FCB, 'M', u'黹'),
    (0x2FCC, 'M', u'黽'),
    (0x2FCD, 'M', u'鼎'),
    (0x2FCE, 'M', u'鼓'),
    (0x2FCF, 'M', u'鼠'),
    (0x2FD0, 'M', u'鼻'),
    (0x2FD1, 'M', u'齊'),
    (0x2FD2, 'M', u'齒'),
    (0x2FD3, 'M', u'龍'),
    (0x2FD4, 'M', u'龜'),
    (0x2FD5, 'M', u'龠'),
    (0x2FD6, 'X'),
    (0x3000, '3', u' '),
    (0x3001, 'V'),
    (0x3002, 'M', u'.'),
    (0x3003, 'V'),
    (0x3036, 'M', u'〒'),
    (0x3037, 'V'),
    (0x3038, 'M', u'十'),
    (0x3039, 'M', u'卄'),
    (0x303A, 'M', u'卅'),
    (0x303B, 'V'),
    (0x3040, 'X'),
    (0x3041, 'V'),
    (0x3097, 'X'),
    (0x3099, 'V'),
    (0x309B, '3', u' ゙'),
    (0x309C, '3', u' ゚'),
    (0x309D, 'V'),
    (0x309F, 'M', u'より'),
    (0x30A0, 'V'),
    (0x30FF, 'M', u'コト'),
    (0x3100, 'X'),
    (0x3105, 'V'),
    (0x3130, 'X'),
    (0x3131, 'M', u'ᄀ'),
    (0x3132, 'M', u'ᄁ'),
    (0x3133, 'M', u'ᆪ'),
    (0x3134, 'M', u'ᄂ'),
    (0x3135, 'M', u'ᆬ'),
    (0x3136, 'M', u'ᆭ'),
    (0x3137, 'M', u'ᄃ'),
    (0x3138, 'M', u'ᄄ'),
    ]

def _seg_29():
    return [
    (0x3139, 'M', u'ᄅ'),
    (0x313A, 'M', u'ᆰ'),
    (0x313B, 'M', u'ᆱ'),
    (0x313C, 'M', u'ᆲ'),
    (0x313D, 'M', u'ᆳ'),
    (0x313E, 'M', u'ᆴ'),
    (0x313F, 'M', u'ᆵ'),
    (0x3140, 'M', u'ᄚ'),
    (0x3141, 'M', u'ᄆ'),
    (0x3142, 'M', u'ᄇ'),
    (0x3143, 'M', u'ᄈ'),
    (0x3144, 'M', u'ᄡ'),
    (0x3145, 'M', u'ᄉ'),
    (0x3146, 'M', u'ᄊ'),
    (0x3147, 'M', u'ᄋ'),
    (0x3148, 'M', u'ᄌ'),
    (0x3149, 'M', u'ᄍ'),
    (0x314A, 'M', u'ᄎ'),
    (0x314B, 'M', u'ᄏ'),
    (0x314C, 'M', u'ᄐ'),
    (0x314D, 'M', u'ᄑ'),
    (0x314E, 'M', u'ᄒ'),
    (0x314F, 'M', u'ᅡ'),
    (0x3150, 'M', u'ᅢ'),
    (0x3151, 'M', u'ᅣ'),
    (0x3152, 'M', u'ᅤ'),
    (0x3153, 'M', u'ᅥ'),
    (0x3154, 'M', u'ᅦ'),
    (0x3155, 'M', u'ᅧ'),
    (0x3156, 'M', u'ᅨ'),
    (0x3157, 'M', u'ᅩ'),
    (0x3158, 'M', u'ᅪ'),
    (0x3159, 'M', u'ᅫ'),
    (0x315A, 'M', u'ᅬ'),
    (0x315B, 'M', u'ᅭ'),
    (0x315C, 'M', u'ᅮ'),
    (0x315D, 'M', u'ᅯ'),
    (0x315E, 'M', u'ᅰ'),
    (0x315F, 'M', u'ᅱ'),
    (0x3160, 'M', u'ᅲ'),
    (0x3161, 'M', u'ᅳ'),
    (0x3162, 'M', u'ᅴ'),
    (0x3163, 'M', u'ᅵ'),
    (0x3164, 'X'),
    (0x3165, 'M', u'ᄔ'),
    (0x3166, 'M', u'ᄕ'),
    (0x3167, 'M', u'ᇇ'),
    (0x3168, 'M', u'ᇈ'),
    (0x3169, 'M', u'ᇌ'),
    (0x316A, 'M', u'ᇎ'),
    (0x316B, 'M', u'ᇓ'),
    (0x316C, 'M', u'ᇗ'),
    (0x316D, 'M', u'ᇙ'),
    (0x316E, 'M', u'ᄜ'),
    (0x316F, 'M', u'ᇝ'),
    (0x3170, 'M', u'ᇟ'),
    (0x3171, 'M', u'ᄝ'),
    (0x3172, 'M', u'ᄞ'),
    (0x3173, 'M', u'ᄠ'),
    (0x3174, 'M', u'ᄢ'),
    (0x3175, 'M', u'ᄣ'),
    (0x3176, 'M', u'ᄧ'),
    (0x3177, 'M', u'ᄩ'),
    (0x3178, 'M', u'ᄫ'),
    (0x3179, 'M', u'ᄬ'),
    (0x317A, 'M', u'ᄭ'),
    (0x317B, 'M', u'ᄮ'),
    (0x317C, 'M', u'ᄯ'),
    (0x317D, 'M', u'ᄲ'),
    (0x317E, 'M', u'ᄶ'),
    (0x317F, 'M', u'ᅀ'),
    (0x3180, 'M', u'ᅇ'),
    (0x3181, 'M', u'ᅌ'),
    (0x3182, 'M', u'ᇱ'),
    (0x3183, 'M', u'ᇲ'),
    (0x3184, 'M', u'ᅗ'),
    (0x3185, 'M', u'ᅘ'),
    (0x3186, 'M', u'ᅙ'),
    (0x3187, 'M', u'ᆄ'),
    (0x3188, 'M', u'ᆅ'),
    (0x3189, 'M', u'ᆈ'),
    (0x318A, 'M', u'ᆑ'),
    (0x318B, 'M', u'ᆒ'),
    (0x318C, 'M', u'ᆔ'),
    (0x318D, 'M', u'ᆞ'),
    (0x318E, 'M', u'ᆡ'),
    (0x318F, 'X'),
    (0x3190, 'V'),
    (0x3192, 'M', u'一'),
    (0x3193, 'M', u'二'),
    (0x3194, 'M', u'三'),
    (0x3195, 'M', u'四'),
    (0x3196, 'M', u'上'),
    (0x3197, 'M', u'中'),
    (0x3198, 'M', u'下'),
    (0x3199, 'M', u'甲'),
    (0x319A, 'M', u'乙'),
    (0x319B, 'M', u'丙'),
    (0x319C, 'M', u'丁'),
    (0x319D, 'M', u'天'),
    ]

def _seg_30():
    return [
    (0x319E, 'M', u'地'),
    (0x319F, 'M', u'人'),
    (0x31A0, 'V'),
    (0x31BB, 'X'),
    (0x31C0, 'V'),
    (0x31E4, 'X'),
    (0x31F0, 'V'),
    (0x3200, '3', u'(ᄀ)'),
    (0x3201, '3', u'(ᄂ)'),
    (0x3202, '3', u'(ᄃ)'),
    (0x3203, '3', u'(ᄅ)'),
    (0x3204, '3', u'(ᄆ)'),
    (0x3205, '3', u'(ᄇ)'),
    (0x3206, '3', u'(ᄉ)'),
    (0x3207, '3', u'(ᄋ)'),
    (0x3208, '3', u'(ᄌ)'),
    (0x3209, '3', u'(ᄎ)'),
    (0x320A, '3', u'(ᄏ)'),
    (0x320B, '3', u'(ᄐ)'),
    (0x320C, '3', u'(ᄑ)'),
    (0x320D, '3', u'(ᄒ)'),
    (0x320E, '3', u'(가)'),
    (0x320F, '3', u'(나)'),
    (0x3210, '3', u'(다)'),
    (0x3211, '3', u'(라)'),
    (0x3212, '3', u'(마)'),
    (0x3213, '3', u'(바)'),
    (0x3214, '3', u'(사)'),
    (0x3215, '3', u'(아)'),
    (0x3216, '3', u'(자)'),
    (0x3217, '3', u'(차)'),
    (0x3218, '3', u'(카)'),
    (0x3219, '3', u'(타)'),
    (0x321A, '3', u'(파)'),
    (0x321B, '3', u'(하)'),
    (0x321C, '3', u'(주)'),
    (0x321D, '3', u'(오전)'),
    (0x321E, '3', u'(오후)'),
    (0x321F, 'X'),
    (0x3220, '3', u'(一)'),
    (0x3221, '3', u'(二)'),
    (0x3222, '3', u'(三)'),
    (0x3223, '3', u'(四)'),
    (0x3224, '3', u'(五)'),
    (0x3225, '3', u'(六)'),
    (0x3226, '3', u'(七)'),
    (0x3227, '3', u'(八)'),
    (0x3228, '3', u'(九)'),
    (0x3229, '3', u'(十)'),
    (0x322A, '3', u'(月)'),
    (0x322B, '3', u'(火)'),
    (0x322C, '3', u'(水)'),
    (0x322D, '3', u'(木)'),
    (0x322E, '3', u'(金)'),
    (0x322F, '3', u'(土)'),
    (0x3230, '3', u'(日)'),
    (0x3231, '3', u'(株)'),
    (0x3232, '3', u'(有)'),
    (0x3233, '3', u'(社)'),
    (0x3234, '3', u'(名)'),
    (0x3235, '3', u'(特)'),
    (0x3236, '3', u'(財)'),
    (0x3237, '3', u'(祝)'),
    (0x3238, '3', u'(労)'),
    (0x3239, '3', u'(代)'),
    (0x323A, '3', u'(呼)'),
    (0x323B, '3', u'(学)'),
    (0x323C, '3', u'(監)'),
    (0x323D, '3', u'(企)'),
    (0x323E, '3', u'(資)'),
    (0x323F, '3', u'(協)'),
    (0x3240, '3', u'(祭)'),
    (0x3241, '3', u'(休)'),
    (0x3242, '3', u'(自)'),
    (0x3243, '3', u'(至)'),
    (0x3244, 'M', u'問'),
    (0x3245, 'M', u'幼'),
    (0x3246, 'M', u'文'),
    (0x3247, 'M', u'箏'),
    (0x3248, 'V'),
    (0x3250, 'M', u'pte'),
    (0x3251, 'M', u'21'),
    (0x3252, 'M', u'22'),
    (0x3253, 'M', u'23'),
    (0x3254, 'M', u'24'),
    (0x3255, 'M', u'25'),
    (0x3256, 'M', u'26'),
    (0x3257, 'M', u'27'),
    (0x3258, 'M', u'28'),
    (0x3259, 'M', u'29'),
    (0x325A, 'M', u'30'),
    (0x325B, 'M', u'31'),
    (0x325C, 'M', u'32'),
    (0x325D, 'M', u'33'),
    (0x325E, 'M', u'34'),
    (0x325F, 'M', u'35'),
    (0x3260, 'M', u'ᄀ'),
    (0x3261, 'M', u'ᄂ'),
    (0x3262, 'M', u'ᄃ'),
    (0x3263, 'M', u'ᄅ'),
    ]

def _seg_31():
    return [
    (0x3264, 'M', u'ᄆ'),
    (0x3265, 'M', u'ᄇ'),
    (0x3266, 'M', u'ᄉ'),
    (0x3267, 'M', u'ᄋ'),
    (0x3268, 'M', u'ᄌ'),
    (0x3269, 'M', u'ᄎ'),
    (0x326A, 'M', u'ᄏ'),
    (0x326B, 'M', u'ᄐ'),
    (0x326C, 'M', u'ᄑ'),
    (0x326D, 'M', u'ᄒ'),
    (0x326E, 'M', u'가'),
    (0x326F, 'M', u'나'),
    (0x3270, 'M', u'다'),
    (0x3271, 'M', u'라'),
    (0x3272, 'M', u'마'),
    (0x3273, 'M', u'바'),
    (0x3274, 'M', u'사'),
    (0x3275, 'M', u'아'),
    (0x3276, 'M', u'자'),
    (0x3277, 'M', u'차'),
    (0x3278, 'M', u'카'),
    (0x3279, 'M', u'타'),
    (0x327A, 'M', u'파'),
    (0x327B, 'M', u'하'),
    (0x327C, 'M', u'참고'),
    (0x327D, 'M', u'주의'),
    (0x327E, 'M', u'우'),
    (0x327F, 'V'),
    (0x3280, 'M', u'一'),
    (0x3281, 'M', u'二'),
    (0x3282, 'M', u'三'),
    (0x3283, 'M', u'四'),
    (0x3284, 'M', u'五'),
    (0x3285, 'M', u'六'),
    (0x3286, 'M', u'七'),
    (0x3287, 'M', u'八'),
    (0x3288, 'M', u'九'),
    (0x3289, 'M', u'十'),
    (0x328A, 'M', u'月'),
    (0x328B, 'M', u'火'),
    (0x328C, 'M', u'水'),
    (0x328D, 'M', u'木'),
    (0x328E, 'M', u'金'),
    (0x328F, 'M', u'土'),
    (0x3290, 'M', u'日'),
    (0x3291, 'M', u'株'),
    (0x3292, 'M', u'有'),
    (0x3293, 'M', u'社'),
    (0x3294, 'M', u'名'),
    (0x3295, 'M', u'特'),
    (0x3296, 'M', u'財'),
    (0x3297, 'M', u'祝'),
    (0x3298, 'M', u'労'),
    (0x3299, 'M', u'秘'),
    (0x329A, 'M', u'男'),
    (0x329B, 'M', u'女'),
    (0x329C, 'M', u'適'),
    (0x329D, 'M', u'優'),
    (0x329E, 'M', u'印'),
    (0x329F, 'M', u'注'),
    (0x32A0, 'M', u'項'),
    (0x32A1, 'M', u'休'),
    (0x32A2, 'M', u'写'),
    (0x32A3, 'M', u'正'),
    (0x32A4, 'M', u'上'),
    (0x32A5, 'M', u'中'),
    (0x32A6, 'M', u'下'),
    (0x32A7, 'M', u'左'),
    (0x32A8, 'M', u'右'),
    (0x32A9, 'M', u'医'),
    (0x32AA, 'M', u'宗'),
    (0x32AB, 'M', u'学'),
    (0x32AC, 'M', u'監'),
    (0x32AD, 'M', u'企'),
    (0x32AE, 'M', u'資'),
    (0x32AF, 'M', u'協'),
    (0x32B0, 'M', u'夜'),
    (0x32B1, 'M', u'36'),
    (0x32B2, 'M', u'37'),
    (0x32B3, 'M', u'38'),
    (0x32B4, 'M', u'39'),
    (0x32B5, 'M', u'40'),
    (0x32B6, 'M', u'41'),
    (0x32B7, 'M', u'42'),
    (0x32B8, 'M', u'43'),
    (0x32B9, 'M', u'44'),
    (0x32BA, 'M', u'45'),
    (0x32BB, 'M', u'46'),
    (0x32BC, 'M', u'47'),
    (0x32BD, 'M', u'48'),
    (0x32BE, 'M', u'49'),
    (0x32BF, 'M', u'50'),
    (0x32C0, 'M', u'1月'),
    (0x32C1, 'M', u'2月'),
    (0x32C2, 'M', u'3月'),
    (0x32C3, 'M', u'4月'),
    (0x32C4, 'M', u'5月'),
    (0x32C5, 'M', u'6月'),
    (0x32C6, 'M', u'7月'),
    (0x32C7, 'M', u'8月'),
    ]

def _seg_32():
    return [
    (0x32C8, 'M', u'9月'),
    (0x32C9, 'M', u'10月'),
    (0x32CA, 'M', u'11月'),
    (0x32CB, 'M', u'12月'),
    (0x32CC, 'M', u'hg'),
    (0x32CD, 'M', u'erg'),
    (0x32CE, 'M', u'ev'),
    (0x32CF, 'M', u'ltd'),
    (0x32D0, 'M', u'ア'),
    (0x32D1, 'M', u'イ'),
    (0x32D2, 'M', u'ウ'),
    (0x32D3, 'M', u'エ'),
    (0x32D4, 'M', u'オ'),
    (0x32D5, 'M', u'カ'),
    (0x32D6, 'M', u'キ'),
    (0x32D7, 'M', u'ク'),
    (0x32D8, 'M', u'ケ'),
    (0x32D9, 'M', u'コ'),
    (0x32DA, 'M', u'サ'),
    (0x32DB, 'M', u'シ'),
    (0x32DC, 'M', u'ス'),
    (0x32DD, 'M', u'セ'),
    (0x32DE, 'M', u'ソ'),
    (0x32DF, 'M', u'タ'),
    (0x32E0, 'M', u'チ'),
    (0x32E1, 'M', u'ツ'),
    (0x32E2, 'M', u'テ'),
    (0x32E3, 'M', u'ト'),
    (0x32E4, 'M', u'ナ'),
    (0x32E5, 'M', u'ニ'),
    (0x32E6, 'M', u'ヌ'),
    (0x32E7, 'M', u'ネ'),
    (0x32E8, 'M', u'ノ'),
    (0x32E9, 'M', u'ハ'),
    (0x32EA, 'M', u'ヒ'),
    (0x32EB, 'M', u'フ'),
    (0x32EC, 'M', u'ヘ'),
    (0x32ED, 'M', u'ホ'),
    (0x32EE, 'M', u'マ'),
    (0x32EF, 'M', u'ミ'),
    (0x32F0, 'M', u'ム'),
    (0x32F1, 'M', u'メ'),
    (0x32F2, 'M', u'モ'),
    (0x32F3, 'M', u'ヤ'),
    (0x32F4, 'M', u'ユ'),
    (0x32F5, 'M', u'ヨ'),
    (0x32F6, 'M', u'ラ'),
    (0x32F7, 'M', u'リ'),
    (0x32F8, 'M', u'ル'),
    (0x32F9, 'M', u'レ'),
    (0x32FA, 'M', u'ロ'),
    (0x32FB, 'M', u'ワ'),
    (0x32FC, 'M', u'ヰ'),
    (0x32FD, 'M', u'ヱ'),
    (0x32FE, 'M', u'ヲ'),
    (0x32FF, 'X'),
    (0x3300, 'M', u'アパート'),
    (0x3301, 'M', u'アルファ'),
    (0x3302, 'M', u'アンペア'),
    (0x3303, 'M', u'アール'),
    (0x3304, 'M', u'イニング'),
    (0x3305, 'M', u'インチ'),
    (0x3306, 'M', u'ウォン'),
    (0x3307, 'M', u'エスクード'),
    (0x3308, 'M', u'エーカー'),
    (0x3309, 'M', u'オンス'),
    (0x330A, 'M', u'オーム'),
    (0x330B, 'M', u'カイリ'),
    (0x330C, 'M', u'カラット'),
    (0x330D, 'M', u'カロリー'),
    (0x330E, 'M', u'ガロン'),
    (0x330F, 'M', u'ガンマ'),
    (0x3310, 'M', u'ギガ'),
    (0x3311, 'M', u'ギニー'),
    (0x3312, 'M', u'キュリー'),
    (0x3313, 'M', u'ギルダー'),
    (0x3314, 'M', u'キロ'),
    (0x3315, 'M', u'キログラム'),
    (0x3316, 'M', u'キロメートル'),
    (0x3317, 'M', u'キロワット'),
    (0x3318, 'M', u'グラム'),
    (0x3319, 'M', u'グラムトン'),
    (0x331A, 'M', u'クルゼイロ'),
    (0x331B, 'M', u'クローネ'),
    (0x331C, 'M', u'ケース'),
    (0x331D, 'M', u'コルナ'),
    (0x331E, 'M', u'コーポ'),
    (0x331F, 'M', u'サイクル'),
    (0x3320, 'M', u'サンチーム'),
    (0x3321, 'M', u'シリング'),
    (0x3322, 'M', u'センチ'),
    (0x3323, 'M', u'セント'),
    (0x3324, 'M', u'ダース'),
    (0x3325, 'M', u'デシ'),
    (0x3326, 'M', u'ドル'),
    (0x3327, 'M', u'トン'),
    (0x3328, 'M', u'ナノ'),
    (0x3329, 'M', u'ノット'),
    (0x332A, 'M', u'ハイツ'),
    (0x332B, 'M', u'パーセント'),
    ]

def _seg_33():
    return [
    (0x332C, 'M', u'パーツ'),
    (0x332D, 'M', u'バーレル'),
    (0x332E, 'M', u'ピアストル'),
    (0x332F, 'M', u'ピクル'),
    (0x3330, 'M', u'ピコ'),
    (0x3331, 'M', u'ビル'),
    (0x3332, 'M', u'ファラッド'),
    (0x3333, 'M', u'フィート'),
    (0x3334, 'M', u'ブッシェル'),
    (0x3335, 'M', u'フラン'),
    (0x3336, 'M', u'ヘクタール'),
    (0x3337, 'M', u'ペソ'),
    (0x3338, 'M', u'ペニヒ'),
    (0x3339, 'M', u'ヘルツ'),
    (0x333A, 'M', u'ペンス'),
    (0x333B, 'M', u'ページ'),
    (0x333C, 'M', u'ベータ'),
    (0x333D, 'M', u'ポイント'),
    (0x333E, 'M', u'ボルト'),
    (0x333F, 'M', u'ホン'),
    (0x3340, 'M', u'ポンド'),
    (0x3341, 'M', u'ホール'),
    (0x3342, 'M', u'ホーン'),
    (0x3343, 'M', u'マイクロ'),
    (0x3344, 'M', u'マイル'),
    (0x3345, 'M', u'マッハ'),
    (0x3346, 'M', u'マルク'),
    (0x3347, 'M', u'マンション'),
    (0x3348, 'M', u'ミクロン'),
    (0x3349, 'M', u'ミリ'),
    (0x334A, 'M', u'ミリバール'),
    (0x334B, 'M', u'メガ'),
    (0x334C, 'M', u'メガトン'),
    (0x334D, 'M', u'メートル'),
    (0x334E, 'M', u'ヤード'),
    (0x334F, 'M', u'ヤール'),
    (0x3350, 'M', u'ユアン'),
    (0x3351, 'M', u'リットル'),
    (0x3352, 'M', u'リラ'),
    (0x3353, 'M', u'ルピー'),
    (0x3354, 'M', u'ルーブル'),
    (0x3355, 'M', u'レム'),
    (0x3356, 'M', u'レントゲン'),
    (0x3357, 'M', u'ワット'),
    (0x3358, 'M', u'0点'),
    (0x3359, 'M', u'1点'),
    (0x335A, 'M', u'2点'),
    (0x335B, 'M', u'3点'),
    (0x335C, 'M', u'4点'),
    (0x335D, 'M', u'5点'),
    (0x335E, 'M', u'6点'),
    (0x335F, 'M', u'7点'),
    (0x3360, 'M', u'8点'),
    (0x3361, 'M', u'9点'),
    (0x3362, 'M', u'10点'),
    (0x3363, 'M', u'11点'),
    (0x3364, 'M', u'12点'),
    (0x3365, 'M', u'13点'),
    (0x3366, 'M', u'14点'),
    (0x3367, 'M', u'15点'),
    (0x3368, 'M', u'16点'),
    (0x3369, 'M', u'17点'),
    (0x336A, 'M', u'18点'),
    (0x336B, 'M', u'19点'),
    (0x336C, 'M', u'20点'),
    (0x336D, 'M', u'21点'),
    (0x336E, 'M', u'22点'),
    (0x336F, 'M', u'23点'),
    (0x3370, 'M', u'24点'),
    (0x3371, 'M', u'hpa'),
    (0x3372, 'M', u'da'),
    (0x3373, 'M', u'au'),
    (0x3374, 'M', u'bar'),
    (0x3375, 'M', u'ov'),
    (0x3376, 'M', u'pc'),
    (0x3377, 'M', u'dm'),
    (0x3378, 'M', u'dm2'),
    (0x3379, 'M', u'dm3'),
    (0x337A, 'M', u'iu'),
    (0x337B, 'M', u'平成'),
    (0x337C, 'M', u'昭和'),
    (0x337D, 'M', u'大正'),
    (0x337E, 'M', u'明治'),
    (0x337F, 'M', u'株式会社'),
    (0x3380, 'M', u'pa'),
    (0x3381, 'M', u'na'),
    (0x3382, 'M', u'μa'),
    (0x3383, 'M', u'ma'),
    (0x3384, 'M', u'ka'),
    (0x3385, 'M', u'kb'),
    (0x3386, 'M', u'mb'),
    (0x3387, 'M', u'gb'),
    (0x3388, 'M', u'cal'),
    (0x3389, 'M', u'kcal'),
    (0x338A, 'M', u'pf'),
    (0x338B, 'M', u'nf'),
    (0x338C, 'M', u'μf'),
    (0x338D, 'M', u'μg'),
    (0x338E, 'M', u'mg'),
    (0x338F, 'M', u'kg'),
    ]

def _seg_34():
    return [
    (0x3390, 'M', u'hz'),
    (0x3391, 'M', u'khz'),
    (0x3392, 'M', u'mhz'),
    (0x3393, 'M', u'ghz'),
    (0x3394, 'M', u'thz'),
    (0x3395, 'M', u'μl'),
    (0x3396, 'M', u'ml'),
    (0x3397, 'M', u'dl'),
    (0x3398, 'M', u'kl'),
    (0x3399, 'M', u'fm'),
    (0x339A, 'M', u'nm'),
    (0x339B, 'M', u'μm'),
    (0x339C, 'M', u'mm'),
    (0x339D, 'M', u'cm'),
    (0x339E, 'M', u'km'),
    (0x339F, 'M', u'mm2'),
    (0x33A0, 'M', u'cm2'),
    (0x33A1, 'M', u'm2'),
    (0x33A2, 'M', u'km2'),
    (0x33A3, 'M', u'mm3'),
    (0x33A4, 'M', u'cm3'),
    (0x33A5, 'M', u'm3'),
    (0x33A6, 'M', u'km3'),
    (0x33A7, 'M', u'm∕s'),
    (0x33A8, 'M', u'm∕s2'),
    (0x33A9, 'M', u'pa'),
    (0x33AA, 'M', u'kpa'),
    (0x33AB, 'M', u'mpa'),
    (0x33AC, 'M', u'gpa'),
    (0x33AD, 'M', u'rad'),
    (0x33AE, 'M', u'rad∕s'),
    (0x33AF, 'M', u'rad∕s2'),
    (0x33B0, 'M', u'ps'),
    (0x33B1, 'M', u'ns'),
    (0x33B2, 'M', u'μs'),
    (0x33B3, 'M', u'ms'),
    (0x33B4, 'M', u'pv'),
    (0x33B5, 'M', u'nv'),
    (0x33B6, 'M', u'μv'),
    (0x33B7, 'M', u'mv'),
    (0x33B8, 'M', u'kv'),
    (0x33B9, 'M', u'mv'),
    (0x33BA, 'M', u'pw'),
    (0x33BB, 'M', u'nw'),
    (0x33BC, 'M', u'μw'),
    (0x33BD, 'M', u'mw'),
    (0x33BE, 'M', u'kw'),
    (0x33BF, 'M', u'mw'),
    (0x33C0, 'M', u'kω'),
    (0x33C1, 'M', u'mω'),
    (0x33C2, 'X'),
    (0x33C3, 'M', u'bq'),
    (0x33C4, 'M', u'cc'),
    (0x33C5, 'M', u'cd'),
    (0x33C6, 'M', u'c∕kg'),
    (0x33C7, 'X'),
    (0x33C8, 'M', u'db'),
    (0x33C9, 'M', u'gy'),
    (0x33CA, 'M', u'ha'),
    (0x33CB, 'M', u'hp'),
    (0x33CC, 'M', u'in'),
    (0x33CD, 'M', u'kk'),
    (0x33CE, 'M', u'km'),
    (0x33CF, 'M', u'kt'),
    (0x33D0, 'M', u'lm'),
    (0x33D1, 'M', u'ln'),
    (0x33D2, 'M', u'log'),
    (0x33D3, 'M', u'lx'),
    (0x33D4, 'M', u'mb'),
    (0x33D5, 'M', u'mil'),
    (0x33D6, 'M', u'mol'),
    (0x33D7, 'M', u'ph'),
    (0x33D8, 'X'),
    (0x33D9, 'M', u'ppm'),
    (0x33DA, 'M', u'pr'),
    (0x33DB, 'M', u'sr'),
    (0x33DC, 'M', u'sv'),
    (0x33DD, 'M', u'wb'),
    (0x33DE, 'M', u'v∕m'),
    (0x33DF, 'M', u'a∕m'),
    (0x33E0, 'M', u'1日'),
    (0x33E1, 'M', u'2日'),
    (0x33E2, 'M', u'3日'),
    (0x33E3, 'M', u'4日'),
    (0x33E4, 'M', u'5日'),
    (0x33E5, 'M', u'6日'),
    (0x33E6, 'M', u'7日'),
    (0x33E7, 'M', u'8日'),
    (0x33E8, 'M', u'9日'),
    (0x33E9, 'M', u'10日'),
    (0x33EA, 'M', u'11日'),
    (0x33EB, 'M', u'12日'),
    (0x33EC, 'M', u'13日'),
    (0x33ED, 'M', u'14日'),
    (0x33EE, 'M', u'15日'),
    (0x33EF, 'M', u'16日'),
    (0x33F0, 'M', u'17日'),
    (0x33F1, 'M', u'18日'),
    (0x33F2, 'M', u'19日'),
    (0x33F3, 'M', u'20日'),
    ]

def _seg_35():
    return [
    (0x33F4, 'M', u'21日'),
    (0x33F5, 'M', u'22日'),
    (0x33F6, 'M', u'23日'),
    (0x33F7, 'M', u'24日'),
    (0x33F8, 'M', u'25日'),
    (0x33F9, 'M', u'26日'),
    (0x33FA, 'M', u'27日'),
    (0x33FB, 'M', u'28日'),
    (0x33FC, 'M', u'29日'),
    (0x33FD, 'M', u'30日'),
    (0x33FE, 'M', u'31日'),
    (0x33FF, 'M', u'gal'),
    (0x3400, 'V'),
    (0x4DB6, 'X'),
    (0x4DC0, 'V'),
    (0x9FF0, 'X'),
    (0xA000, 'V'),
    (0xA48D, 'X'),
    (0xA490, 'V'),
    (0xA4C7, 'X'),
    (0xA4D0, 'V'),
    (0xA62C, 'X'),
    (0xA640, 'M', u'ꙁ'),
    (0xA641, 'V'),
    (0xA642, 'M', u'ꙃ'),
    (0xA643, 'V'),
    (0xA644, 'M', u'ꙅ'),
    (0xA645, 'V'),
    (0xA646, 'M', u'ꙇ'),
    (0xA647, 'V'),
    (0xA648, 'M', u'ꙉ'),
    (0xA649, 'V'),
    (0xA64A, 'M', u'ꙋ'),
    (0xA64B, 'V'),
    (0xA64C, 'M', u'ꙍ'),
    (0xA64D, 'V'),
    (0xA64E, 'M', u'ꙏ'),
    (0xA64F, 'V'),
    (0xA650, 'M', u'ꙑ'),
    (0xA651, 'V'),
    (0xA652, 'M', u'ꙓ'),
    (0xA653, 'V'),
    (0xA654, 'M', u'ꙕ'),
    (0xA655, 'V'),
    (0xA656, 'M', u'ꙗ'),
    (0xA657, 'V'),
    (0xA658, 'M', u'ꙙ'),
    (0xA659, 'V'),
    (0xA65A, 'M', u'ꙛ'),
    (0xA65B, 'V'),
    (0xA65C, 'M', u'ꙝ'),
    (0xA65D, 'V'),
    (0xA65E, 'M', u'ꙟ'),
    (0xA65F, 'V'),
    (0xA660, 'M', u'ꙡ'),
    (0xA661, 'V'),
    (0xA662, 'M', u'ꙣ'),
    (0xA663, 'V'),
    (0xA664, 'M', u'ꙥ'),
    (0xA665, 'V'),
    (0xA666, 'M', u'ꙧ'),
    (0xA667, 'V'),
    (0xA668, 'M', u'ꙩ'),
    (0xA669, 'V'),
    (0xA66A, 'M', u'ꙫ'),
    (0xA66B, 'V'),
    (0xA66C, 'M', u'ꙭ'),
    (0xA66D, 'V'),
    (0xA680, 'M', u'ꚁ'),
    (0xA681, 'V'),
    (0xA682, 'M', u'ꚃ'),
    (0xA683, 'V'),
    (0xA684, 'M', u'ꚅ'),
    (0xA685, 'V'),
    (0xA686, 'M', u'ꚇ'),
    (0xA687, 'V'),
    (0xA688, 'M', u'ꚉ'),
    (0xA689, 'V'),
    (0xA68A, 'M', u'ꚋ'),
    (0xA68B, 'V'),
    (0xA68C, 'M', u'ꚍ'),
    (0xA68D, 'V'),
    (0xA68E, 'M', u'ꚏ'),
    (0xA68F, 'V'),
    (0xA690, 'M', u'ꚑ'),
    (0xA691, 'V'),
    (0xA692, 'M', u'ꚓ'),
    (0xA693, 'V'),
    (0xA694, 'M', u'ꚕ'),
    (0xA695, 'V'),
    (0xA696, 'M', u'ꚗ'),
    (0xA697, 'V'),
    (0xA698, 'M', u'ꚙ'),
    (0xA699, 'V'),
    (0xA69A, 'M', u'ꚛ'),
    (0xA69B, 'V'),
    (0xA69C, 'M', u'ъ'),
    (0xA69D, 'M', u'ь'),
    (0xA69E, 'V'),
    (0xA6F8, 'X'),
    ]

def _seg_36():
    return [
    (0xA700, 'V'),
    (0xA722, 'M', u'ꜣ'),
    (0xA723, 'V'),
    (0xA724, 'M', u'ꜥ'),
    (0xA725, 'V'),
    (0xA726, 'M', u'ꜧ'),
    (0xA727, 'V'),
    (0xA728, 'M', u'ꜩ'),
    (0xA729, 'V'),
    (0xA72A, 'M', u'ꜫ'),
    (0xA72B, 'V'),
    (0xA72C, 'M', u'ꜭ'),
    (0xA72D, 'V'),
    (0xA72E, 'M', u'ꜯ'),
    (0xA72F, 'V'),
    (0xA732, 'M', u'ꜳ'),
    (0xA733, 'V'),
    (0xA734, 'M', u'ꜵ'),
    (0xA735, 'V'),
    (0xA736, 'M', u'ꜷ'),
    (0xA737, 'V'),
    (0xA738, 'M', u'ꜹ'),
    (0xA739, 'V'),
    (0xA73A, 'M', u'ꜻ'),
    (0xA73B, 'V'),
    (0xA73C, 'M', u'ꜽ'),
    (0xA73D, 'V'),
    (0xA73E, 'M', u'ꜿ'),
    (0xA73F, 'V'),
    (0xA740, 'M', u'ꝁ'),
    (0xA741, 'V'),
    (0xA742, 'M', u'ꝃ'),
    (0xA743, 'V'),
    (0xA744, 'M', u'ꝅ'),
    (0xA745, 'V'),
    (0xA746, 'M', u'ꝇ'),
    (0xA747, 'V'),
    (0xA748, 'M', u'ꝉ'),
    (0xA749, 'V'),
    (0xA74A, 'M', u'ꝋ'),
    (0xA74B, 'V'),
    (0xA74C, 'M', u'ꝍ'),
    (0xA74D, 'V'),
    (0xA74E, 'M', u'ꝏ'),
    (0xA74F, 'V'),
    (0xA750, 'M', u'ꝑ'),
    (0xA751, 'V'),
    (0xA752, 'M', u'ꝓ'),
    (0xA753, 'V'),
    (0xA754, 'M', u'ꝕ'),
    (0xA755, 'V'),
    (0xA756, 'M', u'ꝗ'),
    (0xA757, 'V'),
    (0xA758, 'M', u'ꝙ'),
    (0xA759, 'V'),
    (0xA75A, 'M', u'ꝛ'),
    (0xA75B, 'V'),
    (0xA75C, 'M', u'ꝝ'),
    (0xA75D, 'V'),
    (0xA75E, 'M', u'ꝟ'),
    (0xA75F, 'V'),
    (0xA760, 'M', u'ꝡ'),
    (0xA761, 'V'),
    (0xA762, 'M', u'ꝣ'),
    (0xA763, 'V'),
    (0xA764, 'M', u'ꝥ'),
    (0xA765, 'V'),
    (0xA766, 'M', u'ꝧ'),
    (0xA767, 'V'),
    (0xA768, 'M', u'ꝩ'),
    (0xA769, 'V'),
    (0xA76A, 'M', u'ꝫ'),
    (0xA76B, 'V'),
    (0xA76C, 'M', u'ꝭ'),
    (0xA76D, 'V'),
    (0xA76E, 'M', u'ꝯ'),
    (0xA76F, 'V'),
    (0xA770, 'M', u'ꝯ'),
    (0xA771, 'V'),
    (0xA779, 'M', u'ꝺ'),
    (0xA77A, 'V'),
    (0xA77B, 'M', u'ꝼ'),
    (0xA77C, 'V'),
    (0xA77D, 'M', u'ᵹ'),
    (0xA77E, 'M', u'ꝿ'),
    (0xA77F, 'V'),
    (0xA780, 'M', u'ꞁ'),
    (0xA781, 'V'),
    (0xA782, 'M', u'ꞃ'),
    (0xA783, 'V'),
    (0xA784, 'M', u'ꞅ'),
    (0xA785, 'V'),
    (0xA786, 'M', u'ꞇ'),
    (0xA787, 'V'),
    (0xA78B, 'M', u'ꞌ'),
    (0xA78C, 'V'),
    (0xA78D, 'M', u'ɥ'),
    (0xA78E, 'V'),
    (0xA790, 'M', u'ꞑ'),
    (0xA791, 'V'),
    ]

def _seg_37():
    return [
    (0xA792, 'M', u'ꞓ'),
    (0xA793, 'V'),
    (0xA796, 'M', u'ꞗ'),
    (0xA797, 'V'),
    (0xA798, 'M', u'ꞙ'),
    (0xA799, 'V'),
    (0xA79A, 'M', u'ꞛ'),
    (0xA79B, 'V'),
    (0xA79C, 'M', u'ꞝ'),
    (0xA79D, 'V'),
    (0xA79E, 'M', u'ꞟ'),
    (0xA79F, 'V'),
    (0xA7A0, 'M', u'ꞡ'),
    (0xA7A1, 'V'),
    (0xA7A2, 'M', u'ꞣ'),
    (0xA7A3, 'V'),
    (0xA7A4, 'M', u'ꞥ'),
    (0xA7A5, 'V'),
    (0xA7A6, 'M', u'ꞧ'),
    (0xA7A7, 'V'),
    (0xA7A8, 'M', u'ꞩ'),
    (0xA7A9, 'V'),
    (0xA7AA, 'M', u'ɦ'),
    (0xA7AB, 'M', u'ɜ'),
    (0xA7AC, 'M', u'ɡ'),
    (0xA7AD, 'M', u'ɬ'),
    (0xA7AE, 'M', u'ɪ'),
    (0xA7AF, 'V'),
    (0xA7B0, 'M', u'ʞ'),
    (0xA7B1, 'M', u'ʇ'),
    (0xA7B2, 'M', u'ʝ'),
    (0xA7B3, 'M', u'ꭓ'),
    (0xA7B4, 'M', u'ꞵ'),
    (0xA7B5, 'V'),
    (0xA7B6, 'M', u'ꞷ'),
    (0xA7B7, 'V'),
    (0xA7B8, 'X'),
    (0xA7B9, 'V'),
    (0xA7BA, 'X'),
    (0xA7F7, 'V'),
    (0xA7F8, 'M', u'ħ'),
    (0xA7F9, 'M', u'œ'),
    (0xA7FA, 'V'),
    (0xA82C, 'X'),
    (0xA830, 'V'),
    (0xA83A, 'X'),
    (0xA840, 'V'),
    (0xA878, 'X'),
    (0xA880, 'V'),
    (0xA8C6, 'X'),
    (0xA8CE, 'V'),
    (0xA8DA, 'X'),
    (0xA8E0, 'V'),
    (0xA954, 'X'),
    (0xA95F, 'V'),
    (0xA97D, 'X'),
    (0xA980, 'V'),
    (0xA9CE, 'X'),
    (0xA9CF, 'V'),
    (0xA9DA, 'X'),
    (0xA9DE, 'V'),
    (0xA9FF, 'X'),
    (0xAA00, 'V'),
    (0xAA37, 'X'),
    (0xAA40, 'V'),
    (0xAA4E, 'X'),
    (0xAA50, 'V'),
    (0xAA5A, 'X'),
    (0xAA5C, 'V'),
    (0xAAC3, 'X'),
    (0xAADB, 'V'),
    (0xAAF7, 'X'),
    (0xAB01, 'V'),
    (0xAB07, 'X'),
    (0xAB09, 'V'),
    (0xAB0F, 'X'),
    (0xAB11, 'V'),
    (0xAB17, 'X'),
    (0xAB20, 'V'),
    (0xAB27, 'X'),
    (0xAB28, 'V'),
    (0xAB2F, 'X'),
    (0xAB30, 'V'),
    (0xAB5C, 'M', u'ꜧ'),
    (0xAB5D, 'M', u'ꬷ'),
    (0xAB5E, 'M', u'ɫ'),
    (0xAB5F, 'M', u'ꭒ'),
    (0xAB60, 'V'),
    (0xAB66, 'X'),
    (0xAB70, 'M', u'Ꭰ'),
    (0xAB71, 'M', u'Ꭱ'),
    (0xAB72, 'M', u'Ꭲ'),
    (0xAB73, 'M', u'Ꭳ'),
    (0xAB74, 'M', u'Ꭴ'),
    (0xAB75, 'M', u'Ꭵ'),
    (0xAB76, 'M', u'Ꭶ'),
    (0xAB77, 'M', u'Ꭷ'),
    (0xAB78, 'M', u'Ꭸ'),
    (0xAB79, 'M', u'Ꭹ'),
    (0xAB7A, 'M', u'Ꭺ'),
    ]

def _seg_38():
    return [
    (0xAB7B, 'M', u'Ꭻ'),
    (0xAB7C, 'M', u'Ꭼ'),
    (0xAB7D, 'M', u'Ꭽ'),
    (0xAB7E, 'M', u'Ꭾ'),
    (0xAB7F, 'M', u'Ꭿ'),
    (0xAB80, 'M', u'Ꮀ'),
    (0xAB81, 'M', u'Ꮁ'),
    (0xAB82, 'M', u'Ꮂ'),
    (0xAB83, 'M', u'Ꮃ'),
    (0xAB84, 'M', u'Ꮄ'),
    (0xAB85, 'M', u'Ꮅ'),
    (0xAB86, 'M', u'Ꮆ'),
    (0xAB87, 'M', u'Ꮇ'),
    (0xAB88, 'M', u'Ꮈ'),
    (0xAB89, 'M', u'Ꮉ'),
    (0xAB8A, 'M', u'Ꮊ'),
    (0xAB8B, 'M', u'Ꮋ'),
    (0xAB8C, 'M', u'Ꮌ'),
    (0xAB8D, 'M', u'Ꮍ'),
    (0xAB8E, 'M', u'Ꮎ'),
    (0xAB8F, 'M', u'Ꮏ'),
    (0xAB90, 'M', u'Ꮐ'),
    (0xAB91, 'M', u'Ꮑ'),
    (0xAB92, 'M', u'Ꮒ'),
    (0xAB93, 'M', u'Ꮓ'),
    (0xAB94, 'M', u'Ꮔ'),
    (0xAB95, 'M', u'Ꮕ'),
    (0xAB96, 'M', u'Ꮖ'),
    (0xAB97, 'M', u'Ꮗ'),
    (0xAB98, 'M', u'Ꮘ'),
    (0xAB99, 'M', u'Ꮙ'),
    (0xAB9A, 'M', u'Ꮚ'),
    (0xAB9B, 'M', u'Ꮛ'),
    (0xAB9C, 'M', u'Ꮜ'),
    (0xAB9D, 'M', u'Ꮝ'),
    (0xAB9E, 'M', u'Ꮞ'),
    (0xAB9F, 'M', u'Ꮟ'),
    (0xABA0, 'M', u'Ꮠ'),
    (0xABA1, 'M', u'Ꮡ'),
    (0xABA2, 'M', u'Ꮢ'),
    (0xABA3, 'M', u'Ꮣ'),
    (0xABA4, 'M', u'Ꮤ'),
    (0xABA5, 'M', u'Ꮥ'),
    (0xABA6, 'M', u'Ꮦ'),
    (0xABA7, 'M', u'Ꮧ'),
    (0xABA8, 'M', u'Ꮨ'),
    (0xABA9, 'M', u'Ꮩ'),
    (0xABAA, 'M', u'Ꮪ'),
    (0xABAB, 'M', u'Ꮫ'),
    (0xABAC, 'M', u'Ꮬ'),
    (0xABAD, 'M', u'Ꮭ'),
    (0xABAE, 'M', u'Ꮮ'),
    (0xABAF, 'M', u'Ꮯ'),
    (0xABB0, 'M', u'Ꮰ'),
    (0xABB1, 'M', u'Ꮱ'),
    (0xABB2, 'M', u'Ꮲ'),
    (0xABB3, 'M', u'Ꮳ'),
    (0xABB4, 'M', u'Ꮴ'),
    (0xABB5, 'M', u'Ꮵ'),
    (0xABB6, 'M', u'Ꮶ'),
    (0xABB7, 'M', u'Ꮷ'),
    (0xABB8, 'M', u'Ꮸ'),
    (0xABB9, 'M', u'Ꮹ'),
    (0xABBA, 'M', u'Ꮺ'),
    (0xABBB, 'M', u'Ꮻ'),
    (0xABBC, 'M', u'Ꮼ'),
    (0xABBD, 'M', u'Ꮽ'),
    (0xABBE, 'M', u'Ꮾ'),
    (0xABBF, 'M', u'Ꮿ'),
    (0xABC0, 'V'),
    (0xABEE, 'X'),
    (0xABF0, 'V'),
    (0xABFA, 'X'),
    (0xAC00, 'V'),
    (0xD7A4, 'X'),
    (0xD7B0, 'V'),
    (0xD7C7, 'X'),
    (0xD7CB, 'V'),
    (0xD7FC, 'X'),
    (0xF900, 'M', u'豈'),
    (0xF901, 'M', u'更'),
    (0xF902, 'M', u'車'),
    (0xF903, 'M', u'賈'),
    (0xF904, 'M', u'滑'),
    (0xF905, 'M', u'串'),
    (0xF906, 'M', u'句'),
    (0xF907, 'M', u'龜'),
    (0xF909, 'M', u'契'),
    (0xF90A, 'M', u'金'),
    (0xF90B, 'M', u'喇'),
    (0xF90C, 'M', u'奈'),
    (0xF90D, 'M', u'懶'),
    (0xF90E, 'M', u'癩'),
    (0xF90F, 'M', u'羅'),
    (0xF910, 'M', u'蘿'),
    (0xF911, 'M', u'螺'),
    (0xF912, 'M', u'裸'),
    (0xF913, 'M', u'邏'),
    (0xF914, 'M', u'樂'),
    (0xF915, 'M', u'洛'),
    ]

def _seg_39():
    return [
    (0xF916, 'M', u'烙'),
    (0xF917, 'M', u'珞'),
    (0xF918, 'M', u'落'),
    (0xF919, 'M', u'酪'),
    (0xF91A, 'M', u'駱'),
    (0xF91B, 'M', u'亂'),
    (0xF91C, 'M', u'卵'),
    (0xF91D, 'M', u'欄'),
    (0xF91E, 'M', u'爛'),
    (0xF91F, 'M', u'蘭'),
    (0xF920, 'M', u'鸞'),
    (0xF921, 'M', u'嵐'),
    (0xF922, 'M', u'濫'),
    (0xF923, 'M', u'藍'),
    (0xF924, 'M', u'襤'),
    (0xF925, 'M', u'拉'),
    (0xF926, 'M', u'臘'),
    (0xF927, 'M', u'蠟'),
    (0xF928, 'M', u'廊'),
    (0xF929, 'M', u'朗'),
    (0xF92A, 'M', u'浪'),
    (0xF92B, 'M', u'狼'),
    (0xF92C, 'M', u'郎'),
    (0xF92D, 'M', u'來'),
    (0xF92E, 'M', u'冷'),
    (0xF92F, 'M', u'勞'),
    (0xF930, 'M', u'擄'),
    (0xF931, 'M', u'櫓'),
    (0xF932, 'M', u'爐'),
    (0xF933, 'M', u'盧'),
    (0xF934, 'M', u'老'),
    (0xF935, 'M', u'蘆'),
    (0xF936, 'M', u'虜'),
    (0xF937, 'M', u'路'),
    (0xF938, 'M', u'露'),
    (0xF939, 'M', u'魯'),
    (0xF93A, 'M', u'鷺'),
    (0xF93B, 'M', u'碌'),
    (0xF93C, 'M', u'祿'),
    (0xF93D, 'M', u'綠'),
    (0xF93E, 'M', u'菉'),
    (0xF93F, 'M', u'錄'),
    (0xF940, 'M', u'鹿'),
    (0xF941, 'M', u'論'),
    (0xF942, 'M', u'壟'),
    (0xF943, 'M', u'弄'),
    (0xF944, 'M', u'籠'),
    (0xF945, 'M', u'聾'),
    (0xF946, 'M', u'牢'),
    (0xF947, 'M', u'磊'),
    (0xF948, 'M', u'賂'),
    (0xF949, 'M', u'雷'),
    (0xF94A, 'M', u'壘'),
    (0xF94B, 'M', u'屢'),
    (0xF94C, 'M', u'樓'),
    (0xF94D, 'M', u'淚'),
    (0xF94E, 'M', u'漏'),
    (0xF94F, 'M', u'累'),
    (0xF950, 'M', u'縷'),
    (0xF951, 'M', u'陋'),
    (0xF952, 'M', u'勒'),
    (0xF953, 'M', u'肋'),
    (0xF954, 'M', u'凜'),
    (0xF955, 'M', u'凌'),
    (0xF956, 'M', u'稜'),
    (0xF957, 'M', u'綾'),
    (0xF958, 'M', u'菱'),
    (0xF959, 'M', u'陵'),
    (0xF95A, 'M', u'讀'),
    (0xF95B, 'M', u'拏'),
    (0xF95C, 'M', u'樂'),
    (0xF95D, 'M', u'諾'),
    (0xF95E, 'M', u'丹'),
    (0xF95F, 'M', u'寧'),
    (0xF960, 'M', u'怒'),
    (0xF961, 'M', u'率'),
    (0xF962, 'M', u'異'),
    (0xF963, 'M', u'北'),
    (0xF964, 'M', u'磻'),
    (0xF965, 'M', u'便'),
    (0xF966, 'M', u'復'),
    (0xF967, 'M', u'不'),
    (0xF968, 'M', u'泌'),
    (0xF969, 'M', u'數'),
    (0xF96A, 'M', u'索'),
    (0xF96B, 'M', u'參'),
    (0xF96C, 'M', u'塞'),
    (0xF96D, 'M', u'省'),
    (0xF96E, 'M', u'葉'),
    (0xF96F, 'M', u'說'),
    (0xF970, 'M', u'殺'),
    (0xF971, 'M', u'辰'),
    (0xF972, 'M', u'沈'),
    (0xF973, 'M', u'拾'),
    (0xF974, 'M', u'若'),
    (0xF975, 'M', u'掠'),
    (0xF976, 'M', u'略'),
    (0xF977, 'M', u'亮'),
    (0xF978, 'M', u'兩'),
    (0xF979, 'M', u'凉'),
    ]

def _seg_40():
    return [
    (0xF97A, 'M', u'梁'),
    (0xF97B, 'M', u'糧'),
    (0xF97C, 'M', u'良'),
    (0xF97D, 'M', u'諒'),
    (0xF97E, 'M', u'量'),
    (0xF97F, 'M', u'勵'),
    (0xF980, 'M', u'呂'),
    (0xF981, 'M', u'女'),
    (0xF982, 'M', u'廬'),
    (0xF983, 'M', u'旅'),
    (0xF984, 'M', u'濾'),
    (0xF985, 'M', u'礪'),
    (0xF986, 'M', u'閭'),
    (0xF987, 'M', u'驪'),
    (0xF988, 'M', u'麗'),
    (0xF989, 'M', u'黎'),
    (0xF98A, 'M', u'力'),
    (0xF98B, 'M', u'曆'),
    (0xF98C, 'M', u'歷'),
    (0xF98D, 'M', u'轢'),
    (0xF98E, 'M', u'年'),
    (0xF98F, 'M', u'憐'),
    (0xF990, 'M', u'戀'),
    (0xF991, 'M', u'撚'),
    (0xF992, 'M', u'漣'),
    (0xF993, 'M', u'煉'),
    (0xF994, 'M', u'璉'),
    (0xF995, 'M', u'秊'),
    (0xF996, 'M', u'練'),
    (0xF997, 'M', u'聯'),
    (0xF998, 'M', u'輦'),
    (0xF999, 'M', u'蓮'),
    (0xF99A, 'M', u'連'),
    (0xF99B, 'M', u'鍊'),
    (0xF99C, 'M', u'列'),
    (0xF99D, 'M', u'劣'),
    (0xF99E, 'M', u'咽'),
    (0xF99F, 'M', u'烈'),
    (0xF9A0, 'M', u'裂'),
    (0xF9A1, 'M', u'說'),
    (0xF9A2, 'M', u'廉'),
    (0xF9A3, 'M', u'念'),
    (0xF9A4, 'M', u'捻'),
    (0xF9A5, 'M', u'殮'),
    (0xF9A6, 'M', u'簾'),
    (0xF9A7, 'M', u'獵'),
    (0xF9A8, 'M', u'令'),
    (0xF9A9, 'M', u'囹'),
    (0xF9AA, 'M', u'寧'),
    (0xF9AB, 'M', u'嶺'),
    (0xF9AC, 'M', u'怜'),
    (0xF9AD, 'M', u'玲'),
    (0xF9AE, 'M', u'瑩'),
    (0xF9AF, 'M', u'羚'),
    (0xF9B0, 'M', u'聆'),
    (0xF9B1, 'M', u'鈴'),
    (0xF9B2, 'M', u'零'),
    (0xF9B3, 'M', u'靈'),
    (0xF9B4, 'M', u'領'),
    (0xF9B5, 'M', u'例'),
    (0xF9B6, 'M', u'禮'),
    (0xF9B7, 'M', u'醴'),
    (0xF9B8, 'M', u'隸'),
    (0xF9B9, 'M', u'惡'),
    (0xF9BA, 'M', u'了'),
    (0xF9BB, 'M', u'僚'),
    (0xF9BC, 'M', u'寮'),
    (0xF9BD, 'M', u'尿'),
    (0xF9BE, 'M', u'料'),
    (0xF9BF, 'M', u'樂'),
    (0xF9C0, 'M', u'燎'),
    (0xF9C1, 'M', u'療'),
    (0xF9C2, 'M', u'蓼'),
    (0xF9C3, 'M', u'遼'),
    (0xF9C4, 'M', u'龍'),
    (0xF9C5, 'M', u'暈'),
    (0xF9C6, 'M', u'阮'),
    (0xF9C7, 'M', u'劉'),
    (0xF9C8, 'M', u'杻'),
    (0xF9C9, 'M', u'柳'),
    (0xF9CA, 'M', u'流'),
    (0xF9CB, 'M', u'溜'),
    (0xF9CC, 'M', u'琉'),
    (0xF9CD, 'M', u'留'),
    (0xF9CE, 'M', u'硫'),
    (0xF9CF, 'M', u'紐'),
    (0xF9D0, 'M', u'類'),
    (0xF9D1, 'M', u'六'),
    (0xF9D2, 'M', u'戮'),
    (0xF9D3, 'M', u'陸'),
    (0xF9D4, 'M', u'倫'),
    (0xF9D5, 'M', u'崙'),
    (0xF9D6, 'M', u'淪'),
    (0xF9D7, 'M', u'輪'),
    (0xF9D8, 'M', u'律'),
    (0xF9D9, 'M', u'慄'),
    (0xF9DA, 'M', u'栗'),
    (0xF9DB, 'M', u'率'),
    (0xF9DC, 'M', u'隆'),
    (0xF9DD, 'M', u'利'),
    ]

def _seg_41():
    return [
    (0xF9DE, 'M', u'吏'),
    (0xF9DF, 'M', u'履'),
    (0xF9E0, 'M', u'易'),
    (0xF9E1, 'M', u'李'),
    (0xF9E2, 'M', u'梨'),
    (0xF9E3, 'M', u'泥'),
    (0xF9E4, 'M', u'理'),
    (0xF9E5, 'M', u'痢'),
    (0xF9E6, 'M', u'罹'),
    (0xF9E7, 'M', u'裏'),
    (0xF9E8, 'M', u'裡'),
    (0xF9E9, 'M', u'里'),
    (0xF9EA, 'M', u'離'),
    (0xF9EB, 'M', u'匿'),
    (0xF9EC, 'M', u'溺'),
    (0xF9ED, 'M', u'吝'),
    (0xF9EE, 'M', u'燐'),
    (0xF9EF, 'M', u'璘'),
    (0xF9F0, 'M', u'藺'),
    (0xF9F1, 'M', u'隣'),
    (0xF9F2, 'M', u'鱗'),
    (0xF9F3, 'M', u'麟'),
    (0xF9F4, 'M', u'林'),
    (0xF9F5, 'M', u'淋'),
    (0xF9F6, 'M', u'臨'),
    (0xF9F7, 'M', u'立'),
    (0xF9F8, 'M', u'笠'),
    (0xF9F9, 'M', u'粒'),
    (0xF9FA, 'M', u'狀'),
    (0xF9FB, 'M', u'炙'),
    (0xF9FC, 'M', u'識'),
    (0xF9FD, 'M', u'什'),
    (0xF9FE, 'M', u'茶'),
    (0xF9FF, 'M', u'刺'),
    (0xFA00, 'M', u'切'),
    (0xFA01, 'M', u'度'),
    (0xFA02, 'M', u'拓'),
    (0xFA03, 'M', u'糖'),
    (0xFA04, 'M', u'宅'),
    (0xFA05, 'M', u'洞'),
    (0xFA06, 'M', u'暴'),
    (0xFA07, 'M', u'輻'),
    (0xFA08, 'M', u'行'),
    (0xFA09, 'M', u'降'),
    (0xFA0A, 'M', u'見'),
    (0xFA0B, 'M', u'廓'),
    (0xFA0C, 'M', u'兀'),
    (0xFA0D, 'M', u'嗀'),
    (0xFA0E, 'V'),
    (0xFA10, 'M', u'塚'),
    (0xFA11, 'V'),
    (0xFA12, 'M', u'晴'),
    (0xFA13, 'V'),
    (0xFA15, 'M', u'凞'),
    (0xFA16, 'M', u'猪'),
    (0xFA17, 'M', u'益'),
    (0xFA18, 'M', u'礼'),
    (0xFA19, 'M', u'神'),
    (0xFA1A, 'M', u'祥'),
    (0xFA1B, 'M', u'福'),
    (0xFA1C, 'M', u'靖'),
    (0xFA1D, 'M', u'精'),
    (0xFA1E, 'M', u'羽'),
    (0xFA1F, 'V'),
    (0xFA20, 'M', u'蘒'),
    (0xFA21, 'V'),
    (0xFA22, 'M', u'諸'),
    (0xFA23, 'V'),
    (0xFA25, 'M', u'逸'),
    (0xFA26, 'M', u'都'),
    (0xFA27, 'V'),
    (0xFA2A, 'M', u'飯'),
    (0xFA2B, 'M', u'飼'),
    (0xFA2C, 'M', u'館'),
    (0xFA2D, 'M', u'鶴'),
    (0xFA2E, 'M', u'郞'),
    (0xFA2F, 'M', u'隷'),
    (0xFA30, 'M', u'侮'),
    (0xFA31, 'M', u'僧'),
    (0xFA32, 'M', u'免'),
    (0xFA33, 'M', u'勉'),
    (0xFA34, 'M', u'勤'),
    (0xFA35, 'M', u'卑'),
    (0xFA36, 'M', u'喝'),
    (0xFA37, 'M', u'嘆'),
    (0xFA38, 'M', u'器'),
    (0xFA39, 'M', u'塀'),
    (0xFA3A, 'M', u'墨'),
    (0xFA3B, 'M', u'層'),
    (0xFA3C, 'M', u'屮'),
    (0xFA3D, 'M', u'悔'),
    (0xFA3E, 'M', u'慨'),
    (0xFA3F, 'M', u'憎'),
    (0xFA40, 'M', u'懲'),
    (0xFA41, 'M', u'敏'),
    (0xFA42, 'M', u'既'),
    (0xFA43, 'M', u'暑'),
    (0xFA44, 'M', u'梅'),
    (0xFA45, 'M', u'海'),
    (0xFA46, 'M', u'渚'),
    ]

def _seg_42():
    return [
    (0xFA47, 'M', u'漢'),
    (0xFA48, 'M', u'煮'),
    (0xFA49, 'M', u'爫'),
    (0xFA4A, 'M', u'琢'),
    (0xFA4B, 'M', u'碑'),
    (0xFA4C, 'M', u'社'),
    (0xFA4D, 'M', u'祉'),
    (0xFA4E, 'M', u'祈'),
    (0xFA4F, 'M', u'祐'),
    (0xFA50, 'M', u'祖'),
    (0xFA51, 'M', u'祝'),
    (0xFA52, 'M', u'禍'),
    (0xFA53, 'M', u'禎'),
    (0xFA54, 'M', u'穀'),
    (0xFA55, 'M', u'突'),
    (0xFA56, 'M', u'節'),
    (0xFA57, 'M', u'練'),
    (0xFA58, 'M', u'縉'),
    (0xFA59, 'M', u'繁'),
    (0xFA5A, 'M', u'署'),
    (0xFA5B, 'M', u'者'),
    (0xFA5C, 'M', u'臭'),
    (0xFA5D, 'M', u'艹'),
    (0xFA5F, 'M', u'著'),
    (0xFA60, 'M', u'褐'),
    (0xFA61, 'M', u'視'),
    (0xFA62, 'M', u'謁'),
    (0xFA63, 'M', u'謹'),
    (0xFA64, 'M', u'賓'),
    (0xFA65, 'M', u'贈'),
    (0xFA66, 'M', u'辶'),
    (0xFA67, 'M', u'逸'),
    (0xFA68, 'M', u'難'),
    (0xFA69, 'M', u'響'),
    (0xFA6A, 'M', u'頻'),
    (0xFA6B, 'M', u'恵'),
    (0xFA6C, 'M', u'𤋮'),
    (0xFA6D, 'M', u'舘'),
    (0xFA6E, 'X'),
    (0xFA70, 'M', u'並'),
    (0xFA71, 'M', u'况'),
    (0xFA72, 'M', u'全'),
    (0xFA73, 'M', u'侀'),
    (0xFA74, 'M', u'充'),
    (0xFA75, 'M', u'冀'),
    (0xFA76, 'M', u'勇'),
    (0xFA77, 'M', u'勺'),
    (0xFA78, 'M', u'喝'),
    (0xFA79, 'M', u'啕'),
    (0xFA7A, 'M', u'喙'),
    (0xFA7B, 'M', u'嗢'),
    (0xFA7C, 'M', u'塚'),
    (0xFA7D, 'M', u'墳'),
    (0xFA7E, 'M', u'奄'),
    (0xFA7F, 'M', u'奔'),
    (0xFA80, 'M', u'婢'),
    (0xFA81, 'M', u'嬨'),
    (0xFA82, 'M', u'廒'),
    (0xFA83, 'M', u'廙'),
    (0xFA84, 'M', u'彩'),
    (0xFA85, 'M', u'徭'),
    (0xFA86, 'M', u'惘'),
    (0xFA87, 'M', u'慎'),
    (0xFA88, 'M', u'愈'),
    (0xFA89, 'M', u'憎'),
    (0xFA8A, 'M', u'慠'),
    (0xFA8B, 'M', u'懲'),
    (0xFA8C, 'M', u'戴'),
    (0xFA8D, 'M', u'揄'),
    (0xFA8E, 'M', u'搜'),
    (0xFA8F, 'M', u'摒'),
    (0xFA90, 'M', u'敖'),
    (0xFA91, 'M', u'晴'),
    (0xFA92, 'M', u'朗'),
    (0xFA93, 'M', u'望'),
    (0xFA94, 'M', u'杖'),
    (0xFA95, 'M', u'歹'),
    (0xFA96, 'M', u'殺'),
    (0xFA97, 'M', u'流'),
    (0xFA98, 'M', u'滛'),
    (0xFA99, 'M', u'滋'),
    (0xFA9A, 'M', u'漢'),
    (0xFA9B, 'M', u'瀞'),
    (0xFA9C, 'M', u'煮'),
    (0xFA9D, 'M', u'瞧'),
    (0xFA9E, 'M', u'爵'),
    (0xFA9F, 'M', u'犯'),
    (0xFAA0, 'M', u'猪'),
    (0xFAA1, 'M', u'瑱'),
    (0xFAA2, 'M', u'甆'),
    (0xFAA3, 'M', u'画'),
    (0xFAA4, 'M', u'瘝'),
    (0xFAA5, 'M', u'瘟'),
    (0xFAA6, 'M', u'益'),
    (0xFAA7, 'M', u'盛'),
    (0xFAA8, 'M', u'直'),
    (0xFAA9, 'M', u'睊'),
    (0xFAAA, 'M', u'着'),
    (0xFAAB, 'M', u'磌'),
    (0xFAAC, 'M', u'窱'),
    ]

def _seg_43():
    return [
    (0xFAAD, 'M', u'節'),
    (0xFAAE, 'M', u'类'),
    (0xFAAF, 'M', u'絛'),
    (0xFAB0, 'M', u'練'),
    (0xFAB1, 'M', u'缾'),
    (0xFAB2, 'M', u'者'),
    (0xFAB3, 'M', u'荒'),
    (0xFAB4, 'M', u'華'),
    (0xFAB5, 'M', u'蝹'),
    (0xFAB6, 'M', u'襁'),
    (0xFAB7, 'M', u'覆'),
    (0xFAB8, 'M', u'視'),
    (0xFAB9, 'M', u'調'),
    (0xFABA, 'M', u'諸'),
    (0xFABB, 'M', u'請'),
    (0xFABC, 'M', u'謁'),
    (0xFABD, 'M', u'諾'),
    (0xFABE, 'M', u'諭'),
    (0xFABF, 'M', u'謹'),
    (0xFAC0, 'M', u'變'),
    (0xFAC1, 'M', u'贈'),
    (0xFAC2, 'M', u'輸'),
    (0xFAC3, 'M', u'遲'),
    (0xFAC4, 'M', u'醙'),
    (0xFAC5, 'M', u'鉶'),
    (0xFAC6, 'M', u'陼'),
    (0xFAC7, 'M', u'難'),
    (0xFAC8, 'M', u'靖'),
    (0xFAC9, 'M', u'韛'),
    (0xFACA, 'M', u'響'),
    (0xFACB, 'M', u'頋'),
    (0xFACC, 'M', u'頻'),
    (0xFACD, 'M', u'鬒'),
    (0xFACE, 'M', u'龜'),
    (0xFACF, 'M', u'𢡊'),
    (0xFAD0, 'M', u'𢡄'),
    (0xFAD1, 'M', u'𣏕'),
    (0xFAD2, 'M', u'㮝'),
    (0xFAD3, 'M', u'䀘'),
    (0xFAD4, 'M', u'䀹'),
    (0xFAD5, 'M', u'𥉉'),
    (0xFAD6, 'M', u'𥳐'),
    (0xFAD7, 'M', u'𧻓'),
    (0xFAD8, 'M', u'齃'),
    (0xFAD9, 'M', u'龎'),
    (0xFADA, 'X'),
    (0xFB00, 'M', u'ff'),
    (0xFB01, 'M', u'fi'),
    (0xFB02, 'M', u'fl'),
    (0xFB03, 'M', u'ffi'),
    (0xFB04, 'M', u'ffl'),
    (0xFB05, 'M', u'st'),
    (0xFB07, 'X'),
    (0xFB13, 'M', u'մն'),
    (0xFB14, 'M', u'մե'),
    (0xFB15, 'M', u'մի'),
    (0xFB16, 'M', u'վն'),
    (0xFB17, 'M', u'մխ'),
    (0xFB18, 'X'),
    (0xFB1D, 'M', u'יִ'),
    (0xFB1E, 'V'),
    (0xFB1F, 'M', u'ײַ'),
    (0xFB20, 'M', u'ע'),
    (0xFB21, 'M', u'א'),
    (0xFB22, 'M', u'ד'),
    (0xFB23, 'M', u'ה'),
    (0xFB24, 'M', u'כ'),
    (0xFB25, 'M', u'ל'),
    (0xFB26, 'M', u'ם'),
    (0xFB27, 'M', u'ר'),
    (0xFB28, 'M', u'ת'),
    (0xFB29, '3', u'+'),
    (0xFB2A, 'M', u'שׁ'),
    (0xFB2B, 'M', u'שׂ'),
    (0xFB2C, 'M', u'שּׁ'),
    (0xFB2D, 'M', u'שּׂ'),
    (0xFB2E, 'M', u'אַ'),
    (0xFB2F, 'M', u'אָ'),
    (0xFB30, 'M', u'אּ'),
    (0xFB31, 'M', u'בּ'),
    (0xFB32, 'M', u'גּ'),
    (0xFB33, 'M', u'דּ'),
    (0xFB34, 'M', u'הּ'),
    (0xFB35, 'M', u'וּ'),
    (0xFB36, 'M', u'זּ'),
    (0xFB37, 'X'),
    (0xFB38, 'M', u'טּ'),
    (0xFB39, 'M', u'יּ'),
    (0xFB3A, 'M', u'ךּ'),
    (0xFB3B, 'M', u'כּ'),
    (0xFB3C, 'M', u'לּ'),
    (0xFB3D, 'X'),
    (0xFB3E, 'M', u'מּ'),
    (0xFB3F, 'X'),
    (0xFB40, 'M', u'נּ'),
    (0xFB41, 'M', u'סּ'),
    (0xFB42, 'X'),
    (0xFB43, 'M', u'ףּ'),
    (0xFB44, 'M', u'פּ'),
    (0xFB45, 'X'),
    ]

def _seg_44():
    return [
    (0xFB46, 'M', u'צּ'),
    (0xFB47, 'M', u'קּ'),
    (0xFB48, 'M', u'רּ'),
    (0xFB49, 'M', u'שּ'),
    (0xFB4A, 'M', u'תּ'),
    (0xFB4B, 'M', u'וֹ'),
    (0xFB4C, 'M', u'בֿ'),
    (0xFB4D, 'M', u'כֿ'),
    (0xFB4E, 'M', u'פֿ'),
    (0xFB4F, 'M', u'אל'),
    (0xFB50, 'M', u'ٱ'),
    (0xFB52, 'M', u'ٻ'),
    (0xFB56, 'M', u'پ'),
    (0xFB5A, 'M', u'ڀ'),
    (0xFB5E, 'M', u'ٺ'),
    (0xFB62, 'M', u'ٿ'),
    (0xFB66, 'M', u'ٹ'),
    (0xFB6A, 'M', u'ڤ'),
    (0xFB6E, 'M', u'ڦ'),
    (0xFB72, 'M', u'ڄ'),
    (0xFB76, 'M', u'ڃ'),
    (0xFB7A, 'M', u'چ'),
    (0xFB7E, 'M', u'ڇ'),
    (0xFB82, 'M', u'ڍ'),
    (0xFB84, 'M', u'ڌ'),
    (0xFB86, 'M', u'ڎ'),
    (0xFB88, 'M', u'ڈ'),
    (0xFB8A, 'M', u'ژ'),
    (0xFB8C, 'M', u'ڑ'),
    (0xFB8E, 'M', u'ک'),
    (0xFB92, 'M', u'گ'),
    (0xFB96, 'M', u'ڳ'),
    (0xFB9A, 'M', u'ڱ'),
    (0xFB9E, 'M', u'ں'),
    (0xFBA0, 'M', u'ڻ'),
    (0xFBA4, 'M', u'ۀ'),
    (0xFBA6, 'M', u'ہ'),
    (0xFBAA, 'M', u'ھ'),
    (0xFBAE, 'M', u'ے'),
    (0xFBB0, 'M', u'ۓ'),
    (0xFBB2, 'V'),
    (0xFBC2, 'X'),
    (0xFBD3, 'M', u'ڭ'),
    (0xFBD7, 'M', u'ۇ'),
    (0xFBD9, 'M', u'ۆ'),
    (0xFBDB, 'M', u'ۈ'),
    (0xFBDD, 'M', u'ۇٴ'),
    (0xFBDE, 'M', u'ۋ'),
    (0xFBE0, 'M', u'ۅ'),
    (0xFBE2, 'M', u'ۉ'),
    (0xFBE4, 'M', u'ې'),
    (0xFBE8, 'M', u'ى'),
    (0xFBEA, 'M', u'ئا'),
    (0xFBEC, 'M', u'ئە'),
    (0xFBEE, 'M', u'ئو'),
    (0xFBF0, 'M', u'ئۇ'),
    (0xFBF2, 'M', u'ئۆ'),
    (0xFBF4, 'M', u'ئۈ'),
    (0xFBF6, 'M', u'ئې'),
    (0xFBF9, 'M', u'ئى'),
    (0xFBFC, 'M', u'ی'),
    (0xFC00, 'M', u'ئج'),
    (0xFC01, 'M', u'ئح'),
    (0xFC02, 'M', u'ئم'),
    (0xFC03, 'M', u'ئى'),
    (0xFC04, 'M', u'ئي'),
    (0xFC05, 'M', u'بج'),
    (0xFC06, 'M', u'بح'),
    (0xFC07, 'M', u'بخ'),
    (0xFC08, 'M', u'بم'),
    (0xFC09, 'M', u'بى'),
    (0xFC0A, 'M', u'بي'),
    (0xFC0B, 'M', u'تج'),
    (0xFC0C, 'M', u'تح'),
    (0xFC0D, 'M', u'تخ'),
    (0xFC0E, 'M', u'تم'),
    (0xFC0F, 'M', u'تى'),
    (0xFC10, 'M', u'تي'),
    (0xFC11, 'M', u'ثج'),
    (0xFC12, 'M', u'ثم'),
    (0xFC13, 'M', u'ثى'),
    (0xFC14, 'M', u'ثي'),
    (0xFC15, 'M', u'جح'),
    (0xFC16, 'M', u'جم'),
    (0xFC17, 'M', u'حج'),
    (0xFC18, 'M', u'حم'),
    (0xFC19, 'M', u'خج'),
    (0xFC1A, 'M', u'خح'),
    (0xFC1B, 'M', u'خم'),
    (0xFC1C, 'M', u'سج'),
    (0xFC1D, 'M', u'سح'),
    (0xFC1E, 'M', u'سخ'),
    (0xFC1F, 'M', u'سم'),
    (0xFC20, 'M', u'صح'),
    (0xFC21, 'M', u'صم'),
    (0xFC22, 'M', u'ضج'),
    (0xFC23, 'M', u'ضح'),
    (0xFC24, 'M', u'ضخ'),
    (0xFC25, 'M', u'ضم'),
    (0xFC26, 'M', u'طح'),
    ]

def _seg_45():
    return [
    (0xFC27, 'M', u'طم'),
    (0xFC28, 'M', u'ظم'),
    (0xFC29, 'M', u'عج'),
    (0xFC2A, 'M', u'عم'),
    (0xFC2B, 'M', u'غج'),
    (0xFC2C, 'M', u'غم'),
    (0xFC2D, 'M', u'فج'),
    (0xFC2E, 'M', u'فح'),
    (0xFC2F, 'M', u'فخ'),
    (0xFC30, 'M', u'فم'),
    (0xFC31, 'M', u'فى'),
    (0xFC32, 'M', u'في'),
    (0xFC33, 'M', u'قح'),
    (0xFC34, 'M', u'قم'),
    (0xFC35, 'M', u'قى'),
    (0xFC36, 'M', u'قي'),
    (0xFC37, 'M', u'كا'),
    (0xFC38, 'M', u'كج'),
    (0xFC39, 'M', u'كح'),
    (0xFC3A, 'M', u'كخ'),
    (0xFC3B, 'M', u'كل'),
    (0xFC3C, 'M', u'كم'),
    (0xFC3D, 'M', u'كى'),
    (0xFC3E, 'M', u'كي'),
    (0xFC3F, 'M', u'لج'),
    (0xFC40, 'M', u'لح'),
    (0xFC41, 'M', u'لخ'),
    (0xFC42, 'M', u'لم'),
    (0xFC43, 'M', u'لى'),
    (0xFC44, 'M', u'لي'),
    (0xFC45, 'M', u'مج'),
    (0xFC46, 'M', u'مح'),
    (0xFC47, 'M', u'مخ'),
    (0xFC48, 'M', u'مم'),
    (0xFC49, 'M', u'مى'),
    (0xFC4A, 'M', u'مي'),
    (0xFC4B, 'M', u'نج'),
    (0xFC4C, 'M', u'نح'),
    (0xFC4D, 'M', u'نخ'),
    (0xFC4E, 'M', u'نم'),
    (0xFC4F, 'M', u'نى'),
    (0xFC50, 'M', u'ني'),
    (0xFC51, 'M', u'هج'),
    (0xFC52, 'M', u'هم'),
    (0xFC53, 'M', u'هى'),
    (0xFC54, 'M', u'هي'),
    (0xFC55, 'M', u'يج'),
    (0xFC56, 'M', u'يح'),
    (0xFC57, 'M', u'يخ'),
    (0xFC58, 'M', u'يم'),
    (0xFC59, 'M', u'يى'),
    (0xFC5A, 'M', u'يي'),
    (0xFC5B, 'M', u'ذٰ'),
    (0xFC5C, 'M', u'رٰ'),
    (0xFC5D, 'M', u'ىٰ'),
    (0xFC5E, '3', u' ٌّ'),
    (0xFC5F, '3', u' ٍّ'),
    (0xFC60, '3', u' َّ'),
    (0xFC61, '3', u' ُّ'),
    (0xFC62, '3', u' ِّ'),
    (0xFC63, '3', u' ّٰ'),
    (0xFC64, 'M', u'ئر'),
    (0xFC65, 'M', u'ئز'),
    (0xFC66, 'M', u'ئم'),
    (0xFC67, 'M', u'ئن'),
    (0xFC68, 'M', u'ئى'),
    (0xFC69, 'M', u'ئي'),
    (0xFC6A, 'M', u'بر'),
    (0xFC6B, 'M', u'بز'),
    (0xFC6C, 'M', u'بم'),
    (0xFC6D, 'M', u'بن'),
    (0xFC6E, 'M', u'بى'),
    (0xFC6F, 'M', u'بي'),
    (0xFC70, 'M', u'تر'),
    (0xFC71, 'M', u'تز'),
    (0xFC72, 'M', u'تم'),
    (0xFC73, 'M', u'تن'),
    (0xFC74, 'M', u'تى'),
    (0xFC75, 'M', u'تي'),
    (0xFC76, 'M', u'ثر'),
    (0xFC77, 'M', u'ثز'),
    (0xFC78, 'M', u'ثم'),
    (0xFC79, 'M', u'ثن'),
    (0xFC7A, 'M', u'ثى'),
    (0xFC7B, 'M', u'ثي'),
    (0xFC7C, 'M', u'فى'),
    (0xFC7D, 'M', u'في'),
    (0xFC7E, 'M', u'قى'),
    (0xFC7F, 'M', u'قي'),
    (0xFC80, 'M', u'كا'),
    (0xFC81, 'M', u'كل'),
    (0xFC82, 'M', u'كم'),
    (0xFC83, 'M', u'كى'),
    (0xFC84, 'M', u'كي'),
    (0xFC85, 'M', u'لم'),
    (0xFC86, 'M', u'لى'),
    (0xFC87, 'M', u'لي'),
    (0xFC88, 'M', u'ما'),
    (0xFC89, 'M', u'مم'),
    (0xFC8A, 'M', u'نر'),
    ]

def _seg_46():
    return [
    (0xFC8B, 'M', u'نز'),
    (0xFC8C, 'M', u'نم'),
    (0xFC8D, 'M', u'نن'),
    (0xFC8E, 'M', u'نى'),
    (0xFC8F, 'M', u'ني'),
    (0xFC90, 'M', u'ىٰ'),
    (0xFC91, 'M', u'ير'),
    (0xFC92, 'M', u'يز'),
    (0xFC93, 'M', u'يم'),
    (0xFC94, 'M', u'ين'),
    (0xFC95, 'M', u'يى'),
    (0xFC96, 'M', u'يي'),
    (0xFC97, 'M', u'ئج'),
    (0xFC98, 'M', u'ئح'),
    (0xFC99, 'M', u'ئخ'),
    (0xFC9A, 'M', u'ئم'),
    (0xFC9B, 'M', u'ئه'),
    (0xFC9C, 'M', u'بج'),
    (0xFC9D, 'M', u'بح'),
    (0xFC9E, 'M', u'بخ'),
    (0xFC9F, 'M', u'بم'),
    (0xFCA0, 'M', u'به'),
    (0xFCA1, 'M', u'تج'),
    (0xFCA2, 'M', u'تح'),
    (0xFCA3, 'M', u'تخ'),
    (0xFCA4, 'M', u'تم'),
    (0xFCA5, 'M', u'ته'),
    (0xFCA6, 'M', u'ثم'),
    (0xFCA7, 'M', u'جح'),
    (0xFCA8, 'M', u'جم'),
    (0xFCA9, 'M', u'حج'),
    (0xFCAA, 'M', u'حم'),
    (0xFCAB, 'M', u'خج'),
    (0xFCAC, 'M', u'خم'),
    (0xFCAD, 'M', u'سج'),
    (0xFCAE, 'M', u'سح'),
    (0xFCAF, 'M', u'سخ'),
    (0xFCB0, 'M', u'سم'),
    (0xFCB1, 'M', u'صح'),
    (0xFCB2, 'M', u'صخ'),
    (0xFCB3, 'M', u'صم'),
    (0xFCB4, 'M', u'ضج'),
    (0xFCB5, 'M', u'ضح'),
    (0xFCB6, 'M', u'ضخ'),
    (0xFCB7, 'M', u'ضم'),
    (0xFCB8, 'M', u'طح'),
    (0xFCB9, 'M', u'ظم'),
    (0xFCBA, 'M', u'عج'),
    (0xFCBB, 'M', u'عم'),
    (0xFCBC, 'M', u'غج'),
    (0xFCBD, 'M', u'غم'),
    (0xFCBE, 'M', u'فج'),
    (0xFCBF, 'M', u'فح'),
    (0xFCC0, 'M', u'فخ'),
    (0xFCC1, 'M', u'فم'),
    (0xFCC2, 'M', u'قح'),
    (0xFCC3, 'M', u'قم'),
    (0xFCC4, 'M', u'كج'),
    (0xFCC5, 'M', u'كح'),
    (0xFCC6, 'M', u'كخ'),
    (0xFCC7, 'M', u'كل'),
    (0xFCC8, 'M', u'كم'),
    (0xFCC9, 'M', u'لج'),
    (0xFCCA, 'M', u'لح'),
    (0xFCCB, 'M', u'لخ'),
    (0xFCCC, 'M', u'لم'),
    (0xFCCD, 'M', u'له'),
    (0xFCCE, 'M', u'مج'),
    (0xFCCF, 'M', u'مح'),
    (0xFCD0, 'M', u'مخ'),
    (0xFCD1, 'M', u'مم'),
    (0xFCD2, 'M', u'نج'),
    (0xFCD3, 'M', u'نح'),
    (0xFCD4, 'M', u'نخ'),
    (0xFCD5, 'M', u'نم'),
    (0xFCD6, 'M', u'نه'),
    (0xFCD7, 'M', u'هج'),
    (0xFCD8, 'M', u'هم'),
    (0xFCD9, 'M', u'هٰ'),
    (0xFCDA, 'M', u'يج'),
    (0xFCDB, 'M', u'يح'),
    (0xFCDC, 'M', u'يخ'),
    (0xFCDD, 'M', u'يم'),
    (0xFCDE, 'M', u'يه'),
    (0xFCDF, 'M', u'ئم'),
    (0xFCE0, 'M', u'ئه'),
    (0xFCE1, 'M', u'بم'),
    (0xFCE2, 'M', u'به'),
    (0xFCE3, 'M', u'تم'),
    (0xFCE4, 'M', u'ته'),
    (0xFCE5, 'M', u'ثم'),
    (0xFCE6, 'M', u'ثه'),
    (0xFCE7, 'M', u'سم'),
    (0xFCE8, 'M', u'سه'),
    (0xFCE9, 'M', u'شم'),
    (0xFCEA, 'M', u'شه'),
    (0xFCEB, 'M', u'كل'),
    (0xFCEC, 'M', u'كم'),
    (0xFCED, 'M', u'لم'),
    (0xFCEE, 'M', u'نم'),
    ]

def _seg_47():
    return [
    (0xFCEF, 'M', u'نه'),
    (0xFCF0, 'M', u'يم'),
    (0xFCF1, 'M', u'يه'),
    (0xFCF2, 'M', u'ـَّ'),
    (0xFCF3, 'M', u'ـُّ'),
    (0xFCF4, 'M', u'ـِّ'),
    (0xFCF5, 'M', u'طى'),
    (0xFCF6, 'M', u'طي'),
    (0xFCF7, 'M', u'عى'),
    (0xFCF8, 'M', u'عي'),
    (0xFCF9, 'M', u'غى'),
    (0xFCFA, 'M', u'غي'),
    (0xFCFB, 'M', u'سى'),
    (0xFCFC, 'M', u'سي'),
    (0xFCFD, 'M', u'شى'),
    (0xFCFE, 'M', u'شي'),
    (0xFCFF, 'M', u'حى'),
    (0xFD00, 'M', u'حي'),
    (0xFD01, 'M', u'جى'),
    (0xFD02, 'M', u'جي'),
    (0xFD03, 'M', u'خى'),
    (0xFD04, 'M', u'خي'),
    (0xFD05, 'M', u'صى'),
    (0xFD06, 'M', u'صي'),
    (0xFD07, 'M', u'ضى'),
    (0xFD08, 'M', u'ضي'),
    (0xFD09, 'M', u'شج'),
    (0xFD0A, 'M', u'شح'),
    (0xFD0B, 'M', u'شخ'),
    (0xFD0C, 'M', u'شم'),
    (0xFD0D, 'M', u'شر'),
    (0xFD0E, 'M', u'سر'),
    (0xFD0F, 'M', u'صر'),
    (0xFD10, 'M', u'ضر'),
    (0xFD11, 'M', u'طى'),
    (0xFD12, 'M', u'طي'),
    (0xFD13, 'M', u'عى'),
    (0xFD14, 'M', u'عي'),
    (0xFD15, 'M', u'غى'),
    (0xFD16, 'M', u'غي'),
    (0xFD17, 'M', u'سى'),
    (0xFD18, 'M', u'سي'),
    (0xFD19, 'M', u'شى'),
    (0xFD1A, 'M', u'شي'),
    (0xFD1B, 'M', u'حى'),
    (0xFD1C, 'M', u'حي'),
    (0xFD1D, 'M', u'جى'),
    (0xFD1E, 'M', u'جي'),
    (0xFD1F, 'M', u'خى'),
    (0xFD20, 'M', u'خي'),
    (0xFD21, 'M', u'صى'),
    (0xFD22, 'M', u'صي'),
    (0xFD23, 'M', u'ضى'),
    (0xFD24, 'M', u'ضي'),
    (0xFD25, 'M', u'شج'),
    (0xFD26, 'M', u'شح'),
    (0xFD27, 'M', u'شخ'),
    (0xFD28, 'M', u'شم'),
    (0xFD29, 'M', u'شر'),
    (0xFD2A, 'M', u'سر'),
    (0xFD2B, 'M', u'صر'),
    (0xFD2C, 'M', u'ضر'),
    (0xFD2D, 'M', u'شج'),
    (0xFD2E, 'M', u'شح'),
    (0xFD2F, 'M', u'شخ'),
    (0xFD30, 'M', u'شم'),
    (0xFD31, 'M', u'سه'),
    (0xFD32, 'M', u'شه'),
    (0xFD33, 'M', u'طم'),
    (0xFD34, 'M', u'سج'),
    (0xFD35, 'M', u'سح'),
    (0xFD36, 'M', u'سخ'),
    (0xFD37, 'M', u'شج'),
    (0xFD38, 'M', u'شح'),
    (0xFD39, 'M', u'شخ'),
    (0xFD3A, 'M', u'طم'),
    (0xFD3B, 'M', u'ظم'),
    (0xFD3C, 'M', u'اً'),
    (0xFD3E, 'V'),
    (0xFD40, 'X'),
    (0xFD50, 'M', u'تجم'),
    (0xFD51, 'M', u'تحج'),
    (0xFD53, 'M', u'تحم'),
    (0xFD54, 'M', u'تخم'),
    (0xFD55, 'M', u'تمج'),
    (0xFD56, 'M', u'تمح'),
    (0xFD57, 'M', u'تمخ'),
    (0xFD58, 'M', u'جمح'),
    (0xFD5A, 'M', u'حمي'),
    (0xFD5B, 'M', u'حمى'),
    (0xFD5C, 'M', u'سحج'),
    (0xFD5D, 'M', u'سجح'),
    (0xFD5E, 'M', u'سجى'),
    (0xFD5F, 'M', u'سمح'),
    (0xFD61, 'M', u'سمج'),
    (0xFD62, 'M', u'سمم'),
    (0xFD64, 'M', u'صحح'),
    (0xFD66, 'M', u'صمم'),
    (0xFD67, 'M', u'شحم'),
    (0xFD69, 'M', u'شجي'),
    ]

def _seg_48():
    return [
    (0xFD6A, 'M', u'شمخ'),
    (0xFD6C, 'M', u'شمم'),
    (0xFD6E, 'M', u'ضحى'),
    (0xFD6F, 'M', u'ضخم'),
    (0xFD71, 'M', u'طمح'),
    (0xFD73, 'M', u'طمم'),
    (0xFD74, 'M', u'طمي'),
    (0xFD75, 'M', u'عجم'),
    (0xFD76, 'M', u'عمم'),
    (0xFD78, 'M', u'عمى'),
    (0xFD79, 'M', u'غمم'),
    (0xFD7A, 'M', u'غمي'),
    (0xFD7B, 'M', u'غمى'),
    (0xFD7C, 'M', u'فخم'),
    (0xFD7E, 'M', u'قمح'),
    (0xFD7F, 'M', u'قمم'),
    (0xFD80, 'M', u'لحم'),
    (0xFD81, 'M', u'لحي'),
    (0xFD82, 'M', u'لحى'),
    (0xFD83, 'M', u'لجج'),
    (0xFD85, 'M', u'لخم'),
    (0xFD87, 'M', u'لمح'),
    (0xFD89, 'M', u'محج'),
    (0xFD8A, 'M', u'محم'),
    (0xFD8B, 'M', u'محي'),
    (0xFD8C, 'M', u'مجح'),
    (0xFD8D, 'M', u'مجم'),
    (0xFD8E, 'M', u'مخج'),
    (0xFD8F, 'M', u'مخم'),
    (0xFD90, 'X'),
    (0xFD92, 'M', u'مجخ'),
    (0xFD93, 'M', u'همج'),
    (0xFD94, 'M', u'همم'),
    (0xFD95, 'M', u'نحم'),
    (0xFD96, 'M', u'نحى'),
    (0xFD97, 'M', u'نجم'),
    (0xFD99, 'M', u'نجى'),
    (0xFD9A, 'M', u'نمي'),
    (0xFD9B, 'M', u'نمى'),
    (0xFD9C, 'M', u'يمم'),
    (0xFD9E, 'M', u'بخي'),
    (0xFD9F, 'M', u'تجي'),
    (0xFDA0, 'M', u'تجى'),
    (0xFDA1, 'M', u'تخي'),
    (0xFDA2, 'M', u'تخى'),
    (0xFDA3, 'M', u'تمي'),
    (0xFDA4, 'M', u'تمى'),
    (0xFDA5, 'M', u'جمي'),
    (0xFDA6, 'M', u'جحى'),
    (0xFDA7, 'M', u'جمى'),
    (0xFDA8, 'M', u'سخى'),
    (0xFDA9, 'M', u'صحي'),
    (0xFDAA, 'M', u'شحي'),
    (0xFDAB, 'M', u'ضحي'),
    (0xFDAC, 'M', u'لجي'),
    (0xFDAD, 'M', u'لمي'),
    (0xFDAE, 'M', u'يحي'),
    (0xFDAF, 'M', u'يجي'),
    (0xFDB0, 'M', u'يمي'),
    (0xFDB1, 'M', u'ممي'),
    (0xFDB2, 'M', u'قمي'),
    (0xFDB3, 'M', u'نحي'),
    (0xFDB4, 'M', u'قمح'),
    (0xFDB5, 'M', u'لحم'),
    (0xFDB6, 'M', u'عمي'),
    (0xFDB7, 'M', u'كمي'),
    (0xFDB8, 'M', u'نجح'),
    (0xFDB9, 'M', u'مخي'),
    (0xFDBA, 'M', u'لجم'),
    (0xFDBB, 'M', u'كمم'),
    (0xFDBC, 'M', u'لجم'),
    (0xFDBD, 'M', u'نجح'),
    (0xFDBE, 'M', u'جحي'),
    (0xFDBF, 'M', u'حجي'),
    (0xFDC0, 'M', u'مجي'),
    (0xFDC1, 'M', u'فمي'),
    (0xFDC2, 'M', u'بحي'),
    (0xFDC3, 'M', u'كمم'),
    (0xFDC4, 'M', u'عجم'),
    (0xFDC5, 'M', u'صمم'),
    (0xFDC6, 'M', u'سخي'),
    (0xFDC7, 'M', u'نجي'),
    (0xFDC8, 'X'),
    (0xFDF0, 'M', u'صلے'),
    (0xFDF1, 'M', u'قلے'),
    (0xFDF2, 'M', u'الله'),
    (0xFDF3, 'M', u'اكبر'),
    (0xFDF4, 'M', u'محمد'),
    (0xFDF5, 'M', u'صلعم'),
    (0xFDF6, 'M', u'رسول'),
    (0xFDF7, 'M', u'عليه'),
    (0xFDF8, 'M', u'وسلم'),
    (0xFDF9, 'M', u'صلى'),
    (0xFDFA, '3', u'صلى الله عليه وسلم'),
    (0xFDFB, '3', u'جل جلاله'),
    (0xFDFC, 'M', u'ریال'),
    (0xFDFD, 'V'),
    (0xFDFE, 'X'),
    (0xFE00, 'I'),
    (0xFE10, '3', u','),
    ]

def _seg_49():
    return [
    (0xFE11, 'M', u'、'),
    (0xFE12, 'X'),
    (0xFE13, '3', u':'),
    (0xFE14, '3', u';'),
    (0xFE15, '3', u'!'),
    (0xFE16, '3', u'?'),
    (0xFE17, 'M', u'〖'),
    (0xFE18, 'M', u'〗'),
    (0xFE19, 'X'),
    (0xFE20, 'V'),
    (0xFE30, 'X'),
    (0xFE31, 'M', u'—'),
    (0xFE32, 'M', u'–'),
    (0xFE33, '3', u'_'),
    (0xFE35, '3', u'('),
    (0xFE36, '3', u')'),
    (0xFE37, '3', u'{'),
    (0xFE38, '3', u'}'),
    (0xFE39, 'M', u'〔'),
    (0xFE3A, 'M', u'〕'),
    (0xFE3B, 'M', u'【'),
    (0xFE3C, 'M', u'】'),
    (0xFE3D, 'M', u'《'),
    (0xFE3E, 'M', u'》'),
    (0xFE3F, 'M', u'〈'),
    (0xFE40, 'M', u'〉'),
    (0xFE41, 'M', u'「'),
    (0xFE42, 'M', u'」'),
    (0xFE43, 'M', u'『'),
    (0xFE44, 'M', u'』'),
    (0xFE45, 'V'),
    (0xFE47, '3', u'['),
    (0xFE48, '3', u']'),
    (0xFE49, '3', u' ̅'),
    (0xFE4D, '3', u'_'),
    (0xFE50, '3', u','),
    (0xFE51, 'M', u'、'),
    (0xFE52, 'X'),
    (0xFE54, '3', u';'),
    (0xFE55, '3', u':'),
    (0xFE56, '3', u'?'),
    (0xFE57, '3', u'!'),
    (0xFE58, 'M', u'—'),
    (0xFE59, '3', u'('),
    (0xFE5A, '3', u')'),
    (0xFE5B, '3', u'{'),
    (0xFE5C, '3', u'}'),
    (0xFE5D, 'M', u'〔'),
    (0xFE5E, 'M', u'〕'),
    (0xFE5F, '3', u'#'),
    (0xFE60, '3', u'&'),
    (0xFE61, '3', u'*'),
    (0xFE62, '3', u'+'),
    (0xFE63, 'M', u'-'),
    (0xFE64, '3', u'<'),
    (0xFE65, '3', u'>'),
    (0xFE66, '3', u'='),
    (0xFE67, 'X'),
    (0xFE68, '3', u'\\'),
    (0xFE69, '3', u'$'),
    (0xFE6A, '3', u'%'),
    (0xFE6B, '3', u'@'),
    (0xFE6C, 'X'),
    (0xFE70, '3', u' ً'),
    (0xFE71, 'M', u'ـً'),
    (0xFE72, '3', u' ٌ'),
    (0xFE73, 'V'),
    (0xFE74, '3', u' ٍ'),
    (0xFE75, 'X'),
    (0xFE76, '3', u' َ'),
    (0xFE77, 'M', u'ـَ'),
    (0xFE78, '3', u' ُ'),
    (0xFE79, 'M', u'ـُ'),
    (0xFE7A, '3', u' ِ'),
    (0xFE7B, 'M', u'ـِ'),
    (0xFE7C, '3', u' ّ'),
    (0xFE7D, 'M', u'ـّ'),
    (0xFE7E, '3', u' ْ'),
    (0xFE7F, 'M', u'ـْ'),
    (0xFE80, 'M', u'ء'),
    (0xFE81, 'M', u'آ'),
    (0xFE83, 'M', u'أ'),
    (0xFE85, 'M', u'ؤ'),
    (0xFE87, 'M', u'إ'),
    (0xFE89, 'M', u'ئ'),
    (0xFE8D, 'M', u'ا'),
    (0xFE8F, 'M', u'ب'),
    (0xFE93, 'M', u'ة'),
    (0xFE95, 'M', u'ت'),
    (0xFE99, 'M', u'ث'),
    (0xFE9D, 'M', u'ج'),
    (0xFEA1, 'M', u'ح'),
    (0xFEA5, 'M', u'خ'),
    (0xFEA9, 'M', u'د'),
    (0xFEAB, 'M', u'ذ'),
    (0xFEAD, 'M', u'ر'),
    (0xFEAF, 'M', u'ز'),
    (0xFEB1, 'M', u'س'),
    (0xFEB5, 'M', u'ش'),
    (0xFEB9, 'M', u'ص'),
    ]

def _seg_50():
    return [
    (0xFEBD, 'M', u'ض'),
    (0xFEC1, 'M', u'ط'),
    (0xFEC5, 'M', u'ظ'),
    (0xFEC9, 'M', u'ع'),
    (0xFECD, 'M', u'غ'),
    (0xFED1, 'M', u'ف'),
    (0xFED5, 'M', u'ق'),
    (0xFED9, 'M', u'ك'),
    (0xFEDD, 'M', u'ل'),
    (0xFEE1, 'M', u'م'),
    (0xFEE5, 'M', u'ن'),
    (0xFEE9, 'M', u'ه'),
    (0xFEED, 'M', u'و'),
    (0xFEEF, 'M', u'ى'),
    (0xFEF1, 'M', u'ي'),
    (0xFEF5, 'M', u'لآ'),
    (0xFEF7, 'M', u'لأ'),
    (0xFEF9, 'M', u'لإ'),
    (0xFEFB, 'M', u'لا'),
    (0xFEFD, 'X'),
    (0xFEFF, 'I'),
    (0xFF00, 'X'),
    (0xFF01, '3', u'!'),
    (0xFF02, '3', u'"'),
    (0xFF03, '3', u'#'),
    (0xFF04, '3', u'$'),
    (0xFF05, '3', u'%'),
    (0xFF06, '3', u'&'),
    (0xFF07, '3', u'\''),
    (0xFF08, '3', u'('),
    (0xFF09, '3', u')'),
    (0xFF0A, '3', u'*'),
    (0xFF0B, '3', u'+'),
    (0xFF0C, '3', u','),
    (0xFF0D, 'M', u'-'),
    (0xFF0E, 'M', u'.'),
    (0xFF0F, '3', u'/'),
    (0xFF10, 'M', u'0'),
    (0xFF11, 'M', u'1'),
    (0xFF12, 'M', u'2'),
    (0xFF13, 'M', u'3'),
    (0xFF14, 'M', u'4'),
    (0xFF15, 'M', u'5'),
    (0xFF16, 'M', u'6'),
    (0xFF17, 'M', u'7'),
    (0xFF18, 'M', u'8'),
    (0xFF19, 'M', u'9'),
    (0xFF1A, '3', u':'),
    (0xFF1B, '3', u';'),
    (0xFF1C, '3', u'<'),
    (0xFF1D, '3', u'='),
    (0xFF1E, '3', u'>'),
    (0xFF1F, '3', u'?'),
    (0xFF20, '3', u'@'),
    (0xFF21, 'M', u'a'),
    (0xFF22, 'M', u'b'),
    (0xFF23, 'M', u'c'),
    (0xFF24, 'M', u'd'),
    (0xFF25, 'M', u'e'),
    (0xFF26, 'M', u'f'),
    (0xFF27, 'M', u'g'),
    (0xFF28, 'M', u'h'),
    (0xFF29, 'M', u'i'),
    (0xFF2A, 'M', u'j'),
    (0xFF2B, 'M', u'k'),
    (0xFF2C, 'M', u'l'),
    (0xFF2D, 'M', u'm'),
    (0xFF2E, 'M', u'n'),
    (0xFF2F, 'M', u'o'),
    (0xFF30, 'M', u'p'),
    (0xFF31, 'M', u'q'),
    (0xFF32, 'M', u'r'),
    (0xFF33, 'M', u's'),
    (0xFF34, 'M', u't'),
    (0xFF35, 'M', u'u'),
    (0xFF36, 'M', u'v'),
    (0xFF37, 'M', u'w'),
    (0xFF38, 'M', u'x'),
    (0xFF39, 'M', u'y'),
    (0xFF3A, 'M', u'z'),
    (0xFF3B, '3', u'['),
    (0xFF3C, '3', u'\\'),
    (0xFF3D, '3', u']'),
    (0xFF3E, '3', u'^'),
    (0xFF3F, '3', u'_'),
    (0xFF40, '3', u'`'),
    (0xFF41, 'M', u'a'),
    (0xFF42, 'M', u'b'),
    (0xFF43, 'M', u'c'),
    (0xFF44, 'M', u'd'),
    (0xFF45, 'M', u'e'),
    (0xFF46, 'M', u'f'),
    (0xFF47, 'M', u'g'),
    (0xFF48, 'M', u'h'),
    (0xFF49, 'M', u'i'),
    (0xFF4A, 'M', u'j'),
    (0xFF4B, 'M', u'k'),
    (0xFF4C, 'M', u'l'),
    (0xFF4D, 'M', u'm'),
    (0xFF4E, 'M', u'n'),
    ]

def _seg_51():
    return [
    (0xFF4F, 'M', u'o'),
    (0xFF50, 'M', u'p'),
    (0xFF51, 'M', u'q'),
    (0xFF52, 'M', u'r'),
    (0xFF53, 'M', u's'),
    (0xFF54, 'M', u't'),
    (0xFF55, 'M', u'u'),
    (0xFF56, 'M', u'v'),
    (0xFF57, 'M', u'w'),
    (0xFF58, 'M', u'x'),
    (0xFF59, 'M', u'y'),
    (0xFF5A, 'M', u'z'),
    (0xFF5B, '3', u'{'),
    (0xFF5C, '3', u'|'),
    (0xFF5D, '3', u'}'),
    (0xFF5E, '3', u'~'),
    (0xFF5F, 'M', u'⦅'),
    (0xFF60, 'M', u'⦆'),
    (0xFF61, 'M', u'.'),
    (0xFF62, 'M', u'「'),
    (0xFF63, 'M', u'」'),
    (0xFF64, 'M', u'、'),
    (0xFF65, 'M', u'・'),
    (0xFF66, 'M', u'ヲ'),
    (0xFF67, 'M', u'ァ'),
    (0xFF68, 'M', u'ィ'),
    (0xFF69, 'M', u'ゥ'),
    (0xFF6A, 'M', u'ェ'),
    (0xFF6B, 'M', u'ォ'),
    (0xFF6C, 'M', u'ャ'),
    (0xFF6D, 'M', u'ュ'),
    (0xFF6E, 'M', u'ョ'),
    (0xFF6F, 'M', u'ッ'),
    (0xFF70, 'M', u'ー'),
    (0xFF71, 'M', u'ア'),
    (0xFF72, 'M', u'イ'),
    (0xFF73, 'M', u'ウ'),
    (0xFF74, 'M', u'エ'),
    (0xFF75, 'M', u'オ'),
    (0xFF76, 'M', u'カ'),
    (0xFF77, 'M', u'キ'),
    (0xFF78, 'M', u'ク'),
    (0xFF79, 'M', u'ケ'),
    (0xFF7A, 'M', u'コ'),
    (0xFF7B, 'M', u'サ'),
    (0xFF7C, 'M', u'シ'),
    (0xFF7D, 'M', u'ス'),
    (0xFF7E, 'M', u'セ'),
    (0xFF7F, 'M', u'ソ'),
    (0xFF80, 'M', u'タ'),
    (0xFF81, 'M', u'チ'),
    (0xFF82, 'M', u'ツ'),
    (0xFF83, 'M', u'テ'),
    (0xFF84, 'M', u'ト'),
    (0xFF85, 'M', u'ナ'),
    (0xFF86, 'M', u'ニ'),
    (0xFF87, 'M', u'ヌ'),
    (0xFF88, 'M', u'ネ'),
    (0xFF89, 'M', u'ノ'),
    (0xFF8A, 'M', u'ハ'),
    (0xFF8B, 'M', u'ヒ'),
    (0xFF8C, 'M', u'フ'),
    (0xFF8D, 'M', u'ヘ'),
    (0xFF8E, 'M', u'ホ'),
    (0xFF8F, 'M', u'マ'),
    (0xFF90, 'M', u'ミ'),
    (0xFF91, 'M', u'ム'),
    (0xFF92, 'M', u'メ'),
    (0xFF93, 'M', u'モ'),
    (0xFF94, 'M', u'ヤ'),
    (0xFF95, 'M', u'ユ'),
    (0xFF96, 'M', u'ヨ'),
    (0xFF97, 'M', u'ラ'),
    (0xFF98, 'M', u'リ'),
    (0xFF99, 'M', u'ル'),
    (0xFF9A, 'M', u'レ'),
    (0xFF9B, 'M', u'ロ'),
    (0xFF9C, 'M', u'ワ'),
    (0xFF9D, 'M', u'ン'),
    (0xFF9E, 'M', u'゙'),
    (0xFF9F, 'M', u'゚'),
    (0xFFA0, 'X'),
    (0xFFA1, 'M', u'ᄀ'),
    (0xFFA2, 'M', u'ᄁ'),
    (0xFFA3, 'M', u'ᆪ'),
    (0xFFA4, 'M', u'ᄂ'),
    (0xFFA5, 'M', u'ᆬ'),
    (0xFFA6, 'M', u'ᆭ'),
    (0xFFA7, 'M', u'ᄃ'),
    (0xFFA8, 'M', u'ᄄ'),
    (0xFFA9, 'M', u'ᄅ'),
    (0xFFAA, 'M', u'ᆰ'),
    (0xFFAB, 'M', u'ᆱ'),
    (0xFFAC, 'M', u'ᆲ'),
    (0xFFAD, 'M', u'ᆳ'),
    (0xFFAE, 'M', u'ᆴ'),
    (0xFFAF, 'M', u'ᆵ'),
    (0xFFB0, 'M', u'ᄚ'),
    (0xFFB1, 'M', u'ᄆ'),
    (0xFFB2, 'M', u'ᄇ'),
    ]

def _seg_52():
    return [
    (0xFFB3, 'M', u'ᄈ'),
    (0xFFB4, 'M', u'ᄡ'),
    (0xFFB5, 'M', u'ᄉ'),
    (0xFFB6, 'M', u'ᄊ'),
    (0xFFB7, 'M', u'ᄋ'),
    (0xFFB8, 'M', u'ᄌ'),
    (0xFFB9, 'M', u'ᄍ'),
    (0xFFBA, 'M', u'ᄎ'),
    (0xFFBB, 'M', u'ᄏ'),
    (0xFFBC, 'M', u'ᄐ'),
    (0xFFBD, 'M', u'ᄑ'),
    (0xFFBE, 'M', u'ᄒ'),
    (0xFFBF, 'X'),
    (0xFFC2, 'M', u'ᅡ'),
    (0xFFC3, 'M', u'ᅢ'),
    (0xFFC4, 'M', u'ᅣ'),
    (0xFFC5, 'M', u'ᅤ'),
    (0xFFC6, 'M', u'ᅥ'),
    (0xFFC7, 'M', u'ᅦ'),
    (0xFFC8, 'X'),
    (0xFFCA, 'M', u'ᅧ'),
    (0xFFCB, 'M', u'ᅨ'),
    (0xFFCC, 'M', u'ᅩ'),
    (0xFFCD, 'M', u'ᅪ'),
    (0xFFCE, 'M', u'ᅫ'),
    (0xFFCF, 'M', u'ᅬ'),
    (0xFFD0, 'X'),
    (0xFFD2, 'M', u'ᅭ'),
    (0xFFD3, 'M', u'ᅮ'),
    (0xFFD4, 'M', u'ᅯ'),
    (0xFFD5, 'M', u'ᅰ'),
    (0xFFD6, 'M', u'ᅱ'),
    (0xFFD7, 'M', u'ᅲ'),
    (0xFFD8, 'X'),
    (0xFFDA, 'M', u'ᅳ'),
    (0xFFDB, 'M', u'ᅴ'),
    (0xFFDC, 'M', u'ᅵ'),
    (0xFFDD, 'X'),
    (0xFFE0, 'M', u'¢'),
    (0xFFE1, 'M', u'£'),
    (0xFFE2, 'M', u'¬'),
    (0xFFE3, '3', u' ̄'),
    (0xFFE4, 'M', u'¦'),
    (0xFFE5, 'M', u'¥'),
    (0xFFE6, 'M', u'₩'),
    (0xFFE7, 'X'),
    (0xFFE8, 'M', u'│'),
    (0xFFE9, 'M', u'←'),
    (0xFFEA, 'M', u'↑'),
    (0xFFEB, 'M', u'→'),
    (0xFFEC, 'M', u'↓'),
    (0xFFED, 'M', u'■'),
    (0xFFEE, 'M', u'○'),
    (0xFFEF, 'X'),
    (0x10000, 'V'),
    (0x1000C, 'X'),
    (0x1000D, 'V'),
    (0x10027, 'X'),
    (0x10028, 'V'),
    (0x1003B, 'X'),
    (0x1003C, 'V'),
    (0x1003E, 'X'),
    (0x1003F, 'V'),
    (0x1004E, 'X'),
    (0x10050, 'V'),
    (0x1005E, 'X'),
    (0x10080, 'V'),
    (0x100FB, 'X'),
    (0x10100, 'V'),
    (0x10103, 'X'),
    (0x10107, 'V'),
    (0x10134, 'X'),
    (0x10137, 'V'),
    (0x1018F, 'X'),
    (0x10190, 'V'),
    (0x1019C, 'X'),
    (0x101A0, 'V'),
    (0x101A1, 'X'),
    (0x101D0, 'V'),
    (0x101FE, 'X'),
    (0x10280, 'V'),
    (0x1029D, 'X'),
    (0x102A0, 'V'),
    (0x102D1, 'X'),
    (0x102E0, 'V'),
    (0x102FC, 'X'),
    (0x10300, 'V'),
    (0x10324, 'X'),
    (0x1032D, 'V'),
    (0x1034B, 'X'),
    (0x10350, 'V'),
    (0x1037B, 'X'),
    (0x10380, 'V'),
    (0x1039E, 'X'),
    (0x1039F, 'V'),
    (0x103C4, 'X'),
    (0x103C8, 'V'),
    (0x103D6, 'X'),
    (0x10400, 'M', u'𐐨'),
    (0x10401, 'M', u'𐐩'),
    ]

def _seg_53():
    return [
    (0x10402, 'M', u'𐐪'),
    (0x10403, 'M', u'𐐫'),
    (0x10404, 'M', u'𐐬'),
    (0x10405, 'M', u'𐐭'),
    (0x10406, 'M', u'𐐮'),
    (0x10407, 'M', u'𐐯'),
    (0x10408, 'M', u'𐐰'),
    (0x10409, 'M', u'𐐱'),
    (0x1040A, 'M', u'𐐲'),
    (0x1040B, 'M', u'𐐳'),
    (0x1040C, 'M', u'𐐴'),
    (0x1040D, 'M', u'𐐵'),
    (0x1040E, 'M', u'𐐶'),
    (0x1040F, 'M', u'𐐷'),
    (0x10410, 'M', u'𐐸'),
    (0x10411, 'M', u'𐐹'),
    (0x10412, 'M', u'𐐺'),
    (0x10413, 'M', u'𐐻'),
    (0x10414, 'M', u'𐐼'),
    (0x10415, 'M', u'𐐽'),
    (0x10416, 'M', u'𐐾'),
    (0x10417, 'M', u'𐐿'),
    (0x10418, 'M', u'𐑀'),
    (0x10419, 'M', u'𐑁'),
    (0x1041A, 'M', u'𐑂'),
    (0x1041B, 'M', u'𐑃'),
    (0x1041C, 'M', u'𐑄'),
    (0x1041D, 'M', u'𐑅'),
    (0x1041E, 'M', u'𐑆'),
    (0x1041F, 'M', u'𐑇'),
    (0x10420, 'M', u'𐑈'),
    (0x10421, 'M', u'𐑉'),
    (0x10422, 'M', u'𐑊'),
    (0x10423, 'M', u'𐑋'),
    (0x10424, 'M', u'𐑌'),
    (0x10425, 'M', u'𐑍'),
    (0x10426, 'M', u'𐑎'),
    (0x10427, 'M', u'𐑏'),
    (0x10428, 'V'),
    (0x1049E, 'X'),
    (0x104A0, 'V'),
    (0x104AA, 'X'),
    (0x104B0, 'M', u'𐓘'),
    (0x104B1, 'M', u'𐓙'),
    (0x104B2, 'M', u'𐓚'),
    (0x104B3, 'M', u'𐓛'),
    (0x104B4, 'M', u'𐓜'),
    (0x104B5, 'M', u'𐓝'),
    (0x104B6, 'M', u'𐓞'),
    (0x104B7, 'M', u'𐓟'),
    (0x104B8, 'M', u'𐓠'),
    (0x104B9, 'M', u'𐓡'),
    (0x104BA, 'M', u'𐓢'),
    (0x104BB, 'M', u'𐓣'),
    (0x104BC, 'M', u'𐓤'),
    (0x104BD, 'M', u'𐓥'),
    (0x104BE, 'M', u'𐓦'),
    (0x104BF, 'M', u'𐓧'),
    (0x104C0, 'M', u'𐓨'),
    (0x104C1, 'M', u'𐓩'),
    (0x104C2, 'M', u'𐓪'),
    (0x104C3, 'M', u'𐓫'),
    (0x104C4, 'M', u'𐓬'),
    (0x104C5, 'M', u'𐓭'),
    (0x104C6, 'M', u'𐓮'),
    (0x104C7, 'M', u'𐓯'),
    (0x104C8, 'M', u'𐓰'),
    (0x104C9, 'M', u'𐓱'),
    (0x104CA, 'M', u'𐓲'),
    (0x104CB, 'M', u'𐓳'),
    (0x104CC, 'M', u'𐓴'),
    (0x104CD, 'M', u'𐓵'),
    (0x104CE, 'M', u'𐓶'),
    (0x104CF, 'M', u'𐓷'),
    (0x104D0, 'M', u'𐓸'),
    (0x104D1, 'M', u'𐓹'),
    (0x104D2, 'M', u'𐓺'),
    (0x104D3, 'M', u'𐓻'),
    (0x104D4, 'X'),
    (0x104D8, 'V'),
    (0x104FC, 'X'),
    (0x10500, 'V'),
    (0x10528, 'X'),
    (0x10530, 'V'),
    (0x10564, 'X'),
    (0x1056F, 'V'),
    (0x10570, 'X'),
    (0x10600, 'V'),
    (0x10737, 'X'),
    (0x10740, 'V'),
    (0x10756, 'X'),
    (0x10760, 'V'),
    (0x10768, 'X'),
    (0x10800, 'V'),
    (0x10806, 'X'),
    (0x10808, 'V'),
    (0x10809, 'X'),
    (0x1080A, 'V'),
    (0x10836, 'X'),
    (0x10837, 'V'),
    ]

def _seg_54():
    return [
    (0x10839, 'X'),
    (0x1083C, 'V'),
    (0x1083D, 'X'),
    (0x1083F, 'V'),
    (0x10856, 'X'),
    (0x10857, 'V'),
    (0x1089F, 'X'),
    (0x108A7, 'V'),
    (0x108B0, 'X'),
    (0x108E0, 'V'),
    (0x108F3, 'X'),
    (0x108F4, 'V'),
    (0x108F6, 'X'),
    (0x108FB, 'V'),
    (0x1091C, 'X'),
    (0x1091F, 'V'),
    (0x1093A, 'X'),
    (0x1093F, 'V'),
    (0x10940, 'X'),
    (0x10980, 'V'),
    (0x109B8, 'X'),
    (0x109BC, 'V'),
    (0x109D0, 'X'),
    (0x109D2, 'V'),
    (0x10A04, 'X'),
    (0x10A05, 'V'),
    (0x10A07, 'X'),
    (0x10A0C, 'V'),
    (0x10A14, 'X'),
    (0x10A15, 'V'),
    (0x10A18, 'X'),
    (0x10A19, 'V'),
    (0x10A36, 'X'),
    (0x10A38, 'V'),
    (0x10A3B, 'X'),
    (0x10A3F, 'V'),
    (0x10A49, 'X'),
    (0x10A50, 'V'),
    (0x10A59, 'X'),
    (0x10A60, 'V'),
    (0x10AA0, 'X'),
    (0x10AC0, 'V'),
    (0x10AE7, 'X'),
    (0x10AEB, 'V'),
    (0x10AF7, 'X'),
    (0x10B00, 'V'),
    (0x10B36, 'X'),
    (0x10B39, 'V'),
    (0x10B56, 'X'),
    (0x10B58, 'V'),
    (0x10B73, 'X'),
    (0x10B78, 'V'),
    (0x10B92, 'X'),
    (0x10B99, 'V'),
    (0x10B9D, 'X'),
    (0x10BA9, 'V'),
    (0x10BB0, 'X'),
    (0x10C00, 'V'),
    (0x10C49, 'X'),
    (0x10C80, 'M', u'𐳀'),
    (0x10C81, 'M', u'𐳁'),
    (0x10C82, 'M', u'𐳂'),
    (0x10C83, 'M', u'𐳃'),
    (0x10C84, 'M', u'𐳄'),
    (0x10C85, 'M', u'𐳅'),
    (0x10C86, 'M', u'𐳆'),
    (0x10C87, 'M', u'𐳇'),
    (0x10C88, 'M', u'𐳈'),
    (0x10C89, 'M', u'𐳉'),
    (0x10C8A, 'M', u'𐳊'),
    (0x10C8B, 'M', u'𐳋'),
    (0x10C8C, 'M', u'𐳌'),
    (0x10C8D, 'M', u'𐳍'),
    (0x10C8E, 'M', u'𐳎'),
    (0x10C8F, 'M', u'𐳏'),
    (0x10C90, 'M', u'𐳐'),
    (0x10C91, 'M', u'𐳑'),
    (0x10C92, 'M', u'𐳒'),
    (0x10C93, 'M', u'𐳓'),
    (0x10C94, 'M', u'𐳔'),
    (0x10C95, 'M', u'𐳕'),
    (0x10C96, 'M', u'𐳖'),
    (0x10C97, 'M', u'𐳗'),
    (0x10C98, 'M', u'𐳘'),
    (0x10C99, 'M', u'𐳙'),
    (0x10C9A, 'M', u'𐳚'),
    (0x10C9B, 'M', u'𐳛'),
    (0x10C9C, 'M', u'𐳜'),
    (0x10C9D, 'M', u'𐳝'),
    (0x10C9E, 'M', u'𐳞'),
    (0x10C9F, 'M', u'𐳟'),
    (0x10CA0, 'M', u'𐳠'),
    (0x10CA1, 'M', u'𐳡'),
    (0x10CA2, 'M', u'𐳢'),
    (0x10CA3, 'M', u'𐳣'),
    (0x10CA4, 'M', u'𐳤'),
    (0x10CA5, 'M', u'𐳥'),
    (0x10CA6, 'M', u'𐳦'),
    (0x10CA7, 'M', u'𐳧'),
    (0x10CA8, 'M', u'𐳨'),
    ]

def _seg_55():
    return [
    (0x10CA9, 'M', u'𐳩'),
    (0x10CAA, 'M', u'𐳪'),
    (0x10CAB, 'M', u'𐳫'),
    (0x10CAC, 'M', u'𐳬'),
    (0x10CAD, 'M', u'𐳭'),
    (0x10CAE, 'M', u'𐳮'),
    (0x10CAF, 'M', u'𐳯'),
    (0x10CB0, 'M', u'𐳰'),
    (0x10CB1, 'M', u'𐳱'),
    (0x10CB2, 'M', u'𐳲'),
    (0x10CB3, 'X'),
    (0x10CC0, 'V'),
    (0x10CF3, 'X'),
    (0x10CFA, 'V'),
    (0x10D28, 'X'),
    (0x10D30, 'V'),
    (0x10D3A, 'X'),
    (0x10E60, 'V'),
    (0x10E7F, 'X'),
    (0x10F00, 'V'),
    (0x10F28, 'X'),
    (0x10F30, 'V'),
    (0x10F5A, 'X'),
    (0x11000, 'V'),
    (0x1104E, 'X'),
    (0x11052, 'V'),
    (0x11070, 'X'),
    (0x1107F, 'V'),
    (0x110BD, 'X'),
    (0x110BE, 'V'),
    (0x110C2, 'X'),
    (0x110D0, 'V'),
    (0x110E9, 'X'),
    (0x110F0, 'V'),
    (0x110FA, 'X'),
    (0x11100, 'V'),
    (0x11135, 'X'),
    (0x11136, 'V'),
    (0x11147, 'X'),
    (0x11150, 'V'),
    (0x11177, 'X'),
    (0x11180, 'V'),
    (0x111CE, 'X'),
    (0x111D0, 'V'),
    (0x111E0, 'X'),
    (0x111E1, 'V'),
    (0x111F5, 'X'),
    (0x11200, 'V'),
    (0x11212, 'X'),
    (0x11213, 'V'),
    (0x1123F, 'X'),
    (0x11280, 'V'),
    (0x11287, 'X'),
    (0x11288, 'V'),
    (0x11289, 'X'),
    (0x1128A, 'V'),
    (0x1128E, 'X'),
    (0x1128F, 'V'),
    (0x1129E, 'X'),
    (0x1129F, 'V'),
    (0x112AA, 'X'),
    (0x112B0, 'V'),
    (0x112EB, 'X'),
    (0x112F0, 'V'),
    (0x112FA, 'X'),
    (0x11300, 'V'),
    (0x11304, 'X'),
    (0x11305, 'V'),
    (0x1130D, 'X'),
    (0x1130F, 'V'),
    (0x11311, 'X'),
    (0x11313, 'V'),
    (0x11329, 'X'),
    (0x1132A, 'V'),
    (0x11331, 'X'),
    (0x11332, 'V'),
    (0x11334, 'X'),
    (0x11335, 'V'),
    (0x1133A, 'X'),
    (0x1133B, 'V'),
    (0x11345, 'X'),
    (0x11347, 'V'),
    (0x11349, 'X'),
    (0x1134B, 'V'),
    (0x1134E, 'X'),
    (0x11350, 'V'),
    (0x11351, 'X'),
    (0x11357, 'V'),
    (0x11358, 'X'),
    (0x1135D, 'V'),
    (0x11364, 'X'),
    (0x11366, 'V'),
    (0x1136D, 'X'),
    (0x11370, 'V'),
    (0x11375, 'X'),
    (0x11400, 'V'),
    (0x1145A, 'X'),
    (0x1145B, 'V'),
    (0x1145C, 'X'),
    (0x1145D, 'V'),
    ]

def _seg_56():
    return [
    (0x1145F, 'X'),
    (0x11480, 'V'),
    (0x114C8, 'X'),
    (0x114D0, 'V'),
    (0x114DA, 'X'),
    (0x11580, 'V'),
    (0x115B6, 'X'),
    (0x115B8, 'V'),
    (0x115DE, 'X'),
    (0x11600, 'V'),
    (0x11645, 'X'),
    (0x11650, 'V'),
    (0x1165A, 'X'),
    (0x11660, 'V'),
    (0x1166D, 'X'),
    (0x11680, 'V'),
    (0x116B8, 'X'),
    (0x116C0, 'V'),
    (0x116CA, 'X'),
    (0x11700, 'V'),
    (0x1171B, 'X'),
    (0x1171D, 'V'),
    (0x1172C, 'X'),
    (0x11730, 'V'),
    (0x11740, 'X'),
    (0x11800, 'V'),
    (0x1183C, 'X'),
    (0x118A0, 'M', u'𑣀'),
    (0x118A1, 'M', u'𑣁'),
    (0x118A2, 'M', u'𑣂'),
    (0x118A3, 'M', u'𑣃'),
    (0x118A4, 'M', u'𑣄'),
    (0x118A5, 'M', u'𑣅'),
    (0x118A6, 'M', u'𑣆'),
    (0x118A7, 'M', u'𑣇'),
    (0x118A8, 'M', u'𑣈'),
    (0x118A9, 'M', u'𑣉'),
    (0x118AA, 'M', u'𑣊'),
    (0x118AB, 'M', u'𑣋'),
    (0x118AC, 'M', u'𑣌'),
    (0x118AD, 'M', u'𑣍'),
    (0x118AE, 'M', u'𑣎'),
    (0x118AF, 'M', u'𑣏'),
    (0x118B0, 'M', u'𑣐'),
    (0x118B1, 'M', u'𑣑'),
    (0x118B2, 'M', u'𑣒'),
    (0x118B3, 'M', u'𑣓'),
    (0x118B4, 'M', u'𑣔'),
    (0x118B5, 'M', u'𑣕'),
    (0x118B6, 'M', u'𑣖'),
    (0x118B7, 'M', u'𑣗'),
    (0x118B8, 'M', u'𑣘'),
    (0x118B9, 'M', u'𑣙'),
    (0x118BA, 'M', u'𑣚'),
    (0x118BB, 'M', u'𑣛'),
    (0x118BC, 'M', u'𑣜'),
    (0x118BD, 'M', u'𑣝'),
    (0x118BE, 'M', u'𑣞'),
    (0x118BF, 'M', u'𑣟'),
    (0x118C0, 'V'),
    (0x118F3, 'X'),
    (0x118FF, 'V'),
    (0x11900, 'X'),
    (0x11A00, 'V'),
    (0x11A48, 'X'),
    (0x11A50, 'V'),
    (0x11A84, 'X'),
    (0x11A86, 'V'),
    (0x11AA3, 'X'),
    (0x11AC0, 'V'),
    (0x11AF9, 'X'),
    (0x11C00, 'V'),
    (0x11C09, 'X'),
    (0x11C0A, 'V'),
    (0x11C37, 'X'),
    (0x11C38, 'V'),
    (0x11C46, 'X'),
    (0x11C50, 'V'),
    (0x11C6D, 'X'),
    (0x11C70, 'V'),
    (0x11C90, 'X'),
    (0x11C92, 'V'),
    (0x11CA8, 'X'),
    (0x11CA9, 'V'),
    (0x11CB7, 'X'),
    (0x11D00, 'V'),
    (0x11D07, 'X'),
    (0x11D08, 'V'),
    (0x11D0A, 'X'),
    (0x11D0B, 'V'),
    (0x11D37, 'X'),
    (0x11D3A, 'V'),
    (0x11D3B, 'X'),
    (0x11D3C, 'V'),
    (0x11D3E, 'X'),
    (0x11D3F, 'V'),
    (0x11D48, 'X'),
    (0x11D50, 'V'),
    (0x11D5A, 'X'),
    (0x11D60, 'V'),
    ]

def _seg_57():
    return [
    (0x11D66, 'X'),
    (0x11D67, 'V'),
    (0x11D69, 'X'),
    (0x11D6A, 'V'),
    (0x11D8F, 'X'),
    (0x11D90, 'V'),
    (0x11D92, 'X'),
    (0x11D93, 'V'),
    (0x11D99, 'X'),
    (0x11DA0, 'V'),
    (0x11DAA, 'X'),
    (0x11EE0, 'V'),
    (0x11EF9, 'X'),
    (0x12000, 'V'),
    (0x1239A, 'X'),
    (0x12400, 'V'),
    (0x1246F, 'X'),
    (0x12470, 'V'),
    (0x12475, 'X'),
    (0x12480, 'V'),
    (0x12544, 'X'),
    (0x13000, 'V'),
    (0x1342F, 'X'),
    (0x14400, 'V'),
    (0x14647, 'X'),
    (0x16800, 'V'),
    (0x16A39, 'X'),
    (0x16A40, 'V'),
    (0x16A5F, 'X'),
    (0x16A60, 'V'),
    (0x16A6A, 'X'),
    (0x16A6E, 'V'),
    (0x16A70, 'X'),
    (0x16AD0, 'V'),
    (0x16AEE, 'X'),
    (0x16AF0, 'V'),
    (0x16AF6, 'X'),
    (0x16B00, 'V'),
    (0x16B46, 'X'),
    (0x16B50, 'V'),
    (0x16B5A, 'X'),
    (0x16B5B, 'V'),
    (0x16B62, 'X'),
    (0x16B63, 'V'),
    (0x16B78, 'X'),
    (0x16B7D, 'V'),
    (0x16B90, 'X'),
    (0x16E60, 'V'),
    (0x16E9B, 'X'),
    (0x16F00, 'V'),
    (0x16F45, 'X'),
    (0x16F50, 'V'),
    (0x16F7F, 'X'),
    (0x16F8F, 'V'),
    (0x16FA0, 'X'),
    (0x16FE0, 'V'),
    (0x16FE2, 'X'),
    (0x17000, 'V'),
    (0x187F2, 'X'),
    (0x18800, 'V'),
    (0x18AF3, 'X'),
    (0x1B000, 'V'),
    (0x1B11F, 'X'),
    (0x1B170, 'V'),
    (0x1B2FC, 'X'),
    (0x1BC00, 'V'),
    (0x1BC6B, 'X'),
    (0x1BC70, 'V'),
    (0x1BC7D, 'X'),
    (0x1BC80, 'V'),
    (0x1BC89, 'X'),
    (0x1BC90, 'V'),
    (0x1BC9A, 'X'),
    (0x1BC9C, 'V'),
    (0x1BCA0, 'I'),
    (0x1BCA4, 'X'),
    (0x1D000, 'V'),
    (0x1D0F6, 'X'),
    (0x1D100, 'V'),
    (0x1D127, 'X'),
    (0x1D129, 'V'),
    (0x1D15E, 'M', u'𝅗𝅥'),
    (0x1D15F, 'M', u'𝅘𝅥'),
    (0x1D160, 'M', u'𝅘𝅥𝅮'),
    (0x1D161, 'M', u'𝅘𝅥𝅯'),
    (0x1D162, 'M', u'𝅘𝅥𝅰'),
    (0x1D163, 'M', u'𝅘𝅥𝅱'),
    (0x1D164, 'M', u'𝅘𝅥𝅲'),
    (0x1D165, 'V'),
    (0x1D173, 'X'),
    (0x1D17B, 'V'),
    (0x1D1BB, 'M', u'𝆹𝅥'),
    (0x1D1BC, 'M', u'𝆺𝅥'),
    (0x1D1BD, 'M', u'𝆹𝅥𝅮'),
    (0x1D1BE, 'M', u'𝆺𝅥𝅮'),
    (0x1D1BF, 'M', u'𝆹𝅥𝅯'),
    (0x1D1C0, 'M', u'𝆺𝅥𝅯'),
    (0x1D1C1, 'V'),
    (0x1D1E9, 'X'),
    (0x1D200, 'V'),
    ]

def _seg_58():
    return [
    (0x1D246, 'X'),
    (0x1D2E0, 'V'),
    (0x1D2F4, 'X'),
    (0x1D300, 'V'),
    (0x1D357, 'X'),
    (0x1D360, 'V'),
    (0x1D379, 'X'),
    (0x1D400, 'M', u'a'),
    (0x1D401, 'M', u'b'),
    (0x1D402, 'M', u'c'),
    (0x1D403, 'M', u'd'),
    (0x1D404, 'M', u'e'),
    (0x1D405, 'M', u'f'),
    (0x1D406, 'M', u'g'),
    (0x1D407, 'M', u'h'),
    (0x1D408, 'M', u'i'),
    (0x1D409, 'M', u'j'),
    (0x1D40A, 'M', u'k'),
    (0x1D40B, 'M', u'l'),
    (0x1D40C, 'M', u'm'),
    (0x1D40D, 'M', u'n'),
    (0x1D40E, 'M', u'o'),
    (0x1D40F, 'M', u'p'),
    (0x1D410, 'M', u'q'),
    (0x1D411, 'M', u'r'),
    (0x1D412, 'M', u's'),
    (0x1D413, 'M', u't'),
    (0x1D414, 'M', u'u'),
    (0x1D415, 'M', u'v'),
    (0x1D416, 'M', u'w'),
    (0x1D417, 'M', u'x'),
    (0x1D418, 'M', u'y'),
    (0x1D419, 'M', u'z'),
    (0x1D41A, 'M', u'a'),
    (0x1D41B, 'M', u'b'),
    (0x1D41C, 'M', u'c'),
    (0x1D41D, 'M', u'd'),
    (0x1D41E, 'M', u'e'),
    (0x1D41F, 'M', u'f'),
    (0x1D420, 'M', u'g'),
    (0x1D421, 'M', u'h'),
    (0x1D422, 'M', u'i'),
    (0x1D423, 'M', u'j'),
    (0x1D424, 'M', u'k'),
    (0x1D425, 'M', u'l'),
    (0x1D426, 'M', u'm'),
    (0x1D427, 'M', u'n'),
    (0x1D428, 'M', u'o'),
    (0x1D429, 'M', u'p'),
    (0x1D42A, 'M', u'q'),
    (0x1D42B, 'M', u'r'),
    (0x1D42C, 'M', u's'),
    (0x1D42D, 'M', u't'),
    (0x1D42E, 'M', u'u'),
    (0x1D42F, 'M', u'v'),
    (0x1D430, 'M', u'w'),
    (0x1D431, 'M', u'x'),
    (0x1D432, 'M', u'y'),
    (0x1D433, 'M', u'z'),
    (0x1D434, 'M', u'a'),
    (0x1D435, 'M', u'b'),
    (0x1D436, 'M', u'c'),
    (0x1D437, 'M', u'd'),
    (0x1D438, 'M', u'e'),
    (0x1D439, 'M', u'f'),
    (0x1D43A, 'M', u'g'),
    (0x1D43B, 'M', u'h'),
    (0x1D43C, 'M', u'i'),
    (0x1D43D, 'M', u'j'),
    (0x1D43E, 'M', u'k'),
    (0x1D43F, 'M', u'l'),
    (0x1D440, 'M', u'm'),
    (0x1D441, 'M', u'n'),
    (0x1D442, 'M', u'o'),
    (0x1D443, 'M', u'p'),
    (0x1D444, 'M', u'q'),
    (0x1D445, 'M', u'r'),
    (0x1D446, 'M', u's'),
    (0x1D447, 'M', u't'),
    (0x1D448, 'M', u'u'),
    (0x1D449, 'M', u'v'),
    (0x1D44A, 'M', u'w'),
    (0x1D44B, 'M', u'x'),
    (0x1D44C, 'M', u'y'),
    (0x1D44D, 'M', u'z'),
    (0x1D44E, 'M', u'a'),
    (0x1D44F, 'M', u'b'),
    (0x1D450, 'M', u'c'),
    (0x1D451, 'M', u'd'),
    (0x1D452, 'M', u'e'),
    (0x1D453, 'M', u'f'),
    (0x1D454, 'M', u'g'),
    (0x1D455, 'X'),
    (0x1D456, 'M', u'i'),
    (0x1D457, 'M', u'j'),
    (0x1D458, 'M', u'k'),
    (0x1D459, 'M', u'l'),
    (0x1D45A, 'M', u'm'),
    (0x1D45B, 'M', u'n'),
    (0x1D45C, 'M', u'o'),
    ]

def _seg_59():
    return [
    (0x1D45D, 'M', u'p'),
    (0x1D45E, 'M', u'q'),
    (0x1D45F, 'M', u'r'),
    (0x1D460, 'M', u's'),
    (0x1D461, 'M', u't'),
    (0x1D462, 'M', u'u'),
    (0x1D463, 'M', u'v'),
    (0x1D464, 'M', u'w'),
    (0x1D465, 'M', u'x'),
    (0x1D466, 'M', u'y'),
    (0x1D467, 'M', u'z'),
    (0x1D468, 'M', u'a'),
    (0x1D469, 'M', u'b'),
    (0x1D46A, 'M', u'c'),
    (0x1D46B, 'M', u'd'),
    (0x1D46C, 'M', u'e'),
    (0x1D46D, 'M', u'f'),
    (0x1D46E, 'M', u'g'),
    (0x1D46F, 'M', u'h'),
    (0x1D470, 'M', u'i'),
    (0x1D471, 'M', u'j'),
    (0x1D472, 'M', u'k'),
    (0x1D473, 'M', u'l'),
    (0x1D474, 'M', u'm'),
    (0x1D475, 'M', u'n'),
    (0x1D476, 'M', u'o'),
    (0x1D477, 'M', u'p'),
    (0x1D478, 'M', u'q'),
    (0x1D479, 'M', u'r'),
    (0x1D47A, 'M', u's'),
    (0x1D47B, 'M', u't'),
    (0x1D47C, 'M', u'u'),
    (0x1D47D, 'M', u'v'),
    (0x1D47E, 'M', u'w'),
    (0x1D47F, 'M', u'x'),
    (0x1D480, 'M', u'y'),
    (0x1D481, 'M', u'z'),
    (0x1D482, 'M', u'a'),
    (0x1D483, 'M', u'b'),
    (0x1D484, 'M', u'c'),
    (0x1D485, 'M', u'd'),
    (0x1D486, 'M', u'e'),
    (0x1D487, 'M', u'f'),
    (0x1D488, 'M', u'g'),
    (0x1D489, 'M', u'h'),
    (0x1D48A, 'M', u'i'),
    (0x1D48B, 'M', u'j'),
    (0x1D48C, 'M', u'k'),
    (0x1D48D, 'M', u'l'),
    (0x1D48E, 'M', u'm'),
    (0x1D48F, 'M', u'n'),
    (0x1D490, 'M', u'o'),
    (0x1D491, 'M', u'p'),
    (0x1D492, 'M', u'q'),
    (0x1D493, 'M', u'r'),
    (0x1D494, 'M', u's'),
    (0x1D495, 'M', u't'),
    (0x1D496, 'M', u'u'),
    (0x1D497, 'M', u'v'),
    (0x1D498, 'M', u'w'),
    (0x1D499, 'M', u'x'),
    (0x1D49A, 'M', u'y'),
    (0x1D49B, 'M', u'z'),
    (0x1D49C, 'M', u'a'),
    (0x1D49D, 'X'),
    (0x1D49E, 'M', u'c'),
    (0x1D49F, 'M', u'd'),
    (0x1D4A0, 'X'),
    (0x1D4A2, 'M', u'g'),
    (0x1D4A3, 'X'),
    (0x1D4A5, 'M', u'j'),
    (0x1D4A6, 'M', u'k'),
    (0x1D4A7, 'X'),
    (0x1D4A9, 'M', u'n'),
    (0x1D4AA, 'M', u'o'),
    (0x1D4AB, 'M', u'p'),
    (0x1D4AC, 'M', u'q'),
    (0x1D4AD, 'X'),
    (0x1D4AE, 'M', u's'),
    (0x1D4AF, 'M', u't'),
    (0x1D4B0, 'M', u'u'),
    (0x1D4B1, 'M', u'v'),
    (0x1D4B2, 'M', u'w'),
    (0x1D4B3, 'M', u'x'),
    (0x1D4B4, 'M', u'y'),
    (0x1D4B5, 'M', u'z'),
    (0x1D4B6, 'M', u'a'),
    (0x1D4B7, 'M', u'b'),
    (0x1D4B8, 'M', u'c'),
    (0x1D4B9, 'M', u'd'),
    (0x1D4BA, 'X'),
    (0x1D4BB, 'M', u'f'),
    (0x1D4BC, 'X'),
    (0x1D4BD, 'M', u'h'),
    (0x1D4BE, 'M', u'i'),
    (0x1D4BF, 'M', u'j'),
    (0x1D4C0, 'M', u'k'),
    (0x1D4C1, 'M', u'l'),
    (0x1D4C2, 'M', u'm'),
    (0x1D4C3, 'M', u'n'),
    ]

def _seg_60():
    return [
    (0x1D4C4, 'X'),
    (0x1D4C5, 'M', u'p'),
    (0x1D4C6, 'M', u'q'),
    (0x1D4C7, 'M', u'r'),
    (0x1D4C8, 'M', u's'),
    (0x1D4C9, 'M', u't'),
    (0x1D4CA, 'M', u'u'),
    (0x1D4CB, 'M', u'v'),
    (0x1D4CC, 'M', u'w'),
    (0x1D4CD, 'M', u'x'),
    (0x1D4CE, 'M', u'y'),
    (0x1D4CF, 'M', u'z'),
    (0x1D4D0, 'M', u'a'),
    (0x1D4D1, 'M', u'b'),
    (0x1D4D2, 'M', u'c'),
    (0x1D4D3, 'M', u'd'),
    (0x1D4D4, 'M', u'e'),
    (0x1D4D5, 'M', u'f'),
    (0x1D4D6, 'M', u'g'),
    (0x1D4D7, 'M', u'h'),
    (0x1D4D8, 'M', u'i'),
    (0x1D4D9, 'M', u'j'),
    (0x1D4DA, 'M', u'k'),
    (0x1D4DB, 'M', u'l'),
    (0x1D4DC, 'M', u'm'),
    (0x1D4DD, 'M', u'n'),
    (0x1D4DE, 'M', u'o'),
    (0x1D4DF, 'M', u'p'),
    (0x1D4E0, 'M', u'q'),
    (0x1D4E1, 'M', u'r'),
    (0x1D4E2, 'M', u's'),
    (0x1D4E3, 'M', u't'),
    (0x1D4E4, 'M', u'u'),
    (0x1D4E5, 'M', u'v'),
    (0x1D4E6, 'M', u'w'),
    (0x1D4E7, 'M', u'x'),
    (0x1D4E8, 'M', u'y'),
    (0x1D4E9, 'M', u'z'),
    (0x1D4EA, 'M', u'a'),
    (0x1D4EB, 'M', u'b'),
    (0x1D4EC, 'M', u'c'),
    (0x1D4ED, 'M', u'd'),
    (0x1D4EE, 'M', u'e'),
    (0x1D4EF, 'M', u'f'),
    (0x1D4F0, 'M', u'g'),
    (0x1D4F1, 'M', u'h'),
    (0x1D4F2, 'M', u'i'),
    (0x1D4F3, 'M', u'j'),
    (0x1D4F4, 'M', u'k'),
    (0x1D4F5, 'M', u'l'),
    (0x1D4F6, 'M', u'm'),
    (0x1D4F7, 'M', u'n'),
    (0x1D4F8, 'M', u'o'),
    (0x1D4F9, 'M', u'p'),
    (0x1D4FA, 'M', u'q'),
    (0x1D4FB, 'M', u'r'),
    (0x1D4FC, 'M', u's'),
    (0x1D4FD, 'M', u't'),
    (0x1D4FE, 'M', u'u'),
    (0x1D4FF, 'M', u'v'),
    (0x1D500, 'M', u'w'),
    (0x1D501, 'M', u'x'),
    (0x1D502, 'M', u'y'),
    (0x1D503, 'M', u'z'),
    (0x1D504, 'M', u'a'),
    (0x1D505, 'M', u'b'),
    (0x1D506, 'X'),
    (0x1D507, 'M', u'd'),
    (0x1D508, 'M', u'e'),
    (0x1D509, 'M', u'f'),
    (0x1D50A, 'M', u'g'),
    (0x1D50B, 'X'),
    (0x1D50D, 'M', u'j'),
    (0x1D50E, 'M', u'k'),
    (0x1D50F, 'M', u'l'),
    (0x1D510, 'M', u'm'),
    (0x1D511, 'M', u'n'),
    (0x1D512, 'M', u'o'),
    (0x1D513, 'M', u'p'),
    (0x1D514, 'M', u'q'),
    (0x1D515, 'X'),
    (0x1D516, 'M', u's'),
    (0x1D517, 'M', u't'),
    (0x1D518, 'M', u'u'),
    (0x1D519, 'M', u'v'),
    (0x1D51A, 'M', u'w'),
    (0x1D51B, 'M', u'x'),
    (0x1D51C, 'M', u'y'),
    (0x1D51D, 'X'),
    (0x1D51E, 'M', u'a'),
    (0x1D51F, 'M', u'b'),
    (0x1D520, 'M', u'c'),
    (0x1D521, 'M', u'd'),
    (0x1D522, 'M', u'e'),
    (0x1D523, 'M', u'f'),
    (0x1D524, 'M', u'g'),
    (0x1D525, 'M', u'h'),
    (0x1D526, 'M', u'i'),
    (0x1D527, 'M', u'j'),
    (0x1D528, 'M', u'k'),
    ]

def _seg_61():
    return [
    (0x1D529, 'M', u'l'),
    (0x1D52A, 'M', u'm'),
    (0x1D52B, 'M', u'n'),
    (0x1D52C, 'M', u'o'),
    (0x1D52D, 'M', u'p'),
    (0x1D52E, 'M', u'q'),
    (0x1D52F, 'M', u'r'),
    (0x1D530, 'M', u's'),
    (0x1D531, 'M', u't'),
    (0x1D532, 'M', u'u'),
    (0x1D533, 'M', u'v'),
    (0x1D534, 'M', u'w'),
    (0x1D535, 'M', u'x'),
    (0x1D536, 'M', u'y'),
    (0x1D537, 'M', u'z'),
    (0x1D538, 'M', u'a'),
    (0x1D539, 'M', u'b'),
    (0x1D53A, 'X'),
    (0x1D53B, 'M', u'd'),
    (0x1D53C, 'M', u'e'),
    (0x1D53D, 'M', u'f'),
    (0x1D53E, 'M', u'g'),
    (0x1D53F, 'X'),
    (0x1D540, 'M', u'i'),
    (0x1D541, 'M', u'j'),
    (0x1D542, 'M', u'k'),
    (0x1D543, 'M', u'l'),
    (0x1D544, 'M', u'm'),
    (0x1D545, 'X'),
    (0x1D546, 'M', u'o'),
    (0x1D547, 'X'),
    (0x1D54A, 'M', u's'),
    (0x1D54B, 'M', u't'),
    (0x1D54C, 'M', u'u'),
    (0x1D54D, 'M', u'v'),
    (0x1D54E, 'M', u'w'),
    (0x1D54F, 'M', u'x'),
    (0x1D550, 'M', u'y'),
    (0x1D551, 'X'),
    (0x1D552, 'M', u'a'),
    (0x1D553, 'M', u'b'),
    (0x1D554, 'M', u'c'),
    (0x1D555, 'M', u'd'),
    (0x1D556, 'M', u'e'),
    (0x1D557, 'M', u'f'),
    (0x1D558, 'M', u'g'),
    (0x1D559, 'M', u'h'),
    (0x1D55A, 'M', u'i'),
    (0x1D55B, 'M', u'j'),
    (0x1D55C, 'M', u'k'),
    (0x1D55D, 'M', u'l'),
    (0x1D55E, 'M', u'm'),
    (0x1D55F, 'M', u'n'),
    (0x1D560, 'M', u'o'),
    (0x1D561, 'M', u'p'),
    (0x1D562, 'M', u'q'),
    (0x1D563, 'M', u'r'),
    (0x1D564, 'M', u's'),
    (0x1D565, 'M', u't'),
    (0x1D566, 'M', u'u'),
    (0x1D567, 'M', u'v'),
    (0x1D568, 'M', u'w'),
    (0x1D569, 'M', u'x'),
    (0x1D56A, 'M', u'y'),
    (0x1D56B, 'M', u'z'),
    (0x1D56C, 'M', u'a'),
    (0x1D56D, 'M', u'b'),
    (0x1D56E, 'M', u'c'),
    (0x1D56F, 'M', u'd'),
    (0x1D570, 'M', u'e'),
    (0x1D571, 'M', u'f'),
    (0x1D572, 'M', u'g'),
    (0x1D573, 'M', u'h'),
    (0x1D574, 'M', u'i'),
    (0x1D575, 'M', u'j'),
    (0x1D576, 'M', u'k'),
    (0x1D577, 'M', u'l'),
    (0x1D578, 'M', u'm'),
    (0x1D579, 'M', u'n'),
    (0x1D57A, 'M', u'o'),
    (0x1D57B, 'M', u'p'),
    (0x1D57C, 'M', u'q'),
    (0x1D57D, 'M', u'r'),
    (0x1D57E, 'M', u's'),
    (0x1D57F, 'M', u't'),
    (0x1D580, 'M', u'u'),
    (0x1D581, 'M', u'v'),
    (0x1D582, 'M', u'w'),
    (0x1D583, 'M', u'x'),
    (0x1D584, 'M', u'y'),
    (0x1D585, 'M', u'z'),
    (0x1D586, 'M', u'a'),
    (0x1D587, 'M', u'b'),
    (0x1D588, 'M', u'c'),
    (0x1D589, 'M', u'd'),
    (0x1D58A, 'M', u'e'),
    (0x1D58B, 'M', u'f'),
    (0x1D58C, 'M', u'g'),
    (0x1D58D, 'M', u'h'),
    (0x1D58E, 'M', u'i'),
    ]

def _seg_62():
    return [
    (0x1D58F, 'M', u'j'),
    (0x1D590, 'M', u'k'),
    (0x1D591, 'M', u'l'),
    (0x1D592, 'M', u'm'),
    (0x1D593, 'M', u'n'),
    (0x1D594, 'M', u'o'),
    (0x1D595, 'M', u'p'),
    (0x1D596, 'M', u'q'),
    (0x1D597, 'M', u'r'),
    (0x1D598, 'M', u's'),
    (0x1D599, 'M', u't'),
    (0x1D59A, 'M', u'u'),
    (0x1D59B, 'M', u'v'),
    (0x1D59C, 'M', u'w'),
    (0x1D59D, 'M', u'x'),
    (0x1D59E, 'M', u'y'),
    (0x1D59F, 'M', u'z'),
    (0x1D5A0, 'M', u'a'),
    (0x1D5A1, 'M', u'b'),
    (0x1D5A2, 'M', u'c'),
    (0x1D5A3, 'M', u'd'),
    (0x1D5A4, 'M', u'e'),
    (0x1D5A5, 'M', u'f'),
    (0x1D5A6, 'M', u'g'),
    (0x1D5A7, 'M', u'h'),
    (0x1D5A8, 'M', u'i'),
    (0x1D5A9, 'M', u'j'),
    (0x1D5AA, 'M', u'k'),
    (0x1D5AB, 'M', u'l'),
    (0x1D5AC, 'M', u'm'),
    (0x1D5AD, 'M', u'n'),
    (0x1D5AE, 'M', u'o'),
    (0x1D5AF, 'M', u'p'),
    (0x1D5B0, 'M', u'q'),
    (0x1D5B1, 'M', u'r'),
    (0x1D5B2, 'M', u's'),
    (0x1D5B3, 'M', u't'),
    (0x1D5B4, 'M', u'u'),
    (0x1D5B5, 'M', u'v'),
    (0x1D5B6, 'M', u'w'),
    (0x1D5B7, 'M', u'x'),
    (0x1D5B8, 'M', u'y'),
    (0x1D5B9, 'M', u'z'),
    (0x1D5BA, 'M', u'a'),
    (0x1D5BB, 'M', u'b'),
    (0x1D5BC, 'M', u'c'),
    (0x1D5BD, 'M', u'd'),
    (0x1D5BE, 'M', u'e'),
    (0x1D5BF, 'M', u'f'),
    (0x1D5C0, 'M', u'g'),
    (0x1D5C1, 'M', u'h'),
    (0x1D5C2, 'M', u'i'),
    (0x1D5C3, 'M', u'j'),
    (0x1D5C4, 'M', u'k'),
    (0x1D5C5, 'M', u'l'),
    (0x1D5C6, 'M', u'm'),
    (0x1D5C7, 'M', u'n'),
    (0x1D5C8, 'M', u'o'),
    (0x1D5C9, 'M', u'p'),
    (0x1D5CA, 'M', u'q'),
    (0x1D5CB, 'M', u'r'),
    (0x1D5CC, 'M', u's'),
    (0x1D5CD, 'M', u't'),
    (0x1D5CE, 'M', u'u'),
    (0x1D5CF, 'M', u'v'),
    (0x1D5D0, 'M', u'w'),
    (0x1D5D1, 'M', u'x'),
    (0x1D5D2, 'M', u'y'),
    (0x1D5D3, 'M', u'z'),
    (0x1D5D4, 'M', u'a'),
    (0x1D5D5, 'M', u'b'),
    (0x1D5D6, 'M', u'c'),
    (0x1D5D7, 'M', u'd'),
    (0x1D5D8, 'M', u'e'),
    (0x1D5D9, 'M', u'f'),
    (0x1D5DA, 'M', u'g'),
    (0x1D5DB, 'M', u'h'),
    (0x1D5DC, 'M', u'i'),
    (0x1D5DD, 'M', u'j'),
    (0x1D5DE, 'M', u'k'),
    (0x1D5DF, 'M', u'l'),
    (0x1D5E0, 'M', u'm'),
    (0x1D5E1, 'M', u'n'),
    (0x1D5E2, 'M', u'o'),
    (0x1D5E3, 'M', u'p'),
    (0x1D5E4, 'M', u'q'),
    (0x1D5E5, 'M', u'r'),
    (0x1D5E6, 'M', u's'),
    (0x1D5E7, 'M', u't'),
    (0x1D5E8, 'M', u'u'),
    (0x1D5E9, 'M', u'v'),
    (0x1D5EA, 'M', u'w'),
    (0x1D5EB, 'M', u'x'),
    (0x1D5EC, 'M', u'y'),
    (0x1D5ED, 'M', u'z'),
    (0x1D5EE, 'M', u'a'),
    (0x1D5EF, 'M', u'b'),
    (0x1D5F0, 'M', u'c'),
    (0x1D5F1, 'M', u'd'),
    (0x1D5F2, 'M', u'e'),
    ]

def _seg_63():
    return [
    (0x1D5F3, 'M', u'f'),
    (0x1D5F4, 'M', u'g'),
    (0x1D5F5, 'M', u'h'),
    (0x1D5F6, 'M', u'i'),
    (0x1D5F7, 'M', u'j'),
    (0x1D5F8, 'M', u'k'),
    (0x1D5F9, 'M', u'l'),
    (0x1D5FA, 'M', u'm'),
    (0x1D5FB, 'M', u'n'),
    (0x1D5FC, 'M', u'o'),
    (0x1D5FD, 'M', u'p'),
    (0x1D5FE, 'M', u'q'),
    (0x1D5FF, 'M', u'r'),
    (0x1D600, 'M', u's'),
    (0x1D601, 'M', u't'),
    (0x1D602, 'M', u'u'),
    (0x1D603, 'M', u'v'),
    (0x1D604, 'M', u'w'),
    (0x1D605, 'M', u'x'),
    (0x1D606, 'M', u'y'),
    (0x1D607, 'M', u'z'),
    (0x1D608, 'M', u'a'),
    (0x1D609, 'M', u'b'),
    (0x1D60A, 'M', u'c'),
    (0x1D60B, 'M', u'd'),
    (0x1D60C, 'M', u'e'),
    (0x1D60D, 'M', u'f'),
    (0x1D60E, 'M', u'g'),
    (0x1D60F, 'M', u'h'),
    (0x1D610, 'M', u'i'),
    (0x1D611, 'M', u'j'),
    (0x1D612, 'M', u'k'),
    (0x1D613, 'M', u'l'),
    (0x1D614, 'M', u'm'),
    (0x1D615, 'M', u'n'),
    (0x1D616, 'M', u'o'),
    (0x1D617, 'M', u'p'),
    (0x1D618, 'M', u'q'),
    (0x1D619, 'M', u'r'),
    (0x1D61A, 'M', u's'),
    (0x1D61B, 'M', u't'),
    (0x1D61C, 'M', u'u'),
    (0x1D61D, 'M', u'v'),
    (0x1D61E, 'M', u'w'),
    (0x1D61F, 'M', u'x'),
    (0x1D620, 'M', u'y'),
    (0x1D621, 'M', u'z'),
    (0x1D622, 'M', u'a'),
    (0x1D623, 'M', u'b'),
    (0x1D624, 'M', u'c'),
    (0x1D625, 'M', u'd'),
    (0x1D626, 'M', u'e'),
    (0x1D627, 'M', u'f'),
    (0x1D628, 'M', u'g'),
    (0x1D629, 'M', u'h'),
    (0x1D62A, 'M', u'i'),
    (0x1D62B, 'M', u'j'),
    (0x1D62C, 'M', u'k'),
    (0x1D62D, 'M', u'l'),
    (0x1D62E, 'M', u'm'),
    (0x1D62F, 'M', u'n'),
    (0x1D630, 'M', u'o'),
    (0x1D631, 'M', u'p'),
    (0x1D632, 'M', u'q'),
    (0x1D633, 'M', u'r'),
    (0x1D634, 'M', u's'),
    (0x1D635, 'M', u't'),
    (0x1D636, 'M', u'u'),
    (0x1D637, 'M', u'v'),
    (0x1D638, 'M', u'w'),
    (0x1D639, 'M', u'x'),
    (0x1D63A, 'M', u'y'),
    (0x1D63B, 'M', u'z'),
    (0x1D63C, 'M', u'a'),
    (0x1D63D, 'M', u'b'),
    (0x1D63E, 'M', u'c'),
    (0x1D63F, 'M', u'd'),
    (0x1D640, 'M', u'e'),
    (0x1D641, 'M', u'f'),
    (0x1D642, 'M', u'g'),
    (0x1D643, 'M', u'h'),
    (0x1D644, 'M', u'i'),
    (0x1D645, 'M', u'j'),
    (0x1D646, 'M', u'k'),
    (0x1D647, 'M', u'l'),
    (0x1D648, 'M', u'm'),
    (0x1D649, 'M', u'n'),
    (0x1D64A, 'M', u'o'),
    (0x1D64B, 'M', u'p'),
    (0x1D64C, 'M', u'q'),
    (0x1D64D, 'M', u'r'),
    (0x1D64E, 'M', u's'),
    (0x1D64F, 'M', u't'),
    (0x1D650, 'M', u'u'),
    (0x1D651, 'M', u'v'),
    (0x1D652, 'M', u'w'),
    (0x1D653, 'M', u'x'),
    (0x1D654, 'M', u'y'),
    (0x1D655, 'M', u'z'),
    (0x1D656, 'M', u'a'),
    ]

def _seg_64():
    return [
    (0x1D657, 'M', u'b'),
    (0x1D658, 'M', u'c'),
    (0x1D659, 'M', u'd'),
    (0x1D65A, 'M', u'e'),
    (0x1D65B, 'M', u'f'),
    (0x1D65C, 'M', u'g'),
    (0x1D65D, 'M', u'h'),
    (0x1D65E, 'M', u'i'),
    (0x1D65F, 'M', u'j'),
    (0x1D660, 'M', u'k'),
    (0x1D661, 'M', u'l'),
    (0x1D662, 'M', u'm'),
    (0x1D663, 'M', u'n'),
    (0x1D664, 'M', u'o'),
    (0x1D665, 'M', u'p'),
    (0x1D666, 'M', u'q'),
    (0x1D667, 'M', u'r'),
    (0x1D668, 'M', u's'),
    (0x1D669, 'M', u't'),
    (0x1D66A, 'M', u'u'),
    (0x1D66B, 'M', u'v'),
    (0x1D66C, 'M', u'w'),
    (0x1D66D, 'M', u'x'),
    (0x1D66E, 'M', u'y'),
    (0x1D66F, 'M', u'z'),
    (0x1D670, 'M', u'a'),
    (0x1D671, 'M', u'b'),
    (0x1D672, 'M', u'c'),
    (0x1D673, 'M', u'd'),
    (0x1D674, 'M', u'e'),
    (0x1D675, 'M', u'f'),
    (0x1D676, 'M', u'g'),
    (0x1D677, 'M', u'h'),
    (0x1D678, 'M', u'i'),
    (0x1D679, 'M', u'j'),
    (0x1D67A, 'M', u'k'),
    (0x1D67B, 'M', u'l'),
    (0x1D67C, 'M', u'm'),
    (0x1D67D, 'M', u'n'),
    (0x1D67E, 'M', u'o'),
    (0x1D67F, 'M', u'p'),
    (0x1D680, 'M', u'q'),
    (0x1D681, 'M', u'r'),
    (0x1D682, 'M', u's'),
    (0x1D683, 'M', u't'),
    (0x1D684, 'M', u'u'),
    (0x1D685, 'M', u'v'),
    (0x1D686, 'M', u'w'),
    (0x1D687, 'M', u'x'),
    (0x1D688, 'M', u'y'),
    (0x1D689, 'M', u'z'),
    (0x1D68A, 'M', u'a'),
    (0x1D68B, 'M', u'b'),
    (0x1D68C, 'M', u'c'),
    (0x1D68D, 'M', u'd'),
    (0x1D68E, 'M', u'e'),
    (0x1D68F, 'M', u'f'),
    (0x1D690, 'M', u'g'),
    (0x1D691, 'M', u'h'),
    (0x1D692, 'M', u'i'),
    (0x1D693, 'M', u'j'),
    (0x1D694, 'M', u'k'),
    (0x1D695, 'M', u'l'),
    (0x1D696, 'M', u'm'),
    (0x1D697, 'M', u'n'),
    (0x1D698, 'M', u'o'),
    (0x1D699, 'M', u'p'),
    (0x1D69A, 'M', u'q'),
    (0x1D69B, 'M', u'r'),
    (0x1D69C, 'M', u's'),
    (0x1D69D, 'M', u't'),
    (0x1D69E, 'M', u'u'),
    (0x1D69F, 'M', u'v'),
    (0x1D6A0, 'M', u'w'),
    (0x1D6A1, 'M', u'x'),
    (0x1D6A2, 'M', u'y'),
    (0x1D6A3, 'M', u'z'),
    (0x1D6A4, 'M', u'ı'),
    (0x1D6A5, 'M', u'ȷ'),
    (0x1D6A6, 'X'),
    (0x1D6A8, 'M', u'α'),
    (0x1D6A9, 'M', u'β'),
    (0x1D6AA, 'M', u'γ'),
    (0x1D6AB, 'M', u'δ'),
    (0x1D6AC, 'M', u'ε'),
    (0x1D6AD, 'M', u'ζ'),
    (0x1D6AE, 'M', u'η'),
    (0x1D6AF, 'M', u'θ'),
    (0x1D6B0, 'M', u'ι'),
    (0x1D6B1, 'M', u'κ'),
    (0x1D6B2, 'M', u'λ'),
    (0x1D6B3, 'M', u'μ'),
    (0x1D6B4, 'M', u'ν'),
    (0x1D6B5, 'M', u'ξ'),
    (0x1D6B6, 'M', u'ο'),
    (0x1D6B7, 'M', u'π'),
    (0x1D6B8, 'M', u'ρ'),
    (0x1D6B9, 'M', u'θ'),
    (0x1D6BA, 'M', u'σ'),
    (0x1D6BB, 'M', u'τ'),
    ]

def _seg_65():
    return [
    (0x1D6BC, 'M', u'υ'),
    (0x1D6BD, 'M', u'φ'),
    (0x1D6BE, 'M', u'χ'),
    (0x1D6BF, 'M', u'ψ'),
    (0x1D6C0, 'M', u'ω'),
    (0x1D6C1, 'M', u'∇'),
    (0x1D6C2, 'M', u'α'),
    (0x1D6C3, 'M', u'β'),
    (0x1D6C4, 'M', u'γ'),
    (0x1D6C5, 'M', u'δ'),
    (0x1D6C6, 'M', u'ε'),
    (0x1D6C7, 'M', u'ζ'),
    (0x1D6C8, 'M', u'η'),
    (0x1D6C9, 'M', u'θ'),
    (0x1D6CA, 'M', u'ι'),
    (0x1D6CB, 'M', u'κ'),
    (0x1D6CC, 'M', u'λ'),
    (0x1D6CD, 'M', u'μ'),
    (0x1D6CE, 'M', u'ν'),
    (0x1D6CF, 'M', u'ξ'),
    (0x1D6D0, 'M', u'ο'),
    (0x1D6D1, 'M', u'π'),
    (0x1D6D2, 'M', u'ρ'),
    (0x1D6D3, 'M', u'σ'),
    (0x1D6D5, 'M', u'τ'),
    (0x1D6D6, 'M', u'υ'),
    (0x1D6D7, 'M', u'φ'),
    (0x1D6D8, 'M', u'χ'),
    (0x1D6D9, 'M', u'ψ'),
    (0x1D6DA, 'M', u'ω'),
    (0x1D6DB, 'M', u'∂'),
    (0x1D6DC, 'M', u'ε'),
    (0x1D6DD, 'M', u'θ'),
    (0x1D6DE, 'M', u'κ'),
    (0x1D6DF, 'M', u'φ'),
    (0x1D6E0, 'M', u'ρ'),
    (0x1D6E1, 'M', u'π'),
    (0x1D6E2, 'M', u'α'),
    (0x1D6E3, 'M', u'β'),
    (0x1D6E4, 'M', u'γ'),
    (0x1D6E5, 'M', u'δ'),
    (0x1D6E6, 'M', u'ε'),
    (0x1D6E7, 'M', u'ζ'),
    (0x1D6E8, 'M', u'η'),
    (0x1D6E9, 'M', u'θ'),
    (0x1D6EA, 'M', u'ι'),
    (0x1D6EB, 'M', u'κ'),
    (0x1D6EC, 'M', u'λ'),
    (0x1D6ED, 'M', u'μ'),
    (0x1D6EE, 'M', u'ν'),
    (0x1D6EF, 'M', u'ξ'),
    (0x1D6F0, 'M', u'ο'),
    (0x1D6F1, 'M', u'π'),
    (0x1D6F2, 'M', u'ρ'),
    (0x1D6F3, 'M', u'θ'),
    (0x1D6F4, 'M', u'σ'),
    (0x1D6F5, 'M', u'τ'),
    (0x1D6F6, 'M', u'υ'),
    (0x1D6F7, 'M', u'φ'),
    (0x1D6F8, 'M', u'χ'),
    (0x1D6F9, 'M', u'ψ'),
    (0x1D6FA, 'M', u'ω'),
    (0x1D6FB, 'M', u'∇'),
    (0x1D6FC, 'M', u'α'),
    (0x1D6FD, 'M', u'β'),
    (0x1D6FE, 'M', u'γ'),
    (0x1D6FF, 'M', u'δ'),
    (0x1D700, 'M', u'ε'),
    (0x1D701, 'M', u'ζ'),
    (0x1D702, 'M', u'η'),
    (0x1D703, 'M', u'θ'),
    (0x1D704, 'M', u'ι'),
    (0x1D705, 'M', u'κ'),
    (0x1D706, 'M', u'λ'),
    (0x1D707, 'M', u'μ'),
    (0x1D708, 'M', u'ν'),
    (0x1D709, 'M', u'ξ'),
    (0x1D70A, 'M', u'ο'),
    (0x1D70B, 'M', u'π'),
    (0x1D70C, 'M', u'ρ'),
    (0x1D70D, 'M', u'σ'),
    (0x1D70F, 'M', u'τ'),
    (0x1D710, 'M', u'υ'),
    (0x1D711, 'M', u'φ'),
    (0x1D712, 'M', u'χ'),
    (0x1D713, 'M', u'ψ'),
    (0x1D714, 'M', u'ω'),
    (0x1D715, 'M', u'∂'),
    (0x1D716, 'M', u'ε'),
    (0x1D717, 'M', u'θ'),
    (0x1D718, 'M', u'κ'),
    (0x1D719, 'M', u'φ'),
    (0x1D71A, 'M', u'ρ'),
    (0x1D71B, 'M', u'π'),
    (0x1D71C, 'M', u'α'),
    (0x1D71D, 'M', u'β'),
    (0x1D71E, 'M', u'γ'),
    (0x1D71F, 'M', u'δ'),
    (0x1D720, 'M', u'ε'),
    (0x1D721, 'M', u'ζ'),
    ]

def _seg_66():
    return [
    (0x1D722, 'M', u'η'),
    (0x1D723, 'M', u'θ'),
    (0x1D724, 'M', u'ι'),
    (0x1D725, 'M', u'κ'),
    (0x1D726, 'M', u'λ'),
    (0x1D727, 'M', u'μ'),
    (0x1D728, 'M', u'ν'),
    (0x1D729, 'M', u'ξ'),
    (0x1D72A, 'M', u'ο'),
    (0x1D72B, 'M', u'π'),
    (0x1D72C, 'M', u'ρ'),
    (0x1D72D, 'M', u'θ'),
    (0x1D72E, 'M', u'σ'),
    (0x1D72F, 'M', u'τ'),
    (0x1D730, 'M', u'υ'),
    (0x1D731, 'M', u'φ'),
    (0x1D732, 'M', u'χ'),
    (0x1D733, 'M', u'ψ'),
    (0x1D734, 'M', u'ω'),
    (0x1D735, 'M', u'∇'),
    (0x1D736, 'M', u'α'),
    (0x1D737, 'M', u'β'),
    (0x1D738, 'M', u'γ'),
    (0x1D739, 'M', u'δ'),
    (0x1D73A, 'M', u'ε'),
    (0x1D73B, 'M', u'ζ'),
    (0x1D73C, 'M', u'η'),
    (0x1D73D, 'M', u'θ'),
    (0x1D73E, 'M', u'ι'),
    (0x1D73F, 'M', u'κ'),
    (0x1D740, 'M', u'λ'),
    (0x1D741, 'M', u'μ'),
    (0x1D742, 'M', u'ν'),
    (0x1D743, 'M', u'ξ'),
    (0x1D744, 'M', u'ο'),
    (0x1D745, 'M', u'π'),
    (0x1D746, 'M', u'ρ'),
    (0x1D747, 'M', u'σ'),
    (0x1D749, 'M', u'τ'),
    (0x1D74A, 'M', u'υ'),
    (0x1D74B, 'M', u'φ'),
    (0x1D74C, 'M', u'χ'),
    (0x1D74D, 'M', u'ψ'),
    (0x1D74E, 'M', u'ω'),
    (0x1D74F, 'M', u'∂'),
    (0x1D750, 'M', u'ε'),
    (0x1D751, 'M', u'θ'),
    (0x1D752, 'M', u'κ'),
    (0x1D753, 'M', u'φ'),
    (0x1D754, 'M', u'ρ'),
    (0x1D755, 'M', u'π'),
    (0x1D756, 'M', u'α'),
    (0x1D757, 'M', u'β'),
    (0x1D758, 'M', u'γ'),
    (0x1D759, 'M', u'δ'),
    (0x1D75A, 'M', u'ε'),
    (0x1D75B, 'M', u'ζ'),
    (0x1D75C, 'M', u'η'),
    (0x1D75D, 'M', u'θ'),
    (0x1D75E, 'M', u'ι'),
    (0x1D75F, 'M', u'κ'),
    (0x1D760, 'M', u'λ'),
    (0x1D761, 'M', u'μ'),
    (0x1D762, 'M', u'ν'),
    (0x1D763, 'M', u'ξ'),
    (0x1D764, 'M', u'ο'),
    (0x1D765, 'M', u'π'),
    (0x1D766, 'M', u'ρ'),
    (0x1D767, 'M', u'θ'),
    (0x1D768, 'M', u'σ'),
    (0x1D769, 'M', u'τ'),
    (0x1D76A, 'M', u'υ'),
    (0x1D76B, 'M', u'φ'),
    (0x1D76C, 'M', u'χ'),
    (0x1D76D, 'M', u'ψ'),
    (0x1D76E, 'M', u'ω'),
    (0x1D76F, 'M', u'∇'),
    (0x1D770, 'M', u'α'),
    (0x1D771, 'M', u'β'),
    (0x1D772, 'M', u'γ'),
    (0x1D773, 'M', u'δ'),
    (0x1D774, 'M', u'ε'),
    (0x1D775, 'M', u'ζ'),
    (0x1D776, 'M', u'η'),
    (0x1D777, 'M', u'θ'),
    (0x1D778, 'M', u'ι'),
    (0x1D779, 'M', u'κ'),
    (0x1D77A, 'M', u'λ'),
    (0x1D77B, 'M', u'μ'),
    (0x1D77C, 'M', u'ν'),
    (0x1D77D, 'M', u'ξ'),
    (0x1D77E, 'M', u'ο'),
    (0x1D77F, 'M', u'π'),
    (0x1D780, 'M', u'ρ'),
    (0x1D781, 'M', u'σ'),
    (0x1D783, 'M', u'τ'),
    (0x1D784, 'M', u'υ'),
    (0x1D785, 'M', u'φ'),
    (0x1D786, 'M', u'χ'),
    (0x1D787, 'M', u'ψ'),
    ]

def _seg_67():
    return [
    (0x1D788, 'M', u'ω'),
    (0x1D789, 'M', u'∂'),
    (0x1D78A, 'M', u'ε'),
    (0x1D78B, 'M', u'θ'),
    (0x1D78C, 'M', u'κ'),
    (0x1D78D, 'M', u'φ'),
    (0x1D78E, 'M', u'ρ'),
    (0x1D78F, 'M', u'π'),
    (0x1D790, 'M', u'α'),
    (0x1D791, 'M', u'β'),
    (0x1D792, 'M', u'γ'),
    (0x1D793, 'M', u'δ'),
    (0x1D794, 'M', u'ε'),
    (0x1D795, 'M', u'ζ'),
    (0x1D796, 'M', u'η'),
    (0x1D797, 'M', u'θ'),
    (0x1D798, 'M', u'ι'),
    (0x1D799, 'M', u'κ'),
    (0x1D79A, 'M', u'λ'),
    (0x1D79B, 'M', u'μ'),
    (0x1D79C, 'M', u'ν'),
    (0x1D79D, 'M', u'ξ'),
    (0x1D79E, 'M', u'ο'),
    (0x1D79F, 'M', u'π'),
    (0x1D7A0, 'M', u'ρ'),
    (0x1D7A1, 'M', u'θ'),
    (0x1D7A2, 'M', u'σ'),
    (0x1D7A3, 'M', u'τ'),
    (0x1D7A4, 'M', u'υ'),
    (0x1D7A5, 'M', u'φ'),
    (0x1D7A6, 'M', u'χ'),
    (0x1D7A7, 'M', u'ψ'),
    (0x1D7A8, 'M', u'ω'),
    (0x1D7A9, 'M', u'∇'),
    (0x1D7AA, 'M', u'α'),
    (0x1D7AB, 'M', u'β'),
    (0x1D7AC, 'M', u'γ'),
    (0x1D7AD, 'M', u'δ'),
    (0x1D7AE, 'M', u'ε'),
    (0x1D7AF, 'M', u'ζ'),
    (0x1D7B0, 'M', u'η'),
    (0x1D7B1, 'M', u'θ'),
    (0x1D7B2, 'M', u'ι'),
    (0x1D7B3, 'M', u'κ'),
    (0x1D7B4, 'M', u'λ'),
    (0x1D7B5, 'M', u'μ'),
    (0x1D7B6, 'M', u'ν'),
    (0x1D7B7, 'M', u'ξ'),
    (0x1D7B8, 'M', u'ο'),
    (0x1D7B9, 'M', u'π'),
    (0x1D7BA, 'M', u'ρ'),
    (0x1D7BB, 'M', u'σ'),
    (0x1D7BD, 'M', u'τ'),
    (0x1D7BE, 'M', u'υ'),
    (0x1D7BF, 'M', u'φ'),
    (0x1D7C0, 'M', u'χ'),
    (0x1D7C1, 'M', u'ψ'),
    (0x1D7C2, 'M', u'ω'),
    (0x1D7C3, 'M', u'∂'),
    (0x1D7C4, 'M', u'ε'),
    (0x1D7C5, 'M', u'θ'),
    (0x1D7C6, 'M', u'κ'),
    (0x1D7C7, 'M', u'φ'),
    (0x1D7C8, 'M', u'ρ'),
    (0x1D7C9, 'M', u'π'),
    (0x1D7CA, 'M', u'ϝ'),
    (0x1D7CC, 'X'),
    (0x1D7CE, 'M', u'0'),
    (0x1D7CF, 'M', u'1'),
    (0x1D7D0, 'M', u'2'),
    (0x1D7D1, 'M', u'3'),
    (0x1D7D2, 'M', u'4'),
    (0x1D7D3, 'M', u'5'),
    (0x1D7D4, 'M', u'6'),
    (0x1D7D5, 'M', u'7'),
    (0x1D7D6, 'M', u'8'),
    (0x1D7D7, 'M', u'9'),
    (0x1D7D8, 'M', u'0'),
    (0x1D7D9, 'M', u'1'),
    (0x1D7DA, 'M', u'2'),
    (0x1D7DB, 'M', u'3'),
    (0x1D7DC, 'M', u'4'),
    (0x1D7DD, 'M', u'5'),
    (0x1D7DE, 'M', u'6'),
    (0x1D7DF, 'M', u'7'),
    (0x1D7E0, 'M', u'8'),
    (0x1D7E1, 'M', u'9'),
    (0x1D7E2, 'M', u'0'),
    (0x1D7E3, 'M', u'1'),
    (0x1D7E4, 'M', u'2'),
    (0x1D7E5, 'M', u'3'),
    (0x1D7E6, 'M', u'4'),
    (0x1D7E7, 'M', u'5'),
    (0x1D7E8, 'M', u'6'),
    (0x1D7E9, 'M', u'7'),
    (0x1D7EA, 'M', u'8'),
    (0x1D7EB, 'M', u'9'),
    (0x1D7EC, 'M', u'0'),
    (0x1D7ED, 'M', u'1'),
    (0x1D7EE, 'M', u'2'),
    ]

def _seg_68():
    return [
    (0x1D7EF, 'M', u'3'),
    (0x1D7F0, 'M', u'4'),
    (0x1D7F1, 'M', u'5'),
    (0x1D7F2, 'M', u'6'),
    (0x1D7F3, 'M', u'7'),
    (0x1D7F4, 'M', u'8'),
    (0x1D7F5, 'M', u'9'),
    (0x1D7F6, 'M', u'0'),
    (0x1D7F7, 'M', u'1'),
    (0x1D7F8, 'M', u'2'),
    (0x1D7F9, 'M', u'3'),
    (0x1D7FA, 'M', u'4'),
    (0x1D7FB, 'M', u'5'),
    (0x1D7FC, 'M', u'6'),
    (0x1D7FD, 'M', u'7'),
    (0x1D7FE, 'M', u'8'),
    (0x1D7FF, 'M', u'9'),
    (0x1D800, 'V'),
    (0x1DA8C, 'X'),
    (0x1DA9B, 'V'),
    (0x1DAA0, 'X'),
    (0x1DAA1, 'V'),
    (0x1DAB0, 'X'),
    (0x1E000, 'V'),
    (0x1E007, 'X'),
    (0x1E008, 'V'),
    (0x1E019, 'X'),
    (0x1E01B, 'V'),
    (0x1E022, 'X'),
    (0x1E023, 'V'),
    (0x1E025, 'X'),
    (0x1E026, 'V'),
    (0x1E02B, 'X'),
    (0x1E800, 'V'),
    (0x1E8C5, 'X'),
    (0x1E8C7, 'V'),
    (0x1E8D7, 'X'),
    (0x1E900, 'M', u'𞤢'),
    (0x1E901, 'M', u'𞤣'),
    (0x1E902, 'M', u'𞤤'),
    (0x1E903, 'M', u'𞤥'),
    (0x1E904, 'M', u'𞤦'),
    (0x1E905, 'M', u'𞤧'),
    (0x1E906, 'M', u'𞤨'),
    (0x1E907, 'M', u'𞤩'),
    (0x1E908, 'M', u'𞤪'),
    (0x1E909, 'M', u'𞤫'),
    (0x1E90A, 'M', u'𞤬'),
    (0x1E90B, 'M', u'𞤭'),
    (0x1E90C, 'M', u'𞤮'),
    (0x1E90D, 'M', u'𞤯'),
    (0x1E90E, 'M', u'𞤰'),
    (0x1E90F, 'M', u'𞤱'),
    (0x1E910, 'M', u'𞤲'),
    (0x1E911, 'M', u'𞤳'),
    (0x1E912, 'M', u'𞤴'),
    (0x1E913, 'M', u'𞤵'),
    (0x1E914, 'M', u'𞤶'),
    (0x1E915, 'M', u'𞤷'),
    (0x1E916, 'M', u'𞤸'),
    (0x1E917, 'M', u'𞤹'),
    (0x1E918, 'M', u'𞤺'),
    (0x1E919, 'M', u'𞤻'),
    (0x1E91A, 'M', u'𞤼'),
    (0x1E91B, 'M', u'𞤽'),
    (0x1E91C, 'M', u'𞤾'),
    (0x1E91D, 'M', u'𞤿'),
    (0x1E91E, 'M', u'𞥀'),
    (0x1E91F, 'M', u'𞥁'),
    (0x1E920, 'M', u'𞥂'),
    (0x1E921, 'M', u'𞥃'),
    (0x1E922, 'V'),
    (0x1E94B, 'X'),
    (0x1E950, 'V'),
    (0x1E95A, 'X'),
    (0x1E95E, 'V'),
    (0x1E960, 'X'),
    (0x1EC71, 'V'),
    (0x1ECB5, 'X'),
    (0x1EE00, 'M', u'ا'),
    (0x1EE01, 'M', u'ب'),
    (0x1EE02, 'M', u'ج'),
    (0x1EE03, 'M', u'د'),
    (0x1EE04, 'X'),
    (0x1EE05, 'M', u'و'),
    (0x1EE06, 'M', u'ز'),
    (0x1EE07, 'M', u'ح'),
    (0x1EE08, 'M', u'ط'),
    (0x1EE09, 'M', u'ي'),
    (0x1EE0A, 'M', u'ك'),
    (0x1EE0B, 'M', u'ل'),
    (0x1EE0C, 'M', u'م'),
    (0x1EE0D, 'M', u'ن'),
    (0x1EE0E, 'M', u'س'),
    (0x1EE0F, 'M', u'ع'),
    (0x1EE10, 'M', u'ف'),
    (0x1EE11, 'M', u'ص'),
    (0x1EE12, 'M', u'ق'),
    (0x1EE13, 'M', u'ر'),
    (0x1EE14, 'M', u'ش'),
    ]

def _seg_69():
    return [
    (0x1EE15, 'M', u'ت'),
    (0x1EE16, 'M', u'ث'),
    (0x1EE17, 'M', u'خ'),
    (0x1EE18, 'M', u'ذ'),
    (0x1EE19, 'M', u'ض'),
    (0x1EE1A, 'M', u'ظ'),
    (0x1EE1B, 'M', u'غ'),
    (0x1EE1C, 'M', u'ٮ'),
    (0x1EE1D, 'M', u'ں'),
    (0x1EE1E, 'M', u'ڡ'),
    (0x1EE1F, 'M', u'ٯ'),
    (0x1EE20, 'X'),
    (0x1EE21, 'M', u'ب'),
    (0x1EE22, 'M', u'ج'),
    (0x1EE23, 'X'),
    (0x1EE24, 'M', u'ه'),
    (0x1EE25, 'X'),
    (0x1EE27, 'M', u'ح'),
    (0x1EE28, 'X'),
    (0x1EE29, 'M', u'ي'),
    (0x1EE2A, 'M', u'ك'),
    (0x1EE2B, 'M', u'ل'),
    (0x1EE2C, 'M', u'م'),
    (0x1EE2D, 'M', u'ن'),
    (0x1EE2E, 'M', u'س'),
    (0x1EE2F, 'M', u'ع'),
    (0x1EE30, 'M', u'ف'),
    (0x1EE31, 'M', u'ص'),
    (0x1EE32, 'M', u'ق'),
    (0x1EE33, 'X'),
    (0x1EE34, 'M', u'ش'),
    (0x1EE35, 'M', u'ت'),
    (0x1EE36, 'M', u'ث'),
    (0x1EE37, 'M', u'خ'),
    (0x1EE38, 'X'),
    (0x1EE39, 'M', u'ض'),
    (0x1EE3A, 'X'),
    (0x1EE3B, 'M', u'غ'),
    (0x1EE3C, 'X'),
    (0x1EE42, 'M', u'ج'),
    (0x1EE43, 'X'),
    (0x1EE47, 'M', u'ح'),
    (0x1EE48, 'X'),
    (0x1EE49, 'M', u'ي'),
    (0x1EE4A, 'X'),
    (0x1EE4B, 'M', u'ل'),
    (0x1EE4C, 'X'),
    (0x1EE4D, 'M', u'ن'),
    (0x1EE4E, 'M', u'س'),
    (0x1EE4F, 'M', u'ع'),
    (0x1EE50, 'X'),
    (0x1EE51, 'M', u'ص'),
    (0x1EE52, 'M', u'ق'),
    (0x1EE53, 'X'),
    (0x1EE54, 'M', u'ش'),
    (0x1EE55, 'X'),
    (0x1EE57, 'M', u'خ'),
    (0x1EE58, 'X'),
    (0x1EE59, 'M', u'ض'),
    (0x1EE5A, 'X'),
    (0x1EE5B, 'M', u'غ'),
    (0x1EE5C, 'X'),
    (0x1EE5D, 'M', u'ں'),
    (0x1EE5E, 'X'),
    (0x1EE5F, 'M', u'ٯ'),
    (0x1EE60, 'X'),
    (0x1EE61, 'M', u'ب'),
    (0x1EE62, 'M', u'ج'),
    (0x1EE63, 'X'),
    (0x1EE64, 'M', u'ه'),
    (0x1EE65, 'X'),
    (0x1EE67, 'M', u'ح'),
    (0x1EE68, 'M', u'ط'),
    (0x1EE69, 'M', u'ي'),
    (0x1EE6A, 'M', u'ك'),
    (0x1EE6B, 'X'),
    (0x1EE6C, 'M', u'م'),
    (0x1EE6D, 'M', u'ن'),
    (0x1EE6E, 'M', u'س'),
    (0x1EE6F, 'M', u'ع'),
    (0x1EE70, 'M', u'ف'),
    (0x1EE71, 'M', u'ص'),
    (0x1EE72, 'M', u'ق'),
    (0x1EE73, 'X'),
    (0x1EE74, 'M', u'ش'),
    (0x1EE75, 'M', u'ت'),
    (0x1EE76, 'M', u'ث'),
    (0x1EE77, 'M', u'خ'),
    (0x1EE78, 'X'),
    (0x1EE79, 'M', u'ض'),
    (0x1EE7A, 'M', u'ظ'),
    (0x1EE7B, 'M', u'غ'),
    (0x1EE7C, 'M', u'ٮ'),
    (0x1EE7D, 'X'),
    (0x1EE7E, 'M', u'ڡ'),
    (0x1EE7F, 'X'),
    (0x1EE80, 'M', u'ا'),
    (0x1EE81, 'M', u'ب'),
    (0x1EE82, 'M', u'ج'),
    (0x1EE83, 'M', u'د'),
    ]

def _seg_70():
    return [
    (0x1EE84, 'M', u'ه'),
    (0x1EE85, 'M', u'و'),
    (0x1EE86, 'M', u'ز'),
    (0x1EE87, 'M', u'ح'),
    (0x1EE88, 'M', u'ط'),
    (0x1EE89, 'M', u'ي'),
    (0x1EE8A, 'X'),
    (0x1EE8B, 'M', u'ل'),
    (0x1EE8C, 'M', u'م'),
    (0x1EE8D, 'M', u'ن'),
    (0x1EE8E, 'M', u'س'),
    (0x1EE8F, 'M', u'ع'),
    (0x1EE90, 'M', u'ف'),
    (0x1EE91, 'M', u'ص'),
    (0x1EE92, 'M', u'ق'),
    (0x1EE93, 'M', u'ر'),
    (0x1EE94, 'M', u'ش'),
    (0x1EE95, 'M', u'ت'),
    (0x1EE96, 'M', u'ث'),
    (0x1EE97, 'M', u'خ'),
    (0x1EE98, 'M', u'ذ'),
    (0x1EE99, 'M', u'ض'),
    (0x1EE9A, 'M', u'ظ'),
    (0x1EE9B, 'M', u'غ'),
    (0x1EE9C, 'X'),
    (0x1EEA1, 'M', u'ب'),
    (0x1EEA2, 'M', u'ج'),
    (0x1EEA3, 'M', u'د'),
    (0x1EEA4, 'X'),
    (0x1EEA5, 'M', u'و'),
    (0x1EEA6, 'M', u'ز'),
    (0x1EEA7, 'M', u'ح'),
    (0x1EEA8, 'M', u'ط'),
    (0x1EEA9, 'M', u'ي'),
    (0x1EEAA, 'X'),
    (0x1EEAB, 'M', u'ل'),
    (0x1EEAC, 'M', u'م'),
    (0x1EEAD, 'M', u'ن'),
    (0x1EEAE, 'M', u'س'),
    (0x1EEAF, 'M', u'ع'),
    (0x1EEB0, 'M', u'ف'),
    (0x1EEB1, 'M', u'ص'),
    (0x1EEB2, 'M', u'ق'),
    (0x1EEB3, 'M', u'ر'),
    (0x1EEB4, 'M', u'ش'),
    (0x1EEB5, 'M', u'ت'),
    (0x1EEB6, 'M', u'ث'),
    (0x1EEB7, 'M', u'خ'),
    (0x1EEB8, 'M', u'ذ'),
    (0x1EEB9, 'M', u'ض'),
    (0x1EEBA, 'M', u'ظ'),
    (0x1EEBB, 'M', u'غ'),
    (0x1EEBC, 'X'),
    (0x1EEF0, 'V'),
    (0x1EEF2, 'X'),
    (0x1F000, 'V'),
    (0x1F02C, 'X'),
    (0x1F030, 'V'),
    (0x1F094, 'X'),
    (0x1F0A0, 'V'),
    (0x1F0AF, 'X'),
    (0x1F0B1, 'V'),
    (0x1F0C0, 'X'),
    (0x1F0C1, 'V'),
    (0x1F0D0, 'X'),
    (0x1F0D1, 'V'),
    (0x1F0F6, 'X'),
    (0x1F101, '3', u'0,'),
    (0x1F102, '3', u'1,'),
    (0x1F103, '3', u'2,'),
    (0x1F104, '3', u'3,'),
    (0x1F105, '3', u'4,'),
    (0x1F106, '3', u'5,'),
    (0x1F107, '3', u'6,'),
    (0x1F108, '3', u'7,'),
    (0x1F109, '3', u'8,'),
    (0x1F10A, '3', u'9,'),
    (0x1F10B, 'V'),
    (0x1F10D, 'X'),
    (0x1F110, '3', u'(a)'),
    (0x1F111, '3', u'(b)'),
    (0x1F112, '3', u'(c)'),
    (0x1F113, '3', u'(d)'),
    (0x1F114, '3', u'(e)'),
    (0x1F115, '3', u'(f)'),
    (0x1F116, '3', u'(g)'),
    (0x1F117, '3', u'(h)'),
    (0x1F118, '3', u'(i)'),
    (0x1F119, '3', u'(j)'),
    (0x1F11A, '3', u'(k)'),
    (0x1F11B, '3', u'(l)'),
    (0x1F11C, '3', u'(m)'),
    (0x1F11D, '3', u'(n)'),
    (0x1F11E, '3', u'(o)'),
    (0x1F11F, '3', u'(p)'),
    (0x1F120, '3', u'(q)'),
    (0x1F121, '3', u'(r)'),
    (0x1F122, '3', u'(s)'),
    (0x1F123, '3', u'(t)'),
    (0x1F124, '3', u'(u)'),
    ]

def _seg_71():
    return [
    (0x1F125, '3', u'(v)'),
    (0x1F126, '3', u'(w)'),
    (0x1F127, '3', u'(x)'),
    (0x1F128, '3', u'(y)'),
    (0x1F129, '3', u'(z)'),
    (0x1F12A, 'M', u'〔s〕'),
    (0x1F12B, 'M', u'c'),
    (0x1F12C, 'M', u'r'),
    (0x1F12D, 'M', u'cd'),
    (0x1F12E, 'M', u'wz'),
    (0x1F12F, 'V'),
    (0x1F130, 'M', u'a'),
    (0x1F131, 'M', u'b'),
    (0x1F132, 'M', u'c'),
    (0x1F133, 'M', u'd'),
    (0x1F134, 'M', u'e'),
    (0x1F135, 'M', u'f'),
    (0x1F136, 'M', u'g'),
    (0x1F137, 'M', u'h'),
    (0x1F138, 'M', u'i'),
    (0x1F139, 'M', u'j'),
    (0x1F13A, 'M', u'k'),
    (0x1F13B, 'M', u'l'),
    (0x1F13C, 'M', u'm'),
    (0x1F13D, 'M', u'n'),
    (0x1F13E, 'M', u'o'),
    (0x1F13F, 'M', u'p'),
    (0x1F140, 'M', u'q'),
    (0x1F141, 'M', u'r'),
    (0x1F142, 'M', u's'),
    (0x1F143, 'M', u't'),
    (0x1F144, 'M', u'u'),
    (0x1F145, 'M', u'v'),
    (0x1F146, 'M', u'w'),
    (0x1F147, 'M', u'x'),
    (0x1F148, 'M', u'y'),
    (0x1F149, 'M', u'z'),
    (0x1F14A, 'M', u'hv'),
    (0x1F14B, 'M', u'mv'),
    (0x1F14C, 'M', u'sd'),
    (0x1F14D, 'M', u'ss'),
    (0x1F14E, 'M', u'ppv'),
    (0x1F14F, 'M', u'wc'),
    (0x1F150, 'V'),
    (0x1F16A, 'M', u'mc'),
    (0x1F16B, 'M', u'md'),
    (0x1F16C, 'X'),
    (0x1F170, 'V'),
    (0x1F190, 'M', u'dj'),
    (0x1F191, 'V'),
    (0x1F1AD, 'X'),
    (0x1F1E6, 'V'),
    (0x1F200, 'M', u'ほか'),
    (0x1F201, 'M', u'ココ'),
    (0x1F202, 'M', u'サ'),
    (0x1F203, 'X'),
    (0x1F210, 'M', u'手'),
    (0x1F211, 'M', u'字'),
    (0x1F212, 'M', u'双'),
    (0x1F213, 'M', u'デ'),
    (0x1F214, 'M', u'二'),
    (0x1F215, 'M', u'多'),
    (0x1F216, 'M', u'解'),
    (0x1F217, 'M', u'天'),
    (0x1F218, 'M', u'交'),
    (0x1F219, 'M', u'映'),
    (0x1F21A, 'M', u'無'),
    (0x1F21B, 'M', u'料'),
    (0x1F21C, 'M', u'前'),
    (0x1F21D, 'M', u'後'),
    (0x1F21E, 'M', u'再'),
    (0x1F21F, 'M', u'新'),
    (0x1F220, 'M', u'初'),
    (0x1F221, 'M', u'終'),
    (0x1F222, 'M', u'生'),
    (0x1F223, 'M', u'販'),
    (0x1F224, 'M', u'声'),
    (0x1F225, 'M', u'吹'),
    (0x1F226, 'M', u'演'),
    (0x1F227, 'M', u'投'),
    (0x1F228, 'M', u'捕'),
    (0x1F229, 'M', u'一'),
    (0x1F22A, 'M', u'三'),
    (0x1F22B, 'M', u'遊'),
    (0x1F22C, 'M', u'左'),
    (0x1F22D, 'M', u'中'),
    (0x1F22E, 'M', u'右'),
    (0x1F22F, 'M', u'指'),
    (0x1F230, 'M', u'走'),
    (0x1F231, 'M', u'打'),
    (0x1F232, 'M', u'禁'),
    (0x1F233, 'M', u'空'),
    (0x1F234, 'M', u'合'),
    (0x1F235, 'M', u'満'),
    (0x1F236, 'M', u'有'),
    (0x1F237, 'M', u'月'),
    (0x1F238, 'M', u'申'),
    (0x1F239, 'M', u'割'),
    (0x1F23A, 'M', u'営'),
    (0x1F23B, 'M', u'配'),
    ]

def _seg_72():
    return [
    (0x1F23C, 'X'),
    (0x1F240, 'M', u'〔本〕'),
    (0x1F241, 'M', u'〔三〕'),
    (0x1F242, 'M', u'〔二〕'),
    (0x1F243, 'M', u'〔安〕'),
    (0x1F244, 'M', u'〔点〕'),
    (0x1F245, 'M', u'〔打〕'),
    (0x1F246, 'M', u'〔盗〕'),
    (0x1F247, 'M', u'〔勝〕'),
    (0x1F248, 'M', u'〔敗〕'),
    (0x1F249, 'X'),
    (0x1F250, 'M', u'得'),
    (0x1F251, 'M', u'可'),
    (0x1F252, 'X'),
    (0x1F260, 'V'),
    (0x1F266, 'X'),
    (0x1F300, 'V'),
    (0x1F6D5, 'X'),
    (0x1F6E0, 'V'),
    (0x1F6ED, 'X'),
    (0x1F6F0, 'V'),
    (0x1F6FA, 'X'),
    (0x1F700, 'V'),
    (0x1F774, 'X'),
    (0x1F780, 'V'),
    (0x1F7D9, 'X'),
    (0x1F800, 'V'),
    (0x1F80C, 'X'),
    (0x1F810, 'V'),
    (0x1F848, 'X'),
    (0x1F850, 'V'),
    (0x1F85A, 'X'),
    (0x1F860, 'V'),
    (0x1F888, 'X'),
    (0x1F890, 'V'),
    (0x1F8AE, 'X'),
    (0x1F900, 'V'),
    (0x1F90C, 'X'),
    (0x1F910, 'V'),
    (0x1F93F, 'X'),
    (0x1F940, 'V'),
    (0x1F971, 'X'),
    (0x1F973, 'V'),
    (0x1F977, 'X'),
    (0x1F97A, 'V'),
    (0x1F97B, 'X'),
    (0x1F97C, 'V'),
    (0x1F9A3, 'X'),
    (0x1F9B0, 'V'),
    (0x1F9BA, 'X'),
    (0x1F9C0, 'V'),
    (0x1F9C3, 'X'),
    (0x1F9D0, 'V'),
    (0x1FA00, 'X'),
    (0x1FA60, 'V'),
    (0x1FA6E, 'X'),
    (0x20000, 'V'),
    (0x2A6D7, 'X'),
    (0x2A700, 'V'),
    (0x2B735, 'X'),
    (0x2B740, 'V'),
    (0x2B81E, 'X'),
    (0x2B820, 'V'),
    (0x2CEA2, 'X'),
    (0x2CEB0, 'V'),
    (0x2EBE1, 'X'),
    (0x2F800, 'M', u'丽'),
    (0x2F801, 'M', u'丸'),
    (0x2F802, 'M', u'乁'),
    (0x2F803, 'M', u'𠄢'),
    (0x2F804, 'M', u'你'),
    (0x2F805, 'M', u'侮'),
    (0x2F806, 'M', u'侻'),
    (0x2F807, 'M', u'倂'),
    (0x2F808, 'M', u'偺'),
    (0x2F809, 'M', u'備'),
    (0x2F80A, 'M', u'僧'),
    (0x2F80B, 'M', u'像'),
    (0x2F80C, 'M', u'㒞'),
    (0x2F80D, 'M', u'𠘺'),
    (0x2F80E, 'M', u'免'),
    (0x2F80F, 'M', u'兔'),
    (0x2F810, 'M', u'兤'),
    (0x2F811, 'M', u'具'),
    (0x2F812, 'M', u'𠔜'),
    (0x2F813, 'M', u'㒹'),
    (0x2F814, 'M', u'內'),
    (0x2F815, 'M', u'再'),
    (0x2F816, 'M', u'𠕋'),
    (0x2F817, 'M', u'冗'),
    (0x2F818, 'M', u'冤'),
    (0x2F819, 'M', u'仌'),
    (0x2F81A, 'M', u'冬'),
    (0x2F81B, 'M', u'况'),
    (0x2F81C, 'M', u'𩇟'),
    (0x2F81D, 'M', u'凵'),
    (0x2F81E, 'M', u'刃'),
    (0x2F81F, 'M', u'㓟'),
    (0x2F820, 'M', u'刻'),
    (0x2F821, 'M', u'剆'),
    ]

def _seg_73():
    return [
    (0x2F822, 'M', u'割'),
    (0x2F823, 'M', u'剷'),
    (0x2F824, 'M', u'㔕'),
    (0x2F825, 'M', u'勇'),
    (0x2F826, 'M', u'勉'),
    (0x2F827, 'M', u'勤'),
    (0x2F828, 'M', u'勺'),
    (0x2F829, 'M', u'包'),
    (0x2F82A, 'M', u'匆'),
    (0x2F82B, 'M', u'北'),
    (0x2F82C, 'M', u'卉'),
    (0x2F82D, 'M', u'卑'),
    (0x2F82E, 'M', u'博'),
    (0x2F82F, 'M', u'即'),
    (0x2F830, 'M', u'卽'),
    (0x2F831, 'M', u'卿'),
    (0x2F834, 'M', u'𠨬'),
    (0x2F835, 'M', u'灰'),
    (0x2F836, 'M', u'及'),
    (0x2F837, 'M', u'叟'),
    (0x2F838, 'M', u'𠭣'),
    (0x2F839, 'M', u'叫'),
    (0x2F83A, 'M', u'叱'),
    (0x2F83B, 'M', u'吆'),
    (0x2F83C, 'M', u'咞'),
    (0x2F83D, 'M', u'吸'),
    (0x2F83E, 'M', u'呈'),
    (0x2F83F, 'M', u'周'),
    (0x2F840, 'M', u'咢'),
    (0x2F841, 'M', u'哶'),
    (0x2F842, 'M', u'唐'),
    (0x2F843, 'M', u'啓'),
    (0x2F844, 'M', u'啣'),
    (0x2F845, 'M', u'善'),
    (0x2F847, 'M', u'喙'),
    (0x2F848, 'M', u'喫'),
    (0x2F849, 'M', u'喳'),
    (0x2F84A, 'M', u'嗂'),
    (0x2F84B, 'M', u'圖'),
    (0x2F84C, 'M', u'嘆'),
    (0x2F84D, 'M', u'圗'),
    (0x2F84E, 'M', u'噑'),
    (0x2F84F, 'M', u'噴'),
    (0x2F850, 'M', u'切'),
    (0x2F851, 'M', u'壮'),
    (0x2F852, 'M', u'城'),
    (0x2F853, 'M', u'埴'),
    (0x2F854, 'M', u'堍'),
    (0x2F855, 'M', u'型'),
    (0x2F856, 'M', u'堲'),
    (0x2F857, 'M', u'報'),
    (0x2F858, 'M', u'墬'),
    (0x2F859, 'M', u'𡓤'),
    (0x2F85A, 'M', u'売'),
    (0x2F85B, 'M', u'壷'),
    (0x2F85C, 'M', u'夆'),
    (0x2F85D, 'M', u'多'),
    (0x2F85E, 'M', u'夢'),
    (0x2F85F, 'M', u'奢'),
    (0x2F860, 'M', u'𡚨'),
    (0x2F861, 'M', u'𡛪'),
    (0x2F862, 'M', u'姬'),
    (0x2F863, 'M', u'娛'),
    (0x2F864, 'M', u'娧'),
    (0x2F865, 'M', u'姘'),
    (0x2F866, 'M', u'婦'),
    (0x2F867, 'M', u'㛮'),
    (0x2F868, 'X'),
    (0x2F869, 'M', u'嬈'),
    (0x2F86A, 'M', u'嬾'),
    (0x2F86C, 'M', u'𡧈'),
    (0x2F86D, 'M', u'寃'),
    (0x2F86E, 'M', u'寘'),
    (0x2F86F, 'M', u'寧'),
    (0x2F870, 'M', u'寳'),
    (0x2F871, 'M', u'𡬘'),
    (0x2F872, 'M', u'寿'),
    (0x2F873, 'M', u'将'),
    (0x2F874, 'X'),
    (0x2F875, 'M', u'尢'),
    (0x2F876, 'M', u'㞁'),
    (0x2F877, 'M', u'屠'),
    (0x2F878, 'M', u'屮'),
    (0x2F879, 'M', u'峀'),
    (0x2F87A, 'M', u'岍'),
    (0x2F87B, 'M', u'𡷤'),
    (0x2F87C, 'M', u'嵃'),
    (0x2F87D, 'M', u'𡷦'),
    (0x2F87E, 'M', u'嵮'),
    (0x2F87F, 'M', u'嵫'),
    (0x2F880, 'M', u'嵼'),
    (0x2F881, 'M', u'巡'),
    (0x2F882, 'M', u'巢'),
    (0x2F883, 'M', u'㠯'),
    (0x2F884, 'M', u'巽'),
    (0x2F885, 'M', u'帨'),
    (0x2F886, 'M', u'帽'),
    (0x2F887, 'M', u'幩'),
    (0x2F888, 'M', u'㡢'),
    (0x2F889, 'M', u'𢆃'),
    ]

def _seg_74():
    return [
    (0x2F88A, 'M', u'㡼'),
    (0x2F88B, 'M', u'庰'),
    (0x2F88C, 'M', u'庳'),
    (0x2F88D, 'M', u'庶'),
    (0x2F88E, 'M', u'廊'),
    (0x2F88F, 'M', u'𪎒'),
    (0x2F890, 'M', u'廾'),
    (0x2F891, 'M', u'𢌱'),
    (0x2F893, 'M', u'舁'),
    (0x2F894, 'M', u'弢'),
    (0x2F896, 'M', u'㣇'),
    (0x2F897, 'M', u'𣊸'),
    (0x2F898, 'M', u'𦇚'),
    (0x2F899, 'M', u'形'),
    (0x2F89A, 'M', u'彫'),
    (0x2F89B, 'M', u'㣣'),
    (0x2F89C, 'M', u'徚'),
    (0x2F89D, 'M', u'忍'),
    (0x2F89E, 'M', u'志'),
    (0x2F89F, 'M', u'忹'),
    (0x2F8A0, 'M', u'悁'),
    (0x2F8A1, 'M', u'㤺'),
    (0x2F8A2, 'M', u'㤜'),
    (0x2F8A3, 'M', u'悔'),
    (0x2F8A4, 'M', u'𢛔'),
    (0x2F8A5, 'M', u'惇'),
    (0x2F8A6, 'M', u'慈'),
    (0x2F8A7, 'M', u'慌'),
    (0x2F8A8, 'M', u'慎'),
    (0x2F8A9, 'M', u'慌'),
    (0x2F8AA, 'M', u'慺'),
    (0x2F8AB, 'M', u'憎'),
    (0x2F8AC, 'M', u'憲'),
    (0x2F8AD, 'M', u'憤'),
    (0x2F8AE, 'M', u'憯'),
    (0x2F8AF, 'M', u'懞'),
    (0x2F8B0, 'M', u'懲'),
    (0x2F8B1, 'M', u'懶'),
    (0x2F8B2, 'M', u'成'),
    (0x2F8B3, 'M', u'戛'),
    (0x2F8B4, 'M', u'扝'),
    (0x2F8B5, 'M', u'抱'),
    (0x2F8B6, 'M', u'拔'),
    (0x2F8B7, 'M', u'捐'),
    (0x2F8B8, 'M', u'𢬌'),
    (0x2F8B9, 'M', u'挽'),
    (0x2F8BA, 'M', u'拼'),
    (0x2F8BB, 'M', u'捨'),
    (0x2F8BC, 'M', u'掃'),
    (0x2F8BD, 'M', u'揤'),
    (0x2F8BE, 'M', u'𢯱'),
    (0x2F8BF, 'M', u'搢'),
    (0x2F8C0, 'M', u'揅'),
    (0x2F8C1, 'M', u'掩'),
    (0x2F8C2, 'M', u'㨮'),
    (0x2F8C3, 'M', u'摩'),
    (0x2F8C4, 'M', u'摾'),
    (0x2F8C5, 'M', u'撝'),
    (0x2F8C6, 'M', u'摷'),
    (0x2F8C7, 'M', u'㩬'),
    (0x2F8C8, 'M', u'敏'),
    (0x2F8C9, 'M', u'敬'),
    (0x2F8CA, 'M', u'𣀊'),
    (0x2F8CB, 'M', u'旣'),
    (0x2F8CC, 'M', u'書'),
    (0x2F8CD, 'M', u'晉'),
    (0x2F8CE, 'M', u'㬙'),
    (0x2F8CF, 'M', u'暑'),
    (0x2F8D0, 'M', u'㬈'),
    (0x2F8D1, 'M', u'㫤'),
    (0x2F8D2, 'M', u'冒'),
    (0x2F8D3, 'M', u'冕'),
    (0x2F8D4, 'M', u'最'),
    (0x2F8D5, 'M', u'暜'),
    (0x2F8D6, 'M', u'肭'),
    (0x2F8D7, 'M', u'䏙'),
    (0x2F8D8, 'M', u'朗'),
    (0x2F8D9, 'M', u'望'),
    (0x2F8DA, 'M', u'朡'),
    (0x2F8DB, 'M', u'杞'),
    (0x2F8DC, 'M', u'杓'),
    (0x2F8DD, 'M', u'𣏃'),
    (0x2F8DE, 'M', u'㭉'),
    (0x2F8DF, 'M', u'柺'),
    (0x2F8E0, 'M', u'枅'),
    (0x2F8E1, 'M', u'桒'),
    (0x2F8E2, 'M', u'梅'),
    (0x2F8E3, 'M', u'𣑭'),
    (0x2F8E4, 'M', u'梎'),
    (0x2F8E5, 'M', u'栟'),
    (0x2F8E6, 'M', u'椔'),
    (0x2F8E7, 'M', u'㮝'),
    (0x2F8E8, 'M', u'楂'),
    (0x2F8E9, 'M', u'榣'),
    (0x2F8EA, 'M', u'槪'),
    (0x2F8EB, 'M', u'檨'),
    (0x2F8EC, 'M', u'𣚣'),
    (0x2F8ED, 'M', u'櫛'),
    (0x2F8EE, 'M', u'㰘'),
    (0x2F8EF, 'M', u'次'),
    ]

def _seg_75():
    return [
    (0x2F8F0, 'M', u'𣢧'),
    (0x2F8F1, 'M', u'歔'),
    (0x2F8F2, 'M', u'㱎'),
    (0x2F8F3, 'M', u'歲'),
    (0x2F8F4, 'M', u'殟'),
    (0x2F8F5, 'M', u'殺'),
    (0x2F8F6, 'M', u'殻'),
    (0x2F8F7, 'M', u'𣪍'),
    (0x2F8F8, 'M', u'𡴋'),
    (0x2F8F9, 'M', u'𣫺'),
    (0x2F8FA, 'M', u'汎'),
    (0x2F8FB, 'M', u'𣲼'),
    (0x2F8FC, 'M', u'沿'),
    (0x2F8FD, 'M', u'泍'),
    (0x2F8FE, 'M', u'汧'),
    (0x2F8FF, 'M', u'洖'),
    (0x2F900, 'M', u'派'),
    (0x2F901, 'M', u'海'),
    (0x2F902, 'M', u'流'),
    (0x2F903, 'M', u'浩'),
    (0x2F904, 'M', u'浸'),
    (0x2F905, 'M', u'涅'),
    (0x2F906, 'M', u'𣴞'),
    (0x2F907, 'M', u'洴'),
    (0x2F908, 'M', u'港'),
    (0x2F909, 'M', u'湮'),
    (0x2F90A, 'M', u'㴳'),
    (0x2F90B, 'M', u'滋'),
    (0x2F90C, 'M', u'滇'),
    (0x2F90D, 'M', u'𣻑'),
    (0x2F90E, 'M', u'淹'),
    (0x2F90F, 'M', u'潮'),
    (0x2F910, 'M', u'𣽞'),
    (0x2F911, 'M', u'𣾎'),
    (0x2F912, 'M', u'濆'),
    (0x2F913, 'M', u'瀹'),
    (0x2F914, 'M', u'瀞'),
    (0x2F915, 'M', u'瀛'),
    (0x2F916, 'M', u'㶖'),
    (0x2F917, 'M', u'灊'),
    (0x2F918, 'M', u'災'),
    (0x2F919, 'M', u'灷'),
    (0x2F91A, 'M', u'炭'),
    (0x2F91B, 'M', u'𠔥'),
    (0x2F91C, 'M', u'煅'),
    (0x2F91D, 'M', u'𤉣'),
    (0x2F91E, 'M', u'熜'),
    (0x2F91F, 'X'),
    (0x2F920, 'M', u'爨'),
    (0x2F921, 'M', u'爵'),
    (0x2F922, 'M', u'牐'),
    (0x2F923, 'M', u'𤘈'),
    (0x2F924, 'M', u'犀'),
    (0x2F925, 'M', u'犕'),
    (0x2F926, 'M', u'𤜵'),
    (0x2F927, 'M', u'𤠔'),
    (0x2F928, 'M', u'獺'),
    (0x2F929, 'M', u'王'),
    (0x2F92A, 'M', u'㺬'),
    (0x2F92B, 'M', u'玥'),
    (0x2F92C, 'M', u'㺸'),
    (0x2F92E, 'M', u'瑇'),
    (0x2F92F, 'M', u'瑜'),
    (0x2F930, 'M', u'瑱'),
    (0x2F931, 'M', u'璅'),
    (0x2F932, 'M', u'瓊'),
    (0x2F933, 'M', u'㼛'),
    (0x2F934, 'M', u'甤'),
    (0x2F935, 'M', u'𤰶'),
    (0x2F936, 'M', u'甾'),
    (0x2F937, 'M', u'𤲒'),
    (0x2F938, 'M', u'異'),
    (0x2F939, 'M', u'𢆟'),
    (0x2F93A, 'M', u'瘐'),
    (0x2F93B, 'M', u'𤾡'),
    (0x2F93C, 'M', u'𤾸'),
    (0x2F93D, 'M', u'𥁄'),
    (0x2F93E, 'M', u'㿼'),
    (0x2F93F, 'M', u'䀈'),
    (0x2F940, 'M', u'直'),
    (0x2F941, 'M', u'𥃳'),
    (0x2F942, 'M', u'𥃲'),
    (0x2F943, 'M', u'𥄙'),
    (0x2F944, 'M', u'𥄳'),
    (0x2F945, 'M', u'眞'),
    (0x2F946, 'M', u'真'),
    (0x2F948, 'M', u'睊'),
    (0x2F949, 'M', u'䀹'),
    (0x2F94A, 'M', u'瞋'),
    (0x2F94B, 'M', u'䁆'),
    (0x2F94C, 'M', u'䂖'),
    (0x2F94D, 'M', u'𥐝'),
    (0x2F94E, 'M', u'硎'),
    (0x2F94F, 'M', u'碌'),
    (0x2F950, 'M', u'磌'),
    (0x2F951, 'M', u'䃣'),
    (0x2F952, 'M', u'𥘦'),
    (0x2F953, 'M', u'祖'),
    (0x2F954, 'M', u'𥚚'),
    (0x2F955, 'M', u'𥛅'),
    ]

def _seg_76():
    return [
    (0x2F956, 'M', u'福'),
    (0x2F957, 'M', u'秫'),
    (0x2F958, 'M', u'䄯'),
    (0x2F959, 'M', u'穀'),
    (0x2F95A, 'M', u'穊'),
    (0x2F95B, 'M', u'穏'),
    (0x2F95C, 'M', u'𥥼'),
    (0x2F95D, 'M', u'𥪧'),
    (0x2F95F, 'X'),
    (0x2F960, 'M', u'䈂'),
    (0x2F961, 'M', u'𥮫'),
    (0x2F962, 'M', u'篆'),
    (0x2F963, 'M', u'築'),
    (0x2F964, 'M', u'䈧'),
    (0x2F965, 'M', u'𥲀'),
    (0x2F966, 'M', u'糒'),
    (0x2F967, 'M', u'䊠'),
    (0x2F968, 'M', u'糨'),
    (0x2F969, 'M', u'糣'),
    (0x2F96A, 'M', u'紀'),
    (0x2F96B, 'M', u'𥾆'),
    (0x2F96C, 'M', u'絣'),
    (0x2F96D, 'M', u'䌁'),
    (0x2F96E, 'M', u'緇'),
    (0x2F96F, 'M', u'縂'),
    (0x2F970, 'M', u'繅'),
    (0x2F971, 'M', u'䌴'),
    (0x2F972, 'M', u'𦈨'),
    (0x2F973, 'M', u'𦉇'),
    (0x2F974, 'M', u'䍙'),
    (0x2F975, 'M', u'𦋙'),
    (0x2F976, 'M', u'罺'),
    (0x2F977, 'M', u'𦌾'),
    (0x2F978, 'M', u'羕'),
    (0x2F979, 'M', u'翺'),
    (0x2F97A, 'M', u'者'),
    (0x2F97B, 'M', u'𦓚'),
    (0x2F97C, 'M', u'𦔣'),
    (0x2F97D, 'M', u'聠'),
    (0x2F97E, 'M', u'𦖨'),
    (0x2F97F, 'M', u'聰'),
    (0x2F980, 'M', u'𣍟'),
    (0x2F981, 'M', u'䏕'),
    (0x2F982, 'M', u'育'),
    (0x2F983, 'M', u'脃'),
    (0x2F984, 'M', u'䐋'),
    (0x2F985, 'M', u'脾'),
    (0x2F986, 'M', u'媵'),
    (0x2F987, 'M', u'𦞧'),
    (0x2F988, 'M', u'𦞵'),
    (0x2F989, 'M', u'𣎓'),
    (0x2F98A, 'M', u'𣎜'),
    (0x2F98B, 'M', u'舁'),
    (0x2F98C, 'M', u'舄'),
    (0x2F98D, 'M', u'辞'),
    (0x2F98E, 'M', u'䑫'),
    (0x2F98F, 'M', u'芑'),
    (0x2F990, 'M', u'芋'),
    (0x2F991, 'M', u'芝'),
    (0x2F992, 'M', u'劳'),
    (0x2F993, 'M', u'花'),
    (0x2F994, 'M', u'芳'),
    (0x2F995, 'M', u'芽'),
    (0x2F996, 'M', u'苦'),
    (0x2F997, 'M', u'𦬼'),
    (0x2F998, 'M', u'若'),
    (0x2F999, 'M', u'茝'),
    (0x2F99A, 'M', u'荣'),
    (0x2F99B, 'M', u'莭'),
    (0x2F99C, 'M', u'茣'),
    (0x2F99D, 'M', u'莽'),
    (0x2F99E, 'M', u'菧'),
    (0x2F99F, 'M', u'著'),
    (0x2F9A0, 'M', u'荓'),
    (0x2F9A1, 'M', u'菊'),
    (0x2F9A2, 'M', u'菌'),
    (0x2F9A3, 'M', u'菜'),
    (0x2F9A4, 'M', u'𦰶'),
    (0x2F9A5, 'M', u'𦵫'),
    (0x2F9A6, 'M', u'𦳕'),
    (0x2F9A7, 'M', u'䔫'),
    (0x2F9A8, 'M', u'蓱'),
    (0x2F9A9, 'M', u'蓳'),
    (0x2F9AA, 'M', u'蔖'),
    (0x2F9AB, 'M', u'𧏊'),
    (0x2F9AC, 'M', u'蕤'),
    (0x2F9AD, 'M', u'𦼬'),
    (0x2F9AE, 'M', u'䕝'),
    (0x2F9AF, 'M', u'䕡'),
    (0x2F9B0, 'M', u'𦾱'),
    (0x2F9B1, 'M', u'𧃒'),
    (0x2F9B2, 'M', u'䕫'),
    (0x2F9B3, 'M', u'虐'),
    (0x2F9B4, 'M', u'虜'),
    (0x2F9B5, 'M', u'虧'),
    (0x2F9B6, 'M', u'虩'),
    (0x2F9B7, 'M', u'蚩'),
    (0x2F9B8, 'M', u'蚈'),
    (0x2F9B9, 'M', u'蜎'),
    (0x2F9BA, 'M', u'蛢'),
    ]

def _seg_77():
    return [
    (0x2F9BB, 'M', u'蝹'),
    (0x2F9BC, 'M', u'蜨'),
    (0x2F9BD, 'M', u'蝫'),
    (0x2F9BE, 'M', u'螆'),
    (0x2F9BF, 'X'),
    (0x2F9C0, 'M', u'蟡'),
    (0x2F9C1, 'M', u'蠁'),
    (0x2F9C2, 'M', u'䗹'),
    (0x2F9C3, 'M', u'衠'),
    (0x2F9C4, 'M', u'衣'),
    (0x2F9C5, 'M', u'𧙧'),
    (0x2F9C6, 'M', u'裗'),
    (0x2F9C7, 'M', u'裞'),
    (0x2F9C8, 'M', u'䘵'),
    (0x2F9C9, 'M', u'裺'),
    (0x2F9CA, 'M', u'㒻'),
    (0x2F9CB, 'M', u'𧢮'),
    (0x2F9CC, 'M', u'𧥦'),
    (0x2F9CD, 'M', u'䚾'),
    (0x2F9CE, 'M', u'䛇'),
    (0x2F9CF, 'M', u'誠'),
    (0x2F9D0, 'M', u'諭'),
    (0x2F9D1, 'M', u'變'),
    (0x2F9D2, 'M', u'豕'),
    (0x2F9D3, 'M', u'𧲨'),
    (0x2F9D4, 'M', u'貫'),
    (0x2F9D5, 'M', u'賁'),
    (0x2F9D6, 'M', u'贛'),
    (0x2F9D7, 'M', u'起'),
    (0x2F9D8, 'M', u'𧼯'),
    (0x2F9D9, 'M', u'𠠄'),
    (0x2F9DA, 'M', u'跋'),
    (0x2F9DB, 'M', u'趼'),
    (0x2F9DC, 'M', u'跰'),
    (0x2F9DD, 'M', u'𠣞'),
    (0x2F9DE, 'M', u'軔'),
    (0x2F9DF, 'M', u'輸'),
    (0x2F9E0, 'M', u'𨗒'),
    (0x2F9E1, 'M', u'𨗭'),
    (0x2F9E2, 'M', u'邔'),
    (0x2F9E3, 'M', u'郱'),
    (0x2F9E4, 'M', u'鄑'),
    (0x2F9E5, 'M', u'𨜮'),
    (0x2F9E6, 'M', u'鄛'),
    (0x2F9E7, 'M', u'鈸'),
    (0x2F9E8, 'M', u'鋗'),
    (0x2F9E9, 'M', u'鋘'),
    (0x2F9EA, 'M', u'鉼'),
    (0x2F9EB, 'M', u'鏹'),
    (0x2F9EC, 'M', u'鐕'),
    (0x2F9ED, 'M', u'𨯺'),
    (0x2F9EE, 'M', u'開'),
    (0x2F9EF, 'M', u'䦕'),
    (0x2F9F0, 'M', u'閷'),
    (0x2F9F1, 'M', u'𨵷'),
    (0x2F9F2, 'M', u'䧦'),
    (0x2F9F3, 'M', u'雃'),
    (0x2F9F4, 'M', u'嶲'),
    (0x2F9F5, 'M', u'霣'),
    (0x2F9F6, 'M', u'𩅅'),
    (0x2F9F7, 'M', u'𩈚'),
    (0x2F9F8, 'M', u'䩮'),
    (0x2F9F9, 'M', u'䩶'),
    (0x2F9FA, 'M', u'韠'),
    (0x2F9FB, 'M', u'𩐊'),
    (0x2F9FC, 'M', u'䪲'),
    (0x2F9FD, 'M', u'𩒖'),
    (0x2F9FE, 'M', u'頋'),
    (0x2FA00, 'M', u'頩'),
    (0x2FA01, 'M', u'𩖶'),
    (0x2FA02, 'M', u'飢'),
    (0x2FA03, 'M', u'䬳'),
    (0x2FA04, 'M', u'餩'),
    (0x2FA05, 'M', u'馧'),
    (0x2FA06, 'M', u'駂'),
    (0x2FA07, 'M', u'駾'),
    (0x2FA08, 'M', u'䯎'),
    (0x2FA09, 'M', u'𩬰'),
    (0x2FA0A, 'M', u'鬒'),
    (0x2FA0B, 'M', u'鱀'),
    (0x2FA0C, 'M', u'鳽'),
    (0x2FA0D, 'M', u'䳎'),
    (0x2FA0E, 'M', u'䳭'),
    (0x2FA0F, 'M', u'鵧'),
    (0x2FA10, 'M', u'𪃎'),
    (0x2FA11, 'M', u'䳸'),
    (0x2FA12, 'M', u'𪄅'),
    (0x2FA13, 'M', u'𪈎'),
    (0x2FA14, 'M', u'𪊑'),
    (0x2FA15, 'M', u'麻'),
    (0x2FA16, 'M', u'䵖'),
    (0x2FA17, 'M', u'黹'),
    (0x2FA18, 'M', u'黾'),
    (0x2FA19, 'M', u'鼅'),
    (0x2FA1A, 'M', u'鼏'),
    (0x2FA1B, 'M', u'鼖'),
    (0x2FA1C, 'M', u'鼻'),
    (0x2FA1D, 'M', u'𪘀'),
    (0x2FA1E, 'X'),
    (0xE0100, 'I'),
    ]

def _seg_78():
    return [
    (0xE01F0, 'X'),
    ]

uts46data = tuple(
    _seg_0()
    + _seg_1()
    + _seg_2()
    + _seg_3()
    + _seg_4()
    + _seg_5()
    + _seg_6()
    + _seg_7()
    + _seg_8()
    + _seg_9()
    + _seg_10()
    + _seg_11()
    + _seg_12()
    + _seg_13()
    + _seg_14()
    + _seg_15()
    + _seg_16()
    + _seg_17()
    + _seg_18()
    + _seg_19()
    + _seg_20()
    + _seg_21()
    + _seg_22()
    + _seg_23()
    + _seg_24()
    + _seg_25()
    + _seg_26()
    + _seg_27()
    + _seg_28()
    + _seg_29()
    + _seg_30()
    + _seg_31()
    + _seg_32()
    + _seg_33()
    + _seg_34()
    + _seg_35()
    + _seg_36()
    + _seg_37()
    + _seg_38()
    + _seg_39()
    + _seg_40()
    + _seg_41()
    + _seg_42()
    + _seg_43()
    + _seg_44()
    + _seg_45()
    + _seg_46()
    + _seg_47()
    + _seg_48()
    + _seg_49()
    + _seg_50()
    + _seg_51()
    + _seg_52()
    + _seg_53()
    + _seg_54()
    + _seg_55()
    + _seg_56()
    + _seg_57()
    + _seg_58()
    + _seg_59()
    + _seg_60()
    + _seg_61()
    + _seg_62()
    + _seg_63()
    + _seg_64()
    + _seg_65()
    + _seg_66()
    + _seg_67()
    + _seg_68()
    + _seg_69()
    + _seg_70()
    + _seg_71()
    + _seg_72()
    + _seg_73()
    + _seg_74()
    + _seg_75()
    + _seg_76()
    + _seg_77()
    + _seg_78()
)
site-packages/pip/_vendor/idna/idnadata.py000064400000117703151464452710014564 0ustar00# This file is automatically generated by tools/idna-data

__version__ = "11.0.0"
scripts = {
    'Greek': (
        0x37000000374,
        0x37500000378,
        0x37a0000037e,
        0x37f00000380,
        0x38400000385,
        0x38600000387,
        0x3880000038b,
        0x38c0000038d,
        0x38e000003a2,
        0x3a3000003e2,
        0x3f000000400,
        0x1d2600001d2b,
        0x1d5d00001d62,
        0x1d6600001d6b,
        0x1dbf00001dc0,
        0x1f0000001f16,
        0x1f1800001f1e,
        0x1f2000001f46,
        0x1f4800001f4e,
        0x1f5000001f58,
        0x1f5900001f5a,
        0x1f5b00001f5c,
        0x1f5d00001f5e,
        0x1f5f00001f7e,
        0x1f8000001fb5,
        0x1fb600001fc5,
        0x1fc600001fd4,
        0x1fd600001fdc,
        0x1fdd00001ff0,
        0x1ff200001ff5,
        0x1ff600001fff,
        0x212600002127,
        0xab650000ab66,
        0x101400001018f,
        0x101a0000101a1,
        0x1d2000001d246,
    ),
    'Han': (
        0x2e8000002e9a,
        0x2e9b00002ef4,
        0x2f0000002fd6,
        0x300500003006,
        0x300700003008,
        0x30210000302a,
        0x30380000303c,
        0x340000004db6,
        0x4e0000009ff0,
        0xf9000000fa6e,
        0xfa700000fada,
        0x200000002a6d7,
        0x2a7000002b735,
        0x2b7400002b81e,
        0x2b8200002cea2,
        0x2ceb00002ebe1,
        0x2f8000002fa1e,
    ),
    'Hebrew': (
        0x591000005c8,
        0x5d0000005eb,
        0x5ef000005f5,
        0xfb1d0000fb37,
        0xfb380000fb3d,
        0xfb3e0000fb3f,
        0xfb400000fb42,
        0xfb430000fb45,
        0xfb460000fb50,
    ),
    'Hiragana': (
        0x304100003097,
        0x309d000030a0,
        0x1b0010001b11f,
        0x1f2000001f201,
    ),
    'Katakana': (
        0x30a1000030fb,
        0x30fd00003100,
        0x31f000003200,
        0x32d0000032ff,
        0x330000003358,
        0xff660000ff70,
        0xff710000ff9e,
        0x1b0000001b001,
    ),
}
joining_types = {
    0x600: 85,
    0x601: 85,
    0x602: 85,
    0x603: 85,
    0x604: 85,
    0x605: 85,
    0x608: 85,
    0x60b: 85,
    0x620: 68,
    0x621: 85,
    0x622: 82,
    0x623: 82,
    0x624: 82,
    0x625: 82,
    0x626: 68,
    0x627: 82,
    0x628: 68,
    0x629: 82,
    0x62a: 68,
    0x62b: 68,
    0x62c: 68,
    0x62d: 68,
    0x62e: 68,
    0x62f: 82,
    0x630: 82,
    0x631: 82,
    0x632: 82,
    0x633: 68,
    0x634: 68,
    0x635: 68,
    0x636: 68,
    0x637: 68,
    0x638: 68,
    0x639: 68,
    0x63a: 68,
    0x63b: 68,
    0x63c: 68,
    0x63d: 68,
    0x63e: 68,
    0x63f: 68,
    0x640: 67,
    0x641: 68,
    0x642: 68,
    0x643: 68,
    0x644: 68,
    0x645: 68,
    0x646: 68,
    0x647: 68,
    0x648: 82,
    0x649: 68,
    0x64a: 68,
    0x66e: 68,
    0x66f: 68,
    0x671: 82,
    0x672: 82,
    0x673: 82,
    0x674: 85,
    0x675: 82,
    0x676: 82,
    0x677: 82,
    0x678: 68,
    0x679: 68,
    0x67a: 68,
    0x67b: 68,
    0x67c: 68,
    0x67d: 68,
    0x67e: 68,
    0x67f: 68,
    0x680: 68,
    0x681: 68,
    0x682: 68,
    0x683: 68,
    0x684: 68,
    0x685: 68,
    0x686: 68,
    0x687: 68,
    0x688: 82,
    0x689: 82,
    0x68a: 82,
    0x68b: 82,
    0x68c: 82,
    0x68d: 82,
    0x68e: 82,
    0x68f: 82,
    0x690: 82,
    0x691: 82,
    0x692: 82,
    0x693: 82,
    0x694: 82,
    0x695: 82,
    0x696: 82,
    0x697: 82,
    0x698: 82,
    0x699: 82,
    0x69a: 68,
    0x69b: 68,
    0x69c: 68,
    0x69d: 68,
    0x69e: 68,
    0x69f: 68,
    0x6a0: 68,
    0x6a1: 68,
    0x6a2: 68,
    0x6a3: 68,
    0x6a4: 68,
    0x6a5: 68,
    0x6a6: 68,
    0x6a7: 68,
    0x6a8: 68,
    0x6a9: 68,
    0x6aa: 68,
    0x6ab: 68,
    0x6ac: 68,
    0x6ad: 68,
    0x6ae: 68,
    0x6af: 68,
    0x6b0: 68,
    0x6b1: 68,
    0x6b2: 68,
    0x6b3: 68,
    0x6b4: 68,
    0x6b5: 68,
    0x6b6: 68,
    0x6b7: 68,
    0x6b8: 68,
    0x6b9: 68,
    0x6ba: 68,
    0x6bb: 68,
    0x6bc: 68,
    0x6bd: 68,
    0x6be: 68,
    0x6bf: 68,
    0x6c0: 82,
    0x6c1: 68,
    0x6c2: 68,
    0x6c3: 82,
    0x6c4: 82,
    0x6c5: 82,
    0x6c6: 82,
    0x6c7: 82,
    0x6c8: 82,
    0x6c9: 82,
    0x6ca: 82,
    0x6cb: 82,
    0x6cc: 68,
    0x6cd: 82,
    0x6ce: 68,
    0x6cf: 82,
    0x6d0: 68,
    0x6d1: 68,
    0x6d2: 82,
    0x6d3: 82,
    0x6d5: 82,
    0x6dd: 85,
    0x6ee: 82,
    0x6ef: 82,
    0x6fa: 68,
    0x6fb: 68,
    0x6fc: 68,
    0x6ff: 68,
    0x70f: 84,
    0x710: 82,
    0x712: 68,
    0x713: 68,
    0x714: 68,
    0x715: 82,
    0x716: 82,
    0x717: 82,
    0x718: 82,
    0x719: 82,
    0x71a: 68,
    0x71b: 68,
    0x71c: 68,
    0x71d: 68,
    0x71e: 82,
    0x71f: 68,
    0x720: 68,
    0x721: 68,
    0x722: 68,
    0x723: 68,
    0x724: 68,
    0x725: 68,
    0x726: 68,
    0x727: 68,
    0x728: 82,
    0x729: 68,
    0x72a: 82,
    0x72b: 68,
    0x72c: 82,
    0x72d: 68,
    0x72e: 68,
    0x72f: 82,
    0x74d: 82,
    0x74e: 68,
    0x74f: 68,
    0x750: 68,
    0x751: 68,
    0x752: 68,
    0x753: 68,
    0x754: 68,
    0x755: 68,
    0x756: 68,
    0x757: 68,
    0x758: 68,
    0x759: 82,
    0x75a: 82,
    0x75b: 82,
    0x75c: 68,
    0x75d: 68,
    0x75e: 68,
    0x75f: 68,
    0x760: 68,
    0x761: 68,
    0x762: 68,
    0x763: 68,
    0x764: 68,
    0x765: 68,
    0x766: 68,
    0x767: 68,
    0x768: 68,
    0x769: 68,
    0x76a: 68,
    0x76b: 82,
    0x76c: 82,
    0x76d: 68,
    0x76e: 68,
    0x76f: 68,
    0x770: 68,
    0x771: 82,
    0x772: 68,
    0x773: 82,
    0x774: 82,
    0x775: 68,
    0x776: 68,
    0x777: 68,
    0x778: 82,
    0x779: 82,
    0x77a: 68,
    0x77b: 68,
    0x77c: 68,
    0x77d: 68,
    0x77e: 68,
    0x77f: 68,
    0x7ca: 68,
    0x7cb: 68,
    0x7cc: 68,
    0x7cd: 68,
    0x7ce: 68,
    0x7cf: 68,
    0x7d0: 68,
    0x7d1: 68,
    0x7d2: 68,
    0x7d3: 68,
    0x7d4: 68,
    0x7d5: 68,
    0x7d6: 68,
    0x7d7: 68,
    0x7d8: 68,
    0x7d9: 68,
    0x7da: 68,
    0x7db: 68,
    0x7dc: 68,
    0x7dd: 68,
    0x7de: 68,
    0x7df: 68,
    0x7e0: 68,
    0x7e1: 68,
    0x7e2: 68,
    0x7e3: 68,
    0x7e4: 68,
    0x7e5: 68,
    0x7e6: 68,
    0x7e7: 68,
    0x7e8: 68,
    0x7e9: 68,
    0x7ea: 68,
    0x7fa: 67,
    0x840: 82,
    0x841: 68,
    0x842: 68,
    0x843: 68,
    0x844: 68,
    0x845: 68,
    0x846: 82,
    0x847: 82,
    0x848: 68,
    0x849: 82,
    0x84a: 68,
    0x84b: 68,
    0x84c: 68,
    0x84d: 68,
    0x84e: 68,
    0x84f: 68,
    0x850: 68,
    0x851: 68,
    0x852: 68,
    0x853: 68,
    0x854: 82,
    0x855: 68,
    0x856: 85,
    0x857: 85,
    0x858: 85,
    0x860: 68,
    0x861: 85,
    0x862: 68,
    0x863: 68,
    0x864: 68,
    0x865: 68,
    0x866: 85,
    0x867: 82,
    0x868: 68,
    0x869: 82,
    0x86a: 82,
    0x8a0: 68,
    0x8a1: 68,
    0x8a2: 68,
    0x8a3: 68,
    0x8a4: 68,
    0x8a5: 68,
    0x8a6: 68,
    0x8a7: 68,
    0x8a8: 68,
    0x8a9: 68,
    0x8aa: 82,
    0x8ab: 82,
    0x8ac: 82,
    0x8ad: 85,
    0x8ae: 82,
    0x8af: 68,
    0x8b0: 68,
    0x8b1: 82,
    0x8b2: 82,
    0x8b3: 68,
    0x8b4: 68,
    0x8b6: 68,
    0x8b7: 68,
    0x8b8: 68,
    0x8b9: 82,
    0x8ba: 68,
    0x8bb: 68,
    0x8bc: 68,
    0x8bd: 68,
    0x8e2: 85,
    0x1806: 85,
    0x1807: 68,
    0x180a: 67,
    0x180e: 85,
    0x1820: 68,
    0x1821: 68,
    0x1822: 68,
    0x1823: 68,
    0x1824: 68,
    0x1825: 68,
    0x1826: 68,
    0x1827: 68,
    0x1828: 68,
    0x1829: 68,
    0x182a: 68,
    0x182b: 68,
    0x182c: 68,
    0x182d: 68,
    0x182e: 68,
    0x182f: 68,
    0x1830: 68,
    0x1831: 68,
    0x1832: 68,
    0x1833: 68,
    0x1834: 68,
    0x1835: 68,
    0x1836: 68,
    0x1837: 68,
    0x1838: 68,
    0x1839: 68,
    0x183a: 68,
    0x183b: 68,
    0x183c: 68,
    0x183d: 68,
    0x183e: 68,
    0x183f: 68,
    0x1840: 68,
    0x1841: 68,
    0x1842: 68,
    0x1843: 68,
    0x1844: 68,
    0x1845: 68,
    0x1846: 68,
    0x1847: 68,
    0x1848: 68,
    0x1849: 68,
    0x184a: 68,
    0x184b: 68,
    0x184c: 68,
    0x184d: 68,
    0x184e: 68,
    0x184f: 68,
    0x1850: 68,
    0x1851: 68,
    0x1852: 68,
    0x1853: 68,
    0x1854: 68,
    0x1855: 68,
    0x1856: 68,
    0x1857: 68,
    0x1858: 68,
    0x1859: 68,
    0x185a: 68,
    0x185b: 68,
    0x185c: 68,
    0x185d: 68,
    0x185e: 68,
    0x185f: 68,
    0x1860: 68,
    0x1861: 68,
    0x1862: 68,
    0x1863: 68,
    0x1864: 68,
    0x1865: 68,
    0x1866: 68,
    0x1867: 68,
    0x1868: 68,
    0x1869: 68,
    0x186a: 68,
    0x186b: 68,
    0x186c: 68,
    0x186d: 68,
    0x186e: 68,
    0x186f: 68,
    0x1870: 68,
    0x1871: 68,
    0x1872: 68,
    0x1873: 68,
    0x1874: 68,
    0x1875: 68,
    0x1876: 68,
    0x1877: 68,
    0x1878: 68,
    0x1880: 85,
    0x1881: 85,
    0x1882: 85,
    0x1883: 85,
    0x1884: 85,
    0x1885: 84,
    0x1886: 84,
    0x1887: 68,
    0x1888: 68,
    0x1889: 68,
    0x188a: 68,
    0x188b: 68,
    0x188c: 68,
    0x188d: 68,
    0x188e: 68,
    0x188f: 68,
    0x1890: 68,
    0x1891: 68,
    0x1892: 68,
    0x1893: 68,
    0x1894: 68,
    0x1895: 68,
    0x1896: 68,
    0x1897: 68,
    0x1898: 68,
    0x1899: 68,
    0x189a: 68,
    0x189b: 68,
    0x189c: 68,
    0x189d: 68,
    0x189e: 68,
    0x189f: 68,
    0x18a0: 68,
    0x18a1: 68,
    0x18a2: 68,
    0x18a3: 68,
    0x18a4: 68,
    0x18a5: 68,
    0x18a6: 68,
    0x18a7: 68,
    0x18a8: 68,
    0x18aa: 68,
    0x200c: 85,
    0x200d: 67,
    0x202f: 85,
    0x2066: 85,
    0x2067: 85,
    0x2068: 85,
    0x2069: 85,
    0xa840: 68,
    0xa841: 68,
    0xa842: 68,
    0xa843: 68,
    0xa844: 68,
    0xa845: 68,
    0xa846: 68,
    0xa847: 68,
    0xa848: 68,
    0xa849: 68,
    0xa84a: 68,
    0xa84b: 68,
    0xa84c: 68,
    0xa84d: 68,
    0xa84e: 68,
    0xa84f: 68,
    0xa850: 68,
    0xa851: 68,
    0xa852: 68,
    0xa853: 68,
    0xa854: 68,
    0xa855: 68,
    0xa856: 68,
    0xa857: 68,
    0xa858: 68,
    0xa859: 68,
    0xa85a: 68,
    0xa85b: 68,
    0xa85c: 68,
    0xa85d: 68,
    0xa85e: 68,
    0xa85f: 68,
    0xa860: 68,
    0xa861: 68,
    0xa862: 68,
    0xa863: 68,
    0xa864: 68,
    0xa865: 68,
    0xa866: 68,
    0xa867: 68,
    0xa868: 68,
    0xa869: 68,
    0xa86a: 68,
    0xa86b: 68,
    0xa86c: 68,
    0xa86d: 68,
    0xa86e: 68,
    0xa86f: 68,
    0xa870: 68,
    0xa871: 68,
    0xa872: 76,
    0xa873: 85,
    0x10ac0: 68,
    0x10ac1: 68,
    0x10ac2: 68,
    0x10ac3: 68,
    0x10ac4: 68,
    0x10ac5: 82,
    0x10ac6: 85,
    0x10ac7: 82,
    0x10ac8: 85,
    0x10ac9: 82,
    0x10aca: 82,
    0x10acb: 85,
    0x10acc: 85,
    0x10acd: 76,
    0x10ace: 82,
    0x10acf: 82,
    0x10ad0: 82,
    0x10ad1: 82,
    0x10ad2: 82,
    0x10ad3: 68,
    0x10ad4: 68,
    0x10ad5: 68,
    0x10ad6: 68,
    0x10ad7: 76,
    0x10ad8: 68,
    0x10ad9: 68,
    0x10ada: 68,
    0x10adb: 68,
    0x10adc: 68,
    0x10add: 82,
    0x10ade: 68,
    0x10adf: 68,
    0x10ae0: 68,
    0x10ae1: 82,
    0x10ae2: 85,
    0x10ae3: 85,
    0x10ae4: 82,
    0x10aeb: 68,
    0x10aec: 68,
    0x10aed: 68,
    0x10aee: 68,
    0x10aef: 82,
    0x10b80: 68,
    0x10b81: 82,
    0x10b82: 68,
    0x10b83: 82,
    0x10b84: 82,
    0x10b85: 82,
    0x10b86: 68,
    0x10b87: 68,
    0x10b88: 68,
    0x10b89: 82,
    0x10b8a: 68,
    0x10b8b: 68,
    0x10b8c: 82,
    0x10b8d: 68,
    0x10b8e: 82,
    0x10b8f: 82,
    0x10b90: 68,
    0x10b91: 82,
    0x10ba9: 82,
    0x10baa: 82,
    0x10bab: 82,
    0x10bac: 82,
    0x10bad: 68,
    0x10bae: 68,
    0x10baf: 85,
    0x10d00: 76,
    0x10d01: 68,
    0x10d02: 68,
    0x10d03: 68,
    0x10d04: 68,
    0x10d05: 68,
    0x10d06: 68,
    0x10d07: 68,
    0x10d08: 68,
    0x10d09: 68,
    0x10d0a: 68,
    0x10d0b: 68,
    0x10d0c: 68,
    0x10d0d: 68,
    0x10d0e: 68,
    0x10d0f: 68,
    0x10d10: 68,
    0x10d11: 68,
    0x10d12: 68,
    0x10d13: 68,
    0x10d14: 68,
    0x10d15: 68,
    0x10d16: 68,
    0x10d17: 68,
    0x10d18: 68,
    0x10d19: 68,
    0x10d1a: 68,
    0x10d1b: 68,
    0x10d1c: 68,
    0x10d1d: 68,
    0x10d1e: 68,
    0x10d1f: 68,
    0x10d20: 68,
    0x10d21: 68,
    0x10d22: 82,
    0x10d23: 68,
    0x10f30: 68,
    0x10f31: 68,
    0x10f32: 68,
    0x10f33: 82,
    0x10f34: 68,
    0x10f35: 68,
    0x10f36: 68,
    0x10f37: 68,
    0x10f38: 68,
    0x10f39: 68,
    0x10f3a: 68,
    0x10f3b: 68,
    0x10f3c: 68,
    0x10f3d: 68,
    0x10f3e: 68,
    0x10f3f: 68,
    0x10f40: 68,
    0x10f41: 68,
    0x10f42: 68,
    0x10f43: 68,
    0x10f44: 68,
    0x10f45: 85,
    0x10f51: 68,
    0x10f52: 68,
    0x10f53: 68,
    0x10f54: 82,
    0x110bd: 85,
    0x110cd: 85,
    0x1e900: 68,
    0x1e901: 68,
    0x1e902: 68,
    0x1e903: 68,
    0x1e904: 68,
    0x1e905: 68,
    0x1e906: 68,
    0x1e907: 68,
    0x1e908: 68,
    0x1e909: 68,
    0x1e90a: 68,
    0x1e90b: 68,
    0x1e90c: 68,
    0x1e90d: 68,
    0x1e90e: 68,
    0x1e90f: 68,
    0x1e910: 68,
    0x1e911: 68,
    0x1e912: 68,
    0x1e913: 68,
    0x1e914: 68,
    0x1e915: 68,
    0x1e916: 68,
    0x1e917: 68,
    0x1e918: 68,
    0x1e919: 68,
    0x1e91a: 68,
    0x1e91b: 68,
    0x1e91c: 68,
    0x1e91d: 68,
    0x1e91e: 68,
    0x1e91f: 68,
    0x1e920: 68,
    0x1e921: 68,
    0x1e922: 68,
    0x1e923: 68,
    0x1e924: 68,
    0x1e925: 68,
    0x1e926: 68,
    0x1e927: 68,
    0x1e928: 68,
    0x1e929: 68,
    0x1e92a: 68,
    0x1e92b: 68,
    0x1e92c: 68,
    0x1e92d: 68,
    0x1e92e: 68,
    0x1e92f: 68,
    0x1e930: 68,
    0x1e931: 68,
    0x1e932: 68,
    0x1e933: 68,
    0x1e934: 68,
    0x1e935: 68,
    0x1e936: 68,
    0x1e937: 68,
    0x1e938: 68,
    0x1e939: 68,
    0x1e93a: 68,
    0x1e93b: 68,
    0x1e93c: 68,
    0x1e93d: 68,
    0x1e93e: 68,
    0x1e93f: 68,
    0x1e940: 68,
    0x1e941: 68,
    0x1e942: 68,
    0x1e943: 68,
}
codepoint_classes = {
    'PVALID': (
        0x2d0000002e,
        0x300000003a,
        0x610000007b,
        0xdf000000f7,
        0xf800000100,
        0x10100000102,
        0x10300000104,
        0x10500000106,
        0x10700000108,
        0x1090000010a,
        0x10b0000010c,
        0x10d0000010e,
        0x10f00000110,
        0x11100000112,
        0x11300000114,
        0x11500000116,
        0x11700000118,
        0x1190000011a,
        0x11b0000011c,
        0x11d0000011e,
        0x11f00000120,
        0x12100000122,
        0x12300000124,
        0x12500000126,
        0x12700000128,
        0x1290000012a,
        0x12b0000012c,
        0x12d0000012e,
        0x12f00000130,
        0x13100000132,
        0x13500000136,
        0x13700000139,
        0x13a0000013b,
        0x13c0000013d,
        0x13e0000013f,
        0x14200000143,
        0x14400000145,
        0x14600000147,
        0x14800000149,
        0x14b0000014c,
        0x14d0000014e,
        0x14f00000150,
        0x15100000152,
        0x15300000154,
        0x15500000156,
        0x15700000158,
        0x1590000015a,
        0x15b0000015c,
        0x15d0000015e,
        0x15f00000160,
        0x16100000162,
        0x16300000164,
        0x16500000166,
        0x16700000168,
        0x1690000016a,
        0x16b0000016c,
        0x16d0000016e,
        0x16f00000170,
        0x17100000172,
        0x17300000174,
        0x17500000176,
        0x17700000178,
        0x17a0000017b,
        0x17c0000017d,
        0x17e0000017f,
        0x18000000181,
        0x18300000184,
        0x18500000186,
        0x18800000189,
        0x18c0000018e,
        0x19200000193,
        0x19500000196,
        0x1990000019c,
        0x19e0000019f,
        0x1a1000001a2,
        0x1a3000001a4,
        0x1a5000001a6,
        0x1a8000001a9,
        0x1aa000001ac,
        0x1ad000001ae,
        0x1b0000001b1,
        0x1b4000001b5,
        0x1b6000001b7,
        0x1b9000001bc,
        0x1bd000001c4,
        0x1ce000001cf,
        0x1d0000001d1,
        0x1d2000001d3,
        0x1d4000001d5,
        0x1d6000001d7,
        0x1d8000001d9,
        0x1da000001db,
        0x1dc000001de,
        0x1df000001e0,
        0x1e1000001e2,
        0x1e3000001e4,
        0x1e5000001e6,
        0x1e7000001e8,
        0x1e9000001ea,
        0x1eb000001ec,
        0x1ed000001ee,
        0x1ef000001f1,
        0x1f5000001f6,
        0x1f9000001fa,
        0x1fb000001fc,
        0x1fd000001fe,
        0x1ff00000200,
        0x20100000202,
        0x20300000204,
        0x20500000206,
        0x20700000208,
        0x2090000020a,
        0x20b0000020c,
        0x20d0000020e,
        0x20f00000210,
        0x21100000212,
        0x21300000214,
        0x21500000216,
        0x21700000218,
        0x2190000021a,
        0x21b0000021c,
        0x21d0000021e,
        0x21f00000220,
        0x22100000222,
        0x22300000224,
        0x22500000226,
        0x22700000228,
        0x2290000022a,
        0x22b0000022c,
        0x22d0000022e,
        0x22f00000230,
        0x23100000232,
        0x2330000023a,
        0x23c0000023d,
        0x23f00000241,
        0x24200000243,
        0x24700000248,
        0x2490000024a,
        0x24b0000024c,
        0x24d0000024e,
        0x24f000002b0,
        0x2b9000002c2,
        0x2c6000002d2,
        0x2ec000002ed,
        0x2ee000002ef,
        0x30000000340,
        0x34200000343,
        0x3460000034f,
        0x35000000370,
        0x37100000372,
        0x37300000374,
        0x37700000378,
        0x37b0000037e,
        0x39000000391,
        0x3ac000003cf,
        0x3d7000003d8,
        0x3d9000003da,
        0x3db000003dc,
        0x3dd000003de,
        0x3df000003e0,
        0x3e1000003e2,
        0x3e3000003e4,
        0x3e5000003e6,
        0x3e7000003e8,
        0x3e9000003ea,
        0x3eb000003ec,
        0x3ed000003ee,
        0x3ef000003f0,
        0x3f3000003f4,
        0x3f8000003f9,
        0x3fb000003fd,
        0x43000000460,
        0x46100000462,
        0x46300000464,
        0x46500000466,
        0x46700000468,
        0x4690000046a,
        0x46b0000046c,
        0x46d0000046e,
        0x46f00000470,
        0x47100000472,
        0x47300000474,
        0x47500000476,
        0x47700000478,
        0x4790000047a,
        0x47b0000047c,
        0x47d0000047e,
        0x47f00000480,
        0x48100000482,
        0x48300000488,
        0x48b0000048c,
        0x48d0000048e,
        0x48f00000490,
        0x49100000492,
        0x49300000494,
        0x49500000496,
        0x49700000498,
        0x4990000049a,
        0x49b0000049c,
        0x49d0000049e,
        0x49f000004a0,
        0x4a1000004a2,
        0x4a3000004a4,
        0x4a5000004a6,
        0x4a7000004a8,
        0x4a9000004aa,
        0x4ab000004ac,
        0x4ad000004ae,
        0x4af000004b0,
        0x4b1000004b2,
        0x4b3000004b4,
        0x4b5000004b6,
        0x4b7000004b8,
        0x4b9000004ba,
        0x4bb000004bc,
        0x4bd000004be,
        0x4bf000004c0,
        0x4c2000004c3,
        0x4c4000004c5,
        0x4c6000004c7,
        0x4c8000004c9,
        0x4ca000004cb,
        0x4cc000004cd,
        0x4ce000004d0,
        0x4d1000004d2,
        0x4d3000004d4,
        0x4d5000004d6,
        0x4d7000004d8,
        0x4d9000004da,
        0x4db000004dc,
        0x4dd000004de,
        0x4df000004e0,
        0x4e1000004e2,
        0x4e3000004e4,
        0x4e5000004e6,
        0x4e7000004e8,
        0x4e9000004ea,
        0x4eb000004ec,
        0x4ed000004ee,
        0x4ef000004f0,
        0x4f1000004f2,
        0x4f3000004f4,
        0x4f5000004f6,
        0x4f7000004f8,
        0x4f9000004fa,
        0x4fb000004fc,
        0x4fd000004fe,
        0x4ff00000500,
        0x50100000502,
        0x50300000504,
        0x50500000506,
        0x50700000508,
        0x5090000050a,
        0x50b0000050c,
        0x50d0000050e,
        0x50f00000510,
        0x51100000512,
        0x51300000514,
        0x51500000516,
        0x51700000518,
        0x5190000051a,
        0x51b0000051c,
        0x51d0000051e,
        0x51f00000520,
        0x52100000522,
        0x52300000524,
        0x52500000526,
        0x52700000528,
        0x5290000052a,
        0x52b0000052c,
        0x52d0000052e,
        0x52f00000530,
        0x5590000055a,
        0x56000000587,
        0x58800000589,
        0x591000005be,
        0x5bf000005c0,
        0x5c1000005c3,
        0x5c4000005c6,
        0x5c7000005c8,
        0x5d0000005eb,
        0x5ef000005f3,
        0x6100000061b,
        0x62000000640,
        0x64100000660,
        0x66e00000675,
        0x679000006d4,
        0x6d5000006dd,
        0x6df000006e9,
        0x6ea000006f0,
        0x6fa00000700,
        0x7100000074b,
        0x74d000007b2,
        0x7c0000007f6,
        0x7fd000007fe,
        0x8000000082e,
        0x8400000085c,
        0x8600000086b,
        0x8a0000008b5,
        0x8b6000008be,
        0x8d3000008e2,
        0x8e300000958,
        0x96000000964,
        0x96600000970,
        0x97100000984,
        0x9850000098d,
        0x98f00000991,
        0x993000009a9,
        0x9aa000009b1,
        0x9b2000009b3,
        0x9b6000009ba,
        0x9bc000009c5,
        0x9c7000009c9,
        0x9cb000009cf,
        0x9d7000009d8,
        0x9e0000009e4,
        0x9e6000009f2,
        0x9fc000009fd,
        0x9fe000009ff,
        0xa0100000a04,
        0xa0500000a0b,
        0xa0f00000a11,
        0xa1300000a29,
        0xa2a00000a31,
        0xa3200000a33,
        0xa3500000a36,
        0xa3800000a3a,
        0xa3c00000a3d,
        0xa3e00000a43,
        0xa4700000a49,
        0xa4b00000a4e,
        0xa5100000a52,
        0xa5c00000a5d,
        0xa6600000a76,
        0xa8100000a84,
        0xa8500000a8e,
        0xa8f00000a92,
        0xa9300000aa9,
        0xaaa00000ab1,
        0xab200000ab4,
        0xab500000aba,
        0xabc00000ac6,
        0xac700000aca,
        0xacb00000ace,
        0xad000000ad1,
        0xae000000ae4,
        0xae600000af0,
        0xaf900000b00,
        0xb0100000b04,
        0xb0500000b0d,
        0xb0f00000b11,
        0xb1300000b29,
        0xb2a00000b31,
        0xb3200000b34,
        0xb3500000b3a,
        0xb3c00000b45,
        0xb4700000b49,
        0xb4b00000b4e,
        0xb5600000b58,
        0xb5f00000b64,
        0xb6600000b70,
        0xb7100000b72,
        0xb8200000b84,
        0xb8500000b8b,
        0xb8e00000b91,
        0xb9200000b96,
        0xb9900000b9b,
        0xb9c00000b9d,
        0xb9e00000ba0,
        0xba300000ba5,
        0xba800000bab,
        0xbae00000bba,
        0xbbe00000bc3,
        0xbc600000bc9,
        0xbca00000bce,
        0xbd000000bd1,
        0xbd700000bd8,
        0xbe600000bf0,
        0xc0000000c0d,
        0xc0e00000c11,
        0xc1200000c29,
        0xc2a00000c3a,
        0xc3d00000c45,
        0xc4600000c49,
        0xc4a00000c4e,
        0xc5500000c57,
        0xc5800000c5b,
        0xc6000000c64,
        0xc6600000c70,
        0xc8000000c84,
        0xc8500000c8d,
        0xc8e00000c91,
        0xc9200000ca9,
        0xcaa00000cb4,
        0xcb500000cba,
        0xcbc00000cc5,
        0xcc600000cc9,
        0xcca00000cce,
        0xcd500000cd7,
        0xcde00000cdf,
        0xce000000ce4,
        0xce600000cf0,
        0xcf100000cf3,
        0xd0000000d04,
        0xd0500000d0d,
        0xd0e00000d11,
        0xd1200000d45,
        0xd4600000d49,
        0xd4a00000d4f,
        0xd5400000d58,
        0xd5f00000d64,
        0xd6600000d70,
        0xd7a00000d80,
        0xd8200000d84,
        0xd8500000d97,
        0xd9a00000db2,
        0xdb300000dbc,
        0xdbd00000dbe,
        0xdc000000dc7,
        0xdca00000dcb,
        0xdcf00000dd5,
        0xdd600000dd7,
        0xdd800000de0,
        0xde600000df0,
        0xdf200000df4,
        0xe0100000e33,
        0xe3400000e3b,
        0xe4000000e4f,
        0xe5000000e5a,
        0xe8100000e83,
        0xe8400000e85,
        0xe8700000e89,
        0xe8a00000e8b,
        0xe8d00000e8e,
        0xe9400000e98,
        0xe9900000ea0,
        0xea100000ea4,
        0xea500000ea6,
        0xea700000ea8,
        0xeaa00000eac,
        0xead00000eb3,
        0xeb400000eba,
        0xebb00000ebe,
        0xec000000ec5,
        0xec600000ec7,
        0xec800000ece,
        0xed000000eda,
        0xede00000ee0,
        0xf0000000f01,
        0xf0b00000f0c,
        0xf1800000f1a,
        0xf2000000f2a,
        0xf3500000f36,
        0xf3700000f38,
        0xf3900000f3a,
        0xf3e00000f43,
        0xf4400000f48,
        0xf4900000f4d,
        0xf4e00000f52,
        0xf5300000f57,
        0xf5800000f5c,
        0xf5d00000f69,
        0xf6a00000f6d,
        0xf7100000f73,
        0xf7400000f75,
        0xf7a00000f81,
        0xf8200000f85,
        0xf8600000f93,
        0xf9400000f98,
        0xf9900000f9d,
        0xf9e00000fa2,
        0xfa300000fa7,
        0xfa800000fac,
        0xfad00000fb9,
        0xfba00000fbd,
        0xfc600000fc7,
        0x10000000104a,
        0x10500000109e,
        0x10d0000010fb,
        0x10fd00001100,
        0x120000001249,
        0x124a0000124e,
        0x125000001257,
        0x125800001259,
        0x125a0000125e,
        0x126000001289,
        0x128a0000128e,
        0x1290000012b1,
        0x12b2000012b6,
        0x12b8000012bf,
        0x12c0000012c1,
        0x12c2000012c6,
        0x12c8000012d7,
        0x12d800001311,
        0x131200001316,
        0x13180000135b,
        0x135d00001360,
        0x138000001390,
        0x13a0000013f6,
        0x14010000166d,
        0x166f00001680,
        0x16810000169b,
        0x16a0000016eb,
        0x16f1000016f9,
        0x17000000170d,
        0x170e00001715,
        0x172000001735,
        0x174000001754,
        0x17600000176d,
        0x176e00001771,
        0x177200001774,
        0x1780000017b4,
        0x17b6000017d4,
        0x17d7000017d8,
        0x17dc000017de,
        0x17e0000017ea,
        0x18100000181a,
        0x182000001879,
        0x1880000018ab,
        0x18b0000018f6,
        0x19000000191f,
        0x19200000192c,
        0x19300000193c,
        0x19460000196e,
        0x197000001975,
        0x1980000019ac,
        0x19b0000019ca,
        0x19d0000019da,
        0x1a0000001a1c,
        0x1a2000001a5f,
        0x1a6000001a7d,
        0x1a7f00001a8a,
        0x1a9000001a9a,
        0x1aa700001aa8,
        0x1ab000001abe,
        0x1b0000001b4c,
        0x1b5000001b5a,
        0x1b6b00001b74,
        0x1b8000001bf4,
        0x1c0000001c38,
        0x1c4000001c4a,
        0x1c4d00001c7e,
        0x1cd000001cd3,
        0x1cd400001cfa,
        0x1d0000001d2c,
        0x1d2f00001d30,
        0x1d3b00001d3c,
        0x1d4e00001d4f,
        0x1d6b00001d78,
        0x1d7900001d9b,
        0x1dc000001dfa,
        0x1dfb00001e00,
        0x1e0100001e02,
        0x1e0300001e04,
        0x1e0500001e06,
        0x1e0700001e08,
        0x1e0900001e0a,
        0x1e0b00001e0c,
        0x1e0d00001e0e,
        0x1e0f00001e10,
        0x1e1100001e12,
        0x1e1300001e14,
        0x1e1500001e16,
        0x1e1700001e18,
        0x1e1900001e1a,
        0x1e1b00001e1c,
        0x1e1d00001e1e,
        0x1e1f00001e20,
        0x1e2100001e22,
        0x1e2300001e24,
        0x1e2500001e26,
        0x1e2700001e28,
        0x1e2900001e2a,
        0x1e2b00001e2c,
        0x1e2d00001e2e,
        0x1e2f00001e30,
        0x1e3100001e32,
        0x1e3300001e34,
        0x1e3500001e36,
        0x1e3700001e38,
        0x1e3900001e3a,
        0x1e3b00001e3c,
        0x1e3d00001e3e,
        0x1e3f00001e40,
        0x1e4100001e42,
        0x1e4300001e44,
        0x1e4500001e46,
        0x1e4700001e48,
        0x1e4900001e4a,
        0x1e4b00001e4c,
        0x1e4d00001e4e,
        0x1e4f00001e50,
        0x1e5100001e52,
        0x1e5300001e54,
        0x1e5500001e56,
        0x1e5700001e58,
        0x1e5900001e5a,
        0x1e5b00001e5c,
        0x1e5d00001e5e,
        0x1e5f00001e60,
        0x1e6100001e62,
        0x1e6300001e64,
        0x1e6500001e66,
        0x1e6700001e68,
        0x1e6900001e6a,
        0x1e6b00001e6c,
        0x1e6d00001e6e,
        0x1e6f00001e70,
        0x1e7100001e72,
        0x1e7300001e74,
        0x1e7500001e76,
        0x1e7700001e78,
        0x1e7900001e7a,
        0x1e7b00001e7c,
        0x1e7d00001e7e,
        0x1e7f00001e80,
        0x1e8100001e82,
        0x1e8300001e84,
        0x1e8500001e86,
        0x1e8700001e88,
        0x1e8900001e8a,
        0x1e8b00001e8c,
        0x1e8d00001e8e,
        0x1e8f00001e90,
        0x1e9100001e92,
        0x1e9300001e94,
        0x1e9500001e9a,
        0x1e9c00001e9e,
        0x1e9f00001ea0,
        0x1ea100001ea2,
        0x1ea300001ea4,
        0x1ea500001ea6,
        0x1ea700001ea8,
        0x1ea900001eaa,
        0x1eab00001eac,
        0x1ead00001eae,
        0x1eaf00001eb0,
        0x1eb100001eb2,
        0x1eb300001eb4,
        0x1eb500001eb6,
        0x1eb700001eb8,
        0x1eb900001eba,
        0x1ebb00001ebc,
        0x1ebd00001ebe,
        0x1ebf00001ec0,
        0x1ec100001ec2,
        0x1ec300001ec4,
        0x1ec500001ec6,
        0x1ec700001ec8,
        0x1ec900001eca,
        0x1ecb00001ecc,
        0x1ecd00001ece,
        0x1ecf00001ed0,
        0x1ed100001ed2,
        0x1ed300001ed4,
        0x1ed500001ed6,
        0x1ed700001ed8,
        0x1ed900001eda,
        0x1edb00001edc,
        0x1edd00001ede,
        0x1edf00001ee0,
        0x1ee100001ee2,
        0x1ee300001ee4,
        0x1ee500001ee6,
        0x1ee700001ee8,
        0x1ee900001eea,
        0x1eeb00001eec,
        0x1eed00001eee,
        0x1eef00001ef0,
        0x1ef100001ef2,
        0x1ef300001ef4,
        0x1ef500001ef6,
        0x1ef700001ef8,
        0x1ef900001efa,
        0x1efb00001efc,
        0x1efd00001efe,
        0x1eff00001f08,
        0x1f1000001f16,
        0x1f2000001f28,
        0x1f3000001f38,
        0x1f4000001f46,
        0x1f5000001f58,
        0x1f6000001f68,
        0x1f7000001f71,
        0x1f7200001f73,
        0x1f7400001f75,
        0x1f7600001f77,
        0x1f7800001f79,
        0x1f7a00001f7b,
        0x1f7c00001f7d,
        0x1fb000001fb2,
        0x1fb600001fb7,
        0x1fc600001fc7,
        0x1fd000001fd3,
        0x1fd600001fd8,
        0x1fe000001fe3,
        0x1fe400001fe8,
        0x1ff600001ff7,
        0x214e0000214f,
        0x218400002185,
        0x2c3000002c5f,
        0x2c6100002c62,
        0x2c6500002c67,
        0x2c6800002c69,
        0x2c6a00002c6b,
        0x2c6c00002c6d,
        0x2c7100002c72,
        0x2c7300002c75,
        0x2c7600002c7c,
        0x2c8100002c82,
        0x2c8300002c84,
        0x2c8500002c86,
        0x2c8700002c88,
        0x2c8900002c8a,
        0x2c8b00002c8c,
        0x2c8d00002c8e,
        0x2c8f00002c90,
        0x2c9100002c92,
        0x2c9300002c94,
        0x2c9500002c96,
        0x2c9700002c98,
        0x2c9900002c9a,
        0x2c9b00002c9c,
        0x2c9d00002c9e,
        0x2c9f00002ca0,
        0x2ca100002ca2,
        0x2ca300002ca4,
        0x2ca500002ca6,
        0x2ca700002ca8,
        0x2ca900002caa,
        0x2cab00002cac,
        0x2cad00002cae,
        0x2caf00002cb0,
        0x2cb100002cb2,
        0x2cb300002cb4,
        0x2cb500002cb6,
        0x2cb700002cb8,
        0x2cb900002cba,
        0x2cbb00002cbc,
        0x2cbd00002cbe,
        0x2cbf00002cc0,
        0x2cc100002cc2,
        0x2cc300002cc4,
        0x2cc500002cc6,
        0x2cc700002cc8,
        0x2cc900002cca,
        0x2ccb00002ccc,
        0x2ccd00002cce,
        0x2ccf00002cd0,
        0x2cd100002cd2,
        0x2cd300002cd4,
        0x2cd500002cd6,
        0x2cd700002cd8,
        0x2cd900002cda,
        0x2cdb00002cdc,
        0x2cdd00002cde,
        0x2cdf00002ce0,
        0x2ce100002ce2,
        0x2ce300002ce5,
        0x2cec00002ced,
        0x2cee00002cf2,
        0x2cf300002cf4,
        0x2d0000002d26,
        0x2d2700002d28,
        0x2d2d00002d2e,
        0x2d3000002d68,
        0x2d7f00002d97,
        0x2da000002da7,
        0x2da800002daf,
        0x2db000002db7,
        0x2db800002dbf,
        0x2dc000002dc7,
        0x2dc800002dcf,
        0x2dd000002dd7,
        0x2dd800002ddf,
        0x2de000002e00,
        0x2e2f00002e30,
        0x300500003008,
        0x302a0000302e,
        0x303c0000303d,
        0x304100003097,
        0x30990000309b,
        0x309d0000309f,
        0x30a1000030fb,
        0x30fc000030ff,
        0x310500003130,
        0x31a0000031bb,
        0x31f000003200,
        0x340000004db6,
        0x4e0000009ff0,
        0xa0000000a48d,
        0xa4d00000a4fe,
        0xa5000000a60d,
        0xa6100000a62c,
        0xa6410000a642,
        0xa6430000a644,
        0xa6450000a646,
        0xa6470000a648,
        0xa6490000a64a,
        0xa64b0000a64c,
        0xa64d0000a64e,
        0xa64f0000a650,
        0xa6510000a652,
        0xa6530000a654,
        0xa6550000a656,
        0xa6570000a658,
        0xa6590000a65a,
        0xa65b0000a65c,
        0xa65d0000a65e,
        0xa65f0000a660,
        0xa6610000a662,
        0xa6630000a664,
        0xa6650000a666,
        0xa6670000a668,
        0xa6690000a66a,
        0xa66b0000a66c,
        0xa66d0000a670,
        0xa6740000a67e,
        0xa67f0000a680,
        0xa6810000a682,
        0xa6830000a684,
        0xa6850000a686,
        0xa6870000a688,
        0xa6890000a68a,
        0xa68b0000a68c,
        0xa68d0000a68e,
        0xa68f0000a690,
        0xa6910000a692,
        0xa6930000a694,
        0xa6950000a696,
        0xa6970000a698,
        0xa6990000a69a,
        0xa69b0000a69c,
        0xa69e0000a6e6,
        0xa6f00000a6f2,
        0xa7170000a720,
        0xa7230000a724,
        0xa7250000a726,
        0xa7270000a728,
        0xa7290000a72a,
        0xa72b0000a72c,
        0xa72d0000a72e,
        0xa72f0000a732,
        0xa7330000a734,
        0xa7350000a736,
        0xa7370000a738,
        0xa7390000a73a,
        0xa73b0000a73c,
        0xa73d0000a73e,
        0xa73f0000a740,
        0xa7410000a742,
        0xa7430000a744,
        0xa7450000a746,
        0xa7470000a748,
        0xa7490000a74a,
        0xa74b0000a74c,
        0xa74d0000a74e,
        0xa74f0000a750,
        0xa7510000a752,
        0xa7530000a754,
        0xa7550000a756,
        0xa7570000a758,
        0xa7590000a75a,
        0xa75b0000a75c,
        0xa75d0000a75e,
        0xa75f0000a760,
        0xa7610000a762,
        0xa7630000a764,
        0xa7650000a766,
        0xa7670000a768,
        0xa7690000a76a,
        0xa76b0000a76c,
        0xa76d0000a76e,
        0xa76f0000a770,
        0xa7710000a779,
        0xa77a0000a77b,
        0xa77c0000a77d,
        0xa77f0000a780,
        0xa7810000a782,
        0xa7830000a784,
        0xa7850000a786,
        0xa7870000a789,
        0xa78c0000a78d,
        0xa78e0000a790,
        0xa7910000a792,
        0xa7930000a796,
        0xa7970000a798,
        0xa7990000a79a,
        0xa79b0000a79c,
        0xa79d0000a79e,
        0xa79f0000a7a0,
        0xa7a10000a7a2,
        0xa7a30000a7a4,
        0xa7a50000a7a6,
        0xa7a70000a7a8,
        0xa7a90000a7aa,
        0xa7af0000a7b0,
        0xa7b50000a7b6,
        0xa7b70000a7b8,
        0xa7b90000a7ba,
        0xa7f70000a7f8,
        0xa7fa0000a828,
        0xa8400000a874,
        0xa8800000a8c6,
        0xa8d00000a8da,
        0xa8e00000a8f8,
        0xa8fb0000a8fc,
        0xa8fd0000a92e,
        0xa9300000a954,
        0xa9800000a9c1,
        0xa9cf0000a9da,
        0xa9e00000a9ff,
        0xaa000000aa37,
        0xaa400000aa4e,
        0xaa500000aa5a,
        0xaa600000aa77,
        0xaa7a0000aac3,
        0xaadb0000aade,
        0xaae00000aaf0,
        0xaaf20000aaf7,
        0xab010000ab07,
        0xab090000ab0f,
        0xab110000ab17,
        0xab200000ab27,
        0xab280000ab2f,
        0xab300000ab5b,
        0xab600000ab66,
        0xabc00000abeb,
        0xabec0000abee,
        0xabf00000abfa,
        0xac000000d7a4,
        0xfa0e0000fa10,
        0xfa110000fa12,
        0xfa130000fa15,
        0xfa1f0000fa20,
        0xfa210000fa22,
        0xfa230000fa25,
        0xfa270000fa2a,
        0xfb1e0000fb1f,
        0xfe200000fe30,
        0xfe730000fe74,
        0x100000001000c,
        0x1000d00010027,
        0x100280001003b,
        0x1003c0001003e,
        0x1003f0001004e,
        0x100500001005e,
        0x10080000100fb,
        0x101fd000101fe,
        0x102800001029d,
        0x102a0000102d1,
        0x102e0000102e1,
        0x1030000010320,
        0x1032d00010341,
        0x103420001034a,
        0x103500001037b,
        0x103800001039e,
        0x103a0000103c4,
        0x103c8000103d0,
        0x104280001049e,
        0x104a0000104aa,
        0x104d8000104fc,
        0x1050000010528,
        0x1053000010564,
        0x1060000010737,
        0x1074000010756,
        0x1076000010768,
        0x1080000010806,
        0x1080800010809,
        0x1080a00010836,
        0x1083700010839,
        0x1083c0001083d,
        0x1083f00010856,
        0x1086000010877,
        0x108800001089f,
        0x108e0000108f3,
        0x108f4000108f6,
        0x1090000010916,
        0x109200001093a,
        0x10980000109b8,
        0x109be000109c0,
        0x10a0000010a04,
        0x10a0500010a07,
        0x10a0c00010a14,
        0x10a1500010a18,
        0x10a1900010a36,
        0x10a3800010a3b,
        0x10a3f00010a40,
        0x10a6000010a7d,
        0x10a8000010a9d,
        0x10ac000010ac8,
        0x10ac900010ae7,
        0x10b0000010b36,
        0x10b4000010b56,
        0x10b6000010b73,
        0x10b8000010b92,
        0x10c0000010c49,
        0x10cc000010cf3,
        0x10d0000010d28,
        0x10d3000010d3a,
        0x10f0000010f1d,
        0x10f2700010f28,
        0x10f3000010f51,
        0x1100000011047,
        0x1106600011070,
        0x1107f000110bb,
        0x110d0000110e9,
        0x110f0000110fa,
        0x1110000011135,
        0x1113600011140,
        0x1114400011147,
        0x1115000011174,
        0x1117600011177,
        0x11180000111c5,
        0x111c9000111cd,
        0x111d0000111db,
        0x111dc000111dd,
        0x1120000011212,
        0x1121300011238,
        0x1123e0001123f,
        0x1128000011287,
        0x1128800011289,
        0x1128a0001128e,
        0x1128f0001129e,
        0x1129f000112a9,
        0x112b0000112eb,
        0x112f0000112fa,
        0x1130000011304,
        0x113050001130d,
        0x1130f00011311,
        0x1131300011329,
        0x1132a00011331,
        0x1133200011334,
        0x113350001133a,
        0x1133b00011345,
        0x1134700011349,
        0x1134b0001134e,
        0x1135000011351,
        0x1135700011358,
        0x1135d00011364,
        0x113660001136d,
        0x1137000011375,
        0x114000001144b,
        0x114500001145a,
        0x1145e0001145f,
        0x11480000114c6,
        0x114c7000114c8,
        0x114d0000114da,
        0x11580000115b6,
        0x115b8000115c1,
        0x115d8000115de,
        0x1160000011641,
        0x1164400011645,
        0x116500001165a,
        0x11680000116b8,
        0x116c0000116ca,
        0x117000001171b,
        0x1171d0001172c,
        0x117300001173a,
        0x118000001183b,
        0x118c0000118ea,
        0x118ff00011900,
        0x11a0000011a3f,
        0x11a4700011a48,
        0x11a5000011a84,
        0x11a8600011a9a,
        0x11a9d00011a9e,
        0x11ac000011af9,
        0x11c0000011c09,
        0x11c0a00011c37,
        0x11c3800011c41,
        0x11c5000011c5a,
        0x11c7200011c90,
        0x11c9200011ca8,
        0x11ca900011cb7,
        0x11d0000011d07,
        0x11d0800011d0a,
        0x11d0b00011d37,
        0x11d3a00011d3b,
        0x11d3c00011d3e,
        0x11d3f00011d48,
        0x11d5000011d5a,
        0x11d6000011d66,
        0x11d6700011d69,
        0x11d6a00011d8f,
        0x11d9000011d92,
        0x11d9300011d99,
        0x11da000011daa,
        0x11ee000011ef7,
        0x120000001239a,
        0x1248000012544,
        0x130000001342f,
        0x1440000014647,
        0x1680000016a39,
        0x16a4000016a5f,
        0x16a6000016a6a,
        0x16ad000016aee,
        0x16af000016af5,
        0x16b0000016b37,
        0x16b4000016b44,
        0x16b5000016b5a,
        0x16b6300016b78,
        0x16b7d00016b90,
        0x16e6000016e80,
        0x16f0000016f45,
        0x16f5000016f7f,
        0x16f8f00016fa0,
        0x16fe000016fe2,
        0x17000000187f2,
        0x1880000018af3,
        0x1b0000001b11f,
        0x1b1700001b2fc,
        0x1bc000001bc6b,
        0x1bc700001bc7d,
        0x1bc800001bc89,
        0x1bc900001bc9a,
        0x1bc9d0001bc9f,
        0x1da000001da37,
        0x1da3b0001da6d,
        0x1da750001da76,
        0x1da840001da85,
        0x1da9b0001daa0,
        0x1daa10001dab0,
        0x1e0000001e007,
        0x1e0080001e019,
        0x1e01b0001e022,
        0x1e0230001e025,
        0x1e0260001e02b,
        0x1e8000001e8c5,
        0x1e8d00001e8d7,
        0x1e9220001e94b,
        0x1e9500001e95a,
        0x200000002a6d7,
        0x2a7000002b735,
        0x2b7400002b81e,
        0x2b8200002cea2,
        0x2ceb00002ebe1,
    ),
    'CONTEXTJ': (
        0x200c0000200e,
    ),
    'CONTEXTO': (
        0xb7000000b8,
        0x37500000376,
        0x5f3000005f5,
        0x6600000066a,
        0x6f0000006fa,
        0x30fb000030fc,
    ),
}
site-packages/pip/_vendor/idna/__pycache__/idnadata.cpython-38.opt-1.pyc000064400000051541151464452710022006 0ustar00U

.eß@sdZddddddZddddddddddd	d	d	d	dd	dd	dddddd	d	d	d	dddddddddddddd
dddddddd	ddddd	d	d	dd	d	d	ddddddddddddddddd	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	ddddddddddddddddddddddddddddddddddddddd	ddd	d	d	d	d	d	d	d	d	dd	dd	ddd	d	d	dd	d	dddddd	dddd	d	d	d	d	ddddd	dddddddddd	dd	dd	ddd	d	dddddddddddd	d	d	dddddddddddddddd	d	ddddd	dd	d	dddd	d	dddddddddddddddddddddddddddddddddddddddd
d	dddddd	d	dd	ddddddddddd	dddddddddddd	dd	d	ddddddddddd	d	d	dd	ddd	d	dddddd	dddddddd
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd
ddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	dd	dd	d	dddd	d	d	d	d	ddddddddddd	dddd	ddd	ddddd	dd	dd	d	d	dddd	ddd	dd	d	dd	d	d	d	d	dddddddddddddddddddddddddddddddddddddd	ddddd	dddddddddddddddddddddd	ddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd
ZddddZdS)z11.0.0)$lt
lx
l~
l
lll l0l8lll+tlbtulkulvl|l`|lF|lN }X@}lZd}l\l}l^t}l~|}l~l~llXltlll'!lf+-lllFRH)l.:l.l:l/<l0@l0@l*0@l<0@MP8lnzdlzi&578]N`k:lz`)	lD@ll7{tll={ll?{llB{mlE{mlP{m)0Al0tBl1@
lrH)0Bl1C2Gl2@KlX3Llp}l}l0@
)ZGreekZHanZHebrewZHiraganaZKatakanaUDRCTL(iiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJinioiqirisitiuiviwixiyizi{i|i}i~iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/iMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiyizi{i|i}i~iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXi`iaibicidieifigihiiijiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii
ii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii i
 i/ if ig ih ii i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisi
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
iiiiiiiiiiiiiiiiiiiiiiiiii
i
i
i
i
i
i
i
i
i	
i

i
i
i

i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i 
i!
i"
i#
i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiQiRiSiTiiiiiiiiiiii	i
iii
iiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiC(ql.l:l{l|llllll
$l,l4l<lDlLlTl\ldllltl |l"l$l&l(l*l,l.l0l2l6l9l;l=l?lClElGlI lL,lN4lP<lRDlTLlVTlX\lZdl\ll^tl`|lbldlflhljlllnlplrltlvlxl{l}lllll l0lHlTldlxllllllllllll8l@lHlPlXl`lhlpl|llllllllllllllllll
$l,l4l<lDlLlTl\ldllltl |l"l$l&l(l*l,l.l0l2l:l=lAlC	lH	lJ$	lL,	lN4	l<	l
llll@lC
lO
lp@
lr
lt
lx
l~
l@ll\ldllltl|llllllllllll`lbldlflhljlllnlplrltlvlxlzl|l~llll,l4l<lDlLlTl\ldllltl|llllllllllllllllllll l(l0l8lDlLlTl\ldllltl|lllllllllllllllllllll
$l,l4l<lDlLlTl\ldllltl |l"l$l&l(l*l,l.l0lZdll lDllllr	ll@l@l`lullTl|lllK@l4lll. l\!lk!l"l"lL#lX	#ld	%lp	%l	%l	&l	<&l	L&l	&l	&l	&l	&l	'l	,'l	\'l	'l	'l	'l	'l
(l
(l
<(l)
L(l1
(l3
(l6
(l:
(l=
(lC
(lI
)lN
,)lR
D)l]
p)lv
)l
*l
*l
<*l
L*l
*l
*l
*l
*l
+l
,+l
@+l
+l
+l+l,l
,l<,l)L,l1,l4,l:,lE,lI-lN,-lXX-ld|-lp-lr-l.l.l8.lH.ld.lp.lx.l.l.l.l.l/l(/l@/l\/l/l
0l80l)H0l:0lE0lI1lN(1lWT1l[`1ld1lp1l2l2l82lH2l2l2l2l3l(3lT3lx3l3l3l3l
4l

4l
84lE
H4lI
5lO
(5lX
P5ld
|5lp
5l
5l
6l
6l
h6l
6l
6l
7l
(7l
<7l
X7l
`7l
7l
7l38l;8lO9lZ@9l:l:l:l(:l4:lP:ld:l:l:l:l:l:l:l:l;l;l ;l@;lx;ll>lP>ld>lx>l>l>l>l>l?lJ@l@Al@ClClIHlN(IlW@IlY`Il^hIlIl(Jl@JlJlJlKlKl Kl`KlHLl[`Ll`tMlNlNlmPlYlZlZl[l
\l8\l5\lT]lm]lq]lt]l^l^l\_lp_l_l@`ly`lblbldl,dl<dlneluelflfl@glhl_hl}ilil@jljljlLllZ@mltmlnl8plJql~4ql@slPsl,tl0tl<tlO8ulxululwlwlxlxlxlxl
$xl,xl4xlxl@xlBylDylFylHylJ$ylL,ylN4ylP'l@'lB'lD'lF'lH'lJ'$lL',lN'4lP'<lR'DlT'LlV'TlX'\lZ'dl\'ll^'tl`'|lb'ld'lf'lh'lj'll'ln'lp'ly'l{'l}'l'l'l'l'l'l'0l'8l'Dl'Ll'\l'dl'll'tl'|l'l'l'l'l'l'l'l'l'l'l((lt(!l("l(@#l(#l(#l.)#lT)$l)&l)<'l)'l7*(lN*)lZ*@)lw*)l*)l*l+l*+l*+l+,l+$,l+D,l'+,l/+,l[+,lf+-l+/l+/l+/lW0lz8hlzDhlzLhl z|hl"zhl%zhl*zhl{xll0~xlt~yll'4l;l>lNl^@lll
l
ll lAlJ
l{@
lll lll`l(ldl7lVlhl l	  l6( l9 l= lV lw!l"l#l#l	$l:	$l	&l	&l
(l
(l
0(l
T(l6
d(l;
(l@
(l}
)l
*l
+l
$+l6,lV-ls-l.lI0l3l(
4l:
4l<l(<lQ<lG@lpAlAl@ClCl5Dl@DlGElt@ElwElFl$Gl@GlpGlHl8LHl?HlJl Jl(JltlHtlZ@ulfuliulul@vlLvlvl{l#	lD%	l/4@	lGF
l9j l_j)ljj)lj@+lj+l7k,lDk-lZk@-lxk-lk-ln9lEo<lo@=lo<>lo?l@l
 l1@
l2E
lk<p
l}<q
l<r
l<@r
l<tr
l7ZhlmZhlvZilZjlZljlZjl`l` l"`ll%`l+`lh lh@#lKi$lZi@%rrrrr)l 0)llv
lljll0C)ZPVALIDZCONTEXTJZCONTEXTON)__version__ZscriptsZ
joining_typesZcodepoint_classesrr=/usr/lib/python3.8/site-packages/pip/_vendor/idna/idnadata.pys
&Wh{site-packages/pip/_vendor/idna/__pycache__/core.cpython-38.pyc000064400000021461151464452710020230 0ustar00U

.e-@s>ddlmZddlZddlZddlZddlZddlmZdZdZ	e
dZejddkr`e
ZeZGd	d
d
eZGdddeZGd
ddeZGdddeZddZddZddZddZddZddZd7ddZd d!Zd"d#Zd$d%Zd&d'Z d8d(d)Z!d*d+Z"d,d-Z#d.d/Z$d9d1d2Z%d:d3d4Z&d;d5d6Z'dS)<)idnadataN)intranges_contain	sxn--u[.。.。]c@seZdZdZdS)	IDNAErrorz7 Base exception for all IDNA-encoding related problems N__name__
__module____qualname____doc__r
r
9/usr/lib/python3.8/site-packages/pip/_vendor/idna/core.pyrsrc@seZdZdZdS)
IDNABidiErrorz= Exception when bidirectional requirements are not satisfied Nrr
r
r
rrsrc@seZdZdZdS)InvalidCodepointz> Exception when a disallowed or unallocated codepoint is used Nrr
r
r
rrsrc@seZdZdZdS)InvalidCodepointContextzE Exception when the codepoint is not valid in the context it is used Nrr
r
r
rrsrcCs0tt|}|dkr,tt|s,td|S)Nrz Unknown character in unicodedata)unicodedataZ	combiningunichrname
ValueError)cpvr
r
r_combining_class$s
rcCstt|tj|S)N)rordrZscripts)rZscriptr
r
r
_is_script+srcCs
|dS)Npunycode)encodesr
r
r	_punycode.srcCs
d|S)Nz	U+{0:04X})formatrr
r
r_unot1sr!cCst|dkrdSdS)N?FTlenlabelr
r
rvalid_label_length5sr'cCst||rdndkrdSdS)NFTr#)r&trailing_dotr
r
rvalid_string_length<sr+Fc	Cs^d}t|dD]:\}}t|}|dkrd}q|dkrd}q|sZtddS)NFrz3Unknown directionality in label {0} at position {1})RALANTr)r-r.Lz>First codepoint in label {0} must be directionality L, R or AL)
r-r.r/ENESCSETONBNNSMzHInvalid direction for codepoint at position {0} in a right-to-left label)r-r.r1r/r7)r/r1z2Can not mix numeral types in a right-to-left label)r0r1r2r3r4r5r6r7zHInvalid direction for codepoint at position {0} in a left-to-right label)r0r1z0Label ends with illegal codepoint directionality)	enumeraterZ
bidirectionalrr repr)	r&Z	check_ltrZ
bidi_labelidxr	directionZrtlZvalid_endingZnumber_typer
r
r
check_bidiCsP






r<cCs"t|dddkrtddS)NrMz0Label begins with an illegal combining characterT)rcategoryrr%r
r
rcheck_initial_combinersr?cCs<|dddkrtd|ddks0|ddkr8tdd	S)
Nz--z4Label has disallowed hyphens in 3rd and 4th positionr-z)Label must not start or end with a hyphenT)rr%r
r
rcheck_hyphen_oks
rDcCstd||krtddS)NNFCz%Label must be in Normalization Form C)r	normalizerr%r
r
r	check_nfcsrGcCs4t||}|dkr|dkr8tt||dtkr8dSd}t|dddD]B}tjt||}|tdkrrqL|tdtd	fkrLd}qqL|sdSd}t|dt|D]B}tjt||}|tdkrq|td
td	fkrd}qq|S|dkr,|dkr(tt||dtkr(dSdSdSdS)Ni rrTFrCTr0Dr-i
 )rr_virama_combining_classrangerZ
joining_typesgetr$)r&poscp_valueokiZjoining_typer
r
rvalid_contextjs<

rQcCst||}|dkrdd|kr0t|dkr`nn,t||ddkr`t||ddkr`dSdS|dkr|t|dkrt|dkrt||ddSdS|d	ks|d
kr|dkrt||ddSdS|dkr|D]6}|d
krqt|ds
t|ds
t|drdSqdSd|kr.dkrhnn6|D],}dt|krVdkr6nndSq6dSd|krdkrnn6|D],}dt|krdkrnndSqdSdS)NrrlTFiuZGreekiiZHebrewi0u・ZHiraganaZKatakanaZHani`iiii)rr$r)r&rMZ	exceptionrNrr
r
rvalid_contextos> (
" 
 
rTcCsLt|ttfr|d}t|dkr,tdt|t|t|t	|D]\}}t
|}t|tj
drpqLqLt|tj
drz,t||stdt||dt|Wn2tk
rtdt||dt|YnXqLt|tj
d	r t||s>td
t||dt|qLtdt||dt|qLt|dS)Nzutf-8rzEmpty LabelZPVALIDZCONTEXTJz-Joiner {0} not allowed at position {1} in {2}rz?Unknown codepoint adjacent to joiner {0} at position {1} in {2}ZCONTEXTO0Codepoint {0} not allowed at position {1} in {2}z0Codepoint {0} at position {1} of {2} not allowed)
isinstancebytes	bytearraydecoder$rrGrDr?r8rrrZcodepoint_classesrQrr r!r9rrTrr<)r&rMrrNr
r
rcheck_labels<


  rZcCs~z(|d}t|t|s$td|WStk
r<YnX|sJtdt|}t|t|}t|}t|sztd|S)NasciizLabel too longzNo Input)	rulabelr'rUnicodeEncodeErrorunicoderZr_alabel_prefixr%r
r
ralabel
s"
r`cCst|ttfs>z|d}Wn tk
r<t||YSX|}|trb|t	td}nt||
dS|
d}t||S)Nr[r)rVrWrXrr]rZlower
startswithr_r$rYr%r
r
rr\#s



r\Tc
Cs(ddlm}d}zt|D]\}}t|}||dkr8|nt||dfd}|d}	t|dkrj|dnd}
|	d	ks|	d
kr|r|	dkr|s|
dkr||7}q|
dk	r|	dks|	dkr|r|	d
kr|r||
7}q|	d
krtqtd|WStk
r"t	d
t||dt|YnXdS)zBRe-map the characters in the string according to UTS46 processing.r)	uts46datar,Zrr@NVrI3r=IrErU)
rcr8rbisectZbisect_leftr$
IndexErrorrrFrr r!r9)Zdomain
std3_rulestransitionalrcoutputrMcharZ
code_pointZuts46rowZstatusZreplacementr
r
ruts46_remap8sV

roc	Cst|ttfr|d}|r(t|||}d}g}|r@|d}n
t|}|rX|dgkr`td|ddkrv|d=d}|D]$}t|}|r|	|qztdqz|r|	d	d

|}t||std|S)Nr[F.r,Empty domainrCTEmpty label.zDomain too long)rVrWrXrYrosplit_unicode_dots_rerr`appendjoinr+)	rstrictuts46rkrlr*resultlabelsr&r
r
rrTs2





rcCst|ttfr|d}|r(t||d}d}g}|s@t|}n
|d}|rX|dgkr`td|dsr|d=d}|D]$}t|}|r|	|qvtdqv|r|	dd
|S)	Nr[Frpr,rqrCTrr)rVrWrXrYrorvrurr\rwrx)rryrzrkr*r{r|r&r
r
rrYss,



rY)F)F)TF)FFFF)FFF)(r,rrirresysZ	intrangesrrJr_compilervversion_infostrr^chrrUnicodeErrorrrrrrrrr!r'r+r<r?rDrGrQrTrZr`r\rorrYr
r
r
rsB

<	,
) 

site-packages/pip/_vendor/idna/__pycache__/uts46data.cpython-38.pyc000064400000524104151464452710021121 0ustar00U

.e@sbdZdZddZddZddZdd	Zd
dZdd
ZddZddZ	ddZ
ddZddZddZ
ddZddZddZd d!Zd"d#Zd$d%Zd&d'Zd(d)Zd*d+Zd,d-Zd.d/Zd0d1Zd2d3Zd4d5Zd6d7Zd8d9Zd:d;Zdd?Z d@dAZ!dBdCZ"dDdEZ#dFdGZ$dHdIZ%dJdKZ&dLdMZ'dNdOZ(dPdQZ)dRdSZ*dTdUZ+dVdWZ,dXdYZ-dZd[Z.d\d]Z/d^d_Z0d`daZ1dbdcZ2dddeZ3dfdgZ4dhdiZ5djdkZ6dldmZ7dndoZ8dpdqZ9drdsZ:dtduZ;dvdwZd|d}Z?d~dZ@ddZAddZBddZCddZDddZEddZFddZGddZHddZIddZJddZKddZLddZMddZNddZOddZPeQeeeeeeee	e
eee
eeeeeeeeeeeeeeeeeee e!e"e#e$e%e&e'e(e)e*e+e,e-e.e/e0e1e2e3e4e5e6e7e8e9e:e;e<e=e>e?e@eAeBeCeDeEeFeGeHeIeJeKeLeMeNeOePZRdS)zIDNA Mapping Table from UTS46.z11.0.0cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)3)r)r)r)r)r)r)r)r)	r)
r)r)r)
r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r) r)!r)"r)#r)$r)%r)&r)'r)(r))r)*r)+r),r)-V).r0)/r)0r0)1r0)2r0)3r0)4r0)5r0)6r0)7r0)8r0)9r0):r);r)<r)=r)>r)?r)@r)AMa)BrEb)CrEc)DrEd)ErEe)FrEf)GrEg)HrEh)IrEi)JrEj)KrEk)LrEl)MrEm)NrEn)OrEo)PrEp)QrEq)RrEr)SrEs)TrEt)UrEu)VrEv)WrEw)XrEx)YrEy)ZrEz)[r)\r)]r)^r)_r)`r)ar0)br0)cr0rrr>/usr/lib/python3.8/site-packages/pip/_vendor/idna/uts46data.py_seg_0srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)dr0)er0)fr0)gr0)hr0)ir0)jr0)kr0)lr0)mr0)nr0)or0)pr0)qr0)rr0)sr0)tr0)ur0)vr0)wr0)xr0)yr0)zr0){r)|r)}r)~r)r)X)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r )r0)r0)r0)r0)r0)r0)r0)ru ̈)r0)rErF)r0)r0)I)r0)r ̄)r0)r0)rE2)rEr)r ́)rEμ)r0)r0)ru ̧)rE1)rErb)r0)rEu1⁄4)rEu1⁄2)rEu3⁄4)r0)rEà)rEá)rEâ)rEã)rEä)rEå)rEæ)rEçrrrrr_seg_1psrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)rEè)rEé)rEê)rEë)rEì)rEí)rEî)rEï)rEð)rEñ)rEò)rEó)rEô)rEõ)rEö)r0)rEø)rEù)rEú)rEû)rEü)rEý)rEþ)Dss)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)rEuā)ir0)irEuă)ir0)irEuą)ir0)irEuć)ir0)irEuĉ)i	r0)i
rEuċ)ir0)irEuč)i
r0)irEuď)ir0)irEuđ)ir0)irEuē)ir0)irEuĕ)ir0)irEuė)ir0)irEuę)ir0)irEuě)ir0)irEuĝ)ir0)irEuğ)ir0)i rEuġ)i!r0)i"rEuģ)i#r0)i$rEuĥ)i%r0)i&rEħ)i'r0)i(rEuĩ)i)r0)i*rEuī)i+r0rrrrr_seg_2srLcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i,rEuĭ)i-r0)i.rEuį)i/r0)i0rEui̇)i1r0)i2rEZij)i4rEuĵ)i5r0)i6rEuķ)i7r0)i9rEuĺ)i:r0)i;rEuļ)i<r0)i=rEuľ)i>r0)i?rEul·)iArEuł)iBr0)iCrEuń)iDr0)iErEuņ)iFr0)iGrEuň)iHr0)iIrEuʼn)iJrEŋ)iKr0)iLrEuō)iMr0)iNrEuŏ)iOr0)iPrEuő)iQr0)iRrEœ)iSr0)iTrEuŕ)iUr0)iVrEuŗ)iWr0)iXrEuř)iYr0)iZrEuś)i[r0)i\rEuŝ)i]r0)i^rEuş)i_r0)i`rEuš)iar0)ibrEuţ)icr0)idrEuť)ier0)ifrEuŧ)igr0)ihrEuũ)iir0)ijrEuū)ikr0)ilrEuŭ)imr0)inrEuů)ior0)iprEuű)iqr0)irrEuų)isr0)itrEuŵ)iur0)ivrEuŷ)iwr0)ixrEÿ)iyrEuź)izr0)i{rEuż)i|r0)i}rEuž)i~r0)irErj)ir0)irEuɓ)irEuƃ)ir0)irEuƅ)ir0)irEɔ)irEuƈ)ir0)irEuɖ)irEuɗ)irEuƌ)ir0)irEǝ)irEə)irEɛ)irEuƒ)ir0)irEuɠrrrrr_seg_3@srTcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEɣ)ir0)irEɩ)irEɨ)irEuƙ)ir0)irEɯ)irEɲ)ir0)irEɵ)irEuơ)ir0)irEuƣ)ir0)irEuƥ)ir0)irEuʀ)irEuƨ)ir0)irEʃ)ir0)irEuƭ)ir0)irEuʈ)irEuư)ir0)irEʊ)irEʋ)irEuƴ)ir0)irEuƶ)ir0)irEʒ)irEuƹ)ir0)irEuƽ)ir0)irEudž)irEZlj)irEZnj)irEuǎ)ir0)irEuǐ)ir0)irEuǒ)ir0)irEuǔ)ir0)irEuǖ)ir0)irEuǘ)ir0)irEuǚ)ir0)irEuǜ)ir0)irEuǟ)ir0)irEuǡ)ir0)irEuǣ)ir0)irEuǥ)ir0)irEuǧ)ir0)irEuǩ)ir0)irEuǫ)ir0)irEuǭ)ir0)irEuǯ)ir0)irEZdz)irEuǵ)ir0)irEuƕ)irEuƿ)irEuǹ)ir0)irEuǻ)ir0)irEuǽ)ir0)irEuǿ)ir0)irEuȁ)ir0)irEuȃ)ir0)irEuȅ)ir0)irEuȇ)ir0)irEuȉ)i	r0)i
rEuȋ)ir0)irEuȍrrrrr_seg_4sr_cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i
r0)irEuȏ)ir0)irEuȑ)ir0)irEuȓ)ir0)irEuȕ)ir0)irEuȗ)ir0)irEuș)ir0)irEuț)ir0)irEuȝ)ir0)irEuȟ)ir0)i rEuƞ)i!r0)i"rEȣ)i#r0)i$rEuȥ)i%r0)i&rEuȧ)i'r0)i(rEuȩ)i)r0)i*rEuȫ)i+r0)i,rEuȭ)i-r0)i.rEuȯ)i/r0)i0rEuȱ)i1r0)i2rEuȳ)i3r0)i:rEuⱥ)i;rEuȼ)i<r0)i=rEuƚ)i>rEuⱦ)i?r0)iArEuɂ)iBr0)iCrEuƀ)iDrEʉ)iErEʌ)iFrEuɇ)iGr0)iHrEuɉ)iIr0)iJrEuɋ)iKr0)iLrEuɍ)iMr0)iNrEuɏ)iOr0)irErT)irEɦ)irErX)irErh)irEuɹ)irEuɻ)irEuʁ)irErr)irErv)ir0)iru ̆)iru ̇)iru ̊)iru ̨)iru ̃)iru ̋)ir0)irErU)irEr\)irErj)irErt)irEuʕ)ir0)i@rEù)iArEú)iBr0)iCrEu̓)iDrEǘ)iErEι)iFr0)iOr)iPr0)iprEuͱ)iqr0)irrEuͳ)isr0)itrEuʹ)iur0)ivrEuͷ)iwr0rrrrr_seg_5srecdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ixr)izru ι)i{r0)i~r;)irEuϳ)ir)irr)ir ̈́)irEά)irE·)irEέ)irEή)irEί)ir)irEό)ir)irEύ)irEώ)ir0)irEα)irEβ)irEγ)irEδ)irEε)irEζ)irEη)irEθ)irErd)irEκ)irEλ)irEr)irEν)irEξ)irEο)irEπ)irEρ)ir)irEσ)irEτ)irEυ)irEφ)irEχ)irEψ)irEω)irEuϊ)irEϋ)ir0)ir(r)ir0)irEuϗ)irErq)irErw)irEr)irErn)irEr)irEr)irEr})ir0)irEuϙ)ir0)irEuϛ)ir0)irEϝ)ir0)irEuϟ)ir0)irEuϡ)ir0)irEuϣ)ir0)irEuϥ)ir0)irEuϧ)ir0)irEuϩ)ir0)irEuϫ)ir0)irEuϭ)ir0)irEuϯ)ir0)irErx)irEr~)irEr)ir0)irErw)irErt)ir0)irEuϸ)ir0)irEr)irEuϻ)ir0)irEuͻ)irEuͼ)irEuͽ)irEuѐ)irEuё)irEuђrrrrr_seg_6xsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuѓ)irEuє)irEuѕ)irEuі)irEuї)irEuј)i	rEuљ)i
rEuњ)irEuћ)irEuќ)i
rEuѝ)irEuў)irEuџ)irEuа)irEuб)irEв)irEuг)irEд)irEuе)irEuж)irEuз)irEuи)irEuй)irEuк)irEuл)irEuм)irEн)irEо)irEuп)i rEuр)i!rEс)i"rEт)i#rEuу)i$rEuф)i%rEuх)i&rEuц)i'rEuч)i(rEuш)i)rEuщ)i*rEъ)i+rEuы)i,rEь)i-rEuэ)i.rEuю)i/rEuя)i0r0)i`rEuѡ)iar0)ibrEѣ)icr0)idrEuѥ)ier0)ifrEuѧ)igr0)ihrEuѩ)iir0)ijrEuѫ)ikr0)ilrEuѭ)imr0)inrEuѯ)ior0)iprEuѱ)iqr0)irrEuѳ)isr0)itrEuѵ)iur0)ivrEuѷ)iwr0)ixrEuѹ)iyr0)izrEuѻ)i{r0)i|rEuѽ)i}r0)i~rEuѿ)ir0)irEuҁ)ir0)irEuҋ)ir0)irEuҍ)ir0)irEuҏ)ir0)irEuґ)ir0)irEuғ)ir0)irEuҕ)ir0)irEuҗ)ir0)irEuҙ)ir0)irEuқ)ir0)irEuҝ)ir0rrrrr_seg_7srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuҟ)ir0)irEuҡ)ir0)irEuң)ir0)irEuҥ)ir0)irEuҧ)ir0)irEuҩ)ir0)irEuҫ)ir0)irEuҭ)ir0)irEuү)ir0)irEuұ)ir0)irEuҳ)ir0)irEuҵ)ir0)irEuҷ)ir0)irEuҹ)ir0)irEuһ)ir0)irEuҽ)ir0)irEuҿ)ir0)ir)irEuӂ)ir0)irEuӄ)ir0)irEuӆ)ir0)irEuӈ)ir0)irEuӊ)ir0)irEuӌ)ir0)irEuӎ)ir0)irEuӑ)ir0)irEuӓ)ir0)irEuӕ)ir0)irEuӗ)ir0)irEuә)ir0)irEuӛ)ir0)irEuӝ)ir0)irEuӟ)ir0)irEuӡ)ir0)irEuӣ)ir0)irEuӥ)ir0)irEuӧ)ir0)irEuө)ir0)irEuӫ)ir0)irEuӭ)ir0)irEuӯ)ir0)irEuӱ)ir0)irEuӳ)ir0)irEuӵ)ir0)irEuӷ)ir0)irEuӹ)ir0)irEuӻ)ir0)irEuӽ)ir0)irEuӿ)ir0)irEuԁ)ir0)irEuԃrrrrr_seg_8HsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir0)irEuԅ)ir0)irEuԇ)ir0)irEuԉ)i	r0)i
rEuԋ)ir0)irEuԍ)i
r0)irEuԏ)ir0)irEuԑ)ir0)irEuԓ)ir0)irEuԕ)ir0)irEuԗ)ir0)irEuԙ)ir0)irEuԛ)ir0)irEuԝ)ir0)irEuԟ)ir0)i rEuԡ)i!r0)i"rEuԣ)i#r0)i$rEuԥ)i%r0)i&rEuԧ)i'r0)i(rEuԩ)i)r0)i*rEuԫ)i+r0)i,rEuԭ)i-r0)i.rEuԯ)i/r0)i0r)i1rEuա)i2rEuբ)i3rEuգ)i4rEuդ)i5rEuե)i6rEuզ)i7rEuէ)i8rEuը)i9rEuթ)i:rEuժ)i;rEuի)i<rEuլ)i=rEuխ)i>rEuծ)i?rEuկ)i@rEuհ)iArEuձ)iBrEuղ)iCrEuճ)iDrEuմ)iErEuյ)iFrEuն)iGrEuշ)iHrEuո)iIrEuչ)iJrEuպ)iKrEuջ)iLrEuռ)iMrEuս)iNrEuվ)iOrEuտ)iPrEuր)iQrEuց)iRrEuւ)iSrEuփ)iTrEuք)iUrEuօ)iVrEuֆ)iWr)iYr0)irEuեւ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0rrrrr_seg_9srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iurEuاٴ)ivrEuوٴ)iwrEۇٴ)ixrEuيٴ)iyr0)ir)ir0)ir)ir0)iKr)iMr0)ir)ir0)ir)ir0)i.r)i0r0)i?r)i@r0)i\r)i^r0)i_r)i`r0)ikr)ir0)ir)ir0)ir)ir0)ir)ir0)iX	rEuक़)iY	rEuख़)iZ	rEuग़)i[	rEuज़)i\	rEuड़)i]	rEuढ़)i^	rEuफ़)i_	rEuय़)i`	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	rEuড়)i	rEuঢ়)i	r)i	rEuয়)i	r0)i	r)i	r0)i	r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i)
r)i*
r0)i1
r)i2
r0)i3
rEuਲ਼)i4
r)i5
r0)i6
rEuਸ਼)i7
r)i8
r0)i:
r)i<
r0)i=
r)i>
r0)iC
r)iG
r0)iI
r)iK
r0)iN
r)iQ
r0)iR
r)iY
rEuਖ਼)iZ
rEuਗ਼)i[
rEuਜ਼rrrrr_seg_10srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i\
r0)i]
r)i^
rEuਫ਼)i_
r)if
r0)iw
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)ir)ir0)ir)ir0)i
r)ir0)ir)ir0)i)r)i*r0)i1r)i2r0)i4r)i5r0)i:r)i<r0)iEr)iGr0)iIr)iKr0)iNr)iVr0)iXr)i\rEuଡ଼)i]rEuଢ଼)i^r)i_r0)idr)ifr0)ixr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)ir)ir0rrrrr_seg_11srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i)r)i*r0)i:r)i=r0)iEr)iFr0)iIr)iJr0)iNr)iUr0)iWr)iXr0)i[r)i`r0)idr)ifr0)ipr)ixr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)i
r0)i
r)i
r0)i

r)i
r0)i
r)i
r0)iE
r)iF
r0)iI
r)iJ
r0)iP
r)iT
r0)id
r)if
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)ir0)i3rEuํา)i4r0)i;r)i?r0)i\r)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0rrrrr_seg_12srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)irEuໍາ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irEuຫນ)irEuຫມ)ir0)ir)ir0)irEu་)i
r0)iCrEuགྷ)iDr0)iHr)iIr0)iMrEuཌྷ)iNr0)iRrEuདྷ)iSr0)iWrEuབྷ)iXr0)i\rEuཛྷ)i]r0)iirEuཀྵ)ijr0)imr)iqr0)isrEuཱི)itr0)iurEuཱུ)ivrEuྲྀ)iwrEu	ྲཱྀ)ixrEuླྀ)iyrEu	ླཱྀ)izr0)irEuཱྀ)ir0)irEuྒྷ)ir0)ir)ir0)irEuྜྷ)ir0)irEuྡྷ)ir0)irEuྦྷ)ir0)irEuྫྷ)ir0)irEuྐྵ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irEuⴧ)ir)irEuⴭ)ir)ir0)irEuნ)ir0)i_r)iar0)iIr)iJr0)iNr)iPr0)iWr)iXr0)iYr)iZr0)i^r)i`r0)ir)ir0rrrrr_seg_13PsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i[r)i]r0)i}r)ir0)ir)ir0)ir)irEuᏰ)irEuᏱ)irEuᏲ)irEuᏳ)irEuᏴ)irEuᏵ)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)ir)i r0)i7r)i@r0)iTr)i`r0)imr)inr0)iqr)irr0)itr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir)ir0)ir)i r0)iyr)ir0)ir)ir0)ir)ir0)ir)i r0)i,r)i0r0)i<r)i@r0)iAr)iDr0)inr)ipr0)iur)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i_r)i`r0)i}r)ir0)ir)ir0rrrrr_seg_14srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)ir0)ir)ir0)ir)ir0)iLr)iPr0)i}r)ir0)ir)ir0)i8r)i;r0)iJr)iMr0)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEꙋ)ir)ir0)ir)ir0)ir)ir0)i,rErF)i-rEr)i.rErH)i/r0)i0rErL)i1rErN)i2rErQ)i3rErR)i4rErT)i5rErV)i6rErX)i7rErZ)i8rEr\)i9rEr^)i:rEr`)i;r0)i<rErb)i=rEr`)i>rErd)i?rErh)i@rErl)iArErn)iBrErr)iCrErF)iDrEɐ)iErEɑ)iFrEuᴂ)iGrErH)iHrErL)iIrErN)iJrErR)iKrErS)iLrEɜ)iMrErR)iNr0)iOrErZ)iPrEr^)iQrErM)iRrErb)iSrErP)iTrEuᴖ)iUrEuᴗ)iVrErd)iWrErl)iXrErn)iYrEuᴝ)iZrErX)i[rErp)i\rEuᴥ)i]rErq)i^rErr)i_rErs)i`rEr)iarEr)ibrErV)icrErh)idrErn)ierErp)ifrErq)igrErr)ihrEr~)iirEr)ijrEr)ikr0)ixrEr)iyr0)irEɒ)irErJ)irEuɕ)irErrrrrr_seg_15 srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irErP)irEuɟ)irEɡ)irEɥ)irErW)irErV)irEɪ)irEuᵻ)irEʝ)irEuɭ)irEuᶅ)irEuʟ)irEɱ)irEuɰ)irErY)irEuɳ)irEuɴ)irErZ)irEuɸ)irEuʂ)irEr[)irEuƫ)irEra)irEr\)irEuᴜ)irEr])irErb)irErx)irEuʐ)irEuʑ)irEr^)irErw)ir0)ir)ir0)irEuḁ)ir0)irEuḃ)ir0)irEuḅ)ir0)irEuḇ)ir0)irEuḉ)i	r0)i
rEuḋ)ir0)irEuḍ)i
r0)irEuḏ)ir0)irEuḑ)ir0)irEuḓ)ir0)irEuḕ)ir0)irEuḗ)ir0)irEuḙ)ir0)irEuḛ)ir0)irEuḝ)ir0)irEuḟ)ir0)i rEuḡ)i!r0)i"rEuḣ)i#r0)i$rEuḥ)i%r0)i&rEuḧ)i'r0)i(rEuḩ)i)r0)i*rEuḫ)i+r0)i,rEuḭ)i-r0)i.rEuḯ)i/r0)i0rEuḱ)i1r0)i2rEuḳ)i3r0)i4rEuḵ)i5r0)i6rEuḷ)i7r0)i8rEuḹ)i9r0)i:rEuḻ)i;r0)i<rEuḽ)i=r0)i>rEuḿ)i?r0rrrrr_seg_16srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i@rEuṁ)iAr0)iBrEuṃ)iCr0)iDrEuṅ)iEr0)iFrEuṇ)iGr0)iHrEuṉ)iIr0)iJrEuṋ)iKr0)iLrEuṍ)iMr0)iNrEuṏ)iOr0)iPrEuṑ)iQr0)iRrEuṓ)iSr0)iTrEuṕ)iUr0)iVrEuṗ)iWr0)iXrEuṙ)iYr0)iZrEuṛ)i[r0)i\rEuṝ)i]r0)i^rEuṟ)i_r0)i`rEṡ)iar0)ibrEuṣ)icr0)idrEuṥ)ier0)ifrEuṧ)igr0)ihrEuṩ)iir0)ijrEuṫ)ikr0)ilrEuṭ)imr0)inrEuṯ)ior0)iprEuṱ)iqr0)irrEuṳ)isr0)itrEuṵ)iur0)ivrEuṷ)iwr0)ixrEuṹ)iyr0)izrEuṻ)i{r0)i|rEuṽ)i}r0)i~rEuṿ)ir0)irEuẁ)ir0)irEuẃ)ir0)irEuẅ)ir0)irEuẇ)ir0)irEuẉ)ir0)irEuẋ)ir0)irEuẍ)ir0)irEuẏ)ir0)irEuẑ)ir0)irEuẓ)ir0)irEuẕ)ir0)irEuaʾ)irEr)ir0)irEr))ir0)irEuạ)ir0)irEuả)ir0)irEuấ)ir0)irEuầ)ir0)irEuẩrrrrr_seg_17srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir0)irEuẫ)ir0)irEuậ)ir0)irEuắ)ir0)irEuằ)ir0)irEuẳ)ir0)irEuẵ)ir0)irEuặ)ir0)irEuẹ)ir0)irEuẻ)ir0)irEuẽ)ir0)irEuế)ir0)irEuề)ir0)irEuể)ir0)irEuễ)ir0)irEuệ)ir0)irEuỉ)ir0)irEuị)ir0)irEuọ)ir0)irEuỏ)ir0)irEuố)ir0)irEuồ)ir0)irEuổ)ir0)irEuỗ)ir0)irEuộ)ir0)irEuớ)ir0)irEuờ)ir0)irEuở)ir0)irEuỡ)ir0)irEuợ)ir0)irEuụ)ir0)irEuủ)ir0)irEuứ)ir0)irEuừ)ir0)irEuử)ir0)irEuữ)ir0)irEuự)ir0)irEuỳ)ir0)irEuỵ)ir0)irEuỷ)ir0)irEuỹ)ir0)irEuỻ)ir0)irEuỽ)ir0)irEuỿ)ir0)irEuἀ)i	rEuἁ)i
rEuἂ)irEuἃ)irEuἄ)i
rEuἅ)irEuἆ)irEuἇ)ir0)ir)irEuἐ)irEuἑ)irEuἒrrrrr_seg_18XsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuἓ)irEuἔ)irEuἕ)ir)i r0)i(rEuἠ)i)rEuἡ)i*rEuἢ)i+rEuἣ)i,rEuἤ)i-rEuἥ)i.rEuἦ)i/rEuἧ)i0r0)i8rEuἰ)i9rEuἱ)i:rEuἲ)i;rEuἳ)i<rEuἴ)i=rEuἵ)i>rEuἶ)i?rEuἷ)i@r0)iFr)iHrEuὀ)iIrEuὁ)iJrEuὂ)iKrEuὃ)iLrEuὄ)iMrEuὅ)iNr)iPr0)iXr)iYrEuὑ)iZr)i[rEuὓ)i\r)i]rEuὕ)i^r)i_rEuὗ)i`r0)ihrEuὠ)iirEuὡ)ijrEuὢ)ikrEuὣ)ilrEuὤ)imrEuὥ)inrEuὦ)iorEuὧ)ipr0)iqrErh)irr0)isrErj)itr0)iurErk)ivr0)iwrErl)ixr0)iyrErm)izr0)i{rErn)i|r0)i}rEro)i~r)irEἀι)irEἁι)irEἂι)irEἃι)irEἄι)irEἅι)irEἆι)irEἇι)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEἠι)irEἡι)irEἢι)irEἣι)irEἤι)irEἥι)irEἦι)irEἧι)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEὠι)irEὡι)irEὢι)irEὣιrrrrr_seg_19srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEὤι)irEὥι)irEὦι)irEὧι)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir0)irEuὰι)irEαι)irEuάι)ir)ir0)irEuᾶι)irEuᾰ)irEuᾱ)irEuὰ)irErh)irEr)ir ̓)irErd)irr)iru ͂)iru ̈͂)irEuὴι)irEηι)irEuήι)ir)ir0)irEuῆι)irEuὲ)irErj)irEuὴ)irErk)irEr)iru ̓̀)iru ̓́)iru ̓͂)ir0)irEuΐ)ir)ir0)irEuῐ)irEuῑ)irEuὶ)irErl)ir)iru ̔̀)iru ̔́)iru ̔͂)ir0)irEuΰ)ir0)irEuῠ)irEuῡ)irEuὺ)irErn)irEuῥ)iru ̈̀)irrg)ir`)ir)irEuὼι)irEωι)irEuώι)ir)ir0)irEuῶι)irEuὸ)irErm)irEuὼ)irEro)irEr)irr)iru ̔)ir)i rr)i r)i r()i r)i r0)i rEu‐)i r0)i ru ̳)i r0)i$ r)i' r0)i( r)i/ rr)i0 r0)i3 rEu′′)i4 rEu	′′′)i5 r0)i6 rEu‵‵)i7 rEu	‵‵‵rrrrr_seg_20(srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i8 r0)i< rz!!)i= r0)i> r ̅)i? r0)iG rz??)iH rz?!)iI rz!?)iJ r0)iW rEu′′′′)iX r0)i_ rr)i` r)ia r)id r)ie r)ip rE0)iq rErV)ir r)it rE4)iu rE5)iv rE6)iw rE7)ix rE8)iy rE9)iz r+)i{ rE−)i| r=)i} r()i~ r))i rEr`)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rr)i rEr)i rr)i rr)i rr)i r)i rErF)i rErN)i rErb)i rErt)i rErR)i rErT)i rErZ)i rEr\)i rEr^)i rEr`)i rErd)i rErj)i rErl)i r)i r0)i rEZrs)i r0)i r)i r0)i r)i!rza/c)i!rza/s)i!rErJ)i!rEu°c)i!r0)i!rzc/o)i!rzc/u)i!rErS)i!r0)i	!rEu°f)i
!rErR)i!rErT)i!rErK)i!rErV)i!rEr\)i!r0)i!rEr`)i!rEZno)i!r0)i!rErd)i!rErf)i!rErh)i!r0)i !rEZsm)i!!rEZtel)i"!rEZtm)i#!r0)i$!rErx)i%!r0)i&!rEr)i'!r0)i(!rErx)i)!r0rrrrr_seg_21srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i*!rErZ)i+!rEr)i,!rErH)i-!rErJ)i.!r0)i/!rErN)i1!rErP)i2!r)i3!rEr^)i4!rErb)i5!rEא)i6!rEuב)i7!rEuג)i8!rEד)i9!rErV)i:!r0)i;!rEZfax)id?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ib$rEr)ic$rEr)id$rEr)ie$rEr)if$rEr)ig$rEr)ih$rEr)ii$rEZ10)ij$rEZ11)ik$rEZ12)il$rEZ13)im$rEZ14)in$rEZ15)io$rEZ16)ip$rEZ17)iq$rEZ18)ir$rEZ19)is$rEZ20)it$rz(1))iu$rz(2))iv$rz(3))iw$rz(4))ix$rz(5))iy$rz(6))iz$rz(7))i{$rz(8))i|$rz(9))i}$rz(10))i~$rz(11))i$rz(12))i$rz(13))i$rz(14))i$rz(15))i$rz(16))i$rz(17))i$rz(18))i$rz(19))i$rz(20))i$r)i$r(a))i$r(b))i$r(c))i$r(d))i$r(e))i$r(f))i$r(g))i$r(h))i$r(i))i$r(j))i$r(k))i$r(l))i$r(m))i$r(n))i$r(o))i$r(p))i$r(q))i$r(r))i$r(s))i$r(t))i$r(u))i$r(v))i$r(w))i$r(x))i$r(y))i$r(z))i$rErF)i$rErH)i$rErJ)i$rErL)i$rErN)i$rErP)i$rErR)i$rErT)i$rErV)i$rErX)i$rErZ)i$rEr\)i$rEr^)i$rEr`)i$rErb)i$rErd)i$rErf)i$rErh)i$rErj)i$rErl)i$rErn)i$rErp)i$rErr)i$rErt)i$rErv)i$rErx)i$rErF)i$rErH)i$rErJ)i$rErL)i$rErN)i$rErP)i$rErR)i$rErT)i$rErVrrrrr_seg_23`	srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i$rErX)i$rErZ)i$rEr\)i$rEr^)i$rEr`)i$rErb)i$rErd)i$rErf)i$rErh)i$rErj)i$rErl)i$rErn)i$rErp)i$rErr)i$rErt)i$rErv)i$rErx)i$rEr)i$r0)i*rEu∫∫∫∫)i
*r0)it*rz::=)iu*rz==)iv*rz===)iw*r0)i*rEu⫝̸)i*r0)it+r)iv+r0)i+r)i+r0)i+r)i+r0)i+r)i,rEuⰰ)i,rEuⰱ)i,rEuⰲ)i,rEuⰳ)i,rEuⰴ)i,rEuⰵ)i,rEuⰶ)i,rEuⰷ)i,rEuⰸ)i	,rEuⰹ)i
,rEuⰺ)i,rEuⰻ)i,rEuⰼ)i
,rEuⰽ)i,rEuⰾ)i,rEuⰿ)i,rEuⱀ)i,rEuⱁ)i,rEuⱂ)i,rEuⱃ)i,rEuⱄ)i,rEuⱅ)i,rEuⱆ)i,rEuⱇ)i,rEuⱈ)i,rEuⱉ)i,rEuⱊ)i,rEuⱋ)i,rEuⱌ)i,rEuⱍ)i,rEuⱎ)i,rEuⱏ)i ,rEuⱐ)i!,rEuⱑ)i",rEuⱒ)i#,rEuⱓ)i$,rEuⱔ)i%,rEuⱕ)i&,rEuⱖ)i',rEuⱗ)i(,rEuⱘ)i),rEuⱙ)i*,rEuⱚ)i+,rEuⱛ)i,,rEuⱜ)i-,rEuⱝ)i.,rEuⱞ)i/,r)i0,r0)i_,r)i`,rEuⱡ)ia,r0)ib,rEɫ)ic,rEuᵽ)id,rEuɽ)ie,r0)ig,rEuⱨ)ih,r0)ii,rEuⱪ)ij,r0)ik,rEuⱬ)il,r0)im,rEr)in,rEr)io,rEr)ip,rErrrrrr_seg_24	srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iq,r0)ir,rEuⱳ)is,r0)iu,rEuⱶ)iv,r0)i|,rErX)i},rErp)i~,rEuȿ)i,rEuɀ)i,rEuⲁ)i,r0)i,rEuⲃ)i,r0)i,rEuⲅ)i,r0)i,rEuⲇ)i,r0)i,rEuⲉ)i,r0)i,rEuⲋ)i,r0)i,rEuⲍ)i,r0)i,rEuⲏ)i,r0)i,rEuⲑ)i,r0)i,rEuⲓ)i,r0)i,rEuⲕ)i,r0)i,rEuⲗ)i,r0)i,rEuⲙ)i,r0)i,rEuⲛ)i,r0)i,rEuⲝ)i,r0)i,rEuⲟ)i,r0)i,rEuⲡ)i,r0)i,rEuⲣ)i,r0)i,rEuⲥ)i,r0)i,rEuⲧ)i,r0)i,rEuⲩ)i,r0)i,rEuⲫ)i,r0)i,rEuⲭ)i,r0)i,rEuⲯ)i,r0)i,rEuⲱ)i,r0)i,rEuⲳ)i,r0)i,rEuⲵ)i,r0)i,rEuⲷ)i,r0)i,rEuⲹ)i,r0)i,rEuⲻ)i,r0)i,rEuⲽ)i,r0)i,rEuⲿ)i,r0)i,rEuⳁ)i,r0)i,rEuⳃ)i,r0)i,rEuⳅ)i,r0)i,rEuⳇ)i,r0)i,rEuⳉ)i,r0)i,rEuⳋ)i,r0)i,rEuⳍ)i,r0)i,rEuⳏ)i,r0)i,rEuⳑ)i,r0)i,rEuⳓ)i,r0)i,rEuⳕ)i,r0)i,rEuⳗ)i,r0)i,rEuⳙ)i,r0)i,rEuⳛrrrrr_seg_250
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i,r0)i,rEuⳝ)i,r0)i,rEuⳟ)i,r0)i,rEuⳡ)i,r0)i,rEuⳣ)i,r0)i,rEuⳬ)i,r0)i,rEuⳮ)i,r0)i,rEuⳳ)i,r0)i,r)i,r0)i&-r)i'-r0)i(-r)i--r0)i.-r)i0-r0)ih-r)io-rEuⵡ)ip-r0)iq-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)iO.r)i.r0)i.r)i.r0)i.rEu母)i.r0)i.rEu龟)i.r)i/rE一)i/rEu丨)i/rEu丶)i/rEu丿)i/rE乙)i/rEu亅)i/rE二)i/rEu亠)i/rE人)i	/rEu儿)i
/rEu入)i/rE八)i/rEu冂)i
/rEu冖)i/rEu冫)i/rEu几)i/rE凵)i/rEu刀)i/rE力)i/rEu勹)i/rEu匕)i/rEu匚)i/rEu匸)i/rE十)i/rEu卜)i/rEu卩)i/rEu厂)i/rEu厶)i/rEu又)i/rEu口)i/rEu囗)i/rE土)i /rEu士)i!/rEu夂)i"/rEu夊)i#/rEu夕)i$/rEu大)i%/rE女)i&/rEu子)i'/rEu宀)i(/rEu寸)i)/rEu小)i*/rE尢)i+/rEu尸)i,/rE屮)i-/rEu山rrrrr_seg_26
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i./rEu巛)i//rEu工)i0/rEu己)i1/rEu巾)i2/rEu干)i3/rEu幺)i4/rEu广)i5/rEu廴)i6/rE廾)i7/rEu弋)i8/rEu弓)i9/rEu彐)i:/rEu彡)i;/rEu彳)i/rEu戶)i?/rE手)i@/rEu支)iA/rEu攴)iB/rE文)iC/rEu斗)iD/rEu斤)iE/rEu方)iF/rEu无)iG/rE日)iH/rEu曰)iI/rE月)iJ/rE木)iK/rEu欠)iL/rEu止)iM/rE歹)iN/rEu殳)iO/rEu毋)iP/rEu比)iQ/rEu毛)iR/rEu氏)iS/rEu气)iT/rE水)iU/rE火)iV/rEu爪)iW/rEu父)iX/rEu爻)iY/rEu爿)iZ/rEu片)i[/rEu牙)i\/rEu牛)i]/rEu犬)i^/rEu玄)i_/rEu玉)i`/rEu瓜)ia/rEu瓦)ib/rEu甘)ic/rE生)id/rEu用)ie/rEu田)if/rEu疋)ig/rEu疒)ih/rEu癶)ii/rEu白)ij/rEu皮)ik/rEu皿)il/rEu目)im/rEu矛)in/rEu矢)io/rEu石)ip/rEu示)iq/rEu禸)ir/rEu禾)is/rEu穴)it/rE立)iu/rEu竹)iv/rEu米)iw/rEu糸)ix/rEu缶)iy/rEu网)iz/rEu羊)i{/rE羽)i|/rE老)i}/rEu而)i~/rEu耒)i/rEu耳)i/rEu聿)i/rEu肉)i/rEu臣)i/rEu自)i/rEu至)i/rEu臼)i/rEu舌)i/rEu舛)i/rEu舟)i/rEu艮)i/rEu色)i/rEu艸)i/rEu虍)i/rEu虫)i/rEu血)i/rE行)i/rE衣)i/rEu襾rrrrr_seg_27sr cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i/rE見)i/rEu角)i/rEu言)i/rEu谷)i/rEu豆)i/rE豕)i/rEu豸)i/rEu貝)i/rEu赤)i/rE走)i/rEu足)i/rEu身)i/rE車)i/rEu辛)i/rE辰)i/rEu辵)i/rEu邑)i/rEu酉)i/rEu釆)i/rE里)i/rE金)i/rEu長)i/rEu門)i/rEu阜)i/rEu隶)i/rEu隹)i/rEu雨)i/rEu靑)i/rEu非)i/rEu面)i/rEu革)i/rEu韋)i/rEu韭)i/rEu音)i/rEu頁)i/rEu風)i/rEu飛)i/rEu食)i/rEu首)i/rEu香)i/rEu馬)i/rEu骨)i/rEu高)i/rEu髟)i/rEu鬥)i/rEu鬯)i/rEu鬲)i/rEu鬼)i/rEu魚)i/rEu鳥)i/rEu鹵)i/rE鹿)i/rEu麥)i/rE麻)i/rEu黃)i/rEu黍)i/rEu黑)i/rE黹)i/rEu黽)i/rEu鼎)i/rEu鼓)i/rEu鼠)i/rE鼻)i/rEu齊)i/rEu齒)i/rE龍)i/rE龜)i/rEu龠)i/r)i0rr)i0r0)i0rE.)i0r0)i60rEu〒)i70r0)i80rEr)i90rEu卄)i:0rEu卅)i;0r0)i@0r)iA0r0)i0r)i0r0)i0ru ゙)i0ru ゚)i0r0)i0rEuより)i0r0)i0rEuコト)i1r)i1r0)i01r)i11rEᄀ)i21rEᄁ)i31rEᆪ)i41rEᄂ)i51rEᆬ)i61rEᆭ)i71rEᄃ)i81rEᄄrrrrr_seg_28hsr7cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i91rEᄅ)i:1rEᆰ)i;1rEᆱ)i<1rEᆲ)i=1rEᆳ)i>1rEᆴ)i?1rEᆵ)i@1rEᄚ)iA1rEᄆ)iB1rEᄇ)iC1rEᄈ)iD1rEᄡ)iE1rEᄉ)iF1rEᄊ)iG1rEᄋ)iH1rEᄌ)iI1rEᄍ)iJ1rEᄎ)iK1rEᄏ)iL1rEᄐ)iM1rEᄑ)iN1rEᄒ)iO1rEᅡ)iP1rEᅢ)iQ1rEᅣ)iR1rEᅤ)iS1rEᅥ)iT1rEᅦ)iU1rEᅧ)iV1rEᅨ)iW1rEᅩ)iX1rEᅪ)iY1rEᅫ)iZ1rEᅬ)i[1rEᅭ)i\1rEᅮ)i]1rEᅯ)i^1rEᅰ)i_1rEᅱ)i`1rEᅲ)ia1rEᅳ)ib1rEᅴ)ic1rEᅵ)id1r)ie1rEuᄔ)if1rEuᄕ)ig1rEuᇇ)ih1rEuᇈ)ii1rEuᇌ)ij1rEuᇎ)ik1rEuᇓ)il1rEuᇗ)im1rEuᇙ)in1rEuᄜ)io1rEuᇝ)ip1rEuᇟ)iq1rEuᄝ)ir1rEuᄞ)is1rEuᄠ)it1rEuᄢ)iu1rEuᄣ)iv1rEuᄧ)iw1rEuᄩ)ix1rEuᄫ)iy1rEuᄬ)iz1rEuᄭ)i{1rEuᄮ)i|1rEuᄯ)i}1rEuᄲ)i~1rEuᄶ)i1rEuᅀ)i1rEuᅇ)i1rEuᅌ)i1rEuᇱ)i1rEuᇲ)i1rEuᅗ)i1rEuᅘ)i1rEuᅙ)i1rEuᆄ)i1rEuᆅ)i1rEuᆈ)i1rEuᆑ)i1rEuᆒ)i1rEuᆔ)i1rEuᆞ)i1rEuᆡ)i1r)i1r0)i1rEr)i1rEr)i1rE三)i1rE四)i1rE上)i1rE中)i1rE下)i1rEu甲)i1rEr)i1rEu丙)i1rEu丁)i1rE天rrrrr_seg_29sricdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i1rEu地)i1rEr)i1r0)i1r)i1r0)i1r)i1r0)i2ru(ᄀ))i2ru(ᄂ))i2ru(ᄃ))i2ru(ᄅ))i2ru(ᄆ))i2ru(ᄇ))i2ru(ᄉ))i2ru(ᄋ))i2ru(ᄌ))i	2ru(ᄎ))i
2ru(ᄏ))i2ru(ᄐ))i2ru(ᄑ))i
2ru(ᄒ))i2ru(가))i2ru(나))i2ru(다))i2ru(라))i2ru(마))i2ru(바))i2ru(사))i2ru(아))i2ru(자))i2ru(차))i2ru(카))i2ru(타))i2ru(파))i2ru(하))i2ru(주))i2ru(오전))i2ru(오후))i2r)i 2ru(一))i!2ru(二))i"2ru(三))i#2ru(四))i$2ru(五))i%2ru(六))i&2ru(七))i'2ru(八))i(2ru(九))i)2ru(十))i*2ru(月))i+2ru(火))i,2ru(水))i-2ru(木))i.2ru(金))i/2ru(土))i02ru(日))i12ru(株))i22ru(有))i32ru(社))i42ru(名))i52ru(特))i62ru(財))i72ru(祝))i82ru(労))i92ru(代))i:2ru(呼))i;2ru(学))i<2ru(監))i=2ru(企))i>2ru(資))i?2ru(協))i@2ru(祭))iA2ru(休))iB2ru(自))iC2ru(至))iD2rEu問)iE2rEu幼)iF2rEr)iG2rEu箏)iH2r0)iP2rEZpte)iQ2rEZ21)iR2rEZ22)iS2rEZ23)iT2rEZ24)iU2rEZ25)iV2rEZ26)iW2rEZ27)iX2rEZ28)iY2rEZ29)iZ2rEZ30)i[2rEZ31)i\2rEZ32)i]2rEZ33)i^2rEZ34)i_2rEZ35)i`2rEr/)ia2rEr2)ib2rEr5)ic2rEr8rrrrr_seg_308srjcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)id2rEr@)ie2rErA)if2rErD)ig2rErF)ih2rErG)ii2rErI)ij2rErJ)ik2rErK)il2rErL)im2rErM)in2rEu가)io2rEu나)ip2rEu다)iq2rEu라)ir2rEu마)is2rEu바)it2rEu사)iu2rEu아)iv2rEu자)iw2rEu차)ix2rEu카)iy2rEu타)iz2rEu파)i{2rEu하)i|2rEu참고)i}2rEu주의)i~2rEu우)i2r0)i2rEr)i2rEr)i2rErc)i2rErd)i2rEu五)i2rE六)i2rEu七)i2rEr)i2rEu九)i2rEr)i2rEr)i2rEr)i2rEr)i2rEr)i2rEr')i2rEr)i2rEr)i2rEu株)i2rE有)i2rE社)i2rEu名)i2rEu特)i2rEu財)i2rE祝)i2rEu労)i2rEu秘)i2rEu男)i2rEr
)i2rEu適)i2rEu優)i2rEu印)i2rEu注)i2rEu項)i2rEu休)i2rEu写)i2rEu正)i2rEre)i2rErf)i2rErg)i2rE左)i2rE右)i2rEu医)i2rEu宗)i2rEu学)i2rEu監)i2rEu企)i2rEu資)i2rEu協)i2rEu夜)i2rEZ36)i2rEZ37)i2rEZ38)i2rEZ39)i2rEZ40)i2rEZ41)i2rEZ42)i2rEZ43)i2rEZ44)i2rEZ45)i2rEZ46)i2rEZ47)i2rEZ48)i2rEZ49)i2rEZ50)i2rEu1月)i2rEu2月)i2rEu3月)i2rEu4月)i2rEu5月)i2rEu6月)i2rEu7月)i2rEu8月rrrrr_seg_31srqcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i2rEu9月)i2rEu10月)i2rEu11月)i2rEu12月)i2rEZhg)i2rEZerg)i2rEZev)i2rEZltd)i2rEア)i2rEイ)i2rEウ)i2rEエ)i2rEオ)i2rEカ)i2rEキ)i2rEク)i2rEケ)i2rEコ)i2rEサ)i2rEシ)i2rEス)i2rEセ)i2rEソ)i2rEタ)i2rEチ)i2rEツ)i2rEテ)i2rEト)i2rEナ)i2rEニ)i2rEヌ)i2rEネ)i2rEノ)i2rEハ)i2rEヒ)i2rEフ)i2rEヘ)i2rEホ)i2rEマ)i2rEミ)i2rEム)i2rEメ)i2rEモ)i2rEヤ)i2rEユ)i2rEヨ)i2rEラ)i2rEリ)i2rEル)i2rEレ)i2rEロ)i2rEワ)i2rEuヰ)i2rEuヱ)i2rEヲ)i2r)i3rEuアパート)i3rEuアルファ)i3rEuアンペア)i3rEu	アール)i3rEuイニング)i3rEu	インチ)i3rEu	ウォン)i3rEuエスクード)i3rEuエーカー)i	3rEu	オンス)i
3rEu	オーム)i3rEu	カイリ)i3rEuカラット)i
3rEuカロリー)i3rEu	ガロン)i3rEu	ガンマ)i3rEuギガ)i3rEu	ギニー)i3rEuキュリー)i3rEuギルダー)i3rEuキロ)i3rEuキログラム)i3rEuキロメートル)i3rEuキロワット)i3rEu	グラム)i3rEuグラムトン)i3rEuクルゼイロ)i3rEuクローネ)i3rEu	ケース)i3rEu	コルナ)i3rEu	コーポ)i3rEuサイクル)i 3rEuサンチーム)i!3rEuシリング)i"3rEu	センチ)i#3rEu	セント)i$3rEu	ダース)i%3rEuデシ)i&3rEuドル)i'3rEuトン)i(3rEuナノ)i)3rEu	ノット)i*3rEu	ハイツ)i+3rEuパーセントrrrrr_seg_32
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i,3rEu	パーツ)i-3rEuバーレル)i.3rEuピアストル)i/3rEu	ピクル)i03rEuピコ)i13rEuビル)i23rEuファラッド)i33rEuフィート)i43rEuブッシェル)i53rEu	フラン)i63rEuヘクタール)i73rEuペソ)i83rEu	ペニヒ)i93rEu	ヘルツ)i:3rEu	ペンス)i;3rEu	ページ)i<3rEu	ベータ)i=3rEuポイント)i>3rEu	ボルト)i?3rEuホン)i@3rEu	ポンド)iA3rEu	ホール)iB3rEu	ホーン)iC3rEuマイクロ)iD3rEu	マイル)iE3rEu	マッハ)iF3rEu	マルク)iG3rEuマンション)iH3rEuミクロン)iI3rEuミリ)iJ3rEuミリバール)iK3rEuメガ)iL3rEuメガトン)iM3rEuメートル)iN3rEu	ヤード)iO3rEu	ヤール)iP3rEu	ユアン)iQ3rEuリットル)iR3rEuリラ)iS3rEu	ルピー)iT3rEuルーブル)iU3rEuレム)iV3rEuレントゲン)iW3rEu	ワット)iX3rEu0点)iY3rEu1点)iZ3rEu2点)i[3rEu3点)i\3rEu4点)i]3rEu5点)i^3rEu6点)i_3rEu7点)i`3rEu8点)ia3rEu9点)ib3rEu10点)ic3rEu11点)id3rEu12点)ie3rEu13点)if3rEu14点)ig3rEu15点)ih3rEu16点)ii3rEu17点)ij3rEu18点)ik3rEu19点)il3rEu20点)im3rEu21点)in3rEu22点)io3rEu23点)ip3rEu24点)iq3rEZhpa)ir3rEZda)is3rEZau)it3rEZbar)iu3rEZov)iv3rEZpc)iw3rEZdm)ix3rEZdm2)iy3rEZdm3)iz3rEZiu)i{3rEu平成)i|3rEu昭和)i}3rEu大正)i~3rEu明治)i3rEu株式会社)i3rEpa)i3rEZna)i3rEuμa)i3rEZma)i3rEZka)i3rEZkb)i3rEmb)i3rEZgb)i3rEZcal)i3rEZkcal)i3rEZpf)i3rEZnf)i3rEuμf)i3rEuμg)i3rEZmg)i3rEZkgrrrrr_seg_33p
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i3rEhz)i3rEZkhz)i3rEZmhz)i3rEZghz)i3rEZthz)i3rEuμl)i3rEZml)i3rEZdl)i3rEZkl)i3rEZfm)i3rEZnm)i3rEuμm)i3rEZmm)i3rEcm)i3rEkm)i3rEZmm2)i3rEZcm2)i3rEZm2)i3rEZkm2)i3rEZmm3)i3rEZcm3)i3rEZm3)i3rEZkm3)i3rEum∕s)i3rEum∕s2)i3rEr)i3rEZkpa)i3rEZmpa)i3rEZgpa)i3rEZrad)i3rEurad∕s)i3rEurad∕s2)i3rEZps)i3rEns)i3rEuμs)i3rEZms)i3rEZpv)i3rEZnv)i3rEuμv)i3rEmv)i3rEZkv)i3rEr)i3rEZpw)i3rEZnw)i3rEuμw)i3rEmw)i3rEkw)i3rEr)i3rEukω)i3rEumω)i3r)i3rEZbq)i3rEZcc)i3rEcd)i3rEuc∕kg)i3r)i3rEZdb)i3rEZgy)i3rEZha)i3rEZhp)i3rEin)i3rEZkk)i3rEr)i3rEZkt)i3rEZlm)i3rEZln)i3rElog)i3rEZlx)i3rEr)i3rEZmil)i3rEZmol)i3rEZph)i3r)i3rEZppm)i3rEZpr)i3rEsr)i3rEZsv)i3rEwb)i3rEuv∕m)i3rEua∕m)i3rEu1日)i3rEu2日)i3rEu3日)i3rEu4日)i3rEu5日)i3rEu6日)i3rEu7日)i3rEu8日)i3rEu9日)i3rEu10日)i3rEu11日)i3rEu12日)i3rEu13日)i3rEu14日)i3rEu15日)i3rEu16日)i3rEu17日)i3rEu18日)i3rEu19日)i3rEu20日rrrrr_seg_34
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i3rEu21日)i3rEu22日)i3rEu23日)i3rEu24日)i3rEu25日)i3rEu26日)i3rEu27日)i3rEu28日)i3rEu29日)i3rEu30日)i3rEu31日)i3rEZgal)i4r0)iMr)iMr0)ir)ir0)ir)ir0)iǤr)iФr0)i,r)i@rEuꙁ)iAr0)iBrEuꙃ)iCr0)iDrEuꙅ)iEr0)iFrEuꙇ)iGr0)iHrEuꙉ)iIr0)iJrEr)iKr0)iLrEuꙍ)iMr0)iNrEuꙏ)iOr0)iPrEuꙑ)iQr0)iRrEuꙓ)iSr0)iTrEuꙕ)iUr0)iVrEuꙗ)iWr0)iXrEuꙙ)iYr0)iZrEuꙛ)i[r0)i\rEuꙝ)i]r0)i^rEuꙟ)i_r0)i`rEuꙡ)iar0)ibrEuꙣ)icr0)idrEuꙥ)ier0)ifrEuꙧ)igr0)ihrEuꙩ)iir0)ijrEuꙫ)ikr0)ilrEuꙭ)imr0)irEuꚁ)ir0)irEuꚃ)ir0)irEuꚅ)ir0)irEuꚇ)ir0)irEuꚉ)ir0)irEuꚋ)ir0)irEuꚍ)ir0)irEuꚏ)ir0)irEuꚑ)ir0)irEuꚓ)ir0)irEuꚕ)ir0)irEuꚗ)ir0)irEuꚙ)ir0)irEuꚛ)ir0)irEr)irEr)ir0)irrrrrr_seg_35@srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir0)i"rEuꜣ)i#r0)i$rEuꜥ)i%r0)i&rEꜧ)i'r0)i(rEuꜩ)i)r0)i*rEuꜫ)i+r0)i,rEuꜭ)i-r0)i.rEuꜯ)i/r0)i2rEuꜳ)i3r0)i4rEuꜵ)i5r0)i6rEuꜷ)i7r0)i8rEuꜹ)i9r0)i:rEuꜻ)i;r0)i<rEuꜽ)i=r0)i>rEuꜿ)i?r0)i@rEuꝁ)iAr0)iBrEuꝃ)iCr0)iDrEuꝅ)iEr0)iFrEuꝇ)iGr0)iHrEuꝉ)iIr0)iJrEuꝋ)iKr0)iLrEuꝍ)iMr0)iNrEuꝏ)iOr0)iPrEuꝑ)iQr0)iRrEuꝓ)iSr0)iTrEuꝕ)iUr0)iVrEuꝗ)iWr0)iXrEuꝙ)iYr0)iZrEuꝛ)i[r0)i\rEuꝝ)i]r0)i^rEuꝟ)i_r0)i`rEuꝡ)iar0)ibrEuꝣ)icr0)idrEuꝥ)ier0)ifrEuꝧ)igr0)ihrEuꝩ)iir0)ijrEuꝫ)ikr0)ilrEuꝭ)imr0)inrEꝯ)ior0)iprEr)iqr0)iyrEuꝺ)izr0)i{rEuꝼ)i|r0)i}rEuᵹ)i~rEuꝿ)ir0)irEuꞁ)ir0)irEuꞃ)ir0)irEuꞅ)ir0)irEuꞇ)ir0)irEuꞌ)ir0)irEr)ir0)irEuꞑ)ir0rrrrr_seg_36srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuꞓ)ir0)irEuꞗ)ir0)irEuꞙ)ir0)irEuꞛ)ir0)irEuꞝ)ir0)irEuꞟ)ir0)irEuꞡ)ir0)irEuꞣ)ir0)irEuꞥ)ir0)irEuꞧ)ir0)irEuꞩ)ir0)irErc)irEr)irEr)irEuɬ)irEr)ir0)irEuʞ)irEuʇ)irEr)irEuꭓ)irEuꞵ)ir0)irEuꞷ)ir0)ir)ir0)ir)ir0)irErK)irErN)ir0)i,r)i0r0)i:r)i@r0)ixr)ir0)iƨr)iΨr0)iڨr)ir0)iTr)i_r0)i}r)ir0)iΩr)iϩr0)iکr)iީr0)ir)ir0)i7r)i@r0)iNr)iPr0)iZr)i\r0)iêr)i۪r0)ir)ir0)ir)i	r0)ir)ir0)ir)i r0)i'r)i(r0)i/r)i0r0)i\rEr)i]rEuꬷ)i^rEr)i_rEuꭒ)i`r0)ifr)iprEuᎠ)iqrEuᎡ)irrEuᎢ)isrEuᎣ)itrEuᎤ)iurEuᎥ)ivrEuᎦ)iwrEuᎧ)ixrEuᎨ)iyrEuᎩ)izrEuᎪrrrrr_seg_37srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i{rEuᎫ)i|rEuᎬ)i}rEuᎭ)i~rEuᎮ)irEuᎯ)irEuᎰ)irEuᎱ)irEuᎲ)irEuᎳ)irEuᎴ)irEuᎵ)irEuᎶ)irEuᎷ)irEuᎸ)irEuᎹ)irEuᎺ)irEuᎻ)irEuᎼ)irEuᎽ)irEuᎾ)irEuᎿ)irEuᏀ)irEuᏁ)irEuᏂ)irEuᏃ)irEuᏄ)irEuᏅ)irEuᏆ)irEuᏇ)irEuᏈ)irEuᏉ)irEuᏊ)irEuᏋ)irEuᏌ)irEuᏍ)irEuᏎ)irEuᏏ)irEuᏐ)irEuᏑ)irEuᏒ)irEuᏓ)irEuᏔ)irEuᏕ)irEuᏖ)irEuᏗ)irEuᏘ)irEuᏙ)irEuᏚ)irEuᏛ)irEuᏜ)irEuᏝ)irEuᏞ)irEuᏟ)irEuᏠ)irEuᏡ)irEuᏢ)irEuᏣ)irEuᏤ)irEuᏥ)irEuᏦ)irEuᏧ)irEuᏨ)irEuᏩ)irEuᏪ)irEuᏫ)irEuᏬ)irEuᏭ)irEuᏮ)irEuᏯ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irEu豈)irEu更)irEr$)irEu賈)irEu滑)irEu串)irEu句)irEr-)i	rEu契)i
rEr')irEu喇)irEu奈)i
rE懶)irEu癩)irEu羅)irEu蘿)irEu螺)irEu裸)irEu邏)irE樂)irEu洛rrrrr_seg_38xsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu烙)irEu珞)irEu落)irEu酪)irEu駱)irEu亂)irEu卵)irEu欄)irEu爛)irEu蘭)i rEu鸞)i!rEu嵐)i"rEu濫)i#rEu藍)i$rEu襤)i%rEu拉)i&rEu臘)i'rEu蠟)i(rE廊)i)rE朗)i*rEu浪)i+rEu狼)i,rEu郎)i-rEu來)i.rEu冷)i/rEu勞)i0rEu擄)i1rEu櫓)i2rEu爐)i3rEu盧)i4rEr)i5rEu蘆)i6rE虜)i7rEu路)i8rEu露)i9rEu魯)i:rEu鷺)i;rE碌)i<rEu祿)i=rEu綠)i>rEu菉)i?rEu錄)i@rEr()iArEu論)iBrEu壟)iCrEu弄)iDrEu籠)iErEu聾)iFrEu牢)iGrEu磊)iHrEu賂)iIrEu雷)iJrEu壘)iKrEu屢)iLrEu樓)iMrEu淚)iNrEu漏)iOrEu累)iPrEu縷)iQrEu陋)iRrEu勒)iSrEu肋)iTrEu凜)iUrEu凌)iVrEu稜)iWrEu綾)iXrEu菱)iYrEu陵)iZrEu讀)i[rEu拏)i\rEr)i]rE諾)i^rEu丹)i_rE寧)i`rEu怒)iarE率)ibrE異)icrE北)idrEu磻)ierEu便)ifrEu復)igrEu不)ihrEu泌)iirEu數)ijrEu索)ikrEu參)ilrEu塞)imrEu省)inrEu葉)iorE說)iprE殺)iqrEr%)irrEu沈)isrEu拾)itrE若)iurEu掠)ivrEu略)iwrEu亮)ixrEu兩)iyrEu凉rrrrr_seg_39srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)izrEu梁)i{rEu糧)i|rEu良)i}rEu諒)i~rEu量)irEu勵)irEu呂)irEr
)irEu廬)irEu旅)irEu濾)irEu礪)irEu閭)irEu驪)irEu麗)irEu黎)irEr
)irEu曆)irEu歷)irEu轢)irEu年)irEu憐)irEu戀)irEu撚)irEu漣)irEu煉)irEu璉)irEu秊)irE練)irEu聯)irEu輦)irEu蓮)irEu連)irEu鍊)irEu列)irEu劣)irEu咽)irEu烈)irEu裂)irEr)irEu廉)irEu念)irEu捻)irEu殮)irEu簾)irEu獵)irEu令)irEu囹)irEr)irEu嶺)irEu怜)irEu玲)irEu瑩)irEu羚)irEu聆)irEu鈴)irEu零)irEu靈)irEu領)irEu例)irEu禮)irEu醴)irEu隸)irEu惡)irEu了)irEu僚)irEu寮)irEu尿)irE料)irEr)irEu燎)irEu療)irEu蓼)irEu遼)irEr,)irEu暈)irEu阮)irEu劉)irEu杻)irEu柳)irE流)irEu溜)irEu琉)irEu留)irEu硫)irEu紐)irEu類)irErk)irEu戮)irEu陸)irEu倫)irEu崙)irEu淪)irEu輪)irEu律)irEu慄)irEu栗)irEr)irEu隆)irEu利rrrrr_seg_40HsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu吏)irEu履)irEu易)irEu李)irEu梨)irEu泥)irEu理)irEu痢)irEu罹)irEu裏)irEu裡)irEr&)irEu離)irEu匿)irEu溺)irEu吝)irEu燐)irEu璘)irEu藺)irEu隣)irEu鱗)irEu麟)irEu林)irEu淋)irEu臨)irEr)irEu笠)irEu粒)irEu狀)irEu炙)irEu識)irEu什)irEu茶)irEu刺)irE切)irEu度)irEu拓)irEu糖)irEu宅)irEu洞)irEu暴)irEu輻)irEr)i	rEu降)i
rEr!)irEu廓)irEu兀)i
rEu嗀)ir0)irE塚)ir0)irE晴)ir0)irEu凞)irE猪)irE益)irEu礼)irEu神)irEu祥)irE福)irE靖)irEu精)irEr)ir0)i rEu蘒)i!r0)i"rE諸)i#r0)i%rE逸)i&rEu都)i'r0)i*rEu飯)i+rEu飼)i,rEu館)i-rEu鶴)i.rEu郞)i/rEu隷)i0rE侮)i1rE僧)i2rE免)i3rE勉)i4rE勤)i5rE卑)i6rE喝)i7rE嘆)i8rEu器)i9rEu塀)i:rEu墨)i;rEu層)i<rEr)i=rE悔)i>rEu慨)i?rE憎)i@rE懲)iArE敏)iBrEu既)iCrE暑)iDrE梅)iErE海)iFrEu渚rrrrr_seg_41srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iGrE漢)iHrE煮)iIrEu爫)iJrEu琢)iKrEu碑)iLrErm)iMrEu祉)iNrEu祈)iOrEu祐)iPrE祖)iQrErn)iRrEu禍)iSrEu禎)iTrE穀)iUrEu突)iVrE節)iWrEr)iXrEu縉)iYrEu繁)iZrEu署)i[rE者)i\rEu臭)i]rEu艹)i_rE著)i`rEu褐)iarE視)ibrE謁)icrE謹)idrEu賓)ierE贈)ifrEu辶)igrEr)ihrE難)iirE響)ijrE頻)ikrEu恵)ilrEu𤋮)imrEu舘)inr)iprEu並)iqrE况)irrEu全)isrEu侀)itrEu充)iurEu冀)ivrE勇)iwrE勺)ixrEr)iyrEu啕)izrE喙)i{rEu嗢)i|rEr)i}rEu墳)i~rEu奄)irEu奔)irEu婢)irEu嬨)irEu廒)irEu廙)irEu彩)irEu徭)irEu惘)irE慎)irEu愈)irEr)irEu慠)irEr)irEu戴)irEu揄)irEu搜)irEu摒)irEu敖)irEr)irEr)irE望)irEu杖)irEr)irEr)irEr)irEu滛)irE滋)irEr)irE瀞)irEr)irEu瞧)irE爵)irEu犯)irEr)irE瑱)irEu甆)irEu画)irEu瘝)irEu瘟)irEr)irEu盛)irE直)irE睊)irEu着)irE磌)irEu窱rrrrr_seg_42srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEu类)irEu絛)irEr)irEu缾)irEr)irEu荒)irEu華)irE蝹)irEu襁)irEu覆)irEr)irEu調)irEr)irEu請)irEr)irEr)irE諭)irEr)irE變)irEr)irE輸)irEu遲)irEu醙)irEu鉶)irEu陼)irEr)irEr)irEu韛)irEr)irE頋)irEr)irE鬒)irEr-)irEu𢡊)irEu𢡄)irEu𣏕)irE㮝)irEu䀘)irE䀹)irEu𥉉)irEu𥳐)irEu𧻓)irEu齃)irEu龎)ir)irEZff)irEZfi)irEZfl)irEZffi)irEZffl)irEst)ir)irEuմն)irEuմե)irEuմի)irEuվն)irEuմխ)ir)irEuיִ)ir0)irEuײַ)i rEuע)i!rEr)i"rEr)i#rEuה)i$rEuכ)i%rEuל)i&rEuם)i'rEuר)i(rEuת)i)rr)i*rEuשׁ)i+rEuשׂ)i,rEuשּׁ)i-rEuשּׂ)i.rEuאַ)i/rEuאָ)i0rEuאּ)i1rEuבּ)i2rEuגּ)i3rEuדּ)i4rEuהּ)i5rEuוּ)i6rEuזּ)i7r)i8rEuטּ)i9rEuיּ)i:rEuךּ)i;rEuכּ)i<rEuלּ)i=r)i>rEuמּ)i?r)i@rEuנּ)iArEuסּ)iBr)iCrEuףּ)iDrEuפּ)iErrrrrr_seg_43srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iFrEuצּ)iGrEuקּ)iHrEuרּ)iIrEuשּ)iJrEuתּ)iKrEuוֹ)iLrEuבֿ)iMrEuכֿ)iNrEuפֿ)iOrEuאל)iPrEuٱ)iRrEuٻ)iVrEuپ)iZrEuڀ)i^rEuٺ)ibrEuٿ)ifrEuٹ)ijrEuڤ)inrEuڦ)irrEuڄ)ivrEuڃ)izrEuچ)i~rEuڇ)irEuڍ)irEuڌ)irEuڎ)irEuڈ)irEuژ)irEuڑ)irEuک)irEuگ)irEuڳ)irEuڱ)irEں)irEuڻ)irEuۀ)irEuہ)irEuھ)irEuے)irEuۓ)ir0)ir)irEuڭ)irEuۇ)irEuۆ)irEuۈ)irEr)irEuۋ)irEuۅ)irEuۉ)irEuې)irEى)irEuئا)irEuئە)irEuئو)irEuئۇ)irEuئۆ)irEuئۈ)irEuئې)irEئى)irEuی)irEئج)irEئح)irEئم)irEr
)irEئي)irEبج)irEبح)irEبخ)irEبم)i	rEبى)i
rEبي)irEتج)irEتح)i
rEتخ)irEتم)irEتى)irEتي)irEuثج)irEثم)irEثى)irEثي)irEجح)irEجم)irEحج)irEحم)irEخج)irEuخح)irEخم)irEسج)irEسح)irEسخ)irEسم)i rEصح)i!rEصم)i"rEضج)i#rEضح)i$rEضخ)i%rEضم)i&rEطحrrrrr_seg_44sr/cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i'rEطم)i(rEظم)i)rEعج)i*rEعم)i+rEغج)i,rEغم)i-rEفج)i.rEفح)i/rEفخ)i0rEفم)i1rEفى)i2rEفي)i3rEقح)i4rEقم)i5rEقى)i6rEقي)i7rEكا)i8rEكج)i9rEكح)i:rEكخ)i;rEكل)i<rEكم)i=rEكى)i>rEكي)i?rEلج)i@rEلح)iArEلخ)iBrEلم)iCrEلى)iDrEلي)iErEمج)iFrEمح)iGrEمخ)iHrEمم)iIrEuمى)iJrEuمي)iKrEنج)iLrEنح)iMrEنخ)iNrEنم)iOrEنى)iPrEني)iQrEهج)iRrEهم)iSrEuهى)iTrEuهي)iUrEيج)iVrEيح)iWrEيخ)iXrEيم)iYrEيى)iZrEيي)i[rEuذٰ)i\rEuرٰ)i]rEىٰ)i^ru ٌّ)i_ru ٍّ)i`ru َّ)iaru ُّ)ibru ِّ)icru ّٰ)idrEuئر)ierEuئز)ifrEr
)igrEuئن)ihrEr
)iirEr)ijrEuبر)ikrEuبز)ilrEr)imrEuبن)inrEr)iorEr)iprEuتر)iqrEuتز)irrEr)isrEuتن)itrEr)iurEr)ivrEuثر)iwrEuثز)ixrEr)iyrEuثن)izrEr)i{rEr)i|rEr:)i}rEr;)i~rEr>)irEr?)irEr@)irErD)irErE)irErF)irErG)irErK)irErL)irErM)irEuما)irErQ)irEuنرrrrrr_seg_45PsracdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuنز)irErU)irEuنن)irErV)irErW)irEr`)irEuير)irEuيز)irEr])irEuين)irEr^)irEr_)irEr)irEr)irEuئخ)irEr
)irEئه)irEr)irEr)irEr)irEr)irEبه)irEr)irEr)irEr)irEr)irEته)irEr)irEr)irEr)irEr )irEr!)irEr")irEr#)irEr$)irEr%)irEr&)irEr')irEr()irEuصخ)irEr))irEr*)irEr+)irEr,)irEr-)irEr.)irEr1)irEr2)irEr3)irEr4)irEr5)irEr6)irEr7)irEr8)irEr9)irEr<)irEr=)irErA)irErB)irErC)irErD)irErE)irErH)irErI)irErJ)irErK)irEuله)irErN)irErO)irErP)irErQ)irErR)irErS)irErT)irErU)irEنه)irErX)irErY)irEuهٰ)irErZ)irEr[)irEr\)irEr])irEيه)irEr
)irErb)irEr)irErc)irEr)irErd)irEr)irEuثه)irEr')irEسه)irEشم)irEشه)irErD)irErE)irErK)irErUrrrrr_seg_46srjcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEre)irEr])irErf)irEuـَّ)irEuـُّ)irEuـِّ)irEطى)irEطي)irEعى)irEعي)irEغى)irEغي)irEسى)irEسي)irEشى)irEشي)irEحى)irEحي)irEجى)irEجي)irEخى)irEخي)irEصى)irEصي)irEضى)irEضي)i	rEشج)i
rEشح)irEشخ)irErh)i
rEشر)irEسر)irEصر)irEضر)irErk)irErl)irErm)irErn)irEro)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErx)irEry)i rErz)i!rEr{)i"rEr|)i#rEr})i$rEr~)i%rEr)i&rEr)i'rEr)i(rErh)i)rEr)i*rEr)i+rEr)i,rEr)i-rEr)i.rEr)i/rEr)i0rErh)i1rErg)i2rEri)i3rEr0)i4rEr$)i5rEr%)i6rEr&)i7rEr)i8rEr)i9rEr)i:rEr0)i;rEr1)i<rEuاً)i>r0)i@r)iPrEuتجم)iQrEuتحج)iSrEuتحم)iTrEuتخم)iUrEuتمج)iVrEuتمح)iWrEuتمخ)iXrEuجمح)iZrEuحمي)i[rEuحمى)i\rEuسحج)i]rEuسجح)i^rEuسجى)i_rEuسمح)iarEuسمج)ibrEuسمم)idrEuصحح)ifrEصمم)igrEuشحم)iirEuشجيrrrrr_seg_47 srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ijrEuشمخ)ilrEuشمم)inrEuضحى)iorEuضخم)iqrEuطمح)isrEuطمم)itrEuطمي)iurEعجم)ivrEuعمم)ixrEuعمى)iyrEuغمم)izrEuغمي)i{rEuغمى)i|rEuفخم)i~rEقمح)irEuقمم)irEلحم)irEuلحي)irEuلحى)irEuلجج)irEuلخم)irEuلمح)irEuمحج)irEuمحم)irEuمحي)irEuمجح)irEuمجم)irEuمخج)irEuمخم)ir)irEuمجخ)irEuهمج)irEuهمم)irEuنحم)irEuنحى)irEuنجم)irEuنجى)irEuنمي)irEuنمى)irEuيمم)irEuبخي)irEuتجي)irEuتجى)irEuتخي)irEuتخى)irEuتمي)irEuتمى)irEuجمي)irEuجحى)irEuجمى)irEuسخى)irEuصحي)irEuشحي)irEuضحي)irEuلجي)irEuلمي)irEuيحي)irEuيجي)irEuيمي)irEuممي)irEuقمي)irEuنحي)irEr)irEr)irEuعمي)irEuكمي)irEنجح)irEuمخي)irEلجم)irEكمم)irEr)irEr)irEuجحي)irEuحجي)irEuمجي)irEuفمي)irEuبحي)irEr)irEr)irEr)irEuسخي)irEuنجي)ir)irEuصلے)irEuقلے)irEuالله)irEuاكبر)irEuمحمد)irEuصلعم)irEuرسول)irEuعليه)irEuوسلم)irEuصلى)iru!صلى الله عليه وسلم)iruجل جلاله)irEuریال)ir0)ir)ir)ir,rrrrr_seg_48srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irE、)ir)ir:)irrf)ir!)ir?)irEu〖)irEu〗)ir)i r0)i0r)i1rE—)i2rEu–)i3r_)i5rr)i6rr)i7r{)i8r})i9rE〔)i:rE〕)i;rEu【)i<rEu】)i=rEu《)i>rEu》)i?rEr)i@rEr)iArE「)iBrE」)iCrEu『)iDrEu』)iEr0)iGr[)iHr])iIrr)iMrr)iPrr)iQrEr)iRr)iTrrf)iUrr)iVrr)iWrr)iXrEr)iYrr)iZrr)i[rr)i\rr)i]rEr)i^rEr)i_r#)i`r&)iar*)ibrr)icrE-)idr<)ier>)ifrr)igr)ihr\)iir$)ijr%)ikr@)ilr)ipru ً)iqrEuـً)irru ٌ)isr0)itru ٍ)iur)ivru َ)iwrEuـَ)ixru ُ)iyrEuـُ)izru ِ)i{rEuـِ)i|ru ّ)i}rEuـّ)i~ru ْ)irEuـْ)irEuء)irEuآ)irEuأ)irEuؤ)irEuإ)irEuئ)irEا)irEب)irEuة)irEت)irEث)irEج)irEح)irEخ)irEد)irEذ)irEر)irEز)irEس)irEش)irEصrrrrr_seg_49srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEض)irEط)irEظ)irEع)irEغ)irEف)irEق)irEك)irEل)irEم)irEن)irEه)irEو)irEr	)irEي)irEuلآ)irEuلأ)irEuلإ)irEuلا)ir)ir)ir)irr)ir")irr)irr)irr)irr)ir')irr)i	rr)i
rr)irr)irr)i
rEr)irEr.)ir/)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irr)irrf)irr)irr)irr)irr)i rr)i!rErF)i"rErH)i#rErJ)i$rErL)i%rErN)i&rErP)i'rErR)i(rErT)i)rErV)i*rErX)i+rErZ)i,rEr\)i-rEr^)i.rEr`)i/rErb)i0rErd)i1rErf)i2rErh)i3rErj)i4rErl)i5rErn)i6rErp)i7rErr)i8rErt)i9rErv)i:rErx)i;rr)i<rr)i=rr)i>r^)i?rr)i@rr)iArErF)iBrErH)iCrErJ)iDrErL)iErErN)iFrErP)iGrErR)iHrErT)iIrErV)iJrErX)iKrErZ)iLrEr\)iMrEr^)iNrEr`rrrrr_seg_50XsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iOrErb)iPrErd)iQrErf)iRrErh)iSrErj)iTrErl)iUrErn)iVrErp)iWrErr)iXrErt)iYrErv)iZrErx)i[rr)i\r|)i]rr)i^r~)i_rEu⦅)i`rEu⦆)iarEr.)ibrEr)icrEr)idrEr)ierEu・)ifrEr)igrEuァ)ihrEuィ)iirEuゥ)ijrEuェ)ikrEuォ)ilrEuャ)imrEuュ)inrEuョ)iorEuッ)iprEuー)iqrErr)irrErs)isrErt)itrEru)iurErv)ivrErw)iwrErx)ixrEry)iyrErz)izrEr{)i{rEr|)i|rEr})i}rEr~)i~rEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEuン)irEu゙)irEu゚)ir)irEr/)irEr0)irEr1)irEr2)irEr3)irEr4)irEr5)irEr6)irEr8)irEr9)irEr:)irEr;)irEr<)irEr=)irEr>)irEr?)irEr@)irErArrrrr_seg_51srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irErB)irErC)irErD)irErE)irErF)irErG)irErH)irErI)irErJ)irErK)irErL)irErM)ir)irErN)irErO)irErP)irErQ)irErR)irErS)ir)irErT)irErU)irErV)irErW)irErX)irErY)ir)irErZ)irEr[)irEr\)irEr])irEr^)irEr_)ir)irEr`)irEra)irErb)ir)irE¢)irE£)irE¬)irr)irE¦)irE¥)irEu₩)ir)irEu│)irEu←)irEu↑)irEu→)irEu↓)irEu■)irEu○)ir)ir0)ir)i
r0)i'r)i(r0)i;r)i<r0)i>r)i?r0)iNr)iPr0)i^r)ir0)ir)ir0)ir)ir0)i4r)i7r0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i$r)i-r0)iKr)iPr0)i{r)ir0)ir)ir0)ir)ir0)ir)irEu𐐨)irEu𐐩rrrrr_seg_52(srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu𐐪)irEu𐐫)irEu𐐬)irEu𐐭)irEu𐐮)irEu𐐯)irEu𐐰)i	rEu𐐱)i
rEu𐐲)irEu𐐳)irEu𐐴)i
rEu𐐵)irEu𐐶)irEu𐐷)irEu𐐸)irEu𐐹)irEu𐐺)irEu𐐻)irEu𐐼)irEu𐐽)irEu𐐾)irEu𐐿)irEu𐑀)irEu𐑁)irEu𐑂)irEu𐑃)irEu𐑄)irEu𐑅)irEu𐑆)irEu𐑇)i rEu𐑈)i!rEu𐑉)i"rEu𐑊)i#rEu𐑋)i$rEu𐑌)i%rEu𐑍)i&rEu𐑎)i'rEu𐑏)i(r0)ir)ir0)ir)irEu𐓘)irEu𐓙)irEu𐓚)irEu𐓛)irEu𐓜)irEu𐓝)irEu𐓞)irEu𐓟)irEu𐓠)irEu𐓡)irEu𐓢)irEu𐓣)irEu𐓤)irEu𐓥)irEu𐓦)irEu𐓧)irEu𐓨)irEu𐓩)irEu𐓪)irEu𐓫)irEu𐓬)irEu𐓭)irEu𐓮)irEu𐓯)irEu𐓰)irEu𐓱)irEu𐓲)irEu𐓳)irEu𐓴)irEu𐓵)irEu𐓶)irEu𐓷)irEu𐓸)irEu𐓹)irEu𐓺)irEu𐓻)ir)ir0)ir)ir0)i(r)i0r0)idr)ior0)ipr)ir0)i7r)i@r0)iVr)i`r0)ihr)ir0)ir)ir0)i	r)i
r0)i6r)i7r0rrrrr_seg_53srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i9r)i<r0)i=r)i?r0)iVr)iWr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i	r)i	r0)i:	r)i?	r0)i@	r)i	r0)i	r)i	r0)i	r)i	r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i6
r)i8
r0)i;
r)i?
r0)iI
r)iP
r0)iY
r)i`
r0)i
r)i
r0)i
r)i
r0)i
r)ir0)i6r)i9r0)iVr)iXr0)isr)ixr0)ir)ir0)ir)ir0)ir)ir0)iIr)irEu𐳀)irEu𐳁)irEu𐳂)irEu𐳃)irEu𐳄)irEu𐳅)irEu𐳆)irEu𐳇)irEu𐳈)irEu𐳉)irEu𐳊)irEu𐳋)irEu𐳌)irEu𐳍)irEu𐳎)irEu𐳏)irEu𐳐)irEu𐳑)irEu𐳒)irEu𐳓)irEu𐳔)irEu𐳕)irEu𐳖)irEu𐳗)irEu𐳘)irEu𐳙)irEu𐳚)irEu𐳛)irEu𐳜)irEu𐳝)irEu𐳞)irEu𐳟)irEu𐳠)irEu𐳡)irEu𐳢)irEu𐳣)irEu𐳤)irEu𐳥)irEu𐳦)irEu𐳧)irEu𐳨rrrrr_seg_54srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu𐳩)irEu𐳪)irEu𐳫)irEu𐳬)irEu𐳭)irEu𐳮)irEu𐳯)irEu𐳰)irEu𐳱)irEu𐳲)ir)ir0)ir)ir0)i(
r)i0
r0)i:
r)i`r0)ir)ir0)i(r)i0r0)iZr)ir0)iNr)iRr0)ipr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i5r)i6r0)iGr)iPr0)iwr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i?r)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)ir)ir0)i)r)i*r0)i1r)i2r0)i4r)i5r0)i:r)i;r0)iEr)iGr0)iIr)iKr0)iNr)iPr0)iQr)iWr0)iXr)i]r0)idr)ifr0)imr)ipr0)iur)ir0)iZr)i[r0)i\r)i]r0rrrrr_seg_55`srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i_r)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)iEr)iPr0)iZr)i`r0)imr)ir0)ir)ir0)ir)ir0)ir)ir0)i,r)i0r0)i@r)ir0)i<r)irEu𑣀)irEu𑣁)irEu𑣂)irEu𑣃)irEu𑣄)irEu𑣅)irEu𑣆)irEu𑣇)irEu𑣈)irEu𑣉)irEu𑣊)irEu𑣋)irEu𑣌)irEu𑣍)irEu𑣎)irEu𑣏)irEu𑣐)irEu𑣑)irEu𑣒)irEu𑣓)irEu𑣔)irEu𑣕)irEu𑣖)irEu𑣗)irEu𑣘)irEu𑣙)irEu𑣚)irEu𑣛)irEu𑣜)irEu𑣝)irEu𑣞)irEu𑣟)ir0)ir)ir0)ir)ir0)iHr)iPr0)ir)ir0)ir)ir0)ir)ir0)i	r)i
r0)i7r)i8r0)iFr)iPr0)imr)ipr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)i7r)i:r0)i;r)i<r0)i>r)i?r0)iHr)iPr0)iZr)i`r0rrrrr_seg_56srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ifr)igr0)iir)ijr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)i r0)i#r)i$r0)io$r)ip$r0)iu$r)i$r0)iD%r)i0r0)i/4r)iDr0)iGFr)ihr0)i9jr)i@jr0)i_jr)i`jr0)ijjr)injr0)ipjr)ijr0)ijr)ijr0)ijr)ikr0)iFkr)iPkr0)iZkr)i[kr0)ibkr)ickr0)ixkr)i}kr0)ikr)i`nr0)inr)ior0)iEor)iPor0)ior)ior0)ior)ior0)ior)ipr0)ir)ir0)ir)ir0)ir)ipr0)ir)ir0)ikr)ipr0)i}r)ir0)ir)ir0)ir)ir0)ir)ir)ir0)ir)ir0)i'r)i)r0)i^rEu𝅗𝅥)i_rEu𝅘𝅥)i`rEu𝅘𝅥𝅮)iarEu𝅘𝅥𝅯)ibrEu𝅘𝅥𝅰)icrEu𝅘𝅥𝅱)idrEu𝅘𝅥𝅲)ier0)isr)i{r0)irEu𝆹𝅥)irEu𝆺𝅥)irEu𝆹𝅥𝅮)irEu𝆺𝅥𝅮)irEu𝆹𝅥𝅯)irEu𝆺𝅥𝅯)ir0)ir)ir0rrrrr_seg_570srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iFr)ir0)ir)ir0)iWr)i`r0)iyr)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)i	rErX)i
rErZ)irEr\)irEr^)i
rEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)i rErR)i!rErT)i"rErV)i#rErX)i$rErZ)i%rEr\)i&rEr^)i'rEr`)i(rErb)i)rErd)i*rErf)i+rErh)i,rErj)i-rErl)i.rErn)i/rErp)i0rErr)i1rErt)i2rErv)i3rErx)i4rErF)i5rErH)i6rErJ)i7rErL)i8rErN)i9rErP)i:rErR)i;rErT)i<rErV)i=rErX)i>rErZ)i?rEr\)i@rEr^)iArEr`)iBrErb)iCrErd)iDrErf)iErErh)iFrErj)iGrErl)iHrErn)iIrErp)iJrErr)iKrErt)iLrErv)iMrErx)iNrErF)iOrErH)iPrErJ)iQrErL)iRrErN)iSrErP)iTrErR)iUr)iVrErV)iWrErX)iXrErZ)iYrEr\)iZrEr^)i[rEr`)i\rErbrrrrr_seg_58srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i]rErd)i^rErf)i_rErh)i`rErj)iarErl)ibrErn)icrErp)idrErr)ierErt)ifrErv)igrErx)ihrErF)iirErH)ijrErJ)ikrErL)ilrErN)imrErP)inrErR)iorErT)iprErV)iqrErX)irrErZ)isrEr\)itrEr^)iurEr`)ivrErb)iwrErd)ixrErf)iyrErh)izrErj)i{rErl)i|rErn)i}rErp)i~rErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)ir)irErJ)irErL)ir)irErR)ir)irErX)irErZ)ir)irEr`)irErb)irErd)irErf)ir)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)ir)irErP)ir)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`rrrrr_seg_59srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)ir)irErL)irErN)i	rErP)i
rErR)ir)i
rErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)ir)irErj)irErl)irErn)irErp)irErr)irErt)irErv)ir)irErF)irErH)i rErJ)i!rErL)i"rErN)i#rErP)i$rErR)i%rErT)i&rErV)i'rErX)i(rErZrrrrr_seg_60hsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i)rEr\)i*rEr^)i+rEr`)i,rErb)i-rErd)i.rErf)i/rErh)i0rErj)i1rErl)i2rErn)i3rErp)i4rErr)i5rErt)i6rErv)i7rErx)i8rErF)i9rErH)i:r)i;rErL)i<rErN)i=rErP)i>rErR)i?r)i@rErV)iArErX)iBrErZ)iCrEr\)iDrEr^)iEr)iFrErb)iGr)iJrErj)iKrErl)iLrErn)iMrErp)iNrErr)iOrErt)iPrErv)iQr)iRrErF)iSrErH)iTrErJ)iUrErL)iVrErN)iWrErP)iXrErR)iYrErT)iZrErV)i[rErX)i\rErZ)i]rEr\)i^rEr^)i_rEr`)i`rErb)iarErd)ibrErf)icrErh)idrErj)ierErl)ifrErn)igrErp)ihrErr)iirErt)ijrErv)ikrErx)ilrErF)imrErH)inrErJ)iorErL)iprErN)iqrErP)irrErR)isrErT)itrErV)iurErX)ivrErZ)iwrEr\)ixrEr^)iyrEr`)izrErb)i{rErd)i|rErf)i}rErh)i~rErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErVrrrrr_seg_61srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErNrrrrr_seg_628srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)i	rErH)i
rErJ)irErL)irErN)i
rErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)i rErv)i!rErx)i"rErF)i#rErH)i$rErJ)i%rErL)i&rErN)i'rErP)i(rErR)i)rErT)i*rErV)i+rErX)i,rErZ)i-rEr\)i.rEr^)i/rEr`)i0rErb)i1rErd)i2rErf)i3rErh)i4rErj)i5rErl)i6rErn)i7rErp)i8rErr)i9rErt)i:rErv)i;rErx)i<rErF)i=rErH)i>rErJ)i?rErL)i@rErN)iArErP)iBrErR)iCrErT)iDrErV)iErErX)iFrErZ)iGrEr\)iHrEr^)iIrEr`)iJrErb)iKrErd)iLrErf)iMrErh)iNrErj)iOrErl)iPrErn)iQrErp)iRrErr)iSrErt)iTrErv)iUrErx)iVrErFrrrrr_seg_63srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iWrErH)iXrErJ)iYrErL)iZrErN)i[rErP)i\rErR)i]rErT)i^rErV)i_rErX)i`rErZ)iarEr\)ibrEr^)icrEr`)idrErb)ierErd)ifrErf)igrErh)ihrErj)iirErl)ijrErn)ikrErp)ilrErr)imrErt)inrErv)iorErx)iprErF)iqrErH)irrErJ)isrErL)itrErN)iurErP)ivrErR)iwrErT)ixrErV)iyrErX)izrErZ)i{rEr\)i|rEr^)i}rEr`)i~rErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irEuı)irEuȷ)ir)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irErw)irEr)irErrrrrr_seg_64srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irE∇)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irE∂)irErt)irErw)irErx)irEr)irEr~)irEr})irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irErw)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)i	rEr{)i
rEr|)irEr})irEr~)i
rEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErt)irErw)irErx)irEr)irEr~)irEr})irErp)irErq)irErr)irErs)i rErt)i!rErurrrrr_seg_65psrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i"rErv)i#rErw)i$rErd)i%rErx)i&rEry)i'rEr)i(rErz)i)rEr{)i*rEr|)i+rEr})i,rEr~)i-rErw)i.rEr)i/rEr)i0rEr)i1rEr)i2rEr)i3rEr)i4rEr)i5rEr)i6rErp)i7rErq)i8rErr)i9rErs)i:rErt)i;rEru)i<rErv)i=rErw)i>rErd)i?rErx)i@rEry)iArEr)iBrErz)iCrEr{)iDrEr|)iErEr})iFrEr~)iGrEr)iIrEr)iJrEr)iKrEr)iLrEr)iMrEr)iNrEr)iOrEr)iPrErt)iQrErw)iRrErx)iSrEr)iTrEr~)iUrEr})iVrErp)iWrErq)iXrErr)iYrErs)iZrErt)i[rEru)i\rErv)i]rErw)i^rErd)i_rErx)i`rEry)iarEr)ibrErz)icrEr{)idrEr|)ierEr})ifrEr~)igrErw)ihrEr)iirEr)ijrEr)ikrEr)ilrEr)imrEr)inrEr)iorEr)iprErp)iqrErq)irrErr)isrErs)itrErt)iurEru)ivrErv)iwrErw)ixrErd)iyrErx)izrEry)i{rEr)i|rErz)i}rEr{)i~rEr|)irEr})irEr~)irEr)irEr)irEr)irEr)irEr)irErrrrrr_seg_66srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irErt)irErw)irErx)irEr)irEr~)irEr})irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irErw)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErt)irErw)irErx)irEr)irEr~)irEr})irEr)ir)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErrrrrr_seg_67@srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i"r)i#r0)i%r)i&r0)i+r)ir0)ir)ir0)ir)irEu𞤢)irEu𞤣)irEu𞤤)irEu𞤥)irEu𞤦)irEu𞤧)irEu𞤨)irEu𞤩)irEu𞤪)i	rEu𞤫)i
rEu𞤬)irEu𞤭)irEu𞤮)i
rEu𞤯)irEu𞤰)irEu𞤱)irEu𞤲)irEu𞤳)irEu𞤴)irEu𞤵)irEu𞤶)irEu𞤷)irEu𞤸)irEu𞤹)irEu𞤺)irEu𞤻)irEu𞤼)irEu𞤽)irEu𞤾)irEu𞤿)irEu𞥀)irEu𞥁)i rEu𞥂)i!rEu𞥃)i"r0)iKr)iPr0)iZr)i^r0)i`r)iqr0)ir)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)i	rEr)i
rEr)irEr)irEr)i
rEr)irEr)irEr)irEr)irEr)irEr)irEr)irErrrrrr_seg_68srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEٮ)irEr)irEڡ)irEٯ)i r)i!rEr)i"rEr)i#r)i$rEr)i%r)i'rEr)i(r)i)rEr)i*rEr)i+rEr)i,rEr)i-rEr)i.rEr)i/rEr)i0rEr)i1rEr)i2rEr)i3r)i4rEr)i5rEr)i6rEr)i7rEr)i8r)i9rEr)i:r)i;rEr)i<r)iBrEr)iCr)iGrEr)iHr)iIrEr)iJr)iKrEr)iLr)iMrEr)iNrEr)iOrEr)iPr)iQrEr)iRrEr)iSr)iTrEr)iUr)iWrEr)iXr)iYrEr)iZr)i[rEr)i\r)i]rEr)i^r)i_rEr)i`r)iarEr)ibrEr)icr)idrEr)ier)igrEr)ihrEr)iirEr)ijrEr)ikr)ilrEr)imrEr)inrEr)iorEr)iprEr)iqrEr)irrEr)isr)itrEr)iurEr)ivrEr)iwrEr)ixr)iyrEr)izrEr)i{rEr)i|rEr)i}r)i~rEr)ir)irEr)irEr)irEr)irErrrrrr_seg_69srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir)ir0)ir)ir0)i,r)i0r0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irz0,)irz1,)irz2,)irz3,)irz4,)irz5,)irz6,)irz7,)i	rz8,)i
rz9,)ir0)i
r)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)i rr)i!rr)i"rr)i#rr)i$rrrrrrr_seg_70xsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i%rr)i&rr)i'rr)i(rr)i)rr)i*rEu〔s〕)i+rErJ)i,rErh)i-rEr)i.rEZwz)i/r0)i0rErF)i1rErH)i2rErJ)i3rErL)i4rErN)i5rErP)i6rErR)i7rErT)i8rErV)i9rErX)i:rErZ)i;rEr\)i<rEr^)i=rEr`)i>rErb)i?rErd)i@rErf)iArErh)iBrErj)iCrErl)iDrErn)iErErp)iFrErr)iGrErt)iHrErv)iIrErx)iJrEZhv)iKrEr)iLrEZsd)iMrEr))iNrEZppv)iOrEZwc)iPr0)ijrEZmc)ikrEZmd)ilr)ipr0)irEZdj)ir0)ir)ir0)irEuほか)irEuココ)irEr|)ir)irEr)irEu字)irEu双)irEuデ)irEr)irE多)irEu解)irErh)irEu交)irEu映)irEu無)irEr)irEu前)irEu後)irE再)irEu新)i rEu初)i!rEu終)i"rEr)i#rEu販)i$rEu声)i%rEu吹)i&rEu演)i'rEu投)i(rEu捕)i)rEr)i*rErc)i+rEu遊)i,rEro)i-rErf)i.rErp)i/rEu指)i0rEr#)i1rEu打)i2rEu禁)i3rEu空)i4rEu合)i5rEu満)i6rErl)i7rEr)i8rEu申)i9rE割)i:rEu営)i;rEu配rrrrr_seg_71srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i<r)i@rEu	〔本〕)iArEu	〔三〕)iBrEu	〔二〕)iCrEu	〔安〕)iDrEu	〔点〕)iErEu	〔打〕)iFrEu	〔盗〕)iGrEu	〔勝〕)iHrEu	〔敗〕)iIr)iPrEu得)iQrEu可)iRr)i`r0)ifr)ir0)ir)ir0)ir)ir0)ir)ir0)itr)ir0)ir)ir0)ir)ir0)iHr)iPr0)iZr)i`r0)ir)ir0)ir)ir0)ir)ir0)i?r)i@r0)iqr)isr0)iwr)izr0)i{r)i|r0)ir)ir0)ir)ir0)ir)ir0)ir)i`r0)inr)ir0)iצr)ir0)i5r)i@r0)ir)i r0)ir)ir0)ir)irEu丽)irEu丸)irEu乁)irEu𠄢)irEu你)irEr)irEu侻)irEu倂)irEu偺)i	rEu備)i
rEr)irEu像)irEu㒞)i
rEu𠘺)irEr)irEu兔)irEu兤)irEu具)irEu𠔜)irEu㒹)irEu內)irEr)irEu𠕋)irEu冗)irEu冤)irEu仌)irEu冬)irEr)irEu𩇟)irEr	)irEu刃)irEu㓟)i rEu刻)i!rEu剆rrrrr_seg_72HsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i"rEr)i#rEu剷)i$rEu㔕)i%rEr)i&rEr)i'rEr)i(rEr)i)rEu包)i*rEu匆)i+rEr)i,rEu卉)i-rEr)i.rEu博)i/rEu即)i0rEu卽)i1rEu卿)i4rEu𠨬)i5rEu灰)i6rEu及)i7rEu叟)i8rEu𠭣)i9rEu叫)i:rEu叱)i;rEu吆)i<rEu咞)i=rEu吸)i>rEu呈)i?rEu周)i@rEu咢)iArEu哶)iBrEu唐)iCrEu啓)iDrEu啣)iErEu善)iGrEr)iHrEu喫)iIrEu喳)iJrEu嗂)iKrEu圖)iLrEr)iMrEu圗)iNrEu噑)iOrEu噴)iPrEr)iQrEu壮)iRrEu城)iSrEu埴)iTrEu堍)iUrEu型)iVrEu堲)iWrEu報)iXrEu墬)iYrEu𡓤)iZrEu売)i[rEu壷)i\rEu夆)i]rEr)i^rEu夢)i_rEu奢)i`rEu𡚨)iarEu𡛪)ibrEu姬)icrEu娛)idrEu娧)ierEu姘)ifrEu婦)igrEu㛮)ihr)iirEu嬈)ijrEu嬾)ilrEu𡧈)imrEu寃)inrEu寘)iorEr)iprEu寳)iqrEu𡬘)irrEu寿)isrEu将)itr)iurEr)ivrEu㞁)iwrEu屠)ixrEr)iyrEu峀)izrEu岍)i{rEu𡷤)i|rEu嵃)i}rEu𡷦)i~rEu嵮)irEu嵫)irEu嵼)irEu巡)irEu巢)irEu㠯)irEu巽)irEu帨)irEu帽)irEu幩)irEu㡢)irEu𢆃rrrrr_seg_73srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu㡼)irEu庰)irEu庳)irEu庶)irEr)irEu𪎒)irEr)irEu𢌱)irE舁)irEu弢)irEu㣇)irEu𣊸)irEu𦇚)irEu形)irEu彫)irEu㣣)irEu徚)irEu忍)irEu志)irEu忹)irEu悁)irEu㤺)irEu㤜)irEr)irEu𢛔)irEu惇)irEu慈)irE慌)irEr)irEr)irEu慺)irEr)irEu憲)irEu憤)irEu憯)irEu懞)irEr)irEr)irEu成)irEu戛)irEu扝)irEu抱)irEu拔)irEu捐)irEu𢬌)irEu挽)irEu拼)irEu捨)irEu掃)irEu揤)irEu𢯱)irEu搢)irEu揅)irEu掩)irEu㨮)irEu摩)irEu摾)irEu撝)irEu摷)irEu㩬)irEr)irEu敬)irEu𣀊)irEu旣)irEu書)irEu晉)irEu㬙)irEr)irEu㬈)irEu㫤)irEu冒)irEu冕)irEu最)irEu暜)irEu肭)irEu䏙)irEr)irEr)irEu朡)irEu杞)irEu杓)irEu𣏃)irEu㭉)irEu柺)irEu枅)irEu桒)irEr)irEu𣑭)irEu梎)irEu栟)irEu椔)irEr)irEu楂)irEu榣)irEu槪)irEu檨)irEu𣚣)irEu櫛)irEu㰘)irEu次rrrrr_seg_74srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu𣢧)irEu歔)irEu㱎)irEu歲)irEu殟)irEr)irEu殻)irEu𣪍)irEu𡴋)irEu𣫺)irEu汎)irEu𣲼)irEu沿)irEu泍)irEu汧)irEu洖)irEu派)irEr)irEr)irEu浩)irEu浸)irEu涅)irEu𣴞)irEu洴)irEu港)i	rEu湮)i
rEu㴳)irEr)irEu滇)i
rEu𣻑)irEu淹)irEu潮)irEu𣽞)irEu𣾎)irEu濆)irEu瀹)irEr)irEu瀛)irEu㶖)irEu灊)irEu災)irEu灷)irEu炭)irEu𠔥)irEu煅)irEu𤉣)irEu熜)ir)i rEu爨)i!rEr)i"rEu牐)i#rEu𤘈)i$rEu犀)i%rEu犕)i&rEu𤜵)i'rEu𤠔)i(rEu獺)i)rEu王)i*rEu㺬)i+rEu玥)i,rEu㺸)i.rEu瑇)i/rEu瑜)i0rEr)i1rEu璅)i2rEu瓊)i3rEu㼛)i4rEu甤)i5rEu𤰶)i6rEu甾)i7rEu𤲒)i8rEr)i9rEu𢆟)i:rEu瘐)i;rEu𤾡)i<rEu𤾸)i=rEu𥁄)i>rEu㿼)i?rEu䀈)i@rEr)iArEu𥃳)iBrEu𥃲)iCrEu𥄙)iDrEu𥄳)iErEu眞)iFrEu真)iHrEr)iIrEr)iJrEu瞋)iKrEu䁆)iLrEu䂖)iMrEu𥐝)iNrEu硎)iOrEr)iPrEr)iQrEu䃣)iRrEu𥘦)iSrEr)iTrEu𥚚)iUrEu𥛅rrrrr_seg_75srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iVrEr)iWrEu秫)iXrEu䄯)iYrEr)iZrEu穊)i[rEu穏)i\rEu𥥼)i]rEu𥪧)i_r)i`rEu䈂)iarEu𥮫)ibrEu篆)icrEu築)idrEu䈧)ierEu𥲀)ifrEu糒)igrEu䊠)ihrEu糨)iirEu糣)ijrEu紀)ikrEu𥾆)ilrEu絣)imrEu䌁)inrEu緇)iorEu縂)iprEu繅)iqrEu䌴)irrEu𦈨)isrEu𦉇)itrEu䍙)iurEu𦋙)ivrEu罺)iwrEu𦌾)ixrEu羕)iyrEu翺)izrEr)i{rEu𦓚)i|rEu𦔣)i}rEu聠)i~rEu𦖨)irEu聰)irEu𣍟)irEu䏕)irEu育)irEu脃)irEu䐋)irEu脾)irEu媵)irEu𦞧)irEu𦞵)irEu𣎓)irEu𣎜)irEr)irEu舄)irEu辞)irEu䑫)irEu芑)irEu芋)irEu芝)irEu劳)irEu花)irEu芳)irEu芽)irEu苦)irEu𦬼)irEr)irEu茝)irEu荣)irEu莭)irEu茣)irEu莽)irEu菧)irEr)irEu荓)irEu菊)irEu菌)irEu菜)irEu𦰶)irEu𦵫)irEu𦳕)irEu䔫)irEu蓱)irEu蓳)irEu蔖)irEu𧏊)irEu蕤)irEu𦼬)irEu䕝)irEu䕡)irEu𦾱)irEu𧃒)irEu䕫)irEu虐)irEr)irEu虧)irEu虩)irEu蚩)irEu蚈)irEu蜎)irEu蛢rrrrr_seg_76srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEu蜨)irEu蝫)irEu螆)ir)irEu蟡)irEu蠁)irEu䗹)irEu衠)irEr)irEu𧙧)irEu裗)irEu裞)irEu䘵)irEu裺)irEu㒻)irEu𧢮)irEu𧥦)irEu䚾)irEu䛇)irEu誠)irEr)irEr)irEr")irEu𧲨)irEu貫)irEu賁)irEu贛)irEu起)irEu𧼯)irEu𠠄)irEu跋)irEu趼)irEu跰)irEu𠣞)irEu軔)irEr)irEu𨗒)irEu𨗭)irEu邔)irEu郱)irEu鄑)irEu𨜮)irEu鄛)irEu鈸)irEu鋗)irEu鋘)irEu鉼)irEu鏹)irEu鐕)irEu𨯺)irEu開)irEu䦕)irEu閷)irEu𨵷)irEu䧦)irEu雃)irEu嶲)irEu霣)irEu𩅅)irEu𩈚)irEu䩮)irEu䩶)irEu韠)irEu𩐊)irEu䪲)irEu𩒖)irEr)irEu頩)irEu𩖶)irEu飢)irEu䬳)irEu餩)irEu馧)irEu駂)irEu駾)irEu䯎)i	rEu𩬰)i
rEr)irEu鱀)irEu鳽)i
rEu䳎)irEu䳭)irEu鵧)irEu𪃎)irEu䳸)irEu𪄅)irEu𪈎)irEu𪊑)irEr))irEu䵖)irEr*)irEu黾)irEu鼅)irEu鼏)irEu鼖)irEr+)irEu𪘀)ir)irrrrrr_seg_77PsrcCsdgS)N)irrrrrr_seg_78srN)S__doc____version__rrrLrTr_rerrrrrrrrrrrrrrrrrrrrrr r7rirjrqrrrrrrrrrrrrr/rarjrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrtupleZ	uts46datarrrrshhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhh	

 !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNsite-packages/pip/_vendor/idna/__pycache__/uts46data.cpython-38.opt-1.pyc000064400000524104151464452710022060 0ustar00U

.e@sbdZdZddZddZddZdd	Zd
dZdd
ZddZddZ	ddZ
ddZddZddZ
ddZddZddZd d!Zd"d#Zd$d%Zd&d'Zd(d)Zd*d+Zd,d-Zd.d/Zd0d1Zd2d3Zd4d5Zd6d7Zd8d9Zd:d;Zdd?Z d@dAZ!dBdCZ"dDdEZ#dFdGZ$dHdIZ%dJdKZ&dLdMZ'dNdOZ(dPdQZ)dRdSZ*dTdUZ+dVdWZ,dXdYZ-dZd[Z.d\d]Z/d^d_Z0d`daZ1dbdcZ2dddeZ3dfdgZ4dhdiZ5djdkZ6dldmZ7dndoZ8dpdqZ9drdsZ:dtduZ;dvdwZd|d}Z?d~dZ@ddZAddZBddZCddZDddZEddZFddZGddZHddZIddZJddZKddZLddZMddZNddZOddZPeQeeeeeeee	e
eee
eeeeeeeeeeeeeeeeeee e!e"e#e$e%e&e'e(e)e*e+e,e-e.e/e0e1e2e3e4e5e6e7e8e9e:e;e<e=e>e?e@eAeBeCeDeEeFeGeHeIeJeKeLeMeNeOePZRdS)zIDNA Mapping Table from UTS46.z11.0.0cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)3)r)r)r)r)r)r)r)r)	r)
r)r)r)
r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r) r)!r)"r)#r)$r)%r)&r)'r)(r))r)*r)+r),r)-V).r0)/r)0r0)1r0)2r0)3r0)4r0)5r0)6r0)7r0)8r0)9r0):r);r)<r)=r)>r)?r)@r)AMa)BrEb)CrEc)DrEd)ErEe)FrEf)GrEg)HrEh)IrEi)JrEj)KrEk)LrEl)MrEm)NrEn)OrEo)PrEp)QrEq)RrEr)SrEs)TrEt)UrEu)VrEv)WrEw)XrEx)YrEy)ZrEz)[r)\r)]r)^r)_r)`r)ar0)br0)cr0rrr>/usr/lib/python3.8/site-packages/pip/_vendor/idna/uts46data.py_seg_0srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)dr0)er0)fr0)gr0)hr0)ir0)jr0)kr0)lr0)mr0)nr0)or0)pr0)qr0)rr0)sr0)tr0)ur0)vr0)wr0)xr0)yr0)zr0){r)|r)}r)~r)r)X)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r)r )r0)r0)r0)r0)r0)r0)r0)ru ̈)r0)rErF)r0)r0)I)r0)r ̄)r0)r0)rE2)rEr)r ́)rEμ)r0)r0)ru ̧)rE1)rErb)r0)rEu1⁄4)rEu1⁄2)rEu3⁄4)r0)rEà)rEá)rEâ)rEã)rEä)rEå)rEæ)rEçrrrrr_seg_1psrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)rEè)rEé)rEê)rEë)rEì)rEí)rEî)rEï)rEð)rEñ)rEò)rEó)rEô)rEõ)rEö)r0)rEø)rEù)rEú)rEû)rEü)rEý)rEþ)Dss)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)r0)rEuā)ir0)irEuă)ir0)irEuą)ir0)irEuć)ir0)irEuĉ)i	r0)i
rEuċ)ir0)irEuč)i
r0)irEuď)ir0)irEuđ)ir0)irEuē)ir0)irEuĕ)ir0)irEuė)ir0)irEuę)ir0)irEuě)ir0)irEuĝ)ir0)irEuğ)ir0)i rEuġ)i!r0)i"rEuģ)i#r0)i$rEuĥ)i%r0)i&rEħ)i'r0)i(rEuĩ)i)r0)i*rEuī)i+r0rrrrr_seg_2srLcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i,rEuĭ)i-r0)i.rEuį)i/r0)i0rEui̇)i1r0)i2rEZij)i4rEuĵ)i5r0)i6rEuķ)i7r0)i9rEuĺ)i:r0)i;rEuļ)i<r0)i=rEuľ)i>r0)i?rEul·)iArEuł)iBr0)iCrEuń)iDr0)iErEuņ)iFr0)iGrEuň)iHr0)iIrEuʼn)iJrEŋ)iKr0)iLrEuō)iMr0)iNrEuŏ)iOr0)iPrEuő)iQr0)iRrEœ)iSr0)iTrEuŕ)iUr0)iVrEuŗ)iWr0)iXrEuř)iYr0)iZrEuś)i[r0)i\rEuŝ)i]r0)i^rEuş)i_r0)i`rEuš)iar0)ibrEuţ)icr0)idrEuť)ier0)ifrEuŧ)igr0)ihrEuũ)iir0)ijrEuū)ikr0)ilrEuŭ)imr0)inrEuů)ior0)iprEuű)iqr0)irrEuų)isr0)itrEuŵ)iur0)ivrEuŷ)iwr0)ixrEÿ)iyrEuź)izr0)i{rEuż)i|r0)i}rEuž)i~r0)irErj)ir0)irEuɓ)irEuƃ)ir0)irEuƅ)ir0)irEɔ)irEuƈ)ir0)irEuɖ)irEuɗ)irEuƌ)ir0)irEǝ)irEə)irEɛ)irEuƒ)ir0)irEuɠrrrrr_seg_3@srTcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEɣ)ir0)irEɩ)irEɨ)irEuƙ)ir0)irEɯ)irEɲ)ir0)irEɵ)irEuơ)ir0)irEuƣ)ir0)irEuƥ)ir0)irEuʀ)irEuƨ)ir0)irEʃ)ir0)irEuƭ)ir0)irEuʈ)irEuư)ir0)irEʊ)irEʋ)irEuƴ)ir0)irEuƶ)ir0)irEʒ)irEuƹ)ir0)irEuƽ)ir0)irEudž)irEZlj)irEZnj)irEuǎ)ir0)irEuǐ)ir0)irEuǒ)ir0)irEuǔ)ir0)irEuǖ)ir0)irEuǘ)ir0)irEuǚ)ir0)irEuǜ)ir0)irEuǟ)ir0)irEuǡ)ir0)irEuǣ)ir0)irEuǥ)ir0)irEuǧ)ir0)irEuǩ)ir0)irEuǫ)ir0)irEuǭ)ir0)irEuǯ)ir0)irEZdz)irEuǵ)ir0)irEuƕ)irEuƿ)irEuǹ)ir0)irEuǻ)ir0)irEuǽ)ir0)irEuǿ)ir0)irEuȁ)ir0)irEuȃ)ir0)irEuȅ)ir0)irEuȇ)ir0)irEuȉ)i	r0)i
rEuȋ)ir0)irEuȍrrrrr_seg_4sr_cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i
r0)irEuȏ)ir0)irEuȑ)ir0)irEuȓ)ir0)irEuȕ)ir0)irEuȗ)ir0)irEuș)ir0)irEuț)ir0)irEuȝ)ir0)irEuȟ)ir0)i rEuƞ)i!r0)i"rEȣ)i#r0)i$rEuȥ)i%r0)i&rEuȧ)i'r0)i(rEuȩ)i)r0)i*rEuȫ)i+r0)i,rEuȭ)i-r0)i.rEuȯ)i/r0)i0rEuȱ)i1r0)i2rEuȳ)i3r0)i:rEuⱥ)i;rEuȼ)i<r0)i=rEuƚ)i>rEuⱦ)i?r0)iArEuɂ)iBr0)iCrEuƀ)iDrEʉ)iErEʌ)iFrEuɇ)iGr0)iHrEuɉ)iIr0)iJrEuɋ)iKr0)iLrEuɍ)iMr0)iNrEuɏ)iOr0)irErT)irEɦ)irErX)irErh)irEuɹ)irEuɻ)irEuʁ)irErr)irErv)ir0)iru ̆)iru ̇)iru ̊)iru ̨)iru ̃)iru ̋)ir0)irErU)irEr\)irErj)irErt)irEuʕ)ir0)i@rEù)iArEú)iBr0)iCrEu̓)iDrEǘ)iErEι)iFr0)iOr)iPr0)iprEuͱ)iqr0)irrEuͳ)isr0)itrEuʹ)iur0)ivrEuͷ)iwr0rrrrr_seg_5srecdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ixr)izru ι)i{r0)i~r;)irEuϳ)ir)irr)ir ̈́)irEά)irE·)irEέ)irEή)irEί)ir)irEό)ir)irEύ)irEώ)ir0)irEα)irEβ)irEγ)irEδ)irEε)irEζ)irEη)irEθ)irErd)irEκ)irEλ)irEr)irEν)irEξ)irEο)irEπ)irEρ)ir)irEσ)irEτ)irEυ)irEφ)irEχ)irEψ)irEω)irEuϊ)irEϋ)ir0)ir(r)ir0)irEuϗ)irErq)irErw)irEr)irErn)irEr)irEr)irEr})ir0)irEuϙ)ir0)irEuϛ)ir0)irEϝ)ir0)irEuϟ)ir0)irEuϡ)ir0)irEuϣ)ir0)irEuϥ)ir0)irEuϧ)ir0)irEuϩ)ir0)irEuϫ)ir0)irEuϭ)ir0)irEuϯ)ir0)irErx)irEr~)irEr)ir0)irErw)irErt)ir0)irEuϸ)ir0)irEr)irEuϻ)ir0)irEuͻ)irEuͼ)irEuͽ)irEuѐ)irEuё)irEuђrrrrr_seg_6xsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuѓ)irEuє)irEuѕ)irEuі)irEuї)irEuј)i	rEuљ)i
rEuњ)irEuћ)irEuќ)i
rEuѝ)irEuў)irEuџ)irEuа)irEuб)irEв)irEuг)irEд)irEuе)irEuж)irEuз)irEuи)irEuй)irEuк)irEuл)irEuм)irEн)irEо)irEuп)i rEuр)i!rEс)i"rEт)i#rEuу)i$rEuф)i%rEuх)i&rEuц)i'rEuч)i(rEuш)i)rEuщ)i*rEъ)i+rEuы)i,rEь)i-rEuэ)i.rEuю)i/rEuя)i0r0)i`rEuѡ)iar0)ibrEѣ)icr0)idrEuѥ)ier0)ifrEuѧ)igr0)ihrEuѩ)iir0)ijrEuѫ)ikr0)ilrEuѭ)imr0)inrEuѯ)ior0)iprEuѱ)iqr0)irrEuѳ)isr0)itrEuѵ)iur0)ivrEuѷ)iwr0)ixrEuѹ)iyr0)izrEuѻ)i{r0)i|rEuѽ)i}r0)i~rEuѿ)ir0)irEuҁ)ir0)irEuҋ)ir0)irEuҍ)ir0)irEuҏ)ir0)irEuґ)ir0)irEuғ)ir0)irEuҕ)ir0)irEuҗ)ir0)irEuҙ)ir0)irEuқ)ir0)irEuҝ)ir0rrrrr_seg_7srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuҟ)ir0)irEuҡ)ir0)irEuң)ir0)irEuҥ)ir0)irEuҧ)ir0)irEuҩ)ir0)irEuҫ)ir0)irEuҭ)ir0)irEuү)ir0)irEuұ)ir0)irEuҳ)ir0)irEuҵ)ir0)irEuҷ)ir0)irEuҹ)ir0)irEuһ)ir0)irEuҽ)ir0)irEuҿ)ir0)ir)irEuӂ)ir0)irEuӄ)ir0)irEuӆ)ir0)irEuӈ)ir0)irEuӊ)ir0)irEuӌ)ir0)irEuӎ)ir0)irEuӑ)ir0)irEuӓ)ir0)irEuӕ)ir0)irEuӗ)ir0)irEuә)ir0)irEuӛ)ir0)irEuӝ)ir0)irEuӟ)ir0)irEuӡ)ir0)irEuӣ)ir0)irEuӥ)ir0)irEuӧ)ir0)irEuө)ir0)irEuӫ)ir0)irEuӭ)ir0)irEuӯ)ir0)irEuӱ)ir0)irEuӳ)ir0)irEuӵ)ir0)irEuӷ)ir0)irEuӹ)ir0)irEuӻ)ir0)irEuӽ)ir0)irEuӿ)ir0)irEuԁ)ir0)irEuԃrrrrr_seg_8HsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir0)irEuԅ)ir0)irEuԇ)ir0)irEuԉ)i	r0)i
rEuԋ)ir0)irEuԍ)i
r0)irEuԏ)ir0)irEuԑ)ir0)irEuԓ)ir0)irEuԕ)ir0)irEuԗ)ir0)irEuԙ)ir0)irEuԛ)ir0)irEuԝ)ir0)irEuԟ)ir0)i rEuԡ)i!r0)i"rEuԣ)i#r0)i$rEuԥ)i%r0)i&rEuԧ)i'r0)i(rEuԩ)i)r0)i*rEuԫ)i+r0)i,rEuԭ)i-r0)i.rEuԯ)i/r0)i0r)i1rEuա)i2rEuբ)i3rEuգ)i4rEuդ)i5rEuե)i6rEuզ)i7rEuէ)i8rEuը)i9rEuթ)i:rEuժ)i;rEuի)i<rEuլ)i=rEuխ)i>rEuծ)i?rEuկ)i@rEuհ)iArEuձ)iBrEuղ)iCrEuճ)iDrEuմ)iErEuյ)iFrEuն)iGrEuշ)iHrEuո)iIrEuչ)iJrEuպ)iKrEuջ)iLrEuռ)iMrEuս)iNrEuվ)iOrEuտ)iPrEuր)iQrEuց)iRrEuւ)iSrEuփ)iTrEuք)iUrEuօ)iVrEuֆ)iWr)iYr0)irEuեւ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0rrrrr_seg_9srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iurEuاٴ)ivrEuوٴ)iwrEۇٴ)ixrEuيٴ)iyr0)ir)ir0)ir)ir0)iKr)iMr0)ir)ir0)ir)ir0)i.r)i0r0)i?r)i@r0)i\r)i^r0)i_r)i`r0)ikr)ir0)ir)ir0)ir)ir0)ir)ir0)iX	rEuक़)iY	rEuख़)iZ	rEuग़)i[	rEuज़)i\	rEuड़)i]	rEuढ़)i^	rEuफ़)i_	rEuय़)i`	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	r0)i	r)i	rEuড়)i	rEuঢ়)i	r)i	rEuয়)i	r0)i	r)i	r0)i	r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i)
r)i*
r0)i1
r)i2
r0)i3
rEuਲ਼)i4
r)i5
r0)i6
rEuਸ਼)i7
r)i8
r0)i:
r)i<
r0)i=
r)i>
r0)iC
r)iG
r0)iI
r)iK
r0)iN
r)iQ
r0)iR
r)iY
rEuਖ਼)iZ
rEuਗ਼)i[
rEuਜ਼rrrrr_seg_10srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i\
r0)i]
r)i^
rEuਫ਼)i_
r)if
r0)iw
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)ir)ir0)ir)ir0)i
r)ir0)ir)ir0)i)r)i*r0)i1r)i2r0)i4r)i5r0)i:r)i<r0)iEr)iGr0)iIr)iKr0)iNr)iVr0)iXr)i\rEuଡ଼)i]rEuଢ଼)i^r)i_r0)idr)ifr0)ixr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)ir)ir0rrrrr_seg_11srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i)r)i*r0)i:r)i=r0)iEr)iFr0)iIr)iJr0)iNr)iUr0)iWr)iXr0)i[r)i`r0)idr)ifr0)ipr)ixr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)i
r0)i
r)i
r0)i

r)i
r0)i
r)i
r0)iE
r)iF
r0)iI
r)iJ
r0)iP
r)iT
r0)id
r)if
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)ir0)i3rEuํา)i4r0)i;r)i?r0)i\r)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0rrrrr_seg_12srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)irEuໍາ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irEuຫນ)irEuຫມ)ir0)ir)ir0)irEu་)i
r0)iCrEuགྷ)iDr0)iHr)iIr0)iMrEuཌྷ)iNr0)iRrEuདྷ)iSr0)iWrEuབྷ)iXr0)i\rEuཛྷ)i]r0)iirEuཀྵ)ijr0)imr)iqr0)isrEuཱི)itr0)iurEuཱུ)ivrEuྲྀ)iwrEu	ྲཱྀ)ixrEuླྀ)iyrEu	ླཱྀ)izr0)irEuཱྀ)ir0)irEuྒྷ)ir0)ir)ir0)irEuྜྷ)ir0)irEuྡྷ)ir0)irEuྦྷ)ir0)irEuྫྷ)ir0)irEuྐྵ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irEuⴧ)ir)irEuⴭ)ir)ir0)irEuნ)ir0)i_r)iar0)iIr)iJr0)iNr)iPr0)iWr)iXr0)iYr)iZr0)i^r)i`r0)ir)ir0rrrrr_seg_13PsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i[r)i]r0)i}r)ir0)ir)ir0)ir)irEuᏰ)irEuᏱ)irEuᏲ)irEuᏳ)irEuᏴ)irEuᏵ)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)ir)i r0)i7r)i@r0)iTr)i`r0)imr)inr0)iqr)irr0)itr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir)ir0)ir)i r0)iyr)ir0)ir)ir0)ir)ir0)ir)i r0)i,r)i0r0)i<r)i@r0)iAr)iDr0)inr)ipr0)iur)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i_r)i`r0)i}r)ir0)ir)ir0rrrrr_seg_14srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)ir0)ir)ir0)ir)ir0)iLr)iPr0)i}r)ir0)ir)ir0)i8r)i;r0)iJr)iMr0)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEꙋ)ir)ir0)ir)ir0)ir)ir0)i,rErF)i-rEr)i.rErH)i/r0)i0rErL)i1rErN)i2rErQ)i3rErR)i4rErT)i5rErV)i6rErX)i7rErZ)i8rEr\)i9rEr^)i:rEr`)i;r0)i<rErb)i=rEr`)i>rErd)i?rErh)i@rErl)iArErn)iBrErr)iCrErF)iDrEɐ)iErEɑ)iFrEuᴂ)iGrErH)iHrErL)iIrErN)iJrErR)iKrErS)iLrEɜ)iMrErR)iNr0)iOrErZ)iPrEr^)iQrErM)iRrErb)iSrErP)iTrEuᴖ)iUrEuᴗ)iVrErd)iWrErl)iXrErn)iYrEuᴝ)iZrErX)i[rErp)i\rEuᴥ)i]rErq)i^rErr)i_rErs)i`rEr)iarEr)ibrErV)icrErh)idrErn)ierErp)ifrErq)igrErr)ihrEr~)iirEr)ijrEr)ikr0)ixrEr)iyr0)irEɒ)irErJ)irEuɕ)irErrrrrr_seg_15 srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irErP)irEuɟ)irEɡ)irEɥ)irErW)irErV)irEɪ)irEuᵻ)irEʝ)irEuɭ)irEuᶅ)irEuʟ)irEɱ)irEuɰ)irErY)irEuɳ)irEuɴ)irErZ)irEuɸ)irEuʂ)irEr[)irEuƫ)irEra)irEr\)irEuᴜ)irEr])irErb)irErx)irEuʐ)irEuʑ)irEr^)irErw)ir0)ir)ir0)irEuḁ)ir0)irEuḃ)ir0)irEuḅ)ir0)irEuḇ)ir0)irEuḉ)i	r0)i
rEuḋ)ir0)irEuḍ)i
r0)irEuḏ)ir0)irEuḑ)ir0)irEuḓ)ir0)irEuḕ)ir0)irEuḗ)ir0)irEuḙ)ir0)irEuḛ)ir0)irEuḝ)ir0)irEuḟ)ir0)i rEuḡ)i!r0)i"rEuḣ)i#r0)i$rEuḥ)i%r0)i&rEuḧ)i'r0)i(rEuḩ)i)r0)i*rEuḫ)i+r0)i,rEuḭ)i-r0)i.rEuḯ)i/r0)i0rEuḱ)i1r0)i2rEuḳ)i3r0)i4rEuḵ)i5r0)i6rEuḷ)i7r0)i8rEuḹ)i9r0)i:rEuḻ)i;r0)i<rEuḽ)i=r0)i>rEuḿ)i?r0rrrrr_seg_16srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i@rEuṁ)iAr0)iBrEuṃ)iCr0)iDrEuṅ)iEr0)iFrEuṇ)iGr0)iHrEuṉ)iIr0)iJrEuṋ)iKr0)iLrEuṍ)iMr0)iNrEuṏ)iOr0)iPrEuṑ)iQr0)iRrEuṓ)iSr0)iTrEuṕ)iUr0)iVrEuṗ)iWr0)iXrEuṙ)iYr0)iZrEuṛ)i[r0)i\rEuṝ)i]r0)i^rEuṟ)i_r0)i`rEṡ)iar0)ibrEuṣ)icr0)idrEuṥ)ier0)ifrEuṧ)igr0)ihrEuṩ)iir0)ijrEuṫ)ikr0)ilrEuṭ)imr0)inrEuṯ)ior0)iprEuṱ)iqr0)irrEuṳ)isr0)itrEuṵ)iur0)ivrEuṷ)iwr0)ixrEuṹ)iyr0)izrEuṻ)i{r0)i|rEuṽ)i}r0)i~rEuṿ)ir0)irEuẁ)ir0)irEuẃ)ir0)irEuẅ)ir0)irEuẇ)ir0)irEuẉ)ir0)irEuẋ)ir0)irEuẍ)ir0)irEuẏ)ir0)irEuẑ)ir0)irEuẓ)ir0)irEuẕ)ir0)irEuaʾ)irEr)ir0)irEr))ir0)irEuạ)ir0)irEuả)ir0)irEuấ)ir0)irEuầ)ir0)irEuẩrrrrr_seg_17srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir0)irEuẫ)ir0)irEuậ)ir0)irEuắ)ir0)irEuằ)ir0)irEuẳ)ir0)irEuẵ)ir0)irEuặ)ir0)irEuẹ)ir0)irEuẻ)ir0)irEuẽ)ir0)irEuế)ir0)irEuề)ir0)irEuể)ir0)irEuễ)ir0)irEuệ)ir0)irEuỉ)ir0)irEuị)ir0)irEuọ)ir0)irEuỏ)ir0)irEuố)ir0)irEuồ)ir0)irEuổ)ir0)irEuỗ)ir0)irEuộ)ir0)irEuớ)ir0)irEuờ)ir0)irEuở)ir0)irEuỡ)ir0)irEuợ)ir0)irEuụ)ir0)irEuủ)ir0)irEuứ)ir0)irEuừ)ir0)irEuử)ir0)irEuữ)ir0)irEuự)ir0)irEuỳ)ir0)irEuỵ)ir0)irEuỷ)ir0)irEuỹ)ir0)irEuỻ)ir0)irEuỽ)ir0)irEuỿ)ir0)irEuἀ)i	rEuἁ)i
rEuἂ)irEuἃ)irEuἄ)i
rEuἅ)irEuἆ)irEuἇ)ir0)ir)irEuἐ)irEuἑ)irEuἒrrrrr_seg_18XsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuἓ)irEuἔ)irEuἕ)ir)i r0)i(rEuἠ)i)rEuἡ)i*rEuἢ)i+rEuἣ)i,rEuἤ)i-rEuἥ)i.rEuἦ)i/rEuἧ)i0r0)i8rEuἰ)i9rEuἱ)i:rEuἲ)i;rEuἳ)i<rEuἴ)i=rEuἵ)i>rEuἶ)i?rEuἷ)i@r0)iFr)iHrEuὀ)iIrEuὁ)iJrEuὂ)iKrEuὃ)iLrEuὄ)iMrEuὅ)iNr)iPr0)iXr)iYrEuὑ)iZr)i[rEuὓ)i\r)i]rEuὕ)i^r)i_rEuὗ)i`r0)ihrEuὠ)iirEuὡ)ijrEuὢ)ikrEuὣ)ilrEuὤ)imrEuὥ)inrEuὦ)iorEuὧ)ipr0)iqrErh)irr0)isrErj)itr0)iurErk)ivr0)iwrErl)ixr0)iyrErm)izr0)i{rErn)i|r0)i}rEro)i~r)irEἀι)irEἁι)irEἂι)irEἃι)irEἄι)irEἅι)irEἆι)irEἇι)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEἠι)irEἡι)irEἢι)irEἣι)irEἤι)irEἥι)irEἦι)irEἧι)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEὠι)irEὡι)irEὢι)irEὣιrrrrr_seg_19srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEὤι)irEὥι)irEὦι)irEὧι)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir0)irEuὰι)irEαι)irEuάι)ir)ir0)irEuᾶι)irEuᾰ)irEuᾱ)irEuὰ)irErh)irEr)ir ̓)irErd)irr)iru ͂)iru ̈͂)irEuὴι)irEηι)irEuήι)ir)ir0)irEuῆι)irEuὲ)irErj)irEuὴ)irErk)irEr)iru ̓̀)iru ̓́)iru ̓͂)ir0)irEuΐ)ir)ir0)irEuῐ)irEuῑ)irEuὶ)irErl)ir)iru ̔̀)iru ̔́)iru ̔͂)ir0)irEuΰ)ir0)irEuῠ)irEuῡ)irEuὺ)irErn)irEuῥ)iru ̈̀)irrg)ir`)ir)irEuὼι)irEωι)irEuώι)ir)ir0)irEuῶι)irEuὸ)irErm)irEuὼ)irEro)irEr)irr)iru ̔)ir)i rr)i r)i r()i r)i r0)i rEu‐)i r0)i ru ̳)i r0)i$ r)i' r0)i( r)i/ rr)i0 r0)i3 rEu′′)i4 rEu	′′′)i5 r0)i6 rEu‵‵)i7 rEu	‵‵‵rrrrr_seg_20(srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i8 r0)i< rz!!)i= r0)i> r ̅)i? r0)iG rz??)iH rz?!)iI rz!?)iJ r0)iW rEu′′′′)iX r0)i_ rr)i` r)ia r)id r)ie r)ip rE0)iq rErV)ir r)it rE4)iu rE5)iv rE6)iw rE7)ix rE8)iy rE9)iz r+)i{ rE−)i| r=)i} r()i~ r))i rEr`)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rEr)i rr)i rEr)i rr)i rr)i rr)i r)i rErF)i rErN)i rErb)i rErt)i rErR)i rErT)i rErZ)i rEr\)i rEr^)i rEr`)i rErd)i rErj)i rErl)i r)i r0)i rEZrs)i r0)i r)i r0)i r)i!rza/c)i!rza/s)i!rErJ)i!rEu°c)i!r0)i!rzc/o)i!rzc/u)i!rErS)i!r0)i	!rEu°f)i
!rErR)i!rErT)i!rErK)i!rErV)i!rEr\)i!r0)i!rEr`)i!rEZno)i!r0)i!rErd)i!rErf)i!rErh)i!r0)i !rEZsm)i!!rEZtel)i"!rEZtm)i#!r0)i$!rErx)i%!r0)i&!rEr)i'!r0)i(!rErx)i)!r0rrrrr_seg_21srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i*!rErZ)i+!rEr)i,!rErH)i-!rErJ)i.!r0)i/!rErN)i1!rErP)i2!r)i3!rEr^)i4!rErb)i5!rEא)i6!rEuב)i7!rEuג)i8!rEד)i9!rErV)i:!r0)i;!rEZfax)id?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ib$rEr)ic$rEr)id$rEr)ie$rEr)if$rEr)ig$rEr)ih$rEr)ii$rEZ10)ij$rEZ11)ik$rEZ12)il$rEZ13)im$rEZ14)in$rEZ15)io$rEZ16)ip$rEZ17)iq$rEZ18)ir$rEZ19)is$rEZ20)it$rz(1))iu$rz(2))iv$rz(3))iw$rz(4))ix$rz(5))iy$rz(6))iz$rz(7))i{$rz(8))i|$rz(9))i}$rz(10))i~$rz(11))i$rz(12))i$rz(13))i$rz(14))i$rz(15))i$rz(16))i$rz(17))i$rz(18))i$rz(19))i$rz(20))i$r)i$r(a))i$r(b))i$r(c))i$r(d))i$r(e))i$r(f))i$r(g))i$r(h))i$r(i))i$r(j))i$r(k))i$r(l))i$r(m))i$r(n))i$r(o))i$r(p))i$r(q))i$r(r))i$r(s))i$r(t))i$r(u))i$r(v))i$r(w))i$r(x))i$r(y))i$r(z))i$rErF)i$rErH)i$rErJ)i$rErL)i$rErN)i$rErP)i$rErR)i$rErT)i$rErV)i$rErX)i$rErZ)i$rEr\)i$rEr^)i$rEr`)i$rErb)i$rErd)i$rErf)i$rErh)i$rErj)i$rErl)i$rErn)i$rErp)i$rErr)i$rErt)i$rErv)i$rErx)i$rErF)i$rErH)i$rErJ)i$rErL)i$rErN)i$rErP)i$rErR)i$rErT)i$rErVrrrrr_seg_23`	srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i$rErX)i$rErZ)i$rEr\)i$rEr^)i$rEr`)i$rErb)i$rErd)i$rErf)i$rErh)i$rErj)i$rErl)i$rErn)i$rErp)i$rErr)i$rErt)i$rErv)i$rErx)i$rEr)i$r0)i*rEu∫∫∫∫)i
*r0)it*rz::=)iu*rz==)iv*rz===)iw*r0)i*rEu⫝̸)i*r0)it+r)iv+r0)i+r)i+r0)i+r)i+r0)i+r)i,rEuⰰ)i,rEuⰱ)i,rEuⰲ)i,rEuⰳ)i,rEuⰴ)i,rEuⰵ)i,rEuⰶ)i,rEuⰷ)i,rEuⰸ)i	,rEuⰹ)i
,rEuⰺ)i,rEuⰻ)i,rEuⰼ)i
,rEuⰽ)i,rEuⰾ)i,rEuⰿ)i,rEuⱀ)i,rEuⱁ)i,rEuⱂ)i,rEuⱃ)i,rEuⱄ)i,rEuⱅ)i,rEuⱆ)i,rEuⱇ)i,rEuⱈ)i,rEuⱉ)i,rEuⱊ)i,rEuⱋ)i,rEuⱌ)i,rEuⱍ)i,rEuⱎ)i,rEuⱏ)i ,rEuⱐ)i!,rEuⱑ)i",rEuⱒ)i#,rEuⱓ)i$,rEuⱔ)i%,rEuⱕ)i&,rEuⱖ)i',rEuⱗ)i(,rEuⱘ)i),rEuⱙ)i*,rEuⱚ)i+,rEuⱛ)i,,rEuⱜ)i-,rEuⱝ)i.,rEuⱞ)i/,r)i0,r0)i_,r)i`,rEuⱡ)ia,r0)ib,rEɫ)ic,rEuᵽ)id,rEuɽ)ie,r0)ig,rEuⱨ)ih,r0)ii,rEuⱪ)ij,r0)ik,rEuⱬ)il,r0)im,rEr)in,rEr)io,rEr)ip,rErrrrrr_seg_24	srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iq,r0)ir,rEuⱳ)is,r0)iu,rEuⱶ)iv,r0)i|,rErX)i},rErp)i~,rEuȿ)i,rEuɀ)i,rEuⲁ)i,r0)i,rEuⲃ)i,r0)i,rEuⲅ)i,r0)i,rEuⲇ)i,r0)i,rEuⲉ)i,r0)i,rEuⲋ)i,r0)i,rEuⲍ)i,r0)i,rEuⲏ)i,r0)i,rEuⲑ)i,r0)i,rEuⲓ)i,r0)i,rEuⲕ)i,r0)i,rEuⲗ)i,r0)i,rEuⲙ)i,r0)i,rEuⲛ)i,r0)i,rEuⲝ)i,r0)i,rEuⲟ)i,r0)i,rEuⲡ)i,r0)i,rEuⲣ)i,r0)i,rEuⲥ)i,r0)i,rEuⲧ)i,r0)i,rEuⲩ)i,r0)i,rEuⲫ)i,r0)i,rEuⲭ)i,r0)i,rEuⲯ)i,r0)i,rEuⲱ)i,r0)i,rEuⲳ)i,r0)i,rEuⲵ)i,r0)i,rEuⲷ)i,r0)i,rEuⲹ)i,r0)i,rEuⲻ)i,r0)i,rEuⲽ)i,r0)i,rEuⲿ)i,r0)i,rEuⳁ)i,r0)i,rEuⳃ)i,r0)i,rEuⳅ)i,r0)i,rEuⳇ)i,r0)i,rEuⳉ)i,r0)i,rEuⳋ)i,r0)i,rEuⳍ)i,r0)i,rEuⳏ)i,r0)i,rEuⳑ)i,r0)i,rEuⳓ)i,r0)i,rEuⳕ)i,r0)i,rEuⳗ)i,r0)i,rEuⳙ)i,r0)i,rEuⳛrrrrr_seg_250
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i,r0)i,rEuⳝ)i,r0)i,rEuⳟ)i,r0)i,rEuⳡ)i,r0)i,rEuⳣ)i,r0)i,rEuⳬ)i,r0)i,rEuⳮ)i,r0)i,rEuⳳ)i,r0)i,r)i,r0)i&-r)i'-r0)i(-r)i--r0)i.-r)i0-r0)ih-r)io-rEuⵡ)ip-r0)iq-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)i-r)i-r0)iO.r)i.r0)i.r)i.r0)i.rEu母)i.r0)i.rEu龟)i.r)i/rE一)i/rEu丨)i/rEu丶)i/rEu丿)i/rE乙)i/rEu亅)i/rE二)i/rEu亠)i/rE人)i	/rEu儿)i
/rEu入)i/rE八)i/rEu冂)i
/rEu冖)i/rEu冫)i/rEu几)i/rE凵)i/rEu刀)i/rE力)i/rEu勹)i/rEu匕)i/rEu匚)i/rEu匸)i/rE十)i/rEu卜)i/rEu卩)i/rEu厂)i/rEu厶)i/rEu又)i/rEu口)i/rEu囗)i/rE土)i /rEu士)i!/rEu夂)i"/rEu夊)i#/rEu夕)i$/rEu大)i%/rE女)i&/rEu子)i'/rEu宀)i(/rEu寸)i)/rEu小)i*/rE尢)i+/rEu尸)i,/rE屮)i-/rEu山rrrrr_seg_26
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i./rEu巛)i//rEu工)i0/rEu己)i1/rEu巾)i2/rEu干)i3/rEu幺)i4/rEu广)i5/rEu廴)i6/rE廾)i7/rEu弋)i8/rEu弓)i9/rEu彐)i:/rEu彡)i;/rEu彳)i/rEu戶)i?/rE手)i@/rEu支)iA/rEu攴)iB/rE文)iC/rEu斗)iD/rEu斤)iE/rEu方)iF/rEu无)iG/rE日)iH/rEu曰)iI/rE月)iJ/rE木)iK/rEu欠)iL/rEu止)iM/rE歹)iN/rEu殳)iO/rEu毋)iP/rEu比)iQ/rEu毛)iR/rEu氏)iS/rEu气)iT/rE水)iU/rE火)iV/rEu爪)iW/rEu父)iX/rEu爻)iY/rEu爿)iZ/rEu片)i[/rEu牙)i\/rEu牛)i]/rEu犬)i^/rEu玄)i_/rEu玉)i`/rEu瓜)ia/rEu瓦)ib/rEu甘)ic/rE生)id/rEu用)ie/rEu田)if/rEu疋)ig/rEu疒)ih/rEu癶)ii/rEu白)ij/rEu皮)ik/rEu皿)il/rEu目)im/rEu矛)in/rEu矢)io/rEu石)ip/rEu示)iq/rEu禸)ir/rEu禾)is/rEu穴)it/rE立)iu/rEu竹)iv/rEu米)iw/rEu糸)ix/rEu缶)iy/rEu网)iz/rEu羊)i{/rE羽)i|/rE老)i}/rEu而)i~/rEu耒)i/rEu耳)i/rEu聿)i/rEu肉)i/rEu臣)i/rEu自)i/rEu至)i/rEu臼)i/rEu舌)i/rEu舛)i/rEu舟)i/rEu艮)i/rEu色)i/rEu艸)i/rEu虍)i/rEu虫)i/rEu血)i/rE行)i/rE衣)i/rEu襾rrrrr_seg_27sr cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i/rE見)i/rEu角)i/rEu言)i/rEu谷)i/rEu豆)i/rE豕)i/rEu豸)i/rEu貝)i/rEu赤)i/rE走)i/rEu足)i/rEu身)i/rE車)i/rEu辛)i/rE辰)i/rEu辵)i/rEu邑)i/rEu酉)i/rEu釆)i/rE里)i/rE金)i/rEu長)i/rEu門)i/rEu阜)i/rEu隶)i/rEu隹)i/rEu雨)i/rEu靑)i/rEu非)i/rEu面)i/rEu革)i/rEu韋)i/rEu韭)i/rEu音)i/rEu頁)i/rEu風)i/rEu飛)i/rEu食)i/rEu首)i/rEu香)i/rEu馬)i/rEu骨)i/rEu高)i/rEu髟)i/rEu鬥)i/rEu鬯)i/rEu鬲)i/rEu鬼)i/rEu魚)i/rEu鳥)i/rEu鹵)i/rE鹿)i/rEu麥)i/rE麻)i/rEu黃)i/rEu黍)i/rEu黑)i/rE黹)i/rEu黽)i/rEu鼎)i/rEu鼓)i/rEu鼠)i/rE鼻)i/rEu齊)i/rEu齒)i/rE龍)i/rE龜)i/rEu龠)i/r)i0rr)i0r0)i0rE.)i0r0)i60rEu〒)i70r0)i80rEr)i90rEu卄)i:0rEu卅)i;0r0)i@0r)iA0r0)i0r)i0r0)i0ru ゙)i0ru ゚)i0r0)i0rEuより)i0r0)i0rEuコト)i1r)i1r0)i01r)i11rEᄀ)i21rEᄁ)i31rEᆪ)i41rEᄂ)i51rEᆬ)i61rEᆭ)i71rEᄃ)i81rEᄄrrrrr_seg_28hsr7cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i91rEᄅ)i:1rEᆰ)i;1rEᆱ)i<1rEᆲ)i=1rEᆳ)i>1rEᆴ)i?1rEᆵ)i@1rEᄚ)iA1rEᄆ)iB1rEᄇ)iC1rEᄈ)iD1rEᄡ)iE1rEᄉ)iF1rEᄊ)iG1rEᄋ)iH1rEᄌ)iI1rEᄍ)iJ1rEᄎ)iK1rEᄏ)iL1rEᄐ)iM1rEᄑ)iN1rEᄒ)iO1rEᅡ)iP1rEᅢ)iQ1rEᅣ)iR1rEᅤ)iS1rEᅥ)iT1rEᅦ)iU1rEᅧ)iV1rEᅨ)iW1rEᅩ)iX1rEᅪ)iY1rEᅫ)iZ1rEᅬ)i[1rEᅭ)i\1rEᅮ)i]1rEᅯ)i^1rEᅰ)i_1rEᅱ)i`1rEᅲ)ia1rEᅳ)ib1rEᅴ)ic1rEᅵ)id1r)ie1rEuᄔ)if1rEuᄕ)ig1rEuᇇ)ih1rEuᇈ)ii1rEuᇌ)ij1rEuᇎ)ik1rEuᇓ)il1rEuᇗ)im1rEuᇙ)in1rEuᄜ)io1rEuᇝ)ip1rEuᇟ)iq1rEuᄝ)ir1rEuᄞ)is1rEuᄠ)it1rEuᄢ)iu1rEuᄣ)iv1rEuᄧ)iw1rEuᄩ)ix1rEuᄫ)iy1rEuᄬ)iz1rEuᄭ)i{1rEuᄮ)i|1rEuᄯ)i}1rEuᄲ)i~1rEuᄶ)i1rEuᅀ)i1rEuᅇ)i1rEuᅌ)i1rEuᇱ)i1rEuᇲ)i1rEuᅗ)i1rEuᅘ)i1rEuᅙ)i1rEuᆄ)i1rEuᆅ)i1rEuᆈ)i1rEuᆑ)i1rEuᆒ)i1rEuᆔ)i1rEuᆞ)i1rEuᆡ)i1r)i1r0)i1rEr)i1rEr)i1rE三)i1rE四)i1rE上)i1rE中)i1rE下)i1rEu甲)i1rEr)i1rEu丙)i1rEu丁)i1rE天rrrrr_seg_29sricdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i1rEu地)i1rEr)i1r0)i1r)i1r0)i1r)i1r0)i2ru(ᄀ))i2ru(ᄂ))i2ru(ᄃ))i2ru(ᄅ))i2ru(ᄆ))i2ru(ᄇ))i2ru(ᄉ))i2ru(ᄋ))i2ru(ᄌ))i	2ru(ᄎ))i
2ru(ᄏ))i2ru(ᄐ))i2ru(ᄑ))i
2ru(ᄒ))i2ru(가))i2ru(나))i2ru(다))i2ru(라))i2ru(마))i2ru(바))i2ru(사))i2ru(아))i2ru(자))i2ru(차))i2ru(카))i2ru(타))i2ru(파))i2ru(하))i2ru(주))i2ru(오전))i2ru(오후))i2r)i 2ru(一))i!2ru(二))i"2ru(三))i#2ru(四))i$2ru(五))i%2ru(六))i&2ru(七))i'2ru(八))i(2ru(九))i)2ru(十))i*2ru(月))i+2ru(火))i,2ru(水))i-2ru(木))i.2ru(金))i/2ru(土))i02ru(日))i12ru(株))i22ru(有))i32ru(社))i42ru(名))i52ru(特))i62ru(財))i72ru(祝))i82ru(労))i92ru(代))i:2ru(呼))i;2ru(学))i<2ru(監))i=2ru(企))i>2ru(資))i?2ru(協))i@2ru(祭))iA2ru(休))iB2ru(自))iC2ru(至))iD2rEu問)iE2rEu幼)iF2rEr)iG2rEu箏)iH2r0)iP2rEZpte)iQ2rEZ21)iR2rEZ22)iS2rEZ23)iT2rEZ24)iU2rEZ25)iV2rEZ26)iW2rEZ27)iX2rEZ28)iY2rEZ29)iZ2rEZ30)i[2rEZ31)i\2rEZ32)i]2rEZ33)i^2rEZ34)i_2rEZ35)i`2rEr/)ia2rEr2)ib2rEr5)ic2rEr8rrrrr_seg_308srjcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)id2rEr@)ie2rErA)if2rErD)ig2rErF)ih2rErG)ii2rErI)ij2rErJ)ik2rErK)il2rErL)im2rErM)in2rEu가)io2rEu나)ip2rEu다)iq2rEu라)ir2rEu마)is2rEu바)it2rEu사)iu2rEu아)iv2rEu자)iw2rEu차)ix2rEu카)iy2rEu타)iz2rEu파)i{2rEu하)i|2rEu참고)i}2rEu주의)i~2rEu우)i2r0)i2rEr)i2rEr)i2rErc)i2rErd)i2rEu五)i2rE六)i2rEu七)i2rEr)i2rEu九)i2rEr)i2rEr)i2rEr)i2rEr)i2rEr)i2rEr')i2rEr)i2rEr)i2rEu株)i2rE有)i2rE社)i2rEu名)i2rEu特)i2rEu財)i2rE祝)i2rEu労)i2rEu秘)i2rEu男)i2rEr
)i2rEu適)i2rEu優)i2rEu印)i2rEu注)i2rEu項)i2rEu休)i2rEu写)i2rEu正)i2rEre)i2rErf)i2rErg)i2rE左)i2rE右)i2rEu医)i2rEu宗)i2rEu学)i2rEu監)i2rEu企)i2rEu資)i2rEu協)i2rEu夜)i2rEZ36)i2rEZ37)i2rEZ38)i2rEZ39)i2rEZ40)i2rEZ41)i2rEZ42)i2rEZ43)i2rEZ44)i2rEZ45)i2rEZ46)i2rEZ47)i2rEZ48)i2rEZ49)i2rEZ50)i2rEu1月)i2rEu2月)i2rEu3月)i2rEu4月)i2rEu5月)i2rEu6月)i2rEu7月)i2rEu8月rrrrr_seg_31srqcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i2rEu9月)i2rEu10月)i2rEu11月)i2rEu12月)i2rEZhg)i2rEZerg)i2rEZev)i2rEZltd)i2rEア)i2rEイ)i2rEウ)i2rEエ)i2rEオ)i2rEカ)i2rEキ)i2rEク)i2rEケ)i2rEコ)i2rEサ)i2rEシ)i2rEス)i2rEセ)i2rEソ)i2rEタ)i2rEチ)i2rEツ)i2rEテ)i2rEト)i2rEナ)i2rEニ)i2rEヌ)i2rEネ)i2rEノ)i2rEハ)i2rEヒ)i2rEフ)i2rEヘ)i2rEホ)i2rEマ)i2rEミ)i2rEム)i2rEメ)i2rEモ)i2rEヤ)i2rEユ)i2rEヨ)i2rEラ)i2rEリ)i2rEル)i2rEレ)i2rEロ)i2rEワ)i2rEuヰ)i2rEuヱ)i2rEヲ)i2r)i3rEuアパート)i3rEuアルファ)i3rEuアンペア)i3rEu	アール)i3rEuイニング)i3rEu	インチ)i3rEu	ウォン)i3rEuエスクード)i3rEuエーカー)i	3rEu	オンス)i
3rEu	オーム)i3rEu	カイリ)i3rEuカラット)i
3rEuカロリー)i3rEu	ガロン)i3rEu	ガンマ)i3rEuギガ)i3rEu	ギニー)i3rEuキュリー)i3rEuギルダー)i3rEuキロ)i3rEuキログラム)i3rEuキロメートル)i3rEuキロワット)i3rEu	グラム)i3rEuグラムトン)i3rEuクルゼイロ)i3rEuクローネ)i3rEu	ケース)i3rEu	コルナ)i3rEu	コーポ)i3rEuサイクル)i 3rEuサンチーム)i!3rEuシリング)i"3rEu	センチ)i#3rEu	セント)i$3rEu	ダース)i%3rEuデシ)i&3rEuドル)i'3rEuトン)i(3rEuナノ)i)3rEu	ノット)i*3rEu	ハイツ)i+3rEuパーセントrrrrr_seg_32
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i,3rEu	パーツ)i-3rEuバーレル)i.3rEuピアストル)i/3rEu	ピクル)i03rEuピコ)i13rEuビル)i23rEuファラッド)i33rEuフィート)i43rEuブッシェル)i53rEu	フラン)i63rEuヘクタール)i73rEuペソ)i83rEu	ペニヒ)i93rEu	ヘルツ)i:3rEu	ペンス)i;3rEu	ページ)i<3rEu	ベータ)i=3rEuポイント)i>3rEu	ボルト)i?3rEuホン)i@3rEu	ポンド)iA3rEu	ホール)iB3rEu	ホーン)iC3rEuマイクロ)iD3rEu	マイル)iE3rEu	マッハ)iF3rEu	マルク)iG3rEuマンション)iH3rEuミクロン)iI3rEuミリ)iJ3rEuミリバール)iK3rEuメガ)iL3rEuメガトン)iM3rEuメートル)iN3rEu	ヤード)iO3rEu	ヤール)iP3rEu	ユアン)iQ3rEuリットル)iR3rEuリラ)iS3rEu	ルピー)iT3rEuルーブル)iU3rEuレム)iV3rEuレントゲン)iW3rEu	ワット)iX3rEu0点)iY3rEu1点)iZ3rEu2点)i[3rEu3点)i\3rEu4点)i]3rEu5点)i^3rEu6点)i_3rEu7点)i`3rEu8点)ia3rEu9点)ib3rEu10点)ic3rEu11点)id3rEu12点)ie3rEu13点)if3rEu14点)ig3rEu15点)ih3rEu16点)ii3rEu17点)ij3rEu18点)ik3rEu19点)il3rEu20点)im3rEu21点)in3rEu22点)io3rEu23点)ip3rEu24点)iq3rEZhpa)ir3rEZda)is3rEZau)it3rEZbar)iu3rEZov)iv3rEZpc)iw3rEZdm)ix3rEZdm2)iy3rEZdm3)iz3rEZiu)i{3rEu平成)i|3rEu昭和)i}3rEu大正)i~3rEu明治)i3rEu株式会社)i3rEpa)i3rEZna)i3rEuμa)i3rEZma)i3rEZka)i3rEZkb)i3rEmb)i3rEZgb)i3rEZcal)i3rEZkcal)i3rEZpf)i3rEZnf)i3rEuμf)i3rEuμg)i3rEZmg)i3rEZkgrrrrr_seg_33p
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i3rEhz)i3rEZkhz)i3rEZmhz)i3rEZghz)i3rEZthz)i3rEuμl)i3rEZml)i3rEZdl)i3rEZkl)i3rEZfm)i3rEZnm)i3rEuμm)i3rEZmm)i3rEcm)i3rEkm)i3rEZmm2)i3rEZcm2)i3rEZm2)i3rEZkm2)i3rEZmm3)i3rEZcm3)i3rEZm3)i3rEZkm3)i3rEum∕s)i3rEum∕s2)i3rEr)i3rEZkpa)i3rEZmpa)i3rEZgpa)i3rEZrad)i3rEurad∕s)i3rEurad∕s2)i3rEZps)i3rEns)i3rEuμs)i3rEZms)i3rEZpv)i3rEZnv)i3rEuμv)i3rEmv)i3rEZkv)i3rEr)i3rEZpw)i3rEZnw)i3rEuμw)i3rEmw)i3rEkw)i3rEr)i3rEukω)i3rEumω)i3r)i3rEZbq)i3rEZcc)i3rEcd)i3rEuc∕kg)i3r)i3rEZdb)i3rEZgy)i3rEZha)i3rEZhp)i3rEin)i3rEZkk)i3rEr)i3rEZkt)i3rEZlm)i3rEZln)i3rElog)i3rEZlx)i3rEr)i3rEZmil)i3rEZmol)i3rEZph)i3r)i3rEZppm)i3rEZpr)i3rEsr)i3rEZsv)i3rEwb)i3rEuv∕m)i3rEua∕m)i3rEu1日)i3rEu2日)i3rEu3日)i3rEu4日)i3rEu5日)i3rEu6日)i3rEu7日)i3rEu8日)i3rEu9日)i3rEu10日)i3rEu11日)i3rEu12日)i3rEu13日)i3rEu14日)i3rEu15日)i3rEu16日)i3rEu17日)i3rEu18日)i3rEu19日)i3rEu20日rrrrr_seg_34
srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i3rEu21日)i3rEu22日)i3rEu23日)i3rEu24日)i3rEu25日)i3rEu26日)i3rEu27日)i3rEu28日)i3rEu29日)i3rEu30日)i3rEu31日)i3rEZgal)i4r0)iMr)iMr0)ir)ir0)ir)ir0)iǤr)iФr0)i,r)i@rEuꙁ)iAr0)iBrEuꙃ)iCr0)iDrEuꙅ)iEr0)iFrEuꙇ)iGr0)iHrEuꙉ)iIr0)iJrEr)iKr0)iLrEuꙍ)iMr0)iNrEuꙏ)iOr0)iPrEuꙑ)iQr0)iRrEuꙓ)iSr0)iTrEuꙕ)iUr0)iVrEuꙗ)iWr0)iXrEuꙙ)iYr0)iZrEuꙛ)i[r0)i\rEuꙝ)i]r0)i^rEuꙟ)i_r0)i`rEuꙡ)iar0)ibrEuꙣ)icr0)idrEuꙥ)ier0)ifrEuꙧ)igr0)ihrEuꙩ)iir0)ijrEuꙫ)ikr0)ilrEuꙭ)imr0)irEuꚁ)ir0)irEuꚃ)ir0)irEuꚅ)ir0)irEuꚇ)ir0)irEuꚉ)ir0)irEuꚋ)ir0)irEuꚍ)ir0)irEuꚏ)ir0)irEuꚑ)ir0)irEuꚓ)ir0)irEuꚕ)ir0)irEuꚗ)ir0)irEuꚙ)ir0)irEuꚛ)ir0)irEr)irEr)ir0)irrrrrr_seg_35@srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir0)i"rEuꜣ)i#r0)i$rEuꜥ)i%r0)i&rEꜧ)i'r0)i(rEuꜩ)i)r0)i*rEuꜫ)i+r0)i,rEuꜭ)i-r0)i.rEuꜯ)i/r0)i2rEuꜳ)i3r0)i4rEuꜵ)i5r0)i6rEuꜷ)i7r0)i8rEuꜹ)i9r0)i:rEuꜻ)i;r0)i<rEuꜽ)i=r0)i>rEuꜿ)i?r0)i@rEuꝁ)iAr0)iBrEuꝃ)iCr0)iDrEuꝅ)iEr0)iFrEuꝇ)iGr0)iHrEuꝉ)iIr0)iJrEuꝋ)iKr0)iLrEuꝍ)iMr0)iNrEuꝏ)iOr0)iPrEuꝑ)iQr0)iRrEuꝓ)iSr0)iTrEuꝕ)iUr0)iVrEuꝗ)iWr0)iXrEuꝙ)iYr0)iZrEuꝛ)i[r0)i\rEuꝝ)i]r0)i^rEuꝟ)i_r0)i`rEuꝡ)iar0)ibrEuꝣ)icr0)idrEuꝥ)ier0)ifrEuꝧ)igr0)ihrEuꝩ)iir0)ijrEuꝫ)ikr0)ilrEuꝭ)imr0)inrEꝯ)ior0)iprEr)iqr0)iyrEuꝺ)izr0)i{rEuꝼ)i|r0)i}rEuᵹ)i~rEuꝿ)ir0)irEuꞁ)ir0)irEuꞃ)ir0)irEuꞅ)ir0)irEuꞇ)ir0)irEuꞌ)ir0)irEr)ir0)irEuꞑ)ir0rrrrr_seg_36srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuꞓ)ir0)irEuꞗ)ir0)irEuꞙ)ir0)irEuꞛ)ir0)irEuꞝ)ir0)irEuꞟ)ir0)irEuꞡ)ir0)irEuꞣ)ir0)irEuꞥ)ir0)irEuꞧ)ir0)irEuꞩ)ir0)irErc)irEr)irEr)irEuɬ)irEr)ir0)irEuʞ)irEuʇ)irEr)irEuꭓ)irEuꞵ)ir0)irEuꞷ)ir0)ir)ir0)ir)ir0)irErK)irErN)ir0)i,r)i0r0)i:r)i@r0)ixr)ir0)iƨr)iΨr0)iڨr)ir0)iTr)i_r0)i}r)ir0)iΩr)iϩr0)iکr)iީr0)ir)ir0)i7r)i@r0)iNr)iPr0)iZr)i\r0)iêr)i۪r0)ir)ir0)ir)i	r0)ir)ir0)ir)i r0)i'r)i(r0)i/r)i0r0)i\rEr)i]rEuꬷ)i^rEr)i_rEuꭒ)i`r0)ifr)iprEuᎠ)iqrEuᎡ)irrEuᎢ)isrEuᎣ)itrEuᎤ)iurEuᎥ)ivrEuᎦ)iwrEuᎧ)ixrEuᎨ)iyrEuᎩ)izrEuᎪrrrrr_seg_37srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i{rEuᎫ)i|rEuᎬ)i}rEuᎭ)i~rEuᎮ)irEuᎯ)irEuᎰ)irEuᎱ)irEuᎲ)irEuᎳ)irEuᎴ)irEuᎵ)irEuᎶ)irEuᎷ)irEuᎸ)irEuᎹ)irEuᎺ)irEuᎻ)irEuᎼ)irEuᎽ)irEuᎾ)irEuᎿ)irEuᏀ)irEuᏁ)irEuᏂ)irEuᏃ)irEuᏄ)irEuᏅ)irEuᏆ)irEuᏇ)irEuᏈ)irEuᏉ)irEuᏊ)irEuᏋ)irEuᏌ)irEuᏍ)irEuᏎ)irEuᏏ)irEuᏐ)irEuᏑ)irEuᏒ)irEuᏓ)irEuᏔ)irEuᏕ)irEuᏖ)irEuᏗ)irEuᏘ)irEuᏙ)irEuᏚ)irEuᏛ)irEuᏜ)irEuᏝ)irEuᏞ)irEuᏟ)irEuᏠ)irEuᏡ)irEuᏢ)irEuᏣ)irEuᏤ)irEuᏥ)irEuᏦ)irEuᏧ)irEuᏨ)irEuᏩ)irEuᏪ)irEuᏫ)irEuᏬ)irEuᏭ)irEuᏮ)irEuᏯ)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irEu豈)irEu更)irEr$)irEu賈)irEu滑)irEu串)irEu句)irEr-)i	rEu契)i
rEr')irEu喇)irEu奈)i
rE懶)irEu癩)irEu羅)irEu蘿)irEu螺)irEu裸)irEu邏)irE樂)irEu洛rrrrr_seg_38xsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu烙)irEu珞)irEu落)irEu酪)irEu駱)irEu亂)irEu卵)irEu欄)irEu爛)irEu蘭)i rEu鸞)i!rEu嵐)i"rEu濫)i#rEu藍)i$rEu襤)i%rEu拉)i&rEu臘)i'rEu蠟)i(rE廊)i)rE朗)i*rEu浪)i+rEu狼)i,rEu郎)i-rEu來)i.rEu冷)i/rEu勞)i0rEu擄)i1rEu櫓)i2rEu爐)i3rEu盧)i4rEr)i5rEu蘆)i6rE虜)i7rEu路)i8rEu露)i9rEu魯)i:rEu鷺)i;rE碌)i<rEu祿)i=rEu綠)i>rEu菉)i?rEu錄)i@rEr()iArEu論)iBrEu壟)iCrEu弄)iDrEu籠)iErEu聾)iFrEu牢)iGrEu磊)iHrEu賂)iIrEu雷)iJrEu壘)iKrEu屢)iLrEu樓)iMrEu淚)iNrEu漏)iOrEu累)iPrEu縷)iQrEu陋)iRrEu勒)iSrEu肋)iTrEu凜)iUrEu凌)iVrEu稜)iWrEu綾)iXrEu菱)iYrEu陵)iZrEu讀)i[rEu拏)i\rEr)i]rE諾)i^rEu丹)i_rE寧)i`rEu怒)iarE率)ibrE異)icrE北)idrEu磻)ierEu便)ifrEu復)igrEu不)ihrEu泌)iirEu數)ijrEu索)ikrEu參)ilrEu塞)imrEu省)inrEu葉)iorE說)iprE殺)iqrEr%)irrEu沈)isrEu拾)itrE若)iurEu掠)ivrEu略)iwrEu亮)ixrEu兩)iyrEu凉rrrrr_seg_39srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)izrEu梁)i{rEu糧)i|rEu良)i}rEu諒)i~rEu量)irEu勵)irEu呂)irEr
)irEu廬)irEu旅)irEu濾)irEu礪)irEu閭)irEu驪)irEu麗)irEu黎)irEr
)irEu曆)irEu歷)irEu轢)irEu年)irEu憐)irEu戀)irEu撚)irEu漣)irEu煉)irEu璉)irEu秊)irE練)irEu聯)irEu輦)irEu蓮)irEu連)irEu鍊)irEu列)irEu劣)irEu咽)irEu烈)irEu裂)irEr)irEu廉)irEu念)irEu捻)irEu殮)irEu簾)irEu獵)irEu令)irEu囹)irEr)irEu嶺)irEu怜)irEu玲)irEu瑩)irEu羚)irEu聆)irEu鈴)irEu零)irEu靈)irEu領)irEu例)irEu禮)irEu醴)irEu隸)irEu惡)irEu了)irEu僚)irEu寮)irEu尿)irE料)irEr)irEu燎)irEu療)irEu蓼)irEu遼)irEr,)irEu暈)irEu阮)irEu劉)irEu杻)irEu柳)irE流)irEu溜)irEu琉)irEu留)irEu硫)irEu紐)irEu類)irErk)irEu戮)irEu陸)irEu倫)irEu崙)irEu淪)irEu輪)irEu律)irEu慄)irEu栗)irEr)irEu隆)irEu利rrrrr_seg_40HsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu吏)irEu履)irEu易)irEu李)irEu梨)irEu泥)irEu理)irEu痢)irEu罹)irEu裏)irEu裡)irEr&)irEu離)irEu匿)irEu溺)irEu吝)irEu燐)irEu璘)irEu藺)irEu隣)irEu鱗)irEu麟)irEu林)irEu淋)irEu臨)irEr)irEu笠)irEu粒)irEu狀)irEu炙)irEu識)irEu什)irEu茶)irEu刺)irE切)irEu度)irEu拓)irEu糖)irEu宅)irEu洞)irEu暴)irEu輻)irEr)i	rEu降)i
rEr!)irEu廓)irEu兀)i
rEu嗀)ir0)irE塚)ir0)irE晴)ir0)irEu凞)irE猪)irE益)irEu礼)irEu神)irEu祥)irE福)irE靖)irEu精)irEr)ir0)i rEu蘒)i!r0)i"rE諸)i#r0)i%rE逸)i&rEu都)i'r0)i*rEu飯)i+rEu飼)i,rEu館)i-rEu鶴)i.rEu郞)i/rEu隷)i0rE侮)i1rE僧)i2rE免)i3rE勉)i4rE勤)i5rE卑)i6rE喝)i7rE嘆)i8rEu器)i9rEu塀)i:rEu墨)i;rEu層)i<rEr)i=rE悔)i>rEu慨)i?rE憎)i@rE懲)iArE敏)iBrEu既)iCrE暑)iDrE梅)iErE海)iFrEu渚rrrrr_seg_41srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iGrE漢)iHrE煮)iIrEu爫)iJrEu琢)iKrEu碑)iLrErm)iMrEu祉)iNrEu祈)iOrEu祐)iPrE祖)iQrErn)iRrEu禍)iSrEu禎)iTrE穀)iUrEu突)iVrE節)iWrEr)iXrEu縉)iYrEu繁)iZrEu署)i[rE者)i\rEu臭)i]rEu艹)i_rE著)i`rEu褐)iarE視)ibrE謁)icrE謹)idrEu賓)ierE贈)ifrEu辶)igrEr)ihrE難)iirE響)ijrE頻)ikrEu恵)ilrEu𤋮)imrEu舘)inr)iprEu並)iqrE况)irrEu全)isrEu侀)itrEu充)iurEu冀)ivrE勇)iwrE勺)ixrEr)iyrEu啕)izrE喙)i{rEu嗢)i|rEr)i}rEu墳)i~rEu奄)irEu奔)irEu婢)irEu嬨)irEu廒)irEu廙)irEu彩)irEu徭)irEu惘)irE慎)irEu愈)irEr)irEu慠)irEr)irEu戴)irEu揄)irEu搜)irEu摒)irEu敖)irEr)irEr)irE望)irEu杖)irEr)irEr)irEr)irEu滛)irE滋)irEr)irE瀞)irEr)irEu瞧)irE爵)irEu犯)irEr)irE瑱)irEu甆)irEu画)irEu瘝)irEu瘟)irEr)irEu盛)irE直)irE睊)irEu着)irE磌)irEu窱rrrrr_seg_42srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEu类)irEu絛)irEr)irEu缾)irEr)irEu荒)irEu華)irE蝹)irEu襁)irEu覆)irEr)irEu調)irEr)irEu請)irEr)irEr)irE諭)irEr)irE變)irEr)irE輸)irEu遲)irEu醙)irEu鉶)irEu陼)irEr)irEr)irEu韛)irEr)irE頋)irEr)irE鬒)irEr-)irEu𢡊)irEu𢡄)irEu𣏕)irE㮝)irEu䀘)irE䀹)irEu𥉉)irEu𥳐)irEu𧻓)irEu齃)irEu龎)ir)irEZff)irEZfi)irEZfl)irEZffi)irEZffl)irEst)ir)irEuմն)irEuմե)irEuմի)irEuվն)irEuմխ)ir)irEuיִ)ir0)irEuײַ)i rEuע)i!rEr)i"rEr)i#rEuה)i$rEuכ)i%rEuל)i&rEuם)i'rEuר)i(rEuת)i)rr)i*rEuשׁ)i+rEuשׂ)i,rEuשּׁ)i-rEuשּׂ)i.rEuאַ)i/rEuאָ)i0rEuאּ)i1rEuבּ)i2rEuגּ)i3rEuדּ)i4rEuהּ)i5rEuוּ)i6rEuזּ)i7r)i8rEuטּ)i9rEuיּ)i:rEuךּ)i;rEuכּ)i<rEuלּ)i=r)i>rEuמּ)i?r)i@rEuנּ)iArEuסּ)iBr)iCrEuףּ)iDrEuפּ)iErrrrrr_seg_43srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iFrEuצּ)iGrEuקּ)iHrEuרּ)iIrEuשּ)iJrEuתּ)iKrEuוֹ)iLrEuבֿ)iMrEuכֿ)iNrEuפֿ)iOrEuאל)iPrEuٱ)iRrEuٻ)iVrEuپ)iZrEuڀ)i^rEuٺ)ibrEuٿ)ifrEuٹ)ijrEuڤ)inrEuڦ)irrEuڄ)ivrEuڃ)izrEuچ)i~rEuڇ)irEuڍ)irEuڌ)irEuڎ)irEuڈ)irEuژ)irEuڑ)irEuک)irEuگ)irEuڳ)irEuڱ)irEں)irEuڻ)irEuۀ)irEuہ)irEuھ)irEuے)irEuۓ)ir0)ir)irEuڭ)irEuۇ)irEuۆ)irEuۈ)irEr)irEuۋ)irEuۅ)irEuۉ)irEuې)irEى)irEuئا)irEuئە)irEuئو)irEuئۇ)irEuئۆ)irEuئۈ)irEuئې)irEئى)irEuی)irEئج)irEئح)irEئم)irEr
)irEئي)irEبج)irEبح)irEبخ)irEبم)i	rEبى)i
rEبي)irEتج)irEتح)i
rEتخ)irEتم)irEتى)irEتي)irEuثج)irEثم)irEثى)irEثي)irEجح)irEجم)irEحج)irEحم)irEخج)irEuخح)irEخم)irEسج)irEسح)irEسخ)irEسم)i rEصح)i!rEصم)i"rEضج)i#rEضح)i$rEضخ)i%rEضم)i&rEطحrrrrr_seg_44sr/cdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i'rEطم)i(rEظم)i)rEعج)i*rEعم)i+rEغج)i,rEغم)i-rEفج)i.rEفح)i/rEفخ)i0rEفم)i1rEفى)i2rEفي)i3rEقح)i4rEقم)i5rEقى)i6rEقي)i7rEكا)i8rEكج)i9rEكح)i:rEكخ)i;rEكل)i<rEكم)i=rEكى)i>rEكي)i?rEلج)i@rEلح)iArEلخ)iBrEلم)iCrEلى)iDrEلي)iErEمج)iFrEمح)iGrEمخ)iHrEمم)iIrEuمى)iJrEuمي)iKrEنج)iLrEنح)iMrEنخ)iNrEنم)iOrEنى)iPrEني)iQrEهج)iRrEهم)iSrEuهى)iTrEuهي)iUrEيج)iVrEيح)iWrEيخ)iXrEيم)iYrEيى)iZrEيي)i[rEuذٰ)i\rEuرٰ)i]rEىٰ)i^ru ٌّ)i_ru ٍّ)i`ru َّ)iaru ُّ)ibru ِّ)icru ّٰ)idrEuئر)ierEuئز)ifrEr
)igrEuئن)ihrEr
)iirEr)ijrEuبر)ikrEuبز)ilrEr)imrEuبن)inrEr)iorEr)iprEuتر)iqrEuتز)irrEr)isrEuتن)itrEr)iurEr)ivrEuثر)iwrEuثز)ixrEr)iyrEuثن)izrEr)i{rEr)i|rEr:)i}rEr;)i~rEr>)irEr?)irEr@)irErD)irErE)irErF)irErG)irErK)irErL)irErM)irEuما)irErQ)irEuنرrrrrr_seg_45PsracdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEuنز)irErU)irEuنن)irErV)irErW)irEr`)irEuير)irEuيز)irEr])irEuين)irEr^)irEr_)irEr)irEr)irEuئخ)irEr
)irEئه)irEr)irEr)irEr)irEr)irEبه)irEr)irEr)irEr)irEr)irEته)irEr)irEr)irEr)irEr )irEr!)irEr")irEr#)irEr$)irEr%)irEr&)irEr')irEr()irEuصخ)irEr))irEr*)irEr+)irEr,)irEr-)irEr.)irEr1)irEr2)irEr3)irEr4)irEr5)irEr6)irEr7)irEr8)irEr9)irEr<)irEr=)irErA)irErB)irErC)irErD)irErE)irErH)irErI)irErJ)irErK)irEuله)irErN)irErO)irErP)irErQ)irErR)irErS)irErT)irErU)irEنه)irErX)irErY)irEuهٰ)irErZ)irEr[)irEr\)irEr])irEيه)irEr
)irErb)irEr)irErc)irEr)irErd)irEr)irEuثه)irEr')irEسه)irEشم)irEشه)irErD)irErE)irErK)irErUrrrrr_seg_46srjcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEre)irEr])irErf)irEuـَّ)irEuـُّ)irEuـِّ)irEطى)irEطي)irEعى)irEعي)irEغى)irEغي)irEسى)irEسي)irEشى)irEشي)irEحى)irEحي)irEجى)irEجي)irEخى)irEخي)irEصى)irEصي)irEضى)irEضي)i	rEشج)i
rEشح)irEشخ)irErh)i
rEشر)irEسر)irEصر)irEضر)irErk)irErl)irErm)irErn)irEro)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErx)irEry)i rErz)i!rEr{)i"rEr|)i#rEr})i$rEr~)i%rEr)i&rEr)i'rEr)i(rErh)i)rEr)i*rEr)i+rEr)i,rEr)i-rEr)i.rEr)i/rEr)i0rErh)i1rErg)i2rEri)i3rEr0)i4rEr$)i5rEr%)i6rEr&)i7rEr)i8rEr)i9rEr)i:rEr0)i;rEr1)i<rEuاً)i>r0)i@r)iPrEuتجم)iQrEuتحج)iSrEuتحم)iTrEuتخم)iUrEuتمج)iVrEuتمح)iWrEuتمخ)iXrEuجمح)iZrEuحمي)i[rEuحمى)i\rEuسحج)i]rEuسجح)i^rEuسجى)i_rEuسمح)iarEuسمج)ibrEuسمم)idrEuصحح)ifrEصمم)igrEuشحم)iirEuشجيrrrrr_seg_47 srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ijrEuشمخ)ilrEuشمم)inrEuضحى)iorEuضخم)iqrEuطمح)isrEuطمم)itrEuطمي)iurEعجم)ivrEuعمم)ixrEuعمى)iyrEuغمم)izrEuغمي)i{rEuغمى)i|rEuفخم)i~rEقمح)irEuقمم)irEلحم)irEuلحي)irEuلحى)irEuلجج)irEuلخم)irEuلمح)irEuمحج)irEuمحم)irEuمحي)irEuمجح)irEuمجم)irEuمخج)irEuمخم)ir)irEuمجخ)irEuهمج)irEuهمم)irEuنحم)irEuنحى)irEuنجم)irEuنجى)irEuنمي)irEuنمى)irEuيمم)irEuبخي)irEuتجي)irEuتجى)irEuتخي)irEuتخى)irEuتمي)irEuتمى)irEuجمي)irEuجحى)irEuجمى)irEuسخى)irEuصحي)irEuشحي)irEuضحي)irEuلجي)irEuلمي)irEuيحي)irEuيجي)irEuيمي)irEuممي)irEuقمي)irEuنحي)irEr)irEr)irEuعمي)irEuكمي)irEنجح)irEuمخي)irEلجم)irEكمم)irEr)irEr)irEuجحي)irEuحجي)irEuمجي)irEuفمي)irEuبحي)irEr)irEr)irEr)irEuسخي)irEuنجي)ir)irEuصلے)irEuقلے)irEuالله)irEuاكبر)irEuمحمد)irEuصلعم)irEuرسول)irEuعليه)irEuوسلم)irEuصلى)iru!صلى الله عليه وسلم)iruجل جلاله)irEuریال)ir0)ir)ir)ir,rrrrr_seg_48srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irE、)ir)ir:)irrf)ir!)ir?)irEu〖)irEu〗)ir)i r0)i0r)i1rE—)i2rEu–)i3r_)i5rr)i6rr)i7r{)i8r})i9rE〔)i:rE〕)i;rEu【)i<rEu】)i=rEu《)i>rEu》)i?rEr)i@rEr)iArE「)iBrE」)iCrEu『)iDrEu』)iEr0)iGr[)iHr])iIrr)iMrr)iPrr)iQrEr)iRr)iTrrf)iUrr)iVrr)iWrr)iXrEr)iYrr)iZrr)i[rr)i\rr)i]rEr)i^rEr)i_r#)i`r&)iar*)ibrr)icrE-)idr<)ier>)ifrr)igr)ihr\)iir$)ijr%)ikr@)ilr)ipru ً)iqrEuـً)irru ٌ)isr0)itru ٍ)iur)ivru َ)iwrEuـَ)ixru ُ)iyrEuـُ)izru ِ)i{rEuـِ)i|ru ّ)i}rEuـّ)i~ru ْ)irEuـْ)irEuء)irEuآ)irEuأ)irEuؤ)irEuإ)irEuئ)irEا)irEب)irEuة)irEت)irEث)irEج)irEح)irEخ)irEد)irEذ)irEر)irEز)irEس)irEش)irEصrrrrr_seg_49srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEض)irEط)irEظ)irEع)irEغ)irEف)irEق)irEك)irEل)irEم)irEن)irEه)irEو)irEr	)irEي)irEuلآ)irEuلأ)irEuلإ)irEuلا)ir)ir)ir)irr)ir")irr)irr)irr)irr)ir')irr)i	rr)i
rr)irr)irr)i
rEr)irEr.)ir/)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irr)irrf)irr)irr)irr)irr)i rr)i!rErF)i"rErH)i#rErJ)i$rErL)i%rErN)i&rErP)i'rErR)i(rErT)i)rErV)i*rErX)i+rErZ)i,rEr\)i-rEr^)i.rEr`)i/rErb)i0rErd)i1rErf)i2rErh)i3rErj)i4rErl)i5rErn)i6rErp)i7rErr)i8rErt)i9rErv)i:rErx)i;rr)i<rr)i=rr)i>r^)i?rr)i@rr)iArErF)iBrErH)iCrErJ)iDrErL)iErErN)iFrErP)iGrErR)iHrErT)iIrErV)iJrErX)iKrErZ)iLrEr\)iMrEr^)iNrEr`rrrrr_seg_50XsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iOrErb)iPrErd)iQrErf)iRrErh)iSrErj)iTrErl)iUrErn)iVrErp)iWrErr)iXrErt)iYrErv)iZrErx)i[rr)i\r|)i]rr)i^r~)i_rEu⦅)i`rEu⦆)iarEr.)ibrEr)icrEr)idrEr)ierEu・)ifrEr)igrEuァ)ihrEuィ)iirEuゥ)ijrEuェ)ikrEuォ)ilrEuャ)imrEuュ)inrEuョ)iorEuッ)iprEuー)iqrErr)irrErs)isrErt)itrEru)iurErv)ivrErw)iwrErx)ixrEry)iyrErz)izrEr{)i{rEr|)i|rEr})i}rEr~)i~rEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEuン)irEu゙)irEu゚)ir)irEr/)irEr0)irEr1)irEr2)irEr3)irEr4)irEr5)irEr6)irEr8)irEr9)irEr:)irEr;)irEr<)irEr=)irEr>)irEr?)irEr@)irErArrrrr_seg_51srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irErB)irErC)irErD)irErE)irErF)irErG)irErH)irErI)irErJ)irErK)irErL)irErM)ir)irErN)irErO)irErP)irErQ)irErR)irErS)ir)irErT)irErU)irErV)irErW)irErX)irErY)ir)irErZ)irEr[)irEr\)irEr])irEr^)irEr_)ir)irEr`)irEra)irErb)ir)irE¢)irE£)irE¬)irr)irE¦)irE¥)irEu₩)ir)irEu│)irEu←)irEu↑)irEu→)irEu↓)irEu■)irEu○)ir)ir0)ir)i
r0)i'r)i(r0)i;r)i<r0)i>r)i?r0)iNr)iPr0)i^r)ir0)ir)ir0)ir)ir0)i4r)i7r0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i$r)i-r0)iKr)iPr0)i{r)ir0)ir)ir0)ir)ir0)ir)irEu𐐨)irEu𐐩rrrrr_seg_52(srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu𐐪)irEu𐐫)irEu𐐬)irEu𐐭)irEu𐐮)irEu𐐯)irEu𐐰)i	rEu𐐱)i
rEu𐐲)irEu𐐳)irEu𐐴)i
rEu𐐵)irEu𐐶)irEu𐐷)irEu𐐸)irEu𐐹)irEu𐐺)irEu𐐻)irEu𐐼)irEu𐐽)irEu𐐾)irEu𐐿)irEu𐑀)irEu𐑁)irEu𐑂)irEu𐑃)irEu𐑄)irEu𐑅)irEu𐑆)irEu𐑇)i rEu𐑈)i!rEu𐑉)i"rEu𐑊)i#rEu𐑋)i$rEu𐑌)i%rEu𐑍)i&rEu𐑎)i'rEu𐑏)i(r0)ir)ir0)ir)irEu𐓘)irEu𐓙)irEu𐓚)irEu𐓛)irEu𐓜)irEu𐓝)irEu𐓞)irEu𐓟)irEu𐓠)irEu𐓡)irEu𐓢)irEu𐓣)irEu𐓤)irEu𐓥)irEu𐓦)irEu𐓧)irEu𐓨)irEu𐓩)irEu𐓪)irEu𐓫)irEu𐓬)irEu𐓭)irEu𐓮)irEu𐓯)irEu𐓰)irEu𐓱)irEu𐓲)irEu𐓳)irEu𐓴)irEu𐓵)irEu𐓶)irEu𐓷)irEu𐓸)irEu𐓹)irEu𐓺)irEu𐓻)ir)ir0)ir)ir0)i(r)i0r0)idr)ior0)ipr)ir0)i7r)i@r0)iVr)i`r0)ihr)ir0)ir)ir0)i	r)i
r0)i6r)i7r0rrrrr_seg_53srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i9r)i<r0)i=r)i?r0)iVr)iWr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i	r)i	r0)i:	r)i?	r0)i@	r)i	r0)i	r)i	r0)i	r)i	r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i
r)i
r0)i6
r)i8
r0)i;
r)i?
r0)iI
r)iP
r0)iY
r)i`
r0)i
r)i
r0)i
r)i
r0)i
r)ir0)i6r)i9r0)iVr)iXr0)isr)ixr0)ir)ir0)ir)ir0)ir)ir0)iIr)irEu𐳀)irEu𐳁)irEu𐳂)irEu𐳃)irEu𐳄)irEu𐳅)irEu𐳆)irEu𐳇)irEu𐳈)irEu𐳉)irEu𐳊)irEu𐳋)irEu𐳌)irEu𐳍)irEu𐳎)irEu𐳏)irEu𐳐)irEu𐳑)irEu𐳒)irEu𐳓)irEu𐳔)irEu𐳕)irEu𐳖)irEu𐳗)irEu𐳘)irEu𐳙)irEu𐳚)irEu𐳛)irEu𐳜)irEu𐳝)irEu𐳞)irEu𐳟)irEu𐳠)irEu𐳡)irEu𐳢)irEu𐳣)irEu𐳤)irEu𐳥)irEu𐳦)irEu𐳧)irEu𐳨rrrrr_seg_54srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu𐳩)irEu𐳪)irEu𐳫)irEu𐳬)irEu𐳭)irEu𐳮)irEu𐳯)irEu𐳰)irEu𐳱)irEu𐳲)ir)ir0)ir)ir0)i(
r)i0
r0)i:
r)i`r0)ir)ir0)i(r)i0r0)iZr)ir0)iNr)iRr0)ipr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i5r)i6r0)iGr)iPr0)iwr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i?r)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)ir)ir0)i)r)i*r0)i1r)i2r0)i4r)i5r0)i:r)i;r0)iEr)iGr0)iIr)iKr0)iNr)iPr0)iQr)iWr0)iXr)i]r0)idr)ifr0)imr)ipr0)iur)ir0)iZr)i[r0)i\r)i]r0rrrrr_seg_55`srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i_r)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)iEr)iPr0)iZr)i`r0)imr)ir0)ir)ir0)ir)ir0)ir)ir0)i,r)i0r0)i@r)ir0)i<r)irEu𑣀)irEu𑣁)irEu𑣂)irEu𑣃)irEu𑣄)irEu𑣅)irEu𑣆)irEu𑣇)irEu𑣈)irEu𑣉)irEu𑣊)irEu𑣋)irEu𑣌)irEu𑣍)irEu𑣎)irEu𑣏)irEu𑣐)irEu𑣑)irEu𑣒)irEu𑣓)irEu𑣔)irEu𑣕)irEu𑣖)irEu𑣗)irEu𑣘)irEu𑣙)irEu𑣚)irEu𑣛)irEu𑣜)irEu𑣝)irEu𑣞)irEu𑣟)ir0)ir)ir0)ir)ir0)iHr)iPr0)ir)ir0)ir)ir0)ir)ir0)i	r)i
r0)i7r)i8r0)iFr)iPr0)imr)ipr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i
r)ir0)i7r)i:r0)i;r)i<r0)i>r)i?r0)iHr)iPr0)iZr)i`r0rrrrr_seg_56srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ifr)igr0)iir)ijr0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)i r0)i#r)i$r0)io$r)ip$r0)iu$r)i$r0)iD%r)i0r0)i/4r)iDr0)iGFr)ihr0)i9jr)i@jr0)i_jr)i`jr0)ijjr)injr0)ipjr)ijr0)ijr)ijr0)ijr)ikr0)iFkr)iPkr0)iZkr)i[kr0)ibkr)ickr0)ixkr)i}kr0)ikr)i`nr0)inr)ior0)iEor)iPor0)ior)ior0)ior)ior0)ior)ipr0)ir)ir0)ir)ir0)ir)ipr0)ir)ir0)ikr)ipr0)i}r)ir0)ir)ir0)ir)ir0)ir)ir)ir0)ir)ir0)i'r)i)r0)i^rEu𝅗𝅥)i_rEu𝅘𝅥)i`rEu𝅘𝅥𝅮)iarEu𝅘𝅥𝅯)ibrEu𝅘𝅥𝅰)icrEu𝅘𝅥𝅱)idrEu𝅘𝅥𝅲)ier0)isr)i{r0)irEu𝆹𝅥)irEu𝆺𝅥)irEu𝆹𝅥𝅮)irEu𝆺𝅥𝅮)irEu𝆹𝅥𝅯)irEu𝆺𝅥𝅯)ir0)ir)ir0rrrrr_seg_570srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iFr)ir0)ir)ir0)iWr)i`r0)iyr)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)i	rErX)i
rErZ)irEr\)irEr^)i
rEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)i rErR)i!rErT)i"rErV)i#rErX)i$rErZ)i%rEr\)i&rEr^)i'rEr`)i(rErb)i)rErd)i*rErf)i+rErh)i,rErj)i-rErl)i.rErn)i/rErp)i0rErr)i1rErt)i2rErv)i3rErx)i4rErF)i5rErH)i6rErJ)i7rErL)i8rErN)i9rErP)i:rErR)i;rErT)i<rErV)i=rErX)i>rErZ)i?rEr\)i@rEr^)iArEr`)iBrErb)iCrErd)iDrErf)iErErh)iFrErj)iGrErl)iHrErn)iIrErp)iJrErr)iKrErt)iLrErv)iMrErx)iNrErF)iOrErH)iPrErJ)iQrErL)iRrErN)iSrErP)iTrErR)iUr)iVrErV)iWrErX)iXrErZ)iYrEr\)iZrEr^)i[rEr`)i\rErbrrrrr_seg_58srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i]rErd)i^rErf)i_rErh)i`rErj)iarErl)ibrErn)icrErp)idrErr)ierErt)ifrErv)igrErx)ihrErF)iirErH)ijrErJ)ikrErL)ilrErN)imrErP)inrErR)iorErT)iprErV)iqrErX)irrErZ)isrEr\)itrEr^)iurEr`)ivrErb)iwrErd)ixrErf)iyrErh)izrErj)i{rErl)i|rErn)i}rErp)i~rErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)ir)irErJ)irErL)ir)irErR)ir)irErX)irErZ)ir)irEr`)irErb)irErd)irErf)ir)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)ir)irErP)ir)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`rrrrr_seg_59srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)ir)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)ir)irErL)irErN)i	rErP)i
rErR)ir)i
rErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)ir)irErj)irErl)irErn)irErp)irErr)irErt)irErv)ir)irErF)irErH)i rErJ)i!rErL)i"rErN)i#rErP)i$rErR)i%rErT)i&rErV)i'rErX)i(rErZrrrrr_seg_60hsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i)rEr\)i*rEr^)i+rEr`)i,rErb)i-rErd)i.rErf)i/rErh)i0rErj)i1rErl)i2rErn)i3rErp)i4rErr)i5rErt)i6rErv)i7rErx)i8rErF)i9rErH)i:r)i;rErL)i<rErN)i=rErP)i>rErR)i?r)i@rErV)iArErX)iBrErZ)iCrEr\)iDrEr^)iEr)iFrErb)iGr)iJrErj)iKrErl)iLrErn)iMrErp)iNrErr)iOrErt)iPrErv)iQr)iRrErF)iSrErH)iTrErJ)iUrErL)iVrErN)iWrErP)iXrErR)iYrErT)iZrErV)i[rErX)i\rErZ)i]rEr\)i^rEr^)i_rEr`)i`rErb)iarErd)ibrErf)icrErh)idrErj)ierErl)ifrErn)igrErp)ihrErr)iirErt)ijrErv)ikrErx)ilrErF)imrErH)inrErJ)iorErL)iprErN)iqrErP)irrErR)isrErT)itrErV)iurErX)ivrErZ)iwrEr\)ixrEr^)iyrEr`)izrErb)i{rErd)i|rErf)i}rErh)i~rErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErVrrrrr_seg_61srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErNrrrrr_seg_628srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)i	rErH)i
rErJ)irErL)irErN)i
rErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)i rErv)i!rErx)i"rErF)i#rErH)i$rErJ)i%rErL)i&rErN)i'rErP)i(rErR)i)rErT)i*rErV)i+rErX)i,rErZ)i-rEr\)i.rEr^)i/rEr`)i0rErb)i1rErd)i2rErf)i3rErh)i4rErj)i5rErl)i6rErn)i7rErp)i8rErr)i9rErt)i:rErv)i;rErx)i<rErF)i=rErH)i>rErJ)i?rErL)i@rErN)iArErP)iBrErR)iCrErT)iDrErV)iErErX)iFrErZ)iGrEr\)iHrEr^)iIrEr`)iJrErb)iKrErd)iLrErf)iMrErh)iNrErj)iOrErl)iPrErn)iQrErp)iRrErr)iSrErt)iTrErv)iUrErx)iVrErFrrrrr_seg_63srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iWrErH)iXrErJ)iYrErL)iZrErN)i[rErP)i\rErR)i]rErT)i^rErV)i_rErX)i`rErZ)iarEr\)ibrEr^)icrEr`)idrErb)ierErd)ifrErf)igrErh)ihrErj)iirErl)ijrErn)ikrErp)ilrErr)imrErt)inrErv)iorErx)iprErF)iqrErH)irrErJ)isrErL)itrErN)iurErP)ivrErR)iwrErT)ixrErV)iyrErX)izrErZ)i{rEr\)i|rEr^)i}rEr`)i~rErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irErF)irErH)irErJ)irErL)irErN)irErP)irErR)irErT)irErV)irErX)irErZ)irEr\)irEr^)irEr`)irErb)irErd)irErf)irErh)irErj)irErl)irErn)irErp)irErr)irErt)irErv)irErx)irEuı)irEuȷ)ir)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irErw)irEr)irErrrrrr_seg_64srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irE∇)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irE∂)irErt)irErw)irErx)irEr)irEr~)irEr})irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irErw)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)i	rEr{)i
rEr|)irEr})irEr~)i
rEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErt)irErw)irErx)irEr)irEr~)irEr})irErp)irErq)irErr)irErs)i rErt)i!rErurrrrr_seg_65psrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i"rErv)i#rErw)i$rErd)i%rErx)i&rEry)i'rEr)i(rErz)i)rEr{)i*rEr|)i+rEr})i,rEr~)i-rErw)i.rEr)i/rEr)i0rEr)i1rEr)i2rEr)i3rEr)i4rEr)i5rEr)i6rErp)i7rErq)i8rErr)i9rErs)i:rErt)i;rEru)i<rErv)i=rErw)i>rErd)i?rErx)i@rEry)iArEr)iBrErz)iCrEr{)iDrEr|)iErEr})iFrEr~)iGrEr)iIrEr)iJrEr)iKrEr)iLrEr)iMrEr)iNrEr)iOrEr)iPrErt)iQrErw)iRrErx)iSrEr)iTrEr~)iUrEr})iVrErp)iWrErq)iXrErr)iYrErs)iZrErt)i[rEru)i\rErv)i]rErw)i^rErd)i_rErx)i`rEry)iarEr)ibrErz)icrEr{)idrEr|)ierEr})ifrEr~)igrErw)ihrEr)iirEr)ijrEr)ikrEr)ilrEr)imrEr)inrEr)iorEr)iprErp)iqrErq)irrErr)isrErs)itrErt)iurEru)ivrErv)iwrErw)ixrErd)iyrErx)izrEry)i{rEr)i|rErz)i}rEr{)i~rEr|)irEr})irEr~)irEr)irEr)irEr)irEr)irEr)irErrrrrr_seg_66srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irErt)irErw)irErx)irEr)irEr~)irEr})irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irErw)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErp)irErq)irErr)irErs)irErt)irEru)irErv)irErw)irErd)irErx)irEry)irEr)irErz)irEr{)irEr|)irEr})irEr~)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErt)irErw)irErx)irEr)irEr~)irEr})irEr)ir)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irErrrrrr_seg_67@srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)i"r)i#r0)i%r)i&r0)i+r)ir0)ir)ir0)ir)irEu𞤢)irEu𞤣)irEu𞤤)irEu𞤥)irEu𞤦)irEu𞤧)irEu𞤨)irEu𞤩)irEu𞤪)i	rEu𞤫)i
rEu𞤬)irEu𞤭)irEu𞤮)i
rEu𞤯)irEu𞤰)irEu𞤱)irEu𞤲)irEu𞤳)irEu𞤴)irEu𞤵)irEu𞤶)irEu𞤷)irEu𞤸)irEu𞤹)irEu𞤺)irEu𞤻)irEu𞤼)irEu𞤽)irEu𞤾)irEu𞤿)irEu𞥀)irEu𞥁)i rEu𞥂)i!rEu𞥃)i"r0)iKr)iPr0)iZr)i^r0)i`r)iqr0)ir)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)i	rEr)i
rEr)irEr)irEr)i
rEr)irEr)irEr)irEr)irEr)irEr)irEr)irErrrrrr_seg_68srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEٮ)irEr)irEڡ)irEٯ)i r)i!rEr)i"rEr)i#r)i$rEr)i%r)i'rEr)i(r)i)rEr)i*rEr)i+rEr)i,rEr)i-rEr)i.rEr)i/rEr)i0rEr)i1rEr)i2rEr)i3r)i4rEr)i5rEr)i6rEr)i7rEr)i8r)i9rEr)i:r)i;rEr)i<r)iBrEr)iCr)iGrEr)iHr)iIrEr)iJr)iKrEr)iLr)iMrEr)iNrEr)iOrEr)iPr)iQrEr)iRrEr)iSr)iTrEr)iUr)iWrEr)iXr)iYrEr)iZr)i[rEr)i\r)i]rEr)i^r)i_rEr)i`r)iarEr)ibrEr)icr)idrEr)ier)igrEr)ihrEr)iirEr)ijrEr)ikr)ilrEr)imrEr)inrEr)iorEr)iprEr)iqrEr)irrEr)isr)itrEr)iurEr)ivrEr)iwrEr)ixr)iyrEr)izrEr)i{rEr)i|rEr)i}r)i~rEr)ir)irEr)irEr)irEr)irErrrrrr_seg_69srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEr)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)irEr)ir)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)irEr)ir)ir0)ir)ir0)i,r)i0r0)ir)ir0)ir)ir0)ir)ir0)ir)ir0)ir)irz0,)irz1,)irz2,)irz3,)irz4,)irz5,)irz6,)irz7,)i	rz8,)i
rz9,)ir0)i
r)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)irr)i rr)i!rr)i"rr)i#rr)i$rrrrrrr_seg_70xsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i%rr)i&rr)i'rr)i(rr)i)rr)i*rEu〔s〕)i+rErJ)i,rErh)i-rEr)i.rEZwz)i/r0)i0rErF)i1rErH)i2rErJ)i3rErL)i4rErN)i5rErP)i6rErR)i7rErT)i8rErV)i9rErX)i:rErZ)i;rEr\)i<rEr^)i=rEr`)i>rErb)i?rErd)i@rErf)iArErh)iBrErj)iCrErl)iDrErn)iErErp)iFrErr)iGrErt)iHrErv)iIrErx)iJrEZhv)iKrEr)iLrEZsd)iMrEr))iNrEZppv)iOrEZwc)iPr0)ijrEZmc)ikrEZmd)ilr)ipr0)irEZdj)ir0)ir)ir0)irEuほか)irEuココ)irEr|)ir)irEr)irEu字)irEu双)irEuデ)irEr)irE多)irEu解)irErh)irEu交)irEu映)irEu無)irEr)irEu前)irEu後)irE再)irEu新)i rEu初)i!rEu終)i"rEr)i#rEu販)i$rEu声)i%rEu吹)i&rEu演)i'rEu投)i(rEu捕)i)rEr)i*rErc)i+rEu遊)i,rEro)i-rErf)i.rErp)i/rEu指)i0rEr#)i1rEu打)i2rEu禁)i3rEu空)i4rEu合)i5rEu満)i6rErl)i7rEr)i8rEu申)i9rE割)i:rEu営)i;rEu配rrrrr_seg_71srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i<r)i@rEu	〔本〕)iArEu	〔三〕)iBrEu	〔二〕)iCrEu	〔安〕)iDrEu	〔点〕)iErEu	〔打〕)iFrEu	〔盗〕)iGrEu	〔勝〕)iHrEu	〔敗〕)iIr)iPrEu得)iQrEu可)iRr)i`r0)ifr)ir0)ir)ir0)ir)ir0)ir)ir0)itr)ir0)ir)ir0)ir)ir0)iHr)iPr0)iZr)i`r0)ir)ir0)ir)ir0)ir)ir0)i?r)i@r0)iqr)isr0)iwr)izr0)i{r)i|r0)ir)ir0)ir)ir0)ir)ir0)ir)i`r0)inr)ir0)iצr)ir0)i5r)i@r0)ir)i r0)ir)ir0)ir)irEu丽)irEu丸)irEu乁)irEu𠄢)irEu你)irEr)irEu侻)irEu倂)irEu偺)i	rEu備)i
rEr)irEu像)irEu㒞)i
rEu𠘺)irEr)irEu兔)irEu兤)irEu具)irEu𠔜)irEu㒹)irEu內)irEr)irEu𠕋)irEu冗)irEu冤)irEu仌)irEu冬)irEr)irEu𩇟)irEr	)irEu刃)irEu㓟)i rEu刻)i!rEu剆rrrrr_seg_72HsrcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)i"rEr)i#rEu剷)i$rEu㔕)i%rEr)i&rEr)i'rEr)i(rEr)i)rEu包)i*rEu匆)i+rEr)i,rEu卉)i-rEr)i.rEu博)i/rEu即)i0rEu卽)i1rEu卿)i4rEu𠨬)i5rEu灰)i6rEu及)i7rEu叟)i8rEu𠭣)i9rEu叫)i:rEu叱)i;rEu吆)i<rEu咞)i=rEu吸)i>rEu呈)i?rEu周)i@rEu咢)iArEu哶)iBrEu唐)iCrEu啓)iDrEu啣)iErEu善)iGrEr)iHrEu喫)iIrEu喳)iJrEu嗂)iKrEu圖)iLrEr)iMrEu圗)iNrEu噑)iOrEu噴)iPrEr)iQrEu壮)iRrEu城)iSrEu埴)iTrEu堍)iUrEu型)iVrEu堲)iWrEu報)iXrEu墬)iYrEu𡓤)iZrEu売)i[rEu壷)i\rEu夆)i]rEr)i^rEu夢)i_rEu奢)i`rEu𡚨)iarEu𡛪)ibrEu姬)icrEu娛)idrEu娧)ierEu姘)ifrEu婦)igrEu㛮)ihr)iirEu嬈)ijrEu嬾)ilrEu𡧈)imrEu寃)inrEu寘)iorEr)iprEu寳)iqrEu𡬘)irrEu寿)isrEu将)itr)iurEr)ivrEu㞁)iwrEu屠)ixrEr)iyrEu峀)izrEu岍)i{rEu𡷤)i|rEu嵃)i}rEu𡷦)i~rEu嵮)irEu嵫)irEu嵼)irEu巡)irEu巢)irEu㠯)irEu巽)irEu帨)irEu帽)irEu幩)irEu㡢)irEu𢆃rrrrr_seg_73srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu㡼)irEu庰)irEu庳)irEu庶)irEr)irEu𪎒)irEr)irEu𢌱)irE舁)irEu弢)irEu㣇)irEu𣊸)irEu𦇚)irEu形)irEu彫)irEu㣣)irEu徚)irEu忍)irEu志)irEu忹)irEu悁)irEu㤺)irEu㤜)irEr)irEu𢛔)irEu惇)irEu慈)irE慌)irEr)irEr)irEu慺)irEr)irEu憲)irEu憤)irEu憯)irEu懞)irEr)irEr)irEu成)irEu戛)irEu扝)irEu抱)irEu拔)irEu捐)irEu𢬌)irEu挽)irEu拼)irEu捨)irEu掃)irEu揤)irEu𢯱)irEu搢)irEu揅)irEu掩)irEu㨮)irEu摩)irEu摾)irEu撝)irEu摷)irEu㩬)irEr)irEu敬)irEu𣀊)irEu旣)irEu書)irEu晉)irEu㬙)irEr)irEu㬈)irEu㫤)irEu冒)irEu冕)irEu最)irEu暜)irEu肭)irEu䏙)irEr)irEr)irEu朡)irEu杞)irEu杓)irEu𣏃)irEu㭉)irEu柺)irEu枅)irEu桒)irEr)irEu𣑭)irEu梎)irEu栟)irEu椔)irEr)irEu楂)irEu榣)irEu槪)irEu檨)irEu𣚣)irEu櫛)irEu㰘)irEu次rrrrr_seg_74srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEu𣢧)irEu歔)irEu㱎)irEu歲)irEu殟)irEr)irEu殻)irEu𣪍)irEu𡴋)irEu𣫺)irEu汎)irEu𣲼)irEu沿)irEu泍)irEu汧)irEu洖)irEu派)irEr)irEr)irEu浩)irEu浸)irEu涅)irEu𣴞)irEu洴)irEu港)i	rEu湮)i
rEu㴳)irEr)irEu滇)i
rEu𣻑)irEu淹)irEu潮)irEu𣽞)irEu𣾎)irEu濆)irEu瀹)irEr)irEu瀛)irEu㶖)irEu灊)irEu災)irEu灷)irEu炭)irEu𠔥)irEu煅)irEu𤉣)irEu熜)ir)i rEu爨)i!rEr)i"rEu牐)i#rEu𤘈)i$rEu犀)i%rEu犕)i&rEu𤜵)i'rEu𤠔)i(rEu獺)i)rEu王)i*rEu㺬)i+rEu玥)i,rEu㺸)i.rEu瑇)i/rEu瑜)i0rEr)i1rEu璅)i2rEu瓊)i3rEu㼛)i4rEu甤)i5rEu𤰶)i6rEu甾)i7rEu𤲒)i8rEr)i9rEu𢆟)i:rEu瘐)i;rEu𤾡)i<rEu𤾸)i=rEu𥁄)i>rEu㿼)i?rEu䀈)i@rEr)iArEu𥃳)iBrEu𥃲)iCrEu𥄙)iDrEu𥄳)iErEu眞)iFrEu真)iHrEr)iIrEr)iJrEu瞋)iKrEu䁆)iLrEu䂖)iMrEu𥐝)iNrEu硎)iOrEr)iPrEr)iQrEu䃣)iRrEu𥘦)iSrEr)iTrEu𥚚)iUrEu𥛅rrrrr_seg_75srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)iVrEr)iWrEu秫)iXrEu䄯)iYrEr)iZrEu穊)i[rEu穏)i\rEu𥥼)i]rEu𥪧)i_r)i`rEu䈂)iarEu𥮫)ibrEu篆)icrEu築)idrEu䈧)ierEu𥲀)ifrEu糒)igrEu䊠)ihrEu糨)iirEu糣)ijrEu紀)ikrEu𥾆)ilrEu絣)imrEu䌁)inrEu緇)iorEu縂)iprEu繅)iqrEu䌴)irrEu𦈨)isrEu𦉇)itrEu䍙)iurEu𦋙)ivrEu罺)iwrEu𦌾)ixrEu羕)iyrEu翺)izrEr)i{rEu𦓚)i|rEu𦔣)i}rEu聠)i~rEu𦖨)irEu聰)irEu𣍟)irEu䏕)irEu育)irEu脃)irEu䐋)irEu脾)irEu媵)irEu𦞧)irEu𦞵)irEu𣎓)irEu𣎜)irEr)irEu舄)irEu辞)irEu䑫)irEu芑)irEu芋)irEu芝)irEu劳)irEu花)irEu芳)irEu芽)irEu苦)irEu𦬼)irEr)irEu茝)irEu荣)irEu莭)irEu茣)irEu莽)irEu菧)irEr)irEu荓)irEu菊)irEu菌)irEu菜)irEu𦰶)irEu𦵫)irEu𦳕)irEu䔫)irEu蓱)irEu蓳)irEu蔖)irEu𧏊)irEu蕤)irEu𦼬)irEu䕝)irEu䕡)irEu𦾱)irEu𧃒)irEu䕫)irEu虐)irEr)irEu虧)irEu虩)irEu蚩)irEu蚈)irEu蜎)irEu蛢rrrrr_seg_76srcdCsddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcddgdS)eN)irEr)irEu蜨)irEu蝫)irEu螆)ir)irEu蟡)irEu蠁)irEu䗹)irEu衠)irEr)irEu𧙧)irEu裗)irEu裞)irEu䘵)irEu裺)irEu㒻)irEu𧢮)irEu𧥦)irEu䚾)irEu䛇)irEu誠)irEr)irEr)irEr")irEu𧲨)irEu貫)irEu賁)irEu贛)irEu起)irEu𧼯)irEu𠠄)irEu跋)irEu趼)irEu跰)irEu𠣞)irEu軔)irEr)irEu𨗒)irEu𨗭)irEu邔)irEu郱)irEu鄑)irEu𨜮)irEu鄛)irEu鈸)irEu鋗)irEu鋘)irEu鉼)irEu鏹)irEu鐕)irEu𨯺)irEu開)irEu䦕)irEu閷)irEu𨵷)irEu䧦)irEu雃)irEu嶲)irEu霣)irEu𩅅)irEu𩈚)irEu䩮)irEu䩶)irEu韠)irEu𩐊)irEu䪲)irEu𩒖)irEr)irEu頩)irEu𩖶)irEu飢)irEu䬳)irEu餩)irEu馧)irEu駂)irEu駾)irEu䯎)i	rEu𩬰)i
rEr)irEu鱀)irEu鳽)i
rEu䳎)irEu䳭)irEu鵧)irEu𪃎)irEu䳸)irEu𪄅)irEu𪈎)irEu𪊑)irEr))irEu䵖)irEr*)irEu黾)irEu鼅)irEu鼏)irEu鼖)irEr+)irEu𪘀)ir)irrrrrr_seg_77PsrcCsdgS)N)irrrrrr_seg_78srN)S__doc____version__rrrLrTr_rerrrrrrrrrrrrrrrrrrrrrr r7rirjrqrrrrrrrrrrrrr/rarjrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrtupleZ	uts46datarrrrshhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhh	

 !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNsite-packages/pip/_vendor/idna/__pycache__/compat.cpython-38.pyc000064400000001121151464452710020552 0ustar00U

.e@s,ddlTddlTddZddZddZdS)	)*cCst|SN)encodeZlabelr;/usr/lib/python3.8/site-packages/pip/_vendor/idna/compat.pyToASCIIsrcCst|Sr)decoderrrr	ToUnicodesr
cCstddS)Nz,IDNA 2008 does not utilise nameprep protocol)NotImplementedError)srrrnameprep
sr
N)Zcorecodecrr
r
rrrrssite-packages/pip/_vendor/idna/__pycache__/package_data.cpython-38.opt-1.pyc000064400000000257151464452710022623 0ustar00U

.e@sdZdS)z2.8N)__version__rrA/usr/lib/python3.8/site-packages/pip/_vendor/idna/package_data.pysite-packages/pip/_vendor/idna/__pycache__/__init__.cpython-38.pyc000064400000000335151464452710021034 0ustar00U

.e:@sddlmZddlTdS))__version__)*N)Zpackage_datarZcorerr=/usr/lib/python3.8/site-packages/pip/_vendor/idna/__init__.pyssite-packages/pip/_vendor/idna/__pycache__/core.cpython-38.opt-1.pyc000064400000021461151464452710021167 0ustar00U

.e-@s>ddlmZddlZddlZddlZddlZddlmZdZdZ	e
dZejddkr`e
ZeZGd	d
d
eZGdddeZGd
ddeZGdddeZddZddZddZddZddZddZd7ddZd d!Zd"d#Zd$d%Zd&d'Z d8d(d)Z!d*d+Z"d,d-Z#d.d/Z$d9d1d2Z%d:d3d4Z&d;d5d6Z'dS)<)idnadataN)intranges_contain	sxn--u[.。.。]c@seZdZdZdS)	IDNAErrorz7 Base exception for all IDNA-encoding related problems N__name__
__module____qualname____doc__r
r
9/usr/lib/python3.8/site-packages/pip/_vendor/idna/core.pyrsrc@seZdZdZdS)
IDNABidiErrorz= Exception when bidirectional requirements are not satisfied Nrr
r
r
rrsrc@seZdZdZdS)InvalidCodepointz> Exception when a disallowed or unallocated codepoint is used Nrr
r
r
rrsrc@seZdZdZdS)InvalidCodepointContextzE Exception when the codepoint is not valid in the context it is used Nrr
r
r
rrsrcCs0tt|}|dkr,tt|s,td|S)Nrz Unknown character in unicodedata)unicodedataZ	combiningunichrname
ValueError)cpvr
r
r_combining_class$s
rcCstt|tj|S)N)rordrZscripts)rZscriptr
r
r
_is_script+srcCs
|dS)Npunycode)encodesr
r
r	_punycode.srcCs
d|S)Nz	U+{0:04X})formatrr
r
r_unot1sr!cCst|dkrdSdS)N?FTlenlabelr
r
rvalid_label_length5sr'cCst||rdndkrdSdS)NFTr#)r&trailing_dotr
r
rvalid_string_length<sr+Fc	Cs^d}t|dD]:\}}t|}|dkrd}q|dkrd}q|sZtddS)NFrz3Unknown directionality in label {0} at position {1})RALANTr)r-r.Lz>First codepoint in label {0} must be directionality L, R or AL)
r-r.r/ENESCSETONBNNSMzHInvalid direction for codepoint at position {0} in a right-to-left label)r-r.r1r/r7)r/r1z2Can not mix numeral types in a right-to-left label)r0r1r2r3r4r5r6r7zHInvalid direction for codepoint at position {0} in a left-to-right label)r0r1z0Label ends with illegal codepoint directionality)	enumeraterZ
bidirectionalrr repr)	r&Z	check_ltrZ
bidi_labelidxr	directionZrtlZvalid_endingZnumber_typer
r
r
check_bidiCsP






r<cCs"t|dddkrtddS)NrMz0Label begins with an illegal combining characterT)rcategoryrr%r
r
rcheck_initial_combinersr?cCs<|dddkrtd|ddks0|ddkr8tdd	S)
Nz--z4Label has disallowed hyphens in 3rd and 4th positionr-z)Label must not start or end with a hyphenT)rr%r
r
rcheck_hyphen_oks
rDcCstd||krtddS)NNFCz%Label must be in Normalization Form C)r	normalizerr%r
r
r	check_nfcsrGcCs4t||}|dkr|dkr8tt||dtkr8dSd}t|dddD]B}tjt||}|tdkrrqL|tdtd	fkrLd}qqL|sdSd}t|dt|D]B}tjt||}|tdkrq|td
td	fkrd}qq|S|dkr,|dkr(tt||dtkr(dSdSdSdS)Ni rrTFrCTr0Dr-i
 )rr_virama_combining_classrangerZ
joining_typesgetr$)r&poscp_valueokiZjoining_typer
r
rvalid_contextjs<

rQcCst||}|dkrdd|kr0t|dkr`nn,t||ddkr`t||ddkr`dSdS|dkr|t|dkrt|dkrt||ddSdS|d	ks|d
kr|dkrt||ddSdS|dkr|D]6}|d
krqt|ds
t|ds
t|drdSqdSd|kr.dkrhnn6|D],}dt|krVdkr6nndSq6dSd|krdkrnn6|D],}dt|krdkrnndSqdSdS)NrrlTFiuZGreekiiZHebrewi0u・ZHiraganaZKatakanaZHani`iiii)rr$r)r&rMZ	exceptionrNrr
r
rvalid_contextos> (
" 
 
rTcCsLt|ttfr|d}t|dkr,tdt|t|t|t	|D]\}}t
|}t|tj
drpqLqLt|tj
drz,t||stdt||dt|Wn2tk
rtdt||dt|YnXqLt|tj
d	r t||s>td
t||dt|qLtdt||dt|qLt|dS)Nzutf-8rzEmpty LabelZPVALIDZCONTEXTJz-Joiner {0} not allowed at position {1} in {2}rz?Unknown codepoint adjacent to joiner {0} at position {1} in {2}ZCONTEXTO0Codepoint {0} not allowed at position {1} in {2}z0Codepoint {0} at position {1} of {2} not allowed)
isinstancebytes	bytearraydecoder$rrGrDr?r8rrrZcodepoint_classesrQrr r!r9rrTrr<)r&rMrrNr
r
rcheck_labels<


  rZcCs~z(|d}t|t|s$td|WStk
r<YnX|sJtdt|}t|t|}t|}t|sztd|S)NasciizLabel too longzNo Input)	rulabelr'rUnicodeEncodeErrorunicoderZr_alabel_prefixr%r
r
ralabel
s"
r`cCst|ttfs>z|d}Wn tk
r<t||YSX|}|trb|t	td}nt||
dS|
d}t||S)Nr[r)rVrWrXrr]rZlower
startswithr_r$rYr%r
r
rr\#s



r\Tc
Cs(ddlm}d}zt|D]\}}t|}||dkr8|nt||dfd}|d}	t|dkrj|dnd}
|	d	ks|	d
kr|r|	dkr|s|
dkr||7}q|
dk	r|	dks|	dkr|r|	d
kr|r||
7}q|	d
krtqtd|WStk
r"t	d
t||dt|YnXdS)zBRe-map the characters in the string according to UTS46 processing.r)	uts46datar,Zrr@NVrI3r=IrErU)
rcr8rbisectZbisect_leftr$
IndexErrorrrFrr r!r9)Zdomain
std3_rulestransitionalrcoutputrMcharZ
code_pointZuts46rowZstatusZreplacementr
r
ruts46_remap8sV

roc	Cst|ttfr|d}|r(t|||}d}g}|r@|d}n
t|}|rX|dgkr`td|ddkrv|d=d}|D]$}t|}|r|	|qztdqz|r|	d	d

|}t||std|S)Nr[F.r,Empty domainrCTEmpty label.zDomain too long)rVrWrXrYrosplit_unicode_dots_rerr`appendjoinr+)	rstrictuts46rkrlr*resultlabelsr&r
r
rrTs2





rcCst|ttfr|d}|r(t||d}d}g}|s@t|}n
|d}|rX|dgkr`td|dsr|d=d}|D]$}t|}|r|	|qvtdqv|r|	dd
|S)	Nr[Frpr,rqrCTrr)rVrWrXrYrorvrurr\rwrx)rryrzrkr*r{r|r&r
r
rrYss,



rY)F)F)TF)FFFF)FFF)(r,rrirresysZ	intrangesrrJr_compilervversion_infostrr^chrrUnicodeErrorrrrrrrrr!r'r+r<r?rDrGrQrTrZr`r\rorrYr
r
r
rsB

<	,
) 

site-packages/pip/_vendor/idna/__pycache__/intranges.cpython-38.opt-1.pyc000064400000003355151464452710022233 0ustar00U

.e@s0dZddlZddZddZddZd	d
ZdS)a	
Given a list of integers, made up of (hopefully) a small number of long runs
of consecutive integers, compute a representation of the form
((start1, end1), (start2, end2) ...). Then answer the question "was x present
in the original list?" in time O(log(# runs)).
NcCst|}g}d}tt|D]b}|dt|krJ||||ddkrJq||d|d}|t|d|dd|}qt|S)aRepresent a list of integers as a sequence of ranges:
    ((start_0, end_0), (start_1, end_1), ...), such that the original
    integers are exactly those x such that start_i <= x < end_i for some i.

    Ranges are encoded as single integers (start << 32 | end), not as tuples.
    r)sortedrangelenappend
_encode_rangetuple)Zlist_Zsorted_listrangesZ
last_writeiZ
current_ranger>/usr/lib/python3.8/site-packages/pip/_vendor/idna/intranges.pyintranges_from_list
srcCs|d>|BS)N r)startendrrr
rsrcCs|d?|d@fS)Nrlr)rrrr

_decode_range"srcCszt|d}t||}|dkrNt||d\}}||krF|krNnndS|t|krvt||\}}||krvdSdS)z=Determine if `int_` falls into one of the ranges in `ranges`.rrTF)rbisectZbisect_leftrr)Zint_r
Ztuple_posleftright_rrr
intranges_contain&s
r)__doc__rrrrrrrrr
s
site-packages/pip/_vendor/idna/__pycache__/codec.cpython-38.pyc000064400000005473151464452710020362 0ustar00U

.e@sddlmZmZmZmZmZddlZddlZedZ	Gdddej
Z
GdddejZGd	d
d
ej
ZGddde
ejZGd
dde
ejZddZdS))encodedecodealabelulabel	IDNAErrorNu[.。.。]c@s eZdZdddZdddZdS)	CodecstrictcCs.|dkrtd||sdSt|t|fSNr	 Unsupported error handling "{0}"r)rformatrlenselfdataerrorsr:/usr/lib/python3.8/site-packages/pip/_vendor/idna/codec.pyr	s
zCodec.encodecCs.|dkrtd||sdSt|t|fSr
)rrrrrrrrrs
zCodec.decodeN)r	)r	)__name__
__module____qualname__rrrrrrrs

rc@seZdZddZdS)IncrementalEncoderc	Cs|dkrtd||sdSt|}d}|rV|dsDd}|d=n|sV|d=|rVd}g}d}|D]*}|t||r|d7}|t|7}qbd||}|t|7}||fS)	Nr	rrr
.rr)rr_unicode_dots_resplitappendrrjoin	rrrfinallabelsZtrailing_dotresultsizeZlabelrrr_buffer_encodes0
z!IncrementalEncoder._buffer_encodeN)rrrr%rrrrrsrc@seZdZddZdS)IncrementalDecoderc	Cs|dkrtd||sdSt|tr4t|}nt|}t|d|d}d}|r~|dsld}|d=n|s~|d=|r~d}g}d}|D]*}|t||r|d	7}|t	|7}qd
||}|t	|7}||fS)
Nr	rrasciirr
rrr)rr
isinstanceZunicoderrstrrrrrr rrr_buffer_decode?s8


z!IncrementalDecoder._buffer_decodeN)rrrr*rrrrr&>sr&c@seZdZdS)StreamWriterNrrrrrrrr+gsr+c@seZdZdS)StreamReaderNr,rrrrr-jsr-c	Cs tjdtjtjttttdS)NZidna)namerrincrementalencoderincrementaldecoderstreamwriterstreamreader)	codecs	CodecInforrrrr&r+r-rrrrgetregentrymsr5)Zcorerrrrrr3recompilerrBufferedIncrementalEncoderrBufferedIncrementalDecoderr&r+r-r5rrrrs
!)site-packages/pip/_vendor/idna/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000335151464452710021773 0ustar00U

.e:@sddlmZddlTdS))__version__)*N)Zpackage_datarZcorerr=/usr/lib/python3.8/site-packages/pip/_vendor/idna/__init__.pyssite-packages/pip/_vendor/idna/__pycache__/codec.cpython-38.opt-1.pyc000064400000005473151464452710021321 0ustar00U

.e@sddlmZmZmZmZmZddlZddlZedZ	Gdddej
Z
GdddejZGd	d
d
ej
ZGddde
ejZGd
dde
ejZddZdS))encodedecodealabelulabel	IDNAErrorNu[.。.。]c@s eZdZdddZdddZdS)	CodecstrictcCs.|dkrtd||sdSt|t|fSNr	 Unsupported error handling "{0}"r)rformatrlenselfdataerrorsr:/usr/lib/python3.8/site-packages/pip/_vendor/idna/codec.pyr	s
zCodec.encodecCs.|dkrtd||sdSt|t|fSr
)rrrrrrrrrs
zCodec.decodeN)r	)r	)__name__
__module____qualname__rrrrrrrs

rc@seZdZddZdS)IncrementalEncoderc	Cs|dkrtd||sdSt|}d}|rV|dsDd}|d=n|sV|d=|rVd}g}d}|D]*}|t||r|d7}|t|7}qbd||}|t|7}||fS)	Nr	rrr
.rr)rr_unicode_dots_resplitappendrrjoin	rrrfinallabelsZtrailing_dotresultsizeZlabelrrr_buffer_encodes0
z!IncrementalEncoder._buffer_encodeN)rrrr%rrrrrsrc@seZdZddZdS)IncrementalDecoderc	Cs|dkrtd||sdSt|tr4t|}nt|}t|d|d}d}|r~|dsld}|d=n|s~|d=|r~d}g}d}|D]*}|t||r|d	7}|t	|7}qd
||}|t	|7}||fS)
Nr	rrasciirr
rrr)rr
isinstanceZunicoderrstrrrrrr rrr_buffer_decode?s8


z!IncrementalDecoder._buffer_decodeN)rrrr*rrrrr&>sr&c@seZdZdS)StreamWriterNrrrrrrrr+gsr+c@seZdZdS)StreamReaderNr,rrrrr-jsr-c	Cs tjdtjtjttttdS)NZidna)namerrincrementalencoderincrementaldecoderstreamwriterstreamreader)	codecs	CodecInforrrrr&r+r-rrrrgetregentrymsr5)Zcorerrrrrr3recompilerrBufferedIncrementalEncoderrBufferedIncrementalDecoderr&r+r-r5rrrrs
!)site-packages/pip/_vendor/idna/__pycache__/compat.cpython-38.opt-1.pyc000064400000001121151464452710021511 0ustar00U

.e@s,ddlTddlTddZddZddZdS)	)*cCst|SN)encodeZlabelr;/usr/lib/python3.8/site-packages/pip/_vendor/idna/compat.pyToASCIIsrcCst|Sr)decoderrrr	ToUnicodesr
cCstddS)Nz,IDNA 2008 does not utilise nameprep protocol)NotImplementedError)srrrnameprep
sr
N)Zcorecodecrr
r
rrrrssite-packages/pip/_vendor/idna/__pycache__/intranges.cpython-38.pyc000064400000003355151464452710021274 0ustar00U

.e@s0dZddlZddZddZddZd	d
ZdS)a	
Given a list of integers, made up of (hopefully) a small number of long runs
of consecutive integers, compute a representation of the form
((start1, end1), (start2, end2) ...). Then answer the question "was x present
in the original list?" in time O(log(# runs)).
NcCst|}g}d}tt|D]b}|dt|krJ||||ddkrJq||d|d}|t|d|dd|}qt|S)aRepresent a list of integers as a sequence of ranges:
    ((start_0, end_0), (start_1, end_1), ...), such that the original
    integers are exactly those x such that start_i <= x < end_i for some i.

    Ranges are encoded as single integers (start << 32 | end), not as tuples.
    r)sortedrangelenappend
_encode_rangetuple)Zlist_Zsorted_listrangesZ
last_writeiZ
current_ranger>/usr/lib/python3.8/site-packages/pip/_vendor/idna/intranges.pyintranges_from_list
srcCs|d>|BS)N r)startendrrr
rsrcCs|d?|d@fS)Nrlr)rrrr

_decode_range"srcCszt|d}t||}|dkrNt||d\}}||krF|krNnndS|t|krvt||\}}||krvdSdS)z=Determine if `int_` falls into one of the ranges in `ranges`.rrTF)rbisectZbisect_leftrr)Zint_r
Ztuple_posleftright_rrr
intranges_contain&s
r)__doc__rrrrrrrrr
s
site-packages/pip/_vendor/idna/__pycache__/package_data.cpython-38.pyc000064400000000257151464452710021664 0ustar00U

.e@sdZdS)z2.8N)__version__rrA/usr/lib/python3.8/site-packages/pip/_vendor/idna/package_data.pysite-packages/pip/_vendor/idna/__pycache__/idnadata.cpython-38.pyc000064400000051541151464452710021047 0ustar00U

.eß@sdZddddddZddddddddddd	d	d	d	dd	dd	dddddd	d	d	d	dddddddddddddd
dddddddd	ddddd	d	d	dd	d	d	ddddddddddddddddd	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	ddddddddddddddddddddddddddddddddddddddd	ddd	d	d	d	d	d	d	d	d	dd	dd	ddd	d	d	dd	d	dddddd	dddd	d	d	d	d	ddddd	dddddddddd	dd	dd	ddd	d	dddddddddddd	d	d	dddddddddddddddd	d	ddddd	dd	d	dddd	d	dddddddddddddddddddddddddddddddddddddddd
d	dddddd	d	dd	ddddddddddd	dddddddddddd	dd	d	ddddddddddd	d	d	dd	ddd	d	dddddd	dddddddd
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd
ddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	dd	dd	d	dddd	d	d	d	d	ddddddddddd	dddd	ddd	ddddd	dd	dd	d	d	dddd	ddd	dd	d	dd	d	d	d	d	dddddddddddddddddddddddddddddddddddddd	ddddd	dddddddddddddddddddddd	ddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd
ZddddZdS)z11.0.0)$lt
lx
l~
l
lll l0l8lll+tlbtulkulvl|l`|lF|lN }X@}lZd}l\l}l^t}l~|}l~l~llXltlll'!lf+-lllFRH)l.:l.l:l/<l0@l0@l*0@l<0@MP8lnzdlzi&578]N`k:lz`)	lD@ll7{tll={ll?{llB{mlE{mlP{m)0Al0tBl1@
lrH)0Bl1C2Gl2@KlX3Llp}l}l0@
)ZGreekZHanZHebrewZHiraganaZKatakanaUDRCTL(iiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJinioiqirisitiuiviwixiyizi{i|i}i~iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/iMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiyizi{i|i}i~iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXi`iaibicidieifigihiiijiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii
ii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii i
 i/ if ig ih ii i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisi
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
iiiiiiiiiiiiiiiiiiiiiiiiii
i
i
i
i
i
i
i
i
i	
i

i
i
i

i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i 
i!
i"
i#
i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiQiRiSiTiiiiiiiiiiii	i
iii
iiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiC(ql.l:l{l|llllll
$l,l4l<lDlLlTl\ldllltl |l"l$l&l(l*l,l.l0l2l6l9l;l=l?lClElGlI lL,lN4lP<lRDlTLlVTlX\lZdl\ll^tl`|lbldlflhljlllnlplrltlvlxl{l}lllll l0lHlTldlxllllllllllll8l@lHlPlXl`lhlpl|llllllllllllllllll
$l,l4l<lDlLlTl\ldllltl |l"l$l&l(l*l,l.l0l2l:l=lAlC	lH	lJ$	lL,	lN4	l<	l
llll@lC
lO
lp@
lr
lt
lx
l~
l@ll\ldllltl|llllllllllll`lbldlflhljlllnlplrltlvlxlzl|l~llll,l4l<lDlLlTl\ldllltl|llllllllllllllllllll l(l0l8lDlLlTl\ldllltl|lllllllllllllllllllll
$l,l4l<lDlLlTl\ldllltl |l"l$l&l(l*l,l.l0lZdll lDllllr	ll@l@l`lullTl|lllK@l4lll. l\!lk!l"l"lL#lX	#ld	%lp	%l	%l	&l	<&l	L&l	&l	&l	&l	&l	'l	,'l	\'l	'l	'l	'l	'l
(l
(l
<(l)
L(l1
(l3
(l6
(l:
(l=
(lC
(lI
)lN
,)lR
D)l]
p)lv
)l
*l
*l
<*l
L*l
*l
*l
*l
*l
+l
,+l
@+l
+l
+l+l,l
,l<,l)L,l1,l4,l:,lE,lI-lN,-lXX-ld|-lp-lr-l.l.l8.lH.ld.lp.lx.l.l.l.l.l/l(/l@/l\/l/l
0l80l)H0l:0lE0lI1lN(1lWT1l[`1ld1lp1l2l2l82lH2l2l2l2l3l(3lT3lx3l3l3l3l
4l

4l
84lE
H4lI
5lO
(5lX
P5ld
|5lp
5l
5l
6l
6l
h6l
6l
6l
7l
(7l
<7l
X7l
`7l
7l
7l38l;8lO9lZ@9l:l:l:l(:l4:lP:ld:l:l:l:l:l:l:l:l;l;l ;l@;lx;ll>lP>ld>lx>l>l>l>l>l?lJ@l@Al@ClClIHlN(IlW@IlY`Il^hIlIl(Jl@JlJlJlKlKl Kl`KlHLl[`Ll`tMlNlNlmPlYlZlZl[l
\l8\l5\lT]lm]lq]lt]l^l^l\_lp_l_l@`ly`lblbldl,dl<dlneluelflfl@glhl_hl}ilil@jljljlLllZ@mltmlnl8plJql~4ql@slPsl,tl0tl<tlO8ulxululwlwlxlxlxlxl
$xl,xl4xlxl@xlBylDylFylHylJ$ylL,ylN4ylP'l@'lB'lD'lF'lH'lJ'$lL',lN'4lP'<lR'DlT'LlV'TlX'\lZ'dl\'ll^'tl`'|lb'ld'lf'lh'lj'll'ln'lp'ly'l{'l}'l'l'l'l'l'l'0l'8l'Dl'Ll'\l'dl'll'tl'|l'l'l'l'l'l'l'l'l'l'l((lt(!l("l(@#l(#l(#l.)#lT)$l)&l)<'l)'l7*(lN*)lZ*@)lw*)l*)l*l+l*+l*+l+,l+$,l+D,l'+,l/+,l[+,lf+-l+/l+/l+/lW0lz8hlzDhlzLhl z|hl"zhl%zhl*zhl{xll0~xlt~yll'4l;l>lNl^@lll
l
ll lAlJ
l{@
lll lll`l(ldl7lVlhl l	  l6( l9 l= lV lw!l"l#l#l	$l:	$l	&l	&l
(l
(l
0(l
T(l6
d(l;
(l@
(l}
)l
*l
+l
$+l6,lV-ls-l.lI0l3l(
4l:
4l<l(<lQ<lG@lpAlAl@ClCl5Dl@DlGElt@ElwElFl$Gl@GlpGlHl8LHl?HlJl Jl(JltlHtlZ@ulfuliulul@vlLvlvl{l#	lD%	l/4@	lGF
l9j l_j)ljj)lj@+lj+l7k,lDk-lZk@-lxk-lk-ln9lEo<lo@=lo<>lo?l@l
 l1@
l2E
lk<p
l}<q
l<r
l<@r
l<tr
l7ZhlmZhlvZilZjlZljlZjl`l` l"`ll%`l+`lh lh@#lKi$lZi@%rrrrr)l 0)llv
lljll0C)ZPVALIDZCONTEXTJZCONTEXTON)__version__ZscriptsZ
joining_typesZcodepoint_classesrr=/usr/lib/python3.8/site-packages/pip/_vendor/idna/idnadata.pys
&Wh{site-packages/pip/_vendor/idna/compat.py000064400000000350151464452710014267 0ustar00from .core import *
from .codec import *

def ToASCII(label):
    return encode(label)

def ToUnicode(label):
    return decode(label)

def nameprep(s):
    raise NotImplementedError("IDNA 2008 does not utilise nameprep protocol")

site-packages/pip/_vendor/idna/__init__.py000064400000000072151464452710014544 0ustar00from .package_data import __version__
from .core import *
site-packages/pip/_vendor/idna/intranges.py000064400000003325151464452710015003 0ustar00"""
Given a list of integers, made up of (hopefully) a small number of long runs
of consecutive integers, compute a representation of the form
((start1, end1), (start2, end2) ...). Then answer the question "was x present
in the original list?" in time O(log(# runs)).
"""

import bisect

def intranges_from_list(list_):
    """Represent a list of integers as a sequence of ranges:
    ((start_0, end_0), (start_1, end_1), ...), such that the original
    integers are exactly those x such that start_i <= x < end_i for some i.

    Ranges are encoded as single integers (start << 32 | end), not as tuples.
    """

    sorted_list = sorted(list_)
    ranges = []
    last_write = -1
    for i in range(len(sorted_list)):
        if i+1 < len(sorted_list):
            if sorted_list[i] == sorted_list[i+1]-1:
                continue
        current_range = sorted_list[last_write+1:i+1]
        ranges.append(_encode_range(current_range[0], current_range[-1] + 1))
        last_write = i

    return tuple(ranges)

def _encode_range(start, end):
    return (start << 32) | end

def _decode_range(r):
    return (r >> 32), (r & ((1 << 32) - 1))


def intranges_contain(int_, ranges):
    """Determine if `int_` falls into one of the ranges in `ranges`."""
    tuple_ = _encode_range(int_, 0)
    pos = bisect.bisect_left(ranges, tuple_)
    # we could be immediately ahead of a tuple (start, end)
    # with start < int_ <= end
    if pos > 0:
        left, right = _decode_range(ranges[pos-1])
        if left <= int_ < right:
            return True
    # or we could be immediately behind a tuple (int_, end)
    if pos < len(ranges):
        left, _ = _decode_range(ranges[pos])
        if left == int_:
            return True
    return False
site-packages/pip/_vendor/idna/core.py000064400000026725151464452710013752 0ustar00from . import idnadata
import bisect
import unicodedata
import re
import sys
from .intranges import intranges_contain

_virama_combining_class = 9
_alabel_prefix = b'xn--'
_unicode_dots_re = re.compile(u'[\u002e\u3002\uff0e\uff61]')

if sys.version_info[0] == 3:
    unicode = str
    unichr = chr

class IDNAError(UnicodeError):
    """ Base exception for all IDNA-encoding related problems """
    pass


class IDNABidiError(IDNAError):
    """ Exception when bidirectional requirements are not satisfied """
    pass


class InvalidCodepoint(IDNAError):
    """ Exception when a disallowed or unallocated codepoint is used """
    pass


class InvalidCodepointContext(IDNAError):
    """ Exception when the codepoint is not valid in the context it is used """
    pass


def _combining_class(cp):
    v = unicodedata.combining(unichr(cp))
    if v == 0:
        if not unicodedata.name(unichr(cp)):
            raise ValueError("Unknown character in unicodedata")
    return v

def _is_script(cp, script):
    return intranges_contain(ord(cp), idnadata.scripts[script])

def _punycode(s):
    return s.encode('punycode')

def _unot(s):
    return 'U+{0:04X}'.format(s)


def valid_label_length(label):

    if len(label) > 63:
        return False
    return True


def valid_string_length(label, trailing_dot):

    if len(label) > (254 if trailing_dot else 253):
        return False
    return True


def check_bidi(label, check_ltr=False):

    # Bidi rules should only be applied if string contains RTL characters
    bidi_label = False
    for (idx, cp) in enumerate(label, 1):
        direction = unicodedata.bidirectional(cp)
        if direction == '':
            # String likely comes from a newer version of Unicode
            raise IDNABidiError('Unknown directionality in label {0} at position {1}'.format(repr(label), idx))
        if direction in ['R', 'AL', 'AN']:
            bidi_label = True
    if not bidi_label and not check_ltr:
        return True

    # Bidi rule 1
    direction = unicodedata.bidirectional(label[0])
    if direction in ['R', 'AL']:
        rtl = True
    elif direction == 'L':
        rtl = False
    else:
        raise IDNABidiError('First codepoint in label {0} must be directionality L, R or AL'.format(repr(label)))

    valid_ending = False
    number_type = False
    for (idx, cp) in enumerate(label, 1):
        direction = unicodedata.bidirectional(cp)

        if rtl:
            # Bidi rule 2
            if not direction in ['R', 'AL', 'AN', 'EN', 'ES', 'CS', 'ET', 'ON', 'BN', 'NSM']:
                raise IDNABidiError('Invalid direction for codepoint at position {0} in a right-to-left label'.format(idx))
            # Bidi rule 3
            if direction in ['R', 'AL', 'EN', 'AN']:
                valid_ending = True
            elif direction != 'NSM':
                valid_ending = False
            # Bidi rule 4
            if direction in ['AN', 'EN']:
                if not number_type:
                    number_type = direction
                else:
                    if number_type != direction:
                        raise IDNABidiError('Can not mix numeral types in a right-to-left label')
        else:
            # Bidi rule 5
            if not direction in ['L', 'EN', 'ES', 'CS', 'ET', 'ON', 'BN', 'NSM']:
                raise IDNABidiError('Invalid direction for codepoint at position {0} in a left-to-right label'.format(idx))
            # Bidi rule 6
            if direction in ['L', 'EN']:
                valid_ending = True
            elif direction != 'NSM':
                valid_ending = False

    if not valid_ending:
        raise IDNABidiError('Label ends with illegal codepoint directionality')

    return True


def check_initial_combiner(label):

    if unicodedata.category(label[0])[0] == 'M':
        raise IDNAError('Label begins with an illegal combining character')
    return True


def check_hyphen_ok(label):

    if label[2:4] == '--':
        raise IDNAError('Label has disallowed hyphens in 3rd and 4th position')
    if label[0] == '-' or label[-1] == '-':
        raise IDNAError('Label must not start or end with a hyphen')
    return True


def check_nfc(label):

    if unicodedata.normalize('NFC', label) != label:
        raise IDNAError('Label must be in Normalization Form C')


def valid_contextj(label, pos):

    cp_value = ord(label[pos])

    if cp_value == 0x200c:

        if pos > 0:
            if _combining_class(ord(label[pos - 1])) == _virama_combining_class:
                return True

        ok = False
        for i in range(pos-1, -1, -1):
            joining_type = idnadata.joining_types.get(ord(label[i]))
            if joining_type == ord('T'):
                continue
            if joining_type in [ord('L'), ord('D')]:
                ok = True
                break

        if not ok:
            return False

        ok = False
        for i in range(pos+1, len(label)):
            joining_type = idnadata.joining_types.get(ord(label[i]))
            if joining_type == ord('T'):
                continue
            if joining_type in [ord('R'), ord('D')]:
                ok = True
                break
        return ok

    if cp_value == 0x200d:

        if pos > 0:
            if _combining_class(ord(label[pos - 1])) == _virama_combining_class:
                return True
        return False

    else:

        return False


def valid_contexto(label, pos, exception=False):

    cp_value = ord(label[pos])

    if cp_value == 0x00b7:
        if 0 < pos < len(label)-1:
            if ord(label[pos - 1]) == 0x006c and ord(label[pos + 1]) == 0x006c:
                return True
        return False

    elif cp_value == 0x0375:
        if pos < len(label)-1 and len(label) > 1:
            return _is_script(label[pos + 1], 'Greek')
        return False

    elif cp_value == 0x05f3 or cp_value == 0x05f4:
        if pos > 0:
            return _is_script(label[pos - 1], 'Hebrew')
        return False

    elif cp_value == 0x30fb:
        for cp in label:
            if cp == u'\u30fb':
                continue
            if _is_script(cp, 'Hiragana') or _is_script(cp, 'Katakana') or _is_script(cp, 'Han'):
                return True
        return False

    elif 0x660 <= cp_value <= 0x669:
        for cp in label:
            if 0x6f0 <= ord(cp) <= 0x06f9:
                return False
        return True

    elif 0x6f0 <= cp_value <= 0x6f9:
        for cp in label:
            if 0x660 <= ord(cp) <= 0x0669:
                return False
        return True


def check_label(label):

    if isinstance(label, (bytes, bytearray)):
        label = label.decode('utf-8')
    if len(label) == 0:
        raise IDNAError('Empty Label')

    check_nfc(label)
    check_hyphen_ok(label)
    check_initial_combiner(label)

    for (pos, cp) in enumerate(label):
        cp_value = ord(cp)
        if intranges_contain(cp_value, idnadata.codepoint_classes['PVALID']):
            continue
        elif intranges_contain(cp_value, idnadata.codepoint_classes['CONTEXTJ']):
            try:
                if not valid_contextj(label, pos):
                    raise InvalidCodepointContext('Joiner {0} not allowed at position {1} in {2}'.format(
                        _unot(cp_value), pos+1, repr(label)))
            except ValueError:
                raise IDNAError('Unknown codepoint adjacent to joiner {0} at position {1} in {2}'.format(
                    _unot(cp_value), pos+1, repr(label)))
        elif intranges_contain(cp_value, idnadata.codepoint_classes['CONTEXTO']):
            if not valid_contexto(label, pos):
                raise InvalidCodepointContext('Codepoint {0} not allowed at position {1} in {2}'.format(_unot(cp_value), pos+1, repr(label)))
        else:
            raise InvalidCodepoint('Codepoint {0} at position {1} of {2} not allowed'.format(_unot(cp_value), pos+1, repr(label)))

    check_bidi(label)


def alabel(label):

    try:
        label = label.encode('ascii')
        ulabel(label)
        if not valid_label_length(label):
            raise IDNAError('Label too long')
        return label
    except UnicodeEncodeError:
        pass

    if not label:
        raise IDNAError('No Input')

    label = unicode(label)
    check_label(label)
    label = _punycode(label)
    label = _alabel_prefix + label

    if not valid_label_length(label):
        raise IDNAError('Label too long')

    return label


def ulabel(label):

    if not isinstance(label, (bytes, bytearray)):
        try:
            label = label.encode('ascii')
        except UnicodeEncodeError:
            check_label(label)
            return label

    label = label.lower()
    if label.startswith(_alabel_prefix):
        label = label[len(_alabel_prefix):]
    else:
        check_label(label)
        return label.decode('ascii')

    label = label.decode('punycode')
    check_label(label)
    return label


def uts46_remap(domain, std3_rules=True, transitional=False):
    """Re-map the characters in the string according to UTS46 processing."""
    from .uts46data import uts46data
    output = u""
    try:
        for pos, char in enumerate(domain):
            code_point = ord(char)
            uts46row = uts46data[code_point if code_point < 256 else
                bisect.bisect_left(uts46data, (code_point, "Z")) - 1]
            status = uts46row[1]
            replacement = uts46row[2] if len(uts46row) == 3 else None
            if (status == "V" or
                    (status == "D" and not transitional) or
                    (status == "3" and not std3_rules and replacement is None)):
                output += char
            elif replacement is not None and (status == "M" or
                    (status == "3" and not std3_rules) or
                    (status == "D" and transitional)):
                output += replacement
            elif status != "I":
                raise IndexError()
        return unicodedata.normalize("NFC", output)
    except IndexError:
        raise InvalidCodepoint(
            "Codepoint {0} not allowed at position {1} in {2}".format(
            _unot(code_point), pos + 1, repr(domain)))


def encode(s, strict=False, uts46=False, std3_rules=False, transitional=False):

    if isinstance(s, (bytes, bytearray)):
        s = s.decode("ascii")
    if uts46:
        s = uts46_remap(s, std3_rules, transitional)
    trailing_dot = False
    result = []
    if strict:
        labels = s.split('.')
    else:
        labels = _unicode_dots_re.split(s)
    if not labels or labels == ['']:
        raise IDNAError('Empty domain')
    if labels[-1] == '':
        del labels[-1]
        trailing_dot = True
    for label in labels:
        s = alabel(label)
        if s:
            result.append(s)
        else:
            raise IDNAError('Empty label')
    if trailing_dot:
        result.append(b'')
    s = b'.'.join(result)
    if not valid_string_length(s, trailing_dot):
        raise IDNAError('Domain too long')
    return s


def decode(s, strict=False, uts46=False, std3_rules=False):

    if isinstance(s, (bytes, bytearray)):
        s = s.decode("ascii")
    if uts46:
        s = uts46_remap(s, std3_rules, False)
    trailing_dot = False
    result = []
    if not strict:
        labels = _unicode_dots_re.split(s)
    else:
        labels = s.split(u'.')
    if not labels or labels == ['']:
        raise IDNAError('Empty domain')
    if not labels[-1]:
        del labels[-1]
        trailing_dot = True
    for label in labels:
        s = ulabel(label)
        if s:
            result.append(s)
        else:
            raise IDNAError('Empty label')
    if trailing_dot:
        result.append(u'')
    return u'.'.join(result)
site-packages/pip/_vendor/webencodings/mklabels.py000064400000002431151464452710016334 0ustar00"""

    webencodings.mklabels
    ~~~~~~~~~~~~~~~~~~~~~

    Regenarate the webencodings.labels module.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

import json
try:
    from urllib import urlopen
except ImportError:
    from urllib.request import urlopen


def assert_lower(string):
    assert string == string.lower()
    return string


def generate(url):
    parts = ['''\
"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
''']
    labels = [
        (repr(assert_lower(label)).lstrip('u'),
         repr(encoding['name']).lstrip('u'))
        for category in json.loads(urlopen(url).read().decode('ascii'))
        for encoding in category['encodings']
        for label in encoding['labels']]
    max_len = max(len(label) for label, name in labels)
    parts.extend(
        '    %s:%s %s,\n' % (label, ' ' * (max_len - len(label)), name)
        for label, name in labels)
    parts.append('}')
    return ''.join(parts)


if __name__ == '__main__':
    print(generate('http://encoding.spec.whatwg.org/encodings.json'))
site-packages/pip/_vendor/webencodings/labels.py000064400000021423151464452710016006 0ustar00"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
    'unicode-1-1-utf-8':   'utf-8',
    'utf-8':               'utf-8',
    'utf8':                'utf-8',
    '866':                 'ibm866',
    'cp866':               'ibm866',
    'csibm866':            'ibm866',
    'ibm866':              'ibm866',
    'csisolatin2':         'iso-8859-2',
    'iso-8859-2':          'iso-8859-2',
    'iso-ir-101':          'iso-8859-2',
    'iso8859-2':           'iso-8859-2',
    'iso88592':            'iso-8859-2',
    'iso_8859-2':          'iso-8859-2',
    'iso_8859-2:1987':     'iso-8859-2',
    'l2':                  'iso-8859-2',
    'latin2':              'iso-8859-2',
    'csisolatin3':         'iso-8859-3',
    'iso-8859-3':          'iso-8859-3',
    'iso-ir-109':          'iso-8859-3',
    'iso8859-3':           'iso-8859-3',
    'iso88593':            'iso-8859-3',
    'iso_8859-3':          'iso-8859-3',
    'iso_8859-3:1988':     'iso-8859-3',
    'l3':                  'iso-8859-3',
    'latin3':              'iso-8859-3',
    'csisolatin4':         'iso-8859-4',
    'iso-8859-4':          'iso-8859-4',
    'iso-ir-110':          'iso-8859-4',
    'iso8859-4':           'iso-8859-4',
    'iso88594':            'iso-8859-4',
    'iso_8859-4':          'iso-8859-4',
    'iso_8859-4:1988':     'iso-8859-4',
    'l4':                  'iso-8859-4',
    'latin4':              'iso-8859-4',
    'csisolatincyrillic':  'iso-8859-5',
    'cyrillic':            'iso-8859-5',
    'iso-8859-5':          'iso-8859-5',
    'iso-ir-144':          'iso-8859-5',
    'iso8859-5':           'iso-8859-5',
    'iso88595':            'iso-8859-5',
    'iso_8859-5':          'iso-8859-5',
    'iso_8859-5:1988':     'iso-8859-5',
    'arabic':              'iso-8859-6',
    'asmo-708':            'iso-8859-6',
    'csiso88596e':         'iso-8859-6',
    'csiso88596i':         'iso-8859-6',
    'csisolatinarabic':    'iso-8859-6',
    'ecma-114':            'iso-8859-6',
    'iso-8859-6':          'iso-8859-6',
    'iso-8859-6-e':        'iso-8859-6',
    'iso-8859-6-i':        'iso-8859-6',
    'iso-ir-127':          'iso-8859-6',
    'iso8859-6':           'iso-8859-6',
    'iso88596':            'iso-8859-6',
    'iso_8859-6':          'iso-8859-6',
    'iso_8859-6:1987':     'iso-8859-6',
    'csisolatingreek':     'iso-8859-7',
    'ecma-118':            'iso-8859-7',
    'elot_928':            'iso-8859-7',
    'greek':               'iso-8859-7',
    'greek8':              'iso-8859-7',
    'iso-8859-7':          'iso-8859-7',
    'iso-ir-126':          'iso-8859-7',
    'iso8859-7':           'iso-8859-7',
    'iso88597':            'iso-8859-7',
    'iso_8859-7':          'iso-8859-7',
    'iso_8859-7:1987':     'iso-8859-7',
    'sun_eu_greek':        'iso-8859-7',
    'csiso88598e':         'iso-8859-8',
    'csisolatinhebrew':    'iso-8859-8',
    'hebrew':              'iso-8859-8',
    'iso-8859-8':          'iso-8859-8',
    'iso-8859-8-e':        'iso-8859-8',
    'iso-ir-138':          'iso-8859-8',
    'iso8859-8':           'iso-8859-8',
    'iso88598':            'iso-8859-8',
    'iso_8859-8':          'iso-8859-8',
    'iso_8859-8:1988':     'iso-8859-8',
    'visual':              'iso-8859-8',
    'csiso88598i':         'iso-8859-8-i',
    'iso-8859-8-i':        'iso-8859-8-i',
    'logical':             'iso-8859-8-i',
    'csisolatin6':         'iso-8859-10',
    'iso-8859-10':         'iso-8859-10',
    'iso-ir-157':          'iso-8859-10',
    'iso8859-10':          'iso-8859-10',
    'iso885910':           'iso-8859-10',
    'l6':                  'iso-8859-10',
    'latin6':              'iso-8859-10',
    'iso-8859-13':         'iso-8859-13',
    'iso8859-13':          'iso-8859-13',
    'iso885913':           'iso-8859-13',
    'iso-8859-14':         'iso-8859-14',
    'iso8859-14':          'iso-8859-14',
    'iso885914':           'iso-8859-14',
    'csisolatin9':         'iso-8859-15',
    'iso-8859-15':         'iso-8859-15',
    'iso8859-15':          'iso-8859-15',
    'iso885915':           'iso-8859-15',
    'iso_8859-15':         'iso-8859-15',
    'l9':                  'iso-8859-15',
    'iso-8859-16':         'iso-8859-16',
    'cskoi8r':             'koi8-r',
    'koi':                 'koi8-r',
    'koi8':                'koi8-r',
    'koi8-r':              'koi8-r',
    'koi8_r':              'koi8-r',
    'koi8-u':              'koi8-u',
    'csmacintosh':         'macintosh',
    'mac':                 'macintosh',
    'macintosh':           'macintosh',
    'x-mac-roman':         'macintosh',
    'dos-874':             'windows-874',
    'iso-8859-11':         'windows-874',
    'iso8859-11':          'windows-874',
    'iso885911':           'windows-874',
    'tis-620':             'windows-874',
    'windows-874':         'windows-874',
    'cp1250':              'windows-1250',
    'windows-1250':        'windows-1250',
    'x-cp1250':            'windows-1250',
    'cp1251':              'windows-1251',
    'windows-1251':        'windows-1251',
    'x-cp1251':            'windows-1251',
    'ansi_x3.4-1968':      'windows-1252',
    'ascii':               'windows-1252',
    'cp1252':              'windows-1252',
    'cp819':               'windows-1252',
    'csisolatin1':         'windows-1252',
    'ibm819':              'windows-1252',
    'iso-8859-1':          'windows-1252',
    'iso-ir-100':          'windows-1252',
    'iso8859-1':           'windows-1252',
    'iso88591':            'windows-1252',
    'iso_8859-1':          'windows-1252',
    'iso_8859-1:1987':     'windows-1252',
    'l1':                  'windows-1252',
    'latin1':              'windows-1252',
    'us-ascii':            'windows-1252',
    'windows-1252':        'windows-1252',
    'x-cp1252':            'windows-1252',
    'cp1253':              'windows-1253',
    'windows-1253':        'windows-1253',
    'x-cp1253':            'windows-1253',
    'cp1254':              'windows-1254',
    'csisolatin5':         'windows-1254',
    'iso-8859-9':          'windows-1254',
    'iso-ir-148':          'windows-1254',
    'iso8859-9':           'windows-1254',
    'iso88599':            'windows-1254',
    'iso_8859-9':          'windows-1254',
    'iso_8859-9:1989':     'windows-1254',
    'l5':                  'windows-1254',
    'latin5':              'windows-1254',
    'windows-1254':        'windows-1254',
    'x-cp1254':            'windows-1254',
    'cp1255':              'windows-1255',
    'windows-1255':        'windows-1255',
    'x-cp1255':            'windows-1255',
    'cp1256':              'windows-1256',
    'windows-1256':        'windows-1256',
    'x-cp1256':            'windows-1256',
    'cp1257':              'windows-1257',
    'windows-1257':        'windows-1257',
    'x-cp1257':            'windows-1257',
    'cp1258':              'windows-1258',
    'windows-1258':        'windows-1258',
    'x-cp1258':            'windows-1258',
    'x-mac-cyrillic':      'x-mac-cyrillic',
    'x-mac-ukrainian':     'x-mac-cyrillic',
    'chinese':             'gbk',
    'csgb2312':            'gbk',
    'csiso58gb231280':     'gbk',
    'gb2312':              'gbk',
    'gb_2312':             'gbk',
    'gb_2312-80':          'gbk',
    'gbk':                 'gbk',
    'iso-ir-58':           'gbk',
    'x-gbk':               'gbk',
    'gb18030':             'gb18030',
    'hz-gb-2312':          'hz-gb-2312',
    'big5':                'big5',
    'big5-hkscs':          'big5',
    'cn-big5':             'big5',
    'csbig5':              'big5',
    'x-x-big5':            'big5',
    'cseucpkdfmtjapanese': 'euc-jp',
    'euc-jp':              'euc-jp',
    'x-euc-jp':            'euc-jp',
    'csiso2022jp':         'iso-2022-jp',
    'iso-2022-jp':         'iso-2022-jp',
    'csshiftjis':          'shift_jis',
    'ms_kanji':            'shift_jis',
    'shift-jis':           'shift_jis',
    'shift_jis':           'shift_jis',
    'sjis':                'shift_jis',
    'windows-31j':         'shift_jis',
    'x-sjis':              'shift_jis',
    'cseuckr':             'euc-kr',
    'csksc56011987':       'euc-kr',
    'euc-kr':              'euc-kr',
    'iso-ir-149':          'euc-kr',
    'korean':              'euc-kr',
    'ks_c_5601-1987':      'euc-kr',
    'ks_c_5601-1989':      'euc-kr',
    'ksc5601':             'euc-kr',
    'ksc_5601':            'euc-kr',
    'windows-949':         'euc-kr',
    'csiso2022kr':         'iso-2022-kr',
    'iso-2022-kr':         'iso-2022-kr',
    'utf-16be':            'utf-16be',
    'utf-16':              'utf-16le',
    'utf-16le':            'utf-16le',
    'x-user-defined':      'x-user-defined',
}
site-packages/pip/_vendor/webencodings/tests.py000064400000014643151464452710015714 0ustar00# coding: utf-8
"""

    webencodings.tests
    ~~~~~~~~~~~~~~~~~~

    A basic test suite for Encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

from __future__ import unicode_literals

from . import (lookup, LABELS, decode, encode, iter_decode, iter_encode,
               IncrementalDecoder, IncrementalEncoder, UTF8)


def assert_raises(exception, function, *args, **kwargs):
    try:
        function(*args, **kwargs)
    except exception:
        return
    else:  # pragma: no cover
        raise AssertionError('Did not raise %s.' % exception)


def test_labels():
    assert lookup('utf-8').name == 'utf-8'
    assert lookup('Utf-8').name == 'utf-8'
    assert lookup('UTF-8').name == 'utf-8'
    assert lookup('utf8').name == 'utf-8'
    assert lookup('utf8').name == 'utf-8'
    assert lookup('utf8 ').name == 'utf-8'
    assert lookup(' \r\nutf8\t').name == 'utf-8'
    assert lookup('u8') is None  # Python label.
    assert lookup('utf-8 ') is None  # Non-ASCII white space.

    assert lookup('US-ASCII').name == 'windows-1252'
    assert lookup('iso-8859-1').name == 'windows-1252'
    assert lookup('latin1').name == 'windows-1252'
    assert lookup('LATIN1').name == 'windows-1252'
    assert lookup('latin-1') is None
    assert lookup('LATİN1') is None  # ASCII-only case insensitivity.


def test_all_labels():
    for label in LABELS:
        assert decode(b'', label) == ('', lookup(label))
        assert encode('', label) == b''
        for repeat in [0, 1, 12]:
            output, _ = iter_decode([b''] * repeat, label)
            assert list(output) == []
            assert list(iter_encode([''] * repeat, label)) == []
        decoder = IncrementalDecoder(label)
        assert decoder.decode(b'') == ''
        assert decoder.decode(b'', final=True) == ''
        encoder = IncrementalEncoder(label)
        assert encoder.encode('') == b''
        assert encoder.encode('', final=True) == b''
    # All encoding names are valid labels too:
    for name in set(LABELS.values()):
        assert lookup(name).name == name


def test_invalid_label():
    assert_raises(LookupError, decode, b'\xEF\xBB\xBF\xc3\xa9', 'invalid')
    assert_raises(LookupError, encode, 'é', 'invalid')
    assert_raises(LookupError, iter_decode, [], 'invalid')
    assert_raises(LookupError, iter_encode, [], 'invalid')
    assert_raises(LookupError, IncrementalDecoder, 'invalid')
    assert_raises(LookupError, IncrementalEncoder, 'invalid')


def test_decode():
    assert decode(b'\x80', 'latin1') == ('€', lookup('latin1'))
    assert decode(b'\x80', lookup('latin1')) == ('€', lookup('latin1'))
    assert decode(b'\xc3\xa9', 'utf8') == ('é', lookup('utf8'))
    assert decode(b'\xc3\xa9', UTF8) == ('é', lookup('utf8'))
    assert decode(b'\xc3\xa9', 'ascii') == ('é', lookup('ascii'))
    assert decode(b'\xEF\xBB\xBF\xc3\xa9', 'ascii') == ('é', lookup('utf8'))  # UTF-8 with BOM

    assert decode(b'\xFE\xFF\x00\xe9', 'ascii') == ('é', lookup('utf-16be'))  # UTF-16-BE with BOM
    assert decode(b'\xFF\xFE\xe9\x00', 'ascii') == ('é', lookup('utf-16le'))  # UTF-16-LE with BOM
    assert decode(b'\xFE\xFF\xe9\x00', 'ascii') == ('\ue900', lookup('utf-16be'))
    assert decode(b'\xFF\xFE\x00\xe9', 'ascii') == ('\ue900', lookup('utf-16le'))

    assert decode(b'\x00\xe9', 'UTF-16BE') == ('é', lookup('utf-16be'))
    assert decode(b'\xe9\x00', 'UTF-16LE') == ('é', lookup('utf-16le'))
    assert decode(b'\xe9\x00', 'UTF-16') == ('é', lookup('utf-16le'))

    assert decode(b'\xe9\x00', 'UTF-16BE') == ('\ue900', lookup('utf-16be'))
    assert decode(b'\x00\xe9', 'UTF-16LE') == ('\ue900', lookup('utf-16le'))
    assert decode(b'\x00\xe9', 'UTF-16') == ('\ue900', lookup('utf-16le'))


def test_encode():
    assert encode('é', 'latin1') == b'\xe9'
    assert encode('é', 'utf8') == b'\xc3\xa9'
    assert encode('é', 'utf8') == b'\xc3\xa9'
    assert encode('é', 'utf-16') == b'\xe9\x00'
    assert encode('é', 'utf-16le') == b'\xe9\x00'
    assert encode('é', 'utf-16be') == b'\x00\xe9'


def test_iter_decode():
    def iter_decode_to_string(input, fallback_encoding):
        output, _encoding = iter_decode(input, fallback_encoding)
        return ''.join(output)
    assert iter_decode_to_string([], 'latin1') == ''
    assert iter_decode_to_string([b''], 'latin1') == ''
    assert iter_decode_to_string([b'\xe9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'hello'], 'latin1') == 'hello'
    assert iter_decode_to_string([b'he', b'llo'], 'latin1') == 'hello'
    assert iter_decode_to_string([b'hell', b'o'], 'latin1') == 'hello'
    assert iter_decode_to_string([b'\xc3\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'\xEF\xBB\xBF\xc3\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'\xEF\xBB\xBF', b'\xc3', b'\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'\xEF\xBB\xBF', b'a', b'\xc3'], 'latin1') == 'a\uFFFD'
    assert iter_decode_to_string([
        b'', b'\xEF', b'', b'', b'\xBB\xBF\xc3', b'\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'\xEF\xBB\xBF'], 'latin1') == ''
    assert iter_decode_to_string([b'\xEF\xBB'], 'latin1') == 'ï»'
    assert iter_decode_to_string([b'\xFE\xFF\x00\xe9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'\xFF\xFE\xe9\x00'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'', b'\xFF', b'', b'', b'\xFE\xe9', b'\x00'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'', b'h\xe9', b'llo'], 'x-user-defined') == 'h\uF7E9llo'


def test_iter_encode():
    assert b''.join(iter_encode([], 'latin1')) == b''
    assert b''.join(iter_encode([''], 'latin1')) == b''
    assert b''.join(iter_encode(['é'], 'latin1')) == b'\xe9'
    assert b''.join(iter_encode(['', 'é', '', ''], 'latin1')) == b'\xe9'
    assert b''.join(iter_encode(['', 'é', '', ''], 'utf-16')) == b'\xe9\x00'
    assert b''.join(iter_encode(['', 'é', '', ''], 'utf-16le')) == b'\xe9\x00'
    assert b''.join(iter_encode(['', 'é', '', ''], 'utf-16be')) == b'\x00\xe9'
    assert b''.join(iter_encode([
        '', 'h\uF7E9', '', 'llo'], 'x-user-defined')) == b'h\xe9llo'


def test_x_user_defined():
    encoded = b'2,\x0c\x0b\x1aO\xd9#\xcb\x0f\xc9\xbbt\xcf\xa8\xca'
    decoded = '2,\x0c\x0b\x1aO\uf7d9#\uf7cb\x0f\uf7c9\uf7bbt\uf7cf\uf7a8\uf7ca'
    encoded = b'aa'
    decoded = 'aa'
    assert decode(encoded, 'x-user-defined') == (decoded, lookup('x-user-defined'))
    assert encode(decoded, 'x-user-defined') == encoded
site-packages/pip/_vendor/webencodings/__pycache__/mklabels.cpython-38.opt-1.pyc000064400000003503151464452710023562 0ustar00U

.e@sfdZddlZzddlmZWn ek
r<ddlmZYnXddZddZedkrbe	ed	dS)
z

    webencodings.mklabels
    ~~~~~~~~~~~~~~~~~~~~~

    Regenarate the webencodings.labels module.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

N)urlopencCs|SN)stringrrE/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/mklabels.pyassert_lowersrcsfdg}ddtt|dD}tdd|D|fdd|D|dd	|S)
Na"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
cSsLg|]D}|dD]6}|dD](}tt|dt|ddfqqqS)Z	encodingslabelsuname)reprrlstrip).0categoryencodinglabelrrr
,s

zgenerate..asciicss|]\}}t|VqdSrlenr
rr
rrr	2szgenerate..c3s,|]$\}}d|dt||fVqdS)z    %s:%s %s,
 NrrZmax_lenrrr3s})	jsonloadsrreaddecodemaxextendappendjoin)Zurlpartsrrrrgenerates
r$__main__z.http://encoding.spec.whatwg.org/encodings.json)
__doc__rZurllibrImportErrorZurllib.requestrr$__name__printrrrrs!site-packages/pip/_vendor/webencodings/__pycache__/labels.cpython-38.pyc000064400000007331151464452710022276 0ustar00U

.e#@sdZddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	d	d
d
d
ddddddddddd
d
d
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd d!d!d!d!d!d"d"d"d#d#d$d$d$d$d$d$d$d%d%d%d%d%d%d%d%d%d%d&d&d'd(d(d)d*Zd+S),z

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

utf-8ibm866
iso-8859-2
iso-8859-3
iso-8859-4
iso-8859-5
iso-8859-6
iso-8859-7
iso-8859-8iso-8859-8-iiso-8859-10iso-8859-13iso-8859-14iso-8859-15iso-8859-16koi8-rkoi8-u	macintoshwindows-874windows-1250windows-1251windows-1252windows-1253windows-1254windows-1255windows-1256windows-1257windows-1258x-mac-cyrillicgbkgb18030
hz-gb-2312big5euc-jpiso-2022-jp	shift_jiseuc-kriso-2022-krutf-16beutf-16lex-user-defined)zunicode-1-1-utf-8rutf8866cp866csibm866rcsisolatin2rz
iso-ir-101z	iso8859-2Ziso88592z
iso_8859-2ziso_8859-2:1987l2latin2csisolatin3rz
iso-ir-109z	iso8859-3Ziso88593z
iso_8859-3ziso_8859-3:1988l3latin3csisolatin4rz
iso-ir-110z	iso8859-4Ziso88594z
iso_8859-4ziso_8859-4:1988l4latin4csisolatincyrilliccyrillicrz
iso-ir-144z	iso8859-5Ziso88595z
iso_8859-5ziso_8859-5:1988arabiczasmo-708Zcsiso88596eZcsiso88596icsisolatinarabiczecma-114rziso-8859-6-eziso-8859-6-iz
iso-ir-127z	iso8859-6Ziso88596z
iso_8859-6ziso_8859-6:1987csisolatingreekzecma-118elot_928greekgreek8rz
iso-ir-126z	iso8859-7Ziso88597z
iso_8859-7ziso_8859-7:1987Zsun_eu_greekZcsiso88598ecsisolatinhebrewhebrewr	ziso-8859-8-ez
iso-ir-138z	iso8859-8Ziso88598z
iso_8859-8ziso_8859-8:1988ZvisualZcsiso88598ir
Zlogicalcsisolatin6rz
iso-ir-157z
iso8859-10Z	iso885910l6latin6rz
iso8859-13Z	iso885913r
z
iso8859-14Z	iso885914Zcsisolatin9rz
iso8859-15Z	iso885915ziso_8859-15l9rcskoi8rZkoiZkoi8rkoi8_rrZcsmacintoshZmacrzx-mac-romanzdos-874ziso-8859-11z
iso8859-11Z	iso885911ztis-620rcp1250rzx-cp1250cp1251rzx-cp1251zansi_x3.4-1968asciicp1252cp819csisolatin1ibm819z
iso-8859-1z
iso-ir-100z	iso8859-1Ziso88591z
iso_8859-1ziso_8859-1:1987l1latin1zus-asciirzx-cp1252cp1253rzx-cp1253cp1254csisolatin5z
iso-8859-9z
iso-ir-148z	iso8859-9Ziso88599z
iso_8859-9ziso_8859-9:1989l5latin5rzx-cp1254cp1255rzx-cp1255cp1256rzx-cp1256cp1257rzx-cp1257cp1258rzx-cp1258rzx-mac-ukrainianchineseZcsgb2312csiso58gb231280gb2312Zgb_2312z
gb_2312-80rz	iso-ir-58zx-gbkrr r!z
big5-hkscszcn-big5csbig5zx-x-big5Zcseucpkdfmtjapaneser"zx-euc-jpcsiso2022jpr#
csshiftjisms_kanjiz	shift-jisr$sjiszwindows-31jzx-sjisZcseuckrZ
csksc56011987r%z
iso-ir-149koreanzks_c_5601-1987zks_c_5601-1989ksc5601Zksc_5601zwindows-949csiso2022krr&r'zutf-16r(r)N)__doc__ZLABELSrereC/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/labels.pyssite-packages/pip/_vendor/webencodings/__pycache__/tests.cpython-38.pyc000064400000011707151464452710022200 0ustar00U

.e@sdZddlmZddlmZmZmZmZmZm	Z	m
Z
mZmZddZ
ddZd	d
ZddZd
dZddZddZddZddZdS)z

    webencodings.tests
    ~~~~~~~~~~~~~~~~~~

    A basic test suite for Encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

)unicode_literals)	lookupLABELSdecodeencodeiter_decodeiter_encodeIncrementalDecoderIncrementalEncoderUTF8cOs6z|||Wn|k
r$YdSXtd|dS)NzDid not raise %s.AssertionError)Z	exceptionZfunctionargskwargsrB/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/tests.py
assert_raisess
rcCstdjdksttdjdks$ttdjdks6ttdjdksHttdjdksZttdjdkslttdjdks~ttddksttddksttd	jd
ksttdjd
ksttdjd
ksttd
jd
ksttddksttddkstdS)Nzutf-8zUtf-8zUTF-8utf8zutf8 z 
utf8	u8uutf-8 zUS-ASCIIzwindows-1252z
iso-8859-1latin1ZLATIN1zlatin-1uLATİN1)rnamerrrrrtest_labelssrcCstD]}td|dt|fks"ttd|dks4tdD]D}tdg||\}}t|gks`tttdg||gks8tq8t|}|ddkst|jddddkstt	|}|ddkst|jddddkstqt
tD]}t|j|kstqdS)N)rrT)final)
rrrrrrlistr	r
rsetvaluesr)Zlabelrepeatoutput_decoderencoderrrrrtest_all_labels0sr%cCsTtttddtttddtttgdtttgdtttdtttddS)NéZinvalidé)rLookupErrorrrrr	r
rrrrrtest_invalid_labelCsr)cCstdddtdfksttdtddtdfks8ttdddtdfksRttdtdtdfkslttdddtdfksttd	ddtdfksttd
ddtdfksttdddtd
fksttdddtdfksttdddtd
fks
ttdddtdfks&ttdddtd
fksBttdddtd
fks^ttdddtdfkszttdddtd
fksttdddtd
fkstdS)Nru€érr'asciiér&utf-16beutf-16lesuszUTF-16BEzUTF-16LEzUTF-16)rrrrrrrrtest_decodeLs r4cCsptdddksttdddks$ttdddks6ttdddksHttdddksZttdd	d
ksltdS)Nr'rrr+utf-16r3r1r/r2)rrrrrrtest_encodebsr7cCsdd}|gddkst|dgddks.t|dgddksBt|dgdd	ksVt|d
dgdd	kslt|dd
gdd	kst|dgddkst|dgddkst|dddgddkst|dddgddkst|ddddddgddkst|dgddkst|dgddks$t|dgddks:t|dgddksPt|ddddddgddkspt|dddgd d!kstdS)"NcSst||\}}d|S)Nr)rjoin)inputZfallback_encodingr!Z	_encodingrrriter_decode_to_stringlsz/test_iter_decode..iter_decode_to_stringrrrr5r'shelloZhellosheslloshellor+r-r&saua�ssuï»r.r0sshx-user-defineduhllor
)r:rrrtest_iter_decodeks
rCcCsdtgddkstdtdgddks2tdtdgddksLtdtddddgddksltdtddddgddkstdtddddgddkstdtddddgd	d
kstdtddddgd
dkstdS)Nrrrr'r5r6r3r1r/r2uhZllorBshllo)r8r	rrrrrtest_iter_encodes"    rDcCs@d}d}d}d}t|d|tdfks*tt|d|kss,			site-packages/pip/_vendor/webencodings/__pycache__/__init__.cpython-38.pyc000064400000022747151464452710022603 0ustar00U

.eS)@sdZddlmZddlZddlmZdZddd	d
dZiZdd
Z	ddZ
ddZGdddeZ
e
dZe
dZe
dZd+ddZddZedfddZd,dd Zd!d"Zedfd#d$Zd%d&ZGd'd(d(eZGd)d*d*eZdS)-a

    webencodings
    ~~~~~~~~~~~~

    This is a Python implementation of the `WHATWG Encoding standard
    `. See README for details.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

)unicode_literalsN)LABELSz0.5.1z
iso-8859-8zmac-cyrillicz	mac-romanZcp874)ziso-8859-8-izx-mac-cyrillic	macintoshzwindows-874cCs|ddS)a9Transform (only) ASCII letters to lower case: A-Z is mapped to a-z.

    :param string: An Unicode string.
    :returns: A new Unicode string.

    This is used for `ASCII case-insensitive
    `_
    matching of encoding labels.
    The same matching is also used, among other things,
    for `CSS keywords `_.

    This is different from the :meth:`~py:str.lower` method of Unicode strings
    which also affect non-ASCII characters,
    sometimes mapping them into the ASCII range:

        >>> keyword = u'Bac\N{KELVIN SIGN}ground'
        >>> assert keyword.lower() == u'background'
        >>> assert ascii_lower(keyword) != keyword.lower()
        >>> assert ascii_lower(keyword) == u'bac\N{KELVIN SIGN}ground'

    utf8)encodelowerdecode)stringrE/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/__init__.pyascii_lower#sr
cCsxt|d}t|}|dkr$dSt|}|dkrt|dkrLddlm}nt||}t	|}t
||}|t|<|S)u<
    Look for an encoding by its label.
    This is the spec’s `get an encoding
    `_ algorithm.
    Supported labels are listed there.

    :param label: A string.
    :returns:
        An :class:`Encoding` object, or :obj:`None` for an unknown label.

    z	

 Nzx-user-definedr)
codec_info)r
striprgetCACHEZx_user_definedrPYTHON_NAMEScodecslookupEncoding)ZlabelnameencodingrZpython_namerrrr=s




rcCs.t|dr|St|}|dkr*td||S)z
    Accept either an encoding object or label.

    :param encoding: An :class:`Encoding` object or a label string.
    :returns: An :class:`Encoding` object.
    :raises: :exc:`~exceptions.LookupError` for an unknown label.

    rNzUnknown encoding label: %r)hasattrrLookupError)Zencoding_or_labelrrrr
_get_encoding[s	
rc@s eZdZdZddZddZdS)raOReresents a character encoding such as UTF-8,
    that can be used for decoding or encoding.

    .. attribute:: name

        Canonical name of the encoding

    .. attribute:: codec_info

        The actual implementation of the encoding,
        a stdlib :class:`~codecs.CodecInfo` object.
        See :func:`codecs.register`.

    cCs||_||_dSN)rr)selfrrrrr__init__|szEncoding.__init__cCs
d|jS)Nz
)r)rrrr__repr__szEncoding.__repr__N)__name__
__module____qualname____doc__rrrrrrrmsrzutf-8zutf-16lezutf-16bereplacecCs2t|}t|\}}|p|}|j||d|fS)a
    Decode a single string.

    :param input: A byte string
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return:
        A ``(output, encoding)`` tuple of an Unicode string
        and an :obj:`Encoding`.

    r)r_detect_bomrr	)inputfallback_encodingerrorsZbom_encodingrrrrr	sr	cCsV|drt|ddfS|dr4t|ddfS|drNt|ddfSd|fS)zBReturn (bom_encoding, input), with any BOM removed from the input.sNss)
startswith_UTF16LE_UTF16BEUTF8)r%rrrr$s


r$strictcCst|j||dS)a;
    Encode a single string.

    :param input: An Unicode string.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return: A byte string.

    r)rrr)r%rr'rrrrsrcCs$t||}t||}t|}||fS)a
    "Pull"-based decoder.

    :param input:
        An iterable of byte strings.

        The input is first consumed just enough to determine the encoding
        based on the precense of a BOM,
        then consumed on demand when the return value is.
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns:
        An ``(output, encoding)`` tuple.
        :obj:`output` is an iterable of Unicode strings,
        :obj:`encoding` is the :obj:`Encoding` that is being used.

    )IncrementalDecoder_iter_decode_generatornext)r%r&r'decoder	generatorrrrriter_decodes

r4ccs|j}t|}|D]0}||}|r|jdk	s0t|jV|Vqtq|ddd}|jdk	s^t|jV|rp|VdS|D]}||}|rx|Vqx|ddd}|r|VdS)zqReturn a generator that first yields the :obj:`Encoding`,
    then yields output chukns as Unicode strings.

    NTfinal)r	iterrAssertionError)r%r2r	chunckoutputrrrr0s,r0cCst||j}t||S)uY
    “Pull”-based encoder.

    :param input: An iterable of Unicode strings.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns: An iterable of byte strings.

    )IncrementalEncoderr_iter_encode_generator)r%rr'rrrriter_encodesr>ccs6|D]}||}|r|Vq|ddd}|r2|VdS)NTr6r)r%rr:r;rrrr=sr=c@s$eZdZdZd	ddZd
ddZdS)r/uO
    “Push”-based decoder.

    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    r#cCs&t||_||_d|_d|_d|_dS)Nr5)r_fallback_encoding_errors_buffer_decoderr)rr&r'rrrrs

zIncrementalDecoder.__init__FcCs||j}|dk	r|||S|j|}t|\}}|dkrVt|dkrP|sP||_dS|j}|j|jj}||_||_	|||S)zDecode one chunk of the input.

        :param input: A byte string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: An Unicode string.

        Nr)r?)
rCrBr$lenr@rincrementaldecoderrAr	r)rr%r7r2rrrrr	's


zIncrementalDecoder.decodeN)r#)F)rr r!r"rr	rrrrr/s

r/c@seZdZdZedfddZdS)r<u
    “Push”-based encoder.

    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    .. method:: encode(input, final=False)

        :param input: An Unicode string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: A byte string.

    r.cCst|}|j|j|_dSr)rrincrementalencoderr)rrr'rrrrTszIncrementalEncoder.__init__N)rr r!r"r-rrrrrr<Csr<)r#)r#)r"Z
__future__rrlabelsrZVERSIONrrr
rrobjectrr-r+r,r	r$rr4r0r>r=r/r<rrrrs4


 
3site-packages/pip/_vendor/webencodings/__pycache__/x_user_defined.cpython-38.opt-1.pyc000064400000005107151464452710024755 0ustar00U

.e	@sdZddlmZddlZGdddejZGdddejZGdd	d	ejZGd
ddeejZGdd
d
eejZej	dej
ejeeeedZdZ
ee
ZdS)z

    webencodings.x_user_defined
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~

    An implementation of the x-user-defined encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

)unicode_literalsNc@s eZdZdddZdddZdS)	CodecstrictcCst||tSN)codecscharmap_encodeencoding_tableselfinputerrorsr
K/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/x_user_defined.pyencodeszCodec.encodecCst||tSr)rcharmap_decodedecoding_tabler	r
r
rdecodeszCodec.decodeN)r)r)__name__
__module____qualname__rrr
r
r
rrs
rc@seZdZdddZdS)IncrementalEncoderFcCst||jtdSNr)rrrrr
rfinalr
r
rrszIncrementalEncoder.encodeN)F)rrrrr
r
r
rrsrc@seZdZdddZdS)IncrementalDecoderFcCst||jtdSr)rrrrrr
r
rr$szIncrementalDecoder.decodeN)F)rrrrr
r
r
rr#src@seZdZdS)StreamWriterNrrrr
r
r
rr(src@seZdZdS)StreamReaderNrr
r
r
rr,srzx-user-defined)namerrincrementalencoderincrementaldecoderstreamreaderstreamwriteru	

 !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdefghijklmnopqrstuvwxyz{|}~)__doc__Z
__future__rrrrrrr	CodecInforrZ
codec_infor
charmap_buildrr
r
r
rs*	site-packages/pip/_vendor/webencodings/__pycache__/labels.cpython-38.opt-1.pyc000064400000007331151464452710023235 0ustar00U

.e#@sdZddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	d	d
d
d
ddddddddddd
d
d
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd d!d!d!d!d!d"d"d"d#d#d$d$d$d$d$d$d$d%d%d%d%d%d%d%d%d%d%d&d&d'd(d(d)d*Zd+S),z

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

utf-8ibm866
iso-8859-2
iso-8859-3
iso-8859-4
iso-8859-5
iso-8859-6
iso-8859-7
iso-8859-8iso-8859-8-iiso-8859-10iso-8859-13iso-8859-14iso-8859-15iso-8859-16koi8-rkoi8-u	macintoshwindows-874windows-1250windows-1251windows-1252windows-1253windows-1254windows-1255windows-1256windows-1257windows-1258x-mac-cyrillicgbkgb18030
hz-gb-2312big5euc-jpiso-2022-jp	shift_jiseuc-kriso-2022-krutf-16beutf-16lex-user-defined)zunicode-1-1-utf-8rutf8866cp866csibm866rcsisolatin2rz
iso-ir-101z	iso8859-2Ziso88592z
iso_8859-2ziso_8859-2:1987l2latin2csisolatin3rz
iso-ir-109z	iso8859-3Ziso88593z
iso_8859-3ziso_8859-3:1988l3latin3csisolatin4rz
iso-ir-110z	iso8859-4Ziso88594z
iso_8859-4ziso_8859-4:1988l4latin4csisolatincyrilliccyrillicrz
iso-ir-144z	iso8859-5Ziso88595z
iso_8859-5ziso_8859-5:1988arabiczasmo-708Zcsiso88596eZcsiso88596icsisolatinarabiczecma-114rziso-8859-6-eziso-8859-6-iz
iso-ir-127z	iso8859-6Ziso88596z
iso_8859-6ziso_8859-6:1987csisolatingreekzecma-118elot_928greekgreek8rz
iso-ir-126z	iso8859-7Ziso88597z
iso_8859-7ziso_8859-7:1987Zsun_eu_greekZcsiso88598ecsisolatinhebrewhebrewr	ziso-8859-8-ez
iso-ir-138z	iso8859-8Ziso88598z
iso_8859-8ziso_8859-8:1988ZvisualZcsiso88598ir
Zlogicalcsisolatin6rz
iso-ir-157z
iso8859-10Z	iso885910l6latin6rz
iso8859-13Z	iso885913r
z
iso8859-14Z	iso885914Zcsisolatin9rz
iso8859-15Z	iso885915ziso_8859-15l9rcskoi8rZkoiZkoi8rkoi8_rrZcsmacintoshZmacrzx-mac-romanzdos-874ziso-8859-11z
iso8859-11Z	iso885911ztis-620rcp1250rzx-cp1250cp1251rzx-cp1251zansi_x3.4-1968asciicp1252cp819csisolatin1ibm819z
iso-8859-1z
iso-ir-100z	iso8859-1Ziso88591z
iso_8859-1ziso_8859-1:1987l1latin1zus-asciirzx-cp1252cp1253rzx-cp1253cp1254csisolatin5z
iso-8859-9z
iso-ir-148z	iso8859-9Ziso88599z
iso_8859-9ziso_8859-9:1989l5latin5rzx-cp1254cp1255rzx-cp1255cp1256rzx-cp1256cp1257rzx-cp1257cp1258rzx-cp1258rzx-mac-ukrainianchineseZcsgb2312csiso58gb231280gb2312Zgb_2312z
gb_2312-80rz	iso-ir-58zx-gbkrr r!z
big5-hkscszcn-big5csbig5zx-x-big5Zcseucpkdfmtjapaneser"zx-euc-jpcsiso2022jpr#
csshiftjisms_kanjiz	shift-jisr$sjiszwindows-31jzx-sjisZcseuckrZ
csksc56011987r%z
iso-ir-149koreanzks_c_5601-1987zks_c_5601-1989ksc5601Zksc_5601zwindows-949csiso2022krr&r'zutf-16r(r)N)__doc__ZLABELSrereC/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/labels.pyssite-packages/pip/_vendor/webencodings/__pycache__/mklabels.cpython-38.pyc000064400000003551151464452710022626 0ustar00U

.e@sfdZddlZzddlmZWn ek
r<ddlmZYnXddZddZedkrbe	ed	dS)
z

    webencodings.mklabels
    ~~~~~~~~~~~~~~~~~~~~~

    Regenarate the webencodings.labels module.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

N)urlopencCs||kst|SN)lowerAssertionError)stringrE/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/mklabels.pyassert_lowersr	csfdg}ddtt|dD}tdd|D|fdd|D|dd	|S)
Na"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
cSsLg|]D}|dD]6}|dD](}tt|dt|ddfqqqS)Z	encodingslabelsuname)reprr	lstrip).0categoryencodinglabelrrr
,s

zgenerate..asciicss|]\}}t|VqdSrlenrrrrrr	2szgenerate..c3s,|]$\}}d|dt||fVqdS)z    %s:%s %s,
 NrrZmax_lenrrr3s})	jsonloadsrreaddecodemaxextendappendjoin)Zurlpartsr
rrrgenerates
r&__main__z.http://encoding.spec.whatwg.org/encodings.json)
__doc__rZurllibrImportErrorZurllib.requestr	r&__name__printrrrrs!site-packages/pip/_vendor/webencodings/__pycache__/__init__.cpython-38.opt-1.pyc000064400000022667151464452710023543 0ustar00U

.eS)@sdZddlmZddlZddlmZdZddd	d
dZiZdd
Z	ddZ
ddZGdddeZ
e
dZe
dZe
dZd+ddZddZedfddZd,dd Zd!d"Zedfd#d$Zd%d&ZGd'd(d(eZGd)d*d*eZdS)-a

    webencodings
    ~~~~~~~~~~~~

    This is a Python implementation of the `WHATWG Encoding standard
    `. See README for details.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

)unicode_literalsN)LABELSz0.5.1z
iso-8859-8zmac-cyrillicz	mac-romanZcp874)ziso-8859-8-izx-mac-cyrillic	macintoshzwindows-874cCs|ddS)a9Transform (only) ASCII letters to lower case: A-Z is mapped to a-z.

    :param string: An Unicode string.
    :returns: A new Unicode string.

    This is used for `ASCII case-insensitive
    `_
    matching of encoding labels.
    The same matching is also used, among other things,
    for `CSS keywords `_.

    This is different from the :meth:`~py:str.lower` method of Unicode strings
    which also affect non-ASCII characters,
    sometimes mapping them into the ASCII range:

        >>> keyword = u'Bac\N{KELVIN SIGN}ground'
        >>> assert keyword.lower() == u'background'
        >>> assert ascii_lower(keyword) != keyword.lower()
        >>> assert ascii_lower(keyword) == u'bac\N{KELVIN SIGN}ground'

    utf8)encodelowerdecode)stringrE/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/__init__.pyascii_lower#sr
cCsxt|d}t|}|dkr$dSt|}|dkrt|dkrLddlm}nt||}t	|}t
||}|t|<|S)u<
    Look for an encoding by its label.
    This is the spec’s `get an encoding
    `_ algorithm.
    Supported labels are listed there.

    :param label: A string.
    :returns:
        An :class:`Encoding` object, or :obj:`None` for an unknown label.

    z	

 Nzx-user-definedr)
codec_info)r
striprgetCACHEZx_user_definedrPYTHON_NAMEScodecslookupEncoding)ZlabelnameencodingrZpython_namerrrr=s




rcCs.t|dr|St|}|dkr*td||S)z
    Accept either an encoding object or label.

    :param encoding: An :class:`Encoding` object or a label string.
    :returns: An :class:`Encoding` object.
    :raises: :exc:`~exceptions.LookupError` for an unknown label.

    rNzUnknown encoding label: %r)hasattrrLookupError)Zencoding_or_labelrrrr
_get_encoding[s	
rc@s eZdZdZddZddZdS)raOReresents a character encoding such as UTF-8,
    that can be used for decoding or encoding.

    .. attribute:: name

        Canonical name of the encoding

    .. attribute:: codec_info

        The actual implementation of the encoding,
        a stdlib :class:`~codecs.CodecInfo` object.
        See :func:`codecs.register`.

    cCs||_||_dSN)rr)selfrrrrr__init__|szEncoding.__init__cCs
d|jS)Nz
)r)rrrr__repr__szEncoding.__repr__N)__name__
__module____qualname____doc__rrrrrrrmsrzutf-8zutf-16lezutf-16bereplacecCs2t|}t|\}}|p|}|j||d|fS)a
    Decode a single string.

    :param input: A byte string
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return:
        A ``(output, encoding)`` tuple of an Unicode string
        and an :obj:`Encoding`.

    r)r_detect_bomrr	)inputfallback_encodingerrorsZbom_encodingrrrrr	sr	cCsV|drt|ddfS|dr4t|ddfS|drNt|ddfSd|fS)zBReturn (bom_encoding, input), with any BOM removed from the input.sNss)
startswith_UTF16LE_UTF16BEUTF8)r%rrrr$s


r$strictcCst|j||dS)a;
    Encode a single string.

    :param input: An Unicode string.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return: A byte string.

    r)rrr)r%rr'rrrrsrcCs$t||}t||}t|}||fS)a
    "Pull"-based decoder.

    :param input:
        An iterable of byte strings.

        The input is first consumed just enough to determine the encoding
        based on the precense of a BOM,
        then consumed on demand when the return value is.
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns:
        An ``(output, encoding)`` tuple.
        :obj:`output` is an iterable of Unicode strings,
        :obj:`encoding` is the :obj:`Encoding` that is being used.

    )IncrementalDecoder_iter_decode_generatornext)r%r&r'decoder	generatorrrrriter_decodes

r4ccs|j}t|}|D]"}||}|r|jV|VqXq|ddd}|jV|rT|VdS|D]}||}|r\|Vq\|ddd}|r|VdS)zqReturn a generator that first yields the :obj:`Encoding`,
    then yields output chukns as Unicode strings.

    TfinalN)r	iterr)r%r2r	chunckoutputrrrr0s(r0cCst||j}t||S)uY
    “Pull”-based encoder.

    :param input: An iterable of Unicode strings.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns: An iterable of byte strings.

    )IncrementalEncoderr_iter_encode_generator)r%rr'rrrriter_encodesr=ccs6|D]}||}|r|Vq|ddd}|r2|VdS)NTr6r)r%rr9r:rrrr<sr<c@s$eZdZdZd	ddZd
ddZdS)r/uO
    “Push”-based decoder.

    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    r#cCs&t||_||_d|_d|_d|_dS)Nr5)r_fallback_encoding_errors_buffer_decoderr)rr&r'rrrrs

zIncrementalDecoder.__init__FcCs||j}|dk	r|||S|j|}t|\}}|dkrVt|dkrP|sP||_dS|j}|j|jj}||_||_	|||S)zDecode one chunk of the input.

        :param input: A byte string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: An Unicode string.

        Nr)r>)
rBrAr$lenr?rincrementaldecoderr@r	r)rr%r7r2rrrrr	's


zIncrementalDecoder.decodeN)r#)F)rr r!r"rr	rrrrr/s

r/c@seZdZdZedfddZdS)r;u
    “Push”-based encoder.

    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    .. method:: encode(input, final=False)

        :param input: An Unicode string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: A byte string.

    r.cCst|}|j|j|_dSr)rrincrementalencoderr)rrr'rrrrTszIncrementalEncoder.__init__N)rr r!r"r-rrrrrr;Csr;)r#)r#)r"Z
__future__rrlabelsrZVERSIONrrr
rrobjectrr-r+r,r	r$rr4r0r=r<r/r;rrrrs4


 
3site-packages/pip/_vendor/webencodings/__pycache__/tests.cpython-38.opt-1.pyc000064400000004547151464452710023143 0ustar00U

.e@sdZddlmZddlmZmZmZmZmZm	Z	m
Z
mZmZddZ
ddZd	d
ZddZd
dZddZddZddZddZdS)z

    webencodings.tests
    ~~~~~~~~~~~~~~~~~~

    A basic test suite for Encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

)unicode_literals)	lookupLABELSdecodeencodeiter_decodeiter_encodeIncrementalDecoderIncrementalEncoderUTF8cOs6z|||Wn|k
r$YdSXtd|dS)NzDid not raise %s.)AssertionError)Z	exceptionZfunctionargskwargsrB/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/tests.py
assert_raisess
rcCsdSNrrrrrtest_labelssrcCsNtD]2}dD]}tdg||\}}qt|}t|}qttD]}qDdS)N)rr)rrr
rsetvalues)Zlabelrepeatoutput_decoderencodernamerrrtest_all_labels0srcCsTtttddtttddtttgdtttgdtttdtttddS)NséZinvalidé)rLookupErrorrrrr	r
rrrrrtest_invalid_labelCsr"cCsdSrrrrrrtest_decodeLsr#cCsdSrrrrrrtest_encodebsr$cCsdd}dS)NcSst||\}}d|S)N)rjoin)inputZfallback_encodingrZ	_encodingrrriter_decode_to_stringlsz/test_iter_decode..iter_decode_to_stringr)r(rrrtest_iter_decodeksr)cCsdSrrrrrrtest_iter_encodesr*cCsd}d}d}d}dS)Ns2,O#ɻtϨu2,O#tsaaZaar)ZencodedZdecodedrrrtest_x_user_defineds
r+N)__doc__Z
__future__rr%rrrrrr	r
rrrrrr"r#r$r)r*r+rrrrs,			site-packages/pip/_vendor/webencodings/__pycache__/x_user_defined.cpython-38.pyc000064400000005107151464452710024016 0ustar00U

.e	@sdZddlmZddlZGdddejZGdddejZGdd	d	ejZGd
ddeejZGdd
d
eejZej	dej
ejeeeedZdZ
ee
ZdS)z

    webencodings.x_user_defined
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~

    An implementation of the x-user-defined encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

)unicode_literalsNc@s eZdZdddZdddZdS)	CodecstrictcCst||tSN)codecscharmap_encodeencoding_tableselfinputerrorsr
K/usr/lib/python3.8/site-packages/pip/_vendor/webencodings/x_user_defined.pyencodeszCodec.encodecCst||tSr)rcharmap_decodedecoding_tabler	r
r
rdecodeszCodec.decodeN)r)r)__name__
__module____qualname__rrr
r
r
rrs
rc@seZdZdddZdS)IncrementalEncoderFcCst||jtdSNr)rrrrr
rfinalr
r
rrszIncrementalEncoder.encodeN)F)rrrrr
r
r
rrsrc@seZdZdddZdS)IncrementalDecoderFcCst||jtdSr)rrrrrr
r
rr$szIncrementalDecoder.decodeN)F)rrrrr
r
r
rr#src@seZdZdS)StreamWriterNrrrr
r
r
rr(src@seZdZdS)StreamReaderNrr
r
r
rr,srzx-user-defined)namerrincrementalencoderincrementaldecoderstreamreaderstreamwriteru	

 !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdefghijklmnopqrstuvwxyz{|}~)__doc__Z
__future__rrrrrrr	CodecInforrZ
codec_infor
charmap_buildrr
r
r
rs*	site-packages/pip/_vendor/webencodings/x_user_defined.py000064400000010323151464452710017524 0ustar00# coding: utf-8
"""

    webencodings.x_user_defined
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~

    An implementation of the x-user-defined encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

from __future__ import unicode_literals

import codecs


### Codec APIs

class Codec(codecs.Codec):

    def encode(self, input, errors='strict'):
        return codecs.charmap_encode(input, errors, encoding_table)

    def decode(self, input, errors='strict'):
        return codecs.charmap_decode(input, errors, decoding_table)


class IncrementalEncoder(codecs.IncrementalEncoder):
    def encode(self, input, final=False):
        return codecs.charmap_encode(input, self.errors, encoding_table)[0]


class IncrementalDecoder(codecs.IncrementalDecoder):
    def decode(self, input, final=False):
        return codecs.charmap_decode(input, self.errors, decoding_table)[0]


class StreamWriter(Codec, codecs.StreamWriter):
    pass


class StreamReader(Codec, codecs.StreamReader):
    pass


### encodings module API

codec_info = codecs.CodecInfo(
    name='x-user-defined',
    encode=Codec().encode,
    decode=Codec().decode,
    incrementalencoder=IncrementalEncoder,
    incrementaldecoder=IncrementalDecoder,
    streamreader=StreamReader,
    streamwriter=StreamWriter,
)


### Decoding Table

# Python 3:
# for c in range(256): print('    %r' % chr(c if c < 128 else c + 0xF700))
decoding_table = (
    '\x00'
    '\x01'
    '\x02'
    '\x03'
    '\x04'
    '\x05'
    '\x06'
    '\x07'
    '\x08'
    '\t'
    '\n'
    '\x0b'
    '\x0c'
    '\r'
    '\x0e'
    '\x0f'
    '\x10'
    '\x11'
    '\x12'
    '\x13'
    '\x14'
    '\x15'
    '\x16'
    '\x17'
    '\x18'
    '\x19'
    '\x1a'
    '\x1b'
    '\x1c'
    '\x1d'
    '\x1e'
    '\x1f'
    ' '
    '!'
    '"'
    '#'
    '$'
    '%'
    '&'
    "'"
    '('
    ')'
    '*'
    '+'
    ','
    '-'
    '.'
    '/'
    '0'
    '1'
    '2'
    '3'
    '4'
    '5'
    '6'
    '7'
    '8'
    '9'
    ':'
    ';'
    '<'
    '='
    '>'
    '?'
    '@'
    'A'
    'B'
    'C'
    'D'
    'E'
    'F'
    'G'
    'H'
    'I'
    'J'
    'K'
    'L'
    'M'
    'N'
    'O'
    'P'
    'Q'
    'R'
    'S'
    'T'
    'U'
    'V'
    'W'
    'X'
    'Y'
    'Z'
    '['
    '\\'
    ']'
    '^'
    '_'
    '`'
    'a'
    'b'
    'c'
    'd'
    'e'
    'f'
    'g'
    'h'
    'i'
    'j'
    'k'
    'l'
    'm'
    'n'
    'o'
    'p'
    'q'
    'r'
    's'
    't'
    'u'
    'v'
    'w'
    'x'
    'y'
    'z'
    '{'
    '|'
    '}'
    '~'
    '\x7f'
    '\uf780'
    '\uf781'
    '\uf782'
    '\uf783'
    '\uf784'
    '\uf785'
    '\uf786'
    '\uf787'
    '\uf788'
    '\uf789'
    '\uf78a'
    '\uf78b'
    '\uf78c'
    '\uf78d'
    '\uf78e'
    '\uf78f'
    '\uf790'
    '\uf791'
    '\uf792'
    '\uf793'
    '\uf794'
    '\uf795'
    '\uf796'
    '\uf797'
    '\uf798'
    '\uf799'
    '\uf79a'
    '\uf79b'
    '\uf79c'
    '\uf79d'
    '\uf79e'
    '\uf79f'
    '\uf7a0'
    '\uf7a1'
    '\uf7a2'
    '\uf7a3'
    '\uf7a4'
    '\uf7a5'
    '\uf7a6'
    '\uf7a7'
    '\uf7a8'
    '\uf7a9'
    '\uf7aa'
    '\uf7ab'
    '\uf7ac'
    '\uf7ad'
    '\uf7ae'
    '\uf7af'
    '\uf7b0'
    '\uf7b1'
    '\uf7b2'
    '\uf7b3'
    '\uf7b4'
    '\uf7b5'
    '\uf7b6'
    '\uf7b7'
    '\uf7b8'
    '\uf7b9'
    '\uf7ba'
    '\uf7bb'
    '\uf7bc'
    '\uf7bd'
    '\uf7be'
    '\uf7bf'
    '\uf7c0'
    '\uf7c1'
    '\uf7c2'
    '\uf7c3'
    '\uf7c4'
    '\uf7c5'
    '\uf7c6'
    '\uf7c7'
    '\uf7c8'
    '\uf7c9'
    '\uf7ca'
    '\uf7cb'
    '\uf7cc'
    '\uf7cd'
    '\uf7ce'
    '\uf7cf'
    '\uf7d0'
    '\uf7d1'
    '\uf7d2'
    '\uf7d3'
    '\uf7d4'
    '\uf7d5'
    '\uf7d6'
    '\uf7d7'
    '\uf7d8'
    '\uf7d9'
    '\uf7da'
    '\uf7db'
    '\uf7dc'
    '\uf7dd'
    '\uf7de'
    '\uf7df'
    '\uf7e0'
    '\uf7e1'
    '\uf7e2'
    '\uf7e3'
    '\uf7e4'
    '\uf7e5'
    '\uf7e6'
    '\uf7e7'
    '\uf7e8'
    '\uf7e9'
    '\uf7ea'
    '\uf7eb'
    '\uf7ec'
    '\uf7ed'
    '\uf7ee'
    '\uf7ef'
    '\uf7f0'
    '\uf7f1'
    '\uf7f2'
    '\uf7f3'
    '\uf7f4'
    '\uf7f5'
    '\uf7f6'
    '\uf7f7'
    '\uf7f8'
    '\uf7f9'
    '\uf7fa'
    '\uf7fb'
    '\uf7fc'
    '\uf7fd'
    '\uf7fe'
    '\uf7ff'
)

### Encoding table
encoding_table = codecs.charmap_build(decoding_table)
site-packages/pip/_vendor/webencodings/__init__.py000064400000024523151464452710016307 0ustar00# coding: utf-8
"""

    webencodings
    ~~~~~~~~~~~~

    This is a Python implementation of the `WHATWG Encoding standard
    `. See README for details.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

from __future__ import unicode_literals

import codecs

from .labels import LABELS


VERSION = '0.5.1'


# Some names in Encoding are not valid Python aliases. Remap these.
PYTHON_NAMES = {
    'iso-8859-8-i': 'iso-8859-8',
    'x-mac-cyrillic': 'mac-cyrillic',
    'macintosh': 'mac-roman',
    'windows-874': 'cp874'}

CACHE = {}


def ascii_lower(string):
    r"""Transform (only) ASCII letters to lower case: A-Z is mapped to a-z.

    :param string: An Unicode string.
    :returns: A new Unicode string.

    This is used for `ASCII case-insensitive
    `_
    matching of encoding labels.
    The same matching is also used, among other things,
    for `CSS keywords `_.

    This is different from the :meth:`~py:str.lower` method of Unicode strings
    which also affect non-ASCII characters,
    sometimes mapping them into the ASCII range:

        >>> keyword = u'Bac\N{KELVIN SIGN}ground'
        >>> assert keyword.lower() == u'background'
        >>> assert ascii_lower(keyword) != keyword.lower()
        >>> assert ascii_lower(keyword) == u'bac\N{KELVIN SIGN}ground'

    """
    # This turns out to be faster than unicode.translate()
    return string.encode('utf8').lower().decode('utf8')


def lookup(label):
    """
    Look for an encoding by its label.
    This is the spec’s `get an encoding
    `_ algorithm.
    Supported labels are listed there.

    :param label: A string.
    :returns:
        An :class:`Encoding` object, or :obj:`None` for an unknown label.

    """
    # Only strip ASCII whitespace: U+0009, U+000A, U+000C, U+000D, and U+0020.
    label = ascii_lower(label.strip('\t\n\f\r '))
    name = LABELS.get(label)
    if name is None:
        return None
    encoding = CACHE.get(name)
    if encoding is None:
        if name == 'x-user-defined':
            from .x_user_defined import codec_info
        else:
            python_name = PYTHON_NAMES.get(name, name)
            # Any python_name value that gets to here should be valid.
            codec_info = codecs.lookup(python_name)
        encoding = Encoding(name, codec_info)
        CACHE[name] = encoding
    return encoding


def _get_encoding(encoding_or_label):
    """
    Accept either an encoding object or label.

    :param encoding: An :class:`Encoding` object or a label string.
    :returns: An :class:`Encoding` object.
    :raises: :exc:`~exceptions.LookupError` for an unknown label.

    """
    if hasattr(encoding_or_label, 'codec_info'):
        return encoding_or_label

    encoding = lookup(encoding_or_label)
    if encoding is None:
        raise LookupError('Unknown encoding label: %r' % encoding_or_label)
    return encoding


class Encoding(object):
    """Reresents a character encoding such as UTF-8,
    that can be used for decoding or encoding.

    .. attribute:: name

        Canonical name of the encoding

    .. attribute:: codec_info

        The actual implementation of the encoding,
        a stdlib :class:`~codecs.CodecInfo` object.
        See :func:`codecs.register`.

    """
    def __init__(self, name, codec_info):
        self.name = name
        self.codec_info = codec_info

    def __repr__(self):
        return '' % self.name


#: The UTF-8 encoding. Should be used for new content and formats.
UTF8 = lookup('utf-8')

_UTF16LE = lookup('utf-16le')
_UTF16BE = lookup('utf-16be')


def decode(input, fallback_encoding, errors='replace'):
    """
    Decode a single string.

    :param input: A byte string
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return:
        A ``(output, encoding)`` tuple of an Unicode string
        and an :obj:`Encoding`.

    """
    # Fail early if `encoding` is an invalid label.
    fallback_encoding = _get_encoding(fallback_encoding)
    bom_encoding, input = _detect_bom(input)
    encoding = bom_encoding or fallback_encoding
    return encoding.codec_info.decode(input, errors)[0], encoding


def _detect_bom(input):
    """Return (bom_encoding, input), with any BOM removed from the input."""
    if input.startswith(b'\xFF\xFE'):
        return _UTF16LE, input[2:]
    if input.startswith(b'\xFE\xFF'):
        return _UTF16BE, input[2:]
    if input.startswith(b'\xEF\xBB\xBF'):
        return UTF8, input[3:]
    return None, input


def encode(input, encoding=UTF8, errors='strict'):
    """
    Encode a single string.

    :param input: An Unicode string.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return: A byte string.

    """
    return _get_encoding(encoding).codec_info.encode(input, errors)[0]


def iter_decode(input, fallback_encoding, errors='replace'):
    """
    "Pull"-based decoder.

    :param input:
        An iterable of byte strings.

        The input is first consumed just enough to determine the encoding
        based on the precense of a BOM,
        then consumed on demand when the return value is.
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns:
        An ``(output, encoding)`` tuple.
        :obj:`output` is an iterable of Unicode strings,
        :obj:`encoding` is the :obj:`Encoding` that is being used.

    """

    decoder = IncrementalDecoder(fallback_encoding, errors)
    generator = _iter_decode_generator(input, decoder)
    encoding = next(generator)
    return generator, encoding


def _iter_decode_generator(input, decoder):
    """Return a generator that first yields the :obj:`Encoding`,
    then yields output chukns as Unicode strings.

    """
    decode = decoder.decode
    input = iter(input)
    for chunck in input:
        output = decode(chunck)
        if output:
            assert decoder.encoding is not None
            yield decoder.encoding
            yield output
            break
    else:
        # Input exhausted without determining the encoding
        output = decode(b'', final=True)
        assert decoder.encoding is not None
        yield decoder.encoding
        if output:
            yield output
        return

    for chunck in input:
        output = decode(chunck)
        if output:
            yield output
    output = decode(b'', final=True)
    if output:
        yield output


def iter_encode(input, encoding=UTF8, errors='strict'):
    """
    “Pull”-based encoder.

    :param input: An iterable of Unicode strings.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns: An iterable of byte strings.

    """
    # Fail early if `encoding` is an invalid label.
    encode = IncrementalEncoder(encoding, errors).encode
    return _iter_encode_generator(input, encode)


def _iter_encode_generator(input, encode):
    for chunck in input:
        output = encode(chunck)
        if output:
            yield output
    output = encode('', final=True)
    if output:
        yield output


class IncrementalDecoder(object):
    """
    “Push”-based decoder.

    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    """
    def __init__(self, fallback_encoding, errors='replace'):
        # Fail early if `encoding` is an invalid label.
        self._fallback_encoding = _get_encoding(fallback_encoding)
        self._errors = errors
        self._buffer = b''
        self._decoder = None
        #: The actual :class:`Encoding` that is being used,
        #: or :obj:`None` if that is not determined yet.
        #: (Ie. if there is not enough input yet to determine
        #: if there is a BOM.)
        self.encoding = None  # Not known yet.

    def decode(self, input, final=False):
        """Decode one chunk of the input.

        :param input: A byte string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: An Unicode string.

        """
        decoder = self._decoder
        if decoder is not None:
            return decoder(input, final)

        input = self._buffer + input
        encoding, input = _detect_bom(input)
        if encoding is None:
            if len(input) < 3 and not final:  # Not enough data yet.
                self._buffer = input
                return ''
            else:  # No BOM
                encoding = self._fallback_encoding
        decoder = encoding.codec_info.incrementaldecoder(self._errors).decode
        self._decoder = decoder
        self.encoding = encoding
        return decoder(input, final)


class IncrementalEncoder(object):
    """
    “Push”-based encoder.

    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    .. method:: encode(input, final=False)

        :param input: An Unicode string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: A byte string.

    """
    def __init__(self, encoding=UTF8, errors='strict'):
        encoding = _get_encoding(encoding)
        self.encode = encoding.codec_info.incrementalencoder(errors).encode
site-packages/pip/_vendor/urllib3/response.py000064400000066274151464452710015324 0ustar00from __future__ import absolute_import
from contextlib import contextmanager
import zlib
import io
import logging
from socket import timeout as SocketTimeout
from socket import error as SocketError

try:
    import brotli
except ImportError:
    brotli = None

from ._collections import HTTPHeaderDict
from .exceptions import (
    BodyNotHttplibCompatible,
    ProtocolError,
    DecodeError,
    ReadTimeoutError,
    ResponseNotChunked,
    IncompleteRead,
    InvalidHeader,
)
from .packages.six import string_types as basestring, PY3
from .packages.six.moves import http_client as httplib
from .connection import HTTPException, BaseSSLError
from .util.response import is_fp_closed, is_response_to_head

log = logging.getLogger(__name__)


class DeflateDecoder(object):
    def __init__(self):
        self._first_try = True
        self._data = b""
        self._obj = zlib.decompressobj()

    def __getattr__(self, name):
        return getattr(self._obj, name)

    def decompress(self, data):
        if not data:
            return data

        if not self._first_try:
            return self._obj.decompress(data)

        self._data += data
        try:
            decompressed = self._obj.decompress(data)
            if decompressed:
                self._first_try = False
                self._data = None
            return decompressed
        except zlib.error:
            self._first_try = False
            self._obj = zlib.decompressobj(-zlib.MAX_WBITS)
            try:
                return self.decompress(self._data)
            finally:
                self._data = None


class GzipDecoderState(object):

    FIRST_MEMBER = 0
    OTHER_MEMBERS = 1
    SWALLOW_DATA = 2


class GzipDecoder(object):
    def __init__(self):
        self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)
        self._state = GzipDecoderState.FIRST_MEMBER

    def __getattr__(self, name):
        return getattr(self._obj, name)

    def decompress(self, data):
        ret = bytearray()
        if self._state == GzipDecoderState.SWALLOW_DATA or not data:
            return bytes(ret)
        while True:
            try:
                ret += self._obj.decompress(data)
            except zlib.error:
                previous_state = self._state
                # Ignore data after the first error
                self._state = GzipDecoderState.SWALLOW_DATA
                if previous_state == GzipDecoderState.OTHER_MEMBERS:
                    # Allow trailing garbage acceptable in other gzip clients
                    return bytes(ret)
                raise
            data = self._obj.unused_data
            if not data:
                return bytes(ret)
            self._state = GzipDecoderState.OTHER_MEMBERS
            self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)


if brotli is not None:

    class BrotliDecoder(object):
        # Supports both 'brotlipy' and 'Brotli' packages
        # since they share an import name. The top branches
        # are for 'brotlipy' and bottom branches for 'Brotli'
        def __init__(self):
            self._obj = brotli.Decompressor()

        def decompress(self, data):
            if hasattr(self._obj, "decompress"):
                return self._obj.decompress(data)
            return self._obj.process(data)

        def flush(self):
            if hasattr(self._obj, "flush"):
                return self._obj.flush()
            return b""


class MultiDecoder(object):
    """
    From RFC7231:
        If one or more encodings have been applied to a representation, the
        sender that applied the encodings MUST generate a Content-Encoding
        header field that lists the content codings in the order in which
        they were applied.
    """

    def __init__(self, modes):
        self._decoders = [_get_decoder(m.strip()) for m in modes.split(",")]

    def flush(self):
        return self._decoders[0].flush()

    def decompress(self, data):
        for d in reversed(self._decoders):
            data = d.decompress(data)
        return data


def _get_decoder(mode):
    if "," in mode:
        return MultiDecoder(mode)

    if mode == "gzip":
        return GzipDecoder()

    if brotli is not None and mode == "br":
        return BrotliDecoder()

    return DeflateDecoder()


class HTTPResponse(io.IOBase):
    """
    HTTP Response container.

    Backwards-compatible to httplib's HTTPResponse but the response ``body`` is
    loaded and decoded on-demand when the ``data`` property is accessed.  This
    class is also compatible with the Python standard library's :mod:`io`
    module, and can hence be treated as a readable object in the context of that
    framework.

    Extra parameters for behaviour not present in httplib.HTTPResponse:

    :param preload_content:
        If True, the response's body will be preloaded during construction.

    :param decode_content:
        If True, will attempt to decode the body based on the
        'content-encoding' header.

    :param original_response:
        When this HTTPResponse wrapper is generated from an httplib.HTTPResponse
        object, it's convenient to include the original for debug purposes. It's
        otherwise unused.

    :param retries:
        The retries contains the last :class:`~urllib3.util.retry.Retry` that
        was used during the request.

    :param enforce_content_length:
        Enforce content length checking. Body returned by server must match
        value of Content-Length header, if present. Otherwise, raise error.
    """

    CONTENT_DECODERS = ["gzip", "deflate"]
    if brotli is not None:
        CONTENT_DECODERS += ["br"]
    REDIRECT_STATUSES = [301, 302, 303, 307, 308]

    def __init__(
        self,
        body="",
        headers=None,
        status=0,
        version=0,
        reason=None,
        strict=0,
        preload_content=True,
        decode_content=True,
        original_response=None,
        pool=None,
        connection=None,
        msg=None,
        retries=None,
        enforce_content_length=False,
        request_method=None,
        request_url=None,
        auto_close=True,
    ):

        if isinstance(headers, HTTPHeaderDict):
            self.headers = headers
        else:
            self.headers = HTTPHeaderDict(headers)
        self.status = status
        self.version = version
        self.reason = reason
        self.strict = strict
        self.decode_content = decode_content
        self.retries = retries
        self.enforce_content_length = enforce_content_length
        self.auto_close = auto_close

        self._decoder = None
        self._body = None
        self._fp = None
        self._original_response = original_response
        self._fp_bytes_read = 0
        self.msg = msg
        self._request_url = request_url

        if body and isinstance(body, (basestring, bytes)):
            self._body = body

        self._pool = pool
        self._connection = connection

        if hasattr(body, "read"):
            self._fp = body

        # Are we using the chunked-style of transfer encoding?
        self.chunked = False
        self.chunk_left = None
        tr_enc = self.headers.get("transfer-encoding", "").lower()
        # Don't incur the penalty of creating a list and then discarding it
        encodings = (enc.strip() for enc in tr_enc.split(","))
        if "chunked" in encodings:
            self.chunked = True

        # Determine length of response
        self.length_remaining = self._init_length(request_method)

        # If requested, preload the body.
        if preload_content and not self._body:
            self._body = self.read(decode_content=decode_content)

    def get_redirect_location(self):
        """
        Should we redirect and where to?

        :returns: Truthy redirect location string if we got a redirect status
            code and valid location. ``None`` if redirect status and no
            location. ``False`` if not a redirect status code.
        """
        if self.status in self.REDIRECT_STATUSES:
            return self.headers.get("location")

        return False

    def release_conn(self):
        if not self._pool or not self._connection:
            return

        self._pool._put_conn(self._connection)
        self._connection = None

    @property
    def data(self):
        # For backwords-compat with earlier urllib3 0.4 and earlier.
        if self._body:
            return self._body

        if self._fp:
            return self.read(cache_content=True)

    @property
    def connection(self):
        return self._connection

    def isclosed(self):
        return is_fp_closed(self._fp)

    def tell(self):
        """
        Obtain the number of bytes pulled over the wire so far. May differ from
        the amount of content returned by :meth:``HTTPResponse.read`` if bytes
        are encoded on the wire (e.g, compressed).
        """
        return self._fp_bytes_read

    def _init_length(self, request_method):
        """
        Set initial length value for Response content if available.
        """
        length = self.headers.get("content-length")

        if length is not None:
            if self.chunked:
                # This Response will fail with an IncompleteRead if it can't be
                # received as chunked. This method falls back to attempt reading
                # the response before raising an exception.
                log.warning(
                    "Received response with both Content-Length and "
                    "Transfer-Encoding set. This is expressly forbidden "
                    "by RFC 7230 sec 3.3.2. Ignoring Content-Length and "
                    "attempting to process response as Transfer-Encoding: "
                    "chunked."
                )
                return None

            try:
                # RFC 7230 section 3.3.2 specifies multiple content lengths can
                # be sent in a single Content-Length header
                # (e.g. Content-Length: 42, 42). This line ensures the values
                # are all valid ints and that as long as the `set` length is 1,
                # all values are the same. Otherwise, the header is invalid.
                lengths = set([int(val) for val in length.split(",")])
                if len(lengths) > 1:
                    raise InvalidHeader(
                        "Content-Length contained multiple "
                        "unmatching values (%s)" % length
                    )
                length = lengths.pop()
            except ValueError:
                length = None
            else:
                if length < 0:
                    length = None

        # Convert status to int for comparison
        # In some cases, httplib returns a status of "_UNKNOWN"
        try:
            status = int(self.status)
        except ValueError:
            status = 0

        # Check for responses that shouldn't include a body
        if status in (204, 304) or 100 <= status < 200 or request_method == "HEAD":
            length = 0

        return length

    def _init_decoder(self):
        """
        Set-up the _decoder attribute if necessary.
        """
        # Note: content-encoding value should be case-insensitive, per RFC 7230
        # Section 3.2
        content_encoding = self.headers.get("content-encoding", "").lower()
        if self._decoder is None:
            if content_encoding in self.CONTENT_DECODERS:
                self._decoder = _get_decoder(content_encoding)
            elif "," in content_encoding:
                encodings = [
                    e.strip()
                    for e in content_encoding.split(",")
                    if e.strip() in self.CONTENT_DECODERS
                ]
                if len(encodings):
                    self._decoder = _get_decoder(content_encoding)

    DECODER_ERROR_CLASSES = (IOError, zlib.error)
    if brotli is not None:
        DECODER_ERROR_CLASSES += (brotli.error,)

    def _decode(self, data, decode_content, flush_decoder):
        """
        Decode the data passed in and potentially flush the decoder.
        """
        if not decode_content:
            return data

        try:
            if self._decoder:
                data = self._decoder.decompress(data)
        except self.DECODER_ERROR_CLASSES as e:
            content_encoding = self.headers.get("content-encoding", "").lower()
            raise DecodeError(
                "Received response with content-encoding: %s, but "
                "failed to decode it." % content_encoding,
                e,
            )
        if flush_decoder:
            data += self._flush_decoder()

        return data

    def _flush_decoder(self):
        """
        Flushes the decoder. Should only be called if the decoder is actually
        being used.
        """
        if self._decoder:
            buf = self._decoder.decompress(b"")
            return buf + self._decoder.flush()

        return b""

    @contextmanager
    def _error_catcher(self):
        """
        Catch low-level python exceptions, instead re-raising urllib3
        variants, so that low-level exceptions are not leaked in the
        high-level api.

        On exit, release the connection back to the pool.
        """
        clean_exit = False

        try:
            try:
                yield

            except SocketTimeout:
                # FIXME: Ideally we'd like to include the url in the ReadTimeoutError but
                # there is yet no clean way to get at it from this context.
                raise ReadTimeoutError(self._pool, None, "Read timed out.")

            except BaseSSLError as e:
                # FIXME: Is there a better way to differentiate between SSLErrors?
                if "read operation timed out" not in str(e):  # Defensive:
                    # This shouldn't happen but just in case we're missing an edge
                    # case, let's avoid swallowing SSL errors.
                    raise

                raise ReadTimeoutError(self._pool, None, "Read timed out.")

            except (HTTPException, SocketError) as e:
                # This includes IncompleteRead.
                raise ProtocolError("Connection broken: %r" % e, e)

            # If no exception is thrown, we should avoid cleaning up
            # unnecessarily.
            clean_exit = True
        finally:
            # If we didn't terminate cleanly, we need to throw away our
            # connection.
            if not clean_exit:
                # The response may not be closed but we're not going to use it
                # anymore so close it now to ensure that the connection is
                # released back to the pool.
                if self._original_response:
                    self._original_response.close()

                # Closing the response may not actually be sufficient to close
                # everything, so if we have a hold of the connection close that
                # too.
                if self._connection:
                    self._connection.close()

            # If we hold the original response but it's closed now, we should
            # return the connection back to the pool.
            if self._original_response and self._original_response.isclosed():
                self.release_conn()

    def read(self, amt=None, decode_content=None, cache_content=False):
        """
        Similar to :meth:`httplib.HTTPResponse.read`, but with two additional
        parameters: ``decode_content`` and ``cache_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.

        :param cache_content:
            If True, will save the returned data such that the same result is
            returned despite of the state of the underlying file object. This
            is useful if you want the ``.data`` property to continue working
            after having ``.read()`` the file object. (Overridden if ``amt`` is
            set.)
        """
        self._init_decoder()
        if decode_content is None:
            decode_content = self.decode_content

        if self._fp is None:
            return

        flush_decoder = False
        fp_closed = getattr(self._fp, "closed", False)

        with self._error_catcher():
            if amt is None:
                # cStringIO doesn't like amt=None
                data = self._fp.read() if not fp_closed else b""
                flush_decoder = True
            else:
                cache_content = False
                data = self._fp.read(amt) if not fp_closed else b""
                if (
                    amt != 0 and not data
                ):  # Platform-specific: Buggy versions of Python.
                    # Close the connection when no data is returned
                    #
                    # This is redundant to what httplib/http.client _should_
                    # already do.  However, versions of python released before
                    # December 15, 2012 (http://bugs.python.org/issue16298) do
                    # not properly close the connection in all cases. There is
                    # no harm in redundantly calling close.
                    self._fp.close()
                    flush_decoder = True
                    if self.enforce_content_length and self.length_remaining not in (
                        0,
                        None,
                    ):
                        # This is an edge case that httplib failed to cover due
                        # to concerns of backward compatibility. We're
                        # addressing it here to make sure IncompleteRead is
                        # raised during streaming, so all calls with incorrect
                        # Content-Length are caught.
                        raise IncompleteRead(self._fp_bytes_read, self.length_remaining)

        if data:
            self._fp_bytes_read += len(data)
            if self.length_remaining is not None:
                self.length_remaining -= len(data)

            data = self._decode(data, decode_content, flush_decoder)

            if cache_content:
                self._body = data

        return data

    def stream(self, amt=2 ** 16, decode_content=None):
        """
        A generator wrapper for the read() method. A call will block until
        ``amt`` bytes have been read from the connection or until the
        connection is closed.

        :param amt:
            How much of the content to read. The generator will return up to
            much data per iteration, but may return less. This is particularly
            likely when using compressed data. However, the empty string will
            never be returned.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        """
        if self.chunked and self.supports_chunked_reads():
            for line in self.read_chunked(amt, decode_content=decode_content):
                yield line
        else:
            while not is_fp_closed(self._fp):
                data = self.read(amt=amt, decode_content=decode_content)

                if data:
                    yield data

    @classmethod
    def from_httplib(ResponseCls, r, **response_kw):
        """
        Given an :class:`httplib.HTTPResponse` instance ``r``, return a
        corresponding :class:`urllib3.response.HTTPResponse` object.

        Remaining parameters are passed to the HTTPResponse constructor, along
        with ``original_response=r``.
        """
        headers = r.msg

        if not isinstance(headers, HTTPHeaderDict):
            if PY3:
                headers = HTTPHeaderDict(headers.items())
            else:
                # Python 2.7
                headers = HTTPHeaderDict.from_httplib(headers)

        # HTTPResponse objects in Python 3 don't have a .strict attribute
        strict = getattr(r, "strict", 0)
        resp = ResponseCls(
            body=r,
            headers=headers,
            status=r.status,
            version=r.version,
            reason=r.reason,
            strict=strict,
            original_response=r,
            **response_kw
        )
        return resp

    # Backwards-compatibility methods for httplib.HTTPResponse
    def getheaders(self):
        return self.headers

    def getheader(self, name, default=None):
        return self.headers.get(name, default)

    # Backwards compatibility for http.cookiejar
    def info(self):
        return self.headers

    # Overrides from io.IOBase
    def close(self):
        if not self.closed:
            self._fp.close()

        if self._connection:
            self._connection.close()

        if not self.auto_close:
            io.IOBase.close(self)

    @property
    def closed(self):
        if not self.auto_close:
            return io.IOBase.closed.__get__(self)
        elif self._fp is None:
            return True
        elif hasattr(self._fp, "isclosed"):
            return self._fp.isclosed()
        elif hasattr(self._fp, "closed"):
            return self._fp.closed
        else:
            return True

    def fileno(self):
        if self._fp is None:
            raise IOError("HTTPResponse has no file to get a fileno from")
        elif hasattr(self._fp, "fileno"):
            return self._fp.fileno()
        else:
            raise IOError(
                "The file-like object this HTTPResponse is wrapped "
                "around has no file descriptor"
            )

    def flush(self):
        if (
            self._fp is not None
            and hasattr(self._fp, "flush")
            and not getattr(self._fp, "closed", False)
        ):
            return self._fp.flush()

    def readable(self):
        # This method is required for `io` module compatibility.
        return True

    def readinto(self, b):
        # This method is required for `io` module compatibility.
        temp = self.read(len(b))
        if len(temp) == 0:
            return 0
        else:
            b[: len(temp)] = temp
            return len(temp)

    def supports_chunked_reads(self):
        """
        Checks if the underlying file-like object looks like a
        httplib.HTTPResponse object. We do this by testing for the fp
        attribute. If it is present we assume it returns raw chunks as
        processed by read_chunked().
        """
        return hasattr(self._fp, "fp")

    def _update_chunk_length(self):
        # First, we'll figure out length of a chunk and then
        # we'll try to read it from socket.
        if self.chunk_left is not None:
            return
        line = self._fp.fp.readline()
        line = line.split(b";", 1)[0]
        try:
            self.chunk_left = int(line, 16)
        except ValueError:
            # Invalid chunked protocol response, abort.
            self.close()
            raise httplib.IncompleteRead(line)

    def _handle_chunk(self, amt):
        returned_chunk = None
        if amt is None:
            chunk = self._fp._safe_read(self.chunk_left)
            returned_chunk = chunk
            self._fp._safe_read(2)  # Toss the CRLF at the end of the chunk.
            self.chunk_left = None
        elif amt < self.chunk_left:
            value = self._fp._safe_read(amt)
            self.chunk_left = self.chunk_left - amt
            returned_chunk = value
        elif amt == self.chunk_left:
            value = self._fp._safe_read(amt)
            self._fp._safe_read(2)  # Toss the CRLF at the end of the chunk.
            self.chunk_left = None
            returned_chunk = value
        else:  # amt > self.chunk_left
            returned_chunk = self._fp._safe_read(self.chunk_left)
            self._fp._safe_read(2)  # Toss the CRLF at the end of the chunk.
            self.chunk_left = None
        return returned_chunk

    def read_chunked(self, amt=None, decode_content=None):
        """
        Similar to :meth:`HTTPResponse.read`, but with an additional
        parameter: ``decode_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        """
        self._init_decoder()
        # FIXME: Rewrite this method and make it a class with a better structured logic.
        if not self.chunked:
            raise ResponseNotChunked(
                "Response is not chunked. "
                "Header 'transfer-encoding: chunked' is missing."
            )
        if not self.supports_chunked_reads():
            raise BodyNotHttplibCompatible(
                "Body should be httplib.HTTPResponse like. "
                "It should have have an fp attribute which returns raw chunks."
            )

        with self._error_catcher():
            # Don't bother reading the body of a HEAD request.
            if self._original_response and is_response_to_head(self._original_response):
                self._original_response.close()
                return

            # If a response is already read and closed
            # then return immediately.
            if self._fp.fp is None:
                return

            while True:
                self._update_chunk_length()
                if self.chunk_left == 0:
                    break
                chunk = self._handle_chunk(amt)
                decoded = self._decode(
                    chunk, decode_content=decode_content, flush_decoder=False
                )
                if decoded:
                    yield decoded

            if decode_content:
                # On CPython and PyPy, we should never need to flush the
                # decoder. However, on Jython we *might* need to, so
                # lets defensively do it anyway.
                decoded = self._flush_decoder()
                if decoded:  # Platform-specific: Jython.
                    yield decoded

            # Chunk content ends with \r\n: discard it.
            while True:
                line = self._fp.fp.readline()
                if not line:
                    # Some sites may not end with '\r\n'.
                    break
                if line == b"\r\n":
                    break

            # We read everything; close the "file".
            if self._original_response:
                self._original_response.close()

    def geturl(self):
        """
        Returns the URL that was the source of this response.
        If the request that generated this response redirected, this method
        will return the final redirect location.
        """
        if self.retries is not None and len(self.retries.history):
            return self.retries.history[-1].redirect_location
        else:
            return self._request_url

    def __iter__(self):
        buffer = [b""]
        for chunk in self.stream(decode_content=True):
            if b"\n" in chunk:
                chunk = chunk.split(b"\n")
                yield b"".join(buffer) + chunk[0] + b"\n"
                for x in chunk[1:-1]:
                    yield x + b"\n"
                if chunk[-1]:
                    buffer = [chunk[-1]]
                else:
                    buffer = []
            else:
                buffer.append(chunk)
        if buffer:
            yield b"".join(buffer)
site-packages/pip/_vendor/urllib3/util/response.py000064400000005015151464452710016263 0ustar00from __future__ import absolute_import
from ..packages.six.moves import http_client as httplib

from ..exceptions import HeaderParsingError


def is_fp_closed(obj):
    """
    Checks whether a given file-like object is closed.

    :param obj:
        The file-like object to check.
    """

    try:
        # Check `isclosed()` first, in case Python3 doesn't set `closed`.
        # GH Issue #928
        return obj.isclosed()
    except AttributeError:
        pass

    try:
        # Check via the official file-like-object way.
        return obj.closed
    except AttributeError:
        pass

    try:
        # Check if the object is a container for another file-like object that
        # gets released on exhaustion (e.g. HTTPResponse).
        return obj.fp is None
    except AttributeError:
        pass

    raise ValueError("Unable to determine whether fp is closed.")


def assert_header_parsing(headers):
    """
    Asserts whether all headers have been successfully parsed.
    Extracts encountered errors from the result of parsing headers.

    Only works on Python 3.

    :param headers: Headers to verify.
    :type headers: `httplib.HTTPMessage`.

    :raises urllib3.exceptions.HeaderParsingError:
        If parsing errors are found.
    """

    # This will fail silently if we pass in the wrong kind of parameter.
    # To make debugging easier add an explicit check.
    if not isinstance(headers, httplib.HTTPMessage):
        raise TypeError("expected httplib.Message, got {0}.".format(type(headers)))

    defects = getattr(headers, "defects", None)
    get_payload = getattr(headers, "get_payload", None)

    unparsed_data = None
    if get_payload:
        # get_payload is actually email.message.Message.get_payload;
        # we're only interested in the result if it's not a multipart message
        if not headers.is_multipart():
            payload = get_payload()

            if isinstance(payload, (bytes, str)):
                unparsed_data = payload

    if defects or unparsed_data:
        raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)


def is_response_to_head(response):
    """
    Checks whether the request of a response has been a HEAD-request.
    Handles the quirks of AppEngine.

    :param conn:
    :type conn: :class:`httplib.HTTPResponse`
    """
    # FIXME: Can we do this somehow without accessing private httplib _method?
    method = response._method
    if isinstance(method, int):  # Platform-specific: Appengine
        return method == 3
    return method.upper() == "HEAD"
site-packages/pip/_vendor/urllib3/util/ssl_.py000064400000033523151464452710015372 0ustar00from __future__ import absolute_import
import errno
import warnings
import hmac
import sys

from binascii import hexlify, unhexlify
from hashlib import md5, sha1, sha256

from .url import IPV4_RE, BRACELESS_IPV6_ADDRZ_RE
from ..exceptions import SSLError, InsecurePlatformWarning, SNIMissingWarning
from ..packages import six


SSLContext = None
HAS_SNI = False
IS_PYOPENSSL = False
IS_SECURETRANSPORT = False

# Maps the length of a digest to a possible hash function producing this digest
HASHFUNC_MAP = {32: md5, 40: sha1, 64: sha256}


def _const_compare_digest_backport(a, b):
    """
    Compare two digests of equal length in constant time.

    The digests must be of type str/bytes.
    Returns True if the digests match, and False otherwise.
    """
    result = abs(len(a) - len(b))
    for l, r in zip(bytearray(a), bytearray(b)):
        result |= l ^ r
    return result == 0


_const_compare_digest = getattr(hmac, "compare_digest", _const_compare_digest_backport)

try:  # Test for SSL features
    import ssl
    from ssl import wrap_socket, CERT_REQUIRED
    from ssl import HAS_SNI  # Has SNI?
except ImportError:
    pass

try:  # Platform-specific: Python 3.6
    from ssl import PROTOCOL_TLS

    PROTOCOL_SSLv23 = PROTOCOL_TLS
except ImportError:
    try:
        from ssl import PROTOCOL_SSLv23 as PROTOCOL_TLS

        PROTOCOL_SSLv23 = PROTOCOL_TLS
    except ImportError:
        PROTOCOL_SSLv23 = PROTOCOL_TLS = 2


try:
    from ssl import OP_NO_SSLv2, OP_NO_SSLv3, OP_NO_COMPRESSION
except ImportError:
    OP_NO_SSLv2, OP_NO_SSLv3 = 0x1000000, 0x2000000
    OP_NO_COMPRESSION = 0x20000


# A secure default.
# Sources for more information on TLS ciphers:
#
# - https://wiki.mozilla.org/Security/Server_Side_TLS
# - https://www.ssllabs.com/projects/best-practices/index.html
# - https://hynek.me/articles/hardening-your-web-servers-ssl-ciphers/
#
# The general intent is:
# - prefer cipher suites that offer perfect forward secrecy (DHE/ECDHE),
# - prefer ECDHE over DHE for better performance,
# - prefer any AES-GCM and ChaCha20 over any AES-CBC for better performance and
#   security,
# - prefer AES-GCM over ChaCha20 because hardware-accelerated AES is common,
# - disable NULL authentication, MD5 MACs, DSS, and other
#   insecure ciphers for security reasons.
# - NOTE: TLS 1.3 cipher suites are managed through a different interface
#   not exposed by CPython (yet!) and are enabled by default if they're available.
DEFAULT_CIPHERS = ":".join(
    [
        "ECDHE+AESGCM",
        "ECDHE+CHACHA20",
        "DHE+AESGCM",
        "DHE+CHACHA20",
        "ECDH+AESGCM",
        "DH+AESGCM",
        "ECDH+AES",
        "DH+AES",
        "RSA+AESGCM",
        "RSA+AES",
        "!aNULL",
        "!eNULL",
        "!MD5",
        "!DSS",
    ]
)

try:
    from ssl import SSLContext  # Modern SSL?
except ImportError:

    class SSLContext(object):  # Platform-specific: Python 2
        def __init__(self, protocol_version):
            self.protocol = protocol_version
            # Use default values from a real SSLContext
            self.check_hostname = False
            self.verify_mode = ssl.CERT_NONE
            self.ca_certs = None
            self.options = 0
            self.certfile = None
            self.keyfile = None
            self.ciphers = None

        def load_cert_chain(self, certfile, keyfile):
            self.certfile = certfile
            self.keyfile = keyfile

        def load_verify_locations(self, cafile=None, capath=None):
            self.ca_certs = cafile

            if capath is not None:
                raise SSLError("CA directories not supported in older Pythons")

        def set_ciphers(self, cipher_suite):
            self.ciphers = cipher_suite

        def wrap_socket(self, socket, server_hostname=None, server_side=False):
            warnings.warn(
                "A true SSLContext object is not available. This prevents "
                "urllib3 from configuring SSL appropriately and may cause "
                "certain SSL connections to fail. You can upgrade to a newer "
                "version of Python to solve this. For more information, see "
                "https://urllib3.readthedocs.io/en/latest/advanced-usage.html"
                "#ssl-warnings",
                InsecurePlatformWarning,
            )
            kwargs = {
                "keyfile": self.keyfile,
                "certfile": self.certfile,
                "ca_certs": self.ca_certs,
                "cert_reqs": self.verify_mode,
                "ssl_version": self.protocol,
                "server_side": server_side,
            }
            return wrap_socket(socket, ciphers=self.ciphers, **kwargs)


def assert_fingerprint(cert, fingerprint):
    """
    Checks if given fingerprint matches the supplied certificate.

    :param cert:
        Certificate as bytes object.
    :param fingerprint:
        Fingerprint as string of hexdigits, can be interspersed by colons.
    """

    fingerprint = fingerprint.replace(":", "").lower()
    digest_length = len(fingerprint)
    hashfunc = HASHFUNC_MAP.get(digest_length)
    if not hashfunc:
        raise SSLError("Fingerprint of invalid length: {0}".format(fingerprint))

    # We need encode() here for py32; works on py2 and p33.
    fingerprint_bytes = unhexlify(fingerprint.encode())

    cert_digest = hashfunc(cert).digest()

    if not _const_compare_digest(cert_digest, fingerprint_bytes):
        raise SSLError(
            'Fingerprints did not match. Expected "{0}", got "{1}".'.format(
                fingerprint, hexlify(cert_digest)
            )
        )


def resolve_cert_reqs(candidate):
    """
    Resolves the argument to a numeric constant, which can be passed to
    the wrap_socket function/method from the ssl module.
    Defaults to :data:`ssl.CERT_NONE`.
    If given a string it is assumed to be the name of the constant in the
    :mod:`ssl` module or its abbreviation.
    (So you can specify `REQUIRED` instead of `CERT_REQUIRED`.
    If it's neither `None` nor a string we assume it is already the numeric
    constant which can directly be passed to wrap_socket.
    """
    if candidate is None:
        return CERT_REQUIRED

    if isinstance(candidate, str):
        res = getattr(ssl, candidate, None)
        if res is None:
            res = getattr(ssl, "CERT_" + candidate)
        return res

    return candidate


def resolve_ssl_version(candidate):
    """
    like resolve_cert_reqs
    """
    if candidate is None:
        return PROTOCOL_TLS

    if isinstance(candidate, str):
        res = getattr(ssl, candidate, None)
        if res is None:
            res = getattr(ssl, "PROTOCOL_" + candidate)
        return res

    return candidate


def create_urllib3_context(
    ssl_version=None, cert_reqs=None, options=None, ciphers=None
):
    """All arguments have the same meaning as ``ssl_wrap_socket``.

    By default, this function does a lot of the same work that
    ``ssl.create_default_context`` does on Python 3.4+. It:

    - Disables SSLv2, SSLv3, and compression
    - Sets a restricted set of server ciphers

    If you wish to enable SSLv3, you can do::

        from pip._vendor.urllib3.util import ssl_
        context = ssl_.create_urllib3_context()
        context.options &= ~ssl_.OP_NO_SSLv3

    You can do the same to enable compression (substituting ``COMPRESSION``
    for ``SSLv3`` in the last line above).

    :param ssl_version:
        The desired protocol version to use. This will default to
        PROTOCOL_SSLv23 which will negotiate the highest protocol that both
        the server and your installation of OpenSSL support.
    :param cert_reqs:
        Whether to require the certificate verification. This defaults to
        ``ssl.CERT_REQUIRED``.
    :param options:
        Specific OpenSSL options. These default to ``ssl.OP_NO_SSLv2``,
        ``ssl.OP_NO_SSLv3``, ``ssl.OP_NO_COMPRESSION``.
    :param ciphers:
        Which cipher suites to allow the server to select.
    :returns:
        Constructed SSLContext object with specified options
    :rtype: SSLContext
    """
    context = SSLContext(ssl_version or PROTOCOL_TLS)

    context.set_ciphers(ciphers or DEFAULT_CIPHERS)

    # Setting the default here, as we may have no ssl module on import
    cert_reqs = ssl.CERT_REQUIRED if cert_reqs is None else cert_reqs

    if options is None:
        options = 0
        # SSLv2 is easily broken and is considered harmful and dangerous
        options |= OP_NO_SSLv2
        # SSLv3 has several problems and is now dangerous
        options |= OP_NO_SSLv3
        # Disable compression to prevent CRIME attacks for OpenSSL 1.0+
        # (issue #309)
        options |= OP_NO_COMPRESSION

    context.options |= options

    # Enable post-handshake authentication for TLS 1.3, see GH #1634. PHA is
    # necessary for conditional client cert authentication with TLS 1.3.
    # The attribute is None for OpenSSL <= 1.1.0 or does not exist in older
    # versions of Python.  We only enable on Python 3.7.4+ or if certificate
    # verification is enabled to work around Python issue #37428
    # See: https://bugs.python.org/issue37428
    if (cert_reqs == ssl.CERT_REQUIRED or sys.version_info >= (3, 7, 4)) and getattr(
        context, "post_handshake_auth", None
    ) is not None:
        context.post_handshake_auth = True

    context.verify_mode = cert_reqs
    if (
        getattr(context, "check_hostname", None) is not None
    ):  # Platform-specific: Python 3.2
        # We do our own verification, including fingerprints and alternative
        # hostnames. So disable it here
        context.check_hostname = False
    return context


def ssl_wrap_socket(
    sock,
    keyfile=None,
    certfile=None,
    cert_reqs=None,
    ca_certs=None,
    server_hostname=None,
    ssl_version=None,
    ciphers=None,
    ssl_context=None,
    ca_cert_dir=None,
    key_password=None,
):
    """
    All arguments except for server_hostname, ssl_context, and ca_cert_dir have
    the same meaning as they do when using :func:`ssl.wrap_socket`.

    :param server_hostname:
        When SNI is supported, the expected hostname of the certificate
    :param ssl_context:
        A pre-made :class:`SSLContext` object. If none is provided, one will
        be created using :func:`create_urllib3_context`.
    :param ciphers:
        A string of ciphers we wish the client to support.
    :param ca_cert_dir:
        A directory containing CA certificates in multiple separate files, as
        supported by OpenSSL's -CApath flag or the capath argument to
        SSLContext.load_verify_locations().
    :param key_password:
        Optional password if the keyfile is encrypted.
    """
    context = ssl_context
    if context is None:
        # Note: This branch of code and all the variables in it are no longer
        # used by urllib3 itself. We should consider deprecating and removing
        # this code.
        context = create_urllib3_context(ssl_version, cert_reqs, ciphers=ciphers)

    if ca_certs or ca_cert_dir:
        try:
            context.load_verify_locations(ca_certs, ca_cert_dir)
        except IOError as e:  # Platform-specific: Python 2.7
            raise SSLError(e)
        # Py33 raises FileNotFoundError which subclasses OSError
        # These are not equivalent unless we check the errno attribute
        except OSError as e:  # Platform-specific: Python 3.3 and beyond
            if e.errno == errno.ENOENT:
                raise SSLError(e)
            raise

    elif ssl_context is None and hasattr(context, "load_default_certs"):
        # try to load OS default certs; works well on Windows (require Python3.4+)
        context.load_default_certs()

    # Attempt to detect if we get the goofy behavior of the
    # keyfile being encrypted and OpenSSL asking for the
    # passphrase via the terminal and instead error out.
    if keyfile and key_password is None and _is_key_file_encrypted(keyfile):
        raise SSLError("Client private key is encrypted, password is required")

    if certfile:
        if key_password is None:
            context.load_cert_chain(certfile, keyfile)
        else:
            context.load_cert_chain(certfile, keyfile, key_password)

    # If we detect server_hostname is an IP address then the SNI
    # extension should not be used according to RFC3546 Section 3.1
    # We shouldn't warn the user if SNI isn't available but we would
    # not be using SNI anyways due to IP address for server_hostname.
    if (
        server_hostname is not None and not is_ipaddress(server_hostname)
    ) or IS_SECURETRANSPORT:
        if HAS_SNI and server_hostname is not None:
            return context.wrap_socket(sock, server_hostname=server_hostname)

        warnings.warn(
            "An HTTPS request has been made, but the SNI (Server Name "
            "Indication) extension to TLS is not available on this platform. "
            "This may cause the server to present an incorrect TLS "
            "certificate, which can cause validation failures. You can upgrade to "
            "a newer version of Python to solve this. For more information, see "
            "https://urllib3.readthedocs.io/en/latest/advanced-usage.html"
            "#ssl-warnings",
            SNIMissingWarning,
        )

    return context.wrap_socket(sock)


def is_ipaddress(hostname):
    """Detects whether the hostname given is an IPv4 or IPv6 address.
    Also detects IPv6 addresses with Zone IDs.

    :param str hostname: Hostname to examine.
    :return: True if the hostname is an IP address, False otherwise.
    """
    if not six.PY2 and isinstance(hostname, bytes):
        # IDN A-label bytes are ASCII compatible.
        hostname = hostname.decode("ascii")
    return bool(IPV4_RE.match(hostname) or BRACELESS_IPV6_ADDRZ_RE.match(hostname))


def _is_key_file_encrypted(key_file):
    """Detects if a key file is encrypted or not."""
    with open(key_file, "r") as f:
        for line in f:
            # Look for Proc-Type: 4,ENCRYPTED
            if "ENCRYPTED" in line:
                return True

    return False
site-packages/pip/_vendor/urllib3/util/retry.py000064400000036132151464452710015576 0ustar00from __future__ import absolute_import
import time
import logging
from collections import namedtuple
from itertools import takewhile
import email
import re

from ..exceptions import (
    ConnectTimeoutError,
    MaxRetryError,
    ProtocolError,
    ReadTimeoutError,
    ResponseError,
    InvalidHeader,
)
from ..packages import six


log = logging.getLogger(__name__)


# Data structure for representing the metadata of requests that result in a retry.
RequestHistory = namedtuple(
    "RequestHistory", ["method", "url", "error", "status", "redirect_location"]
)


class Retry(object):
    """ Retry configuration.

    Each retry attempt will create a new Retry object with updated values, so
    they can be safely reused.

    Retries can be defined as a default for a pool::

        retries = Retry(connect=5, read=2, redirect=5)
        http = PoolManager(retries=retries)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', retries=Retry(10))

    Retries can be disabled by passing ``False``::

        response = http.request('GET', 'http://example.com/', retries=False)

    Errors will be wrapped in :class:`~urllib3.exceptions.MaxRetryError` unless
    retries are disabled, in which case the causing exception will be raised.

    :param int total:
        Total number of retries to allow. Takes precedence over other counts.

        Set to ``None`` to remove this constraint and fall back on other
        counts. It's a good idea to set this to some sensibly-high value to
        account for unexpected edge cases and avoid infinite retry loops.

        Set to ``0`` to fail on the first retry.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int connect:
        How many connection-related errors to retry on.

        These are errors raised before the request is sent to the remote server,
        which we assume has not triggered the server to process the request.

        Set to ``0`` to fail on the first retry of this type.

    :param int read:
        How many times to retry on read errors.

        These errors are raised after the request was sent to the server, so the
        request may have side-effects.

        Set to ``0`` to fail on the first retry of this type.

    :param int redirect:
        How many redirects to perform. Limit this to avoid infinite redirect
        loops.

        A redirect is a HTTP response with a status code 301, 302, 303, 307 or
        308.

        Set to ``0`` to fail on the first retry of this type.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int status:
        How many times to retry on bad status codes.

        These are retries made on responses, where status code matches
        ``status_forcelist``.

        Set to ``0`` to fail on the first retry of this type.

    :param iterable method_whitelist:
        Set of uppercased HTTP method verbs that we should retry on.

        By default, we only retry on methods which are considered to be
        idempotent (multiple requests with the same parameters end with the
        same state). See :attr:`Retry.DEFAULT_METHOD_WHITELIST`.

        Set to a ``False`` value to retry on any verb.

    :param iterable status_forcelist:
        A set of integer HTTP status codes that we should force a retry on.
        A retry is initiated if the request method is in ``method_whitelist``
        and the response status code is in ``status_forcelist``.

        By default, this is disabled with ``None``.

    :param float backoff_factor:
        A backoff factor to apply between attempts after the second try
        (most errors are resolved immediately by a second try without a
        delay). urllib3 will sleep for::

            {backoff factor} * (2 ** ({number of total retries} - 1))

        seconds. If the backoff_factor is 0.1, then :func:`.sleep` will sleep
        for [0.0s, 0.2s, 0.4s, ...] between retries. It will never be longer
        than :attr:`Retry.BACKOFF_MAX`.

        By default, backoff is disabled (set to 0).

    :param bool raise_on_redirect: Whether, if the number of redirects is
        exhausted, to raise a MaxRetryError, or to return a response with a
        response code in the 3xx range.

    :param bool raise_on_status: Similar meaning to ``raise_on_redirect``:
        whether we should raise an exception, or return a response,
        if status falls in ``status_forcelist`` range and retries have
        been exhausted.

    :param tuple history: The history of the request encountered during
        each call to :meth:`~Retry.increment`. The list is in the order
        the requests occurred. Each list item is of class :class:`RequestHistory`.

    :param bool respect_retry_after_header:
        Whether to respect Retry-After header on status codes defined as
        :attr:`Retry.RETRY_AFTER_STATUS_CODES` or not.

    :param iterable remove_headers_on_redirect:
        Sequence of headers to remove from the request when a response
        indicating a redirect is returned before firing off the redirected
        request.
    """

    DEFAULT_METHOD_WHITELIST = frozenset(
        ["HEAD", "GET", "PUT", "DELETE", "OPTIONS", "TRACE"]
    )

    RETRY_AFTER_STATUS_CODES = frozenset([413, 429, 503])

    DEFAULT_REDIRECT_HEADERS_BLACKLIST = frozenset(["Authorization"])

    #: Maximum backoff time.
    BACKOFF_MAX = 120

    def __init__(
        self,
        total=10,
        connect=None,
        read=None,
        redirect=None,
        status=None,
        method_whitelist=DEFAULT_METHOD_WHITELIST,
        status_forcelist=None,
        backoff_factor=0,
        raise_on_redirect=True,
        raise_on_status=True,
        history=None,
        respect_retry_after_header=True,
        remove_headers_on_redirect=DEFAULT_REDIRECT_HEADERS_BLACKLIST,
    ):

        self.total = total
        self.connect = connect
        self.read = read
        self.status = status

        if redirect is False or total is False:
            redirect = 0
            raise_on_redirect = False

        self.redirect = redirect
        self.status_forcelist = status_forcelist or set()
        self.method_whitelist = method_whitelist
        self.backoff_factor = backoff_factor
        self.raise_on_redirect = raise_on_redirect
        self.raise_on_status = raise_on_status
        self.history = history or tuple()
        self.respect_retry_after_header = respect_retry_after_header
        self.remove_headers_on_redirect = frozenset(
            [h.lower() for h in remove_headers_on_redirect]
        )

    def new(self, **kw):
        params = dict(
            total=self.total,
            connect=self.connect,
            read=self.read,
            redirect=self.redirect,
            status=self.status,
            method_whitelist=self.method_whitelist,
            status_forcelist=self.status_forcelist,
            backoff_factor=self.backoff_factor,
            raise_on_redirect=self.raise_on_redirect,
            raise_on_status=self.raise_on_status,
            history=self.history,
            remove_headers_on_redirect=self.remove_headers_on_redirect,
            respect_retry_after_header=self.respect_retry_after_header,
        )
        params.update(kw)
        return type(self)(**params)

    @classmethod
    def from_int(cls, retries, redirect=True, default=None):
        """ Backwards-compatibility for the old retries format."""
        if retries is None:
            retries = default if default is not None else cls.DEFAULT

        if isinstance(retries, Retry):
            return retries

        redirect = bool(redirect) and None
        new_retries = cls(retries, redirect=redirect)
        log.debug("Converted retries value: %r -> %r", retries, new_retries)
        return new_retries

    def get_backoff_time(self):
        """ Formula for computing the current backoff

        :rtype: float
        """
        # We want to consider only the last consecutive errors sequence (Ignore redirects).
        consecutive_errors_len = len(
            list(
                takewhile(lambda x: x.redirect_location is None, reversed(self.history))
            )
        )
        if consecutive_errors_len <= 1:
            return 0

        backoff_value = self.backoff_factor * (2 ** (consecutive_errors_len - 1))
        return min(self.BACKOFF_MAX, backoff_value)

    def parse_retry_after(self, retry_after):
        # Whitespace: https://tools.ietf.org/html/rfc7230#section-3.2.4
        if re.match(r"^\s*[0-9]+\s*$", retry_after):
            seconds = int(retry_after)
        else:
            retry_date_tuple = email.utils.parsedate(retry_after)
            if retry_date_tuple is None:
                raise InvalidHeader("Invalid Retry-After header: %s" % retry_after)
            retry_date = time.mktime(retry_date_tuple)
            seconds = retry_date - time.time()

        if seconds < 0:
            seconds = 0

        return seconds

    def get_retry_after(self, response):
        """ Get the value of Retry-After in seconds. """

        retry_after = response.getheader("Retry-After")

        if retry_after is None:
            return None

        return self.parse_retry_after(retry_after)

    def sleep_for_retry(self, response=None):
        retry_after = self.get_retry_after(response)
        if retry_after:
            time.sleep(retry_after)
            return True

        return False

    def _sleep_backoff(self):
        backoff = self.get_backoff_time()
        if backoff <= 0:
            return
        time.sleep(backoff)

    def sleep(self, response=None):
        """ Sleep between retry attempts.

        This method will respect a server's ``Retry-After`` response header
        and sleep the duration of the time requested. If that is not present, it
        will use an exponential backoff. By default, the backoff factor is 0 and
        this method will return immediately.
        """

        if self.respect_retry_after_header and response:
            slept = self.sleep_for_retry(response)
            if slept:
                return

        self._sleep_backoff()

    def _is_connection_error(self, err):
        """ Errors when we're fairly sure that the server did not receive the
        request, so it should be safe to retry.
        """
        return isinstance(err, ConnectTimeoutError)

    def _is_read_error(self, err):
        """ Errors that occur after the request has been started, so we should
        assume that the server began processing it.
        """
        return isinstance(err, (ReadTimeoutError, ProtocolError))

    def _is_method_retryable(self, method):
        """ Checks if a given HTTP method should be retried upon, depending if
        it is included on the method whitelist.
        """
        if self.method_whitelist and method.upper() not in self.method_whitelist:
            return False

        return True

    def is_retry(self, method, status_code, has_retry_after=False):
        """ Is this method/status code retryable? (Based on whitelists and control
        variables such as the number of total retries to allow, whether to
        respect the Retry-After header, whether this header is present, and
        whether the returned status code is on the list of status codes to
        be retried upon on the presence of the aforementioned header)
        """
        if not self._is_method_retryable(method):
            return False

        if self.status_forcelist and status_code in self.status_forcelist:
            return True

        return (
            self.total
            and self.respect_retry_after_header
            and has_retry_after
            and (status_code in self.RETRY_AFTER_STATUS_CODES)
        )

    def is_exhausted(self):
        """ Are we out of retries? """
        retry_counts = (self.total, self.connect, self.read, self.redirect, self.status)
        retry_counts = list(filter(None, retry_counts))
        if not retry_counts:
            return False

        return min(retry_counts) < 0

    def increment(
        self,
        method=None,
        url=None,
        response=None,
        error=None,
        _pool=None,
        _stacktrace=None,
    ):
        """ Return a new Retry object with incremented retry counters.

        :param response: A response object, or None, if the server did not
            return a response.
        :type response: :class:`~urllib3.response.HTTPResponse`
        :param Exception error: An error encountered during the request, or
            None if the response was received successfully.

        :return: A new ``Retry`` object.
        """
        if self.total is False and error:
            # Disabled, indicate to re-raise the error.
            raise six.reraise(type(error), error, _stacktrace)

        total = self.total
        if total is not None:
            total -= 1

        connect = self.connect
        read = self.read
        redirect = self.redirect
        status_count = self.status
        cause = "unknown"
        status = None
        redirect_location = None

        if error and self._is_connection_error(error):
            # Connect retry?
            if connect is False:
                raise six.reraise(type(error), error, _stacktrace)
            elif connect is not None:
                connect -= 1

        elif error and self._is_read_error(error):
            # Read retry?
            if read is False or not self._is_method_retryable(method):
                raise six.reraise(type(error), error, _stacktrace)
            elif read is not None:
                read -= 1

        elif response and response.get_redirect_location():
            # Redirect retry?
            if redirect is not None:
                redirect -= 1
            cause = "too many redirects"
            redirect_location = response.get_redirect_location()
            status = response.status

        else:
            # Incrementing because of a server error like a 500 in
            # status_forcelist and a the given method is in the whitelist
            cause = ResponseError.GENERIC_ERROR
            if response and response.status:
                if status_count is not None:
                    status_count -= 1
                cause = ResponseError.SPECIFIC_ERROR.format(status_code=response.status)
                status = response.status

        history = self.history + (
            RequestHistory(method, url, error, status, redirect_location),
        )

        new_retry = self.new(
            total=total,
            connect=connect,
            read=read,
            redirect=redirect,
            status=status_count,
            history=history,
        )

        if new_retry.is_exhausted():
            raise MaxRetryError(_pool, url, error or ResponseError(cause))

        log.debug("Incremented Retry for (url='%s'): %r", url, new_retry)

        return new_retry

    def __repr__(self):
        return (
            "{cls.__name__}(total={self.total}, connect={self.connect}, "
            "read={self.read}, redirect={self.redirect}, status={self.status})"
        ).format(cls=type(self), self=self)


# For backwards compatibility (equivalent to pre-v1.9):
Retry.DEFAULT = Retry(3)
site-packages/pip/_vendor/urllib3/util/request.py000064400000007355151464452710016126 0ustar00from __future__ import absolute_import
from base64 import b64encode

from ..packages.six import b, integer_types
from ..exceptions import UnrewindableBodyError

ACCEPT_ENCODING = "gzip,deflate"
try:
    import brotli as _unused_module_brotli  # noqa: F401
except ImportError:
    pass
else:
    ACCEPT_ENCODING += ",br"

_FAILEDTELL = object()


def make_headers(
    keep_alive=None,
    accept_encoding=None,
    user_agent=None,
    basic_auth=None,
    proxy_basic_auth=None,
    disable_cache=None,
):
    """
    Shortcuts for generating request headers.

    :param keep_alive:
        If ``True``, adds 'connection: keep-alive' header.

    :param accept_encoding:
        Can be a boolean, list, or string.
        ``True`` translates to 'gzip,deflate'.
        List will get joined by comma.
        String will be used as provided.

    :param user_agent:
        String representing the user-agent you want, such as
        "python-urllib3/0.6"

    :param basic_auth:
        Colon-separated username:password string for 'authorization: basic ...'
        auth header.

    :param proxy_basic_auth:
        Colon-separated username:password string for 'proxy-authorization: basic ...'
        auth header.

    :param disable_cache:
        If ``True``, adds 'cache-control: no-cache' header.

    Example::

        >>> make_headers(keep_alive=True, user_agent="Batman/1.0")
        {'connection': 'keep-alive', 'user-agent': 'Batman/1.0'}
        >>> make_headers(accept_encoding=True)
        {'accept-encoding': 'gzip,deflate'}
    """
    headers = {}
    if accept_encoding:
        if isinstance(accept_encoding, str):
            pass
        elif isinstance(accept_encoding, list):
            accept_encoding = ",".join(accept_encoding)
        else:
            accept_encoding = ACCEPT_ENCODING
        headers["accept-encoding"] = accept_encoding

    if user_agent:
        headers["user-agent"] = user_agent

    if keep_alive:
        headers["connection"] = "keep-alive"

    if basic_auth:
        headers["authorization"] = "Basic " + b64encode(b(basic_auth)).decode("utf-8")

    if proxy_basic_auth:
        headers["proxy-authorization"] = "Basic " + b64encode(
            b(proxy_basic_auth)
        ).decode("utf-8")

    if disable_cache:
        headers["cache-control"] = "no-cache"

    return headers


def set_file_position(body, pos):
    """
    If a position is provided, move file to that point.
    Otherwise, we'll attempt to record a position for future use.
    """
    if pos is not None:
        rewind_body(body, pos)
    elif getattr(body, "tell", None) is not None:
        try:
            pos = body.tell()
        except (IOError, OSError):
            # This differentiates from None, allowing us to catch
            # a failed `tell()` later when trying to rewind the body.
            pos = _FAILEDTELL

    return pos


def rewind_body(body, body_pos):
    """
    Attempt to rewind body to a certain position.
    Primarily used for request redirects and retries.

    :param body:
        File-like object that supports seek.

    :param int pos:
        Position to seek to in file.
    """
    body_seek = getattr(body, "seek", None)
    if body_seek is not None and isinstance(body_pos, integer_types):
        try:
            body_seek(body_pos)
        except (IOError, OSError):
            raise UnrewindableBodyError(
                "An error occurred when rewinding request " "body for redirect/retry."
            )
    elif body_pos is _FAILEDTELL:
        raise UnrewindableBodyError(
            "Unable to record file position for rewinding "
            "request body during a redirect/retry."
        )
    else:
        raise ValueError(
            "body_pos must be of type integer, " "instead it was %s." % type(body_pos)
        )
site-packages/pip/_vendor/urllib3/util/connection.py000064400000011034151464452710016562 0ustar00from __future__ import absolute_import
import socket
from .wait import NoWayToWaitForSocketError, wait_for_read
from ..contrib import _appengine_environ


def is_connection_dropped(conn):  # Platform-specific
    """
    Returns True if the connection is dropped and should be closed.

    :param conn:
        :class:`httplib.HTTPConnection` object.

    Note: For platforms like AppEngine, this will always return ``False`` to
    let the platform handle connection recycling transparently for us.
    """
    sock = getattr(conn, "sock", False)
    if sock is False:  # Platform-specific: AppEngine
        return False
    if sock is None:  # Connection already closed (such as by httplib).
        return True
    try:
        # Returns True if readable, which here means it's been dropped
        return wait_for_read(sock, timeout=0.0)
    except NoWayToWaitForSocketError:  # Platform-specific: AppEngine
        return False


# This function is copied from socket.py in the Python 2.7 standard
# library test suite. Added to its signature is only `socket_options`.
# One additional modification is that we avoid binding to IPv6 servers
# discovered in DNS if the system doesn't have IPv6 functionality.
def create_connection(
    address,
    timeout=socket._GLOBAL_DEFAULT_TIMEOUT,
    source_address=None,
    socket_options=None,
):
    """Connect to *address* and return the socket object.

    Convenience function.  Connect to *address* (a 2-tuple ``(host,
    port)``) and return the socket object.  Passing the optional
    *timeout* parameter will set the timeout on the socket instance
    before attempting to connect.  If no *timeout* is supplied, the
    global default timeout setting returned by :func:`getdefaulttimeout`
    is used.  If *source_address* is set it must be a tuple of (host, port)
    for the socket to bind as a source address before making the connection.
    An host of '' or port 0 tells the OS to use the default.
    """

    host, port = address
    if host.startswith("["):
        host = host.strip("[]")
    err = None

    # Using the value from allowed_gai_family() in the context of getaddrinfo lets
    # us select whether to work with IPv4 DNS records, IPv6 records, or both.
    # The original create_connection function always returns all records.
    family = allowed_gai_family()

    for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
        af, socktype, proto, canonname, sa = res
        sock = None
        try:
            sock = socket.socket(af, socktype, proto)

            # If provided, set socket level options before connecting.
            _set_socket_options(sock, socket_options)

            if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT:
                sock.settimeout(timeout)
            if source_address:
                sock.bind(source_address)
            sock.connect(sa)
            return sock

        except socket.error as e:
            err = e
            if sock is not None:
                sock.close()
                sock = None

    if err is not None:
        raise err

    raise socket.error("getaddrinfo returns an empty list")


def _set_socket_options(sock, options):
    if options is None:
        return

    for opt in options:
        sock.setsockopt(*opt)


def allowed_gai_family():
    """This function is designed to work in the context of
    getaddrinfo, where family=socket.AF_UNSPEC is the default and
    will perform a DNS search for both IPv6 and IPv4 records."""

    family = socket.AF_INET
    if HAS_IPV6:
        family = socket.AF_UNSPEC
    return family


def _has_ipv6(host):
    """ Returns True if the system can bind an IPv6 address. """
    sock = None
    has_ipv6 = False

    # App Engine doesn't support IPV6 sockets and actually has a quota on the
    # number of sockets that can be used, so just early out here instead of
    # creating a socket needlessly.
    # See https://github.com/urllib3/urllib3/issues/1446
    if _appengine_environ.is_appengine_sandbox():
        return False

    if socket.has_ipv6:
        # has_ipv6 returns true if cPython was compiled with IPv6 support.
        # It does not tell us if the system has IPv6 support enabled. To
        # determine that we must bind to an IPv6 address.
        # https://github.com/shazow/urllib3/pull/611
        # https://bugs.python.org/issue658327
        try:
            sock = socket.socket(socket.AF_INET6)
            sock.bind((host, 0))
            has_ipv6 = True
        except Exception:
            pass

    if sock:
        sock.close()
    return has_ipv6


HAS_IPV6 = _has_ipv6("::1")
site-packages/pip/_vendor/urllib3/util/__pycache__/queue.cpython-38.opt-1.pyc000064400000002004151464452710022771 0ustar00U

.e@sDddlZddlmZddlmZejr.ddlZGdddejZdS)N)sixqueuec@s0eZdZddZefddZddZddZd	S)
	LifoQueuecCst|_dSN)collectionsdequer)self_rB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/queue.py_initszLifoQueue._initcCs
||jSrr)r
lenrrr
_qsizeszLifoQueue._qsizecCs|j|dSr)rappend)r
itemrrr
_putszLifoQueue._putcCs
|jSr)rpop)r
rrr
_getszLifoQueue._getN)__name__
__module____qualname__rrrrrrrrr
r
sr)	rZpackagesrZpackages.six.movesrZPY2ZQueueZ_unused_module_Queuerrrrr
s
site-packages/pip/_vendor/urllib3/util/__pycache__/retry.cpython-38.pyc000064400000031221151464452710022056 0ustar00U

.eZ<@sddlmZddlZddlZddlmZddlmZddlZddl	Z	ddl
mZmZm
Z
mZmZmZddlmZeeZedd	d
ddd
gZGdddeZede_dS))absolute_importN)
namedtuple)	takewhile)ConnectTimeoutError
MaxRetryError
ProtocolErrorReadTimeoutError
ResponseError
InvalidHeader)sixRequestHistorymethodurlerrorstatusredirect_locationc
@seZdZdZeddddddgZedd	d
gZedgZdZd
ddddeddddddef
ddZ	ddZ
ed2ddZddZ
ddZddZd3ddZdd Zd4d!d"Zd#d$Zd%d&Zd'd(Zd5d*d+Zd,d-Zd6d.d/Zd0d1ZdS)7Retrya3 Retry configuration.

    Each retry attempt will create a new Retry object with updated values, so
    they can be safely reused.

    Retries can be defined as a default for a pool::

        retries = Retry(connect=5, read=2, redirect=5)
        http = PoolManager(retries=retries)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', retries=Retry(10))

    Retries can be disabled by passing ``False``::

        response = http.request('GET', 'http://example.com/', retries=False)

    Errors will be wrapped in :class:`~urllib3.exceptions.MaxRetryError` unless
    retries are disabled, in which case the causing exception will be raised.

    :param int total:
        Total number of retries to allow. Takes precedence over other counts.

        Set to ``None`` to remove this constraint and fall back on other
        counts. It's a good idea to set this to some sensibly-high value to
        account for unexpected edge cases and avoid infinite retry loops.

        Set to ``0`` to fail on the first retry.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int connect:
        How many connection-related errors to retry on.

        These are errors raised before the request is sent to the remote server,
        which we assume has not triggered the server to process the request.

        Set to ``0`` to fail on the first retry of this type.

    :param int read:
        How many times to retry on read errors.

        These errors are raised after the request was sent to the server, so the
        request may have side-effects.

        Set to ``0`` to fail on the first retry of this type.

    :param int redirect:
        How many redirects to perform. Limit this to avoid infinite redirect
        loops.

        A redirect is a HTTP response with a status code 301, 302, 303, 307 or
        308.

        Set to ``0`` to fail on the first retry of this type.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int status:
        How many times to retry on bad status codes.

        These are retries made on responses, where status code matches
        ``status_forcelist``.

        Set to ``0`` to fail on the first retry of this type.

    :param iterable method_whitelist:
        Set of uppercased HTTP method verbs that we should retry on.

        By default, we only retry on methods which are considered to be
        idempotent (multiple requests with the same parameters end with the
        same state). See :attr:`Retry.DEFAULT_METHOD_WHITELIST`.

        Set to a ``False`` value to retry on any verb.

    :param iterable status_forcelist:
        A set of integer HTTP status codes that we should force a retry on.
        A retry is initiated if the request method is in ``method_whitelist``
        and the response status code is in ``status_forcelist``.

        By default, this is disabled with ``None``.

    :param float backoff_factor:
        A backoff factor to apply between attempts after the second try
        (most errors are resolved immediately by a second try without a
        delay). urllib3 will sleep for::

            {backoff factor} * (2 ** ({number of total retries} - 1))

        seconds. If the backoff_factor is 0.1, then :func:`.sleep` will sleep
        for [0.0s, 0.2s, 0.4s, ...] between retries. It will never be longer
        than :attr:`Retry.BACKOFF_MAX`.

        By default, backoff is disabled (set to 0).

    :param bool raise_on_redirect: Whether, if the number of redirects is
        exhausted, to raise a MaxRetryError, or to return a response with a
        response code in the 3xx range.

    :param bool raise_on_status: Similar meaning to ``raise_on_redirect``:
        whether we should raise an exception, or return a response,
        if status falls in ``status_forcelist`` range and retries have
        been exhausted.

    :param tuple history: The history of the request encountered during
        each call to :meth:`~Retry.increment`. The list is in the order
        the requests occurred. Each list item is of class :class:`RequestHistory`.

    :param bool respect_retry_after_header:
        Whether to respect Retry-After header on status codes defined as
        :attr:`Retry.RETRY_AFTER_STATUS_CODES` or not.

    :param iterable remove_headers_on_redirect:
        Sequence of headers to remove from the request when a response
        indicating a redirect is returned before firing off the redirected
        request.
    ZHEADZGETZPUTZDELETEZOPTIONSZTRACEiiiZ
Authorizationx
NrTcCs||_||_||_||_|dks(|dkr0d}d}	||_|p>t|_||_||_|	|_	|
|_
|pbt|_||_
tdd|
D|_dS)NFrcSsg|]}|qS)lower).0hrrB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/retry.py
sz"Retry.__init__..)totalconnectreadrredirectsetstatus_forcelistmethod_whitelistbackoff_factorraise_on_redirectraise_on_statustuplehistoryrespect_retry_after_header	frozensetremove_headers_on_redirect)selfrrrrrr"r!r#r$r%r'r(r*rrr__init__s$zRetry.__init__cKsTt|j|j|j|j|j|j|j|j|j	|j
|j|j|j
d
}||t|f|S)N)
rrrrrr"r!r#r$r%r'r*r()dictrrrrrr"r!r#r$r%r'r*r(updatetype)r+kwZparamsrrrnews"
z	Retry.newcCsR|dkr|dk	r|n|j}t|tr(|St|o2d}|||d}td|||S)z4 Backwards-compatibility for the old retries format.N)rz!Converted retries value: %r -> %r)DEFAULT
isinstancerboollogdebug)clsZretriesrdefaultZnew_retriesrrrfrom_ints
zRetry.from_intcCsFtttddt|j}|dkr(dS|jd|d}t|j|S)zJ Formula for computing the current backoff

        :rtype: float
        cSs
|jdkS)N)r)xrrrz(Retry.get_backoff_time..rr)lenlistrreversedr'r#minBACKOFF_MAX)r+Zconsecutive_errors_lenZ
backoff_valuerrrget_backoff_timeszRetry.get_backoff_timecCs\td|rt|}n6tj|}|dkr6td|t|}|t}|dkrXd}|S)Nz^\s*[0-9]+\s*$zInvalid Retry-After header: %sr)	rematchintemailZutilsZ	parsedatertimemktime)r+retry_afterZsecondsZretry_date_tupleZ
retry_daterrrparse_retry_afters

zRetry.parse_retry_aftercCs |d}|dkrdS||S)z* Get the value of Retry-After in seconds. zRetry-AfterN)Z	getheaderrKr+responserJrrrget_retry_after	s
zRetry.get_retry_aftercCs ||}|rt|dSdS)NTF)rNrHsleeprLrrrsleep_for_retrys


zRetry.sleep_for_retrycCs"|}|dkrdSt|dS)Nr)rCrHrO)r+Zbackoffrrr_sleep_backoffszRetry._sleep_backoffcCs(|jr|r||}|rdS|dS)aC Sleep between retry attempts.

        This method will respect a server's ``Retry-After`` response header
        and sleep the duration of the time requested. If that is not present, it
        will use an exponential backoff. By default, the backoff factor is 0 and
        this method will return immediately.
        N)r(rPrQ)r+rMZsleptrrrrO!s
	

zRetry.sleepcCs
t|tS)z{ Errors when we're fairly sure that the server did not receive the
        request, so it should be safe to retry.
        )r3rr+errrrr_is_connection_error1szRetry._is_connection_errorcCst|ttfS)z Errors that occur after the request has been started, so we should
        assume that the server began processing it.
        )r3r	rrRrrr_is_read_error7szRetry._is_read_errorcCs|jr||jkrdSdS)z| Checks if a given HTTP method should be retried upon, depending if
        it is included on the method whitelist.
        FT)r"upper)r+rrrr_is_method_retryable=szRetry._is_method_retryableFcCs<||sdS|jr"||jkr"dS|jo:|jo:|o:||jkS)ax Is this method/status code retryable? (Based on whitelists and control
        variables such as the number of total retries to allow, whether to
        respect the Retry-After header, whether this header is present, and
        whether the returned status code is on the list of status codes to
        be retried upon on the presence of the aforementioned header)
        FT)rWr!rr(RETRY_AFTER_STATUS_CODES)r+rstatus_codeZhas_retry_afterrrris_retryFs
zRetry.is_retrycCs:|j|j|j|j|jf}ttd|}|s.dSt|dkS)z Are we out of retries? NFr)rrrrrr?filterrA)r+Zretry_countsrrris_exhaustedZs
zRetry.is_exhaustedcCs|jdkr |r tt||||j}|dk	r6|d8}|j}|j}	|j}
|j}d}d}
d}|r||r|dkrtt|||n|dk	r|d8}n|r|	|r|	dks|
|stt|||n|	dk	r|	d8}	np|r|r|
dk	r|
d8}
d}|}|j}
ns$ 
(site-packages/pip/_vendor/urllib3/util/__pycache__/connection.cpython-38.opt-1.pyc000064400000006124151464452710024013 0ustar00U

.e@snddlmZddlZddlmZmZddlmZddZej	ddfd	d
Z
ddZd
dZddZ
e
dZdS))absolute_importN)NoWayToWaitForSocketError
wait_for_read)_appengine_environcCsNt|dd}|dkrdS|dkr$dSzt|ddWStk
rHYdSXdS)a 
    Returns True if the connection is dropped and should be closed.

    :param conn:
        :class:`httplib.HTTPConnection` object.

    Note: For platforms like AppEngine, this will always return ``False`` to
    let the platform handle connection recycling transparently for us.
    sockFNTg)timeout)getattrrr)ZconnrrG/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/connection.pyis_connection_droppeds
r
cCs|\}}|dr|d}d}t}t|||tjD]}|\}	}
}}}
d}zNt|	|
|}t|||tjk	r||||r|	||
|
|WStjk
r}z|}|dk	r|d}W5d}~XYq8Xq8|dk	r|tddS)adConnect to *address* and return the socket object.

    Convenience function.  Connect to *address* (a 2-tuple ``(host,
    port)``) and return the socket object.  Passing the optional
    *timeout* parameter will set the timeout on the socket instance
    before attempting to connect.  If no *timeout* is supplied, the
    global default timeout setting returned by :func:`getdefaulttimeout`
    is used.  If *source_address* is set it must be a tuple of (host, port)
    for the socket to bind as a source address before making the connection.
    An host of '' or port 0 tells the OS to use the default.
    [z[]Nz!getaddrinfo returns an empty list)
startswithstripallowed_gai_familysocketZgetaddrinfoZSOCK_STREAM_set_socket_options_GLOBAL_DEFAULT_TIMEOUTZ
settimeoutbindZconnecterrorclose)Zaddressr	Zsource_addressZsocket_optionshostZporterrfamilyresZafZsocktypeprotoZ	canonnameZsarerrrcreate_connection!s2







rcCs$|dkrdS|D]}|j|qdS)N)Z
setsockopt)rZoptionsoptrrrrYsrcCstj}trtj}|S)zThis function is designed to work in the context of
    getaddrinfo, where family=socket.AF_UNSPEC is the default and
    will perform a DNS search for both IPv6 and IPv4 records.)rZAF_INETHAS_IPV6Z	AF_UNSPEC)rrrrrasrcCsbd}d}trdStjrRz"ttj}||dfd}Wntk
rPYnX|r^||S)z6 Returns True if the system can bind an IPv6 address. NFrT)rZis_appengine_sandboxrhas_ipv6ZAF_INET6r	Exceptionr)rrr!rrr	_has_ipv6lsr#z::1)Z
__future__rrwaitrrZcontribrr
rrrrr#r rrrrs
8site-packages/pip/_vendor/urllib3/util/__pycache__/ssl_.cpython-38.pyc000064400000023135151464452710021656 0ustar00U

.eS7@s(ddlmZddlZddlZddlZddlZddlmZmZddl	m
Z
mZmZddl
mZmZddlmZmZmZdd	lmZdZd
Zd
Zd
Ze
eedZdd
ZeedeZz(ddlZddlmZm Z ddlmZWne!k
rYnXzddlm"Z"e"Z#WnJe!k
rLzddlm#Z"e"Z#Wne!k
rFdZ#Z"YnXYnXzddlm$Z$m%Z%m&Z&Wn"e!k
rd\Z$Z%dZ&YnXd'dddddddddd d!d"d#d$gZ(zdd%lmZWn&e!k
rGd&d'd'e)ZYnXd(d)Z*d*d+Z+d,d-Z,d6d.d/Z-d7d0d1Z.d2d3Z/d4d5Z0dS)8)absolute_importN)hexlify	unhexlify)md5sha1sha256)IPV4_REBRACELESS_IPV6_ADDRZ_RE)SSLErrorInsecurePlatformWarningSNIMissingWarning)sixF) (@cCsDtt|t|}tt|t|D]\}}|||AO}q&|dkS)z
    Compare two digests of equal length in constant time.

    The digests must be of type str/bytes.
    Returns True if the digests match, and False otherwise.
    r)abslenzip	bytearray)abresultlrrA/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/ssl_.py_const_compare_digest_backportsrZcompare_digest)wrap_socket
CERT_REQUIRED)HAS_SNI)PROTOCOL_TLS)PROTOCOL_SSLv23)OP_NO_SSLv2OP_NO_SSLv3OP_NO_COMPRESSION)iii:zECDHE+AESGCMzECDHE+CHACHA20z
DHE+AESGCMzDHE+CHACHA20zECDH+AESGCMz	DH+AESGCMzECDH+AESzDH+AESz
RSA+AESGCMzRSA+AESz!aNULLz!eNULLz!MD5z!DSS)
SSLContextc@s8eZdZddZddZd
ddZdd	ZdddZdS)r(cCs6||_d|_tj|_d|_d|_d|_d|_d|_	dS)NFr)
protocolcheck_hostnamesslZ	CERT_NONEverify_modeca_certsoptionscertfilekeyfileciphers)selfZprotocol_versionrrr__init__kszSSLContext.__init__cCs||_||_dSN)r/r0)r2r/r0rrrload_cert_chainvszSSLContext.load_cert_chainNcCs||_|dk	rtddS)Nz-CA directories not supported in older Pythons)r-r)r2ZcafileZcapathrrrload_verify_locationszsz SSLContext.load_verify_locationscCs
||_dSr4r1)r2Zcipher_suiterrrset_ciphersszSSLContext.set_ciphersFcCs>tdt|j|j|j|j|j|d}t|fd|j	i|S)Na2A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings)r0r/r-	cert_reqsssl_versionserver_sider1)
warningswarnr
r0r/r-r,r)rr1)r2Zsocketserver_hostnamer;kwargsrrrrs
zSSLContext.wrap_socket)NN)NF)__name__
__module____qualname__r3r5r6r8rrrrrr(js

r(cCsn|dd}t|}t|}|s4td|t|}||	}t
||sjtd|t|dS)z
    Checks if given fingerprint matches the supplied certificate.

    :param cert:
        Certificate as bytes object.
    :param fingerprint:
        Fingerprint as string of hexdigits, can be interspersed by colons.
    r'z"Fingerprint of invalid length: {0}z6Fingerprints did not match. Expected "{0}", got "{1}".N)replacelowerrHASHFUNC_MAPgetrformatrencodeZdigest_const_compare_digestr)ZcertZfingerprintZ
digest_lengthZhashfuncZfingerprint_bytesZcert_digestrrrassert_fingerprints


rKcCs@|dkrtSt|traAn HTTPS request has been made, but the SNI (Server Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings)r\r6IOErrorrOSErrorerrnoZENOENThasattrr]_is_key_file_encryptedr5is_ipaddressIS_SECURETRANSPORTr!rr<r=r)
Zsockr0r/r9r-r>r:r1Zssl_contextZca_cert_dirZkey_passwordr[errrssl_wrap_socket(sBrfcCs2tjst|tr|d}tt|p.t|S)zDetects whether the hostname given is an IPv4 or IPv6 address.
    Also detects IPv6 addresses with Zone IDs.

    :param str hostname: Hostname to examine.
    :return: True if the hostname is an IP address, False otherwise.
    ascii)	rZPY2rLbytesdecodeboolr	matchr
)Zhostnamerrrrcs
rcc	Cs<t|d(}|D]}d|krW5QRdSqW5QRXdS)z*Detects if a key file is encrypted or not.rZ	ENCRYPTEDTF)open)Zkey_fileflinerrrrbs
rb)NNNN)
NNNNNNNNNN)1Z
__future__rr`r<ZhmacrYZbinasciirrZhashlibrrrZurlr	r

exceptionsrr
rZpackagesrr(r!ZIS_PYOPENSSLrdrFrrNrJr+rr ImportErrorr"r#r$r%r&joinrXobjectrKrRrSr\rfrcrbrrrrs

.
N
Z
site-packages/pip/_vendor/urllib3/util/__pycache__/__init__.cpython-38.pyc000064400000001727151464452710022460 0ustar00U

.e@sddlmZddlmZddlmZddlmZddlm	Z	m
Z
mZmZm
Z
mZmZmZmZddlmZmZddlmZdd	lmZmZmZmZdd
lmZmZdZdS)
)absolute_import)is_connection_dropped)make_headers)is_fp_closed)	
SSLContextHAS_SNIIS_PYOPENSSLIS_SECURETRANSPORTassert_fingerprintresolve_cert_reqsresolve_ssl_versionssl_wrap_socketPROTOCOL_TLS)current_timeTimeout)Retry)get_host	parse_urlsplit_firstUrl)
wait_for_readwait_for_write)rr	r
rrrrrrrrrrrrrr
rrrrN) Z
__future__rZ
connectionrZrequestrZresponserZssl_rrr	r
rrr
rrZtimeoutrrZretryrZurlrrrrwaitrr__all__rrE/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/__init__.pys,site-packages/pip/_vendor/urllib3/util/__pycache__/timeout.cpython-38.pyc000064400000021213151464452710022377 0ustar00U

.e&@sTddlmZddlmZddlZddlmZeZe	edejZ
GdddeZdS)	)absolute_import)_GLOBAL_DEFAULT_TIMEOUTN)TimeoutStateError	monotonicc@steZdZdZeZdeefddZddZe	ddZ
e	d	d
ZddZd
dZ
ddZeddZeddZdS)Timeouta Timeout configuration.

    Timeouts can be defined as a default for a pool::

        timeout = Timeout(connect=2.0, read=7.0)
        http = PoolManager(timeout=timeout)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', timeout=Timeout(10))

    Timeouts can be disabled by setting all the parameters to ``None``::

        no_timeout = Timeout(connect=None, read=None)
        response = http.request('GET', 'http://example.com/, timeout=no_timeout)


    :param total:
        This combines the connect and read timeouts into one; the read timeout
        will be set to the time leftover from the connect attempt. In the
        event that both a connect timeout and a total are specified, or a read
        timeout and a total are specified, the shorter timeout will be applied.

        Defaults to None.

    :type total: integer, float, or None

    :param connect:
        The maximum amount of time (in seconds) to wait for a connection
        attempt to a server to succeed. Omitting the parameter will default the
        connect timeout to the system default, probably `the global default
        timeout in socket.py
        `_.
        None will set an infinite timeout for connection attempts.

    :type connect: integer, float, or None

    :param read:
        The maximum amount of time (in seconds) to wait between consecutive
        read operations for a response from the server. Omitting the parameter
        will default the read timeout to the system default, probably `the
        global default timeout in socket.py
        `_.
        None will set an infinite timeout.

    :type read: integer, float, or None

    .. note::

        Many factors can affect the total amount of time for urllib3 to return
        an HTTP response.

        For example, Python's DNS resolver does not obey the timeout specified
        on the socket. Other factors that can affect total request time include
        high CPU load, high swap, the program running at a low priority level,
        or other behaviors.

        In addition, the read and total timeouts only measure the time between
        read operations on the socket connecting the client and the server,
        not the total amount of time for the request to return a complete
        response. For most requests, the timeout is raised because the server
        has not sent the first byte in the specified time. This is not always
        the case; if a server streams one byte every fifteen seconds, a timeout
        of 20 seconds will not trigger, even though the request will take
        several minutes to complete.

        If your goal is to cut off any request after a set amount of wall clock
        time, consider having a second "watcher" thread to cut off a slow
        request.
    NcCs4||d|_||d|_||d|_d|_dS)Nconnectreadtotal)_validate_timeout_connect_readr
_start_connect)selfr
rr	rD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/timeout.py__init___szTimeout.__init__cCsdt|j|j|j|jfS)Nz!%s(connect=%r, read=%r, total=%r))type__name__rr
r
rrrr__str__eszTimeout.__str__c	Cs|tkr|jS|dks ||jkr$|St|tr6tdzt|Wn(ttfk
rjtd||fYnXz|dkrtd||fWn$tk
rtd||fYnX|S)a Check that a timeout attribute is valid.

        :param value: The timeout value to validate
        :param name: The name of the timeout attribute to validate. This is
            used to specify in error messages.
        :return: The validated and casted version of the given value.
        :raises ValueError: If it is a numeric value less than or equal to
            zero, or the type is not an integer, float, or None.
        NzDTimeout cannot be a boolean value. It must be an int, float or None.z>Timeout value %s was %s, but it must be an int, float or None.rzdAttempted to set %s timeout to %s, but the timeout cannot be set to a value less than or equal to 0.)_DefaultDEFAULT_TIMEOUT
isinstancebool
ValueErrorfloat	TypeError)clsvaluenamerrrrms<


zTimeout._validate_timeoutcCst||dS)a Create a new Timeout from a legacy timeout value.

        The timeout value used by httplib.py sets the same timeout on the
        connect(), and recv() socket requests. This creates a :class:`Timeout`
        object that sets the individual timeouts to the ``timeout`` value
        passed to this function.

        :param timeout: The legacy timeout value.
        :type timeout: integer, float, sentinel default object, or None
        :return: Timeout object
        :rtype: :class:`Timeout`
        )r	r)r)rZtimeoutrrr
from_floatszTimeout.from_floatcCst|j|j|jdS)a Create a copy of the timeout object

        Timeout properties are stored per-pool but each request needs a fresh
        Timeout object to ensure each one has its own start/stop configured.

        :return: a copy of the timeout object
        :rtype: :class:`Timeout`
        )rr	r
)rrr
r
rrrrclonesz
Timeout.clonecCs |jdk	rtdt|_|jS)z Start the timeout clock, used during a connect() attempt

        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to start a timer that has been started already.
        Nz'Timeout timer has already been started.rrcurrent_timerrrr
start_connects
zTimeout.start_connectcCs|jdkrtdt|jS)a Gets the time elapsed since the call to :meth:`start_connect`.

        :return: Elapsed time in seconds.
        :rtype: float
        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to get duration for a timer that hasn't been started.
        Nz:Can't get connect duration for timer that has not started.r#rrrrget_connect_durations

zTimeout.get_connect_durationcCs:|jdkr|jS|jdks&|j|jkr,|jSt|j|jS)a" Get the value to use when setting a connection timeout.

        This will be a positive float or integer, the value None
        (never timeout), or the default system timeout.

        :return: Connect timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        N)r
rrminrrrrconnect_timeouts


zTimeout.connect_timeoutcCs|jdk	rX|j|jk	rX|jdk	rX|j|jk	rX|jdkr<|jStdt|j||jS|jdk	r|j|jk	rtd|j|S|jSdS)a Get the value for the read timeout.

        This assumes some time has elapsed in the connection timeout and
        computes the read timeout appropriately.

        If self.total is set, the read timeout is dependent on the amount of
        time taken by the connect timeout. If the connection time has not been
        established, a :exc:`~urllib3.exceptions.TimeoutStateError` will be
        raised.

        :return: Value to use for the read timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        :raises urllib3.exceptions.TimeoutStateError: If :meth:`start_connect`
            has not yet been called on this object.
        Nr)r
rr
rmaxr'r&rrrrread_timeouts


zTimeout.read_timeout)r
__module____qualname____doc__rrrrrclassmethodrr!r"r%r&propertyr(r*rrrrrsI
-

r)Z
__future__rZsocketrtime
exceptionsrobjectrgetattrr$rrrrrssite-packages/pip/_vendor/urllib3/util/__pycache__/wait.cpython-38.pyc000064400000006005151464452710021657 0ustar00U

.e@sddlZddlmZddlZddlZzddlmZWn ek
rTddlmZYnXdddgZGddde	Z
ejd	krd
dZnddZdddZ
dddZddZddZddadddZdddZdS)N)partial)	monotonic)timeNoWayToWaitForSocketError
wait_for_readwait_for_writec@seZdZdS)rN)__name__
__module____qualname__rrA/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/wait.pyrs)cCs||S)Nr)fntimeoutrrr_retry_on_intr*src
Cs|dkrtd}n
t|}z
||WSttjfk
r}zF|jdtjkrRn.|t}|dkrhd}|tdkrxd}WYqW5d}~XYqXqdS)Ninfr)floatrOSErrorselecterrorargserrnoZEINTR)rrZdeadlineerrrr0s



Fc
Csd|s|stdg}g}|r&|||r4||ttj|||}t||\}}}	t|p`|p`|	S)N2must specify at least one of read=True, write=True)RuntimeErrorappendrrrbool)
sockreadwriterZrcheckZwcheckrZrreadyZwreadyZxreadyrrrselect_wait_for_socketGs

r!cs^|s|stdd}|r"|tjO}|r0|tjO}t||fdd}tt||S)Nrrcs|dk	r|d9}|S)Ni)poll)tZpoll_objrrdo_pollfsz%poll_wait_for_socket..do_poll)rrZPOLLINZPOLLOUTr"registerrr)rrr rmaskr%rr$rpoll_wait_for_socketZs

r(cOstddS)Nzno select-equivalent available)rrkwargsrrrnull_wait_for_socketnsr+c	Cs<zt}t|jdWnttfk
r2YdSXdSdS)NrFT)rr"rAttributeErrorrr$rrr_have_working_pollrsr-cOs*trtanttdrtantat||S)Nr)r-r(wait_for_sockethasattrrr!r+r)rrrr.s
r.cCst|d|dS)z Waits for reading to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    T)rrr.rrrrrrscCst|d|dS)z Waits for writing to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    T)r rr0r1rrrrs)FFN)FFN)N)N)r	functoolsrrsysrrImportError__all__	Exceptionrversion_inforr!r(r+r-r.rrrrrrs&






site-packages/pip/_vendor/urllib3/util/__pycache__/wait.cpython-38.opt-1.pyc000064400000006005151464452710022616 0ustar00U

.e@sddlZddlmZddlZddlZzddlmZWn ek
rTddlmZYnXdddgZGddde	Z
ejd	krd
dZnddZdddZ
dddZddZddZddadddZdddZdS)N)partial)	monotonic)timeNoWayToWaitForSocketError
wait_for_readwait_for_writec@seZdZdS)rN)__name__
__module____qualname__rrA/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/wait.pyrs)cCs||S)Nr)fntimeoutrrr_retry_on_intr*src
Cs|dkrtd}n
t|}z
||WSttjfk
r}zF|jdtjkrRn.|t}|dkrhd}|tdkrxd}WYqW5d}~XYqXqdS)Ninfr)floatrOSErrorselecterrorargserrnoZEINTR)rrZdeadlineerrrr0s



Fc
Csd|s|stdg}g}|r&|||r4||ttj|||}t||\}}}	t|p`|p`|	S)N2must specify at least one of read=True, write=True)RuntimeErrorappendrrrbool)
sockreadwriterZrcheckZwcheckrZrreadyZwreadyZxreadyrrrselect_wait_for_socketGs

r!cs^|s|stdd}|r"|tjO}|r0|tjO}t||fdd}tt||S)Nrrcs|dk	r|d9}|S)Ni)poll)tZpoll_objrrdo_pollfsz%poll_wait_for_socket..do_poll)rrZPOLLINZPOLLOUTr"registerrr)rrr rmaskr%rr$rpoll_wait_for_socketZs

r(cOstddS)Nzno select-equivalent available)rrkwargsrrrnull_wait_for_socketnsr+c	Cs<zt}t|jdWnttfk
r2YdSXdSdS)NrFT)rr"rAttributeErrorrr$rrr_have_working_pollrsr-cOs*trtanttdrtantat||S)Nr)r-r(wait_for_sockethasattrrr!r+r)rrrr.s
r.cCst|d|dS)z Waits for reading to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    T)rrr.rrrrrrscCst|d|dS)z Waits for writing to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    T)r rr0r1rrrrs)FFN)FFN)N)N)r	functoolsrrsysrrImportError__all__	Exceptionrversion_inforr!r(r+r-r.rrrrrrs&






site-packages/pip/_vendor/urllib3/util/__pycache__/retry.cpython-38.opt-1.pyc000064400000031221151464452710023015 0ustar00U

.eZ<@sddlmZddlZddlZddlmZddlmZddlZddl	Z	ddl
mZmZm
Z
mZmZmZddlmZeeZedd	d
ddd
gZGdddeZede_dS))absolute_importN)
namedtuple)	takewhile)ConnectTimeoutError
MaxRetryError
ProtocolErrorReadTimeoutError
ResponseError
InvalidHeader)sixRequestHistorymethodurlerrorstatusredirect_locationc
@seZdZdZeddddddgZedd	d
gZedgZdZd
ddddeddddddef
ddZ	ddZ
ed2ddZddZ
ddZddZd3ddZdd Zd4d!d"Zd#d$Zd%d&Zd'd(Zd5d*d+Zd,d-Zd6d.d/Zd0d1ZdS)7Retrya3 Retry configuration.

    Each retry attempt will create a new Retry object with updated values, so
    they can be safely reused.

    Retries can be defined as a default for a pool::

        retries = Retry(connect=5, read=2, redirect=5)
        http = PoolManager(retries=retries)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', retries=Retry(10))

    Retries can be disabled by passing ``False``::

        response = http.request('GET', 'http://example.com/', retries=False)

    Errors will be wrapped in :class:`~urllib3.exceptions.MaxRetryError` unless
    retries are disabled, in which case the causing exception will be raised.

    :param int total:
        Total number of retries to allow. Takes precedence over other counts.

        Set to ``None`` to remove this constraint and fall back on other
        counts. It's a good idea to set this to some sensibly-high value to
        account for unexpected edge cases and avoid infinite retry loops.

        Set to ``0`` to fail on the first retry.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int connect:
        How many connection-related errors to retry on.

        These are errors raised before the request is sent to the remote server,
        which we assume has not triggered the server to process the request.

        Set to ``0`` to fail on the first retry of this type.

    :param int read:
        How many times to retry on read errors.

        These errors are raised after the request was sent to the server, so the
        request may have side-effects.

        Set to ``0`` to fail on the first retry of this type.

    :param int redirect:
        How many redirects to perform. Limit this to avoid infinite redirect
        loops.

        A redirect is a HTTP response with a status code 301, 302, 303, 307 or
        308.

        Set to ``0`` to fail on the first retry of this type.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int status:
        How many times to retry on bad status codes.

        These are retries made on responses, where status code matches
        ``status_forcelist``.

        Set to ``0`` to fail on the first retry of this type.

    :param iterable method_whitelist:
        Set of uppercased HTTP method verbs that we should retry on.

        By default, we only retry on methods which are considered to be
        idempotent (multiple requests with the same parameters end with the
        same state). See :attr:`Retry.DEFAULT_METHOD_WHITELIST`.

        Set to a ``False`` value to retry on any verb.

    :param iterable status_forcelist:
        A set of integer HTTP status codes that we should force a retry on.
        A retry is initiated if the request method is in ``method_whitelist``
        and the response status code is in ``status_forcelist``.

        By default, this is disabled with ``None``.

    :param float backoff_factor:
        A backoff factor to apply between attempts after the second try
        (most errors are resolved immediately by a second try without a
        delay). urllib3 will sleep for::

            {backoff factor} * (2 ** ({number of total retries} - 1))

        seconds. If the backoff_factor is 0.1, then :func:`.sleep` will sleep
        for [0.0s, 0.2s, 0.4s, ...] between retries. It will never be longer
        than :attr:`Retry.BACKOFF_MAX`.

        By default, backoff is disabled (set to 0).

    :param bool raise_on_redirect: Whether, if the number of redirects is
        exhausted, to raise a MaxRetryError, or to return a response with a
        response code in the 3xx range.

    :param bool raise_on_status: Similar meaning to ``raise_on_redirect``:
        whether we should raise an exception, or return a response,
        if status falls in ``status_forcelist`` range and retries have
        been exhausted.

    :param tuple history: The history of the request encountered during
        each call to :meth:`~Retry.increment`. The list is in the order
        the requests occurred. Each list item is of class :class:`RequestHistory`.

    :param bool respect_retry_after_header:
        Whether to respect Retry-After header on status codes defined as
        :attr:`Retry.RETRY_AFTER_STATUS_CODES` or not.

    :param iterable remove_headers_on_redirect:
        Sequence of headers to remove from the request when a response
        indicating a redirect is returned before firing off the redirected
        request.
    ZHEADZGETZPUTZDELETEZOPTIONSZTRACEiiiZ
Authorizationx
NrTcCs||_||_||_||_|dks(|dkr0d}d}	||_|p>t|_||_||_|	|_	|
|_
|pbt|_||_
tdd|
D|_dS)NFrcSsg|]}|qS)lower).0hrrB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/retry.py
sz"Retry.__init__..)totalconnectreadrredirectsetstatus_forcelistmethod_whitelistbackoff_factorraise_on_redirectraise_on_statustuplehistoryrespect_retry_after_header	frozensetremove_headers_on_redirect)selfrrrrrr"r!r#r$r%r'r(r*rrr__init__s$zRetry.__init__cKsTt|j|j|j|j|j|j|j|j|j	|j
|j|j|j
d
}||t|f|S)N)
rrrrrr"r!r#r$r%r'r*r()dictrrrrrr"r!r#r$r%r'r*r(updatetype)r+kwZparamsrrrnews"
z	Retry.newcCsR|dkr|dk	r|n|j}t|tr(|St|o2d}|||d}td|||S)z4 Backwards-compatibility for the old retries format.N)rz!Converted retries value: %r -> %r)DEFAULT
isinstancerboollogdebug)clsZretriesrdefaultZnew_retriesrrrfrom_ints
zRetry.from_intcCsFtttddt|j}|dkr(dS|jd|d}t|j|S)zJ Formula for computing the current backoff

        :rtype: float
        cSs
|jdkS)N)r)xrrrz(Retry.get_backoff_time..rr)lenlistrreversedr'r#minBACKOFF_MAX)r+Zconsecutive_errors_lenZ
backoff_valuerrrget_backoff_timeszRetry.get_backoff_timecCs\td|rt|}n6tj|}|dkr6td|t|}|t}|dkrXd}|S)Nz^\s*[0-9]+\s*$zInvalid Retry-After header: %sr)	rematchintemailZutilsZ	parsedatertimemktime)r+retry_afterZsecondsZretry_date_tupleZ
retry_daterrrparse_retry_afters

zRetry.parse_retry_aftercCs |d}|dkrdS||S)z* Get the value of Retry-After in seconds. zRetry-AfterN)Z	getheaderrKr+responserJrrrget_retry_after	s
zRetry.get_retry_aftercCs ||}|rt|dSdS)NTF)rNrHsleeprLrrrsleep_for_retrys


zRetry.sleep_for_retrycCs"|}|dkrdSt|dS)Nr)rCrHrO)r+Zbackoffrrr_sleep_backoffszRetry._sleep_backoffcCs(|jr|r||}|rdS|dS)aC Sleep between retry attempts.

        This method will respect a server's ``Retry-After`` response header
        and sleep the duration of the time requested. If that is not present, it
        will use an exponential backoff. By default, the backoff factor is 0 and
        this method will return immediately.
        N)r(rPrQ)r+rMZsleptrrrrO!s
	

zRetry.sleepcCs
t|tS)z{ Errors when we're fairly sure that the server did not receive the
        request, so it should be safe to retry.
        )r3rr+errrrr_is_connection_error1szRetry._is_connection_errorcCst|ttfS)z Errors that occur after the request has been started, so we should
        assume that the server began processing it.
        )r3r	rrRrrr_is_read_error7szRetry._is_read_errorcCs|jr||jkrdSdS)z| Checks if a given HTTP method should be retried upon, depending if
        it is included on the method whitelist.
        FT)r"upper)r+rrrr_is_method_retryable=szRetry._is_method_retryableFcCs<||sdS|jr"||jkr"dS|jo:|jo:|o:||jkS)ax Is this method/status code retryable? (Based on whitelists and control
        variables such as the number of total retries to allow, whether to
        respect the Retry-After header, whether this header is present, and
        whether the returned status code is on the list of status codes to
        be retried upon on the presence of the aforementioned header)
        FT)rWr!rr(RETRY_AFTER_STATUS_CODES)r+rstatus_codeZhas_retry_afterrrris_retryFs
zRetry.is_retrycCs:|j|j|j|j|jf}ttd|}|s.dSt|dkS)z Are we out of retries? NFr)rrrrrr?filterrA)r+Zretry_countsrrris_exhaustedZs
zRetry.is_exhaustedcCs|jdkr |r tt||||j}|dk	r6|d8}|j}|j}	|j}
|j}d}d}
d}|r||r|dkrtt|||n|dk	r|d8}n|r|	|r|	dks|
|stt|||n|	dk	r|	d8}	np|r|r|
dk	r|
d8}
d}|}|j}
ns$ 
(site-packages/pip/_vendor/urllib3/util/__pycache__/request.cpython-38.pyc000064400000006372151464452710022412 0ustar00U

.e@sddlmZddlmZddlmZmZddlmZdZ	zddl
ZWnek
rXYn
Xe	d7Z	e
Zdd	d
ZddZd
dZdS))absolute_import)	b64encode)b
integer_types)UnrewindableBodyErrorzgzip,deflateNz,brcCsi}|r6t|trnt|tr*d|}nt}||d<|rB||d<|rNd|d<|rldtt|d|d<|rdtt|d|d	<|rd
|d<|S)a
    Shortcuts for generating request headers.

    :param keep_alive:
        If ``True``, adds 'connection: keep-alive' header.

    :param accept_encoding:
        Can be a boolean, list, or string.
        ``True`` translates to 'gzip,deflate'.
        List will get joined by comma.
        String will be used as provided.

    :param user_agent:
        String representing the user-agent you want, such as
        "python-urllib3/0.6"

    :param basic_auth:
        Colon-separated username:password string for 'authorization: basic ...'
        auth header.

    :param proxy_basic_auth:
        Colon-separated username:password string for 'proxy-authorization: basic ...'
        auth header.

    :param disable_cache:
        If ``True``, adds 'cache-control: no-cache' header.

    Example::

        >>> make_headers(keep_alive=True, user_agent="Batman/1.0")
        {'connection': 'keep-alive', 'user-agent': 'Batman/1.0'}
        >>> make_headers(accept_encoding=True)
        {'accept-encoding': 'gzip,deflate'}
    ,zaccept-encodingz
user-agentz
keep-aliveZ
connectionzBasic zutf-8Z
authorizationzproxy-authorizationzno-cachez
cache-control)
isinstancestrlistjoinACCEPT_ENCODINGrrdecode)Z
keep_aliveZaccept_encodingZ
user_agentZ
basic_authZproxy_basic_authZ
disable_cacheZheadersrD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/request.pymake_headerss.*


rc	CsR|dk	rt||n:t|dddk	rNz|}Wnttfk
rLt}YnX|S)z
    If a position is provided, move file to that point.
    Otherwise, we'll attempt to record a position for future use.
    Ntell)rewind_bodygetattrrIOErrorOSError_FAILEDTELL)bodyposrrrset_file_positionZs
rc	Cstt|dd}|dk	rNt|trNz||Wqpttfk
rJtdYqpXn"|tkr`tdntdt|dS)z
    Attempt to rewind body to a certain position.
    Primarily used for request redirects and retries.

    :param body:
        File-like object that supports seek.

    :param int pos:
        Position to seek to in file.
    seekNzAAn error occurred when rewinding request body for redirect/retry.zRUnable to record file position for rewinding request body during a redirect/retry.z4body_pos must be of type integer, instead it was %s.)	rr	rrrrr
ValueErrortype)rZbody_posZ	body_seekrrrrls
r)NNNNNN)Z
__future__rbase64rZpackages.sixrr
exceptionsrr
ZbrotliZ_unused_module_brotliImportErrorobjectrrrrrrrrs&
Hsite-packages/pip/_vendor/urllib3/util/__pycache__/url.cpython-38.pyc000064400000024752151464452710021526 0ustar00U

.e7	@sddlmZddlZddlmZddlmZddlmZddd	d
ddd
gZ	dZ
edZedZ
edejejBZdZdZdjeedZeedZdddddddddg	Zd Zd!d"d#d$eDd%Zd&ed'Zd(ed!ed)Zd*Zed+Zed,ed-Zed,ed-Zed,ed-Z ed,edd.d-Z!ed/ed0Z"d1eeefZ#ee#ejejBZ$e%d2Z&e%d3Z'e&e'Bd4hBZ(e(d5d6hBZ)e)d7hBZ*Z+Gd8d9d9ed9e	Z,d:d;Z-dKd=d>Z.d?d@Z/dAdBZ0dCdDZ1dEdFZ2dGdHZ3dIdJZ4dS)L)absolute_importN)
namedtuple)LocationParseError)sixschemeauthhostportpathqueryfragment)httpZhttpsNz%[a-fA-F0-9]{2}z^(?:[a-zA-Z][a-zA-Z0-9+-]*:|/)zQ^(?:([a-zA-Z][a-zA-Z0-9+.-]*):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?$z(?:[0-9]{1,3}\.){3}[0-9]{1,3}z[0-9A-Fa-f]{1,4}z(?:{hex}:{hex}|{ipv4}))hexZipv4)rZls32z(?:%(hex)s:){6}%(ls32)sz::(?:%(hex)s:){5}%(ls32)sz%(?:%(hex)s)?::(?:%(hex)s:){4}%(ls32)sz2(?:(?:%(hex)s:)?%(hex)s)?::(?:%(hex)s:){3}%(ls32)sz6(?:(?:%(hex)s:){0,2}%(hex)s)?::(?:%(hex)s:){2}%(ls32)sz/(?:(?:%(hex)s:){0,3}%(hex)s)?::%(hex)s:%(ls32)sz'(?:(?:%(hex)s:){0,4}%(hex)s)?::%(ls32)sz&(?:(?:%(hex)s:){0,5}%(hex)s)?::%(hex)sz(?:(?:%(hex)s:){0,6}%(hex)s)?::zDABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._!\-~z(?:|cCsg|]}|tqS)_subs.0xrr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/url.py
6sr)z
(?:%25|%)(?:[z]|%[a-fA-F0-9]{2})+z\[z)?\]z!(?:[^\[\]%:/?#]|%[a-fA-F0-9]{2})*z"^(/[^?]*)(?:\?([^#]+))?(?:#(.*))?$^$(z)\]$z^(%s|%s|%s)(?::([0-9]{0,5}))?$zBABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._-~z!$&'()*+,;=:@/?cs^eZdZdZdZdfdd	ZeddZedd	Zed
dZ	edd
Z
ddZZS)Urlz
    Data structure for representing an HTTP URL. Used as a return value for
    :func:`parse_url`. Both the scheme and host are normalized as they are
    both case-insensitive according to RFC 3986.
    rNc
sD|r|dsd|}|dk	r&|}tt|||||||||S)Nr)
startswithlowersuperr!__new__)clsrrr	r
rrr
	__class__rrr%[s

zUrl.__new__cCs|jS)z@For backwards-compatibility with urlparse. We're nice like that.)r	selfrrrhostnamemszUrl.hostnamecCs&|jpd}|jdk	r"|d|j7}|S)z)Absolute path including the query string.rNr )rr)r*Zurirrrrequest_urirs

zUrl.request_uricCs|jrd|j|jfS|jS)z(Network location including host and portz%s:%d)r
r	r)rrrnetloc|sz
Url.netlocc	Cs|\}}}}}}}d}|dk	r*||d7}|dk	r>||d7}|dk	rN||7}|dk	rf|dt|7}|dk	rv||7}|dk	r|d|7}|dk	r|d|7}|S)a
        Convert self into a url

        This function should more or less round-trip with :func:`.parse_url`. The
        returned url may not be exactly the same as the url inputted to
        :func:`.parse_url`, but it should be equivalent by the RFC (e.g., urls
        with a blank port will have : removed).

        Example: ::

            >>> U = parse_url('http://google.com/mail/')
            >>> U.url
            'http://google.com/mail/'
            >>> Url('http', 'username:password', 'host.com', 80,
            ... '/path', 'query', 'fragment').url
            'http://username:password@host.com:80/path?query#fragment'
        Nz://rrr #)str)	r*rrr	r
rrr
urlrrrr1s"zUrl.urlcCs|jSN)r1r)rrr__str__szUrl.__str__)NNNNNNN)
__name__
__module____qualname____doc__	__slots__r%propertyr+r,r-r1r3
__classcell__rrr'rr!Rs&

	

'r!cCsvd}d}|D]0}||}|dkr$q|dks4||kr|}|}q|dksN|dkrX|ddfS|d|||dd|fS)a
    .. deprecated:: 1.25

    Given a string and an iterable of delimiters, split on the first found
    delimiter. Return two split parts and the matched delimiter.

    If not found, then the first part is the full input string.

    Example::

        >>> split_first('foo/bar?baz', '?/=')
        ('foo', 'bar?baz', '/')
        >>> split_first('foo/bar?baz', '123')
        ('foo/bar?baz', '', None)

    Scales linearly with number of delims. Not ideal for large number of delims.
    Nrr.)find)sZdelimsZmin_idxZ	min_delimdidxrrrsplit_firsts

r@utf-8cCs|dkr|St|}t|}|D]}|s$|||}q$|dd}t||	dk}t
}tdt|D]p}|||d}	t|	}
|r|	dks|
dkr|	
|kr||	qt|dt|
dddqt|
|S)	zfPercent-encodes a URI component without reapplying
    onto an already percent-encoded component.
    NrA
surrogatepass%rr;r)rensure_text
PERCENT_REfindallisupperreplaceupperencodelencount	bytearrayrangeorddecodeextendrzfill)Z	componentZ
allowed_charsencodingZpercent_encodingsencZ	uri_bytesZis_percent_encodedZencoded_componentiZbyteZbyte_ordrrr_encode_invalid_charss,



*rWcCs|d}g}|D]0}|dkr"qq|dkr6||q|r|q|drf|rZ|drf|dd|drz|dd|S)Nr.z..rr.)z/.z/..)splitappendpopr"insertendswithjoin)rZsegmentsoutputZsegmentrrr_remove_path_dot_segmentss



r`cCs|rt|tjrt|}|tkrt|}|rt|}|r|	d\}}|||}|
drx|dkrx|dd}n|dd}dt|t}|d|
|||dS|
Sn*t|stddd|dDS|S)	Nr;z%25%.cSsg|]}t|qSr)_idna_encode)rZlabelrrrr,sz#_normalize_host..rX)
isinstancerZbinary_type
ensure_strNORMALIZABLE_SCHEMES
IPV6_ADDRZ_REmatch
ZONE_ID_REsearchspanr"rWUNRESERVED_CHARSr#IPV4_REr^rY)r	rZis_ipv6ristartendZzone_idrrr_normalize_hosts*


 

rqcCs|rtdd|Drzddlm}Wn$tk
rJttddYnXz|j|dddWS|j	k
rttd|dYnX|d	S)
NcSsg|]}t|dkqS)rD)rPrrrrr2sz _idna_encode..r)idnaz-Unable to parse URL without the 'idna' moduleT)strictZ
std3_rulesz#Name '%s' is not a valid IDNA labelascii)
anyZpip._vendorrrImportErrorr
raise_fromrrKr#Z	IDNAError)namerrrrrrd1s 


rdcCsl|ds|St|\}}}t|t}t|t}t|t}|dk	rT|d|7}|dk	rh|d|7}|S)zHPercent-encodes a request target so that there are no invalid charactersrNr r/)r"	TARGET_RErigroupsrW
PATH_CHARSQUERY_CHARSFRAGMENT_CHARS)targetrrr
rrr_encode_targetCs



rc	s|s
tS|}t|s d|}zt|\}}}}}|dkpN|tk}|r\|}|r|d\}}	}
|pvd}t	|
\}}|r|rt
|t}|dkrd}n
d\}}}|dk	rt|}d|krdksnt
|t||}|r
|r
t|}t
|t}|r |r t
|t}|r6|r6t
|t}Wn*ttfk
rbtt
|dYSX|s|dk	s~|dk	rd}nd}t|tjrtjntjfdd	}
t|
||
||
|||
||
||
|d
S)a

    Given a url, return a parsed :class:`.Url` namedtuple. Best-effort is
    performed to parse incomplete urls. Fields not provided will be None.
    This parser is RFC 3986 compliant.

    The parser logic and helper functions are based heavily on
    work done in the ``rfc3986`` module.

    :param str url: URL to parse into a :class:`.Url` namedtuple.

    Partly backwards-compatible with :mod:`urlparse`.

    Example::

        >>> parse_url('http://google.com/mail/')
        Url(scheme='http', host='google.com', port=None, path='/mail/', ...)
        >>> parse_url('google.com:80')
        Url(scheme=None, host='google.com', port=80, path=None, ...)
        >>> parse_url('/foo?bar')
        Url(scheme=None, host=None, port=None, path='/foo', query='bar', ...)
    z//Nrr.)NNNrics|dkr|S|Sr2r)rZensure_funcrrensure_typeszparse_url..ensure_type)rrr	r
rrr
)r!	SCHEME_RErkURI_RErirzr#rg
rpartition
_HOST_PORT_RErWUSERINFO_CHARSintrrqr`r{r|r}
ValueErrorAttributeErrorrrwreZ	text_typerErf)r1Z
source_urlrZ	authorityrrr
Z
normalize_urir_Z	host_portr	r
rrrr	parse_urlSsd





rcCst|}|jpd|j|jfS)z4
    Deprecated. Use :func:`parse_url` instead.
    r)rrr+r
)r1prrrget_hostsr)rA)5Z
__future__rrecollectionsr
exceptionsrZpackagesrZ	url_attrsrgcompilerFrUNICODEDOTALLrZIPV4_PATZHEX_PATformatZLS32_PATrZ_variationsZUNRESERVED_PATr^ZIPV6_PATZZONE_ID_PATZIPV6_ADDRZ_PATZREG_NAME_PATryrnZIPV6_RErhZBRACELESS_IPV6_ADDRZ_RErjZ_HOST_PORT_PATrsetrmZSUB_DELIM_CHARSrr{r|r}r!r@rWr`rqrdrrrrrrrsv


	

]#
%asite-packages/pip/_vendor/urllib3/util/__pycache__/ssl_.cpython-38.opt-1.pyc000064400000023135151464452710022615 0ustar00U

.eS7@s(ddlmZddlZddlZddlZddlZddlmZmZddl	m
Z
mZmZddl
mZmZddlmZmZmZdd	lmZdZd
Zd
Zd
Ze
eedZdd
ZeedeZz(ddlZddlmZm Z ddlmZWne!k
rYnXzddlm"Z"e"Z#WnJe!k
rLzddlm#Z"e"Z#Wne!k
rFdZ#Z"YnXYnXzddlm$Z$m%Z%m&Z&Wn"e!k
rd\Z$Z%dZ&YnXd'dddddddddd d!d"d#d$gZ(zdd%lmZWn&e!k
rGd&d'd'e)ZYnXd(d)Z*d*d+Z+d,d-Z,d6d.d/Z-d7d0d1Z.d2d3Z/d4d5Z0dS)8)absolute_importN)hexlify	unhexlify)md5sha1sha256)IPV4_REBRACELESS_IPV6_ADDRZ_RE)SSLErrorInsecurePlatformWarningSNIMissingWarning)sixF) (@cCsDtt|t|}tt|t|D]\}}|||AO}q&|dkS)z
    Compare two digests of equal length in constant time.

    The digests must be of type str/bytes.
    Returns True if the digests match, and False otherwise.
    r)abslenzip	bytearray)abresultlrrA/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/ssl_.py_const_compare_digest_backportsrZcompare_digest)wrap_socket
CERT_REQUIRED)HAS_SNI)PROTOCOL_TLS)PROTOCOL_SSLv23)OP_NO_SSLv2OP_NO_SSLv3OP_NO_COMPRESSION)iii:zECDHE+AESGCMzECDHE+CHACHA20z
DHE+AESGCMzDHE+CHACHA20zECDH+AESGCMz	DH+AESGCMzECDH+AESzDH+AESz
RSA+AESGCMzRSA+AESz!aNULLz!eNULLz!MD5z!DSS)
SSLContextc@s8eZdZddZddZd
ddZdd	ZdddZdS)r(cCs6||_d|_tj|_d|_d|_d|_d|_d|_	dS)NFr)
protocolcheck_hostnamesslZ	CERT_NONEverify_modeca_certsoptionscertfilekeyfileciphers)selfZprotocol_versionrrr__init__kszSSLContext.__init__cCs||_||_dSN)r/r0)r2r/r0rrrload_cert_chainvszSSLContext.load_cert_chainNcCs||_|dk	rtddS)Nz-CA directories not supported in older Pythons)r-r)r2ZcafileZcapathrrrload_verify_locationszsz SSLContext.load_verify_locationscCs
||_dSr4r1)r2Zcipher_suiterrrset_ciphersszSSLContext.set_ciphersFcCs>tdt|j|j|j|j|j|d}t|fd|j	i|S)Na2A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings)r0r/r-	cert_reqsssl_versionserver_sider1)
warningswarnr
r0r/r-r,r)rr1)r2Zsocketserver_hostnamer;kwargsrrrrs
zSSLContext.wrap_socket)NN)NF)__name__
__module____qualname__r3r5r6r8rrrrrr(js

r(cCsn|dd}t|}t|}|s4td|t|}||	}t
||sjtd|t|dS)z
    Checks if given fingerprint matches the supplied certificate.

    :param cert:
        Certificate as bytes object.
    :param fingerprint:
        Fingerprint as string of hexdigits, can be interspersed by colons.
    r'z"Fingerprint of invalid length: {0}z6Fingerprints did not match. Expected "{0}", got "{1}".N)replacelowerrHASHFUNC_MAPgetrformatrencodeZdigest_const_compare_digestr)ZcertZfingerprintZ
digest_lengthZhashfuncZfingerprint_bytesZcert_digestrrrassert_fingerprints


rKcCs@|dkrtSt|traAn HTTPS request has been made, but the SNI (Server Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings)r\r6IOErrorrOSErrorerrnoZENOENThasattrr]_is_key_file_encryptedr5is_ipaddressIS_SECURETRANSPORTr!rr<r=r)
Zsockr0r/r9r-r>r:r1Zssl_contextZca_cert_dirZkey_passwordr[errrssl_wrap_socket(sBrfcCs2tjst|tr|d}tt|p.t|S)zDetects whether the hostname given is an IPv4 or IPv6 address.
    Also detects IPv6 addresses with Zone IDs.

    :param str hostname: Hostname to examine.
    :return: True if the hostname is an IP address, False otherwise.
    ascii)	rZPY2rLbytesdecodeboolr	matchr
)Zhostnamerrrrcs
rcc	Cs<t|d(}|D]}d|krW5QRdSqW5QRXdS)z*Detects if a key file is encrypted or not.rZ	ENCRYPTEDTF)open)Zkey_fileflinerrrrbs
rb)NNNN)
NNNNNNNNNN)1Z
__future__rr`r<ZhmacrYZbinasciirrZhashlibrrrZurlr	r

exceptionsrr
rZpackagesrr(r!ZIS_PYOPENSSLrdrFrrNrJr+rr ImportErrorr"r#r$r%r&joinrXobjectrKrRrSr\rfrcrbrrrrs

.
N
Z
site-packages/pip/_vendor/urllib3/util/__pycache__/connection.cpython-38.pyc000064400000006124151464452710023054 0ustar00U

.e@snddlmZddlZddlmZmZddlmZddZej	ddfd	d
Z
ddZd
dZddZ
e
dZdS))absolute_importN)NoWayToWaitForSocketError
wait_for_read)_appengine_environcCsNt|dd}|dkrdS|dkr$dSzt|ddWStk
rHYdSXdS)a 
    Returns True if the connection is dropped and should be closed.

    :param conn:
        :class:`httplib.HTTPConnection` object.

    Note: For platforms like AppEngine, this will always return ``False`` to
    let the platform handle connection recycling transparently for us.
    sockFNTg)timeout)getattrrr)ZconnrrG/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/connection.pyis_connection_droppeds
r
cCs|\}}|dr|d}d}t}t|||tjD]}|\}	}
}}}
d}zNt|	|
|}t|||tjk	r||||r|	||
|
|WStjk
r}z|}|dk	r|d}W5d}~XYq8Xq8|dk	r|tddS)adConnect to *address* and return the socket object.

    Convenience function.  Connect to *address* (a 2-tuple ``(host,
    port)``) and return the socket object.  Passing the optional
    *timeout* parameter will set the timeout on the socket instance
    before attempting to connect.  If no *timeout* is supplied, the
    global default timeout setting returned by :func:`getdefaulttimeout`
    is used.  If *source_address* is set it must be a tuple of (host, port)
    for the socket to bind as a source address before making the connection.
    An host of '' or port 0 tells the OS to use the default.
    [z[]Nz!getaddrinfo returns an empty list)
startswithstripallowed_gai_familysocketZgetaddrinfoZSOCK_STREAM_set_socket_options_GLOBAL_DEFAULT_TIMEOUTZ
settimeoutbindZconnecterrorclose)Zaddressr	Zsource_addressZsocket_optionshostZporterrfamilyresZafZsocktypeprotoZ	canonnameZsarerrrcreate_connection!s2







rcCs$|dkrdS|D]}|j|qdS)N)Z
setsockopt)rZoptionsoptrrrrYsrcCstj}trtj}|S)zThis function is designed to work in the context of
    getaddrinfo, where family=socket.AF_UNSPEC is the default and
    will perform a DNS search for both IPv6 and IPv4 records.)rZAF_INETHAS_IPV6Z	AF_UNSPEC)rrrrrasrcCsbd}d}trdStjrRz"ttj}||dfd}Wntk
rPYnX|r^||S)z6 Returns True if the system can bind an IPv6 address. NFrT)rZis_appengine_sandboxrhas_ipv6ZAF_INET6r	Exceptionr)rrr!rrr	_has_ipv6lsr#z::1)Z
__future__rrwaitrrZcontribrr
rrrrr#r rrrrs
8site-packages/pip/_vendor/urllib3/util/__pycache__/queue.cpython-38.pyc000064400000002004151464452710022032 0ustar00U

.e@sDddlZddlmZddlmZejr.ddlZGdddejZdS)N)sixqueuec@s0eZdZddZefddZddZddZd	S)
	LifoQueuecCst|_dSN)collectionsdequer)self_rB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/queue.py_initszLifoQueue._initcCs
||jSrr)r
lenrrr
_qsizeszLifoQueue._qsizecCs|j|dSr)rappend)r
itemrrr
_putszLifoQueue._putcCs
|jSr)rpop)r
rrr
_getszLifoQueue._getN)__name__
__module____qualname__rrrrrrrrr
r
sr)	rZpackagesrZpackages.six.movesrZPY2ZQueueZ_unused_module_Queuerrrrr
s
site-packages/pip/_vendor/urllib3/util/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001727151464452710023417 0ustar00U

.e@sddlmZddlmZddlmZddlmZddlm	Z	m
Z
mZmZm
Z
mZmZmZmZddlmZmZddlmZdd	lmZmZmZmZdd
lmZmZdZdS)
)absolute_import)is_connection_dropped)make_headers)is_fp_closed)	
SSLContextHAS_SNIIS_PYOPENSSLIS_SECURETRANSPORTassert_fingerprintresolve_cert_reqsresolve_ssl_versionssl_wrap_socketPROTOCOL_TLS)current_timeTimeout)Retry)get_host	parse_urlsplit_firstUrl)
wait_for_readwait_for_write)rr	r
rrrrrrrrrrrrrr
rrrrN) Z
__future__rZ
connectionrZrequestrZresponserZssl_rrr	r
rrr
rrZtimeoutrrZretryrZurlrrrrwaitrr__all__rrE/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/__init__.pys,site-packages/pip/_vendor/urllib3/util/__pycache__/request.cpython-38.opt-1.pyc000064400000006372151464452710023351 0ustar00U

.e@sddlmZddlmZddlmZmZddlmZdZ	zddl
ZWnek
rXYn
Xe	d7Z	e
Zdd	d
ZddZd
dZdS))absolute_import)	b64encode)b
integer_types)UnrewindableBodyErrorzgzip,deflateNz,brcCsi}|r6t|trnt|tr*d|}nt}||d<|rB||d<|rNd|d<|rldtt|d|d<|rdtt|d|d	<|rd
|d<|S)a
    Shortcuts for generating request headers.

    :param keep_alive:
        If ``True``, adds 'connection: keep-alive' header.

    :param accept_encoding:
        Can be a boolean, list, or string.
        ``True`` translates to 'gzip,deflate'.
        List will get joined by comma.
        String will be used as provided.

    :param user_agent:
        String representing the user-agent you want, such as
        "python-urllib3/0.6"

    :param basic_auth:
        Colon-separated username:password string for 'authorization: basic ...'
        auth header.

    :param proxy_basic_auth:
        Colon-separated username:password string for 'proxy-authorization: basic ...'
        auth header.

    :param disable_cache:
        If ``True``, adds 'cache-control: no-cache' header.

    Example::

        >>> make_headers(keep_alive=True, user_agent="Batman/1.0")
        {'connection': 'keep-alive', 'user-agent': 'Batman/1.0'}
        >>> make_headers(accept_encoding=True)
        {'accept-encoding': 'gzip,deflate'}
    ,zaccept-encodingz
user-agentz
keep-aliveZ
connectionzBasic zutf-8Z
authorizationzproxy-authorizationzno-cachez
cache-control)
isinstancestrlistjoinACCEPT_ENCODINGrrdecode)Z
keep_aliveZaccept_encodingZ
user_agentZ
basic_authZproxy_basic_authZ
disable_cacheZheadersrD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/request.pymake_headerss.*


rc	CsR|dk	rt||n:t|dddk	rNz|}Wnttfk
rLt}YnX|S)z
    If a position is provided, move file to that point.
    Otherwise, we'll attempt to record a position for future use.
    Ntell)rewind_bodygetattrrIOErrorOSError_FAILEDTELL)bodyposrrrset_file_positionZs
rc	Cstt|dd}|dk	rNt|trNz||Wqpttfk
rJtdYqpXn"|tkr`tdntdt|dS)z
    Attempt to rewind body to a certain position.
    Primarily used for request redirects and retries.

    :param body:
        File-like object that supports seek.

    :param int pos:
        Position to seek to in file.
    seekNzAAn error occurred when rewinding request body for redirect/retry.zRUnable to record file position for rewinding request body during a redirect/retry.z4body_pos must be of type integer, instead it was %s.)	rr	rrrrr
ValueErrortype)rZbody_posZ	body_seekrrrrls
r)NNNNNN)Z
__future__rbase64rZpackages.sixrr
exceptionsrr
ZbrotliZ_unused_module_brotliImportErrorobjectrrrrrrrrs&
Hsite-packages/pip/_vendor/urllib3/util/__pycache__/response.cpython-38.opt-1.pyc000064400000003635151464452710023516 0ustar00U

.e
@s@ddlmZddlmZddlmZddZddZd	d
Z	dS))absolute_import)http_client)HeaderParsingErrorcCslz
|WStk
rYnXz|jWStk
r<YnXz|jdkWStk
r^YnXtddS)zt
    Checks whether a given file-like object is closed.

    :param obj:
        The file-like object to check.
    Nz)Unable to determine whether fp is closed.)ZisclosedAttributeErrorclosedfp
ValueError)objrE/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/response.pyis_fp_closeds
r
cCsvt|tjstdt|t|dd}t|dd}d}|r^|s^|}t|tt	fr^|}|sf|rrt
||ddS)aP
    Asserts whether all headers have been successfully parsed.
    Extracts encountered errors from the result of parsing headers.

    Only works on Python 3.

    :param headers: Headers to verify.
    :type headers: `httplib.HTTPMessage`.

    :raises urllib3.exceptions.HeaderParsingError:
        If parsing errors are found.
    z"expected httplib.Message, got {0}.defectsNget_payload)r
unparsed_data)
isinstancehttplibZHTTPMessage	TypeErrorformattypegetattrZis_multipartbytesstrr)ZheadersrrrZpayloadrrrassert_header_parsing&srcCs$|j}t|tr|dkS|dkS)z
    Checks whether the request of a response has been a HEAD-request.
    Handles the quirks of AppEngine.

    :param conn:
    :type conn: :class:`httplib.HTTPResponse`
    ZHEAD)_methodrintupper)Zresponsemethodrrris_response_to_headJs	
rN)
Z
__future__rZpackages.six.movesrr
exceptionsrr
rrrrrrs
$site-packages/pip/_vendor/urllib3/util/__pycache__/response.cpython-38.pyc000064400000003635151464452710022557 0ustar00U

.e
@s@ddlmZddlmZddlmZddZddZd	d
Z	dS))absolute_import)http_client)HeaderParsingErrorcCslz
|WStk
rYnXz|jWStk
r<YnXz|jdkWStk
r^YnXtddS)zt
    Checks whether a given file-like object is closed.

    :param obj:
        The file-like object to check.
    Nz)Unable to determine whether fp is closed.)ZisclosedAttributeErrorclosedfp
ValueError)objrE/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/response.pyis_fp_closeds
r
cCsvt|tjstdt|t|dd}t|dd}d}|r^|s^|}t|tt	fr^|}|sf|rrt
||ddS)aP
    Asserts whether all headers have been successfully parsed.
    Extracts encountered errors from the result of parsing headers.

    Only works on Python 3.

    :param headers: Headers to verify.
    :type headers: `httplib.HTTPMessage`.

    :raises urllib3.exceptions.HeaderParsingError:
        If parsing errors are found.
    z"expected httplib.Message, got {0}.defectsNget_payload)r
unparsed_data)
isinstancehttplibZHTTPMessage	TypeErrorformattypegetattrZis_multipartbytesstrr)ZheadersrrrZpayloadrrrassert_header_parsing&srcCs$|j}t|tr|dkS|dkS)z
    Checks whether the request of a response has been a HEAD-request.
    Handles the quirks of AppEngine.

    :param conn:
    :type conn: :class:`httplib.HTTPResponse`
    ZHEAD)_methodrintupper)Zresponsemethodrrris_response_to_headJs	
rN)
Z
__future__rZpackages.six.movesrr
exceptionsrr
rrrrrrs
$site-packages/pip/_vendor/urllib3/util/__pycache__/timeout.cpython-38.opt-1.pyc000064400000021213151464452710023336 0ustar00U

.e&@sTddlmZddlmZddlZddlmZeZe	edejZ
GdddeZdS)	)absolute_import)_GLOBAL_DEFAULT_TIMEOUTN)TimeoutStateError	monotonicc@steZdZdZeZdeefddZddZe	ddZ
e	d	d
ZddZd
dZ
ddZeddZeddZdS)Timeouta Timeout configuration.

    Timeouts can be defined as a default for a pool::

        timeout = Timeout(connect=2.0, read=7.0)
        http = PoolManager(timeout=timeout)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', timeout=Timeout(10))

    Timeouts can be disabled by setting all the parameters to ``None``::

        no_timeout = Timeout(connect=None, read=None)
        response = http.request('GET', 'http://example.com/, timeout=no_timeout)


    :param total:
        This combines the connect and read timeouts into one; the read timeout
        will be set to the time leftover from the connect attempt. In the
        event that both a connect timeout and a total are specified, or a read
        timeout and a total are specified, the shorter timeout will be applied.

        Defaults to None.

    :type total: integer, float, or None

    :param connect:
        The maximum amount of time (in seconds) to wait for a connection
        attempt to a server to succeed. Omitting the parameter will default the
        connect timeout to the system default, probably `the global default
        timeout in socket.py
        `_.
        None will set an infinite timeout for connection attempts.

    :type connect: integer, float, or None

    :param read:
        The maximum amount of time (in seconds) to wait between consecutive
        read operations for a response from the server. Omitting the parameter
        will default the read timeout to the system default, probably `the
        global default timeout in socket.py
        `_.
        None will set an infinite timeout.

    :type read: integer, float, or None

    .. note::

        Many factors can affect the total amount of time for urllib3 to return
        an HTTP response.

        For example, Python's DNS resolver does not obey the timeout specified
        on the socket. Other factors that can affect total request time include
        high CPU load, high swap, the program running at a low priority level,
        or other behaviors.

        In addition, the read and total timeouts only measure the time between
        read operations on the socket connecting the client and the server,
        not the total amount of time for the request to return a complete
        response. For most requests, the timeout is raised because the server
        has not sent the first byte in the specified time. This is not always
        the case; if a server streams one byte every fifteen seconds, a timeout
        of 20 seconds will not trigger, even though the request will take
        several minutes to complete.

        If your goal is to cut off any request after a set amount of wall clock
        time, consider having a second "watcher" thread to cut off a slow
        request.
    NcCs4||d|_||d|_||d|_d|_dS)Nconnectreadtotal)_validate_timeout_connect_readr
_start_connect)selfr
rr	rD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/timeout.py__init___szTimeout.__init__cCsdt|j|j|j|jfS)Nz!%s(connect=%r, read=%r, total=%r))type__name__rr
r
rrrr__str__eszTimeout.__str__c	Cs|tkr|jS|dks ||jkr$|St|tr6tdzt|Wn(ttfk
rjtd||fYnXz|dkrtd||fWn$tk
rtd||fYnX|S)a Check that a timeout attribute is valid.

        :param value: The timeout value to validate
        :param name: The name of the timeout attribute to validate. This is
            used to specify in error messages.
        :return: The validated and casted version of the given value.
        :raises ValueError: If it is a numeric value less than or equal to
            zero, or the type is not an integer, float, or None.
        NzDTimeout cannot be a boolean value. It must be an int, float or None.z>Timeout value %s was %s, but it must be an int, float or None.rzdAttempted to set %s timeout to %s, but the timeout cannot be set to a value less than or equal to 0.)_DefaultDEFAULT_TIMEOUT
isinstancebool
ValueErrorfloat	TypeError)clsvaluenamerrrrms<


zTimeout._validate_timeoutcCst||dS)a Create a new Timeout from a legacy timeout value.

        The timeout value used by httplib.py sets the same timeout on the
        connect(), and recv() socket requests. This creates a :class:`Timeout`
        object that sets the individual timeouts to the ``timeout`` value
        passed to this function.

        :param timeout: The legacy timeout value.
        :type timeout: integer, float, sentinel default object, or None
        :return: Timeout object
        :rtype: :class:`Timeout`
        )r	r)r)rZtimeoutrrr
from_floatszTimeout.from_floatcCst|j|j|jdS)a Create a copy of the timeout object

        Timeout properties are stored per-pool but each request needs a fresh
        Timeout object to ensure each one has its own start/stop configured.

        :return: a copy of the timeout object
        :rtype: :class:`Timeout`
        )rr	r
)rrr
r
rrrrclonesz
Timeout.clonecCs |jdk	rtdt|_|jS)z Start the timeout clock, used during a connect() attempt

        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to start a timer that has been started already.
        Nz'Timeout timer has already been started.rrcurrent_timerrrr
start_connects
zTimeout.start_connectcCs|jdkrtdt|jS)a Gets the time elapsed since the call to :meth:`start_connect`.

        :return: Elapsed time in seconds.
        :rtype: float
        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to get duration for a timer that hasn't been started.
        Nz:Can't get connect duration for timer that has not started.r#rrrrget_connect_durations

zTimeout.get_connect_durationcCs:|jdkr|jS|jdks&|j|jkr,|jSt|j|jS)a" Get the value to use when setting a connection timeout.

        This will be a positive float or integer, the value None
        (never timeout), or the default system timeout.

        :return: Connect timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        N)r
rrminrrrrconnect_timeouts


zTimeout.connect_timeoutcCs|jdk	rX|j|jk	rX|jdk	rX|j|jk	rX|jdkr<|jStdt|j||jS|jdk	r|j|jk	rtd|j|S|jSdS)a Get the value for the read timeout.

        This assumes some time has elapsed in the connection timeout and
        computes the read timeout appropriately.

        If self.total is set, the read timeout is dependent on the amount of
        time taken by the connect timeout. If the connection time has not been
        established, a :exc:`~urllib3.exceptions.TimeoutStateError` will be
        raised.

        :return: Value to use for the read timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        :raises urllib3.exceptions.TimeoutStateError: If :meth:`start_connect`
            has not yet been called on this object.
        Nr)r
rr
rmaxr'r&rrrrread_timeouts


zTimeout.read_timeout)r
__module____qualname____doc__rrrrrclassmethodrr!r"r%r&propertyr(r*rrrrrsI
-

r)Z
__future__rZsocketrtime
exceptionsrobjectrgetattrr$rrrrrssite-packages/pip/_vendor/urllib3/util/__pycache__/url.cpython-38.opt-1.pyc000064400000024752151464452710022465 0ustar00U

.e7	@sddlmZddlZddlmZddlmZddlmZddd	d
ddd
gZ	dZ
edZedZ
edejejBZdZdZdjeedZeedZdddddddddg	Zd Zd!d"d#d$eDd%Zd&ed'Zd(ed!ed)Zd*Zed+Zed,ed-Zed,ed-Zed,ed-Z ed,edd.d-Z!ed/ed0Z"d1eeefZ#ee#ejejBZ$e%d2Z&e%d3Z'e&e'Bd4hBZ(e(d5d6hBZ)e)d7hBZ*Z+Gd8d9d9ed9e	Z,d:d;Z-dKd=d>Z.d?d@Z/dAdBZ0dCdDZ1dEdFZ2dGdHZ3dIdJZ4dS)L)absolute_importN)
namedtuple)LocationParseError)sixschemeauthhostportpathqueryfragment)httpZhttpsNz%[a-fA-F0-9]{2}z^(?:[a-zA-Z][a-zA-Z0-9+-]*:|/)zQ^(?:([a-zA-Z][a-zA-Z0-9+.-]*):)?(?://([^/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?$z(?:[0-9]{1,3}\.){3}[0-9]{1,3}z[0-9A-Fa-f]{1,4}z(?:{hex}:{hex}|{ipv4}))hexZipv4)rZls32z(?:%(hex)s:){6}%(ls32)sz::(?:%(hex)s:){5}%(ls32)sz%(?:%(hex)s)?::(?:%(hex)s:){4}%(ls32)sz2(?:(?:%(hex)s:)?%(hex)s)?::(?:%(hex)s:){3}%(ls32)sz6(?:(?:%(hex)s:){0,2}%(hex)s)?::(?:%(hex)s:){2}%(ls32)sz/(?:(?:%(hex)s:){0,3}%(hex)s)?::%(hex)s:%(ls32)sz'(?:(?:%(hex)s:){0,4}%(hex)s)?::%(ls32)sz&(?:(?:%(hex)s:){0,5}%(hex)s)?::%(hex)sz(?:(?:%(hex)s:){0,6}%(hex)s)?::zDABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._!\-~z(?:|cCsg|]}|tqS)_subs.0xrr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/util/url.py
6sr)z
(?:%25|%)(?:[z]|%[a-fA-F0-9]{2})+z\[z)?\]z!(?:[^\[\]%:/?#]|%[a-fA-F0-9]{2})*z"^(/[^?]*)(?:\?([^#]+))?(?:#(.*))?$^$(z)\]$z^(%s|%s|%s)(?::([0-9]{0,5}))?$zBABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._-~z!$&'()*+,;=:@/?cs^eZdZdZdZdfdd	ZeddZedd	Zed
dZ	edd
Z
ddZZS)Urlz
    Data structure for representing an HTTP URL. Used as a return value for
    :func:`parse_url`. Both the scheme and host are normalized as they are
    both case-insensitive according to RFC 3986.
    rNc
sD|r|dsd|}|dk	r&|}tt|||||||||S)Nr)
startswithlowersuperr!__new__)clsrrr	r
rrr
	__class__rrr%[s

zUrl.__new__cCs|jS)z@For backwards-compatibility with urlparse. We're nice like that.)r	selfrrrhostnamemszUrl.hostnamecCs&|jpd}|jdk	r"|d|j7}|S)z)Absolute path including the query string.rNr )rr)r*Zurirrrrequest_urirs

zUrl.request_uricCs|jrd|j|jfS|jS)z(Network location including host and portz%s:%d)r
r	r)rrrnetloc|sz
Url.netlocc	Cs|\}}}}}}}d}|dk	r*||d7}|dk	r>||d7}|dk	rN||7}|dk	rf|dt|7}|dk	rv||7}|dk	r|d|7}|dk	r|d|7}|S)a
        Convert self into a url

        This function should more or less round-trip with :func:`.parse_url`. The
        returned url may not be exactly the same as the url inputted to
        :func:`.parse_url`, but it should be equivalent by the RFC (e.g., urls
        with a blank port will have : removed).

        Example: ::

            >>> U = parse_url('http://google.com/mail/')
            >>> U.url
            'http://google.com/mail/'
            >>> Url('http', 'username:password', 'host.com', 80,
            ... '/path', 'query', 'fragment').url
            'http://username:password@host.com:80/path?query#fragment'
        Nz://rrr #)str)	r*rrr	r
rrr
urlrrrr1s"zUrl.urlcCs|jSN)r1r)rrr__str__szUrl.__str__)NNNNNNN)
__name__
__module____qualname____doc__	__slots__r%propertyr+r,r-r1r3
__classcell__rrr'rr!Rs&

	

'r!cCsvd}d}|D]0}||}|dkr$q|dks4||kr|}|}q|dksN|dkrX|ddfS|d|||dd|fS)a
    .. deprecated:: 1.25

    Given a string and an iterable of delimiters, split on the first found
    delimiter. Return two split parts and the matched delimiter.

    If not found, then the first part is the full input string.

    Example::

        >>> split_first('foo/bar?baz', '?/=')
        ('foo', 'bar?baz', '/')
        >>> split_first('foo/bar?baz', '123')
        ('foo/bar?baz', '', None)

    Scales linearly with number of delims. Not ideal for large number of delims.
    Nrr.)find)sZdelimsZmin_idxZ	min_delimdidxrrrsplit_firsts

r@utf-8cCs|dkr|St|}t|}|D]}|s$|||}q$|dd}t||	dk}t
}tdt|D]p}|||d}	t|	}
|r|	dks|
dkr|	
|kr||	qt|dt|
dddqt|
|S)	zfPercent-encodes a URI component without reapplying
    onto an already percent-encoded component.
    NrA
surrogatepass%rr;r)rensure_text
PERCENT_REfindallisupperreplaceupperencodelencount	bytearrayrangeorddecodeextendrzfill)Z	componentZ
allowed_charsencodingZpercent_encodingsencZ	uri_bytesZis_percent_encodedZencoded_componentiZbyteZbyte_ordrrr_encode_invalid_charss,



*rWcCs|d}g}|D]0}|dkr"qq|dkr6||q|r|q|drf|rZ|drf|dd|drz|dd|S)Nr.z..rr.)z/.z/..)splitappendpopr"insertendswithjoin)rZsegmentsoutputZsegmentrrr_remove_path_dot_segmentss



r`cCs|rt|tjrt|}|tkrt|}|rt|}|r|	d\}}|||}|
drx|dkrx|dd}n|dd}dt|t}|d|
|||dS|
Sn*t|stddd|dDS|S)	Nr;z%25%.cSsg|]}t|qSr)_idna_encode)rZlabelrrrr,sz#_normalize_host..rX)
isinstancerZbinary_type
ensure_strNORMALIZABLE_SCHEMES
IPV6_ADDRZ_REmatch
ZONE_ID_REsearchspanr"rWUNRESERVED_CHARSr#IPV4_REr^rY)r	rZis_ipv6ristartendZzone_idrrr_normalize_hosts*


 

rqcCs|rtdd|Drzddlm}Wn$tk
rJttddYnXz|j|dddWS|j	k
rttd|dYnX|d	S)
NcSsg|]}t|dkqS)rD)rPrrrrr2sz _idna_encode..r)idnaz-Unable to parse URL without the 'idna' moduleT)strictZ
std3_rulesz#Name '%s' is not a valid IDNA labelascii)
anyZpip._vendorrrImportErrorr
raise_fromrrKr#Z	IDNAError)namerrrrrrd1s 


rdcCsl|ds|St|\}}}t|t}t|t}t|t}|dk	rT|d|7}|dk	rh|d|7}|S)zHPercent-encodes a request target so that there are no invalid charactersrNr r/)r"	TARGET_RErigroupsrW
PATH_CHARSQUERY_CHARSFRAGMENT_CHARS)targetrrr
rrr_encode_targetCs



rc	s|s
tS|}t|s d|}zt|\}}}}}|dkpN|tk}|r\|}|r|d\}}	}
|pvd}t	|
\}}|r|rt
|t}|dkrd}n
d\}}}|dk	rt|}d|krdksnt
|t||}|r
|r
t|}t
|t}|r |r t
|t}|r6|r6t
|t}Wn*ttfk
rbtt
|dYSX|s|dk	s~|dk	rd}nd}t|tjrtjntjfdd	}
t|
||
||
|||
||
||
|d
S)a

    Given a url, return a parsed :class:`.Url` namedtuple. Best-effort is
    performed to parse incomplete urls. Fields not provided will be None.
    This parser is RFC 3986 compliant.

    The parser logic and helper functions are based heavily on
    work done in the ``rfc3986`` module.

    :param str url: URL to parse into a :class:`.Url` namedtuple.

    Partly backwards-compatible with :mod:`urlparse`.

    Example::

        >>> parse_url('http://google.com/mail/')
        Url(scheme='http', host='google.com', port=None, path='/mail/', ...)
        >>> parse_url('google.com:80')
        Url(scheme=None, host='google.com', port=80, path=None, ...)
        >>> parse_url('/foo?bar')
        Url(scheme=None, host=None, port=None, path='/foo', query='bar', ...)
    z//Nrr.)NNNrics|dkr|S|Sr2r)rZensure_funcrrensure_typeszparse_url..ensure_type)rrr	r
rrr
)r!	SCHEME_RErkURI_RErirzr#rg
rpartition
_HOST_PORT_RErWUSERINFO_CHARSintrrqr`r{r|r}
ValueErrorAttributeErrorrrwreZ	text_typerErf)r1Z
source_urlrZ	authorityrrr
Z
normalize_urir_Z	host_portr	r
rrrr	parse_urlSsd





rcCst|}|jpd|j|jfS)z4
    Deprecated. Use :func:`parse_url` instead.
    r)rrr+r
)r1prrrget_hostsr)rA)5Z
__future__rrecollectionsr
exceptionsrZpackagesrZ	url_attrsrgcompilerFrUNICODEDOTALLrZIPV4_PATZHEX_PATformatZLS32_PATrZ_variationsZUNRESERVED_PATr^ZIPV6_PATZZONE_ID_PATZIPV6_ADDRZ_PATZREG_NAME_PATryrnZIPV6_RErhZBRACELESS_IPV6_ADDRZ_RErjZ_HOST_PORT_PATrsetrmZSUB_DELIM_CHARSrr{r|r}r!r@rWr`rqrdrrrrrrrsv


	

]#
%asite-packages/pip/_vendor/urllib3/util/url.py000064400000033751151464452710015237 0ustar00from __future__ import absolute_import
import re
from collections import namedtuple

from ..exceptions import LocationParseError
from ..packages import six


url_attrs = ["scheme", "auth", "host", "port", "path", "query", "fragment"]

# We only want to normalize urls with an HTTP(S) scheme.
# urllib3 infers URLs without a scheme (None) to be http.
NORMALIZABLE_SCHEMES = ("http", "https", None)

# Almost all of these patterns were derived from the
# 'rfc3986' module: https://github.com/python-hyper/rfc3986
PERCENT_RE = re.compile(r"%[a-fA-F0-9]{2}")
SCHEME_RE = re.compile(r"^(?:[a-zA-Z][a-zA-Z0-9+-]*:|/)")
URI_RE = re.compile(
    r"^(?:([a-zA-Z][a-zA-Z0-9+.-]*):)?"
    r"(?://([^/?#]*))?"
    r"([^?#]*)"
    r"(?:\?([^#]*))?"
    r"(?:#(.*))?$",
    re.UNICODE | re.DOTALL,
)

IPV4_PAT = r"(?:[0-9]{1,3}\.){3}[0-9]{1,3}"
HEX_PAT = "[0-9A-Fa-f]{1,4}"
LS32_PAT = "(?:{hex}:{hex}|{ipv4})".format(hex=HEX_PAT, ipv4=IPV4_PAT)
_subs = {"hex": HEX_PAT, "ls32": LS32_PAT}
_variations = [
    #                            6( h16 ":" ) ls32
    "(?:%(hex)s:){6}%(ls32)s",
    #                       "::" 5( h16 ":" ) ls32
    "::(?:%(hex)s:){5}%(ls32)s",
    # [               h16 ] "::" 4( h16 ":" ) ls32
    "(?:%(hex)s)?::(?:%(hex)s:){4}%(ls32)s",
    # [ *1( h16 ":" ) h16 ] "::" 3( h16 ":" ) ls32
    "(?:(?:%(hex)s:)?%(hex)s)?::(?:%(hex)s:){3}%(ls32)s",
    # [ *2( h16 ":" ) h16 ] "::" 2( h16 ":" ) ls32
    "(?:(?:%(hex)s:){0,2}%(hex)s)?::(?:%(hex)s:){2}%(ls32)s",
    # [ *3( h16 ":" ) h16 ] "::"    h16 ":"   ls32
    "(?:(?:%(hex)s:){0,3}%(hex)s)?::%(hex)s:%(ls32)s",
    # [ *4( h16 ":" ) h16 ] "::"              ls32
    "(?:(?:%(hex)s:){0,4}%(hex)s)?::%(ls32)s",
    # [ *5( h16 ":" ) h16 ] "::"              h16
    "(?:(?:%(hex)s:){0,5}%(hex)s)?::%(hex)s",
    # [ *6( h16 ":" ) h16 ] "::"
    "(?:(?:%(hex)s:){0,6}%(hex)s)?::",
]

UNRESERVED_PAT = r"ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._!\-~"
IPV6_PAT = "(?:" + "|".join([x % _subs for x in _variations]) + ")"
ZONE_ID_PAT = "(?:%25|%)(?:[" + UNRESERVED_PAT + "]|%[a-fA-F0-9]{2})+"
IPV6_ADDRZ_PAT = r"\[" + IPV6_PAT + r"(?:" + ZONE_ID_PAT + r")?\]"
REG_NAME_PAT = r"(?:[^\[\]%:/?#]|%[a-fA-F0-9]{2})*"
TARGET_RE = re.compile(r"^(/[^?]*)(?:\?([^#]+))?(?:#(.*))?$")

IPV4_RE = re.compile("^" + IPV4_PAT + "$")
IPV6_RE = re.compile("^" + IPV6_PAT + "$")
IPV6_ADDRZ_RE = re.compile("^" + IPV6_ADDRZ_PAT + "$")
BRACELESS_IPV6_ADDRZ_RE = re.compile("^" + IPV6_ADDRZ_PAT[2:-2] + "$")
ZONE_ID_RE = re.compile("(" + ZONE_ID_PAT + r")\]$")

_HOST_PORT_PAT = ("^(%s|%s|%s)(?::([0-9]{0,5}))?$") % (
    REG_NAME_PAT,
    IPV4_PAT,
    IPV6_ADDRZ_PAT,
)
_HOST_PORT_RE = re.compile(_HOST_PORT_PAT, re.UNICODE | re.DOTALL)

UNRESERVED_CHARS = set(
    "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._-~"
)
SUB_DELIM_CHARS = set("!$&'()*+,;=")
USERINFO_CHARS = UNRESERVED_CHARS | SUB_DELIM_CHARS | {":"}
PATH_CHARS = USERINFO_CHARS | {"@", "/"}
QUERY_CHARS = FRAGMENT_CHARS = PATH_CHARS | {"?"}


class Url(namedtuple("Url", url_attrs)):
    """
    Data structure for representing an HTTP URL. Used as a return value for
    :func:`parse_url`. Both the scheme and host are normalized as they are
    both case-insensitive according to RFC 3986.
    """

    __slots__ = ()

    def __new__(
        cls,
        scheme=None,
        auth=None,
        host=None,
        port=None,
        path=None,
        query=None,
        fragment=None,
    ):
        if path and not path.startswith("/"):
            path = "/" + path
        if scheme is not None:
            scheme = scheme.lower()
        return super(Url, cls).__new__(
            cls, scheme, auth, host, port, path, query, fragment
        )

    @property
    def hostname(self):
        """For backwards-compatibility with urlparse. We're nice like that."""
        return self.host

    @property
    def request_uri(self):
        """Absolute path including the query string."""
        uri = self.path or "/"

        if self.query is not None:
            uri += "?" + self.query

        return uri

    @property
    def netloc(self):
        """Network location including host and port"""
        if self.port:
            return "%s:%d" % (self.host, self.port)
        return self.host

    @property
    def url(self):
        """
        Convert self into a url

        This function should more or less round-trip with :func:`.parse_url`. The
        returned url may not be exactly the same as the url inputted to
        :func:`.parse_url`, but it should be equivalent by the RFC (e.g., urls
        with a blank port will have : removed).

        Example: ::

            >>> U = parse_url('http://google.com/mail/')
            >>> U.url
            'http://google.com/mail/'
            >>> Url('http', 'username:password', 'host.com', 80,
            ... '/path', 'query', 'fragment').url
            'http://username:password@host.com:80/path?query#fragment'
        """
        scheme, auth, host, port, path, query, fragment = self
        url = u""

        # We use "is not None" we want things to happen with empty strings (or 0 port)
        if scheme is not None:
            url += scheme + u"://"
        if auth is not None:
            url += auth + u"@"
        if host is not None:
            url += host
        if port is not None:
            url += u":" + str(port)
        if path is not None:
            url += path
        if query is not None:
            url += u"?" + query
        if fragment is not None:
            url += u"#" + fragment

        return url

    def __str__(self):
        return self.url


def split_first(s, delims):
    """
    .. deprecated:: 1.25

    Given a string and an iterable of delimiters, split on the first found
    delimiter. Return two split parts and the matched delimiter.

    If not found, then the first part is the full input string.

    Example::

        >>> split_first('foo/bar?baz', '?/=')
        ('foo', 'bar?baz', '/')
        >>> split_first('foo/bar?baz', '123')
        ('foo/bar?baz', '', None)

    Scales linearly with number of delims. Not ideal for large number of delims.
    """
    min_idx = None
    min_delim = None
    for d in delims:
        idx = s.find(d)
        if idx < 0:
            continue

        if min_idx is None or idx < min_idx:
            min_idx = idx
            min_delim = d

    if min_idx is None or min_idx < 0:
        return s, "", None

    return s[:min_idx], s[min_idx + 1 :], min_delim


def _encode_invalid_chars(component, allowed_chars, encoding="utf-8"):
    """Percent-encodes a URI component without reapplying
    onto an already percent-encoded component.
    """
    if component is None:
        return component

    component = six.ensure_text(component)

    # Try to see if the component we're encoding is already percent-encoded
    # so we can skip all '%' characters but still encode all others.
    percent_encodings = PERCENT_RE.findall(component)

    # Normalize existing percent-encoded bytes.
    for enc in percent_encodings:
        if not enc.isupper():
            component = component.replace(enc, enc.upper())

    uri_bytes = component.encode("utf-8", "surrogatepass")
    is_percent_encoded = len(percent_encodings) == uri_bytes.count(b"%")

    encoded_component = bytearray()

    for i in range(0, len(uri_bytes)):
        # Will return a single character bytestring on both Python 2 & 3
        byte = uri_bytes[i : i + 1]
        byte_ord = ord(byte)
        if (is_percent_encoded and byte == b"%") or (
            byte_ord < 128 and byte.decode() in allowed_chars
        ):
            encoded_component.extend(byte)
            continue
        encoded_component.extend(b"%" + (hex(byte_ord)[2:].encode().zfill(2).upper()))

    return encoded_component.decode(encoding)


def _remove_path_dot_segments(path):
    # See http://tools.ietf.org/html/rfc3986#section-5.2.4 for pseudo-code
    segments = path.split("/")  # Turn the path into a list of segments
    output = []  # Initialize the variable to use to store output

    for segment in segments:
        # '.' is the current directory, so ignore it, it is superfluous
        if segment == ".":
            continue
        # Anything other than '..', should be appended to the output
        elif segment != "..":
            output.append(segment)
        # In this case segment == '..', if we can, we should pop the last
        # element
        elif output:
            output.pop()

    # If the path starts with '/' and the output is empty or the first string
    # is non-empty
    if path.startswith("/") and (not output or output[0]):
        output.insert(0, "")

    # If the path starts with '/.' or '/..' ensure we add one more empty
    # string to add a trailing '/'
    if path.endswith(("/.", "/..")):
        output.append("")

    return "/".join(output)


def _normalize_host(host, scheme):
    if host:
        if isinstance(host, six.binary_type):
            host = six.ensure_str(host)

        if scheme in NORMALIZABLE_SCHEMES:
            is_ipv6 = IPV6_ADDRZ_RE.match(host)
            if is_ipv6:
                match = ZONE_ID_RE.search(host)
                if match:
                    start, end = match.span(1)
                    zone_id = host[start:end]

                    if zone_id.startswith("%25") and zone_id != "%25":
                        zone_id = zone_id[3:]
                    else:
                        zone_id = zone_id[1:]
                    zone_id = "%" + _encode_invalid_chars(zone_id, UNRESERVED_CHARS)
                    return host[:start].lower() + zone_id + host[end:]
                else:
                    return host.lower()
            elif not IPV4_RE.match(host):
                return six.ensure_str(
                    b".".join([_idna_encode(label) for label in host.split(".")])
                )
    return host


def _idna_encode(name):
    if name and any([ord(x) > 128 for x in name]):
        try:
            from pip._vendor import idna
        except ImportError:
            six.raise_from(
                LocationParseError("Unable to parse URL without the 'idna' module"),
                None,
            )
        try:
            return idna.encode(name.lower(), strict=True, std3_rules=True)
        except idna.IDNAError:
            six.raise_from(
                LocationParseError(u"Name '%s' is not a valid IDNA label" % name), None
            )
    return name.lower().encode("ascii")


def _encode_target(target):
    """Percent-encodes a request target so that there are no invalid characters"""
    if not target.startswith("/"):
        return target

    path, query, fragment = TARGET_RE.match(target).groups()
    target = _encode_invalid_chars(path, PATH_CHARS)
    query = _encode_invalid_chars(query, QUERY_CHARS)
    fragment = _encode_invalid_chars(fragment, FRAGMENT_CHARS)
    if query is not None:
        target += "?" + query
    if fragment is not None:
        target += "#" + target
    return target


def parse_url(url):
    """
    Given a url, return a parsed :class:`.Url` namedtuple. Best-effort is
    performed to parse incomplete urls. Fields not provided will be None.
    This parser is RFC 3986 compliant.

    The parser logic and helper functions are based heavily on
    work done in the ``rfc3986`` module.

    :param str url: URL to parse into a :class:`.Url` namedtuple.

    Partly backwards-compatible with :mod:`urlparse`.

    Example::

        >>> parse_url('http://google.com/mail/')
        Url(scheme='http', host='google.com', port=None, path='/mail/', ...)
        >>> parse_url('google.com:80')
        Url(scheme=None, host='google.com', port=80, path=None, ...)
        >>> parse_url('/foo?bar')
        Url(scheme=None, host=None, port=None, path='/foo', query='bar', ...)
    """
    if not url:
        # Empty
        return Url()

    source_url = url
    if not SCHEME_RE.search(url):
        url = "//" + url

    try:
        scheme, authority, path, query, fragment = URI_RE.match(url).groups()
        normalize_uri = scheme is None or scheme.lower() in NORMALIZABLE_SCHEMES

        if scheme:
            scheme = scheme.lower()

        if authority:
            auth, _, host_port = authority.rpartition("@")
            auth = auth or None
            host, port = _HOST_PORT_RE.match(host_port).groups()
            if auth and normalize_uri:
                auth = _encode_invalid_chars(auth, USERINFO_CHARS)
            if port == "":
                port = None
        else:
            auth, host, port = None, None, None

        if port is not None:
            port = int(port)
            if not (0 <= port <= 65535):
                raise LocationParseError(url)

        host = _normalize_host(host, scheme)

        if normalize_uri and path:
            path = _remove_path_dot_segments(path)
            path = _encode_invalid_chars(path, PATH_CHARS)
        if normalize_uri and query:
            query = _encode_invalid_chars(query, QUERY_CHARS)
        if normalize_uri and fragment:
            fragment = _encode_invalid_chars(fragment, FRAGMENT_CHARS)

    except (ValueError, AttributeError):
        return six.raise_from(LocationParseError(source_url), None)

    # For the sake of backwards compatibility we put empty
    # string values for path if there are any defined values
    # beyond the path in the URL.
    # TODO: Remove this when we break backwards compatibility.
    if not path:
        if query is not None or fragment is not None:
            path = ""
        else:
            path = None

    # Ensure that each part of the URL is a `str` for
    # backwards compatibility.
    if isinstance(url, six.text_type):
        ensure_func = six.ensure_text
    else:
        ensure_func = six.ensure_str

    def ensure_type(x):
        return x if x is None else ensure_func(x)

    return Url(
        scheme=ensure_type(scheme),
        auth=ensure_type(auth),
        host=ensure_type(host),
        port=port,
        path=ensure_type(path),
        query=ensure_type(query),
        fragment=ensure_type(fragment),
    )


def get_host(url):
    """
    Deprecated. Use :func:`parse_url` instead.
    """
    p = parse_url(url)
    return p.scheme or "http", p.hostname, p.port
site-packages/pip/_vendor/urllib3/util/queue.py000064400000000761151464452710015554 0ustar00import collections
from ..packages import six
from ..packages.six.moves import queue

if six.PY2:
    # Queue is imported for side effects on MS Windows. See issue #229.
    import Queue as _unused_module_Queue  # noqa: F401


class LifoQueue(queue.Queue):
    def _init(self, _):
        self.queue = collections.deque()

    def _qsize(self, len=len):
        return len(self.queue)

    def _put(self, item):
        self.queue.append(item)

    def _get(self):
        return self.queue.pop()
site-packages/pip/_vendor/urllib3/util/wait.py000064400000012436151464452710015376 0ustar00import errno
from functools import partial
import select
import sys

try:
    from time import monotonic
except ImportError:
    from time import time as monotonic

__all__ = ["NoWayToWaitForSocketError", "wait_for_read", "wait_for_write"]


class NoWayToWaitForSocketError(Exception):
    pass


# How should we wait on sockets?
#
# There are two types of APIs you can use for waiting on sockets: the fancy
# modern stateful APIs like epoll/kqueue, and the older stateless APIs like
# select/poll. The stateful APIs are more efficient when you have a lots of
# sockets to keep track of, because you can set them up once and then use them
# lots of times. But we only ever want to wait on a single socket at a time
# and don't want to keep track of state, so the stateless APIs are actually
# more efficient. So we want to use select() or poll().
#
# Now, how do we choose between select() and poll()? On traditional Unixes,
# select() has a strange calling convention that makes it slow, or fail
# altogether, for high-numbered file descriptors. The point of poll() is to fix
# that, so on Unixes, we prefer poll().
#
# On Windows, there is no poll() (or at least Python doesn't provide a wrapper
# for it), but that's OK, because on Windows, select() doesn't have this
# strange calling convention; plain select() works fine.
#
# So: on Windows we use select(), and everywhere else we use poll(). We also
# fall back to select() in case poll() is somehow broken or missing.

if sys.version_info >= (3, 5):
    # Modern Python, that retries syscalls by default
    def _retry_on_intr(fn, timeout):
        return fn(timeout)


else:
    # Old and broken Pythons.
    def _retry_on_intr(fn, timeout):
        if timeout is None:
            deadline = float("inf")
        else:
            deadline = monotonic() + timeout

        while True:
            try:
                return fn(timeout)
            # OSError for 3 <= pyver < 3.5, select.error for pyver <= 2.7
            except (OSError, select.error) as e:
                # 'e.args[0]' incantation works for both OSError and select.error
                if e.args[0] != errno.EINTR:
                    raise
                else:
                    timeout = deadline - monotonic()
                    if timeout < 0:
                        timeout = 0
                    if timeout == float("inf"):
                        timeout = None
                    continue


def select_wait_for_socket(sock, read=False, write=False, timeout=None):
    if not read and not write:
        raise RuntimeError("must specify at least one of read=True, write=True")
    rcheck = []
    wcheck = []
    if read:
        rcheck.append(sock)
    if write:
        wcheck.append(sock)
    # When doing a non-blocking connect, most systems signal success by
    # marking the socket writable. Windows, though, signals success by marked
    # it as "exceptional". We paper over the difference by checking the write
    # sockets for both conditions. (The stdlib selectors module does the same
    # thing.)
    fn = partial(select.select, rcheck, wcheck, wcheck)
    rready, wready, xready = _retry_on_intr(fn, timeout)
    return bool(rready or wready or xready)


def poll_wait_for_socket(sock, read=False, write=False, timeout=None):
    if not read and not write:
        raise RuntimeError("must specify at least one of read=True, write=True")
    mask = 0
    if read:
        mask |= select.POLLIN
    if write:
        mask |= select.POLLOUT
    poll_obj = select.poll()
    poll_obj.register(sock, mask)

    # For some reason, poll() takes timeout in milliseconds
    def do_poll(t):
        if t is not None:
            t *= 1000
        return poll_obj.poll(t)

    return bool(_retry_on_intr(do_poll, timeout))


def null_wait_for_socket(*args, **kwargs):
    raise NoWayToWaitForSocketError("no select-equivalent available")


def _have_working_poll():
    # Apparently some systems have a select.poll that fails as soon as you try
    # to use it, either due to strange configuration or broken monkeypatching
    # from libraries like eventlet/greenlet.
    try:
        poll_obj = select.poll()
        _retry_on_intr(poll_obj.poll, 0)
    except (AttributeError, OSError):
        return False
    else:
        return True


def wait_for_socket(*args, **kwargs):
    # We delay choosing which implementation to use until the first time we're
    # called. We could do it at import time, but then we might make the wrong
    # decision if someone goes wild with monkeypatching select.poll after
    # we're imported.
    global wait_for_socket
    if _have_working_poll():
        wait_for_socket = poll_wait_for_socket
    elif hasattr(select, "select"):
        wait_for_socket = select_wait_for_socket
    else:  # Platform-specific: Appengine.
        wait_for_socket = null_wait_for_socket
    return wait_for_socket(*args, **kwargs)


def wait_for_read(sock, timeout=None):
    """ Waits for reading to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    """
    return wait_for_socket(sock, read=True, timeout=timeout)


def wait_for_write(sock, timeout=None):
    """ Waits for writing to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    """
    return wait_for_socket(sock, write=True, timeout=timeout)
site-packages/pip/_vendor/urllib3/util/timeout.py000064400000023222151464452710016113 0ustar00from __future__ import absolute_import

# The default socket timeout, used by httplib to indicate that no timeout was
# specified by the user
from socket import _GLOBAL_DEFAULT_TIMEOUT
import time

from ..exceptions import TimeoutStateError

# A sentinel value to indicate that no timeout was specified by the user in
# urllib3
_Default = object()


# Use time.monotonic if available.
current_time = getattr(time, "monotonic", time.time)


class Timeout(object):
    """ Timeout configuration.

    Timeouts can be defined as a default for a pool::

        timeout = Timeout(connect=2.0, read=7.0)
        http = PoolManager(timeout=timeout)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', timeout=Timeout(10))

    Timeouts can be disabled by setting all the parameters to ``None``::

        no_timeout = Timeout(connect=None, read=None)
        response = http.request('GET', 'http://example.com/, timeout=no_timeout)


    :param total:
        This combines the connect and read timeouts into one; the read timeout
        will be set to the time leftover from the connect attempt. In the
        event that both a connect timeout and a total are specified, or a read
        timeout and a total are specified, the shorter timeout will be applied.

        Defaults to None.

    :type total: integer, float, or None

    :param connect:
        The maximum amount of time (in seconds) to wait for a connection
        attempt to a server to succeed. Omitting the parameter will default the
        connect timeout to the system default, probably `the global default
        timeout in socket.py
        `_.
        None will set an infinite timeout for connection attempts.

    :type connect: integer, float, or None

    :param read:
        The maximum amount of time (in seconds) to wait between consecutive
        read operations for a response from the server. Omitting the parameter
        will default the read timeout to the system default, probably `the
        global default timeout in socket.py
        `_.
        None will set an infinite timeout.

    :type read: integer, float, or None

    .. note::

        Many factors can affect the total amount of time for urllib3 to return
        an HTTP response.

        For example, Python's DNS resolver does not obey the timeout specified
        on the socket. Other factors that can affect total request time include
        high CPU load, high swap, the program running at a low priority level,
        or other behaviors.

        In addition, the read and total timeouts only measure the time between
        read operations on the socket connecting the client and the server,
        not the total amount of time for the request to return a complete
        response. For most requests, the timeout is raised because the server
        has not sent the first byte in the specified time. This is not always
        the case; if a server streams one byte every fifteen seconds, a timeout
        of 20 seconds will not trigger, even though the request will take
        several minutes to complete.

        If your goal is to cut off any request after a set amount of wall clock
        time, consider having a second "watcher" thread to cut off a slow
        request.
    """

    #: A sentinel object representing the default timeout value
    DEFAULT_TIMEOUT = _GLOBAL_DEFAULT_TIMEOUT

    def __init__(self, total=None, connect=_Default, read=_Default):
        self._connect = self._validate_timeout(connect, "connect")
        self._read = self._validate_timeout(read, "read")
        self.total = self._validate_timeout(total, "total")
        self._start_connect = None

    def __str__(self):
        return "%s(connect=%r, read=%r, total=%r)" % (
            type(self).__name__,
            self._connect,
            self._read,
            self.total,
        )

    @classmethod
    def _validate_timeout(cls, value, name):
        """ Check that a timeout attribute is valid.

        :param value: The timeout value to validate
        :param name: The name of the timeout attribute to validate. This is
            used to specify in error messages.
        :return: The validated and casted version of the given value.
        :raises ValueError: If it is a numeric value less than or equal to
            zero, or the type is not an integer, float, or None.
        """
        if value is _Default:
            return cls.DEFAULT_TIMEOUT

        if value is None or value is cls.DEFAULT_TIMEOUT:
            return value

        if isinstance(value, bool):
            raise ValueError(
                "Timeout cannot be a boolean value. It must "
                "be an int, float or None."
            )
        try:
            float(value)
        except (TypeError, ValueError):
            raise ValueError(
                "Timeout value %s was %s, but it must be an "
                "int, float or None." % (name, value)
            )

        try:
            if value <= 0:
                raise ValueError(
                    "Attempted to set %s timeout to %s, but the "
                    "timeout cannot be set to a value less "
                    "than or equal to 0." % (name, value)
                )
        except TypeError:
            # Python 3
            raise ValueError(
                "Timeout value %s was %s, but it must be an "
                "int, float or None." % (name, value)
            )

        return value

    @classmethod
    def from_float(cls, timeout):
        """ Create a new Timeout from a legacy timeout value.

        The timeout value used by httplib.py sets the same timeout on the
        connect(), and recv() socket requests. This creates a :class:`Timeout`
        object that sets the individual timeouts to the ``timeout`` value
        passed to this function.

        :param timeout: The legacy timeout value.
        :type timeout: integer, float, sentinel default object, or None
        :return: Timeout object
        :rtype: :class:`Timeout`
        """
        return Timeout(read=timeout, connect=timeout)

    def clone(self):
        """ Create a copy of the timeout object

        Timeout properties are stored per-pool but each request needs a fresh
        Timeout object to ensure each one has its own start/stop configured.

        :return: a copy of the timeout object
        :rtype: :class:`Timeout`
        """
        # We can't use copy.deepcopy because that will also create a new object
        # for _GLOBAL_DEFAULT_TIMEOUT, which socket.py uses as a sentinel to
        # detect the user default.
        return Timeout(connect=self._connect, read=self._read, total=self.total)

    def start_connect(self):
        """ Start the timeout clock, used during a connect() attempt

        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to start a timer that has been started already.
        """
        if self._start_connect is not None:
            raise TimeoutStateError("Timeout timer has already been started.")
        self._start_connect = current_time()
        return self._start_connect

    def get_connect_duration(self):
        """ Gets the time elapsed since the call to :meth:`start_connect`.

        :return: Elapsed time in seconds.
        :rtype: float
        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to get duration for a timer that hasn't been started.
        """
        if self._start_connect is None:
            raise TimeoutStateError(
                "Can't get connect duration for timer " "that has not started."
            )
        return current_time() - self._start_connect

    @property
    def connect_timeout(self):
        """ Get the value to use when setting a connection timeout.

        This will be a positive float or integer, the value None
        (never timeout), or the default system timeout.

        :return: Connect timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        """
        if self.total is None:
            return self._connect

        if self._connect is None or self._connect is self.DEFAULT_TIMEOUT:
            return self.total

        return min(self._connect, self.total)

    @property
    def read_timeout(self):
        """ Get the value for the read timeout.

        This assumes some time has elapsed in the connection timeout and
        computes the read timeout appropriately.

        If self.total is set, the read timeout is dependent on the amount of
        time taken by the connect timeout. If the connection time has not been
        established, a :exc:`~urllib3.exceptions.TimeoutStateError` will be
        raised.

        :return: Value to use for the read timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        :raises urllib3.exceptions.TimeoutStateError: If :meth:`start_connect`
            has not yet been called on this object.
        """
        if (
            self.total is not None
            and self.total is not self.DEFAULT_TIMEOUT
            and self._read is not None
            and self._read is not self.DEFAULT_TIMEOUT
        ):
            # In case the connect timeout has not yet been established.
            if self._start_connect is None:
                return self._read
            return max(0, min(self.total - self.get_connect_duration(), self._read))
        elif self.total is not None and self.total is not self.DEFAULT_TIMEOUT:
            return max(0, self.total - self.get_connect_duration())
        else:
            return self._read
site-packages/pip/_vendor/urllib3/util/__init__.py000064400000002016151464452710016162 0ustar00from __future__ import absolute_import

# For backwards compatibility, provide imports that used to be here.
from .connection import is_connection_dropped
from .request import make_headers
from .response import is_fp_closed
from .ssl_ import (
    SSLContext,
    HAS_SNI,
    IS_PYOPENSSL,
    IS_SECURETRANSPORT,
    assert_fingerprint,
    resolve_cert_reqs,
    resolve_ssl_version,
    ssl_wrap_socket,
    PROTOCOL_TLS,
)
from .timeout import current_time, Timeout

from .retry import Retry
from .url import get_host, parse_url, split_first, Url
from .wait import wait_for_read, wait_for_write

__all__ = (
    "HAS_SNI",
    "IS_PYOPENSSL",
    "IS_SECURETRANSPORT",
    "SSLContext",
    "PROTOCOL_TLS",
    "Retry",
    "Timeout",
    "Url",
    "assert_fingerprint",
    "current_time",
    "is_connection_dropped",
    "is_fp_closed",
    "get_host",
    "parse_url",
    "make_headers",
    "resolve_cert_reqs",
    "resolve_ssl_version",
    "split_first",
    "ssl_wrap_socket",
    "wait_for_read",
    "wait_for_write",
)
site-packages/pip/_vendor/urllib3/connectionpool.py000064400000107136151464452710016510 0ustar00from __future__ import absolute_import
import errno
import logging
import sys
import warnings

from socket import error as SocketError, timeout as SocketTimeout
import socket


from .exceptions import (
    ClosedPoolError,
    ProtocolError,
    EmptyPoolError,
    HeaderParsingError,
    HostChangedError,
    LocationValueError,
    MaxRetryError,
    ProxyError,
    ReadTimeoutError,
    SSLError,
    TimeoutError,
    InsecureRequestWarning,
    NewConnectionError,
)
from .packages.ssl_match_hostname import CertificateError
from .packages import six
from .packages.six.moves import queue
from .connection import (
    port_by_scheme,
    DummyConnection,
    HTTPConnection,
    HTTPSConnection,
    VerifiedHTTPSConnection,
    HTTPException,
    BaseSSLError,
)
from .request import RequestMethods
from .response import HTTPResponse

from .util.connection import is_connection_dropped
from .util.request import set_file_position
from .util.response import assert_header_parsing
from .util.retry import Retry
from .util.timeout import Timeout
from .util.url import (
    get_host,
    parse_url,
    Url,
    _normalize_host as normalize_host,
    _encode_target,
)
from .util.queue import LifoQueue


xrange = six.moves.xrange

log = logging.getLogger(__name__)

_Default = object()


# Pool objects
class ConnectionPool(object):
    """
    Base class for all connection pools, such as
    :class:`.HTTPConnectionPool` and :class:`.HTTPSConnectionPool`.
    """

    scheme = None
    QueueCls = LifoQueue

    def __init__(self, host, port=None):
        if not host:
            raise LocationValueError("No host specified.")

        self.host = _normalize_host(host, scheme=self.scheme)
        self._proxy_host = host.lower()
        self.port = port

    def __str__(self):
        return "%s(host=%r, port=%r)" % (type(self).__name__, self.host, self.port)

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        self.close()
        # Return False to re-raise any potential exceptions
        return False

    def close(self):
        """
        Close all pooled connections and disable the pool.
        """
        pass


# This is taken from http://hg.python.org/cpython/file/7aaba721ebc0/Lib/socket.py#l252
_blocking_errnos = {errno.EAGAIN, errno.EWOULDBLOCK}


class HTTPConnectionPool(ConnectionPool, RequestMethods):
    """
    Thread-safe connection pool for one host.

    :param host:
        Host used for this HTTP Connection (e.g. "localhost"), passed into
        :class:`httplib.HTTPConnection`.

    :param port:
        Port used for this HTTP Connection (None is equivalent to 80), passed
        into :class:`httplib.HTTPConnection`.

    :param strict:
        Causes BadStatusLine to be raised if the status line can't be parsed
        as a valid HTTP/1.0 or 1.1 status line, passed into
        :class:`httplib.HTTPConnection`.

        .. note::
           Only works in Python 2. This parameter is ignored in Python 3.

    :param timeout:
        Socket timeout in seconds for each individual connection. This can
        be a float or integer, which sets the timeout for the HTTP request,
        or an instance of :class:`urllib3.util.Timeout` which gives you more
        fine-grained control over request timeouts. After the constructor has
        been parsed, this is always a `urllib3.util.Timeout` object.

    :param maxsize:
        Number of connections to save that can be reused. More than 1 is useful
        in multithreaded situations. If ``block`` is set to False, more
        connections will be created but they will not be saved once they've
        been used.

    :param block:
        If set to True, no more than ``maxsize`` connections will be used at
        a time. When no free connections are available, the call will block
        until a connection has been released. This is a useful side effect for
        particular multithreaded situations where one does not want to use more
        than maxsize connections per host to prevent flooding.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param retries:
        Retry configuration to use by default with requests in this pool.

    :param _proxy:
        Parsed proxy URL, should not be used directly, instead, see
        :class:`urllib3.connectionpool.ProxyManager`"

    :param _proxy_headers:
        A dictionary with proxy headers, should not be used directly,
        instead, see :class:`urllib3.connectionpool.ProxyManager`"

    :param \\**conn_kw:
        Additional parameters are used to create fresh :class:`urllib3.connection.HTTPConnection`,
        :class:`urllib3.connection.HTTPSConnection` instances.
    """

    scheme = "http"
    ConnectionCls = HTTPConnection
    ResponseCls = HTTPResponse

    def __init__(
        self,
        host,
        port=None,
        strict=False,
        timeout=Timeout.DEFAULT_TIMEOUT,
        maxsize=1,
        block=False,
        headers=None,
        retries=None,
        _proxy=None,
        _proxy_headers=None,
        **conn_kw
    ):
        ConnectionPool.__init__(self, host, port)
        RequestMethods.__init__(self, headers)

        self.strict = strict

        if not isinstance(timeout, Timeout):
            timeout = Timeout.from_float(timeout)

        if retries is None:
            retries = Retry.DEFAULT

        self.timeout = timeout
        self.retries = retries

        self.pool = self.QueueCls(maxsize)
        self.block = block

        self.proxy = _proxy
        self.proxy_headers = _proxy_headers or {}

        # Fill the queue up so that doing get() on it will block properly
        for _ in xrange(maxsize):
            self.pool.put(None)

        # These are mostly for testing and debugging purposes.
        self.num_connections = 0
        self.num_requests = 0
        self.conn_kw = conn_kw

        if self.proxy:
            # Enable Nagle's algorithm for proxies, to avoid packet fragmentation.
            # We cannot know if the user has added default socket options, so we cannot replace the
            # list.
            self.conn_kw.setdefault("socket_options", [])

    def _new_conn(self):
        """
        Return a fresh :class:`HTTPConnection`.
        """
        self.num_connections += 1
        log.debug(
            "Starting new HTTP connection (%d): %s:%s",
            self.num_connections,
            self.host,
            self.port or "80",
        )

        conn = self.ConnectionCls(
            host=self.host,
            port=self.port,
            timeout=self.timeout.connect_timeout,
            strict=self.strict,
            **self.conn_kw
        )
        return conn

    def _get_conn(self, timeout=None):
        """
        Get a connection. Will return a pooled connection if one is available.

        If no connections are available and :prop:`.block` is ``False``, then a
        fresh connection is returned.

        :param timeout:
            Seconds to wait before giving up and raising
            :class:`urllib3.exceptions.EmptyPoolError` if the pool is empty and
            :prop:`.block` is ``True``.
        """
        conn = None
        try:
            conn = self.pool.get(block=self.block, timeout=timeout)

        except AttributeError:  # self.pool is None
            raise ClosedPoolError(self, "Pool is closed.")

        except queue.Empty:
            if self.block:
                raise EmptyPoolError(
                    self,
                    "Pool reached maximum size and no more " "connections are allowed.",
                )
            pass  # Oh well, we'll create a new connection then

        # If this is a persistent connection, check if it got disconnected
        if conn and is_connection_dropped(conn):
            log.debug("Resetting dropped connection: %s", self.host)
            conn.close()
            if getattr(conn, "auto_open", 1) == 0:
                # This is a proxied connection that has been mutated by
                # httplib._tunnel() and cannot be reused (since it would
                # attempt to bypass the proxy)
                conn = None

        return conn or self._new_conn()

    def _put_conn(self, conn):
        """
        Put a connection back into the pool.

        :param conn:
            Connection object for the current host and port as returned by
            :meth:`._new_conn` or :meth:`._get_conn`.

        If the pool is already full, the connection is closed and discarded
        because we exceeded maxsize. If connections are discarded frequently,
        then maxsize should be increased.

        If the pool is closed, then the connection will be closed and discarded.
        """
        try:
            self.pool.put(conn, block=False)
            return  # Everything is dandy, done.
        except AttributeError:
            # self.pool is None.
            pass
        except queue.Full:
            # This should never happen if self.block == True
            log.warning("Connection pool is full, discarding connection: %s", self.host)

        # Connection never got put back into the pool, close it.
        if conn:
            conn.close()

    def _validate_conn(self, conn):
        """
        Called right before a request is made, after the socket is created.
        """
        pass

    def _prepare_proxy(self, conn):
        # Nothing to do for HTTP connections.
        pass

    def _get_timeout(self, timeout):
        """ Helper that always returns a :class:`urllib3.util.Timeout` """
        if timeout is _Default:
            return self.timeout.clone()

        if isinstance(timeout, Timeout):
            return timeout.clone()
        else:
            # User passed us an int/float. This is for backwards compatibility,
            # can be removed later
            return Timeout.from_float(timeout)

    def _raise_timeout(self, err, url, timeout_value):
        """Is the error actually a timeout? Will raise a ReadTimeout or pass"""

        if isinstance(err, SocketTimeout):
            raise ReadTimeoutError(
                self, url, "Read timed out. (read timeout=%s)" % timeout_value
            )

        # See the above comment about EAGAIN in Python 3. In Python 2 we have
        # to specifically catch it and throw the timeout error
        if hasattr(err, "errno") and err.errno in _blocking_errnos:
            raise ReadTimeoutError(
                self, url, "Read timed out. (read timeout=%s)" % timeout_value
            )

        # Catch possible read timeouts thrown as SSL errors. If not the
        # case, rethrow the original. We need to do this because of:
        # http://bugs.python.org/issue10272
        if "timed out" in str(err) or "did not complete (read)" in str(
            err
        ):  # Python < 2.7.4
            raise ReadTimeoutError(
                self, url, "Read timed out. (read timeout=%s)" % timeout_value
            )

    def _make_request(
        self, conn, method, url, timeout=_Default, chunked=False, **httplib_request_kw
    ):
        """
        Perform a request on a given urllib connection object taken from our
        pool.

        :param conn:
            a connection from one of our connection pools

        :param timeout:
            Socket timeout in seconds for the request. This can be a
            float or integer, which will set the same timeout value for
            the socket connect and the socket read, or an instance of
            :class:`urllib3.util.Timeout`, which gives you more fine-grained
            control over your timeouts.
        """
        self.num_requests += 1

        timeout_obj = self._get_timeout(timeout)
        timeout_obj.start_connect()
        conn.timeout = timeout_obj.connect_timeout

        # Trigger any extra validation we need to do.
        try:
            self._validate_conn(conn)
        except (SocketTimeout, BaseSSLError) as e:
            # Py2 raises this as a BaseSSLError, Py3 raises it as socket timeout.
            self._raise_timeout(err=e, url=url, timeout_value=conn.timeout)
            raise

        # conn.request() calls httplib.*.request, not the method in
        # urllib3.request. It also calls makefile (recv) on the socket.
        if chunked:
            conn.request_chunked(method, url, **httplib_request_kw)
        else:
            conn.request(method, url, **httplib_request_kw)

        # Reset the timeout for the recv() on the socket
        read_timeout = timeout_obj.read_timeout

        # App Engine doesn't have a sock attr
        if getattr(conn, "sock", None):
            # In Python 3 socket.py will catch EAGAIN and return None when you
            # try and read into the file pointer created by http.client, which
            # instead raises a BadStatusLine exception. Instead of catching
            # the exception and assuming all BadStatusLine exceptions are read
            # timeouts, check for a zero timeout before making the request.
            if read_timeout == 0:
                raise ReadTimeoutError(
                    self, url, "Read timed out. (read timeout=%s)" % read_timeout
                )
            if read_timeout is Timeout.DEFAULT_TIMEOUT:
                conn.sock.settimeout(socket.getdefaulttimeout())
            else:  # None or a value
                conn.sock.settimeout(read_timeout)

        # Receive the response from the server
        try:
            try:
                # Python 2.7, use buffering of HTTP responses
                httplib_response = conn.getresponse(buffering=True)
            except TypeError:
                # Python 3
                try:
                    httplib_response = conn.getresponse()
                except BaseException as e:
                    # Remove the TypeError from the exception chain in
                    # Python 3 (including for exceptions like SystemExit).
                    # Otherwise it looks like a bug in the code.
                    six.raise_from(e, None)
        except (SocketTimeout, BaseSSLError, SocketError) as e:
            self._raise_timeout(err=e, url=url, timeout_value=read_timeout)
            raise

        # AppEngine doesn't have a version attr.
        http_version = getattr(conn, "_http_vsn_str", "HTTP/?")
        log.debug(
            '%s://%s:%s "%s %s %s" %s %s',
            self.scheme,
            self.host,
            self.port,
            method,
            url,
            http_version,
            httplib_response.status,
            httplib_response.length,
        )

        try:
            assert_header_parsing(httplib_response.msg)
        except (HeaderParsingError, TypeError) as hpe:  # Platform-specific: Python 3
            log.warning(
                "Failed to parse headers (url=%s): %s",
                self._absolute_url(url),
                hpe,
                exc_info=True,
            )

        return httplib_response

    def _absolute_url(self, path):
        return Url(scheme=self.scheme, host=self.host, port=self.port, path=path).url

    def close(self):
        """
        Close all pooled connections and disable the pool.
        """
        if self.pool is None:
            return
        # Disable access to the pool
        old_pool, self.pool = self.pool, None

        try:
            while True:
                conn = old_pool.get(block=False)
                if conn:
                    conn.close()

        except queue.Empty:
            pass  # Done.

    def is_same_host(self, url):
        """
        Check if the given ``url`` is a member of the same host as this
        connection pool.
        """
        if url.startswith("/"):
            return True

        # TODO: Add optional support for socket.gethostbyname checking.
        scheme, host, port = get_host(url)
        if host is not None:
            host = _normalize_host(host, scheme=scheme)

        # Use explicit default port for comparison when none is given
        if self.port and not port:
            port = port_by_scheme.get(scheme)
        elif not self.port and port == port_by_scheme.get(scheme):
            port = None

        return (scheme, host, port) == (self.scheme, self.host, self.port)

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        retries=None,
        redirect=True,
        assert_same_host=True,
        timeout=_Default,
        pool_timeout=None,
        release_conn=None,
        chunked=False,
        body_pos=None,
        **response_kw
    ):
        """
        Get a connection from the pool and perform an HTTP request. This is the
        lowest level call for making a request, so you'll need to specify all
        the raw details.

        .. note::

           More commonly, it's appropriate to use a convenience method provided
           by :class:`.RequestMethods`, such as :meth:`request`.

        .. note::

           `release_conn` will only behave as expected if
           `preload_content=False` because we want to make
           `preload_content=False` the default behaviour someday soon without
           breaking backwards compatibility.

        :param method:
            HTTP request method (such as GET, POST, PUT, etc.)

        :param body:
            Data to send in the request body (useful for creating
            POST requests, see HTTPConnectionPool.post_url for
            more convenience).

        :param headers:
            Dictionary of custom headers to send, such as User-Agent,
            If-None-Match, etc. If None, pool headers are used. If provided,
            these headers completely replace any pool-specific headers.

        :param retries:
            Configure the number of retries to allow before raising a
            :class:`~urllib3.exceptions.MaxRetryError` exception.

            Pass ``None`` to retry until you receive a response. Pass a
            :class:`~urllib3.util.retry.Retry` object for fine-grained control
            over different types of retries.
            Pass an integer number to retry connection errors that many times,
            but no other types of errors. Pass zero to never retry.

            If ``False``, then retries are disabled and any exception is raised
            immediately. Also, instead of raising a MaxRetryError on redirects,
            the redirect response will be returned.

        :type retries: :class:`~urllib3.util.retry.Retry`, False, or an int.

        :param redirect:
            If True, automatically handle redirects (status codes 301, 302,
            303, 307, 308). Each redirect counts as a retry. Disabling retries
            will disable redirect, too.

        :param assert_same_host:
            If ``True``, will make sure that the host of the pool requests is
            consistent else will raise HostChangedError. When False, you can
            use the pool on an HTTP proxy and request foreign hosts.

        :param timeout:
            If specified, overrides the default timeout for this one
            request. It may be a float (in seconds) or an instance of
            :class:`urllib3.util.Timeout`.

        :param pool_timeout:
            If set and the pool is set to block=True, then this method will
            block for ``pool_timeout`` seconds and raise EmptyPoolError if no
            connection is available within the time period.

        :param release_conn:
            If False, then the urlopen call will not release the connection
            back into the pool once a response is received (but will release if
            you read the entire contents of the response such as when
            `preload_content=True`). This is useful if you're not preloading
            the response's content immediately. You will need to call
            ``r.release_conn()`` on the response ``r`` to return the connection
            back into the pool. If None, it takes the value of
            ``response_kw.get('preload_content', True)``.

        :param chunked:
            If True, urllib3 will send the body using chunked transfer
            encoding. Otherwise, urllib3 will send the body using the standard
            content-length form. Defaults to False.

        :param int body_pos:
            Position to seek to in file-like body in the event of a retry or
            redirect. Typically this won't need to be set because urllib3 will
            auto-populate the value when needed.

        :param \\**response_kw:
            Additional parameters are passed to
            :meth:`urllib3.response.HTTPResponse.from_httplib`
        """
        if headers is None:
            headers = self.headers

        if not isinstance(retries, Retry):
            retries = Retry.from_int(retries, redirect=redirect, default=self.retries)

        if release_conn is None:
            release_conn = response_kw.get("preload_content", True)

        # Check host
        if assert_same_host and not self.is_same_host(url):
            raise HostChangedError(self, url, retries)

        # Ensure that the URL we're connecting to is properly encoded
        if url.startswith("/"):
            url = six.ensure_str(_encode_target(url))
        else:
            url = six.ensure_str(parse_url(url).url)

        conn = None

        # Track whether `conn` needs to be released before
        # returning/raising/recursing. Update this variable if necessary, and
        # leave `release_conn` constant throughout the function. That way, if
        # the function recurses, the original value of `release_conn` will be
        # passed down into the recursive call, and its value will be respected.
        #
        # See issue #651 [1] for details.
        #
        # [1] 
        release_this_conn = release_conn

        # Merge the proxy headers. Only do this in HTTP. We have to copy the
        # headers dict so we can safely change it without those changes being
        # reflected in anyone else's copy.
        if self.scheme == "http":
            headers = headers.copy()
            headers.update(self.proxy_headers)

        # Must keep the exception bound to a separate variable or else Python 3
        # complains about UnboundLocalError.
        err = None

        # Keep track of whether we cleanly exited the except block. This
        # ensures we do proper cleanup in finally.
        clean_exit = False

        # Rewind body position, if needed. Record current position
        # for future rewinds in the event of a redirect/retry.
        body_pos = set_file_position(body, body_pos)

        try:
            # Request a connection from the queue.
            timeout_obj = self._get_timeout(timeout)
            conn = self._get_conn(timeout=pool_timeout)

            conn.timeout = timeout_obj.connect_timeout

            is_new_proxy_conn = self.proxy is not None and not getattr(
                conn, "sock", None
            )
            if is_new_proxy_conn:
                self._prepare_proxy(conn)

            # Make the request on the httplib connection object.
            httplib_response = self._make_request(
                conn,
                method,
                url,
                timeout=timeout_obj,
                body=body,
                headers=headers,
                chunked=chunked,
            )

            # If we're going to release the connection in ``finally:``, then
            # the response doesn't need to know about the connection. Otherwise
            # it will also try to release it and we'll have a double-release
            # mess.
            response_conn = conn if not release_conn else None

            # Pass method to Response for length checking
            response_kw["request_method"] = method

            # Import httplib's response into our own wrapper object
            response = self.ResponseCls.from_httplib(
                httplib_response,
                pool=self,
                connection=response_conn,
                retries=retries,
                **response_kw
            )

            # Everything went great!
            clean_exit = True

        except queue.Empty:
            # Timed out by queue.
            raise EmptyPoolError(self, "No pool connections are available.")

        except (
            TimeoutError,
            HTTPException,
            SocketError,
            ProtocolError,
            BaseSSLError,
            SSLError,
            CertificateError,
        ) as e:
            # Discard the connection for these exceptions. It will be
            # replaced during the next _get_conn() call.
            clean_exit = False
            if isinstance(e, (BaseSSLError, CertificateError)):
                e = SSLError(e)
            elif isinstance(e, (SocketError, NewConnectionError)) and self.proxy:
                e = ProxyError("Cannot connect to proxy.", e)
            elif isinstance(e, (SocketError, HTTPException)):
                e = ProtocolError("Connection aborted.", e)

            retries = retries.increment(
                method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
            )
            retries.sleep()

            # Keep track of the error for the retry warning.
            err = e

        finally:
            if not clean_exit:
                # We hit some kind of exception, handled or otherwise. We need
                # to throw the connection away unless explicitly told not to.
                # Close the connection, set the variable to None, and make sure
                # we put the None back in the pool to avoid leaking it.
                conn = conn and conn.close()
                release_this_conn = True

            if release_this_conn:
                # Put the connection back to be reused. If the connection is
                # expired then it will be None, which will get replaced with a
                # fresh connection during _get_conn.
                self._put_conn(conn)

        if not conn:
            # Try again
            log.warning(
                "Retrying (%r) after connection " "broken by '%r': %s",
                retries,
                err,
                url,
            )
            return self.urlopen(
                method,
                url,
                body,
                headers,
                retries,
                redirect,
                assert_same_host,
                timeout=timeout,
                pool_timeout=pool_timeout,
                release_conn=release_conn,
                body_pos=body_pos,
                **response_kw
            )

        def drain_and_release_conn(response):
            try:
                # discard any remaining response body, the connection will be
                # released back to the pool once the entire response is read
                response.read()
            except (
                TimeoutError,
                HTTPException,
                SocketError,
                ProtocolError,
                BaseSSLError,
                SSLError,
            ):
                pass

        # Handle redirect?
        redirect_location = redirect and response.get_redirect_location()
        if redirect_location:
            if response.status == 303:
                method = "GET"

            try:
                retries = retries.increment(method, url, response=response, _pool=self)
            except MaxRetryError:
                if retries.raise_on_redirect:
                    # Drain and release the connection for this response, since
                    # we're not returning it to be released manually.
                    drain_and_release_conn(response)
                    raise
                return response

            # drain and return the connection to the pool before recursing
            drain_and_release_conn(response)

            retries.sleep_for_retry(response)
            log.debug("Redirecting %s -> %s", url, redirect_location)
            return self.urlopen(
                method,
                redirect_location,
                body,
                headers,
                retries=retries,
                redirect=redirect,
                assert_same_host=assert_same_host,
                timeout=timeout,
                pool_timeout=pool_timeout,
                release_conn=release_conn,
                body_pos=body_pos,
                **response_kw
            )

        # Check if we should retry the HTTP response.
        has_retry_after = bool(response.getheader("Retry-After"))
        if retries.is_retry(method, response.status, has_retry_after):
            try:
                retries = retries.increment(method, url, response=response, _pool=self)
            except MaxRetryError:
                if retries.raise_on_status:
                    # Drain and release the connection for this response, since
                    # we're not returning it to be released manually.
                    drain_and_release_conn(response)
                    raise
                return response

            # drain and return the connection to the pool before recursing
            drain_and_release_conn(response)

            retries.sleep(response)
            log.debug("Retry: %s", url)
            return self.urlopen(
                method,
                url,
                body,
                headers,
                retries=retries,
                redirect=redirect,
                assert_same_host=assert_same_host,
                timeout=timeout,
                pool_timeout=pool_timeout,
                release_conn=release_conn,
                body_pos=body_pos,
                **response_kw
            )

        return response


class HTTPSConnectionPool(HTTPConnectionPool):
    """
    Same as :class:`.HTTPConnectionPool`, but HTTPS.

    When Python is compiled with the :mod:`ssl` module, then
    :class:`.VerifiedHTTPSConnection` is used, which *can* verify certificates,
    instead of :class:`.HTTPSConnection`.

    :class:`.VerifiedHTTPSConnection` uses one of ``assert_fingerprint``,
    ``assert_hostname`` and ``host`` in this order to verify connections.
    If ``assert_hostname`` is False, no verification is done.

    The ``key_file``, ``cert_file``, ``cert_reqs``, ``ca_certs``,
    ``ca_cert_dir``, ``ssl_version``, ``key_password`` are only used if :mod:`ssl`
    is available and are fed into :meth:`urllib3.util.ssl_wrap_socket` to upgrade
    the connection socket into an SSL socket.
    """

    scheme = "https"
    ConnectionCls = HTTPSConnection

    def __init__(
        self,
        host,
        port=None,
        strict=False,
        timeout=Timeout.DEFAULT_TIMEOUT,
        maxsize=1,
        block=False,
        headers=None,
        retries=None,
        _proxy=None,
        _proxy_headers=None,
        key_file=None,
        cert_file=None,
        cert_reqs=None,
        key_password=None,
        ca_certs=None,
        ssl_version=None,
        assert_hostname=None,
        assert_fingerprint=None,
        ca_cert_dir=None,
        **conn_kw
    ):

        HTTPConnectionPool.__init__(
            self,
            host,
            port,
            strict,
            timeout,
            maxsize,
            block,
            headers,
            retries,
            _proxy,
            _proxy_headers,
            **conn_kw
        )

        self.key_file = key_file
        self.cert_file = cert_file
        self.cert_reqs = cert_reqs
        self.key_password = key_password
        self.ca_certs = ca_certs
        self.ca_cert_dir = ca_cert_dir
        self.ssl_version = ssl_version
        self.assert_hostname = assert_hostname
        self.assert_fingerprint = assert_fingerprint

    def _prepare_conn(self, conn):
        """
        Prepare the ``connection`` for :meth:`urllib3.util.ssl_wrap_socket`
        and establish the tunnel if proxy is used.
        """

        if isinstance(conn, VerifiedHTTPSConnection):
            conn.set_cert(
                key_file=self.key_file,
                key_password=self.key_password,
                cert_file=self.cert_file,
                cert_reqs=self.cert_reqs,
                ca_certs=self.ca_certs,
                ca_cert_dir=self.ca_cert_dir,
                assert_hostname=self.assert_hostname,
                assert_fingerprint=self.assert_fingerprint,
            )
            conn.ssl_version = self.ssl_version
        return conn

    def _prepare_proxy(self, conn):
        """
        Establish tunnel connection early, because otherwise httplib
        would improperly set Host: header to proxy's IP:port.
        """
        conn.set_tunnel(self._proxy_host, self.port, self.proxy_headers)
        conn.connect()

    def _new_conn(self):
        """
        Return a fresh :class:`httplib.HTTPSConnection`.
        """
        self.num_connections += 1
        log.debug(
            "Starting new HTTPS connection (%d): %s:%s",
            self.num_connections,
            self.host,
            self.port or "443",
        )

        if not self.ConnectionCls or self.ConnectionCls is DummyConnection:
            raise SSLError(
                "Can't connect to HTTPS URL because the SSL " "module is not available."
            )

        actual_host = self.host
        actual_port = self.port
        if self.proxy is not None:
            actual_host = self.proxy.host
            actual_port = self.proxy.port

        conn = self.ConnectionCls(
            host=actual_host,
            port=actual_port,
            timeout=self.timeout.connect_timeout,
            strict=self.strict,
            cert_file=self.cert_file,
            key_file=self.key_file,
            key_password=self.key_password,
            **self.conn_kw
        )

        return self._prepare_conn(conn)

    def _validate_conn(self, conn):
        """
        Called right before a request is made, after the socket is created.
        """
        super(HTTPSConnectionPool, self)._validate_conn(conn)

        # Force connect early to allow us to validate the connection.
        if not getattr(conn, "sock", None):  # AppEngine might not have  `.sock`
            conn.connect()

        if not conn.is_verified:
            warnings.warn(
                (
                    "Unverified HTTPS request is being made. "
                    "Adding certificate verification is strongly advised. See: "
                    "https://urllib3.readthedocs.io/en/latest/advanced-usage.html"
                    "#ssl-warnings"
                ),
                InsecureRequestWarning,
            )


def connection_from_url(url, **kw):
    """
    Given a url, return an :class:`.ConnectionPool` instance of its host.

    This is a shortcut for not having to parse out the scheme, host, and port
    of the url before creating an :class:`.ConnectionPool` instance.

    :param url:
        Absolute URL string that must include the scheme. Port is optional.

    :param \\**kw:
        Passes additional parameters to the constructor of the appropriate
        :class:`.ConnectionPool`. Useful for specifying things like
        timeout, maxsize, headers, etc.

    Example::

        >>> conn = connection_from_url('http://google.com/')
        >>> r = conn.request('GET', '/')
    """
    scheme, host, port = get_host(url)
    port = port or port_by_scheme.get(scheme, 80)
    if scheme == "https":
        return HTTPSConnectionPool(host, port=port, **kw)
    else:
        return HTTPConnectionPool(host, port=port, **kw)


def _normalize_host(host, scheme):
    """
    Normalize hosts for comparisons and use with sockets.
    """

    host = normalize_host(host, scheme)

    # httplib doesn't like it when we include brackets in IPv6 addresses
    # Specifically, if we include brackets but also pass the port then
    # httplib crazily doubles up the square brackets on the Host header.
    # Instead, we need to make sure we never pass ``None`` as the port.
    # However, for backward compatibility reasons we can't actually
    # *assert* that.  See http://bugs.python.org/issue28539
    if host.startswith("[") and host.endswith("]"):
        host = host[1:-1]
    return host
site-packages/pip/_vendor/urllib3/request.py000064400000013602151464452710015141 0ustar00from __future__ import absolute_import

from .filepost import encode_multipart_formdata
from .packages.six.moves.urllib.parse import urlencode


__all__ = ["RequestMethods"]


class RequestMethods(object):
    """
    Convenience mixin for classes who implement a :meth:`urlopen` method, such
    as :class:`~urllib3.connectionpool.HTTPConnectionPool` and
    :class:`~urllib3.poolmanager.PoolManager`.

    Provides behavior for making common types of HTTP request methods and
    decides which type of request field encoding to use.

    Specifically,

    :meth:`.request_encode_url` is for sending requests whose fields are
    encoded in the URL (such as GET, HEAD, DELETE).

    :meth:`.request_encode_body` is for sending requests whose fields are
    encoded in the *body* of the request using multipart or www-form-urlencoded
    (such as for POST, PUT, PATCH).

    :meth:`.request` is for making any kind of request, it will look up the
    appropriate encoding format and use one of the above two methods to make
    the request.

    Initializer parameters:

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.
    """

    _encode_url_methods = {"DELETE", "GET", "HEAD", "OPTIONS"}

    def __init__(self, headers=None):
        self.headers = headers or {}

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        encode_multipart=True,
        multipart_boundary=None,
        **kw
    ):  # Abstract
        raise NotImplementedError(
            "Classes extending RequestMethods must implement "
            "their own ``urlopen`` method."
        )

    def request(self, method, url, fields=None, headers=None, **urlopen_kw):
        """
        Make a request using :meth:`urlopen` with the appropriate encoding of
        ``fields`` based on the ``method`` used.

        This is a convenience method that requires the least amount of manual
        effort. It can be used in most situations, while still having the
        option to drop down to more specific methods when necessary, such as
        :meth:`request_encode_url`, :meth:`request_encode_body`,
        or even the lowest level :meth:`urlopen`.
        """
        method = method.upper()

        urlopen_kw["request_url"] = url

        if method in self._encode_url_methods:
            return self.request_encode_url(
                method, url, fields=fields, headers=headers, **urlopen_kw
            )
        else:
            return self.request_encode_body(
                method, url, fields=fields, headers=headers, **urlopen_kw
            )

    def request_encode_url(self, method, url, fields=None, headers=None, **urlopen_kw):
        """
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the url. This is useful for request methods like GET, HEAD, DELETE, etc.
        """
        if headers is None:
            headers = self.headers

        extra_kw = {"headers": headers}
        extra_kw.update(urlopen_kw)

        if fields:
            url += "?" + urlencode(fields)

        return self.urlopen(method, url, **extra_kw)

    def request_encode_body(
        self,
        method,
        url,
        fields=None,
        headers=None,
        encode_multipart=True,
        multipart_boundary=None,
        **urlopen_kw
    ):
        """
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the body. This is useful for request methods like POST, PUT, PATCH, etc.

        When ``encode_multipart=True`` (default), then
        :meth:`urllib3.filepost.encode_multipart_formdata` is used to encode
        the payload with the appropriate content type. Otherwise
        :meth:`urllib.urlencode` is used with the
        'application/x-www-form-urlencoded' content type.

        Multipart encoding must be used when posting files, and it's reasonably
        safe to use it in other times too. However, it may break request
        signing, such as with OAuth.

        Supports an optional ``fields`` parameter of key/value strings AND
        key/filetuple. A filetuple is a (filename, data, MIME type) tuple where
        the MIME type is optional. For example::

            fields = {
                'foo': 'bar',
                'fakefile': ('foofile.txt', 'contents of foofile'),
                'realfile': ('barfile.txt', open('realfile').read()),
                'typedfile': ('bazfile.bin', open('bazfile').read(),
                              'image/jpeg'),
                'nonamefile': 'contents of nonamefile field',
            }

        When uploading a file, providing a filename (the first parameter of the
        tuple) is optional but recommended to best mimic behavior of browsers.

        Note that if ``headers`` are supplied, the 'Content-Type' header will
        be overwritten because it depends on the dynamic random boundary string
        which is used to compose the body of the request. The random boundary
        string can be explicitly set with the ``multipart_boundary`` parameter.
        """
        if headers is None:
            headers = self.headers

        extra_kw = {"headers": {}}

        if fields:
            if "body" in urlopen_kw:
                raise TypeError(
                    "request got values for both 'fields' and 'body', can only specify one."
                )

            if encode_multipart:
                body, content_type = encode_multipart_formdata(
                    fields, boundary=multipart_boundary
                )
            else:
                body, content_type = (
                    urlencode(fields),
                    "application/x-www-form-urlencoded",
                )

            extra_kw["body"] = body
            extra_kw["headers"] = {"Content-Type": content_type}

        extra_kw["headers"].update(headers)
        extra_kw.update(urlopen_kw)

        return self.urlopen(method, url, **extra_kw)
site-packages/pip/_vendor/urllib3/fields.py000064400000020551151464452710014720 0ustar00from __future__ import absolute_import
import email.utils
import mimetypes
import re

from .packages import six


def guess_content_type(filename, default="application/octet-stream"):
    """
    Guess the "Content-Type" of a file.

    :param filename:
        The filename to guess the "Content-Type" of using :mod:`mimetypes`.
    :param default:
        If no "Content-Type" can be guessed, default to `default`.
    """
    if filename:
        return mimetypes.guess_type(filename)[0] or default
    return default


def format_header_param_rfc2231(name, value):
    """
    Helper function to format and quote a single header parameter using the
    strategy defined in RFC 2231.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows RFC 2388 Section 4.4.

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        An RFC-2231-formatted unicode string.
    """
    if isinstance(value, six.binary_type):
        value = value.decode("utf-8")

    if not any(ch in value for ch in '"\\\r\n'):
        result = u'%s="%s"' % (name, value)
        try:
            result.encode("ascii")
        except (UnicodeEncodeError, UnicodeDecodeError):
            pass
        else:
            return result

    if six.PY2:  # Python 2:
        value = value.encode("utf-8")

    # encode_rfc2231 accepts an encoded string and returns an ascii-encoded
    # string in Python 2 but accepts and returns unicode strings in Python 3
    value = email.utils.encode_rfc2231(value, "utf-8")
    value = "%s*=%s" % (name, value)

    if six.PY2:  # Python 2:
        value = value.decode("utf-8")

    return value


_HTML5_REPLACEMENTS = {
    u"\u0022": u"%22",
    # Replace "\" with "\\".
    u"\u005C": u"\u005C\u005C",
    u"\u005C": u"\u005C\u005C",
}

# All control characters from 0x00 to 0x1F *except* 0x1B.
_HTML5_REPLACEMENTS.update(
    {
        six.unichr(cc): u"%{:02X}".format(cc)
        for cc in range(0x00, 0x1F + 1)
        if cc not in (0x1B,)
    }
)


def _replace_multiple(value, needles_and_replacements):
    def replacer(match):
        return needles_and_replacements[match.group(0)]

    pattern = re.compile(
        r"|".join([re.escape(needle) for needle in needles_and_replacements.keys()])
    )

    result = pattern.sub(replacer, value)

    return result


def format_header_param_html5(name, value):
    """
    Helper function to format and quote a single header parameter using the
    HTML5 strategy.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows the `HTML5 Working Draft
    Section 4.10.22.7`_ and matches the behavior of curl and modern browsers.

    .. _HTML5 Working Draft Section 4.10.22.7:
        https://w3c.github.io/html/sec-forms.html#multipart-form-data

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        A unicode string, stripped of troublesome characters.
    """
    if isinstance(value, six.binary_type):
        value = value.decode("utf-8")

    value = _replace_multiple(value, _HTML5_REPLACEMENTS)

    return u'%s="%s"' % (name, value)


# For backwards-compatibility.
format_header_param = format_header_param_html5


class RequestField(object):
    """
    A data container for request body parameters.

    :param name:
        The name of this request field. Must be unicode.
    :param data:
        The data/value body.
    :param filename:
        An optional filename of the request field. Must be unicode.
    :param headers:
        An optional dict-like object of headers to initially use for the field.
    :param header_formatter:
        An optional callable that is used to encode and format the headers. By
        default, this is :func:`format_header_param_html5`.
    """

    def __init__(
        self,
        name,
        data,
        filename=None,
        headers=None,
        header_formatter=format_header_param_html5,
    ):
        self._name = name
        self._filename = filename
        self.data = data
        self.headers = {}
        if headers:
            self.headers = dict(headers)
        self.header_formatter = header_formatter

    @classmethod
    def from_tuples(cls, fieldname, value, header_formatter=format_header_param_html5):
        """
        A :class:`~urllib3.fields.RequestField` factory from old-style tuple parameters.

        Supports constructing :class:`~urllib3.fields.RequestField` from
        parameter of key/value strings AND key/filetuple. A filetuple is a
        (filename, data, MIME type) tuple where the MIME type is optional.
        For example::

            'foo': 'bar',
            'fakefile': ('foofile.txt', 'contents of foofile'),
            'realfile': ('barfile.txt', open('realfile').read()),
            'typedfile': ('bazfile.bin', open('bazfile').read(), 'image/jpeg'),
            'nonamefile': 'contents of nonamefile field',

        Field names and filenames must be unicode.
        """
        if isinstance(value, tuple):
            if len(value) == 3:
                filename, data, content_type = value
            else:
                filename, data = value
                content_type = guess_content_type(filename)
        else:
            filename = None
            content_type = None
            data = value

        request_param = cls(
            fieldname, data, filename=filename, header_formatter=header_formatter
        )
        request_param.make_multipart(content_type=content_type)

        return request_param

    def _render_part(self, name, value):
        """
        Overridable helper function to format a single header parameter. By
        default, this calls ``self.header_formatter``.

        :param name:
            The name of the parameter, a string expected to be ASCII only.
        :param value:
            The value of the parameter, provided as a unicode string.
        """

        return self.header_formatter(name, value)

    def _render_parts(self, header_parts):
        """
        Helper function to format and quote a single header.

        Useful for single headers that are composed of multiple items. E.g.,
        'Content-Disposition' fields.

        :param header_parts:
            A sequence of (k, v) tuples or a :class:`dict` of (k, v) to format
            as `k1="v1"; k2="v2"; ...`.
        """
        parts = []
        iterable = header_parts
        if isinstance(header_parts, dict):
            iterable = header_parts.items()

        for name, value in iterable:
            if value is not None:
                parts.append(self._render_part(name, value))

        return u"; ".join(parts)

    def render_headers(self):
        """
        Renders the headers for this request field.
        """
        lines = []

        sort_keys = ["Content-Disposition", "Content-Type", "Content-Location"]
        for sort_key in sort_keys:
            if self.headers.get(sort_key, False):
                lines.append(u"%s: %s" % (sort_key, self.headers[sort_key]))

        for header_name, header_value in self.headers.items():
            if header_name not in sort_keys:
                if header_value:
                    lines.append(u"%s: %s" % (header_name, header_value))

        lines.append(u"\r\n")
        return u"\r\n".join(lines)

    def make_multipart(
        self, content_disposition=None, content_type=None, content_location=None
    ):
        """
        Makes this request field into a multipart request field.

        This method overrides "Content-Disposition", "Content-Type" and
        "Content-Location" headers to the request parameter.

        :param content_type:
            The 'Content-Type' of the request body.
        :param content_location:
            The 'Content-Location' of the request body.

        """
        self.headers["Content-Disposition"] = content_disposition or u"form-data"
        self.headers["Content-Disposition"] += u"; ".join(
            [
                u"",
                self._render_parts(
                    ((u"name", self._name), (u"filename", self._filename))
                ),
            ]
        )
        self.headers["Content-Type"] = content_type
        self.headers["Content-Location"] = content_location
site-packages/pip/_vendor/urllib3/filepost.py000064400000004610151464452710015275 0ustar00from __future__ import absolute_import
import binascii
import codecs
import os

from io import BytesIO

from .packages import six
from .packages.six import b
from .fields import RequestField

writer = codecs.lookup("utf-8")[3]


def choose_boundary():
    """
    Our embarrassingly-simple replacement for mimetools.choose_boundary.
    """
    boundary = binascii.hexlify(os.urandom(16))
    if not six.PY2:
        boundary = boundary.decode("ascii")
    return boundary


def iter_field_objects(fields):
    """
    Iterate over fields.

    Supports list of (k, v) tuples and dicts, and lists of
    :class:`~urllib3.fields.RequestField`.

    """
    if isinstance(fields, dict):
        i = six.iteritems(fields)
    else:
        i = iter(fields)

    for field in i:
        if isinstance(field, RequestField):
            yield field
        else:
            yield RequestField.from_tuples(*field)


def iter_fields(fields):
    """
    .. deprecated:: 1.6

    Iterate over fields.

    The addition of :class:`~urllib3.fields.RequestField` makes this function
    obsolete. Instead, use :func:`iter_field_objects`, which returns
    :class:`~urllib3.fields.RequestField` objects.

    Supports list of (k, v) tuples and dicts.
    """
    if isinstance(fields, dict):
        return ((k, v) for k, v in six.iteritems(fields))

    return ((k, v) for k, v in fields)


def encode_multipart_formdata(fields, boundary=None):
    """
    Encode a dictionary of ``fields`` using the multipart/form-data MIME format.

    :param fields:
        Dictionary of fields or list of (key, :class:`~urllib3.fields.RequestField`).

    :param boundary:
        If not specified, then a random boundary will be generated using
        :func:`urllib3.filepost.choose_boundary`.
    """
    body = BytesIO()
    if boundary is None:
        boundary = choose_boundary()

    for field in iter_field_objects(fields):
        body.write(b("--%s\r\n" % (boundary)))

        writer(body).write(field.render_headers())
        data = field.data

        if isinstance(data, int):
            data = str(data)  # Backwards compatibility

        if isinstance(data, six.text_type):
            writer(body).write(data)
        else:
            body.write(data)

        body.write(b"\r\n")

    body.write(b("--%s--\r\n" % (boundary)))

    content_type = str("multipart/form-data; boundary=%s" % boundary)

    return body.getvalue(), content_type
site-packages/pip/_vendor/urllib3/connection.py000064400000035502151464452710015613 0ustar00from __future__ import absolute_import
import datetime
import logging
import os
import socket
from socket import error as SocketError, timeout as SocketTimeout
import warnings
from .packages import six
from .packages.six.moves.http_client import HTTPConnection as _HTTPConnection
from .packages.six.moves.http_client import HTTPException  # noqa: F401

try:  # Compiled with SSL?
    import ssl

    BaseSSLError = ssl.SSLError
except (ImportError, AttributeError):  # Platform-specific: No SSL.
    ssl = None

    class BaseSSLError(BaseException):
        pass


try:
    # Python 3: not a no-op, we're adding this to the namespace so it can be imported.
    ConnectionError = ConnectionError
except NameError:
    # Python 2
    class ConnectionError(Exception):
        pass


from .exceptions import (
    NewConnectionError,
    ConnectTimeoutError,
    SubjectAltNameWarning,
    SystemTimeWarning,
)
from .packages.ssl_match_hostname import match_hostname, CertificateError

from .util.ssl_ import (
    resolve_cert_reqs,
    resolve_ssl_version,
    assert_fingerprint,
    create_urllib3_context,
    ssl_wrap_socket,
)


from .util import connection

from ._collections import HTTPHeaderDict

log = logging.getLogger(__name__)

port_by_scheme = {"http": 80, "https": 443}

# When it comes time to update this value as a part of regular maintenance
# (ie test_recent_date is failing) update it to ~6 months before the current date.
RECENT_DATE = datetime.date(2019, 1, 1)


class DummyConnection(object):
    """Used to detect a failed ConnectionCls import."""

    pass


class HTTPConnection(_HTTPConnection, object):
    """
    Based on httplib.HTTPConnection but provides an extra constructor
    backwards-compatibility layer between older and newer Pythons.

    Additional keyword parameters are used to configure attributes of the connection.
    Accepted parameters include:

      - ``strict``: See the documentation on :class:`urllib3.connectionpool.HTTPConnectionPool`
      - ``source_address``: Set the source address for the current connection.
      - ``socket_options``: Set specific options on the underlying socket. If not specified, then
        defaults are loaded from ``HTTPConnection.default_socket_options`` which includes disabling
        Nagle's algorithm (sets TCP_NODELAY to 1) unless the connection is behind a proxy.

        For example, if you wish to enable TCP Keep Alive in addition to the defaults,
        you might pass::

            HTTPConnection.default_socket_options + [
                (socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1),
            ]

        Or you may want to disable the defaults by passing an empty list (e.g., ``[]``).
    """

    default_port = port_by_scheme["http"]

    #: Disable Nagle's algorithm by default.
    #: ``[(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)]``
    default_socket_options = [(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)]

    #: Whether this connection verifies the host's certificate.
    is_verified = False

    def __init__(self, *args, **kw):
        if not six.PY2:
            kw.pop("strict", None)

        # Pre-set source_address.
        self.source_address = kw.get("source_address")

        #: The socket options provided by the user. If no options are
        #: provided, we use the default options.
        self.socket_options = kw.pop("socket_options", self.default_socket_options)

        _HTTPConnection.__init__(self, *args, **kw)

    @property
    def host(self):
        """
        Getter method to remove any trailing dots that indicate the hostname is an FQDN.

        In general, SSL certificates don't include the trailing dot indicating a
        fully-qualified domain name, and thus, they don't validate properly when
        checked against a domain name that includes the dot. In addition, some
        servers may not expect to receive the trailing dot when provided.

        However, the hostname with trailing dot is critical to DNS resolution; doing a
        lookup with the trailing dot will properly only resolve the appropriate FQDN,
        whereas a lookup without a trailing dot will search the system's search domain
        list. Thus, it's important to keep the original host around for use only in
        those cases where it's appropriate (i.e., when doing DNS lookup to establish the
        actual TCP connection across which we're going to send HTTP requests).
        """
        return self._dns_host.rstrip(".")

    @host.setter
    def host(self, value):
        """
        Setter for the `host` property.

        We assume that only urllib3 uses the _dns_host attribute; httplib itself
        only uses `host`, and it seems reasonable that other libraries follow suit.
        """
        self._dns_host = value

    def _new_conn(self):
        """ Establish a socket connection and set nodelay settings on it.

        :return: New socket connection.
        """
        extra_kw = {}
        if self.source_address:
            extra_kw["source_address"] = self.source_address

        if self.socket_options:
            extra_kw["socket_options"] = self.socket_options

        try:
            conn = connection.create_connection(
                (self._dns_host, self.port), self.timeout, **extra_kw
            )

        except SocketTimeout:
            raise ConnectTimeoutError(
                self,
                "Connection to %s timed out. (connect timeout=%s)"
                % (self.host, self.timeout),
            )

        except SocketError as e:
            raise NewConnectionError(
                self, "Failed to establish a new connection: %s" % e
            )

        return conn

    def _prepare_conn(self, conn):
        self.sock = conn
        # Google App Engine's httplib does not define _tunnel_host
        if getattr(self, "_tunnel_host", None):
            # TODO: Fix tunnel so it doesn't depend on self.sock state.
            self._tunnel()
            # Mark this connection as not reusable
            self.auto_open = 0

    def connect(self):
        conn = self._new_conn()
        self._prepare_conn(conn)

    def request_chunked(self, method, url, body=None, headers=None):
        """
        Alternative to the common request method, which sends the
        body with chunked encoding and not as one block
        """
        headers = HTTPHeaderDict(headers if headers is not None else {})
        skip_accept_encoding = "accept-encoding" in headers
        skip_host = "host" in headers
        self.putrequest(
            method, url, skip_accept_encoding=skip_accept_encoding, skip_host=skip_host
        )
        for header, value in headers.items():
            self.putheader(header, value)
        if "transfer-encoding" not in headers:
            self.putheader("Transfer-Encoding", "chunked")
        self.endheaders()

        if body is not None:
            stringish_types = six.string_types + (bytes,)
            if isinstance(body, stringish_types):
                body = (body,)
            for chunk in body:
                if not chunk:
                    continue
                if not isinstance(chunk, bytes):
                    chunk = chunk.encode("utf8")
                len_str = hex(len(chunk))[2:]
                self.send(len_str.encode("utf-8"))
                self.send(b"\r\n")
                self.send(chunk)
                self.send(b"\r\n")

        # After the if clause, to always have a closed body
        self.send(b"0\r\n\r\n")


class HTTPSConnection(HTTPConnection):
    default_port = port_by_scheme["https"]

    ssl_version = None

    def __init__(
        self,
        host,
        port=None,
        key_file=None,
        cert_file=None,
        key_password=None,
        strict=None,
        timeout=socket._GLOBAL_DEFAULT_TIMEOUT,
        ssl_context=None,
        server_hostname=None,
        **kw
    ):

        HTTPConnection.__init__(self, host, port, strict=strict, timeout=timeout, **kw)

        self.key_file = key_file
        self.cert_file = cert_file
        self.key_password = key_password
        self.ssl_context = ssl_context
        self.server_hostname = server_hostname

        # Required property for Google AppEngine 1.9.0 which otherwise causes
        # HTTPS requests to go out as HTTP. (See Issue #356)
        self._protocol = "https"

    def connect(self):
        conn = self._new_conn()
        self._prepare_conn(conn)

        # Wrap socket using verification with the root certs in
        # trusted_root_certs
        default_ssl_context = False
        if self.ssl_context is None:
            default_ssl_context = True
            self.ssl_context = create_urllib3_context(
                ssl_version=resolve_ssl_version(self.ssl_version),
                cert_reqs=resolve_cert_reqs(self.cert_reqs),
            )

        # Try to load OS default certs if none are given.
        # Works well on Windows (requires Python3.4+)
        context = self.ssl_context
        if (
            not self.ca_certs
            and not self.ca_cert_dir
            and default_ssl_context
            and hasattr(context, "load_default_certs")
        ):
            context.load_default_certs()

        self.sock = ssl_wrap_socket(
            sock=conn,
            keyfile=self.key_file,
            certfile=self.cert_file,
            key_password=self.key_password,
            ssl_context=self.ssl_context,
            server_hostname=self.server_hostname,
        )


class VerifiedHTTPSConnection(HTTPSConnection):
    """
    Based on httplib.HTTPSConnection but wraps the socket with
    SSL certification.
    """

    cert_reqs = None
    ca_certs = None
    ca_cert_dir = None
    ssl_version = None
    assert_fingerprint = None

    def set_cert(
        self,
        key_file=None,
        cert_file=None,
        cert_reqs=None,
        key_password=None,
        ca_certs=None,
        assert_hostname=None,
        assert_fingerprint=None,
        ca_cert_dir=None,
    ):
        """
        This method should only be called once, before the connection is used.
        """
        # If cert_reqs is not provided we'll assume CERT_REQUIRED unless we also
        # have an SSLContext object in which case we'll use its verify_mode.
        if cert_reqs is None:
            if self.ssl_context is not None:
                cert_reqs = self.ssl_context.verify_mode
            else:
                cert_reqs = resolve_cert_reqs(None)

        self.key_file = key_file
        self.cert_file = cert_file
        self.cert_reqs = cert_reqs
        self.key_password = key_password
        self.assert_hostname = assert_hostname
        self.assert_fingerprint = assert_fingerprint
        self.ca_certs = ca_certs and os.path.expanduser(ca_certs)
        self.ca_cert_dir = ca_cert_dir and os.path.expanduser(ca_cert_dir)

    def connect(self):
        # Add certificate verification
        conn = self._new_conn()
        hostname = self.host

        # Google App Engine's httplib does not define _tunnel_host
        if getattr(self, "_tunnel_host", None):
            self.sock = conn
            # Calls self._set_hostport(), so self.host is
            # self._tunnel_host below.
            self._tunnel()
            # Mark this connection as not reusable
            self.auto_open = 0

            # Override the host with the one we're requesting data from.
            hostname = self._tunnel_host

        server_hostname = hostname
        if self.server_hostname is not None:
            server_hostname = self.server_hostname

        is_time_off = datetime.date.today() < RECENT_DATE
        if is_time_off:
            warnings.warn(
                (
                    "System time is way off (before {0}). This will probably "
                    "lead to SSL verification errors"
                ).format(RECENT_DATE),
                SystemTimeWarning,
            )

        # Wrap socket using verification with the root certs in
        # trusted_root_certs
        default_ssl_context = False
        if self.ssl_context is None:
            default_ssl_context = True
            self.ssl_context = create_urllib3_context(
                ssl_version=resolve_ssl_version(self.ssl_version),
                cert_reqs=resolve_cert_reqs(self.cert_reqs),
            )

        context = self.ssl_context
        context.verify_mode = resolve_cert_reqs(self.cert_reqs)

        # Try to load OS default certs if none are given.
        # Works well on Windows (requires Python3.4+)
        if (
            not self.ca_certs
            and not self.ca_cert_dir
            and default_ssl_context
            and hasattr(context, "load_default_certs")
        ):
            context.load_default_certs()

        self.sock = ssl_wrap_socket(
            sock=conn,
            keyfile=self.key_file,
            certfile=self.cert_file,
            key_password=self.key_password,
            ca_certs=self.ca_certs,
            ca_cert_dir=self.ca_cert_dir,
            server_hostname=server_hostname,
            ssl_context=context,
        )

        if self.assert_fingerprint:
            assert_fingerprint(
                self.sock.getpeercert(binary_form=True), self.assert_fingerprint
            )
        elif (
            context.verify_mode != ssl.CERT_NONE
            and not getattr(context, "check_hostname", False)
            and self.assert_hostname is not False
        ):
            # While urllib3 attempts to always turn off hostname matching from
            # the TLS library, this cannot always be done. So we check whether
            # the TLS Library still thinks it's matching hostnames.
            cert = self.sock.getpeercert()
            if not cert.get("subjectAltName", ()):
                warnings.warn(
                    (
                        "Certificate for {0} has no `subjectAltName`, falling back to check for a "
                        "`commonName` for now. This feature is being removed by major browsers and "
                        "deprecated by RFC 2818. (See https://github.com/shazow/urllib3/issues/497 "
                        "for details.)".format(hostname)
                    ),
                    SubjectAltNameWarning,
                )
            _match_hostname(cert, self.assert_hostname or server_hostname)

        self.is_verified = (
            context.verify_mode == ssl.CERT_REQUIRED
            or self.assert_fingerprint is not None
        )


def _match_hostname(cert, asserted_hostname):
    try:
        match_hostname(cert, asserted_hostname)
    except CertificateError as e:
        log.warning(
            "Certificate did not match expected hostname: %s. " "Certificate: %s",
            asserted_hostname,
            cert,
        )
        # Add cert to exception and reraise so client code can inspect
        # the cert when catching the exception, if they want to
        e._peer_cert = cert
        raise


if ssl:
    # Make a copy for testing.
    UnverifiedHTTPSConnection = HTTPSConnection
    HTTPSConnection = VerifiedHTTPSConnection
else:
    HTTPSConnection = DummyConnection
site-packages/pip/_vendor/urllib3/_collections.py000064400000025050151464452710016126 0ustar00from __future__ import absolute_import

try:
    from collections.abc import Mapping, MutableMapping
except ImportError:
    from collections import Mapping, MutableMapping
try:
    from threading import RLock
except ImportError:  # Platform-specific: No threads available

    class RLock:
        def __enter__(self):
            pass

        def __exit__(self, exc_type, exc_value, traceback):
            pass


from collections import OrderedDict
from .exceptions import InvalidHeader
from .packages.six import iterkeys, itervalues, PY3


__all__ = ["RecentlyUsedContainer", "HTTPHeaderDict"]


_Null = object()


class RecentlyUsedContainer(MutableMapping):
    """
    Provides a thread-safe dict-like container which maintains up to
    ``maxsize`` keys while throwing away the least-recently-used keys beyond
    ``maxsize``.

    :param maxsize:
        Maximum number of recent elements to retain.

    :param dispose_func:
        Every time an item is evicted from the container,
        ``dispose_func(value)`` is called.  Callback which will get called
    """

    ContainerCls = OrderedDict

    def __init__(self, maxsize=10, dispose_func=None):
        self._maxsize = maxsize
        self.dispose_func = dispose_func

        self._container = self.ContainerCls()
        self.lock = RLock()

    def __getitem__(self, key):
        # Re-insert the item, moving it to the end of the eviction line.
        with self.lock:
            item = self._container.pop(key)
            self._container[key] = item
            return item

    def __setitem__(self, key, value):
        evicted_value = _Null
        with self.lock:
            # Possibly evict the existing value of 'key'
            evicted_value = self._container.get(key, _Null)
            self._container[key] = value

            # If we didn't evict an existing value, we might have to evict the
            # least recently used item from the beginning of the container.
            if len(self._container) > self._maxsize:
                _key, evicted_value = self._container.popitem(last=False)

        if self.dispose_func and evicted_value is not _Null:
            self.dispose_func(evicted_value)

    def __delitem__(self, key):
        with self.lock:
            value = self._container.pop(key)

        if self.dispose_func:
            self.dispose_func(value)

    def __len__(self):
        with self.lock:
            return len(self._container)

    def __iter__(self):
        raise NotImplementedError(
            "Iteration over this class is unlikely to be threadsafe."
        )

    def clear(self):
        with self.lock:
            # Copy pointers to all values, then wipe the mapping
            values = list(itervalues(self._container))
            self._container.clear()

        if self.dispose_func:
            for value in values:
                self.dispose_func(value)

    def keys(self):
        with self.lock:
            return list(iterkeys(self._container))


class HTTPHeaderDict(MutableMapping):
    """
    :param headers:
        An iterable of field-value pairs. Must not contain multiple field names
        when compared case-insensitively.

    :param kwargs:
        Additional field-value pairs to pass in to ``dict.update``.

    A ``dict`` like container for storing HTTP Headers.

    Field names are stored and compared case-insensitively in compliance with
    RFC 7230. Iteration provides the first case-sensitive key seen for each
    case-insensitive pair.

    Using ``__setitem__`` syntax overwrites fields that compare equal
    case-insensitively in order to maintain ``dict``'s api. For fields that
    compare equal, instead create a new ``HTTPHeaderDict`` and use ``.add``
    in a loop.

    If multiple fields that are equal case-insensitively are passed to the
    constructor or ``.update``, the behavior is undefined and some will be
    lost.

    >>> headers = HTTPHeaderDict()
    >>> headers.add('Set-Cookie', 'foo=bar')
    >>> headers.add('set-cookie', 'baz=quxx')
    >>> headers['content-length'] = '7'
    >>> headers['SET-cookie']
    'foo=bar, baz=quxx'
    >>> headers['Content-Length']
    '7'
    """

    def __init__(self, headers=None, **kwargs):
        super(HTTPHeaderDict, self).__init__()
        self._container = OrderedDict()
        if headers is not None:
            if isinstance(headers, HTTPHeaderDict):
                self._copy_from(headers)
            else:
                self.extend(headers)
        if kwargs:
            self.extend(kwargs)

    def __setitem__(self, key, val):
        self._container[key.lower()] = [key, val]
        return self._container[key.lower()]

    def __getitem__(self, key):
        val = self._container[key.lower()]
        return ", ".join(val[1:])

    def __delitem__(self, key):
        del self._container[key.lower()]

    def __contains__(self, key):
        return key.lower() in self._container

    def __eq__(self, other):
        if not isinstance(other, Mapping) and not hasattr(other, "keys"):
            return False
        if not isinstance(other, type(self)):
            other = type(self)(other)
        return dict((k.lower(), v) for k, v in self.itermerged()) == dict(
            (k.lower(), v) for k, v in other.itermerged()
        )

    def __ne__(self, other):
        return not self.__eq__(other)

    if not PY3:  # Python 2
        iterkeys = MutableMapping.iterkeys
        itervalues = MutableMapping.itervalues

    __marker = object()

    def __len__(self):
        return len(self._container)

    def __iter__(self):
        # Only provide the originally cased names
        for vals in self._container.values():
            yield vals[0]

    def pop(self, key, default=__marker):
        """D.pop(k[,d]) -> v, remove specified key and return the corresponding value.
          If key is not found, d is returned if given, otherwise KeyError is raised.
        """
        # Using the MutableMapping function directly fails due to the private marker.
        # Using ordinary dict.pop would expose the internal structures.
        # So let's reinvent the wheel.
        try:
            value = self[key]
        except KeyError:
            if default is self.__marker:
                raise
            return default
        else:
            del self[key]
            return value

    def discard(self, key):
        try:
            del self[key]
        except KeyError:
            pass

    def add(self, key, val):
        """Adds a (name, value) pair, doesn't overwrite the value if it already
        exists.

        >>> headers = HTTPHeaderDict(foo='bar')
        >>> headers.add('Foo', 'baz')
        >>> headers['foo']
        'bar, baz'
        """
        key_lower = key.lower()
        new_vals = [key, val]
        # Keep the common case aka no item present as fast as possible
        vals = self._container.setdefault(key_lower, new_vals)
        if new_vals is not vals:
            vals.append(val)

    def extend(self, *args, **kwargs):
        """Generic import function for any type of header-like object.
        Adapted version of MutableMapping.update in order to insert items
        with self.add instead of self.__setitem__
        """
        if len(args) > 1:
            raise TypeError(
                "extend() takes at most 1 positional "
                "arguments ({0} given)".format(len(args))
            )
        other = args[0] if len(args) >= 1 else ()

        if isinstance(other, HTTPHeaderDict):
            for key, val in other.iteritems():
                self.add(key, val)
        elif isinstance(other, Mapping):
            for key in other:
                self.add(key, other[key])
        elif hasattr(other, "keys"):
            for key in other.keys():
                self.add(key, other[key])
        else:
            for key, value in other:
                self.add(key, value)

        for key, value in kwargs.items():
            self.add(key, value)

    def getlist(self, key, default=__marker):
        """Returns a list of all the values for the named field. Returns an
        empty list if the key doesn't exist."""
        try:
            vals = self._container[key.lower()]
        except KeyError:
            if default is self.__marker:
                return []
            return default
        else:
            return vals[1:]

    # Backwards compatibility for httplib
    getheaders = getlist
    getallmatchingheaders = getlist
    iget = getlist

    # Backwards compatibility for http.cookiejar
    get_all = getlist

    def __repr__(self):
        return "%s(%s)" % (type(self).__name__, dict(self.itermerged()))

    def _copy_from(self, other):
        for key in other:
            val = other.getlist(key)
            if isinstance(val, list):
                # Don't need to convert tuples
                val = list(val)
            self._container[key.lower()] = [key] + val

    def copy(self):
        clone = type(self)()
        clone._copy_from(self)
        return clone

    def iteritems(self):
        """Iterate over all header lines, including duplicate ones."""
        for key in self:
            vals = self._container[key.lower()]
            for val in vals[1:]:
                yield vals[0], val

    def itermerged(self):
        """Iterate over all headers, merging duplicate ones together."""
        for key in self:
            val = self._container[key.lower()]
            yield val[0], ", ".join(val[1:])

    def items(self):
        return list(self.iteritems())

    @classmethod
    def from_httplib(cls, message):  # Python 2
        """Read headers from a Python 2 httplib message object."""
        # python2.7 does not expose a proper API for exporting multiheaders
        # efficiently. This function re-reads raw lines from the message
        # object and extracts the multiheaders properly.
        obs_fold_continued_leaders = (" ", "\t")
        headers = []

        for line in message.headers:
            if line.startswith(obs_fold_continued_leaders):
                if not headers:
                    # We received a header line that starts with OWS as described
                    # in RFC-7230 S3.2.4. This indicates a multiline header, but
                    # there exists no previous header to which we can attach it.
                    raise InvalidHeader(
                        "Header continuation with no previous header: %s" % line
                    )
                else:
                    key, value = headers[-1]
                    headers[-1] = (key, value + " " + line.strip())
                    continue

            key, value = line.split(":", 1)
            headers.append((key, value.strip()))

        return cls(headers)
site-packages/pip/_vendor/urllib3/__pycache__/connection.cpython-38.opt-1.pyc000064400000024462151464452710023043 0ustar00U

.eB;	@sddlmZddlZddlZddlZddlZddlmZmZ	ddl
Z
ddlmZddl
mZddl
mZzddlZejZWn,eefk
rdZGdd	d	eZYnXzeZWn$ek
rGd
ddeZYnXddlmZmZmZmZdd
lm Z m!Z!ddl"m#Z#m$Z$m%Z%m&Z&m'Z'ddl(m)Z)ddl*m+Z+e,e-Z.dddZ/e0dddZ1Gddde2Z3Gdddee2ZGdddeZ4Gddde4Z5ddZ6ere4Z7e5Z4ne3Z4dS))absolute_importN)errortimeout)six)HTTPConnection)
HTTPExceptionc@seZdZdS)BaseSSLErrorN__name__
__module____qualname__rrB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/connection.pyr	sr	c@seZdZdS)ConnectionErrorNr
rrrrrsr)NewConnectionErrorConnectTimeoutErrorSubjectAltNameWarningSystemTimeWarning)match_hostnameCertificateError)resolve_cert_reqsresolve_ssl_versionassert_fingerprintcreate_urllib3_contextssl_wrap_socket)
connection)HTTPHeaderDictPi)httphttpsic@seZdZdZdS)DummyConnectionz-Used to detect a failed ConnectionCls import.N)rrr
__doc__rrrrr!>sr!c@speZdZdZedZejejdfgZ	dZ
ddZeddZ
e
jd	dZ
d
dZdd
ZddZdddZdS)ra$
    Based on httplib.HTTPConnection but provides an extra constructor
    backwards-compatibility layer between older and newer Pythons.

    Additional keyword parameters are used to configure attributes of the connection.
    Accepted parameters include:

      - ``strict``: See the documentation on :class:`urllib3.connectionpool.HTTPConnectionPool`
      - ``source_address``: Set the source address for the current connection.
      - ``socket_options``: Set specific options on the underlying socket. If not specified, then
        defaults are loaded from ``HTTPConnection.default_socket_options`` which includes disabling
        Nagle's algorithm (sets TCP_NODELAY to 1) unless the connection is behind a proxy.

        For example, if you wish to enable TCP Keep Alive in addition to the defaults,
        you might pass::

            HTTPConnection.default_socket_options + [
                (socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1),
            ]

        Or you may want to disable the defaults by passing an empty list (e.g., ``[]``).
    rrFcOsDtjs|dd|d|_|d|j|_tj|f||dS)Nstrictsource_addresssocket_options)	rZPY2popgetr$default_socket_optionsr%_HTTPConnection__init__)selfargskwrrrr*es
zHTTPConnection.__init__cCs|jdS)a
        Getter method to remove any trailing dots that indicate the hostname is an FQDN.

        In general, SSL certificates don't include the trailing dot indicating a
        fully-qualified domain name, and thus, they don't validate properly when
        checked against a domain name that includes the dot. In addition, some
        servers may not expect to receive the trailing dot when provided.

        However, the hostname with trailing dot is critical to DNS resolution; doing a
        lookup with the trailing dot will properly only resolve the appropriate FQDN,
        whereas a lookup without a trailing dot will search the system's search domain
        list. Thus, it's important to keep the original host around for use only in
        those cases where it's appropriate (i.e., when doing DNS lookup to establish the
        actual TCP connection across which we're going to send HTTP requests).
        .)	_dns_hostrstrip)r+rrrhostrszHTTPConnection.hostcCs
||_dS)z
        Setter for the `host` property.

        We assume that only urllib3 uses the _dns_host attribute; httplib itself
        only uses `host`, and it seems reasonable that other libraries follow suit.
        N)r/)r+valuerrrr1sc
Csi}|jr|j|d<|jr$|j|d<ztj|j|jf|jf|}WnXtk
rlt|d|j	|jfYn0t
k
r}zt|d|W5d}~XYnX|S)zp Establish a socket connection and set nodelay settings on it.

        :return: New socket connection.
        r$r%z0Connection to %s timed out. (connect timeout=%s)z(Failed to establish a new connection: %sN)r$r%rZcreate_connectionr/portr
SocketTimeoutrr1SocketErrorr)r+Zextra_kwconnerrr	_new_conns2



zHTTPConnection._new_conncCs$||_t|ddr |d|_dS)N_tunnel_hostr)sockgetattr_tunnel	auto_openr+r6rrr
_prepare_connszHTTPConnection._prepare_conncCs|}||dS)N)r8r?r>rrrconnectszHTTPConnection.connectNcCst|dk	r|ni}d|k}d|k}|j||||d|D]\}}|||q>d|krh|dd||dk	rtjtf}	t||	r|f}|D]`}
|
sqt|
ts|
	d}
t
t|
d	d}||	d
|d||
|dq|ddS)
z
        Alternative to the common request method, which sends the
        body with chunked encoding and not as one block
        Nzaccept-encodingr1)skip_accept_encoding	skip_hostztransfer-encodingzTransfer-EncodingZchunkedutf8zutf-8s
s0

)
rZ
putrequestitemsZ	putheaderZ
endheadersrZstring_typesbytes
isinstanceencodehexlensend)r+methodZurlZbodyZheadersrArBheaderr2Zstringish_typeschunkZlen_strrrrrequest_chunkeds:




zHTTPConnection.request_chunked)NN)rrr
r"port_by_schemedefault_portsocketZIPPROTO_TCPZTCP_NODELAYr(is_verifiedr*propertyr1setterr8r?r@rOrrrrrDs


		rc@s<eZdZedZdZdddddejddfddZddZ	dS)HTTPSConnectionr Nc
KsDtj|||f||d|
||_||_||_||_|	|_d|_dS)N)r#rr )rr*key_file	cert_filekey_passwordssl_contextserver_hostnameZ	_protocol)r+r1r3rWrXrYr#rrZr[r-rrrr*szHTTPSConnection.__init__cCs|}||d}|jdkr>d}tt|jt|jd|_|j}|jsf|j	sf|rft
|drf|t||j
|j|j|j|jd|_dS)NFTssl_version	cert_reqsload_default_certs)r:keyfilecertfilerYrZr[)r8r?rZrrr]rr^ca_certsca_cert_dirhasattrr_rrWrXrYr[r:)r+r6default_ssl_contextcontextrrrr@s6

zHTTPSConnection.connect)
rrr
rPrQr]rRZ_GLOBAL_DEFAULT_TIMEOUTr*r@rrrrrVs
rVc@s6eZdZdZdZdZdZdZdZdddZ	ddZ
dS)VerifiedHTTPSConnectionz[
    Based on httplib.HTTPSConnection but wraps the socket with
    SSL certification.
    Nc		Csp|dkr$|jdk	r|jj}ntd}||_||_||_||_||_||_|oVt	j
||_|oht	j
||_
dS)zX
        This method should only be called once, before the connection is used.
        N)rZverify_moderrWrXr^rYassert_hostnamerospath
expanduserrbrc)	r+rWrXr^rYrbrirrcrrrset_cert-s

z VerifiedHTTPSConnection.set_certc
Cs|}|j}t|ddr4||_|d|_|j}|}|jdk	rH|j}tj	
tk}|rlt
dttd}|jdkrd}tt|jt|jd|_|j}t|j|_|js|js|rt|dr|t||j|j|j|j|j||d|_|jrt|jj dd	|jnd|jt!j"krtt|d
dst|j#dk	rt|j }|$ddsbt
d
|t%t&||j#pp||jt!j'kp|jdk	|_(dS)Nr9rzWSystem time is way off (before {0}). This will probably lead to SSL verification errorsFTr\r_)r:r`rarYrbrcr[rZ)Zbinary_formZcheck_hostnameZsubjectAltNamerzCertificate for {0} has no `subjectAltName`, falling back to check for a `commonName` for now. This feature is being removed by major browsers and deprecated by RFC 2818. (See https://github.com/shazow/urllib3/issues/497 for details.)))r8r1r;r:r<r=r9r[datetimedateZtodayRECENT_DATEwarningswarnformatrrZrrr]rr^rhrbrcrdr_rrWrXrYrZgetpeercertsslZ	CERT_NONErir'r_match_hostnameZ
CERT_REQUIREDrS)r+r6Zhostnamer[Zis_time_offrerfcertrrrr@Ls





	zVerifiedHTTPSConnection.connect)NNNNNNNN)rrr
r"r^rbrcr]rrmr@rrrrrg!s 
rgc
CsLzt||Wn8tk
rF}ztd||||_W5d}~XYnXdS)Nz@Certificate did not match expected hostname: %s. Certificate: %s)rrlogZwarningZ
_peer_cert)rvZasserted_hostnamer7rrrrusru)8Z
__future__rrnZloggingrjrRrr5rr4rqZpackagesrZpackages.six.moves.http_clientrr)rrtZSSLErrorr	ImportErrorAttributeError
BaseExceptionr	NameError	Exception
exceptionsrrrrZpackages.ssl_match_hostnamerrZ	util.ssl_rrrrrutilr_collectionsrZ	getLoggerrrwrProrpobjectr!rVrgruZUnverifiedHTTPSConnectionrrrrsL
	

Bsite-packages/pip/_vendor/urllib3/__pycache__/connectionpool.cpython-38.pyc000064400000056730151464452710023001 0ustar00U

.e^@sddlmZddlZddlZddlZddlZddlmZm	Z
ddlZddlmZm
Z
mZmZmZmZmZmZmZmZmZmZmZddlmZddlmZddlmZdd	lm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&dd
l'm(Z(ddl)m*Z*ddl+m,Z,dd
l-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6m7Z7m8Z8m9Z:m;Z;ddlj?Z?e@eAZBeCZDGdddeCZEejFejGhZHGdddeEe(ZIGdddeIZJddZKddZ9dS))absolute_importN)errortimeout)
ClosedPoolError
ProtocolErrorEmptyPoolErrorHeaderParsingErrorHostChangedErrorLocationValueError
MaxRetryError
ProxyErrorReadTimeoutErrorSSLErrorTimeoutErrorInsecureRequestWarningNewConnectionError)CertificateError)six)queue)port_by_schemeDummyConnectionHTTPConnectionHTTPSConnectionVerifiedHTTPSConnection
HTTPExceptionBaseSSLError)RequestMethods)HTTPResponse)is_connection_dropped)set_file_position)assert_header_parsing)Retry)Timeout)get_host	parse_urlUrl_normalize_host_encode_target)	LifoQueuec@sBeZdZdZdZeZd
ddZddZddZ	d	d
Z
ddZdS)ConnectionPoolzz
    Base class for all connection pools, such as
    :class:`.HTTPConnectionPool` and :class:`.HTTPSConnectionPool`.
    NcCs0|stdt||jd|_||_||_dS)NzNo host specified.scheme)rr'r,hostlower_proxy_hostport)selfr-r0r2F/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/connectionpool.py__init__Is

zConnectionPool.__init__cCsdt|j|j|jfS)Nz%s(host=%r, port=%r))type__name__r-r0r1r2r2r3__str__QszConnectionPool.__str__cCs|SNr2r7r2r2r3	__enter__TszConnectionPool.__enter__cCs|dS)NF)close)r1exc_typeZexc_valZexc_tbr2r2r3__exit__WszConnectionPool.__exit__cCsdS)D
        Close all pooled connections and disable the pool.
        Nr2r7r2r2r3r;\szConnectionPool.close)N)r6
__module____qualname____doc__r,r)QueueClsr4r8r:r=r;r2r2r2r3r*@s
r*c
@seZdZdZdZeZeZdde	j
ddddddf	ddZdd	Zd!d
dZ
dd
ZddZddZddZddZedfddZddZddZddZdddddeddddf
dd ZdS)"HTTPConnectionPoolaN	
    Thread-safe connection pool for one host.

    :param host:
        Host used for this HTTP Connection (e.g. "localhost"), passed into
        :class:`httplib.HTTPConnection`.

    :param port:
        Port used for this HTTP Connection (None is equivalent to 80), passed
        into :class:`httplib.HTTPConnection`.

    :param strict:
        Causes BadStatusLine to be raised if the status line can't be parsed
        as a valid HTTP/1.0 or 1.1 status line, passed into
        :class:`httplib.HTTPConnection`.

        .. note::
           Only works in Python 2. This parameter is ignored in Python 3.

    :param timeout:
        Socket timeout in seconds for each individual connection. This can
        be a float or integer, which sets the timeout for the HTTP request,
        or an instance of :class:`urllib3.util.Timeout` which gives you more
        fine-grained control over request timeouts. After the constructor has
        been parsed, this is always a `urllib3.util.Timeout` object.

    :param maxsize:
        Number of connections to save that can be reused. More than 1 is useful
        in multithreaded situations. If ``block`` is set to False, more
        connections will be created but they will not be saved once they've
        been used.

    :param block:
        If set to True, no more than ``maxsize`` connections will be used at
        a time. When no free connections are available, the call will block
        until a connection has been released. This is a useful side effect for
        particular multithreaded situations where one does not want to use more
        than maxsize connections per host to prevent flooding.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param retries:
        Retry configuration to use by default with requests in this pool.

    :param _proxy:
        Parsed proxy URL, should not be used directly, instead, see
        :class:`urllib3.connectionpool.ProxyManager`"

    :param _proxy_headers:
        A dictionary with proxy headers, should not be used directly,
        instead, see :class:`urllib3.connectionpool.ProxyManager`"

    :param \**conn_kw:
        Additional parameters are used to create fresh :class:`urllib3.connection.HTTPConnection`,
        :class:`urllib3.connection.HTTPSConnection` instances.
    httpNFrc
Kst|||t||||_t|ts4t|}|dkrBtj}||_	||_
|||_||_
|	|_|
pli|_t|D]}|jdqxd|_d|_||_|jr|jdgdS)NrZsocket_options)r*r4rstrict
isinstancer#
from_floatr"ZDEFAULTrretriesrBpoolblockproxy
proxy_headersxrangeputnum_connectionsnum_requestsconn_kw
setdefault)
r1r-r0rErmaxsizerJheadersrH_proxy_proxy_headersrQ_r2r2r3r4s(


zHTTPConnectionPool.__init__cCsR|jd7_td|j|j|jp$d|jf|j|j|jj|jd|j	}|S)z9
        Return a fresh :class:`HTTPConnection`.
        rz(Starting new HTTP connection (%d): %s:%sZ80)r-r0rrE)
rOlogdebugr-r0
ConnectionClsrconnect_timeoutrErQr1connr2r2r3	_new_conns zHTTPConnectionPool._new_conncCsd}z|jj|j|d}WnBtk
r8t|dYn&tjk
r\|jrXt|dYnX|rt|rt	
d|j|t
|dddkrd}|p|S)	a
        Get a connection. Will return a pooled connection if one is available.

        If no connections are available and :prop:`.block` is ``False``, then a
        fresh connection is returned.

        :param timeout:
            Seconds to wait before giving up and raising
            :class:`urllib3.exceptions.EmptyPoolError` if the pool is empty and
            :prop:`.block` is ``True``.
        N)rJrzPool is closed.z>Pool reached maximum size and no more connections are allowed.z Resetting dropped connection: %sZ	auto_openrr)rIgetrJAttributeErrorrrEmptyrrrXrYr-r;getattrr^)r1rr]r2r2r3	_get_conns$zHTTPConnectionPool._get_conncCs^z|jj|ddWdStk
r*Yn$tjk
rLtd|jYnX|rZ|dS)a
        Put a connection back into the pool.

        :param conn:
            Connection object for the current host and port as returned by
            :meth:`._new_conn` or :meth:`._get_conn`.

        If the pool is already full, the connection is closed and discarded
        because we exceeded maxsize. If connections are discarded frequently,
        then maxsize should be increased.

        If the pool is closed, then the connection will be closed and discarded.
        FrJNz2Connection pool is full, discarding connection: %s)	rIrNr`rZFullrXwarningr-r;r\r2r2r3	_put_connszHTTPConnectionPool._put_conncCsdS)U
        Called right before a request is made, after the socket is created.
        Nr2r\r2r2r3_validate_conn0sz!HTTPConnectionPool._validate_conncCsdSr9r2r\r2r2r3_prepare_proxy6sz!HTTPConnectionPool._prepare_proxycCs2|tkr|jSt|tr$|St|SdS)z< Helper that always returns a :class:`urllib3.util.Timeout` N)_DefaultrZclonerFr#rG)r1rr2r2r3_get_timeout:s


zHTTPConnectionPool._get_timeoutcCsjt|trt||d|t|dr>|jtkr>t||d|dt|ksVdt|krft||d|dS)zAIs the error actually a timeout? Will raise a ReadTimeout or pass!Read timed out. (read timeout=%s)errnoz	timed outzdid not complete (read)N)rF
SocketTimeoutrhasattrrm_blocking_errnosstr)r1errurl
timeout_valuer2r2r3_raise_timeoutFs(
z!HTTPConnectionPool._raise_timeoutc

Ks|jd7_||}||j|_z||Wn:ttfk
rp}z|j|||jdW5d}~XYnX|r|j	||f|n|j
||f||j}	t|ddr|	dkrt
||d|	|	tjkr|jtn|j|	zjz|jdd}
WnTtk
rPz|}
Wn0tk
rJ}zt|dW5d}~XYnXYnXWn<tttfk
r}z|j|||	dW5d}~XYnXt|d	d
}td|j|j|j||||
j|
j 	zt!|
j"Wn@t#tfk
r}ztj$d|%||dd
W5d}~XYnX|
S)a
        Perform a request on a given urllib connection object taken from our
        pool.

        :param conn:
            a connection from one of our connection pools

        :param timeout:
            Socket timeout in seconds for the request. This can be a
            float or integer, which will set the same timeout value for
            the socket connect and the socket read, or an instance of
            :class:`urllib3.util.Timeout`, which gives you more fine-grained
            control over your timeouts.
        r)rrrsrtNsockrrlT)	bufferingZ
_http_vsn_strzHTTP/?z%s://%s:%s "%s %s %s" %s %sz$Failed to parse headers (url=%s): %s)exc_info)&rPrkZ
start_connectr[rrhrnrruZrequest_chunkedrequestread_timeoutrbrr#DEFAULT_TIMEOUTrvZ
settimeoutsocketZgetdefaulttimeoutZgetresponse	TypeError
BaseExceptionrZ
raise_fromSocketErrorrXrYr,r-r0statusZlengthr!msgr	re
_absolute_url)
r1r]methodrsrchunkedZhttplib_request_kwtimeout_objerzhttplib_responseZhttp_versionZhper2r2r3
_make_request_sp

(z HTTPConnectionPool._make_requestcCst|j|j|j|djS)N)r,r-r0path)r&r,r-r0rs)r1rr2r2r3rsz HTTPConnectionPool._absolute_urlcCsV|jdkrdS|jd}|_z|jdd}|r|qWntjk
rPYnXdS)r>NFrd)rIr_r;rra)r1Zold_poolr]r2r2r3r;s
zHTTPConnectionPool.closecCsx|drdSt|\}}}|dk	r0t||d}|jrF|sFt|}n|js^|t|kr^d}|||f|j|j|jfkS)zj
        Check if the given ``url`` is a member of the same host as this
        connection pool.
        /TNr+)
startswithr$r'r0rr_r,r-)r1rsr,r-r0r2r2r3is_same_hosts

zHTTPConnectionPool.is_same_hostTc
Ks|dkr|j}t|ts*tj|||jd}|
dkr>|
dd}
|rX||sXt||||drrt	
t|}nt	
t|j
}d}|
}|jdkr|}||jd}d}t||}ztz||}|j|	d}|j|_|jdk	ot|d	d}|r|||j|||||||d
}|
s(|nd}||
d<|jj|f|||d|
}d}Wntj k
rxt!|d
Ynt"t#t$t%t&t't(fk
r,}zd}t|t&t(frt'|}n>t|t$t)fr|jrt*d|}nt|t$t#frt%d|}|j+||||t,-dd}|.|}W5d}~XYnXW5|sJ|oD|}d}|rZ||X|st/0d||||j1|||||||f||	|
|d|
Sdd}|o|2}|rR|j3dkrd}z|j+||||d}Wn,t4k
r|j5r|||YSX|||6|t/7d|||j1||||f|||||	|
|d|
St8|9d}|:||j3|rz|j+||||d}Wn,t4k
r|j;r|||YSX|||.|t/7d||j1||||f|||||	|
|d|
S|S)a
        Get a connection from the pool and perform an HTTP request. This is the
        lowest level call for making a request, so you'll need to specify all
        the raw details.

        .. note::

           More commonly, it's appropriate to use a convenience method provided
           by :class:`.RequestMethods`, such as :meth:`request`.

        .. note::

           `release_conn` will only behave as expected if
           `preload_content=False` because we want to make
           `preload_content=False` the default behaviour someday soon without
           breaking backwards compatibility.

        :param method:
            HTTP request method (such as GET, POST, PUT, etc.)

        :param body:
            Data to send in the request body (useful for creating
            POST requests, see HTTPConnectionPool.post_url for
            more convenience).

        :param headers:
            Dictionary of custom headers to send, such as User-Agent,
            If-None-Match, etc. If None, pool headers are used. If provided,
            these headers completely replace any pool-specific headers.

        :param retries:
            Configure the number of retries to allow before raising a
            :class:`~urllib3.exceptions.MaxRetryError` exception.

            Pass ``None`` to retry until you receive a response. Pass a
            :class:`~urllib3.util.retry.Retry` object for fine-grained control
            over different types of retries.
            Pass an integer number to retry connection errors that many times,
            but no other types of errors. Pass zero to never retry.

            If ``False``, then retries are disabled and any exception is raised
            immediately. Also, instead of raising a MaxRetryError on redirects,
            the redirect response will be returned.

        :type retries: :class:`~urllib3.util.retry.Retry`, False, or an int.

        :param redirect:
            If True, automatically handle redirects (status codes 301, 302,
            303, 307, 308). Each redirect counts as a retry. Disabling retries
            will disable redirect, too.

        :param assert_same_host:
            If ``True``, will make sure that the host of the pool requests is
            consistent else will raise HostChangedError. When False, you can
            use the pool on an HTTP proxy and request foreign hosts.

        :param timeout:
            If specified, overrides the default timeout for this one
            request. It may be a float (in seconds) or an instance of
            :class:`urllib3.util.Timeout`.

        :param pool_timeout:
            If set and the pool is set to block=True, then this method will
            block for ``pool_timeout`` seconds and raise EmptyPoolError if no
            connection is available within the time period.

        :param release_conn:
            If False, then the urlopen call will not release the connection
            back into the pool once a response is received (but will release if
            you read the entire contents of the response such as when
            `preload_content=True`). This is useful if you're not preloading
            the response's content immediately. You will need to call
            ``r.release_conn()`` on the response ``r`` to return the connection
            back into the pool. If None, it takes the value of
            ``response_kw.get('preload_content', True)``.

        :param chunked:
            If True, urllib3 will send the body using chunked transfer
            encoding. Otherwise, urllib3 will send the body using the standard
            content-length form. Defaults to False.

        :param int body_pos:
            Position to seek to in file-like body in the event of a retry or
            redirect. Typically this won't need to be set because urllib3 will
            auto-populate the value when needed.

        :param \**response_kw:
            Additional parameters are passed to
            :meth:`urllib3.response.HTTPResponse.from_httplib`
        N)redirectdefaultZpreload_contentTrrDF)rrv)rbodyrTrZrequest_method)rI
connectionrHz"No pool connections are available.zCannot connect to proxy.zConnection aborted.)r_poolZ_stacktracez1Retrying (%r) after connection broken by '%r': %s)rpool_timeoutrelease_connbody_posc
Ss2z|Wn ttttttfk
r,YnXdSr9)readrrrrrr)responser2r2r3drain_and_release_connsz:HTTPConnectionPool.urlopen..drain_and_release_conni/ZGET)rrzRedirecting %s -> %s)rHrassert_same_hostrrrrzRetry-Afterz	Retry: %s)>> conn = connection_from_url('http://google.com/')
        >>> r = conn.request('GET', '/')
    Prr0N)r$rr_rrC)rskwr,r-r0r2r2r3connection_from_urls
rcCs.t||}|dr*|dr*|dd}|S)z?
    Normalize hosts for comparisons and use with sockets.
    []r)normalize_hostrendswith)r-r,r2r2r3r's
r')LZ
__future__rrmZloggingrrr|rrrrn
exceptionsrrrr	r
rrr
rrrrrZpackages.ssl_match_hostnamerZpackagesrZpackages.six.movesrrrrrrrrrryrrrZutil.connectionrZutil.requestr Z
util.responser!Z
util.retryr"Zutil.timeoutr#Zutil.urlr$r%r&r'rr(Z
util.queuer)ZmovesrMZ	getLoggerr6rXobjectrjr*ZEAGAINZEWOULDBLOCKrprCrrr2r2r2r3sF<$	
$rsite-packages/pip/_vendor/urllib3/__pycache__/fields.cpython-38.pyc000064400000017643151464452710021216 0ustar00U

.ei!@sddlmZddlZddlZddlZddlmZdddZdd	Z	d
dddZ
e
d
deddDddZ
ddZeZGdddeZdS))absolute_importN)sixapplication/octet-streamcCs|rt|dp|S|S)z
    Guess the "Content-Type" of a file.

    :param filename:
        The filename to guess the "Content-Type" of using :mod:`mimetypes`.
    :param default:
        If no "Content-Type" can be guessed, default to `default`.
    r)	mimetypesZ
guess_type)filenamedefaultr	>/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/fields.pyguess_content_type	s	rc	sttjrdtfdddDsdd|f}z|dWnttfk
r^YnX|Stjrtdt	j
dd|ftjrdS)a
    Helper function to format and quote a single header parameter using the
    strategy defined in RFC 2231.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows RFC 2388 Section 4.4.

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        An RFC-2231-formatted unicode string.
    utf-8c3s|]}|kVqdSNr	).0Zchvaluer	r
	)sz.format_header_param_rfc2231..z"\
%s="%s"asciiz%s*=%s)
isinstancerbinary_typedecodeanyencodeUnicodeEncodeErrorUnicodeDecodeErrorZPY2emailZutilsZencode_rfc2231)namerresultr	rr
format_header_param_rfc2231s 


rz%22z\\)"\r cCs&i|]}|dkrt|d|qS))z%{:02X})rZunichrformat)rZccr	r	r

Isr# cs:fdd}tdddD}|||}|S)Ncs|dS)Nr)group)matchneedles_and_replacementsr	r
replacerRsz#_replace_multiple..replacer|cSsg|]}t|qSr	)reescape)rZneedler	r	r

Vsz%_replace_multiple..)r+compilejoinkeyssub)rr(r)patternrr	r'r
_replace_multipleQsr3cCs,t|tjr|d}t|t}d||fS)a
    Helper function to format and quote a single header parameter using the
    HTML5 strategy.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows the `HTML5 Working Draft
    Section 4.10.22.7`_ and matches the behavior of curl and modern browsers.

    .. _HTML5 Working Draft Section 4.10.22.7:
        https://w3c.github.io/html/sec-forms.html#multipart-form-data

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        A unicode string, stripped of troublesome characters.
    rr)rrrrr3_HTML5_REPLACEMENTS)rrr	r	r
format_header_param_html5^s

r5c@sReZdZdZddefddZeefddZddZd	d
Z	ddZ
dd
dZdS)RequestFielda
    A data container for request body parameters.

    :param name:
        The name of this request field. Must be unicode.
    :param data:
        The data/value body.
    :param filename:
        An optional filename of the request field. Must be unicode.
    :param headers:
        An optional dict-like object of headers to initially use for the field.
    :param header_formatter:
        An optional callable that is used to encode and format the headers. By
        default, this is :func:`format_header_param_html5`.
    NcCs0||_||_||_i|_|r&t||_||_dSr
)_name	_filenamedataheadersdictheader_formatter)selfrr9rr:r<r	r	r
__init__s
zRequestField.__init__cCs`t|tr4t|dkr"|\}}}q@|\}}t|}nd}d}|}|||||d}|j|d|S)a
        A :class:`~urllib3.fields.RequestField` factory from old-style tuple parameters.

        Supports constructing :class:`~urllib3.fields.RequestField` from
        parameter of key/value strings AND key/filetuple. A filetuple is a
        (filename, data, MIME type) tuple where the MIME type is optional.
        For example::

            'foo': 'bar',
            'fakefile': ('foofile.txt', 'contents of foofile'),
            'realfile': ('barfile.txt', open('realfile').read()),
            'typedfile': ('bazfile.bin', open('bazfile').read(), 'image/jpeg'),
            'nonamefile': 'contents of nonamefile field',

        Field names and filenames must be unicode.
        N)rr<)content_type)rtuplelenrmake_multipart)clsZ	fieldnamerr<rr9r@Z
request_paramr	r	r
from_tupless 

zRequestField.from_tuplescCs|||S)aI
        Overridable helper function to format a single header parameter. By
        default, this calls ``self.header_formatter``.

        :param name:
            The name of the parameter, a string expected to be ASCII only.
        :param value:
            The value of the parameter, provided as a unicode string.
        )r<)r=rrr	r	r
_render_partszRequestField._render_partcCsLg}|}t|tr|}|D]"\}}|dk	r||||qd|S)aO
        Helper function to format and quote a single header.

        Useful for single headers that are composed of multiple items. E.g.,
        'Content-Disposition' fields.

        :param header_parts:
            A sequence of (k, v) tuples or a :class:`dict` of (k, v) to format
            as `k1="v1"; k2="v2"; ...`.
        N; )rr;itemsappendrFr/)r=Zheader_partspartsiterablerrr	r	r

_render_partss
zRequestField._render_partscCsg}dddg}|D]*}|j|dr|d||j|fq|jD]&\}}||krH|rH|d||fqH|dd|S)z=
        Renders the headers for this request field.
        Content-DispositionContent-TypeContent-LocationFz%s: %sz
)r:getrIrHr/)r=linesZ	sort_keysZsort_keyZheader_nameZheader_valuer	r	r
render_headerss

zRequestField.render_headerscCsX|pd|jd<|jddd|d|jfd|jffg7<||jd<||jd<d	S)
a|
        Makes this request field into a multipart request field.

        This method overrides "Content-Disposition", "Content-Type" and
        "Content-Location" headers to the request parameter.

        :param content_type:
            The 'Content-Type' of the request body.
        :param content_location:
            The 'Content-Location' of the request body.

        z	form-datarMrGrrrNrON)r:r/rLr7r8)r=Zcontent_dispositionr@Zcontent_locationr	r	r
rCs
zRequestField.make_multipart)NNN)__name__
__module____qualname____doc__r5r>classmethodrErFrLrRrCr	r	r	r
r6}s
#
r6)r)Z
__future__rZemail.utilsrrr+Zpackagesrrrr4updateranger3r5Zformat_header_paramobjectr6r	r	r	r
s&
*	
site-packages/pip/_vendor/urllib3/__pycache__/__init__.cpython-38.pyc000064400000004045151464452710021477 0ustar00U

.e{
@s4dZddlmZddlZddlmZmZmZddlm	Z	ddl
mZddlm
Z
mZmZdd	lmZdd
lmZddlmZddlmZdd
lmZddlZddlmZdZdZdZdZe e!"eej#fddZ$[ej%de	j&ddej%de	j'ddej%de	j(ddej%de	j)dde	j*fddZ+dS)z8
urllib3 - Thread-safe connection pooling and re-using.
)absolute_importN)HTTPConnectionPoolHTTPSConnectionPoolconnection_from_url)
exceptions)encode_multipart_formdata)PoolManagerProxyManagerproxy_from_url)HTTPResponse)make_headers)get_host)Timeout)Retry)NullHandlerz(Andrey Petrov (andrey.petrov@shazow.net)ZMITz1.25.6)rrr	r
rrradd_stderr_loggerrdisable_warningsrrr
rcCsFtt}t}|td|||||dt|S)z
    Helper for quickly adding a StreamHandler to the logger. Useful for
    debugging.

    Returns the handler after adding it.
    z%%(asctime)s %(levelname)s %(message)sz,Added a stderr logging handler to logger: %s)	logging	getLogger__name__Z
StreamHandlerZsetFormatterZ	Formatter
addHandlerZsetLeveldebug)levelZloggerZhandlerr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/__init__.pyr/s	


ralwaysT)appenddefaultcCstd|dS)z<
    Helper for quickly disabling all urllib3 warnings.
    ignoreN)warningssimplefilter)categoryrrrrRsr),__doc__Z
__future__rr ZconnectionpoolrrrrZfilepostrZpoolmanagerr	r
rZresponserZutil.requestr
Zutil.urlrZutil.timeoutrZ
util.retryrrr
__author__Z__license____version____all__rrrDEBUGrr!ZSecurityWarningZSubjectAltNameWarningZInsecurePlatformWarningZSNIMissingWarningZHTTPWarningrrrrrs2site-packages/pip/_vendor/urllib3/__pycache__/filepost.cpython-38.pyc000064400000005256151464452710021572 0ustar00U

.e	@sddlmZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZe
dd	Zd
dZdd
ZddZdddZdS))absolute_importN)BytesIO)six)b)RequestFieldzutf-8cCs$ttd}tjs |d}|S)zN
    Our embarrassingly-simple replacement for mimetools.choose_boundary.
    ascii)binasciiZhexlifyosurandomrZPY2decode)boundaryr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/filepost.pychoose_boundarys
rccsJt|trt|}nt|}|D]"}t|tr8|Vq"tj|Vq"dS)z
    Iterate over fields.

    Supports list of (k, v) tuples and dicts, and lists of
    :class:`~urllib3.fields.RequestField`.

    N)
isinstancedictr	iteritemsiterrZfrom_tuples)fieldsifieldrrriter_field_objectss

rcCs,t|trddt|DSdd|DS)a-
    .. deprecated:: 1.6

    Iterate over fields.

    The addition of :class:`~urllib3.fields.RequestField` makes this function
    obsolete. Instead, use :func:`iter_field_objects`, which returns
    :class:`~urllib3.fields.RequestField` objects.

    Supports list of (k, v) tuples and dicts.
    css|]\}}||fVqdSNr.0kvrrr	:sziter_fields..css|]\}}||fVqdSrrrrrrr <s)rrrr)rrrriter_fields-s
r!cCst}|dkrt}t|D]p}|td|t|||j}t|t	r\t
|}t|tjrxt||n
|||dq|td|t
d|}|
|fS)aQ
    Encode a dictionary of ``fields`` using the multipart/form-data MIME format.

    :param fields:
        Dictionary of fields or list of (key, :class:`~urllib3.fields.RequestField`).

    :param boundary:
        If not specified, then a random boundary will be generated using
        :func:`urllib3.filepost.choose_boundary`.
    Nz--%s
s
z--%s--
z multipart/form-data; boundary=%s)rrrwriterwriterZrender_headersdatarintstrrZ	text_typegetvalue)rrZbodyrr$Zcontent_typerrrencode_multipart_formdata?s 

r()N)Z
__future__rrcodecsriorZpackagesrZpackages.sixrrrlookupr#rrr!r(rrrrs
site-packages/pip/_vendor/urllib3/__pycache__/exceptions.cpython-38.pyc000064400000023454151464452710022126 0ustar00U

.e@sLddlmZddlmZGdddeZGdddeZGdd	d	eZ	Gd
dde	Z
Gdd
d
eZGdddeZGdddeZ
GdddeZeZGddde
ZGddde
ZGdddeZGdddeZGdddee
ZGdddeZGd d!d!ee	ZGd"d#d#e	ZGd$d%d%e	ZGd&d'd'eeZGd(d)d)eZGd*d+d+eZGd,d-d-eZGd.d/d/eZGd0d1d1eZGd2d3d3eZ Gd4d5d5eZ!Gd6d7d7eZ"Gd8d9d9eZ#Gd:d;d;eeZ$Gdd?d?eeZGd@dAdAeZ&GdBdCdCe'eZ(GdDdEdEeZ)GdFdGdGeZ*dHS)I)absolute_import)IncompleteReadc@seZdZdZdS)	HTTPErrorz#Base exception used by this module.N__name__
__module____qualname____doc__rrB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/exceptions.pyrsrc@seZdZdZdS)HTTPWarningz!Base warning used by this module.Nrrrrrr
sr
c@s eZdZdZddZddZdS)	PoolErrorz/Base exception for errors caused within a pool.cCs||_t|d||fdS)Nz%s: %s)poolr__init__)selfrmessagerrrrszPoolError.__init__cCs
|jdfS)N)NN	__class__rrrr
__reduce__szPoolError.__reduce__Nrrr	r
rrrrrrrsrc@s eZdZdZddZddZdS)RequestErrorz8Base exception for PoolErrors that have associated URLs.cCs||_t|||dSN)urlrr)rrrrrrrr szRequestError.__init__cCs|jd|jdffSr)rrrrrrr$szRequestError.__reduce__Nrrrrrrsrc@seZdZdZdS)SSLErrorz9Raised when SSL certificate fails in an HTTPS connection.Nrrrrrr)src@seZdZdZdS)
ProxyErrorz,Raised when the connection to a proxy fails.Nrrrrrr.src@seZdZdZdS)DecodeErrorz;Raised when automatic decoding based on Content-Type fails.Nrrrrrr3src@seZdZdZdS)
ProtocolErrorz>Raised when something unexpected happens mid-request/response.Nrrrrrr8src@seZdZdZdddZdS)
MaxRetryErroraRaised when the maximum number of retries is exceeded.

    :param pool: The connection pool
    :type pool: :class:`~urllib3.connectionpool.HTTPConnectionPool`
    :param string url: The requested Url
    :param exceptions.Exception reason: The underlying error

    NcCs&||_d||f}t||||dS)Nz0Max retries exceeded with url: %s (Caused by %r))reasonrr)rrrr rrrrrNszMaxRetryError.__init__)Nrrr	r
rrrrrrDs	rc@seZdZdZdddZdS)HostChangedErrorz?Raised when an existing pool gets a request for a foreign host.cCs"d|}t||||||_dS)Nz)Tried to open a foreign host with url: %s)rrretries)rrrr$rrrrrYszHostChangedError.__init__N)r#r!rrrrr"Vsr"c@seZdZdZdS)TimeoutStateErrorz3 Raised when passing an invalid state to a timeout Nrrrrrr%_sr%c@seZdZdZdS)TimeoutErrorz Raised when a socket timeout error occurs.

    Catching this error will catch both :exc:`ReadTimeoutErrors
    ` and :exc:`ConnectTimeoutErrors `.
    Nrrrrrr&esr&c@seZdZdZdS)ReadTimeoutErrorzFRaised when a socket timeout occurs while receiving data from a serverNrrrrrr'osr'c@seZdZdZdS)ConnectTimeoutErrorz@Raised when a socket timeout occurs while connecting to a serverNrrrrrr(vsr(c@seZdZdZdS)NewConnectionErrorzHRaised when we fail to establish a new connection. Usually ECONNREFUSED.Nrrrrrr){sr)c@seZdZdZdS)EmptyPoolErrorzCRaised when a pool runs out of connections and no more are allowed.Nrrrrrr*sr*c@seZdZdZdS)ClosedPoolErrorzCRaised when a request enters a pool after the pool has been closed.Nrrrrrr+sr+c@seZdZdZdS)LocationValueErrorzsr>cs eZdZdZfddZZS)ProxySchemeUnknownz1ProxyManager does not support the supplied schemecsd|}tt||dS)NzNot supported proxy scheme %s)r9r?r)rZschemerrrrrszProxySchemeUnknown.__init__rrr	r
rr=rrrrr?sr?cs eZdZdZfddZZS)HeaderParsingErrorzNRaised by assert_header_parsing, but we convert it to a log.warning statement.cs$d|pd|f}tt||dS)Nz%s, unparsed data: %rZUnknown)r9rAr)rZdefectsZ
unparsed_datarrrrrszHeaderParsingError.__init__r@rrrrrAsrAc@seZdZdZdS)UnrewindableBodyErrorz9urllib3 encountered an error when trying to rewind a bodyNrrrrrrBsrBN)+Z
__future__rZpackages.six.moves.http_clientrZhttplib_IncompleteRead	ExceptionrWarningr
rrrrrrConnectionErrorrr"r%r&r'r(r)r*r+
ValueErrorr,r-r/r0r1r2r3r4r5r6r7r8r>AssertionErrorr?rArBrrrrsH	

			site-packages/pip/_vendor/urllib3/__pycache__/filepost.cpython-38.opt-1.pyc000064400000005256151464452710022531 0ustar00U

.e	@sddlmZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZe
dd	Zd
dZdd
ZddZdddZdS))absolute_importN)BytesIO)six)b)RequestFieldzutf-8cCs$ttd}tjs |d}|S)zN
    Our embarrassingly-simple replacement for mimetools.choose_boundary.
    ascii)binasciiZhexlifyosurandomrZPY2decode)boundaryr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/filepost.pychoose_boundarys
rccsJt|trt|}nt|}|D]"}t|tr8|Vq"tj|Vq"dS)z
    Iterate over fields.

    Supports list of (k, v) tuples and dicts, and lists of
    :class:`~urllib3.fields.RequestField`.

    N)
isinstancedictr	iteritemsiterrZfrom_tuples)fieldsifieldrrriter_field_objectss

rcCs,t|trddt|DSdd|DS)a-
    .. deprecated:: 1.6

    Iterate over fields.

    The addition of :class:`~urllib3.fields.RequestField` makes this function
    obsolete. Instead, use :func:`iter_field_objects`, which returns
    :class:`~urllib3.fields.RequestField` objects.

    Supports list of (k, v) tuples and dicts.
    css|]\}}||fVqdSNr.0kvrrr	:sziter_fields..css|]\}}||fVqdSrrrrrrr <s)rrrr)rrrriter_fields-s
r!cCst}|dkrt}t|D]p}|td|t|||j}t|t	r\t
|}t|tjrxt||n
|||dq|td|t
d|}|
|fS)aQ
    Encode a dictionary of ``fields`` using the multipart/form-data MIME format.

    :param fields:
        Dictionary of fields or list of (key, :class:`~urllib3.fields.RequestField`).

    :param boundary:
        If not specified, then a random boundary will be generated using
        :func:`urllib3.filepost.choose_boundary`.
    Nz--%s
s
z--%s--
z multipart/form-data; boundary=%s)rrrwriterwriterZrender_headersdatarintstrrZ	text_typegetvalue)rrZbodyrr$Zcontent_typerrrencode_multipart_formdata?s 

r()N)Z
__future__rrcodecsriorZpackagesrZpackages.sixrrrlookupr#rrr!r(rrrrs
site-packages/pip/_vendor/urllib3/__pycache__/request.cpython-38.pyc000064400000012766151464452710021441 0ustar00U

.e@s>ddlmZddlmZddlmZdgZGdddeZdS))absolute_import)encode_multipart_formdata)	urlencodeRequestMethodsc@sNeZdZdZddddhZdddZdd
dZddd
ZdddZdddZ	dS)ra
    Convenience mixin for classes who implement a :meth:`urlopen` method, such
    as :class:`~urllib3.connectionpool.HTTPConnectionPool` and
    :class:`~urllib3.poolmanager.PoolManager`.

    Provides behavior for making common types of HTTP request methods and
    decides which type of request field encoding to use.

    Specifically,

    :meth:`.request_encode_url` is for sending requests whose fields are
    encoded in the URL (such as GET, HEAD, DELETE).

    :meth:`.request_encode_body` is for sending requests whose fields are
    encoded in the *body* of the request using multipart or www-form-urlencoded
    (such as for POST, PUT, PATCH).

    :meth:`.request` is for making any kind of request, it will look up the
    appropriate encoding format and use one of the above two methods to make
    the request.

    Initializer parameters:

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.
    ZDELETEZGETZHEADZOPTIONSNcCs|pi|_dS)N)headers)selfrr	?/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/request.py__init__)szRequestMethods.__init__TcKstddS)NzMClasses extending RequestMethods must implement their own ``urlopen`` method.)NotImplementedError)rmethodurlbodyrencode_multipartmultipart_boundarykwr	r	r
urlopen,s
zRequestMethods.urlopencKsR|}||d<||jkr4|j||f||d|S|j||f||d|SdS)a
        Make a request using :meth:`urlopen` with the appropriate encoding of
        ``fields`` based on the ``method`` used.

        This is a convenience method that requires the least amount of manual
        effort. It can be used in most situations, while still having the
        option to drop down to more specific methods when necessary, such as
        :meth:`request_encode_url`, :meth:`request_encode_body`,
        or even the lowest level :meth:`urlopen`.
        Zrequest_url)fieldsrN)upper_encode_url_methodsrequest_encode_urlrequest_encode_body)rr
rrr
urlopen_kwr	r	r
request;s*
zRequestMethods.requestcKsD|dkr|j}d|i}|||r4|dt|7}|j||f|S)z
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the url. This is useful for request methods like GET, HEAD, DELETE, etc.
        Nr?)rupdaterr)rr
rrrrextra_kwr	r	r
rSs
z!RequestMethods.request_encode_urlcKs|dkr|j}dii}|rbd|kr*td|r@t||d\}	}
nt|d}	}
|	|d<d|
i|d<|d||||j||f|S)a
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the body. This is useful for request methods like POST, PUT, PATCH, etc.

        When ``encode_multipart=True`` (default), then
        :meth:`urllib3.filepost.encode_multipart_formdata` is used to encode
        the payload with the appropriate content type. Otherwise
        :meth:`urllib.urlencode` is used with the
        'application/x-www-form-urlencoded' content type.

        Multipart encoding must be used when posting files, and it's reasonably
        safe to use it in other times too. However, it may break request
        signing, such as with OAuth.

        Supports an optional ``fields`` parameter of key/value strings AND
        key/filetuple. A filetuple is a (filename, data, MIME type) tuple where
        the MIME type is optional. For example::

            fields = {
                'foo': 'bar',
                'fakefile': ('foofile.txt', 'contents of foofile'),
                'realfile': ('barfile.txt', open('realfile').read()),
                'typedfile': ('bazfile.bin', open('bazfile').read(),
                              'image/jpeg'),
                'nonamefile': 'contents of nonamefile field',
            }

        When uploading a file, providing a filename (the first parameter of the
        tuple) is optional but recommended to best mimic behavior of browsers.

        Note that if ``headers`` are supplied, the 'Content-Type' header will
        be overwritten because it depends on the dynamic random boundary string
        which is used to compose the body of the request. The random boundary
        string can be explicitly set with the ``multipart_boundary`` parameter.
        NrrzFrequest got values for both 'fields' and 'body', can only specify one.)boundaryz!application/x-www-form-urlencodedzContent-Type)r	TypeErrorrrrr)rr
rrrrrrrrZcontent_typer	r	r
rcs*-
z"RequestMethods.request_encode_body)N)NNTN)NN)NN)NNTN)
__name__
__module____qualname____doc__rrrrrrr	r	r	r
r
s



N)	Z
__future__rZfilepostrZpackages.six.moves.urllib.parser__all__objectrr	r	r	r
ssite-packages/pip/_vendor/urllib3/__pycache__/poolmanager.cpython-38.opt-1.pyc000064400000031150151464452710023200 0ustar00U

.eB@sddlmZddlZddlZddlZddlmZddlmZm	Z	ddlm
Z
ddlmZm
Z
mZddlmZdd	lmZdd
lmZddlmZddlmZd
ddgZeeZdZdZedeZ ddZ!e"e!e e"e!e dZ#ee	dZ$Gdd
d
eZ%Gddde%Z&ddZ'dS))absolute_importN)RecentlyUsedContainer)HTTPConnectionPoolHTTPSConnectionPool)port_by_scheme)LocationValueError
MaxRetryErrorProxySchemeUnknown)six)urljoin)RequestMethods)	parse_url)RetryPoolManagerProxyManagerproxy_from_url)Zkey_fileZ	cert_fileZ	cert_reqsZca_certsZssl_versionZca_cert_dirZssl_contextZkey_password)Z
key_schemeZkey_hostZkey_portZkey_timeoutZkey_retriesZ
key_strictZ	key_blockZkey_source_addressZkey_key_fileZkey_key_passwordZ
key_cert_fileZ
key_cert_reqsZkey_ca_certsZkey_ssl_versionZkey_ca_cert_dirZkey_ssl_contextZkey_maxsizeZkey_headersZ
key__proxyZkey__proxy_headersZkey_socket_optionsZkey__socks_optionsZkey_assert_hostnameZkey_assert_fingerprintZkey_server_hostnamePoolKeycCs|}|d|d<|d|d<dD],}||kr,||dk	r,t||||<q,|d}|dk	rxt||d<t|D]}|||d|<q|j	D]}||krd||<q|f|S)a
    Create a pool key out of a request context dictionary.

    According to RFC 3986, both the scheme and host are case-insensitive.
    Therefore, this function normalizes both before constructing the pool
    key for an HTTPS request. If you wish to change this behaviour, provide
    alternate callables to ``key_fn_by_scheme``.

    :param key_class:
        The class to use when constructing the key. This should be a namedtuple
        with the ``scheme`` and ``host`` keys at a minimum.
    :type  key_class: namedtuple
    :param request_context:
        A dictionary-like object that contain the context for a request.
    :type  request_context: dict

    :return: A namedtuple that can be used as a connection pool key.
    :rtype:  PoolKey
    schemehost)headers_proxy_headersZ_socks_optionsNZsocket_optionsZkey_)
copylower	frozensetitemsgettuplelistkeyspop_fields)Z	key_classrequest_contextcontextkeyZsocket_optsZfieldr%C/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/poolmanager.py_default_key_normalizerDs


r'httphttpsc@sxeZdZdZdZdddZddZdd	Zdd
dZdd
Z	dddZ
ddZdddZd ddZ
ddZd!ddZdS)"ra$
    Allows for arbitrary requests while transparently keeping track of
    necessary connection pools for you.

    :param num_pools:
        Number of connection pools to cache before discarding the least
        recently used pool.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param \**connection_pool_kw:
        Additional parameters are used to create fresh
        :class:`urllib3.connectionpool.ConnectionPool` instances.

    Example::

        >>> manager = PoolManager(num_pools=2)
        >>> r = manager.request('GET', 'http://google.com/')
        >>> r = manager.request('GET', 'http://google.com/mail')
        >>> r = manager.request('GET', 'http://yahoo.com/')
        >>> len(manager.pools)
        2

    N
cKs8t||||_t|ddd|_t|_t|_dS)NcSs|SN)close)pr%r%r&z&PoolManager.__init__..)Zdispose_func)r
__init__connection_pool_kwrpoolspool_classes_by_schemekey_fn_by_schemer)self	num_poolsrr2r%r%r&r1s
zPoolManager.__init__cCs|Sr,r%r6r%r%r&	__enter__szPoolManager.__enter__cCs|dS)NF)clear)r6exc_typeZexc_valZexc_tbr%r%r&__exit__szPoolManager.__exit__cCs^|j|}|dkr|j}dD]}||dq |dkrPtD]}||dq>|||f|S)a
        Create a new :class:`ConnectionPool` based on host, port, scheme, and
        any additional pool keyword arguments.

        If ``request_context`` is provided, it is provided as keyword arguments
        to the pool class used. This method is used to actually create the
        connection pools handed out by :meth:`connection_from_url` and
        companion methods. It is intended to be overridden for customization.
        N)rrportr))r4r2rr SSL_KEYWORDS)r6rrr=r"Zpool_clsr$kwr%r%r&	_new_pools


zPoolManager._new_poolcCs|jdS)z
        Empty our store of pools and direct them all to close.

        This will not affect in-flight connections, but they will not be
        re-used after completion.
        N)r3r:r8r%r%r&r:szPoolManager.clearr)cCsT|std||}|pd|d<|s:t|dd}||d<||d<||S)a
        Get a :class:`ConnectionPool` based on the host, port, and scheme.

        If ``port`` isn't given, it will be derived from the ``scheme`` using
        ``urllib3.connectionpool.port_by_scheme``. If ``pool_kwargs`` is
        provided, it is merged with the instance's ``connection_pool_kw``
        variable and used to create the new connection pool, if one is
        needed.
        zNo host specified.r)rPr=r)r_merge_pool_kwargsrrrconnection_from_context)r6rr=rpool_kwargsr"r%r%r&connection_from_hosts
z PoolManager.connection_from_hostcCs,|d}|j|}||}|j||dS)z
        Get a :class:`ConnectionPool` based on the request context.

        ``request_context`` must at least contain the ``scheme`` key and its
        value must be a key in ``key_fn_by_scheme`` instance variable.
        rr")rr5connection_from_pool_key)r6r"rZpool_key_constructorpool_keyr%r%r&rCs
z#PoolManager.connection_from_contextc
Csl|jjZ|j|}|r*|W5QRS|d}|d}|d}|j||||d}||j|<W5QRX|S)z
        Get a :class:`ConnectionPool` based on the provided pool key.

        ``pool_key`` should be a namedtuple that only contains immutable
        objects. At a minimum it must have the ``scheme``, ``host``, and
        ``port`` fields.
        rrr=rF)r3lockrr@)r6rHr"Zpoolrrr=r%r%r&rGs
z$PoolManager.connection_from_pool_keycCs t|}|j|j|j|j|dS)a
        Similar to :func:`urllib3.connectionpool.connection_from_url`.

        If ``pool_kwargs`` is not provided and a new pool needs to be
        constructed, ``self.connection_pool_kw`` is used to initialize
        the :class:`urllib3.connectionpool.ConnectionPool`. If ``pool_kwargs``
        is provided, it is used instead. Note that if a new pool does not
        need to be created for the request, the provided ``pool_kwargs`` are
        not used.
        )r=rrD)rrErr=r)r6urlrDur%r%r&connection_from_urlszPoolManager.connection_from_urlc	CsV|j}|rR|D]:\}}|dkrHz
||=WqPtk
rDYqPXq|||<q|S)a
        Merge a dictionary of override values for self.connection_pool_kw.

        This does not modify self.connection_pool_kw and returns a new dict.
        Any keys in the override dictionary with a value of ``None`` are
        removed from the merged dictionary.
        N)r2rrKeyError)r6overrideZbase_pool_kwargsr$valuer%r%r&rB!s


zPoolManager._merge_pool_kwargsTcKst|}|j|j|j|jd}d|d<d|d<d|krD|j|d<|jdk	rj|jdkrj|j||f|}n|j||j	f|}|o|
}|s|St||}|jdkrd	}|
d
}	t|	tstj|	|d}	|	jr||stt|d}
|
D]"}||	jkr|d|dqz|	j||||d}	Wn$tk
rR|	jrJ|YSX|	|d
<||d<td
|||j||f|S)a]
        Same as :meth:`urllib3.connectionpool.HTTPConnectionPool.urlopen`
        with custom cross-host redirect logic and only sends the request-uri
        portion of the ``url``.

        The given ``url`` parameter must be absolute, such that an appropriate
        :class:`urllib3.connectionpool.ConnectionPool` can be chosen for it.
        )r=rFZassert_same_hostredirectrNr)i/ZGETretries)rP)responseZ_poolzRedirecting %s -> %s)rrErr=rrrproxyurlopenZrequest_uriZget_redirect_locationrZstatusr
isinstancerZfrom_intZremove_headers_on_redirectZis_same_hostrrZiterkeysrr Z	incrementr	Zraise_on_redirectloginfo)r6methodrJrPr?rKZconnrRZredirect_locationrQrheaderr%r%r&rT5sF	




zPoolManager.urlopen)r+N)N)Nr)N)N)N)T)__name__
__module____qualname____doc__rSr1r9r<r@r:rErCrGrLrBrTr%r%r%r&rs


	


csHeZdZdZdfdd	Zdfdd	Zdd	d
Zdfdd
	ZZS)rax
    Behaves just like :class:`PoolManager`, but sends all requests through
    the defined proxy, using the CONNECT method for HTTPS URLs.

    :param proxy_url:
        The URL of the proxy to be used.

    :param proxy_headers:
        A dictionary containing headers that will be sent to the proxy. In case
        of HTTP they are being sent with each request, while in the
        HTTPS/CONNECT case they are sent only once. Could be used for proxy
        authentication.

    Example:
        >>> proxy = urllib3.ProxyManager('http://localhost:3128/')
        >>> r1 = proxy.request('GET', 'http://google.com/')
        >>> r2 = proxy.request('GET', 'http://httpbin.org/')
        >>> len(proxy.pools)
        1
        >>> r3 = proxy.request('GET', 'https://httpbin.org/')
        >>> r4 = proxy.request('GET', 'https://twitter.com/')
        >>> len(proxy.pools)
        3

    r+Ncst|trd|j|j|jf}t|}|jsFt|jd}|j|d}|jdkrZt	|j||_
|pfi|_|j
|d<|j|d<tt
|j||f|dS)Nz
%s://%s:%irA)r=r(_proxyr)rUrrrr=rrr_replacer
rS
proxy_headerssuperrr1)r6	proxy_urlr7rr`r2rSr=	__class__r%r&r1s"	





zProxyManager.__init__r)csD|dkr tt|j||||dStt|j|jj|jj|jj|dS)Nr*)rD)rarrErSrr=r)r6rr=rrDrcr%r&rEs

z!ProxyManager.connection_from_hostcCs0ddi}t|j}|r||d<|r,|||S)z
        Sets headers needed by proxies: specifically, the Accept and Host
        headers. Only sets headers not provided by the user.
        ZAcceptz*/*ZHost)rnetlocupdate)r6rJrZheaders_rer%r%r&_set_proxy_headerss

zProxyManager._set_proxy_headersTcsNt|}|jdkr0|d|j}||||d<tt|j||fd|i|S)z@Same as HTTP(S)ConnectionPool.urlopen, ``url`` must be absolute.r)rrP)rrrrrgrarrT)r6rXrJrPr?rKrrcr%r&rTs

zProxyManager.urlopen)r+NN)Nr)N)N)T)	rZr[r\r]r1rErgrT
__classcell__r%r%rcr&rts

cKstfd|i|S)Nrb)r)rJr?r%r%r&rs)(Z
__future__rcollections	functoolsZlogging_collectionsrZconnectionpoolrrr
exceptionsrr	r
ZpackagesrZpackages.six.moves.urllib.parserZrequestr
Zutil.urlrZ
util.retryr__all__Z	getLoggerrZrVr>Z_key_fields
namedtuplerr'partialr5r4rrrr%r%r%r&s4


6


tasite-packages/pip/_vendor/urllib3/__pycache__/connection.cpython-38.pyc000064400000024462151464452710022104 0ustar00U

.eB;	@sddlmZddlZddlZddlZddlZddlmZmZ	ddl
Z
ddlmZddl
mZddl
mZzddlZejZWn,eefk
rdZGdd	d	eZYnXzeZWn$ek
rGd
ddeZYnXddlmZmZmZmZdd
lm Z m!Z!ddl"m#Z#m$Z$m%Z%m&Z&m'Z'ddl(m)Z)ddl*m+Z+e,e-Z.dddZ/e0dddZ1Gddde2Z3Gdddee2ZGdddeZ4Gddde4Z5ddZ6ere4Z7e5Z4ne3Z4dS))absolute_importN)errortimeout)six)HTTPConnection)
HTTPExceptionc@seZdZdS)BaseSSLErrorN__name__
__module____qualname__rrB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/connection.pyr	sr	c@seZdZdS)ConnectionErrorNr
rrrrrsr)NewConnectionErrorConnectTimeoutErrorSubjectAltNameWarningSystemTimeWarning)match_hostnameCertificateError)resolve_cert_reqsresolve_ssl_versionassert_fingerprintcreate_urllib3_contextssl_wrap_socket)
connection)HTTPHeaderDictPi)httphttpsic@seZdZdZdS)DummyConnectionz-Used to detect a failed ConnectionCls import.N)rrr
__doc__rrrrr!>sr!c@speZdZdZedZejejdfgZ	dZ
ddZeddZ
e
jd	dZ
d
dZdd
ZddZdddZdS)ra$
    Based on httplib.HTTPConnection but provides an extra constructor
    backwards-compatibility layer between older and newer Pythons.

    Additional keyword parameters are used to configure attributes of the connection.
    Accepted parameters include:

      - ``strict``: See the documentation on :class:`urllib3.connectionpool.HTTPConnectionPool`
      - ``source_address``: Set the source address for the current connection.
      - ``socket_options``: Set specific options on the underlying socket. If not specified, then
        defaults are loaded from ``HTTPConnection.default_socket_options`` which includes disabling
        Nagle's algorithm (sets TCP_NODELAY to 1) unless the connection is behind a proxy.

        For example, if you wish to enable TCP Keep Alive in addition to the defaults,
        you might pass::

            HTTPConnection.default_socket_options + [
                (socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1),
            ]

        Or you may want to disable the defaults by passing an empty list (e.g., ``[]``).
    rrFcOsDtjs|dd|d|_|d|j|_tj|f||dS)Nstrictsource_addresssocket_options)	rZPY2popgetr$default_socket_optionsr%_HTTPConnection__init__)selfargskwrrrr*es
zHTTPConnection.__init__cCs|jdS)a
        Getter method to remove any trailing dots that indicate the hostname is an FQDN.

        In general, SSL certificates don't include the trailing dot indicating a
        fully-qualified domain name, and thus, they don't validate properly when
        checked against a domain name that includes the dot. In addition, some
        servers may not expect to receive the trailing dot when provided.

        However, the hostname with trailing dot is critical to DNS resolution; doing a
        lookup with the trailing dot will properly only resolve the appropriate FQDN,
        whereas a lookup without a trailing dot will search the system's search domain
        list. Thus, it's important to keep the original host around for use only in
        those cases where it's appropriate (i.e., when doing DNS lookup to establish the
        actual TCP connection across which we're going to send HTTP requests).
        .)	_dns_hostrstrip)r+rrrhostrszHTTPConnection.hostcCs
||_dS)z
        Setter for the `host` property.

        We assume that only urllib3 uses the _dns_host attribute; httplib itself
        only uses `host`, and it seems reasonable that other libraries follow suit.
        N)r/)r+valuerrrr1sc
Csi}|jr|j|d<|jr$|j|d<ztj|j|jf|jf|}WnXtk
rlt|d|j	|jfYn0t
k
r}zt|d|W5d}~XYnX|S)zp Establish a socket connection and set nodelay settings on it.

        :return: New socket connection.
        r$r%z0Connection to %s timed out. (connect timeout=%s)z(Failed to establish a new connection: %sN)r$r%rZcreate_connectionr/portr
SocketTimeoutrr1SocketErrorr)r+Zextra_kwconnerrr	_new_conns2



zHTTPConnection._new_conncCs$||_t|ddr |d|_dS)N_tunnel_hostr)sockgetattr_tunnel	auto_openr+r6rrr
_prepare_connszHTTPConnection._prepare_conncCs|}||dS)N)r8r?r>rrrconnectszHTTPConnection.connectNcCst|dk	r|ni}d|k}d|k}|j||||d|D]\}}|||q>d|krh|dd||dk	rtjtf}	t||	r|f}|D]`}
|
sqt|
ts|
	d}
t
t|
d	d}||	d
|d||
|dq|ddS)
z
        Alternative to the common request method, which sends the
        body with chunked encoding and not as one block
        Nzaccept-encodingr1)skip_accept_encoding	skip_hostztransfer-encodingzTransfer-EncodingZchunkedutf8zutf-8s
s0

)
rZ
putrequestitemsZ	putheaderZ
endheadersrZstring_typesbytes
isinstanceencodehexlensend)r+methodZurlZbodyZheadersrArBheaderr2Zstringish_typeschunkZlen_strrrrrequest_chunkeds:




zHTTPConnection.request_chunked)NN)rrr
r"port_by_schemedefault_portsocketZIPPROTO_TCPZTCP_NODELAYr(is_verifiedr*propertyr1setterr8r?r@rOrrrrrDs


		rc@s<eZdZedZdZdddddejddfddZddZ	dS)HTTPSConnectionr Nc
KsDtj|||f||d|
||_||_||_||_|	|_d|_dS)N)r#rr )rr*key_file	cert_filekey_passwordssl_contextserver_hostnameZ	_protocol)r+r1r3rWrXrYr#rrZr[r-rrrr*szHTTPSConnection.__init__cCs|}||d}|jdkr>d}tt|jt|jd|_|j}|jsf|j	sf|rft
|drf|t||j
|j|j|j|jd|_dS)NFTssl_version	cert_reqsload_default_certs)r:keyfilecertfilerYrZr[)r8r?rZrrr]rr^ca_certsca_cert_dirhasattrr_rrWrXrYr[r:)r+r6default_ssl_contextcontextrrrr@s6

zHTTPSConnection.connect)
rrr
rPrQr]rRZ_GLOBAL_DEFAULT_TIMEOUTr*r@rrrrrVs
rVc@s6eZdZdZdZdZdZdZdZdddZ	ddZ
dS)VerifiedHTTPSConnectionz[
    Based on httplib.HTTPSConnection but wraps the socket with
    SSL certification.
    Nc		Csp|dkr$|jdk	r|jj}ntd}||_||_||_||_||_||_|oVt	j
||_|oht	j
||_
dS)zX
        This method should only be called once, before the connection is used.
        N)rZverify_moderrWrXr^rYassert_hostnamerospath
expanduserrbrc)	r+rWrXr^rYrbrirrcrrrset_cert-s

z VerifiedHTTPSConnection.set_certc
Cs|}|j}t|ddr4||_|d|_|j}|}|jdk	rH|j}tj	
tk}|rlt
dttd}|jdkrd}tt|jt|jd|_|j}t|j|_|js|js|rt|dr|t||j|j|j|j|j||d|_|jrt|jj dd	|jnd|jt!j"krtt|d
dst|j#dk	rt|j }|$ddsbt
d
|t%t&||j#pp||jt!j'kp|jdk	|_(dS)Nr9rzWSystem time is way off (before {0}). This will probably lead to SSL verification errorsFTr\r_)r:r`rarYrbrcr[rZ)Zbinary_formZcheck_hostnameZsubjectAltNamerzCertificate for {0} has no `subjectAltName`, falling back to check for a `commonName` for now. This feature is being removed by major browsers and deprecated by RFC 2818. (See https://github.com/shazow/urllib3/issues/497 for details.)))r8r1r;r:r<r=r9r[datetimedateZtodayRECENT_DATEwarningswarnformatrrZrrr]rr^rhrbrcrdr_rrWrXrYrZgetpeercertsslZ	CERT_NONErir'r_match_hostnameZ
CERT_REQUIREDrS)r+r6Zhostnamer[Zis_time_offrerfcertrrrr@Ls





	zVerifiedHTTPSConnection.connect)NNNNNNNN)rrr
r"r^rbrcr]rrmr@rrrrrg!s 
rgc
CsLzt||Wn8tk
rF}ztd||||_W5d}~XYnXdS)Nz@Certificate did not match expected hostname: %s. Certificate: %s)rrlogZwarningZ
_peer_cert)rvZasserted_hostnamer7rrrrusru)8Z
__future__rrnZloggingrjrRrr5rr4rqZpackagesrZpackages.six.moves.http_clientrr)rrtZSSLErrorr	ImportErrorAttributeError
BaseExceptionr	NameError	Exception
exceptionsrrrrZpackages.ssl_match_hostnamerrZ	util.ssl_rrrrrutilr_collectionsrZ	getLoggerrrwrProrpobjectr!rVrgruZUnverifiedHTTPSConnectionrrrrsL
	

Bsite-packages/pip/_vendor/urllib3/__pycache__/_collections.cpython-38.pyc000064400000024617151464452710022424 0ustar00U

.e(*@sddlmZzddlmZmZWn$ek
rDddlmZmZYnXzddlmZWn"ek
rxGdddZYnXddlm	Z	ddl
mZdd	lm
Z
mZmZd
dgZeZGdd
d
eZGd
ddeZdS))absolute_import)MappingMutableMapping)RLockc@seZdZddZddZdS)rcCsdSNselfrrD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/_collections.py	__enter__szRLock.__enter__cCsdSrr)r	exc_type	exc_value	tracebackrrr
__exit__szRLock.__exit__N)__name__
__module____qualname__rrrrrr
rsr)OrderedDict)
InvalidHeader)iterkeys
itervaluesPY3RecentlyUsedContainerHTTPHeaderDictc@sVeZdZdZeZdddZddZdd	Zd
dZ	dd
Z
ddZddZddZ
dS)ra
    Provides a thread-safe dict-like container which maintains up to
    ``maxsize`` keys while throwing away the least-recently-used keys beyond
    ``maxsize``.

    :param maxsize:
        Maximum number of recent elements to retain.

    :param dispose_func:
        Every time an item is evicted from the container,
        ``dispose_func(value)`` is called.  Callback which will get called
    
NcCs"||_||_||_t|_dSr)_maxsizedispose_funcContainerCls
_containerrlock)r	maxsizerrrr
__init__.s
zRecentlyUsedContainer.__init__c
Cs8|j(|j|}||j|<|W5QRSQRXdSr)r rpop)r	keyitemrrr
__getitem__5s
z!RecentlyUsedContainer.__getitem__c	Cslt}|j@|j|t}||j|<t|j|jkrF|jjdd\}}W5QRX|jrh|tk	rh||dS)NF)Zlast)_Nullr rgetlenrpopitemr)r	r$valueZ
evicted_valueZ_keyrrr
__setitem__<s
z!RecentlyUsedContainer.__setitem__c	Cs2|j|j|}W5QRX|jr.||dSr)r rr#r)r	r$r+rrr
__delitem__Ksz!RecentlyUsedContainer.__delitem__c
Cs(|jt|jW5QRSQRXdSr)r r)rrrrr
__len__RszRecentlyUsedContainer.__len__cCstddS)Nz7Iteration over this class is unlikely to be threadsafe.)NotImplementedErrorrrrr
__iter__VszRecentlyUsedContainer.__iter__c	CsH|jtt|j}|jW5QRX|jrD|D]}||q4dSr)r listrrclearr)r	valuesr+rrr
r2[szRecentlyUsedContainer.clearc
Cs,|jtt|jW5QRSQRXdSr)r r1rrrrrr
keyseszRecentlyUsedContainer.keys)rN)rrr__doc__rrr"r&r,r-r.r0r2r4rrrr
rs


cseZdZdZd-fdd	ZddZddZd	d
ZddZd
dZ	ddZ
esZej
Z
ejZeZddZddZefddZddZddZddZefddZeZeZeZeZdd Zd!d"Zd#d$Zd%d&Zd'd(Z d)d*Z!e"d+d,Z#Z$S).rap
    :param headers:
        An iterable of field-value pairs. Must not contain multiple field names
        when compared case-insensitively.

    :param kwargs:
        Additional field-value pairs to pass in to ``dict.update``.

    A ``dict`` like container for storing HTTP Headers.

    Field names are stored and compared case-insensitively in compliance with
    RFC 7230. Iteration provides the first case-sensitive key seen for each
    case-insensitive pair.

    Using ``__setitem__`` syntax overwrites fields that compare equal
    case-insensitively in order to maintain ``dict``'s api. For fields that
    compare equal, instead create a new ``HTTPHeaderDict`` and use ``.add``
    in a loop.

    If multiple fields that are equal case-insensitively are passed to the
    constructor or ``.update``, the behavior is undefined and some will be
    lost.

    >>> headers = HTTPHeaderDict()
    >>> headers.add('Set-Cookie', 'foo=bar')
    >>> headers.add('set-cookie', 'baz=quxx')
    >>> headers['content-length'] = '7'
    >>> headers['SET-cookie']
    'foo=bar, baz=quxx'
    >>> headers['Content-Length']
    '7'
    NcsPtt|t|_|dk	r>t|tr4||n
|||rL||dSr)superrr"rr
isinstance
_copy_fromextend)r	headerskwargs	__class__rr
r"s

zHTTPHeaderDict.__init__cCs ||g|j|<|j|Srrlowerr	r$valrrr
r,szHTTPHeaderDict.__setitem__cCs |j|}d|ddS)N, rrr?joinr@rrr
r&szHTTPHeaderDict.__getitem__cCs|j|=dSrr>r	r$rrr
r-szHTTPHeaderDict.__delitem__cCs||jkSr)r?rrErrr
__contains__szHTTPHeaderDict.__contains__cCs^t|tst|dsdSt|t|s2t||}tdd|Dtdd|DkS)Nr4Fcss|]\}}||fVqdSr)r?).0kvrrr
	sz(HTTPHeaderDict.__eq__..)r7rhasattrtypedict
itermergedr	otherrrr
__eq__szHTTPHeaderDict.__eq__cCs||Sr)rQrOrrr
__ne__szHTTPHeaderDict.__ne__cCs
t|jSr)r)rrrrr
r.szHTTPHeaderDict.__len__ccs|jD]}|dVq
dS)Nr)rr3)r	valsrrr
r0szHTTPHeaderDict.__iter__cCs@z||}Wn$tk
r0||jkr(|YSX||=|SdS)zD.pop(k[,d]) -> v, remove specified key and return the corresponding value.
          If key is not found, d is returned if given, otherwise KeyError is raised.
        N)KeyError_HTTPHeaderDict__marker)r	r$defaultr+rrr
r#s

zHTTPHeaderDict.popcCs$z
||=Wntk
rYnXdSr)rTrErrr
discards
zHTTPHeaderDict.discardcCs4|}||g}|j||}||k	r0||dS)zAdds a (name, value) pair, doesn't overwrite the value if it already
        exists.

        >>> headers = HTTPHeaderDict(foo='bar')
        >>> headers.add('Foo', 'baz')
        >>> headers['foo']
        'bar, baz'
        N)r?r
setdefaultappend)r	r$rAZ	key_lowerZnew_valsrSrrr
adds
	zHTTPHeaderDict.addcOst|dkrtdt|t|dkr2|dnd}t|tr`|D]\}}|||qHnjt|tr|D]}||||qnnDt|dr|	D]}||||qn|D]\}}|||q|
D]\}}|||qdS)zGeneric import function for any type of header-like object.
        Adapted version of MutableMapping.update in order to insert items
        with self.add instead of self.__setitem__
        rz9extend() takes at most 1 positional arguments ({0} given)rrr4N)r)	TypeErrorformatr7r	iteritemsrZrrKr4items)r	argsr;rPr$rAr+rrr
r9s(


zHTTPHeaderDict.extendcCsNz|j|}Wn*tk
r<||jkr4gYS|YSX|ddSdS)zmReturns a list of all the values for the named field. Returns an
        empty list if the key doesn't exist.rN)rr?rTrU)r	r$rVrSrrr
getlists

zHTTPHeaderDict.getlistcCsdt|jt|fS)Nz%s(%s))rLrrMrNrrrr
__repr__szHTTPHeaderDict.__repr__cCs>|D]4}||}t|tr$t|}|g||j|<qdSr)r`r7r1rr?)r	rPr$rArrr
r8s


zHTTPHeaderDict._copy_fromcCst|}|||Sr)rLr8)r	Zclonerrr
copy!s

zHTTPHeaderDict.copyccs<|D]2}|j|}|ddD]}|d|fVq"qdS)z8Iterate over all header lines, including duplicate ones.rNrr>)r	r$rSrArrr
r]&szHTTPHeaderDict.iteritemsccs8|D].}|j|}|dd|ddfVqdS)z:Iterate over all headers, merging duplicate ones together.rrBrNrCr@rrr
rN-szHTTPHeaderDict.itermergedcCst|Sr)r1r]rrrr
r^3szHTTPHeaderDict.itemscCsd}g}|jD]h}||rT|s.td|n&|d\}}||d|f|d<q|dd\}}|||fq||S)z4Read headers from a Python 2 httplib message object.) 	z/Header continuation with no previous header: %src:r)r:
startswithrstripsplitrY)clsmessageZobs_fold_continued_leadersr:liner$r+rrr
from_httplib6s

zHTTPHeaderDict.from_httplib)N)%rrrr5r"r,r&r-rFrQrRrrrrobjectrUr.r0r#rWrZr9r`Z
getheadersZgetallmatchingheadersZigetZget_allrar8rbr]rNr^classmethodrm
__classcell__rrr<r
rjs>!	
N)Z
__future__rZcollections.abcrrImportErrorcollectionsZ	threadingrr
exceptionsrZpackages.sixrrr__all__rnr'rrrrrr
sLsite-packages/pip/_vendor/urllib3/__pycache__/__init__.cpython-38.opt-1.pyc000064400000004045151464452710022436 0ustar00U

.e{
@s4dZddlmZddlZddlmZmZmZddlm	Z	ddl
mZddlm
Z
mZmZdd	lmZdd
lmZddlmZddlmZdd
lmZddlZddlmZdZdZdZdZe e!"eej#fddZ$[ej%de	j&ddej%de	j'ddej%de	j(ddej%de	j)dde	j*fddZ+dS)z8
urllib3 - Thread-safe connection pooling and re-using.
)absolute_importN)HTTPConnectionPoolHTTPSConnectionPoolconnection_from_url)
exceptions)encode_multipart_formdata)PoolManagerProxyManagerproxy_from_url)HTTPResponse)make_headers)get_host)Timeout)Retry)NullHandlerz(Andrey Petrov (andrey.petrov@shazow.net)ZMITz1.25.6)rrr	r
rrradd_stderr_loggerrdisable_warningsrrr
rcCsFtt}t}|td|||||dt|S)z
    Helper for quickly adding a StreamHandler to the logger. Useful for
    debugging.

    Returns the handler after adding it.
    z%%(asctime)s %(levelname)s %(message)sz,Added a stderr logging handler to logger: %s)	logging	getLogger__name__Z
StreamHandlerZsetFormatterZ	Formatter
addHandlerZsetLeveldebug)levelZloggerZhandlerr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/__init__.pyr/s	


ralwaysT)appenddefaultcCstd|dS)z<
    Helper for quickly disabling all urllib3 warnings.
    ignoreN)warningssimplefilter)categoryrrrrRsr),__doc__Z
__future__rr ZconnectionpoolrrrrZfilepostrZpoolmanagerr	r
rZresponserZutil.requestr
Zutil.urlrZutil.timeoutrZ
util.retryrrr
__author__Z__license____version____all__rrrDEBUGrr!ZSecurityWarningZSubjectAltNameWarningZInsecurePlatformWarningZSNIMissingWarningZHTTPWarningrrrrrs2site-packages/pip/_vendor/urllib3/__pycache__/connectionpool.cpython-38.opt-1.pyc000064400000056730151464452710023740 0ustar00U

.e^@sddlmZddlZddlZddlZddlZddlmZm	Z
ddlZddlmZm
Z
mZmZmZmZmZmZmZmZmZmZmZddlmZddlmZddlmZdd	lm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&dd
l'm(Z(ddl)m*Z*ddl+m,Z,dd
l-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6m7Z7m8Z8m9Z:m;Z;ddlj?Z?e@eAZBeCZDGdddeCZEejFejGhZHGdddeEe(ZIGdddeIZJddZKddZ9dS))absolute_importN)errortimeout)
ClosedPoolError
ProtocolErrorEmptyPoolErrorHeaderParsingErrorHostChangedErrorLocationValueError
MaxRetryError
ProxyErrorReadTimeoutErrorSSLErrorTimeoutErrorInsecureRequestWarningNewConnectionError)CertificateError)six)queue)port_by_schemeDummyConnectionHTTPConnectionHTTPSConnectionVerifiedHTTPSConnection
HTTPExceptionBaseSSLError)RequestMethods)HTTPResponse)is_connection_dropped)set_file_position)assert_header_parsing)Retry)Timeout)get_host	parse_urlUrl_normalize_host_encode_target)	LifoQueuec@sBeZdZdZdZeZd
ddZddZddZ	d	d
Z
ddZdS)ConnectionPoolzz
    Base class for all connection pools, such as
    :class:`.HTTPConnectionPool` and :class:`.HTTPSConnectionPool`.
    NcCs0|stdt||jd|_||_||_dS)NzNo host specified.scheme)rr'r,hostlower_proxy_hostport)selfr-r0r2F/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/connectionpool.py__init__Is

zConnectionPool.__init__cCsdt|j|j|jfS)Nz%s(host=%r, port=%r))type__name__r-r0r1r2r2r3__str__QszConnectionPool.__str__cCs|SNr2r7r2r2r3	__enter__TszConnectionPool.__enter__cCs|dS)NF)close)r1exc_typeZexc_valZexc_tbr2r2r3__exit__WszConnectionPool.__exit__cCsdS)D
        Close all pooled connections and disable the pool.
        Nr2r7r2r2r3r;\szConnectionPool.close)N)r6
__module____qualname____doc__r,r)QueueClsr4r8r:r=r;r2r2r2r3r*@s
r*c
@seZdZdZdZeZeZdde	j
ddddddf	ddZdd	Zd!d
dZ
dd
ZddZddZddZddZedfddZddZddZddZdddddeddddf
dd ZdS)"HTTPConnectionPoolaN	
    Thread-safe connection pool for one host.

    :param host:
        Host used for this HTTP Connection (e.g. "localhost"), passed into
        :class:`httplib.HTTPConnection`.

    :param port:
        Port used for this HTTP Connection (None is equivalent to 80), passed
        into :class:`httplib.HTTPConnection`.

    :param strict:
        Causes BadStatusLine to be raised if the status line can't be parsed
        as a valid HTTP/1.0 or 1.1 status line, passed into
        :class:`httplib.HTTPConnection`.

        .. note::
           Only works in Python 2. This parameter is ignored in Python 3.

    :param timeout:
        Socket timeout in seconds for each individual connection. This can
        be a float or integer, which sets the timeout for the HTTP request,
        or an instance of :class:`urllib3.util.Timeout` which gives you more
        fine-grained control over request timeouts. After the constructor has
        been parsed, this is always a `urllib3.util.Timeout` object.

    :param maxsize:
        Number of connections to save that can be reused. More than 1 is useful
        in multithreaded situations. If ``block`` is set to False, more
        connections will be created but they will not be saved once they've
        been used.

    :param block:
        If set to True, no more than ``maxsize`` connections will be used at
        a time. When no free connections are available, the call will block
        until a connection has been released. This is a useful side effect for
        particular multithreaded situations where one does not want to use more
        than maxsize connections per host to prevent flooding.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param retries:
        Retry configuration to use by default with requests in this pool.

    :param _proxy:
        Parsed proxy URL, should not be used directly, instead, see
        :class:`urllib3.connectionpool.ProxyManager`"

    :param _proxy_headers:
        A dictionary with proxy headers, should not be used directly,
        instead, see :class:`urllib3.connectionpool.ProxyManager`"

    :param \**conn_kw:
        Additional parameters are used to create fresh :class:`urllib3.connection.HTTPConnection`,
        :class:`urllib3.connection.HTTPSConnection` instances.
    httpNFrc
Kst|||t||||_t|ts4t|}|dkrBtj}||_	||_
|||_||_
|	|_|
pli|_t|D]}|jdqxd|_d|_||_|jr|jdgdS)NrZsocket_options)r*r4rstrict
isinstancer#
from_floatr"ZDEFAULTrretriesrBpoolblockproxy
proxy_headersxrangeputnum_connectionsnum_requestsconn_kw
setdefault)
r1r-r0rErmaxsizerJheadersrH_proxy_proxy_headersrQ_r2r2r3r4s(


zHTTPConnectionPool.__init__cCsR|jd7_td|j|j|jp$d|jf|j|j|jj|jd|j	}|S)z9
        Return a fresh :class:`HTTPConnection`.
        rz(Starting new HTTP connection (%d): %s:%sZ80)r-r0rrE)
rOlogdebugr-r0
ConnectionClsrconnect_timeoutrErQr1connr2r2r3	_new_conns zHTTPConnectionPool._new_conncCsd}z|jj|j|d}WnBtk
r8t|dYn&tjk
r\|jrXt|dYnX|rt|rt	
d|j|t
|dddkrd}|p|S)	a
        Get a connection. Will return a pooled connection if one is available.

        If no connections are available and :prop:`.block` is ``False``, then a
        fresh connection is returned.

        :param timeout:
            Seconds to wait before giving up and raising
            :class:`urllib3.exceptions.EmptyPoolError` if the pool is empty and
            :prop:`.block` is ``True``.
        N)rJrzPool is closed.z>Pool reached maximum size and no more connections are allowed.z Resetting dropped connection: %sZ	auto_openrr)rIgetrJAttributeErrorrrEmptyrrrXrYr-r;getattrr^)r1rr]r2r2r3	_get_conns$zHTTPConnectionPool._get_conncCs^z|jj|ddWdStk
r*Yn$tjk
rLtd|jYnX|rZ|dS)a
        Put a connection back into the pool.

        :param conn:
            Connection object for the current host and port as returned by
            :meth:`._new_conn` or :meth:`._get_conn`.

        If the pool is already full, the connection is closed and discarded
        because we exceeded maxsize. If connections are discarded frequently,
        then maxsize should be increased.

        If the pool is closed, then the connection will be closed and discarded.
        FrJNz2Connection pool is full, discarding connection: %s)	rIrNr`rZFullrXwarningr-r;r\r2r2r3	_put_connszHTTPConnectionPool._put_conncCsdS)U
        Called right before a request is made, after the socket is created.
        Nr2r\r2r2r3_validate_conn0sz!HTTPConnectionPool._validate_conncCsdSr9r2r\r2r2r3_prepare_proxy6sz!HTTPConnectionPool._prepare_proxycCs2|tkr|jSt|tr$|St|SdS)z< Helper that always returns a :class:`urllib3.util.Timeout` N)_DefaultrZclonerFr#rG)r1rr2r2r3_get_timeout:s


zHTTPConnectionPool._get_timeoutcCsjt|trt||d|t|dr>|jtkr>t||d|dt|ksVdt|krft||d|dS)zAIs the error actually a timeout? Will raise a ReadTimeout or pass!Read timed out. (read timeout=%s)errnoz	timed outzdid not complete (read)N)rF
SocketTimeoutrhasattrrm_blocking_errnosstr)r1errurl
timeout_valuer2r2r3_raise_timeoutFs(
z!HTTPConnectionPool._raise_timeoutc

Ks|jd7_||}||j|_z||Wn:ttfk
rp}z|j|||jdW5d}~XYnX|r|j	||f|n|j
||f||j}	t|ddr|	dkrt
||d|	|	tjkr|jtn|j|	zjz|jdd}
WnTtk
rPz|}
Wn0tk
rJ}zt|dW5d}~XYnXYnXWn<tttfk
r}z|j|||	dW5d}~XYnXt|d	d
}td|j|j|j||||
j|
j 	zt!|
j"Wn@t#tfk
r}ztj$d|%||dd
W5d}~XYnX|
S)a
        Perform a request on a given urllib connection object taken from our
        pool.

        :param conn:
            a connection from one of our connection pools

        :param timeout:
            Socket timeout in seconds for the request. This can be a
            float or integer, which will set the same timeout value for
            the socket connect and the socket read, or an instance of
            :class:`urllib3.util.Timeout`, which gives you more fine-grained
            control over your timeouts.
        r)rrrsrtNsockrrlT)	bufferingZ
_http_vsn_strzHTTP/?z%s://%s:%s "%s %s %s" %s %sz$Failed to parse headers (url=%s): %s)exc_info)&rPrkZ
start_connectr[rrhrnrruZrequest_chunkedrequestread_timeoutrbrr#DEFAULT_TIMEOUTrvZ
settimeoutsocketZgetdefaulttimeoutZgetresponse	TypeError
BaseExceptionrZ
raise_fromSocketErrorrXrYr,r-r0statusZlengthr!msgr	re
_absolute_url)
r1r]methodrsrchunkedZhttplib_request_kwtimeout_objerzhttplib_responseZhttp_versionZhper2r2r3
_make_request_sp

(z HTTPConnectionPool._make_requestcCst|j|j|j|djS)N)r,r-r0path)r&r,r-r0rs)r1rr2r2r3rsz HTTPConnectionPool._absolute_urlcCsV|jdkrdS|jd}|_z|jdd}|r|qWntjk
rPYnXdS)r>NFrd)rIr_r;rra)r1Zold_poolr]r2r2r3r;s
zHTTPConnectionPool.closecCsx|drdSt|\}}}|dk	r0t||d}|jrF|sFt|}n|js^|t|kr^d}|||f|j|j|jfkS)zj
        Check if the given ``url`` is a member of the same host as this
        connection pool.
        /TNr+)
startswithr$r'r0rr_r,r-)r1rsr,r-r0r2r2r3is_same_hosts

zHTTPConnectionPool.is_same_hostTc
Ks|dkr|j}t|ts*tj|||jd}|
dkr>|
dd}
|rX||sXt||||drrt	
t|}nt	
t|j
}d}|
}|jdkr|}||jd}d}t||}ztz||}|j|	d}|j|_|jdk	ot|d	d}|r|||j|||||||d
}|
s(|nd}||
d<|jj|f|||d|
}d}Wntj k
rxt!|d
Ynt"t#t$t%t&t't(fk
r,}zd}t|t&t(frt'|}n>t|t$t)fr|jrt*d|}nt|t$t#frt%d|}|j+||||t,-dd}|.|}W5d}~XYnXW5|sJ|oD|}d}|rZ||X|st/0d||||j1|||||||f||	|
|d|
Sdd}|o|2}|rR|j3dkrd}z|j+||||d}Wn,t4k
r|j5r|||YSX|||6|t/7d|||j1||||f|||||	|
|d|
St8|9d}|:||j3|rz|j+||||d}Wn,t4k
r|j;r|||YSX|||.|t/7d||j1||||f|||||	|
|d|
S|S)a
        Get a connection from the pool and perform an HTTP request. This is the
        lowest level call for making a request, so you'll need to specify all
        the raw details.

        .. note::

           More commonly, it's appropriate to use a convenience method provided
           by :class:`.RequestMethods`, such as :meth:`request`.

        .. note::

           `release_conn` will only behave as expected if
           `preload_content=False` because we want to make
           `preload_content=False` the default behaviour someday soon without
           breaking backwards compatibility.

        :param method:
            HTTP request method (such as GET, POST, PUT, etc.)

        :param body:
            Data to send in the request body (useful for creating
            POST requests, see HTTPConnectionPool.post_url for
            more convenience).

        :param headers:
            Dictionary of custom headers to send, such as User-Agent,
            If-None-Match, etc. If None, pool headers are used. If provided,
            these headers completely replace any pool-specific headers.

        :param retries:
            Configure the number of retries to allow before raising a
            :class:`~urllib3.exceptions.MaxRetryError` exception.

            Pass ``None`` to retry until you receive a response. Pass a
            :class:`~urllib3.util.retry.Retry` object for fine-grained control
            over different types of retries.
            Pass an integer number to retry connection errors that many times,
            but no other types of errors. Pass zero to never retry.

            If ``False``, then retries are disabled and any exception is raised
            immediately. Also, instead of raising a MaxRetryError on redirects,
            the redirect response will be returned.

        :type retries: :class:`~urllib3.util.retry.Retry`, False, or an int.

        :param redirect:
            If True, automatically handle redirects (status codes 301, 302,
            303, 307, 308). Each redirect counts as a retry. Disabling retries
            will disable redirect, too.

        :param assert_same_host:
            If ``True``, will make sure that the host of the pool requests is
            consistent else will raise HostChangedError. When False, you can
            use the pool on an HTTP proxy and request foreign hosts.

        :param timeout:
            If specified, overrides the default timeout for this one
            request. It may be a float (in seconds) or an instance of
            :class:`urllib3.util.Timeout`.

        :param pool_timeout:
            If set and the pool is set to block=True, then this method will
            block for ``pool_timeout`` seconds and raise EmptyPoolError if no
            connection is available within the time period.

        :param release_conn:
            If False, then the urlopen call will not release the connection
            back into the pool once a response is received (but will release if
            you read the entire contents of the response such as when
            `preload_content=True`). This is useful if you're not preloading
            the response's content immediately. You will need to call
            ``r.release_conn()`` on the response ``r`` to return the connection
            back into the pool. If None, it takes the value of
            ``response_kw.get('preload_content', True)``.

        :param chunked:
            If True, urllib3 will send the body using chunked transfer
            encoding. Otherwise, urllib3 will send the body using the standard
            content-length form. Defaults to False.

        :param int body_pos:
            Position to seek to in file-like body in the event of a retry or
            redirect. Typically this won't need to be set because urllib3 will
            auto-populate the value when needed.

        :param \**response_kw:
            Additional parameters are passed to
            :meth:`urllib3.response.HTTPResponse.from_httplib`
        N)redirectdefaultZpreload_contentTrrDF)rrv)rbodyrTrZrequest_method)rI
connectionrHz"No pool connections are available.zCannot connect to proxy.zConnection aborted.)r_poolZ_stacktracez1Retrying (%r) after connection broken by '%r': %s)rpool_timeoutrelease_connbody_posc
Ss2z|Wn ttttttfk
r,YnXdSr9)readrrrrrr)responser2r2r3drain_and_release_connsz:HTTPConnectionPool.urlopen..drain_and_release_conni/ZGET)rrzRedirecting %s -> %s)rHrassert_same_hostrrrrzRetry-Afterz	Retry: %s)>> conn = connection_from_url('http://google.com/')
        >>> r = conn.request('GET', '/')
    Prr0N)r$rr_rrC)rskwr,r-r0r2r2r3connection_from_urls
rcCs.t||}|dr*|dr*|dd}|S)z?
    Normalize hosts for comparisons and use with sockets.
    []r)normalize_hostrendswith)r-r,r2r2r3r's
r')LZ
__future__rrmZloggingrrr|rrrrn
exceptionsrrrr	r
rrr
rrrrrZpackages.ssl_match_hostnamerZpackagesrZpackages.six.movesrrrrrrrrrryrrrZutil.connectionrZutil.requestr Z
util.responser!Z
util.retryr"Zutil.timeoutr#Zutil.urlr$r%r&r'rr(Z
util.queuer)ZmovesrMZ	getLoggerr6rXobjectrjr*ZEAGAINZEWOULDBLOCKrprCrrr2r2r2r3sF<$	
$rsite-packages/pip/_vendor/urllib3/__pycache__/fields.cpython-38.opt-1.pyc000064400000017643151464452710022155 0ustar00U

.ei!@sddlmZddlZddlZddlZddlmZdddZdd	Z	d
dddZ
e
d
deddDddZ
ddZeZGdddeZdS))absolute_importN)sixapplication/octet-streamcCs|rt|dp|S|S)z
    Guess the "Content-Type" of a file.

    :param filename:
        The filename to guess the "Content-Type" of using :mod:`mimetypes`.
    :param default:
        If no "Content-Type" can be guessed, default to `default`.
    r)	mimetypesZ
guess_type)filenamedefaultr	>/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/fields.pyguess_content_type	s	rc	sttjrdtfdddDsdd|f}z|dWnttfk
r^YnX|Stjrtdt	j
dd|ftjrdS)a
    Helper function to format and quote a single header parameter using the
    strategy defined in RFC 2231.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows RFC 2388 Section 4.4.

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        An RFC-2231-formatted unicode string.
    utf-8c3s|]}|kVqdSNr	).0Zchvaluer	r
	)sz.format_header_param_rfc2231..z"\
%s="%s"asciiz%s*=%s)
isinstancerbinary_typedecodeanyencodeUnicodeEncodeErrorUnicodeDecodeErrorZPY2emailZutilsZencode_rfc2231)namerresultr	rr
format_header_param_rfc2231s 


rz%22z\\)"\r cCs&i|]}|dkrt|d|qS))z%{:02X})rZunichrformat)rZccr	r	r

Isr# cs:fdd}tdddD}|||}|S)Ncs|dS)Nr)group)matchneedles_and_replacementsr	r
replacerRsz#_replace_multiple..replacer|cSsg|]}t|qSr	)reescape)rZneedler	r	r

Vsz%_replace_multiple..)r+compilejoinkeyssub)rr(r)patternrr	r'r
_replace_multipleQsr3cCs,t|tjr|d}t|t}d||fS)a
    Helper function to format and quote a single header parameter using the
    HTML5 strategy.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows the `HTML5 Working Draft
    Section 4.10.22.7`_ and matches the behavior of curl and modern browsers.

    .. _HTML5 Working Draft Section 4.10.22.7:
        https://w3c.github.io/html/sec-forms.html#multipart-form-data

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        A unicode string, stripped of troublesome characters.
    rr)rrrrr3_HTML5_REPLACEMENTS)rrr	r	r
format_header_param_html5^s

r5c@sReZdZdZddefddZeefddZddZd	d
Z	ddZ
dd
dZdS)RequestFielda
    A data container for request body parameters.

    :param name:
        The name of this request field. Must be unicode.
    :param data:
        The data/value body.
    :param filename:
        An optional filename of the request field. Must be unicode.
    :param headers:
        An optional dict-like object of headers to initially use for the field.
    :param header_formatter:
        An optional callable that is used to encode and format the headers. By
        default, this is :func:`format_header_param_html5`.
    NcCs0||_||_||_i|_|r&t||_||_dSr
)_name	_filenamedataheadersdictheader_formatter)selfrr9rr:r<r	r	r
__init__s
zRequestField.__init__cCs`t|tr4t|dkr"|\}}}q@|\}}t|}nd}d}|}|||||d}|j|d|S)a
        A :class:`~urllib3.fields.RequestField` factory from old-style tuple parameters.

        Supports constructing :class:`~urllib3.fields.RequestField` from
        parameter of key/value strings AND key/filetuple. A filetuple is a
        (filename, data, MIME type) tuple where the MIME type is optional.
        For example::

            'foo': 'bar',
            'fakefile': ('foofile.txt', 'contents of foofile'),
            'realfile': ('barfile.txt', open('realfile').read()),
            'typedfile': ('bazfile.bin', open('bazfile').read(), 'image/jpeg'),
            'nonamefile': 'contents of nonamefile field',

        Field names and filenames must be unicode.
        N)rr<)content_type)rtuplelenrmake_multipart)clsZ	fieldnamerr<rr9r@Z
request_paramr	r	r
from_tupless 

zRequestField.from_tuplescCs|||S)aI
        Overridable helper function to format a single header parameter. By
        default, this calls ``self.header_formatter``.

        :param name:
            The name of the parameter, a string expected to be ASCII only.
        :param value:
            The value of the parameter, provided as a unicode string.
        )r<)r=rrr	r	r
_render_partszRequestField._render_partcCsLg}|}t|tr|}|D]"\}}|dk	r||||qd|S)aO
        Helper function to format and quote a single header.

        Useful for single headers that are composed of multiple items. E.g.,
        'Content-Disposition' fields.

        :param header_parts:
            A sequence of (k, v) tuples or a :class:`dict` of (k, v) to format
            as `k1="v1"; k2="v2"; ...`.
        N; )rr;itemsappendrFr/)r=Zheader_partspartsiterablerrr	r	r

_render_partss
zRequestField._render_partscCsg}dddg}|D]*}|j|dr|d||j|fq|jD]&\}}||krH|rH|d||fqH|dd|S)z=
        Renders the headers for this request field.
        Content-DispositionContent-TypeContent-LocationFz%s: %sz
)r:getrIrHr/)r=linesZ	sort_keysZsort_keyZheader_nameZheader_valuer	r	r
render_headerss

zRequestField.render_headerscCsX|pd|jd<|jddd|d|jfd|jffg7<||jd<||jd<d	S)
a|
        Makes this request field into a multipart request field.

        This method overrides "Content-Disposition", "Content-Type" and
        "Content-Location" headers to the request parameter.

        :param content_type:
            The 'Content-Type' of the request body.
        :param content_location:
            The 'Content-Location' of the request body.

        z	form-datarMrGrrrNrON)r:r/rLr7r8)r=Zcontent_dispositionr@Zcontent_locationr	r	r
rCs
zRequestField.make_multipart)NNN)__name__
__module____qualname____doc__r5r>classmethodrErFrLrRrCr	r	r	r
r6}s
#
r6)r)Z
__future__rZemail.utilsrrr+Zpackagesrrrr4updateranger3r5Zformat_header_paramobjectr6r	r	r	r
s&
*	
site-packages/pip/_vendor/urllib3/__pycache__/poolmanager.cpython-38.pyc000064400000031150151464452710022241 0ustar00U

.eB@sddlmZddlZddlZddlZddlmZddlmZm	Z	ddlm
Z
ddlmZm
Z
mZddlmZdd	lmZdd
lmZddlmZddlmZd
ddgZeeZdZdZedeZ ddZ!e"e!e e"e!e dZ#ee	dZ$Gdd
d
eZ%Gddde%Z&ddZ'dS))absolute_importN)RecentlyUsedContainer)HTTPConnectionPoolHTTPSConnectionPool)port_by_scheme)LocationValueError
MaxRetryErrorProxySchemeUnknown)six)urljoin)RequestMethods)	parse_url)RetryPoolManagerProxyManagerproxy_from_url)Zkey_fileZ	cert_fileZ	cert_reqsZca_certsZssl_versionZca_cert_dirZssl_contextZkey_password)Z
key_schemeZkey_hostZkey_portZkey_timeoutZkey_retriesZ
key_strictZ	key_blockZkey_source_addressZkey_key_fileZkey_key_passwordZ
key_cert_fileZ
key_cert_reqsZkey_ca_certsZkey_ssl_versionZkey_ca_cert_dirZkey_ssl_contextZkey_maxsizeZkey_headersZ
key__proxyZkey__proxy_headersZkey_socket_optionsZkey__socks_optionsZkey_assert_hostnameZkey_assert_fingerprintZkey_server_hostnamePoolKeycCs|}|d|d<|d|d<dD],}||kr,||dk	r,t||||<q,|d}|dk	rxt||d<t|D]}|||d|<q|j	D]}||krd||<q|f|S)a
    Create a pool key out of a request context dictionary.

    According to RFC 3986, both the scheme and host are case-insensitive.
    Therefore, this function normalizes both before constructing the pool
    key for an HTTPS request. If you wish to change this behaviour, provide
    alternate callables to ``key_fn_by_scheme``.

    :param key_class:
        The class to use when constructing the key. This should be a namedtuple
        with the ``scheme`` and ``host`` keys at a minimum.
    :type  key_class: namedtuple
    :param request_context:
        A dictionary-like object that contain the context for a request.
    :type  request_context: dict

    :return: A namedtuple that can be used as a connection pool key.
    :rtype:  PoolKey
    schemehost)headers_proxy_headersZ_socks_optionsNZsocket_optionsZkey_)
copylower	frozensetitemsgettuplelistkeyspop_fields)Z	key_classrequest_contextcontextkeyZsocket_optsZfieldr%C/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/poolmanager.py_default_key_normalizerDs


r'httphttpsc@sxeZdZdZdZdddZddZdd	Zdd
dZdd
Z	dddZ
ddZdddZd ddZ
ddZd!ddZdS)"ra$
    Allows for arbitrary requests while transparently keeping track of
    necessary connection pools for you.

    :param num_pools:
        Number of connection pools to cache before discarding the least
        recently used pool.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param \**connection_pool_kw:
        Additional parameters are used to create fresh
        :class:`urllib3.connectionpool.ConnectionPool` instances.

    Example::

        >>> manager = PoolManager(num_pools=2)
        >>> r = manager.request('GET', 'http://google.com/')
        >>> r = manager.request('GET', 'http://google.com/mail')
        >>> r = manager.request('GET', 'http://yahoo.com/')
        >>> len(manager.pools)
        2

    N
cKs8t||||_t|ddd|_t|_t|_dS)NcSs|SN)close)pr%r%r&z&PoolManager.__init__..)Zdispose_func)r
__init__connection_pool_kwrpoolspool_classes_by_schemekey_fn_by_schemer)self	num_poolsrr2r%r%r&r1s
zPoolManager.__init__cCs|Sr,r%r6r%r%r&	__enter__szPoolManager.__enter__cCs|dS)NF)clear)r6exc_typeZexc_valZexc_tbr%r%r&__exit__szPoolManager.__exit__cCs^|j|}|dkr|j}dD]}||dq |dkrPtD]}||dq>|||f|S)a
        Create a new :class:`ConnectionPool` based on host, port, scheme, and
        any additional pool keyword arguments.

        If ``request_context`` is provided, it is provided as keyword arguments
        to the pool class used. This method is used to actually create the
        connection pools handed out by :meth:`connection_from_url` and
        companion methods. It is intended to be overridden for customization.
        N)rrportr))r4r2rr SSL_KEYWORDS)r6rrr=r"Zpool_clsr$kwr%r%r&	_new_pools


zPoolManager._new_poolcCs|jdS)z
        Empty our store of pools and direct them all to close.

        This will not affect in-flight connections, but they will not be
        re-used after completion.
        N)r3r:r8r%r%r&r:szPoolManager.clearr)cCsT|std||}|pd|d<|s:t|dd}||d<||d<||S)a
        Get a :class:`ConnectionPool` based on the host, port, and scheme.

        If ``port`` isn't given, it will be derived from the ``scheme`` using
        ``urllib3.connectionpool.port_by_scheme``. If ``pool_kwargs`` is
        provided, it is merged with the instance's ``connection_pool_kw``
        variable and used to create the new connection pool, if one is
        needed.
        zNo host specified.r)rPr=r)r_merge_pool_kwargsrrrconnection_from_context)r6rr=rpool_kwargsr"r%r%r&connection_from_hosts
z PoolManager.connection_from_hostcCs,|d}|j|}||}|j||dS)z
        Get a :class:`ConnectionPool` based on the request context.

        ``request_context`` must at least contain the ``scheme`` key and its
        value must be a key in ``key_fn_by_scheme`` instance variable.
        rr")rr5connection_from_pool_key)r6r"rZpool_key_constructorpool_keyr%r%r&rCs
z#PoolManager.connection_from_contextc
Csl|jjZ|j|}|r*|W5QRS|d}|d}|d}|j||||d}||j|<W5QRX|S)z
        Get a :class:`ConnectionPool` based on the provided pool key.

        ``pool_key`` should be a namedtuple that only contains immutable
        objects. At a minimum it must have the ``scheme``, ``host``, and
        ``port`` fields.
        rrr=rF)r3lockrr@)r6rHr"Zpoolrrr=r%r%r&rGs
z$PoolManager.connection_from_pool_keycCs t|}|j|j|j|j|dS)a
        Similar to :func:`urllib3.connectionpool.connection_from_url`.

        If ``pool_kwargs`` is not provided and a new pool needs to be
        constructed, ``self.connection_pool_kw`` is used to initialize
        the :class:`urllib3.connectionpool.ConnectionPool`. If ``pool_kwargs``
        is provided, it is used instead. Note that if a new pool does not
        need to be created for the request, the provided ``pool_kwargs`` are
        not used.
        )r=rrD)rrErr=r)r6urlrDur%r%r&connection_from_urlszPoolManager.connection_from_urlc	CsV|j}|rR|D]:\}}|dkrHz
||=WqPtk
rDYqPXq|||<q|S)a
        Merge a dictionary of override values for self.connection_pool_kw.

        This does not modify self.connection_pool_kw and returns a new dict.
        Any keys in the override dictionary with a value of ``None`` are
        removed from the merged dictionary.
        N)r2rrKeyError)r6overrideZbase_pool_kwargsr$valuer%r%r&rB!s


zPoolManager._merge_pool_kwargsTcKst|}|j|j|j|jd}d|d<d|d<d|krD|j|d<|jdk	rj|jdkrj|j||f|}n|j||j	f|}|o|
}|s|St||}|jdkrd	}|
d
}	t|	tstj|	|d}	|	jr||stt|d}
|
D]"}||	jkr|d|dqz|	j||||d}	Wn$tk
rR|	jrJ|YSX|	|d
<||d<td
|||j||f|S)a]
        Same as :meth:`urllib3.connectionpool.HTTPConnectionPool.urlopen`
        with custom cross-host redirect logic and only sends the request-uri
        portion of the ``url``.

        The given ``url`` parameter must be absolute, such that an appropriate
        :class:`urllib3.connectionpool.ConnectionPool` can be chosen for it.
        )r=rFZassert_same_hostredirectrNr)i/ZGETretries)rP)responseZ_poolzRedirecting %s -> %s)rrErr=rrrproxyurlopenZrequest_uriZget_redirect_locationrZstatusr
isinstancerZfrom_intZremove_headers_on_redirectZis_same_hostrrZiterkeysrr Z	incrementr	Zraise_on_redirectloginfo)r6methodrJrPr?rKZconnrRZredirect_locationrQrheaderr%r%r&rT5sF	




zPoolManager.urlopen)r+N)N)Nr)N)N)N)T)__name__
__module____qualname____doc__rSr1r9r<r@r:rErCrGrLrBrTr%r%r%r&rs


	


csHeZdZdZdfdd	Zdfdd	Zdd	d
Zdfdd
	ZZS)rax
    Behaves just like :class:`PoolManager`, but sends all requests through
    the defined proxy, using the CONNECT method for HTTPS URLs.

    :param proxy_url:
        The URL of the proxy to be used.

    :param proxy_headers:
        A dictionary containing headers that will be sent to the proxy. In case
        of HTTP they are being sent with each request, while in the
        HTTPS/CONNECT case they are sent only once. Could be used for proxy
        authentication.

    Example:
        >>> proxy = urllib3.ProxyManager('http://localhost:3128/')
        >>> r1 = proxy.request('GET', 'http://google.com/')
        >>> r2 = proxy.request('GET', 'http://httpbin.org/')
        >>> len(proxy.pools)
        1
        >>> r3 = proxy.request('GET', 'https://httpbin.org/')
        >>> r4 = proxy.request('GET', 'https://twitter.com/')
        >>> len(proxy.pools)
        3

    r+Ncst|trd|j|j|jf}t|}|jsFt|jd}|j|d}|jdkrZt	|j||_
|pfi|_|j
|d<|j|d<tt
|j||f|dS)Nz
%s://%s:%irA)r=r(_proxyr)rUrrrr=rrr_replacer
rS
proxy_headerssuperrr1)r6	proxy_urlr7rr`r2rSr=	__class__r%r&r1s"	





zProxyManager.__init__r)csD|dkr tt|j||||dStt|j|jj|jj|jj|dS)Nr*)rD)rarrErSrr=r)r6rr=rrDrcr%r&rEs

z!ProxyManager.connection_from_hostcCs0ddi}t|j}|r||d<|r,|||S)z
        Sets headers needed by proxies: specifically, the Accept and Host
        headers. Only sets headers not provided by the user.
        ZAcceptz*/*ZHost)rnetlocupdate)r6rJrZheaders_rer%r%r&_set_proxy_headerss

zProxyManager._set_proxy_headersTcsNt|}|jdkr0|d|j}||||d<tt|j||fd|i|S)z@Same as HTTP(S)ConnectionPool.urlopen, ``url`` must be absolute.r)rrP)rrrrrgrarrT)r6rXrJrPr?rKrrcr%r&rTs

zProxyManager.urlopen)r+NN)Nr)N)N)T)	rZr[r\r]r1rErgrT
__classcell__r%r%rcr&rts

cKstfd|i|S)Nrb)r)rJr?r%r%r&rs)(Z
__future__rcollections	functoolsZlogging_collectionsrZconnectionpoolrrr
exceptionsrr	r
ZpackagesrZpackages.six.moves.urllib.parserZrequestr
Zutil.urlrZ
util.retryr__all__Z	getLoggerrZrVr>Z_key_fields
namedtuplerr'partialr5r4rrrr%r%r%r&s4


6


tasite-packages/pip/_vendor/urllib3/__pycache__/request.cpython-38.opt-1.pyc000064400000012766151464452710022400 0ustar00U

.e@s>ddlmZddlmZddlmZdgZGdddeZdS))absolute_import)encode_multipart_formdata)	urlencodeRequestMethodsc@sNeZdZdZddddhZdddZdd
dZddd
ZdddZdddZ	dS)ra
    Convenience mixin for classes who implement a :meth:`urlopen` method, such
    as :class:`~urllib3.connectionpool.HTTPConnectionPool` and
    :class:`~urllib3.poolmanager.PoolManager`.

    Provides behavior for making common types of HTTP request methods and
    decides which type of request field encoding to use.

    Specifically,

    :meth:`.request_encode_url` is for sending requests whose fields are
    encoded in the URL (such as GET, HEAD, DELETE).

    :meth:`.request_encode_body` is for sending requests whose fields are
    encoded in the *body* of the request using multipart or www-form-urlencoded
    (such as for POST, PUT, PATCH).

    :meth:`.request` is for making any kind of request, it will look up the
    appropriate encoding format and use one of the above two methods to make
    the request.

    Initializer parameters:

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.
    ZDELETEZGETZHEADZOPTIONSNcCs|pi|_dS)N)headers)selfrr	?/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/request.py__init__)szRequestMethods.__init__TcKstddS)NzMClasses extending RequestMethods must implement their own ``urlopen`` method.)NotImplementedError)rmethodurlbodyrencode_multipartmultipart_boundarykwr	r	r
urlopen,s
zRequestMethods.urlopencKsR|}||d<||jkr4|j||f||d|S|j||f||d|SdS)a
        Make a request using :meth:`urlopen` with the appropriate encoding of
        ``fields`` based on the ``method`` used.

        This is a convenience method that requires the least amount of manual
        effort. It can be used in most situations, while still having the
        option to drop down to more specific methods when necessary, such as
        :meth:`request_encode_url`, :meth:`request_encode_body`,
        or even the lowest level :meth:`urlopen`.
        Zrequest_url)fieldsrN)upper_encode_url_methodsrequest_encode_urlrequest_encode_body)rr
rrr
urlopen_kwr	r	r
request;s*
zRequestMethods.requestcKsD|dkr|j}d|i}|||r4|dt|7}|j||f|S)z
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the url. This is useful for request methods like GET, HEAD, DELETE, etc.
        Nr?)rupdaterr)rr
rrrrextra_kwr	r	r
rSs
z!RequestMethods.request_encode_urlcKs|dkr|j}dii}|rbd|kr*td|r@t||d\}	}
nt|d}	}
|	|d<d|
i|d<|d||||j||f|S)a
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the body. This is useful for request methods like POST, PUT, PATCH, etc.

        When ``encode_multipart=True`` (default), then
        :meth:`urllib3.filepost.encode_multipart_formdata` is used to encode
        the payload with the appropriate content type. Otherwise
        :meth:`urllib.urlencode` is used with the
        'application/x-www-form-urlencoded' content type.

        Multipart encoding must be used when posting files, and it's reasonably
        safe to use it in other times too. However, it may break request
        signing, such as with OAuth.

        Supports an optional ``fields`` parameter of key/value strings AND
        key/filetuple. A filetuple is a (filename, data, MIME type) tuple where
        the MIME type is optional. For example::

            fields = {
                'foo': 'bar',
                'fakefile': ('foofile.txt', 'contents of foofile'),
                'realfile': ('barfile.txt', open('realfile').read()),
                'typedfile': ('bazfile.bin', open('bazfile').read(),
                              'image/jpeg'),
                'nonamefile': 'contents of nonamefile field',
            }

        When uploading a file, providing a filename (the first parameter of the
        tuple) is optional but recommended to best mimic behavior of browsers.

        Note that if ``headers`` are supplied, the 'Content-Type' header will
        be overwritten because it depends on the dynamic random boundary string
        which is used to compose the body of the request. The random boundary
        string can be explicitly set with the ``multipart_boundary`` parameter.
        NrrzFrequest got values for both 'fields' and 'body', can only specify one.)boundaryz!application/x-www-form-urlencodedzContent-Type)r	TypeErrorrrrr)rr
rrrrrrrrZcontent_typer	r	r
rcs*-
z"RequestMethods.request_encode_body)N)NNTN)NN)NN)NNTN)
__name__
__module____qualname____doc__rrrrrrr	r	r	r
r
s



N)	Z
__future__rZfilepostrZpackages.six.moves.urllib.parser__all__objectrr	r	r	r
ssite-packages/pip/_vendor/urllib3/__pycache__/exceptions.cpython-38.opt-1.pyc000064400000023454151464452710023065 0ustar00U

.e@sLddlmZddlmZGdddeZGdddeZGdd	d	eZ	Gd
dde	Z
Gdd
d
eZGdddeZGdddeZ
GdddeZeZGddde
ZGddde
ZGdddeZGdddeZGdddee
ZGdddeZGd d!d!ee	ZGd"d#d#e	ZGd$d%d%e	ZGd&d'd'eeZGd(d)d)eZGd*d+d+eZGd,d-d-eZGd.d/d/eZGd0d1d1eZGd2d3d3eZ Gd4d5d5eZ!Gd6d7d7eZ"Gd8d9d9eZ#Gd:d;d;eeZ$Gdd?d?eeZGd@dAdAeZ&GdBdCdCe'eZ(GdDdEdEeZ)GdFdGdGeZ*dHS)I)absolute_import)IncompleteReadc@seZdZdZdS)	HTTPErrorz#Base exception used by this module.N__name__
__module____qualname____doc__rrB/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/exceptions.pyrsrc@seZdZdZdS)HTTPWarningz!Base warning used by this module.Nrrrrrr
sr
c@s eZdZdZddZddZdS)	PoolErrorz/Base exception for errors caused within a pool.cCs||_t|d||fdS)Nz%s: %s)poolr__init__)selfrmessagerrrrszPoolError.__init__cCs
|jdfS)N)NN	__class__rrrr
__reduce__szPoolError.__reduce__Nrrr	r
rrrrrrrsrc@s eZdZdZddZddZdS)RequestErrorz8Base exception for PoolErrors that have associated URLs.cCs||_t|||dSN)urlrr)rrrrrrrr szRequestError.__init__cCs|jd|jdffSr)rrrrrrr$szRequestError.__reduce__Nrrrrrrsrc@seZdZdZdS)SSLErrorz9Raised when SSL certificate fails in an HTTPS connection.Nrrrrrr)src@seZdZdZdS)
ProxyErrorz,Raised when the connection to a proxy fails.Nrrrrrr.src@seZdZdZdS)DecodeErrorz;Raised when automatic decoding based on Content-Type fails.Nrrrrrr3src@seZdZdZdS)
ProtocolErrorz>Raised when something unexpected happens mid-request/response.Nrrrrrr8src@seZdZdZdddZdS)
MaxRetryErroraRaised when the maximum number of retries is exceeded.

    :param pool: The connection pool
    :type pool: :class:`~urllib3.connectionpool.HTTPConnectionPool`
    :param string url: The requested Url
    :param exceptions.Exception reason: The underlying error

    NcCs&||_d||f}t||||dS)Nz0Max retries exceeded with url: %s (Caused by %r))reasonrr)rrrr rrrrrNszMaxRetryError.__init__)Nrrr	r
rrrrrrDs	rc@seZdZdZdddZdS)HostChangedErrorz?Raised when an existing pool gets a request for a foreign host.cCs"d|}t||||||_dS)Nz)Tried to open a foreign host with url: %s)rrretries)rrrr$rrrrrYszHostChangedError.__init__N)r#r!rrrrr"Vsr"c@seZdZdZdS)TimeoutStateErrorz3 Raised when passing an invalid state to a timeout Nrrrrrr%_sr%c@seZdZdZdS)TimeoutErrorz Raised when a socket timeout error occurs.

    Catching this error will catch both :exc:`ReadTimeoutErrors
    ` and :exc:`ConnectTimeoutErrors `.
    Nrrrrrr&esr&c@seZdZdZdS)ReadTimeoutErrorzFRaised when a socket timeout occurs while receiving data from a serverNrrrrrr'osr'c@seZdZdZdS)ConnectTimeoutErrorz@Raised when a socket timeout occurs while connecting to a serverNrrrrrr(vsr(c@seZdZdZdS)NewConnectionErrorzHRaised when we fail to establish a new connection. Usually ECONNREFUSED.Nrrrrrr){sr)c@seZdZdZdS)EmptyPoolErrorzCRaised when a pool runs out of connections and no more are allowed.Nrrrrrr*sr*c@seZdZdZdS)ClosedPoolErrorzCRaised when a request enters a pool after the pool has been closed.Nrrrrrr+sr+c@seZdZdZdS)LocationValueErrorzsr>cs eZdZdZfddZZS)ProxySchemeUnknownz1ProxyManager does not support the supplied schemecsd|}tt||dS)NzNot supported proxy scheme %s)r9r?r)rZschemerrrrrszProxySchemeUnknown.__init__rrr	r
rr=rrrrr?sr?cs eZdZdZfddZZS)HeaderParsingErrorzNRaised by assert_header_parsing, but we convert it to a log.warning statement.cs$d|pd|f}tt||dS)Nz%s, unparsed data: %rZUnknown)r9rAr)rZdefectsZ
unparsed_datarrrrrszHeaderParsingError.__init__r@rrrrrAsrAc@seZdZdZdS)UnrewindableBodyErrorz9urllib3 encountered an error when trying to rewind a bodyNrrrrrrBsrBN)+Z
__future__rZpackages.six.moves.http_clientrZhttplib_IncompleteRead	ExceptionrWarningr
rrrrrrConnectionErrorrr"r%r&r'r(r)r*r+
ValueErrorr,r-r/r0r1r2r3r4r5r6r7r8r>AssertionErrorr?rArBrrrrsH	

			site-packages/pip/_vendor/urllib3/__pycache__/_collections.cpython-38.opt-1.pyc000064400000024617151464452710023363 0ustar00U

.e(*@sddlmZzddlmZmZWn$ek
rDddlmZmZYnXzddlmZWn"ek
rxGdddZYnXddlm	Z	ddl
mZdd	lm
Z
mZmZd
dgZeZGdd
d
eZGd
ddeZdS))absolute_import)MappingMutableMapping)RLockc@seZdZddZddZdS)rcCsdSNselfrrD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/_collections.py	__enter__szRLock.__enter__cCsdSrr)r	exc_type	exc_value	tracebackrrr
__exit__szRLock.__exit__N)__name__
__module____qualname__rrrrrr
rsr)OrderedDict)
InvalidHeader)iterkeys
itervaluesPY3RecentlyUsedContainerHTTPHeaderDictc@sVeZdZdZeZdddZddZdd	Zd
dZ	dd
Z
ddZddZddZ
dS)ra
    Provides a thread-safe dict-like container which maintains up to
    ``maxsize`` keys while throwing away the least-recently-used keys beyond
    ``maxsize``.

    :param maxsize:
        Maximum number of recent elements to retain.

    :param dispose_func:
        Every time an item is evicted from the container,
        ``dispose_func(value)`` is called.  Callback which will get called
    
NcCs"||_||_||_t|_dSr)_maxsizedispose_funcContainerCls
_containerrlock)r	maxsizerrrr
__init__.s
zRecentlyUsedContainer.__init__c
Cs8|j(|j|}||j|<|W5QRSQRXdSr)r rpop)r	keyitemrrr
__getitem__5s
z!RecentlyUsedContainer.__getitem__c	Cslt}|j@|j|t}||j|<t|j|jkrF|jjdd\}}W5QRX|jrh|tk	rh||dS)NF)Zlast)_Nullr rgetlenrpopitemr)r	r$valueZ
evicted_valueZ_keyrrr
__setitem__<s
z!RecentlyUsedContainer.__setitem__c	Cs2|j|j|}W5QRX|jr.||dSr)r rr#r)r	r$r+rrr
__delitem__Ksz!RecentlyUsedContainer.__delitem__c
Cs(|jt|jW5QRSQRXdSr)r r)rrrrr
__len__RszRecentlyUsedContainer.__len__cCstddS)Nz7Iteration over this class is unlikely to be threadsafe.)NotImplementedErrorrrrr
__iter__VszRecentlyUsedContainer.__iter__c	CsH|jtt|j}|jW5QRX|jrD|D]}||q4dSr)r listrrclearr)r	valuesr+rrr
r2[szRecentlyUsedContainer.clearc
Cs,|jtt|jW5QRSQRXdSr)r r1rrrrrr
keyseszRecentlyUsedContainer.keys)rN)rrr__doc__rrr"r&r,r-r.r0r2r4rrrr
rs


cseZdZdZd-fdd	ZddZddZd	d
ZddZd
dZ	ddZ
esZej
Z
ejZeZddZddZefddZddZddZddZefddZeZeZeZeZdd Zd!d"Zd#d$Zd%d&Zd'd(Z d)d*Z!e"d+d,Z#Z$S).rap
    :param headers:
        An iterable of field-value pairs. Must not contain multiple field names
        when compared case-insensitively.

    :param kwargs:
        Additional field-value pairs to pass in to ``dict.update``.

    A ``dict`` like container for storing HTTP Headers.

    Field names are stored and compared case-insensitively in compliance with
    RFC 7230. Iteration provides the first case-sensitive key seen for each
    case-insensitive pair.

    Using ``__setitem__`` syntax overwrites fields that compare equal
    case-insensitively in order to maintain ``dict``'s api. For fields that
    compare equal, instead create a new ``HTTPHeaderDict`` and use ``.add``
    in a loop.

    If multiple fields that are equal case-insensitively are passed to the
    constructor or ``.update``, the behavior is undefined and some will be
    lost.

    >>> headers = HTTPHeaderDict()
    >>> headers.add('Set-Cookie', 'foo=bar')
    >>> headers.add('set-cookie', 'baz=quxx')
    >>> headers['content-length'] = '7'
    >>> headers['SET-cookie']
    'foo=bar, baz=quxx'
    >>> headers['Content-Length']
    '7'
    NcsPtt|t|_|dk	r>t|tr4||n
|||rL||dSr)superrr"rr
isinstance
_copy_fromextend)r	headerskwargs	__class__rr
r"s

zHTTPHeaderDict.__init__cCs ||g|j|<|j|Srrlowerr	r$valrrr
r,szHTTPHeaderDict.__setitem__cCs |j|}d|ddS)N, rrr?joinr@rrr
r&szHTTPHeaderDict.__getitem__cCs|j|=dSrr>r	r$rrr
r-szHTTPHeaderDict.__delitem__cCs||jkSr)r?rrErrr
__contains__szHTTPHeaderDict.__contains__cCs^t|tst|dsdSt|t|s2t||}tdd|Dtdd|DkS)Nr4Fcss|]\}}||fVqdSr)r?).0kvrrr
	sz(HTTPHeaderDict.__eq__..)r7rhasattrtypedict
itermergedr	otherrrr
__eq__szHTTPHeaderDict.__eq__cCs||Sr)rQrOrrr
__ne__szHTTPHeaderDict.__ne__cCs
t|jSr)r)rrrrr
r.szHTTPHeaderDict.__len__ccs|jD]}|dVq
dS)Nr)rr3)r	valsrrr
r0szHTTPHeaderDict.__iter__cCs@z||}Wn$tk
r0||jkr(|YSX||=|SdS)zD.pop(k[,d]) -> v, remove specified key and return the corresponding value.
          If key is not found, d is returned if given, otherwise KeyError is raised.
        N)KeyError_HTTPHeaderDict__marker)r	r$defaultr+rrr
r#s

zHTTPHeaderDict.popcCs$z
||=Wntk
rYnXdSr)rTrErrr
discards
zHTTPHeaderDict.discardcCs4|}||g}|j||}||k	r0||dS)zAdds a (name, value) pair, doesn't overwrite the value if it already
        exists.

        >>> headers = HTTPHeaderDict(foo='bar')
        >>> headers.add('Foo', 'baz')
        >>> headers['foo']
        'bar, baz'
        N)r?r
setdefaultappend)r	r$rAZ	key_lowerZnew_valsrSrrr
adds
	zHTTPHeaderDict.addcOst|dkrtdt|t|dkr2|dnd}t|tr`|D]\}}|||qHnjt|tr|D]}||||qnnDt|dr|	D]}||||qn|D]\}}|||q|
D]\}}|||qdS)zGeneric import function for any type of header-like object.
        Adapted version of MutableMapping.update in order to insert items
        with self.add instead of self.__setitem__
        rz9extend() takes at most 1 positional arguments ({0} given)rrr4N)r)	TypeErrorformatr7r	iteritemsrZrrKr4items)r	argsr;rPr$rAr+rrr
r9s(


zHTTPHeaderDict.extendcCsNz|j|}Wn*tk
r<||jkr4gYS|YSX|ddSdS)zmReturns a list of all the values for the named field. Returns an
        empty list if the key doesn't exist.rN)rr?rTrU)r	r$rVrSrrr
getlists

zHTTPHeaderDict.getlistcCsdt|jt|fS)Nz%s(%s))rLrrMrNrrrr
__repr__szHTTPHeaderDict.__repr__cCs>|D]4}||}t|tr$t|}|g||j|<qdSr)r`r7r1rr?)r	rPr$rArrr
r8s


zHTTPHeaderDict._copy_fromcCst|}|||Sr)rLr8)r	Zclonerrr
copy!s

zHTTPHeaderDict.copyccs<|D]2}|j|}|ddD]}|d|fVq"qdS)z8Iterate over all header lines, including duplicate ones.rNrr>)r	r$rSrArrr
r]&szHTTPHeaderDict.iteritemsccs8|D].}|j|}|dd|ddfVqdS)z:Iterate over all headers, merging duplicate ones together.rrBrNrCr@rrr
rN-szHTTPHeaderDict.itermergedcCst|Sr)r1r]rrrr
r^3szHTTPHeaderDict.itemscCsd}g}|jD]h}||rT|s.td|n&|d\}}||d|f|d<q|dd\}}|||fq||S)z4Read headers from a Python 2 httplib message object.) 	z/Header continuation with no previous header: %src:r)r:
startswithrstripsplitrY)clsmessageZobs_fold_continued_leadersr:liner$r+rrr
from_httplib6s

zHTTPHeaderDict.from_httplib)N)%rrrr5r"r,r&r-rFrQrRrrrrobjectrUr.r0r#rWrZr9r`Z
getheadersZgetallmatchingheadersZigetZget_allrar8rbr]rNr^classmethodrm
__classcell__rrr<r
rjs>!	
N)Z
__future__rZcollections.abcrrImportErrorcollectionsZ	threadingrr
exceptionsrZpackages.sixrrr__all__rnr'rrrrrr
sLsite-packages/pip/_vendor/urllib3/__pycache__/response.cpython-38.opt-1.pyc000064400000047557151464452710022554 0ustar00U

.el@s\ddlmZddlmZddlZddlZddlZddlmZ	ddlm
ZzddlZWne
k
rldZYnXddlmZddlmZmZmZmZmZmZmZdd	lmZmZdd
lmZddlm Z m!Z!ddl"m#Z#m$Z$e%e&Z'Gd
dde(Z)Gddde(Z*Gddde(Z+edk	r.Gddde(Z,Gddde(Z-ddZ.Gdddej/Z0dS))absolute_import)contextmanagerN)timeout)error)HTTPHeaderDict)BodyNotHttplibCompatible
ProtocolErrorDecodeErrorReadTimeoutErrorResponseNotChunkedIncompleteRead
InvalidHeader)string_typesPY3)http_client)
HTTPExceptionBaseSSLError)is_fp_closedis_response_to_headc@s$eZdZddZddZddZdS)DeflateDecodercCsd|_d|_t|_dS)NT)
_first_try_datazlib
decompressobj_objselfr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py__init__!szDeflateDecoder.__init__cCst|j|SNgetattrrrnamerrr __getattr__&szDeflateDecoder.__getattr__cCs|s|S|js|j|S|j|7_z"|j|}|rFd|_d|_|WStjk
rd|_ttj|_z||jWYSd|_XYnXdS)NF)rr
decompressrrrr	MAX_WBITS)rdataZdecompressedrrr r()s"zDeflateDecoder.decompressN__name__
__module____qualname__r!r'r(rrrr r src@seZdZdZdZdZdS)GzipDecoderStaterrN)r,r-r.FIRST_MEMBER
OTHER_MEMBERSSWALLOW_DATArrrr r/@sr/c@s$eZdZddZddZddZdS)GzipDecodercCstdtj|_tj|_dSN)rrr)rr/r1_staterrrr r!HszGzipDecoder.__init__cCst|j|Sr"r#r%rrr r'LszGzipDecoder.__getattr__cCst}|jtjks|st|Sz||j|7}Wn<tjk
rn|j}tj|_|tj	krht|YSYnX|jj
}|st|Stj	|_tdtj|_qdSr5)
	bytearrayr7r/r3bytesrr(rrr2Zunused_datarr))rr*retZprevious_staterrr r(Os 
zGzipDecoder.decompressNr+rrrr r4Gsr4c@s$eZdZddZddZddZdS)
BrotliDecodercCst|_dSr")brotliZDecompressorrrrrr r!kszBrotliDecoder.__init__cCs$t|jdr|j|S|j|S)Nr()hasattrrr(Zprocess)rr*rrr r(nszBrotliDecoder.decompresscCst|jdr|jSdS)Nflushr)r=rr>rrrr r>ss
zBrotliDecoder.flushN)r,r-r.r!r(r>rrrr r;gsr;c@s(eZdZdZddZddZddZdS)	MultiDecodera
    From RFC7231:
        If one or more encodings have been applied to a representation, the
        sender that applied the encodings MUST generate a Content-Encoding
        header field that lists the content codings in the order in which
        they were applied.
    cCsdd|dD|_dS)NcSsg|]}t|qSr)_get_decoderstrip).0mrrr 
sz)MultiDecoder.__init__..,)split	_decoders)rZmodesrrr r!szMultiDecoder.__init__cCs|jdSNr)rGr>rrrr r>szMultiDecoder.flushcCst|jD]}||}q
|Sr")reversedrGr()rr*drrr r(szMultiDecoder.decompressN)r,r-r.__doc__r!r>r(rrrr r?ysr?cCs:d|krt|S|dkrtStdk	r4|dkr4tStS)NrEgzipbr)r?r4r<r;r)moderrr r@sr@c@sdeZdZdZddgZedk	r&edg7Zdddd	d
gZdLddZddZddZ	e
ddZe
ddZddZ
ddZddZdd ZeejfZedk	reejf7Zd!d"Zd#d$Zed%d&ZdMd'd(ZdNd*d+Zed,d-Zd.d/ZdOd0d1Zd2d3Zd4d5Z e
d6d7Z!d8d9Z"d:d;Z#dd?Z%d@dAZ&dBdCZ'dDdEZ(dPdFdGZ)dHdIZ*dJdKZ+dS)QHTTPResponsea
    HTTP Response container.

    Backwards-compatible to httplib's HTTPResponse but the response ``body`` is
    loaded and decoded on-demand when the ``data`` property is accessed.  This
    class is also compatible with the Python standard library's :mod:`io`
    module, and can hence be treated as a readable object in the context of that
    framework.

    Extra parameters for behaviour not present in httplib.HTTPResponse:

    :param preload_content:
        If True, the response's body will be preloaded during construction.

    :param decode_content:
        If True, will attempt to decode the body based on the
        'content-encoding' header.

    :param original_response:
        When this HTTPResponse wrapper is generated from an httplib.HTTPResponse
        object, it's convenient to include the original for debug purposes. It's
        otherwise unused.

    :param retries:
        The retries contains the last :class:`~urllib3.util.retry.Retry` that
        was used during the request.

    :param enforce_content_length:
        Enforce content length checking. Body returned by server must match
        value of Content-Length header, if present. Otherwise, raise error.
    rLZdeflateNrMi-i.i/i3i4rTFcCst|tr||_n
t||_||_||_||_||_||_|
|_||_	||_
d|_d|_d|_
|	|_d|_||_||_|rt|ttfr||_|
|_||_t|dr||_
d|_d|_|jdd}dd|dD}d	|krd
|_|||_|r|js|j|d|_dS)NrreadFztransfer-encodingrPcss|]}|VqdSr")rA)rBencrrr 	sz(HTTPResponse.__init__..rEchunkedTdecode_content)
isinstancerheadersstatusversionreasonstrictrVretriesenforce_content_length
auto_close_decoder_body_fp_original_response_fp_bytes_readmsg_request_url
basestringr9_pool_connectionr=rT
chunk_leftgetlowerrF_init_lengthlength_remainingrQ)rbodyrXrYrZr[r\Zpreload_contentrVoriginal_responseZpool
connectionrer]r^request_methodZrequest_urlr_Ztr_enc	encodingsrrr r!sB


zHTTPResponse.__init__cCs|j|jkr|jdSdS)a
        Should we redirect and where to?

        :returns: Truthy redirect location string if we got a redirect status
            code and valid location. ``None`` if redirect status and no
            location. ``False`` if not a redirect status code.
        locationF)rYREDIRECT_STATUSESrXrkrrrr get_redirect_locationsz"HTTPResponse.get_redirect_locationcCs(|jr|jsdS|j|jd|_dSr")rhriZ	_put_connrrrr release_connszHTTPResponse.release_conncCs"|jr|jS|jr|jddSdS)NT)
cache_content)rarbrQrrrr r*szHTTPResponse.datacCs|jSr")rirrrr rq!szHTTPResponse.connectioncCs
t|jSr")rrbrrrr isclosed%szHTTPResponse.isclosedcCs|jS)z
        Obtain the number of bytes pulled over the wire so far. May differ from
        the amount of content returned by :meth:``HTTPResponse.read`` if bytes
        are encoded on the wire (e.g, compressed).
        )rdrrrr tell(szHTTPResponse.tellcCs|jd}|dk	r|jr(tddSz.rErz8Content-Length contained multiple unmatching values (%s)r)i0dZHEAD)
rXrkrTlogZwarningsetrFlenrpop
ValueErrorr{rY)rrrZlengthZlengthsrYrrr rm0s6

&zHTTPResponse._init_lengthcshjdd}jdkrd|jkr2t|_n2d|krdfdd|dD}t|rdt|_dS)z=
        Set-up the _decoder attribute if necessary.
        content-encodingrPNrEcs"g|]}|jkr|qSr)rACONTENT_DECODERS)rBerrr rDosz.HTTPResponse._init_decoder..)rXrkrlr`rr@rFr)rcontent_encodingrsrrr 
_init_decoderds


zHTTPResponse._init_decoderc
Csx|s|Sz|jr|j|}WnD|jk
rb}z$|jdd}td||W5d}~XYnX|rt||7}|S)zN
        Decode the data passed in and potentially flush the decoder.
        rrPzEReceived response with content-encoding: %s, but failed to decode it.N)r`r(DECODER_ERROR_CLASSESrXrkrlr
_flush_decoder)rr*rV
flush_decoderrrrrr _decode{s zHTTPResponse._decodecCs$|jr |jd}||jSdS)zk
        Flushes the decoder. Should only be called if the decoder is actually
        being used.
        r)r`r(r>)rZbufrrr rszHTTPResponse._flush_decoderc
csd}zz
dVWntk
r2t|jddYnptk
rn}z dt	|krPt|jddW5d}~XYn4t
tfk
r}ztd||W5d}~XYnXd}W5|s|jr|j|jr|j|jr|jr|XdS)z
        Catch low-level python exceptions, instead re-raising urllib3
        variants, so that low-level exceptions are not leaked in the
        high-level api.

        On exit, release the connection back to the pool.
        FNzRead timed out.zread operation timed outzConnection broken: %rT)
rccloseriryrw
SocketTimeoutrrhrstrrSocketErrorr	)rZ
clean_exitrrrr _error_catchers(	
 

zHTTPResponse._error_catcherc	Cs||dkr|j}|jdkr$dSd}t|jdd}|v|dkr`|sV|jnd}d}nPd}|st|j|nd}|dkr|s|jd}|jr|jdkrt	|j
|jW5QRX|r|j
t|7_
|jdk	r|jt|8_||||}|r||_
|S)aP
        Similar to :meth:`httplib.HTTPResponse.read`, but with two additional
        parameters: ``decode_content`` and ``cache_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.

        :param cache_content:
            If True, will save the returned data such that the same result is
            returned despite of the state of the underlying file object. This
            is useful if you want the ``.data`` property to continue working
            after having ``.read()`` the file object. (Overridden if ``amt`` is
            set.)
        NFclosedrTr)rN)rrVrbr$rrQrr^rnr
rdrrra)ramtrVrxrZ	fp_closedr*rrr rQs:



	
zHTTPResponse.readccsR|jr*|r*|j||dD]
}|Vqn$t|jsN|j||d}|r*|Vq*dS)a_
        A generator wrapper for the read() method. A call will block until
        ``amt`` bytes have been read from the connection or until the
        connection is closed.

        :param amt:
            How much of the content to read. The generator will return up to
            much data per iteration, but may return less. This is particularly
            likely when using compressed data. However, the empty string will
            never be returned.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        rU)rrVN)rTsupports_chunked_readsread_chunkedrrbrQ)rrrVliner*rrr streams

zHTTPResponse.streamc
Ks`|j}t|ts,tr"t|}n
t|}t|dd}|f|||j|j|j	||d|}|S)a
        Given an :class:`httplib.HTTPResponse` instance ``r``, return a
        corresponding :class:`urllib3.response.HTTPResponse` object.

        Remaining parameters are passed to the HTTPResponse constructor, along
        with ``original_response=r``.
        r\r)rorXrYrZr[r\rp)
rerWrritemsfrom_httplibr$rYrZr[)ZResponseClsrZresponse_kwrXr\Zresprrr r9s$	


zHTTPResponse.from_httplibcCs|jSr"rXrrrr 
getheadersZszHTTPResponse.getheaderscCs|j||Sr")rXrk)rr&defaultrrr 	getheader]szHTTPResponse.getheadercCs|jSr"rrrrr infoaszHTTPResponse.infocCs6|js|j|jr |j|js2tj|dSr")rrbrrir_ioIOBaserrrr res

zHTTPResponse.closecCsT|jstjj|S|jdkr"dSt|jdr8|jSt|jdrL|jjSdSdS)NTryr)r_rrr__get__rbr=ryrrrr ros

zHTTPResponse.closedcCs6|jdkrtdnt|jdr*|jStddS)Nz-HTTPResponse has no file to get a fileno fromfilenozOThe file-like object this HTTPResponse is wrapped around has no file descriptor)rbIOErrorr=rrrrr r|s


zHTTPResponse.filenocCs2|jdk	r.t|jdr.t|jdds.|jSdS)Nr>rF)rbr=r$r>rrrr r>s
zHTTPResponse.flushcCsdS)NTrrrrr readableszHTTPResponse.readablecCs:|t|}t|dkrdS||dt|<t|SdSrH)rQr)rbZtemprrr readintos
zHTTPResponse.readintocCst|jdS)z
        Checks if the underlying file-like object looks like a
        httplib.HTTPResponse object. We do this by testing for the fp
        attribute. If it is present we assume it returns raw chunks as
        processed by read_chunked().
        fp)r=rbrrrr rsz#HTTPResponse.supports_chunked_readscCsf|jdk	rdS|jj}|ddd}zt|d|_Wn&tk
r`|t	|YnXdS)N;rrr6)
rjrbrreadlinerFr{rrhttplibr
)rrrrr _update_chunk_lengths
z!HTTPResponse._update_chunk_lengthcCsd}|dkr2|j|j}|}|jdd|_nv||jkrZ|j|}|j||_|}nN||jkr|j|}|jdd|_|}n |j|j}|jdd|_|S)Nr0)rbZ
_safe_readrj)rrZreturned_chunkchunkvaluerrr 
_handle_chunks&

zHTTPResponse._handle_chunkc	cs||jstd|s&td||jrXt|jrX|jW5QRdS|j	j
dkrrW5QRdS||jdkrq|
|}|j||dd}|rr|Vqr|r|}|r|V|j	j
}|sq|dkrqq|jr|jW5QRXdS)a
        Similar to :meth:`HTTPResponse.read`, but with an additional
        parameter: ``decode_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        zHResponse is not chunked. Header 'transfer-encoding: chunked' is missing.zgBody should be httplib.HTTPResponse like. It should have have an fp attribute which returns raw chunks.NrF)rVrs
)rrTrrrrrcrrrbrrrjrrrr)rrrVrZdecodedrrrr rsJ



zHTTPResponse.read_chunkedcCs.|jdk	r$t|jjr$|jjdjS|jSdS)z
        Returns the URL that was the source of this response.
        If the request that generated this response redirected, this method
        will return the final redirect location.
        N)r]rhistoryZredirect_locationrfrrrr geturlszHTTPResponse.geturlccsdg}|jddD]n}d|krv|d}d||ddV|ddD]}|dVqL|drp|dg}qg}q||q|rd|VdS)NrTrU
rrr)rrFjoinappend)rbufferrxrrr __iter__s
zHTTPResponse.__iter__)rPNrrNrTTNNNNNFNNT)NNF)rN)N)NN),r,r-r.rKrr<rur!rvrwpropertyr*rqryrzrmrrrrrrrrrrQrclassmethodrrrrrrrr>rrrrrrrrrrrr rOsx 

C


4

8
J

 


		
FrO)1Z
__future__r
contextlibrrrZloggingZsocketrrrrr<ImportError_collectionsr
exceptionsrr	r
rrr
rZpackages.sixrrgrZpackages.six.movesrrrqrrZ
util.responserrZ	getLoggerr,robjectrr/r4r;r?r@rrOrrrr s2
$	
 

site-packages/pip/_vendor/urllib3/__pycache__/response.cpython-38.pyc000064400000047557151464452710021615 0ustar00U

.el@s\ddlmZddlmZddlZddlZddlZddlmZ	ddlm
ZzddlZWne
k
rldZYnXddlmZddlmZmZmZmZmZmZmZdd	lmZmZdd
lmZddlm Z m!Z!ddl"m#Z#m$Z$e%e&Z'Gd
dde(Z)Gddde(Z*Gddde(Z+edk	r.Gddde(Z,Gddde(Z-ddZ.Gdddej/Z0dS))absolute_import)contextmanagerN)timeout)error)HTTPHeaderDict)BodyNotHttplibCompatible
ProtocolErrorDecodeErrorReadTimeoutErrorResponseNotChunkedIncompleteRead
InvalidHeader)string_typesPY3)http_client)
HTTPExceptionBaseSSLError)is_fp_closedis_response_to_headc@s$eZdZddZddZddZdS)DeflateDecodercCsd|_d|_t|_dS)NT)
_first_try_datazlib
decompressobj_objselfr@/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/response.py__init__!szDeflateDecoder.__init__cCst|j|SNgetattrrrnamerrr __getattr__&szDeflateDecoder.__getattr__cCs|s|S|js|j|S|j|7_z"|j|}|rFd|_d|_|WStjk
rd|_ttj|_z||jWYSd|_XYnXdS)NF)rr
decompressrrrr	MAX_WBITS)rdataZdecompressedrrr r()s"zDeflateDecoder.decompressN__name__
__module____qualname__r!r'r(rrrr r src@seZdZdZdZdZdS)GzipDecoderStaterrN)r,r-r.FIRST_MEMBER
OTHER_MEMBERSSWALLOW_DATArrrr r/@sr/c@s$eZdZddZddZddZdS)GzipDecodercCstdtj|_tj|_dSN)rrr)rr/r1_staterrrr r!HszGzipDecoder.__init__cCst|j|Sr"r#r%rrr r'LszGzipDecoder.__getattr__cCst}|jtjks|st|Sz||j|7}Wn<tjk
rn|j}tj|_|tj	krht|YSYnX|jj
}|st|Stj	|_tdtj|_qdSr5)
	bytearrayr7r/r3bytesrr(rrr2Zunused_datarr))rr*retZprevious_staterrr r(Os 
zGzipDecoder.decompressNr+rrrr r4Gsr4c@s$eZdZddZddZddZdS)
BrotliDecodercCst|_dSr")brotliZDecompressorrrrrr r!kszBrotliDecoder.__init__cCs$t|jdr|j|S|j|S)Nr()hasattrrr(Zprocess)rr*rrr r(nszBrotliDecoder.decompresscCst|jdr|jSdS)Nflushr)r=rr>rrrr r>ss
zBrotliDecoder.flushN)r,r-r.r!r(r>rrrr r;gsr;c@s(eZdZdZddZddZddZdS)	MultiDecodera
    From RFC7231:
        If one or more encodings have been applied to a representation, the
        sender that applied the encodings MUST generate a Content-Encoding
        header field that lists the content codings in the order in which
        they were applied.
    cCsdd|dD|_dS)NcSsg|]}t|qSr)_get_decoderstrip).0mrrr 
sz)MultiDecoder.__init__..,)split	_decoders)rZmodesrrr r!szMultiDecoder.__init__cCs|jdSNr)rGr>rrrr r>szMultiDecoder.flushcCst|jD]}||}q
|Sr")reversedrGr()rr*drrr r(szMultiDecoder.decompressN)r,r-r.__doc__r!r>r(rrrr r?ysr?cCs:d|krt|S|dkrtStdk	r4|dkr4tStS)NrEgzipbr)r?r4r<r;r)moderrr r@sr@c@sdeZdZdZddgZedk	r&edg7Zdddd	d
gZdLddZddZddZ	e
ddZe
ddZddZ
ddZddZdd ZeejfZedk	reejf7Zd!d"Zd#d$Zed%d&ZdMd'd(ZdNd*d+Zed,d-Zd.d/ZdOd0d1Zd2d3Zd4d5Z e
d6d7Z!d8d9Z"d:d;Z#dd?Z%d@dAZ&dBdCZ'dDdEZ(dPdFdGZ)dHdIZ*dJdKZ+dS)QHTTPResponsea
    HTTP Response container.

    Backwards-compatible to httplib's HTTPResponse but the response ``body`` is
    loaded and decoded on-demand when the ``data`` property is accessed.  This
    class is also compatible with the Python standard library's :mod:`io`
    module, and can hence be treated as a readable object in the context of that
    framework.

    Extra parameters for behaviour not present in httplib.HTTPResponse:

    :param preload_content:
        If True, the response's body will be preloaded during construction.

    :param decode_content:
        If True, will attempt to decode the body based on the
        'content-encoding' header.

    :param original_response:
        When this HTTPResponse wrapper is generated from an httplib.HTTPResponse
        object, it's convenient to include the original for debug purposes. It's
        otherwise unused.

    :param retries:
        The retries contains the last :class:`~urllib3.util.retry.Retry` that
        was used during the request.

    :param enforce_content_length:
        Enforce content length checking. Body returned by server must match
        value of Content-Length header, if present. Otherwise, raise error.
    rLZdeflateNrMi-i.i/i3i4rTFcCst|tr||_n
t||_||_||_||_||_||_|
|_||_	||_
d|_d|_d|_
|	|_d|_||_||_|rt|ttfr||_|
|_||_t|dr||_
d|_d|_|jdd}dd|dD}d	|krd
|_|||_|r|js|j|d|_dS)NrreadFztransfer-encodingrPcss|]}|VqdSr")rA)rBencrrr 	sz(HTTPResponse.__init__..rEchunkedTdecode_content)
isinstancerheadersstatusversionreasonstrictrVretriesenforce_content_length
auto_close_decoder_body_fp_original_response_fp_bytes_readmsg_request_url
basestringr9_pool_connectionr=rT
chunk_leftgetlowerrF_init_lengthlength_remainingrQ)rbodyrXrYrZr[r\Zpreload_contentrVoriginal_responseZpool
connectionrer]r^request_methodZrequest_urlr_Ztr_enc	encodingsrrr r!sB


zHTTPResponse.__init__cCs|j|jkr|jdSdS)a
        Should we redirect and where to?

        :returns: Truthy redirect location string if we got a redirect status
            code and valid location. ``None`` if redirect status and no
            location. ``False`` if not a redirect status code.
        locationF)rYREDIRECT_STATUSESrXrkrrrr get_redirect_locationsz"HTTPResponse.get_redirect_locationcCs(|jr|jsdS|j|jd|_dSr")rhriZ	_put_connrrrr release_connszHTTPResponse.release_conncCs"|jr|jS|jr|jddSdS)NT)
cache_content)rarbrQrrrr r*szHTTPResponse.datacCs|jSr")rirrrr rq!szHTTPResponse.connectioncCs
t|jSr")rrbrrrr isclosed%szHTTPResponse.isclosedcCs|jS)z
        Obtain the number of bytes pulled over the wire so far. May differ from
        the amount of content returned by :meth:``HTTPResponse.read`` if bytes
        are encoded on the wire (e.g, compressed).
        )rdrrrr tell(szHTTPResponse.tellcCs|jd}|dk	r|jr(tddSz.rErz8Content-Length contained multiple unmatching values (%s)r)i0dZHEAD)
rXrkrTlogZwarningsetrFlenrpop
ValueErrorr{rY)rrrZlengthZlengthsrYrrr rm0s6

&zHTTPResponse._init_lengthcshjdd}jdkrd|jkr2t|_n2d|krdfdd|dD}t|rdt|_dS)z=
        Set-up the _decoder attribute if necessary.
        content-encodingrPNrEcs"g|]}|jkr|qSr)rACONTENT_DECODERS)rBerrr rDosz.HTTPResponse._init_decoder..)rXrkrlr`rr@rFr)rcontent_encodingrsrrr 
_init_decoderds


zHTTPResponse._init_decoderc
Csx|s|Sz|jr|j|}WnD|jk
rb}z$|jdd}td||W5d}~XYnX|rt||7}|S)zN
        Decode the data passed in and potentially flush the decoder.
        rrPzEReceived response with content-encoding: %s, but failed to decode it.N)r`r(DECODER_ERROR_CLASSESrXrkrlr
_flush_decoder)rr*rV
flush_decoderrrrrr _decode{s zHTTPResponse._decodecCs$|jr |jd}||jSdS)zk
        Flushes the decoder. Should only be called if the decoder is actually
        being used.
        r)r`r(r>)rZbufrrr rszHTTPResponse._flush_decoderc
csd}zz
dVWntk
r2t|jddYnptk
rn}z dt	|krPt|jddW5d}~XYn4t
tfk
r}ztd||W5d}~XYnXd}W5|s|jr|j|jr|j|jr|jr|XdS)z
        Catch low-level python exceptions, instead re-raising urllib3
        variants, so that low-level exceptions are not leaked in the
        high-level api.

        On exit, release the connection back to the pool.
        FNzRead timed out.zread operation timed outzConnection broken: %rT)
rccloseriryrw
SocketTimeoutrrhrstrrSocketErrorr	)rZ
clean_exitrrrr _error_catchers(	
 

zHTTPResponse._error_catcherc	Cs||dkr|j}|jdkr$dSd}t|jdd}|v|dkr`|sV|jnd}d}nPd}|st|j|nd}|dkr|s|jd}|jr|jdkrt	|j
|jW5QRX|r|j
t|7_
|jdk	r|jt|8_||||}|r||_
|S)aP
        Similar to :meth:`httplib.HTTPResponse.read`, but with two additional
        parameters: ``decode_content`` and ``cache_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.

        :param cache_content:
            If True, will save the returned data such that the same result is
            returned despite of the state of the underlying file object. This
            is useful if you want the ``.data`` property to continue working
            after having ``.read()`` the file object. (Overridden if ``amt`` is
            set.)
        NFclosedrTr)rN)rrVrbr$rrQrr^rnr
rdrrra)ramtrVrxrZ	fp_closedr*rrr rQs:



	
zHTTPResponse.readccsR|jr*|r*|j||dD]
}|Vqn$t|jsN|j||d}|r*|Vq*dS)a_
        A generator wrapper for the read() method. A call will block until
        ``amt`` bytes have been read from the connection or until the
        connection is closed.

        :param amt:
            How much of the content to read. The generator will return up to
            much data per iteration, but may return less. This is particularly
            likely when using compressed data. However, the empty string will
            never be returned.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        rU)rrVN)rTsupports_chunked_readsread_chunkedrrbrQ)rrrVliner*rrr streams

zHTTPResponse.streamc
Ks`|j}t|ts,tr"t|}n
t|}t|dd}|f|||j|j|j	||d|}|S)a
        Given an :class:`httplib.HTTPResponse` instance ``r``, return a
        corresponding :class:`urllib3.response.HTTPResponse` object.

        Remaining parameters are passed to the HTTPResponse constructor, along
        with ``original_response=r``.
        r\r)rorXrYrZr[r\rp)
rerWrritemsfrom_httplibr$rYrZr[)ZResponseClsrZresponse_kwrXr\Zresprrr r9s$	


zHTTPResponse.from_httplibcCs|jSr"rXrrrr 
getheadersZszHTTPResponse.getheaderscCs|j||Sr")rXrk)rr&defaultrrr 	getheader]szHTTPResponse.getheadercCs|jSr"rrrrr infoaszHTTPResponse.infocCs6|js|j|jr |j|js2tj|dSr")rrbrrir_ioIOBaserrrr res

zHTTPResponse.closecCsT|jstjj|S|jdkr"dSt|jdr8|jSt|jdrL|jjSdSdS)NTryr)r_rrr__get__rbr=ryrrrr ros

zHTTPResponse.closedcCs6|jdkrtdnt|jdr*|jStddS)Nz-HTTPResponse has no file to get a fileno fromfilenozOThe file-like object this HTTPResponse is wrapped around has no file descriptor)rbIOErrorr=rrrrr r|s


zHTTPResponse.filenocCs2|jdk	r.t|jdr.t|jdds.|jSdS)Nr>rF)rbr=r$r>rrrr r>s
zHTTPResponse.flushcCsdS)NTrrrrr readableszHTTPResponse.readablecCs:|t|}t|dkrdS||dt|<t|SdSrH)rQr)rbZtemprrr readintos
zHTTPResponse.readintocCst|jdS)z
        Checks if the underlying file-like object looks like a
        httplib.HTTPResponse object. We do this by testing for the fp
        attribute. If it is present we assume it returns raw chunks as
        processed by read_chunked().
        fp)r=rbrrrr rsz#HTTPResponse.supports_chunked_readscCsf|jdk	rdS|jj}|ddd}zt|d|_Wn&tk
r`|t	|YnXdS)N;rrr6)
rjrbrreadlinerFr{rrhttplibr
)rrrrr _update_chunk_lengths
z!HTTPResponse._update_chunk_lengthcCsd}|dkr2|j|j}|}|jdd|_nv||jkrZ|j|}|j||_|}nN||jkr|j|}|jdd|_|}n |j|j}|jdd|_|S)Nr0)rbZ
_safe_readrj)rrZreturned_chunkchunkvaluerrr 
_handle_chunks&

zHTTPResponse._handle_chunkc	cs||jstd|s&td||jrXt|jrX|jW5QRdS|j	j
dkrrW5QRdS||jdkrq|
|}|j||dd}|rr|Vqr|r|}|r|V|j	j
}|sq|dkrqq|jr|jW5QRXdS)a
        Similar to :meth:`HTTPResponse.read`, but with an additional
        parameter: ``decode_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        zHResponse is not chunked. Header 'transfer-encoding: chunked' is missing.zgBody should be httplib.HTTPResponse like. It should have have an fp attribute which returns raw chunks.NrF)rVrs
)rrTrrrrrcrrrbrrrjrrrr)rrrVrZdecodedrrrr rsJ



zHTTPResponse.read_chunkedcCs.|jdk	r$t|jjr$|jjdjS|jSdS)z
        Returns the URL that was the source of this response.
        If the request that generated this response redirected, this method
        will return the final redirect location.
        N)r]rhistoryZredirect_locationrfrrrr geturlszHTTPResponse.geturlccsdg}|jddD]n}d|krv|d}d||ddV|ddD]}|dVqL|drp|dg}qg}q||q|rd|VdS)NrTrU
rrr)rrFjoinappend)rbufferrxrrr __iter__s
zHTTPResponse.__iter__)rPNrrNrTTNNNNNFNNT)NNF)rN)N)NN),r,r-r.rKrr<rur!rvrwpropertyr*rqryrzrmrrrrrrrrrrQrclassmethodrrrrrrrr>rrrrrrrrrrrr rOsx 

C


4

8
J

 


		
FrO)1Z
__future__r
contextlibrrrZloggingZsocketrrrrr<ImportError_collectionsr
exceptionsrr	r
rrr
rZpackages.sixrrgrZpackages.six.movesrrrqrrZ
util.responserrZ	getLoggerr,robjectrr/r4r;r?r@rrOrrrr s2
$	
 

site-packages/pip/_vendor/urllib3/contrib/appengine.py000064400000026062151464452710017063 0ustar00"""
This module provides a pool manager that uses Google App Engine's
`URLFetch Service `_.

Example usage::

    from pip._vendor.urllib3 import PoolManager
    from pip._vendor.urllib3.contrib.appengine import AppEngineManager, is_appengine_sandbox

    if is_appengine_sandbox():
        # AppEngineManager uses AppEngine's URLFetch API behind the scenes
        http = AppEngineManager()
    else:
        # PoolManager uses a socket-level API behind the scenes
        http = PoolManager()

    r = http.request('GET', 'https://google.com/')

There are `limitations `_ to the URLFetch service and it may not be
the best choice for your application. There are three options for using
urllib3 on Google App Engine:

1. You can use :class:`AppEngineManager` with URLFetch. URLFetch is
   cost-effective in many circumstances as long as your usage is within the
   limitations.
2. You can use a normal :class:`~urllib3.PoolManager` by enabling sockets.
   Sockets also have `limitations and restrictions
   `_ and have a lower free quota than URLFetch.
   To use sockets, be sure to specify the following in your ``app.yaml``::

        env_variables:
            GAE_USE_SOCKETS_HTTPLIB : 'true'

3. If you are using `App Engine Flexible
`_, you can use the standard
:class:`PoolManager` without any configuration or special environment variables.
"""

from __future__ import absolute_import
import io
import logging
import warnings
from ..packages.six.moves.urllib.parse import urljoin

from ..exceptions import (
    HTTPError,
    HTTPWarning,
    MaxRetryError,
    ProtocolError,
    TimeoutError,
    SSLError,
)

from ..request import RequestMethods
from ..response import HTTPResponse
from ..util.timeout import Timeout
from ..util.retry import Retry
from . import _appengine_environ

try:
    from google.appengine.api import urlfetch
except ImportError:
    urlfetch = None


log = logging.getLogger(__name__)


class AppEnginePlatformWarning(HTTPWarning):
    pass


class AppEnginePlatformError(HTTPError):
    pass


class AppEngineManager(RequestMethods):
    """
    Connection manager for Google App Engine sandbox applications.

    This manager uses the URLFetch service directly instead of using the
    emulated httplib, and is subject to URLFetch limitations as described in
    the App Engine documentation `here
    `_.

    Notably it will raise an :class:`AppEnginePlatformError` if:
        * URLFetch is not available.
        * If you attempt to use this on App Engine Flexible, as full socket
          support is available.
        * If a request size is more than 10 megabytes.
        * If a response size is more than 32 megabtyes.
        * If you use an unsupported request method such as OPTIONS.

    Beyond those cases, it will raise normal urllib3 errors.
    """

    def __init__(
        self,
        headers=None,
        retries=None,
        validate_certificate=True,
        urlfetch_retries=True,
    ):
        if not urlfetch:
            raise AppEnginePlatformError(
                "URLFetch is not available in this environment."
            )

        if is_prod_appengine_mvms():
            raise AppEnginePlatformError(
                "Use normal urllib3.PoolManager instead of AppEngineManager"
                "on Managed VMs, as using URLFetch is not necessary in "
                "this environment."
            )

        warnings.warn(
            "urllib3 is using URLFetch on Google App Engine sandbox instead "
            "of sockets. To use sockets directly instead of URLFetch see "
            "https://urllib3.readthedocs.io/en/latest/reference/urllib3.contrib.html.",
            AppEnginePlatformWarning,
        )

        RequestMethods.__init__(self, headers)
        self.validate_certificate = validate_certificate
        self.urlfetch_retries = urlfetch_retries

        self.retries = retries or Retry.DEFAULT

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        # Return False to re-raise any potential exceptions
        return False

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        retries=None,
        redirect=True,
        timeout=Timeout.DEFAULT_TIMEOUT,
        **response_kw
    ):

        retries = self._get_retries(retries, redirect)

        try:
            follow_redirects = redirect and retries.redirect != 0 and retries.total
            response = urlfetch.fetch(
                url,
                payload=body,
                method=method,
                headers=headers or {},
                allow_truncated=False,
                follow_redirects=self.urlfetch_retries and follow_redirects,
                deadline=self._get_absolute_timeout(timeout),
                validate_certificate=self.validate_certificate,
            )
        except urlfetch.DeadlineExceededError as e:
            raise TimeoutError(self, e)

        except urlfetch.InvalidURLError as e:
            if "too large" in str(e):
                raise AppEnginePlatformError(
                    "URLFetch request too large, URLFetch only "
                    "supports requests up to 10mb in size.",
                    e,
                )
            raise ProtocolError(e)

        except urlfetch.DownloadError as e:
            if "Too many redirects" in str(e):
                raise MaxRetryError(self, url, reason=e)
            raise ProtocolError(e)

        except urlfetch.ResponseTooLargeError as e:
            raise AppEnginePlatformError(
                "URLFetch response too large, URLFetch only supports"
                "responses up to 32mb in size.",
                e,
            )

        except urlfetch.SSLCertificateError as e:
            raise SSLError(e)

        except urlfetch.InvalidMethodError as e:
            raise AppEnginePlatformError(
                "URLFetch does not support method: %s" % method, e
            )

        http_response = self._urlfetch_response_to_http_response(
            response, retries=retries, **response_kw
        )

        # Handle redirect?
        redirect_location = redirect and http_response.get_redirect_location()
        if redirect_location:
            # Check for redirect response
            if self.urlfetch_retries and retries.raise_on_redirect:
                raise MaxRetryError(self, url, "too many redirects")
            else:
                if http_response.status == 303:
                    method = "GET"

                try:
                    retries = retries.increment(
                        method, url, response=http_response, _pool=self
                    )
                except MaxRetryError:
                    if retries.raise_on_redirect:
                        raise MaxRetryError(self, url, "too many redirects")
                    return http_response

                retries.sleep_for_retry(http_response)
                log.debug("Redirecting %s -> %s", url, redirect_location)
                redirect_url = urljoin(url, redirect_location)
                return self.urlopen(
                    method,
                    redirect_url,
                    body,
                    headers,
                    retries=retries,
                    redirect=redirect,
                    timeout=timeout,
                    **response_kw
                )

        # Check if we should retry the HTTP response.
        has_retry_after = bool(http_response.getheader("Retry-After"))
        if retries.is_retry(method, http_response.status, has_retry_after):
            retries = retries.increment(method, url, response=http_response, _pool=self)
            log.debug("Retry: %s", url)
            retries.sleep(http_response)
            return self.urlopen(
                method,
                url,
                body=body,
                headers=headers,
                retries=retries,
                redirect=redirect,
                timeout=timeout,
                **response_kw
            )

        return http_response

    def _urlfetch_response_to_http_response(self, urlfetch_resp, **response_kw):

        if is_prod_appengine():
            # Production GAE handles deflate encoding automatically, but does
            # not remove the encoding header.
            content_encoding = urlfetch_resp.headers.get("content-encoding")

            if content_encoding == "deflate":
                del urlfetch_resp.headers["content-encoding"]

        transfer_encoding = urlfetch_resp.headers.get("transfer-encoding")
        # We have a full response's content,
        # so let's make sure we don't report ourselves as chunked data.
        if transfer_encoding == "chunked":
            encodings = transfer_encoding.split(",")
            encodings.remove("chunked")
            urlfetch_resp.headers["transfer-encoding"] = ",".join(encodings)

        original_response = HTTPResponse(
            # In order for decoding to work, we must present the content as
            # a file-like object.
            body=io.BytesIO(urlfetch_resp.content),
            msg=urlfetch_resp.header_msg,
            headers=urlfetch_resp.headers,
            status=urlfetch_resp.status_code,
            **response_kw
        )

        return HTTPResponse(
            body=io.BytesIO(urlfetch_resp.content),
            headers=urlfetch_resp.headers,
            status=urlfetch_resp.status_code,
            original_response=original_response,
            **response_kw
        )

    def _get_absolute_timeout(self, timeout):
        if timeout is Timeout.DEFAULT_TIMEOUT:
            return None  # Defer to URLFetch's default.
        if isinstance(timeout, Timeout):
            if timeout._read is not None or timeout._connect is not None:
                warnings.warn(
                    "URLFetch does not support granular timeout settings, "
                    "reverting to total or default URLFetch timeout.",
                    AppEnginePlatformWarning,
                )
            return timeout.total
        return timeout

    def _get_retries(self, retries, redirect):
        if not isinstance(retries, Retry):
            retries = Retry.from_int(retries, redirect=redirect, default=self.retries)

        if retries.connect or retries.read or retries.redirect:
            warnings.warn(
                "URLFetch only supports total retries and does not "
                "recognize connect, read, or redirect retry parameters.",
                AppEnginePlatformWarning,
            )

        return retries


# Alias methods from _appengine_environ to maintain public API interface.

is_appengine = _appengine_environ.is_appengine
is_appengine_sandbox = _appengine_environ.is_appengine_sandbox
is_local_appengine = _appengine_environ.is_local_appengine
is_prod_appengine = _appengine_environ.is_prod_appengine
is_prod_appengine_mvms = _appengine_environ.is_prod_appengine_mvms
site-packages/pip/_vendor/urllib3/contrib/_appengine_environ.py000064400000001303151464452710020751 0ustar00"""
This module provides means to detect the App Engine environment.
"""

import os


def is_appengine():
    return is_local_appengine() or is_prod_appengine() or is_prod_appengine_mvms()


def is_appengine_sandbox():
    return is_appengine() and not is_prod_appengine_mvms()


def is_local_appengine():
    return (
        "APPENGINE_RUNTIME" in os.environ
        and "Development/" in os.environ["SERVER_SOFTWARE"]
    )


def is_prod_appengine():
    return (
        "APPENGINE_RUNTIME" in os.environ
        and "Google App Engine/" in os.environ["SERVER_SOFTWARE"]
        and not is_prod_appengine_mvms()
    )


def is_prod_appengine_mvms():
    return os.environ.get("GAE_VM", False) == "true"
site-packages/pip/_vendor/urllib3/contrib/ntlmpool.py000064400000010151151464452710016751 0ustar00"""
NTLM authenticating pool, contributed by erikcederstran

Issue #10, see: http://code.google.com/p/urllib3/issues/detail?id=10
"""
from __future__ import absolute_import

from logging import getLogger
from ntlm import ntlm

from .. import HTTPSConnectionPool
from ..packages.six.moves.http_client import HTTPSConnection


log = getLogger(__name__)


class NTLMConnectionPool(HTTPSConnectionPool):
    """
    Implements an NTLM authentication version of an urllib3 connection pool
    """

    scheme = "https"

    def __init__(self, user, pw, authurl, *args, **kwargs):
        """
        authurl is a random URL on the server that is protected by NTLM.
        user is the Windows user, probably in the DOMAIN\\username format.
        pw is the password for the user.
        """
        super(NTLMConnectionPool, self).__init__(*args, **kwargs)
        self.authurl = authurl
        self.rawuser = user
        user_parts = user.split("\\", 1)
        self.domain = user_parts[0].upper()
        self.user = user_parts[1]
        self.pw = pw

    def _new_conn(self):
        # Performs the NTLM handshake that secures the connection. The socket
        # must be kept open while requests are performed.
        self.num_connections += 1
        log.debug(
            "Starting NTLM HTTPS connection no. %d: https://%s%s",
            self.num_connections,
            self.host,
            self.authurl,
        )

        headers = {"Connection": "Keep-Alive"}
        req_header = "Authorization"
        resp_header = "www-authenticate"

        conn = HTTPSConnection(host=self.host, port=self.port)

        # Send negotiation message
        headers[req_header] = "NTLM %s" % ntlm.create_NTLM_NEGOTIATE_MESSAGE(
            self.rawuser
        )
        log.debug("Request headers: %s", headers)
        conn.request("GET", self.authurl, None, headers)
        res = conn.getresponse()
        reshdr = dict(res.getheaders())
        log.debug("Response status: %s %s", res.status, res.reason)
        log.debug("Response headers: %s", reshdr)
        log.debug("Response data: %s [...]", res.read(100))

        # Remove the reference to the socket, so that it can not be closed by
        # the response object (we want to keep the socket open)
        res.fp = None

        # Server should respond with a challenge message
        auth_header_values = reshdr[resp_header].split(", ")
        auth_header_value = None
        for s in auth_header_values:
            if s[:5] == "NTLM ":
                auth_header_value = s[5:]
        if auth_header_value is None:
            raise Exception(
                "Unexpected %s response header: %s" % (resp_header, reshdr[resp_header])
            )

        # Send authentication message
        ServerChallenge, NegotiateFlags = ntlm.parse_NTLM_CHALLENGE_MESSAGE(
            auth_header_value
        )
        auth_msg = ntlm.create_NTLM_AUTHENTICATE_MESSAGE(
            ServerChallenge, self.user, self.domain, self.pw, NegotiateFlags
        )
        headers[req_header] = "NTLM %s" % auth_msg
        log.debug("Request headers: %s", headers)
        conn.request("GET", self.authurl, None, headers)
        res = conn.getresponse()
        log.debug("Response status: %s %s", res.status, res.reason)
        log.debug("Response headers: %s", dict(res.getheaders()))
        log.debug("Response data: %s [...]", res.read()[:100])
        if res.status != 200:
            if res.status == 401:
                raise Exception(
                    "Server rejected request: wrong " "username or password"
                )
            raise Exception("Wrong server response: %s %s" % (res.status, res.reason))

        res.fp = None
        log.debug("Connection established")
        return conn

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        retries=3,
        redirect=True,
        assert_same_host=True,
    ):
        if headers is None:
            headers = {}
        headers["Connection"] = "Keep-Alive"
        return super(NTLMConnectionPool, self).urlopen(
            method, url, body, headers, retries, redirect, assert_same_host
        )
site-packages/pip/_vendor/urllib3/contrib/securetransport.py000064400000077746151464452710020377 0ustar00"""
SecureTranport support for urllib3 via ctypes.

This makes platform-native TLS available to urllib3 users on macOS without the
use of a compiler. This is an important feature because the Python Package
Index is moving to become a TLSv1.2-or-higher server, and the default OpenSSL
that ships with macOS is not capable of doing TLSv1.2. The only way to resolve
this is to give macOS users an alternative solution to the problem, and that
solution is to use SecureTransport.

We use ctypes here because this solution must not require a compiler. That's
because pip is not allowed to require a compiler either.

This is not intended to be a seriously long-term solution to this problem.
The hope is that PEP 543 will eventually solve this issue for us, at which
point we can retire this contrib module. But in the short term, we need to
solve the impending tire fire that is Python on Mac without this kind of
contrib module. So...here we are.

To use this module, simply import and inject it::

    import urllib3.contrib.securetransport
    urllib3.contrib.securetransport.inject_into_urllib3()

Happy TLSing!

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

    Copyright (c) 2015-2016 Will Bond 

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
"""
from __future__ import absolute_import

import contextlib
import ctypes
import errno
import os.path
import shutil
import socket
import ssl
import threading
import weakref

from .. import util
from ._securetransport.bindings import Security, SecurityConst, CoreFoundation
from ._securetransport.low_level import (
    _assert_no_error,
    _cert_array_from_pem,
    _temporary_keychain,
    _load_client_cert_chain,
)

try:  # Platform-specific: Python 2
    from socket import _fileobject
except ImportError:  # Platform-specific: Python 3
    _fileobject = None
    from ..packages.backports.makefile import backport_makefile

__all__ = ["inject_into_urllib3", "extract_from_urllib3"]

# SNI always works
HAS_SNI = True

orig_util_HAS_SNI = util.HAS_SNI
orig_util_SSLContext = util.ssl_.SSLContext

# This dictionary is used by the read callback to obtain a handle to the
# calling wrapped socket. This is a pretty silly approach, but for now it'll
# do. I feel like I should be able to smuggle a handle to the wrapped socket
# directly in the SSLConnectionRef, but for now this approach will work I
# guess.
#
# We need to lock around this structure for inserts, but we don't do it for
# reads/writes in the callbacks. The reasoning here goes as follows:
#
#    1. It is not possible to call into the callbacks before the dictionary is
#       populated, so once in the callback the id must be in the dictionary.
#    2. The callbacks don't mutate the dictionary, they only read from it, and
#       so cannot conflict with any of the insertions.
#
# This is good: if we had to lock in the callbacks we'd drastically slow down
# the performance of this code.
_connection_refs = weakref.WeakValueDictionary()
_connection_ref_lock = threading.Lock()

# Limit writes to 16kB. This is OpenSSL's limit, but we'll cargo-cult it over
# for no better reason than we need *a* limit, and this one is right there.
SSL_WRITE_BLOCKSIZE = 16384

# This is our equivalent of util.ssl_.DEFAULT_CIPHERS, but expanded out to
# individual cipher suites. We need to do this because this is how
# SecureTransport wants them.
CIPHER_SUITES = [
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,
    SecurityConst.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256,
    SecurityConst.TLS_DHE_RSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_DHE_RSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA,
    SecurityConst.TLS_DHE_RSA_WITH_AES_256_CBC_SHA256,
    SecurityConst.TLS_DHE_RSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_DHE_RSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_DHE_RSA_WITH_AES_128_CBC_SHA,
    SecurityConst.TLS_AES_256_GCM_SHA384,
    SecurityConst.TLS_AES_128_GCM_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_RSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_AES_128_CCM_8_SHA256,
    SecurityConst.TLS_AES_128_CCM_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_256_CBC_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_RSA_WITH_AES_128_CBC_SHA,
]

# Basically this is simple: for PROTOCOL_SSLv23 we turn it into a low of
# TLSv1 and a high of TLSv1.3. For everything else, we pin to that version.
# TLSv1 to 1.2 are supported on macOS 10.8+ and TLSv1.3 is macOS 10.13+
_protocol_to_min_max = {
    util.PROTOCOL_TLS: (
        SecurityConst.kTLSProtocol1,
        SecurityConst.kTLSProtocolMaxSupported,
    )
}

if hasattr(ssl, "PROTOCOL_SSLv2"):
    _protocol_to_min_max[ssl.PROTOCOL_SSLv2] = (
        SecurityConst.kSSLProtocol2,
        SecurityConst.kSSLProtocol2,
    )
if hasattr(ssl, "PROTOCOL_SSLv3"):
    _protocol_to_min_max[ssl.PROTOCOL_SSLv3] = (
        SecurityConst.kSSLProtocol3,
        SecurityConst.kSSLProtocol3,
    )
if hasattr(ssl, "PROTOCOL_TLSv1"):
    _protocol_to_min_max[ssl.PROTOCOL_TLSv1] = (
        SecurityConst.kTLSProtocol1,
        SecurityConst.kTLSProtocol1,
    )
if hasattr(ssl, "PROTOCOL_TLSv1_1"):
    _protocol_to_min_max[ssl.PROTOCOL_TLSv1_1] = (
        SecurityConst.kTLSProtocol11,
        SecurityConst.kTLSProtocol11,
    )
if hasattr(ssl, "PROTOCOL_TLSv1_2"):
    _protocol_to_min_max[ssl.PROTOCOL_TLSv1_2] = (
        SecurityConst.kTLSProtocol12,
        SecurityConst.kTLSProtocol12,
    )


def inject_into_urllib3():
    """
    Monkey-patch urllib3 with SecureTransport-backed SSL-support.
    """
    util.SSLContext = SecureTransportContext
    util.ssl_.SSLContext = SecureTransportContext
    util.HAS_SNI = HAS_SNI
    util.ssl_.HAS_SNI = HAS_SNI
    util.IS_SECURETRANSPORT = True
    util.ssl_.IS_SECURETRANSPORT = True


def extract_from_urllib3():
    """
    Undo monkey-patching by :func:`inject_into_urllib3`.
    """
    util.SSLContext = orig_util_SSLContext
    util.ssl_.SSLContext = orig_util_SSLContext
    util.HAS_SNI = orig_util_HAS_SNI
    util.ssl_.HAS_SNI = orig_util_HAS_SNI
    util.IS_SECURETRANSPORT = False
    util.ssl_.IS_SECURETRANSPORT = False


def _read_callback(connection_id, data_buffer, data_length_pointer):
    """
    SecureTransport read callback. This is called by ST to request that data
    be returned from the socket.
    """
    wrapped_socket = None
    try:
        wrapped_socket = _connection_refs.get(connection_id)
        if wrapped_socket is None:
            return SecurityConst.errSSLInternal
        base_socket = wrapped_socket.socket

        requested_length = data_length_pointer[0]

        timeout = wrapped_socket.gettimeout()
        error = None
        read_count = 0

        try:
            while read_count < requested_length:
                if timeout is None or timeout >= 0:
                    if not util.wait_for_read(base_socket, timeout):
                        raise socket.error(errno.EAGAIN, "timed out")

                remaining = requested_length - read_count
                buffer = (ctypes.c_char * remaining).from_address(
                    data_buffer + read_count
                )
                chunk_size = base_socket.recv_into(buffer, remaining)
                read_count += chunk_size
                if not chunk_size:
                    if not read_count:
                        return SecurityConst.errSSLClosedGraceful
                    break
        except (socket.error) as e:
            error = e.errno

            if error is not None and error != errno.EAGAIN:
                data_length_pointer[0] = read_count
                if error == errno.ECONNRESET or error == errno.EPIPE:
                    return SecurityConst.errSSLClosedAbort
                raise

        data_length_pointer[0] = read_count

        if read_count != requested_length:
            return SecurityConst.errSSLWouldBlock

        return 0
    except Exception as e:
        if wrapped_socket is not None:
            wrapped_socket._exception = e
        return SecurityConst.errSSLInternal


def _write_callback(connection_id, data_buffer, data_length_pointer):
    """
    SecureTransport write callback. This is called by ST to request that data
    actually be sent on the network.
    """
    wrapped_socket = None
    try:
        wrapped_socket = _connection_refs.get(connection_id)
        if wrapped_socket is None:
            return SecurityConst.errSSLInternal
        base_socket = wrapped_socket.socket

        bytes_to_write = data_length_pointer[0]
        data = ctypes.string_at(data_buffer, bytes_to_write)

        timeout = wrapped_socket.gettimeout()
        error = None
        sent = 0

        try:
            while sent < bytes_to_write:
                if timeout is None or timeout >= 0:
                    if not util.wait_for_write(base_socket, timeout):
                        raise socket.error(errno.EAGAIN, "timed out")
                chunk_sent = base_socket.send(data)
                sent += chunk_sent

                # This has some needless copying here, but I'm not sure there's
                # much value in optimising this data path.
                data = data[chunk_sent:]
        except (socket.error) as e:
            error = e.errno

            if error is not None and error != errno.EAGAIN:
                data_length_pointer[0] = sent
                if error == errno.ECONNRESET or error == errno.EPIPE:
                    return SecurityConst.errSSLClosedAbort
                raise

        data_length_pointer[0] = sent

        if sent != bytes_to_write:
            return SecurityConst.errSSLWouldBlock

        return 0
    except Exception as e:
        if wrapped_socket is not None:
            wrapped_socket._exception = e
        return SecurityConst.errSSLInternal


# We need to keep these two objects references alive: if they get GC'd while
# in use then SecureTransport could attempt to call a function that is in freed
# memory. That would be...uh...bad. Yeah, that's the word. Bad.
_read_callback_pointer = Security.SSLReadFunc(_read_callback)
_write_callback_pointer = Security.SSLWriteFunc(_write_callback)


class WrappedSocket(object):
    """
    API-compatibility wrapper for Python's OpenSSL wrapped socket object.

    Note: _makefile_refs, _drop(), and _reuse() are needed for the garbage
    collector of PyPy.
    """

    def __init__(self, socket):
        self.socket = socket
        self.context = None
        self._makefile_refs = 0
        self._closed = False
        self._exception = None
        self._keychain = None
        self._keychain_dir = None
        self._client_cert_chain = None

        # We save off the previously-configured timeout and then set it to
        # zero. This is done because we use select and friends to handle the
        # timeouts, but if we leave the timeout set on the lower socket then
        # Python will "kindly" call select on that socket again for us. Avoid
        # that by forcing the timeout to zero.
        self._timeout = self.socket.gettimeout()
        self.socket.settimeout(0)

    @contextlib.contextmanager
    def _raise_on_error(self):
        """
        A context manager that can be used to wrap calls that do I/O from
        SecureTransport. If any of the I/O callbacks hit an exception, this
        context manager will correctly propagate the exception after the fact.
        This avoids silently swallowing those exceptions.

        It also correctly forces the socket closed.
        """
        self._exception = None

        # We explicitly don't catch around this yield because in the unlikely
        # event that an exception was hit in the block we don't want to swallow
        # it.
        yield
        if self._exception is not None:
            exception, self._exception = self._exception, None
            self.close()
            raise exception

    def _set_ciphers(self):
        """
        Sets up the allowed ciphers. By default this matches the set in
        util.ssl_.DEFAULT_CIPHERS, at least as supported by macOS. This is done
        custom and doesn't allow changing at this time, mostly because parsing
        OpenSSL cipher strings is going to be a freaking nightmare.
        """
        ciphers = (Security.SSLCipherSuite * len(CIPHER_SUITES))(*CIPHER_SUITES)
        result = Security.SSLSetEnabledCiphers(
            self.context, ciphers, len(CIPHER_SUITES)
        )
        _assert_no_error(result)

    def _custom_validate(self, verify, trust_bundle):
        """
        Called when we have set custom validation. We do this in two cases:
        first, when cert validation is entirely disabled; and second, when
        using a custom trust DB.
        """
        # If we disabled cert validation, just say: cool.
        if not verify:
            return

        # We want data in memory, so load it up.
        if os.path.isfile(trust_bundle):
            with open(trust_bundle, "rb") as f:
                trust_bundle = f.read()

        cert_array = None
        trust = Security.SecTrustRef()

        try:
            # Get a CFArray that contains the certs we want.
            cert_array = _cert_array_from_pem(trust_bundle)

            # Ok, now the hard part. We want to get the SecTrustRef that ST has
            # created for this connection, shove our CAs into it, tell ST to
            # ignore everything else it knows, and then ask if it can build a
            # chain. This is a buuuunch of code.
            result = Security.SSLCopyPeerTrust(self.context, ctypes.byref(trust))
            _assert_no_error(result)
            if not trust:
                raise ssl.SSLError("Failed to copy trust reference")

            result = Security.SecTrustSetAnchorCertificates(trust, cert_array)
            _assert_no_error(result)

            result = Security.SecTrustSetAnchorCertificatesOnly(trust, True)
            _assert_no_error(result)

            trust_result = Security.SecTrustResultType()
            result = Security.SecTrustEvaluate(trust, ctypes.byref(trust_result))
            _assert_no_error(result)
        finally:
            if trust:
                CoreFoundation.CFRelease(trust)

            if cert_array is not None:
                CoreFoundation.CFRelease(cert_array)

        # Ok, now we can look at what the result was.
        successes = (
            SecurityConst.kSecTrustResultUnspecified,
            SecurityConst.kSecTrustResultProceed,
        )
        if trust_result.value not in successes:
            raise ssl.SSLError(
                "certificate verify failed, error code: %d" % trust_result.value
            )

    def handshake(
        self,
        server_hostname,
        verify,
        trust_bundle,
        min_version,
        max_version,
        client_cert,
        client_key,
        client_key_passphrase,
    ):
        """
        Actually performs the TLS handshake. This is run automatically by
        wrapped socket, and shouldn't be needed in user code.
        """
        # First, we do the initial bits of connection setup. We need to create
        # a context, set its I/O funcs, and set the connection reference.
        self.context = Security.SSLCreateContext(
            None, SecurityConst.kSSLClientSide, SecurityConst.kSSLStreamType
        )
        result = Security.SSLSetIOFuncs(
            self.context, _read_callback_pointer, _write_callback_pointer
        )
        _assert_no_error(result)

        # Here we need to compute the handle to use. We do this by taking the
        # id of self modulo 2**31 - 1. If this is already in the dictionary, we
        # just keep incrementing by one until we find a free space.
        with _connection_ref_lock:
            handle = id(self) % 2147483647
            while handle in _connection_refs:
                handle = (handle + 1) % 2147483647
            _connection_refs[handle] = self

        result = Security.SSLSetConnection(self.context, handle)
        _assert_no_error(result)

        # If we have a server hostname, we should set that too.
        if server_hostname:
            if not isinstance(server_hostname, bytes):
                server_hostname = server_hostname.encode("utf-8")

            result = Security.SSLSetPeerDomainName(
                self.context, server_hostname, len(server_hostname)
            )
            _assert_no_error(result)

        # Setup the ciphers.
        self._set_ciphers()

        # Set the minimum and maximum TLS versions.
        result = Security.SSLSetProtocolVersionMin(self.context, min_version)
        _assert_no_error(result)

        # TLS 1.3 isn't necessarily enabled by the OS
        # so we have to detect when we error out and try
        # setting TLS 1.3 if it's allowed. kTLSProtocolMaxSupported
        # was added in macOS 10.13 along with kTLSProtocol13.
        result = Security.SSLSetProtocolVersionMax(self.context, max_version)
        if result != 0 and max_version == SecurityConst.kTLSProtocolMaxSupported:
            result = Security.SSLSetProtocolVersionMax(
                self.context, SecurityConst.kTLSProtocol12
            )
        _assert_no_error(result)

        # If there's a trust DB, we need to use it. We do that by telling
        # SecureTransport to break on server auth. We also do that if we don't
        # want to validate the certs at all: we just won't actually do any
        # authing in that case.
        if not verify or trust_bundle is not None:
            result = Security.SSLSetSessionOption(
                self.context, SecurityConst.kSSLSessionOptionBreakOnServerAuth, True
            )
            _assert_no_error(result)

        # If there's a client cert, we need to use it.
        if client_cert:
            self._keychain, self._keychain_dir = _temporary_keychain()
            self._client_cert_chain = _load_client_cert_chain(
                self._keychain, client_cert, client_key
            )
            result = Security.SSLSetCertificate(self.context, self._client_cert_chain)
            _assert_no_error(result)

        while True:
            with self._raise_on_error():
                result = Security.SSLHandshake(self.context)

                if result == SecurityConst.errSSLWouldBlock:
                    raise socket.timeout("handshake timed out")
                elif result == SecurityConst.errSSLServerAuthCompleted:
                    self._custom_validate(verify, trust_bundle)
                    continue
                else:
                    _assert_no_error(result)
                    break

    def fileno(self):
        return self.socket.fileno()

    # Copy-pasted from Python 3.5 source code
    def _decref_socketios(self):
        if self._makefile_refs > 0:
            self._makefile_refs -= 1
        if self._closed:
            self.close()

    def recv(self, bufsiz):
        buffer = ctypes.create_string_buffer(bufsiz)
        bytes_read = self.recv_into(buffer, bufsiz)
        data = buffer[:bytes_read]
        return data

    def recv_into(self, buffer, nbytes=None):
        # Read short on EOF.
        if self._closed:
            return 0

        if nbytes is None:
            nbytes = len(buffer)

        buffer = (ctypes.c_char * nbytes).from_buffer(buffer)
        processed_bytes = ctypes.c_size_t(0)

        with self._raise_on_error():
            result = Security.SSLRead(
                self.context, buffer, nbytes, ctypes.byref(processed_bytes)
            )

        # There are some result codes that we want to treat as "not always
        # errors". Specifically, those are errSSLWouldBlock,
        # errSSLClosedGraceful, and errSSLClosedNoNotify.
        if result == SecurityConst.errSSLWouldBlock:
            # If we didn't process any bytes, then this was just a time out.
            # However, we can get errSSLWouldBlock in situations when we *did*
            # read some data, and in those cases we should just read "short"
            # and return.
            if processed_bytes.value == 0:
                # Timed out, no data read.
                raise socket.timeout("recv timed out")
        elif result in (
            SecurityConst.errSSLClosedGraceful,
            SecurityConst.errSSLClosedNoNotify,
        ):
            # The remote peer has closed this connection. We should do so as
            # well. Note that we don't actually return here because in
            # principle this could actually be fired along with return data.
            # It's unlikely though.
            self.close()
        else:
            _assert_no_error(result)

        # Ok, we read and probably succeeded. We should return whatever data
        # was actually read.
        return processed_bytes.value

    def settimeout(self, timeout):
        self._timeout = timeout

    def gettimeout(self):
        return self._timeout

    def send(self, data):
        processed_bytes = ctypes.c_size_t(0)

        with self._raise_on_error():
            result = Security.SSLWrite(
                self.context, data, len(data), ctypes.byref(processed_bytes)
            )

        if result == SecurityConst.errSSLWouldBlock and processed_bytes.value == 0:
            # Timed out
            raise socket.timeout("send timed out")
        else:
            _assert_no_error(result)

        # We sent, and probably succeeded. Tell them how much we sent.
        return processed_bytes.value

    def sendall(self, data):
        total_sent = 0
        while total_sent < len(data):
            sent = self.send(data[total_sent : total_sent + SSL_WRITE_BLOCKSIZE])
            total_sent += sent

    def shutdown(self):
        with self._raise_on_error():
            Security.SSLClose(self.context)

    def close(self):
        # TODO: should I do clean shutdown here? Do I have to?
        if self._makefile_refs < 1:
            self._closed = True
            if self.context:
                CoreFoundation.CFRelease(self.context)
                self.context = None
            if self._client_cert_chain:
                CoreFoundation.CFRelease(self._client_cert_chain)
                self._client_cert_chain = None
            if self._keychain:
                Security.SecKeychainDelete(self._keychain)
                CoreFoundation.CFRelease(self._keychain)
                shutil.rmtree(self._keychain_dir)
                self._keychain = self._keychain_dir = None
            return self.socket.close()
        else:
            self._makefile_refs -= 1

    def getpeercert(self, binary_form=False):
        # Urgh, annoying.
        #
        # Here's how we do this:
        #
        # 1. Call SSLCopyPeerTrust to get hold of the trust object for this
        #    connection.
        # 2. Call SecTrustGetCertificateAtIndex for index 0 to get the leaf.
        # 3. To get the CN, call SecCertificateCopyCommonName and process that
        #    string so that it's of the appropriate type.
        # 4. To get the SAN, we need to do something a bit more complex:
        #    a. Call SecCertificateCopyValues to get the data, requesting
        #       kSecOIDSubjectAltName.
        #    b. Mess about with this dictionary to try to get the SANs out.
        #
        # This is gross. Really gross. It's going to be a few hundred LoC extra
        # just to repeat something that SecureTransport can *already do*. So my
        # operating assumption at this time is that what we want to do is
        # instead to just flag to urllib3 that it shouldn't do its own hostname
        # validation when using SecureTransport.
        if not binary_form:
            raise ValueError("SecureTransport only supports dumping binary certs")
        trust = Security.SecTrustRef()
        certdata = None
        der_bytes = None

        try:
            # Grab the trust store.
            result = Security.SSLCopyPeerTrust(self.context, ctypes.byref(trust))
            _assert_no_error(result)
            if not trust:
                # Probably we haven't done the handshake yet. No biggie.
                return None

            cert_count = Security.SecTrustGetCertificateCount(trust)
            if not cert_count:
                # Also a case that might happen if we haven't handshaked.
                # Handshook? Handshaken?
                return None

            leaf = Security.SecTrustGetCertificateAtIndex(trust, 0)
            assert leaf

            # Ok, now we want the DER bytes.
            certdata = Security.SecCertificateCopyData(leaf)
            assert certdata

            data_length = CoreFoundation.CFDataGetLength(certdata)
            data_buffer = CoreFoundation.CFDataGetBytePtr(certdata)
            der_bytes = ctypes.string_at(data_buffer, data_length)
        finally:
            if certdata:
                CoreFoundation.CFRelease(certdata)
            if trust:
                CoreFoundation.CFRelease(trust)

        return der_bytes

    def version(self):
        protocol = Security.SSLProtocol()
        result = Security.SSLGetNegotiatedProtocolVersion(
            self.context, ctypes.byref(protocol)
        )
        _assert_no_error(result)
        if protocol.value == SecurityConst.kTLSProtocol13:
            return "TLSv1.3"
        elif protocol.value == SecurityConst.kTLSProtocol12:
            return "TLSv1.2"
        elif protocol.value == SecurityConst.kTLSProtocol11:
            return "TLSv1.1"
        elif protocol.value == SecurityConst.kTLSProtocol1:
            return "TLSv1"
        elif protocol.value == SecurityConst.kSSLProtocol3:
            return "SSLv3"
        elif protocol.value == SecurityConst.kSSLProtocol2:
            return "SSLv2"
        else:
            raise ssl.SSLError("Unknown TLS version: %r" % protocol)

    def _reuse(self):
        self._makefile_refs += 1

    def _drop(self):
        if self._makefile_refs < 1:
            self.close()
        else:
            self._makefile_refs -= 1


if _fileobject:  # Platform-specific: Python 2

    def makefile(self, mode, bufsize=-1):
        self._makefile_refs += 1
        return _fileobject(self, mode, bufsize, close=True)


else:  # Platform-specific: Python 3

    def makefile(self, mode="r", buffering=None, *args, **kwargs):
        # We disable buffering with SecureTransport because it conflicts with
        # the buffering that ST does internally (see issue #1153 for more).
        buffering = 0
        return backport_makefile(self, mode, buffering, *args, **kwargs)


WrappedSocket.makefile = makefile


class SecureTransportContext(object):
    """
    I am a wrapper class for the SecureTransport library, to translate the
    interface of the standard library ``SSLContext`` object to calls into
    SecureTransport.
    """

    def __init__(self, protocol):
        self._min_version, self._max_version = _protocol_to_min_max[protocol]
        self._options = 0
        self._verify = False
        self._trust_bundle = None
        self._client_cert = None
        self._client_key = None
        self._client_key_passphrase = None

    @property
    def check_hostname(self):
        """
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        """
        return True

    @check_hostname.setter
    def check_hostname(self, value):
        """
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        """
        pass

    @property
    def options(self):
        # TODO: Well, crap.
        #
        # So this is the bit of the code that is the most likely to cause us
        # trouble. Essentially we need to enumerate all of the SSL options that
        # users might want to use and try to see if we can sensibly translate
        # them, or whether we should just ignore them.
        return self._options

    @options.setter
    def options(self, value):
        # TODO: Update in line with above.
        self._options = value

    @property
    def verify_mode(self):
        return ssl.CERT_REQUIRED if self._verify else ssl.CERT_NONE

    @verify_mode.setter
    def verify_mode(self, value):
        self._verify = True if value == ssl.CERT_REQUIRED else False

    def set_default_verify_paths(self):
        # So, this has to do something a bit weird. Specifically, what it does
        # is nothing.
        #
        # This means that, if we had previously had load_verify_locations
        # called, this does not undo that. We need to do that because it turns
        # out that the rest of the urllib3 code will attempt to load the
        # default verify paths if it hasn't been told about any paths, even if
        # the context itself was sometime earlier. We resolve that by just
        # ignoring it.
        pass

    def load_default_certs(self):
        return self.set_default_verify_paths()

    def set_ciphers(self, ciphers):
        # For now, we just require the default cipher string.
        if ciphers != util.ssl_.DEFAULT_CIPHERS:
            raise ValueError("SecureTransport doesn't support custom cipher strings")

    def load_verify_locations(self, cafile=None, capath=None, cadata=None):
        # OK, we only really support cadata and cafile.
        if capath is not None:
            raise ValueError("SecureTransport does not support cert directories")

        self._trust_bundle = cafile or cadata

    def load_cert_chain(self, certfile, keyfile=None, password=None):
        self._client_cert = certfile
        self._client_key = keyfile
        self._client_cert_passphrase = password

    def wrap_socket(
        self,
        sock,
        server_side=False,
        do_handshake_on_connect=True,
        suppress_ragged_eofs=True,
        server_hostname=None,
    ):
        # So, what do we do here? Firstly, we assert some properties. This is a
        # stripped down shim, so there is some functionality we don't support.
        # See PEP 543 for the real deal.
        assert not server_side
        assert do_handshake_on_connect
        assert suppress_ragged_eofs

        # Ok, we're good to go. Now we want to create the wrapped socket object
        # and store it in the appropriate place.
        wrapped_socket = WrappedSocket(sock)

        # Now we can handshake
        wrapped_socket.handshake(
            server_hostname,
            self._verify,
            self._trust_bundle,
            self._min_version,
            self._max_version,
            self._client_cert,
            self._client_key,
            self._client_key_passphrase,
        )
        return wrapped_socket
site-packages/pip/_vendor/urllib3/contrib/pyopenssl.py000064400000040070151464452710017144 0ustar00"""
SSL with SNI_-support for Python 2. Follow these instructions if you would
like to verify SSL certificates in Python 2. Note, the default libraries do
*not* do certificate checking; you need to do additional work to validate
certificates yourself.

This needs the following packages installed:

* pyOpenSSL (tested with 16.0.0)
* cryptography (minimum 1.3.4, from pyopenssl)
* idna (minimum 2.0, from cryptography)

However, pyopenssl depends on cryptography, which depends on idna, so while we
use all three directly here we end up having relatively few packages required.

You can install them with the following command:

    pip install pyopenssl cryptography idna

To activate certificate checking, call
:func:`~urllib3.contrib.pyopenssl.inject_into_urllib3` from your Python code
before you begin making HTTP requests. This can be done in a ``sitecustomize``
module, or at any other time before your application begins using ``urllib3``,
like this::

    try:
        import urllib3.contrib.pyopenssl
        urllib3.contrib.pyopenssl.inject_into_urllib3()
    except ImportError:
        pass

Now you can use :mod:`urllib3` as you normally would, and it will support SNI
when the required modules are installed.

Activating this module also has the positive side effect of disabling SSL/TLS
compression in Python 2 (see `CRIME attack`_).

If you want to configure the default list of supported cipher suites, you can
set the ``urllib3.contrib.pyopenssl.DEFAULT_SSL_CIPHER_LIST`` variable.

.. _sni: https://en.wikipedia.org/wiki/Server_Name_Indication
.. _crime attack: https://en.wikipedia.org/wiki/CRIME_(security_exploit)
"""
from __future__ import absolute_import

import OpenSSL.SSL
from cryptography import x509
from cryptography.hazmat.backends.openssl import backend as openssl_backend
from cryptography.hazmat.backends.openssl.x509 import _Certificate

try:
    from cryptography.x509 import UnsupportedExtension
except ImportError:
    # UnsupportedExtension is gone in cryptography >= 2.1.0
    class UnsupportedExtension(Exception):
        pass


from socket import timeout, error as SocketError
from io import BytesIO

try:  # Platform-specific: Python 2
    from socket import _fileobject
except ImportError:  # Platform-specific: Python 3
    _fileobject = None
    from ..packages.backports.makefile import backport_makefile

import logging
import ssl
from ..packages import six
import sys

from .. import util


__all__ = ["inject_into_urllib3", "extract_from_urllib3"]

# SNI always works.
HAS_SNI = True

# Map from urllib3 to PyOpenSSL compatible parameter-values.
_openssl_versions = {
    util.PROTOCOL_TLS: OpenSSL.SSL.SSLv23_METHOD,
    ssl.PROTOCOL_TLSv1: OpenSSL.SSL.TLSv1_METHOD,
}

if hasattr(ssl, "PROTOCOL_SSLv3") and hasattr(OpenSSL.SSL, "SSLv3_METHOD"):
    _openssl_versions[ssl.PROTOCOL_SSLv3] = OpenSSL.SSL.SSLv3_METHOD

if hasattr(ssl, "PROTOCOL_TLSv1_1") and hasattr(OpenSSL.SSL, "TLSv1_1_METHOD"):
    _openssl_versions[ssl.PROTOCOL_TLSv1_1] = OpenSSL.SSL.TLSv1_1_METHOD

if hasattr(ssl, "PROTOCOL_TLSv1_2") and hasattr(OpenSSL.SSL, "TLSv1_2_METHOD"):
    _openssl_versions[ssl.PROTOCOL_TLSv1_2] = OpenSSL.SSL.TLSv1_2_METHOD


_stdlib_to_openssl_verify = {
    ssl.CERT_NONE: OpenSSL.SSL.VERIFY_NONE,
    ssl.CERT_OPTIONAL: OpenSSL.SSL.VERIFY_PEER,
    ssl.CERT_REQUIRED: OpenSSL.SSL.VERIFY_PEER
    + OpenSSL.SSL.VERIFY_FAIL_IF_NO_PEER_CERT,
}
_openssl_to_stdlib_verify = dict((v, k) for k, v in _stdlib_to_openssl_verify.items())

# OpenSSL will only write 16K at a time
SSL_WRITE_BLOCKSIZE = 16384

orig_util_HAS_SNI = util.HAS_SNI
orig_util_SSLContext = util.ssl_.SSLContext


log = logging.getLogger(__name__)


def inject_into_urllib3():
    "Monkey-patch urllib3 with PyOpenSSL-backed SSL-support."

    _validate_dependencies_met()

    util.SSLContext = PyOpenSSLContext
    util.ssl_.SSLContext = PyOpenSSLContext
    util.HAS_SNI = HAS_SNI
    util.ssl_.HAS_SNI = HAS_SNI
    util.IS_PYOPENSSL = True
    util.ssl_.IS_PYOPENSSL = True


def extract_from_urllib3():
    "Undo monkey-patching by :func:`inject_into_urllib3`."

    util.SSLContext = orig_util_SSLContext
    util.ssl_.SSLContext = orig_util_SSLContext
    util.HAS_SNI = orig_util_HAS_SNI
    util.ssl_.HAS_SNI = orig_util_HAS_SNI
    util.IS_PYOPENSSL = False
    util.ssl_.IS_PYOPENSSL = False


def _validate_dependencies_met():
    """
    Verifies that PyOpenSSL's package-level dependencies have been met.
    Throws `ImportError` if they are not met.
    """
    # Method added in `cryptography==1.1`; not available in older versions
    from cryptography.x509.extensions import Extensions

    if getattr(Extensions, "get_extension_for_class", None) is None:
        raise ImportError(
            "'cryptography' module missing required functionality.  "
            "Try upgrading to v1.3.4 or newer."
        )

    # pyOpenSSL 0.14 and above use cryptography for OpenSSL bindings. The _x509
    # attribute is only present on those versions.
    from OpenSSL.crypto import X509

    x509 = X509()
    if getattr(x509, "_x509", None) is None:
        raise ImportError(
            "'pyOpenSSL' module missing required functionality. "
            "Try upgrading to v0.14 or newer."
        )


def _dnsname_to_stdlib(name):
    """
    Converts a dNSName SubjectAlternativeName field to the form used by the
    standard library on the given Python version.

    Cryptography produces a dNSName as a unicode string that was idna-decoded
    from ASCII bytes. We need to idna-encode that string to get it back, and
    then on Python 3 we also need to convert to unicode via UTF-8 (the stdlib
    uses PyUnicode_FromStringAndSize on it, which decodes via UTF-8).

    If the name cannot be idna-encoded then we return None signalling that
    the name given should be skipped.
    """

    def idna_encode(name):
        """
        Borrowed wholesale from the Python Cryptography Project. It turns out
        that we can't just safely call `idna.encode`: it can explode for
        wildcard names. This avoids that problem.
        """
        from pip._vendor import idna

        try:
            for prefix in [u"*.", u"."]:
                if name.startswith(prefix):
                    name = name[len(prefix) :]
                    return prefix.encode("ascii") + idna.encode(name)
            return idna.encode(name)
        except idna.core.IDNAError:
            return None

    # Don't send IPv6 addresses through the IDNA encoder.
    if ":" in name:
        return name

    name = idna_encode(name)
    if name is None:
        return None
    elif sys.version_info >= (3, 0):
        name = name.decode("utf-8")
    return name


def get_subj_alt_name(peer_cert):
    """
    Given an PyOpenSSL certificate, provides all the subject alternative names.
    """
    # Pass the cert to cryptography, which has much better APIs for this.
    if hasattr(peer_cert, "to_cryptography"):
        cert = peer_cert.to_cryptography()
    else:
        # This is technically using private APIs, but should work across all
        # relevant versions before PyOpenSSL got a proper API for this.
        cert = _Certificate(openssl_backend, peer_cert._x509)

    # We want to find the SAN extension. Ask Cryptography to locate it (it's
    # faster than looping in Python)
    try:
        ext = cert.extensions.get_extension_for_class(x509.SubjectAlternativeName).value
    except x509.ExtensionNotFound:
        # No such extension, return the empty list.
        return []
    except (
        x509.DuplicateExtension,
        UnsupportedExtension,
        x509.UnsupportedGeneralNameType,
        UnicodeError,
    ) as e:
        # A problem has been found with the quality of the certificate. Assume
        # no SAN field is present.
        log.warning(
            "A problem was encountered with the certificate that prevented "
            "urllib3 from finding the SubjectAlternativeName field. This can "
            "affect certificate validation. The error was %s",
            e,
        )
        return []

    # We want to return dNSName and iPAddress fields. We need to cast the IPs
    # back to strings because the match_hostname function wants them as
    # strings.
    # Sadly the DNS names need to be idna encoded and then, on Python 3, UTF-8
    # decoded. This is pretty frustrating, but that's what the standard library
    # does with certificates, and so we need to attempt to do the same.
    # We also want to skip over names which cannot be idna encoded.
    names = [
        ("DNS", name)
        for name in map(_dnsname_to_stdlib, ext.get_values_for_type(x509.DNSName))
        if name is not None
    ]
    names.extend(
        ("IP Address", str(name)) for name in ext.get_values_for_type(x509.IPAddress)
    )

    return names


class WrappedSocket(object):
    """API-compatibility wrapper for Python OpenSSL's Connection-class.

    Note: _makefile_refs, _drop() and _reuse() are needed for the garbage
    collector of pypy.
    """

    def __init__(self, connection, socket, suppress_ragged_eofs=True):
        self.connection = connection
        self.socket = socket
        self.suppress_ragged_eofs = suppress_ragged_eofs
        self._makefile_refs = 0
        self._closed = False

    def fileno(self):
        return self.socket.fileno()

    # Copy-pasted from Python 3.5 source code
    def _decref_socketios(self):
        if self._makefile_refs > 0:
            self._makefile_refs -= 1
        if self._closed:
            self.close()

    def recv(self, *args, **kwargs):
        try:
            data = self.connection.recv(*args, **kwargs)
        except OpenSSL.SSL.SysCallError as e:
            if self.suppress_ragged_eofs and e.args == (-1, "Unexpected EOF"):
                return b""
            else:
                raise SocketError(str(e))
        except OpenSSL.SSL.ZeroReturnError:
            if self.connection.get_shutdown() == OpenSSL.SSL.RECEIVED_SHUTDOWN:
                return b""
            else:
                raise
        except OpenSSL.SSL.WantReadError:
            if not util.wait_for_read(self.socket, self.socket.gettimeout()):
                raise timeout("The read operation timed out")
            else:
                return self.recv(*args, **kwargs)

        # TLS 1.3 post-handshake authentication
        except OpenSSL.SSL.Error as e:
            raise ssl.SSLError("read error: %r" % e)
        else:
            return data

    def recv_into(self, *args, **kwargs):
        try:
            return self.connection.recv_into(*args, **kwargs)
        except OpenSSL.SSL.SysCallError as e:
            if self.suppress_ragged_eofs and e.args == (-1, "Unexpected EOF"):
                return 0
            else:
                raise SocketError(str(e))
        except OpenSSL.SSL.ZeroReturnError:
            if self.connection.get_shutdown() == OpenSSL.SSL.RECEIVED_SHUTDOWN:
                return 0
            else:
                raise
        except OpenSSL.SSL.WantReadError:
            if not util.wait_for_read(self.socket, self.socket.gettimeout()):
                raise timeout("The read operation timed out")
            else:
                return self.recv_into(*args, **kwargs)

        # TLS 1.3 post-handshake authentication
        except OpenSSL.SSL.Error as e:
            raise ssl.SSLError("read error: %r" % e)

    def settimeout(self, timeout):
        return self.socket.settimeout(timeout)

    def _send_until_done(self, data):
        while True:
            try:
                return self.connection.send(data)
            except OpenSSL.SSL.WantWriteError:
                if not util.wait_for_write(self.socket, self.socket.gettimeout()):
                    raise timeout()
                continue
            except OpenSSL.SSL.SysCallError as e:
                raise SocketError(str(e))

    def sendall(self, data):
        total_sent = 0
        while total_sent < len(data):
            sent = self._send_until_done(
                data[total_sent : total_sent + SSL_WRITE_BLOCKSIZE]
            )
            total_sent += sent

    def shutdown(self):
        # FIXME rethrow compatible exceptions should we ever use this
        self.connection.shutdown()

    def close(self):
        if self._makefile_refs < 1:
            try:
                self._closed = True
                return self.connection.close()
            except OpenSSL.SSL.Error:
                return
        else:
            self._makefile_refs -= 1

    def getpeercert(self, binary_form=False):
        x509 = self.connection.get_peer_certificate()

        if not x509:
            return x509

        if binary_form:
            return OpenSSL.crypto.dump_certificate(OpenSSL.crypto.FILETYPE_ASN1, x509)

        return {
            "subject": ((("commonName", x509.get_subject().CN),),),
            "subjectAltName": get_subj_alt_name(x509),
        }

    def version(self):
        return self.connection.get_protocol_version_name()

    def _reuse(self):
        self._makefile_refs += 1

    def _drop(self):
        if self._makefile_refs < 1:
            self.close()
        else:
            self._makefile_refs -= 1


if _fileobject:  # Platform-specific: Python 2

    def makefile(self, mode, bufsize=-1):
        self._makefile_refs += 1
        return _fileobject(self, mode, bufsize, close=True)


else:  # Platform-specific: Python 3
    makefile = backport_makefile

WrappedSocket.makefile = makefile


class PyOpenSSLContext(object):
    """
    I am a wrapper class for the PyOpenSSL ``Context`` object. I am responsible
    for translating the interface of the standard library ``SSLContext`` object
    to calls into PyOpenSSL.
    """

    def __init__(self, protocol):
        self.protocol = _openssl_versions[protocol]
        self._ctx = OpenSSL.SSL.Context(self.protocol)
        self._options = 0
        self.check_hostname = False

    @property
    def options(self):
        return self._options

    @options.setter
    def options(self, value):
        self._options = value
        self._ctx.set_options(value)

    @property
    def verify_mode(self):
        return _openssl_to_stdlib_verify[self._ctx.get_verify_mode()]

    @verify_mode.setter
    def verify_mode(self, value):
        self._ctx.set_verify(_stdlib_to_openssl_verify[value], _verify_callback)

    def set_default_verify_paths(self):
        self._ctx.set_default_verify_paths()

    def set_ciphers(self, ciphers):
        if isinstance(ciphers, six.text_type):
            ciphers = ciphers.encode("utf-8")
        self._ctx.set_cipher_list(ciphers)

    def load_verify_locations(self, cafile=None, capath=None, cadata=None):
        if cafile is not None:
            cafile = cafile.encode("utf-8")
        if capath is not None:
            capath = capath.encode("utf-8")
        self._ctx.load_verify_locations(cafile, capath)
        if cadata is not None:
            self._ctx.load_verify_locations(BytesIO(cadata))

    def load_cert_chain(self, certfile, keyfile=None, password=None):
        self._ctx.use_certificate_chain_file(certfile)
        if password is not None:
            if not isinstance(password, six.binary_type):
                password = password.encode("utf-8")
            self._ctx.set_passwd_cb(lambda *_: password)
        self._ctx.use_privatekey_file(keyfile or certfile)

    def wrap_socket(
        self,
        sock,
        server_side=False,
        do_handshake_on_connect=True,
        suppress_ragged_eofs=True,
        server_hostname=None,
    ):
        cnx = OpenSSL.SSL.Connection(self._ctx, sock)

        if isinstance(server_hostname, six.text_type):  # Platform-specific: Python 3
            server_hostname = server_hostname.encode("utf-8")

        if server_hostname is not None:
            cnx.set_tlsext_host_name(server_hostname)

        cnx.set_connect_state()

        while True:
            try:
                cnx.do_handshake()
            except OpenSSL.SSL.WantReadError:
                if not util.wait_for_read(sock, sock.gettimeout()):
                    raise timeout("select timed out")
                continue
            except OpenSSL.SSL.Error as e:
                raise ssl.SSLError("bad handshake: %r" % e)
            break

        return WrappedSocket(cnx, sock)


def _verify_callback(cnx, x509, err_no, err_depth, return_code):
    return err_no == 0
site-packages/pip/_vendor/urllib3/contrib/__pycache__/securetransport.cpython-38.opt-1.pyc000064400000046476151464452710025620 0ustar00U

.e@sdZddlmZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
mZddlmZmZmZddlmZmZmZmZzdd	l	mZWn$ek
rdZdd
lmZYnXddgZd
ZejZejj Z!e"Z#e$Z%dZ&ej'ej(ej)ej*ej+ej,ej-ej.ej/ej0ej1ej2ej3ej4ej5ej6ej7ej8ej9ej:ej;ejej?ej@ejAejBejCejDgZEejFejGejHfiZIeJe
drejKejKfeIe
jL<eJe
drejMejMfeIe
jN<eJe
drejGejGfeIe
jO<eJe
drejPejPfeIe
jQ<eJe
drejRejRfeIe
jS<ddZTddZUddZVddZWeXeVZYeZeWZ[Gddde\Z]ernd#ddZ^n
d$d dZ^e^e]_^Gd!d"d"e\Z_dS)%a	
SecureTranport support for urllib3 via ctypes.

This makes platform-native TLS available to urllib3 users on macOS without the
use of a compiler. This is an important feature because the Python Package
Index is moving to become a TLSv1.2-or-higher server, and the default OpenSSL
that ships with macOS is not capable of doing TLSv1.2. The only way to resolve
this is to give macOS users an alternative solution to the problem, and that
solution is to use SecureTransport.

We use ctypes here because this solution must not require a compiler. That's
because pip is not allowed to require a compiler either.

This is not intended to be a seriously long-term solution to this problem.
The hope is that PEP 543 will eventually solve this issue for us, at which
point we can retire this contrib module. But in the short term, we need to
solve the impending tire fire that is Python on Mac without this kind of
contrib module. So...here we are.

To use this module, simply import and inject it::

    import urllib3.contrib.securetransport
    urllib3.contrib.securetransport.inject_into_urllib3()

Happy TLSing!

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

    Copyright (c) 2015-2016 Will Bond 

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
)absolute_importN)util)Security
SecurityConstCoreFoundation)_assert_no_error_cert_array_from_pem_temporary_keychain_load_client_cert_chain)_fileobjectbackport_makefileinject_into_urllib3extract_from_urllib3Ti@PROTOCOL_SSLv2PROTOCOL_SSLv3PROTOCOL_TLSv1PROTOCOL_TLSv1_1PROTOCOL_TLSv1_2cCs.tt_ttj_tt_ttj_dt_dtj_dS)zG
    Monkey-patch urllib3 with SecureTransport-backed SSL-support.
    TN)SecureTransportContextr
SSLContextssl_HAS_SNIIS_SECURETRANSPORTrrO/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/securetransport.pyrscCs.tt_ttj_tt_ttj_dt_dtj_dS)z>
    Undo monkey-patching by :func:`inject_into_urllib3`.
    FN)orig_util_SSLContextrrrorig_util_HAS_SNIrrrrrrrsc

Csd}zJt|}|dkr"tjWS|j}|d}|}d}d}z|||kr|dksZ|dkrtt||sttt	j
d||}	tj|	
||}
||
|	}||7}|sB|stjWWSqqBWnttjk
r0}zR|j	}|dk	r |t	j
kr ||d<|t	jks|t	jkrtjWY
WSW5d}~XYnX||d<||krLtjWSWdStk
r}z|dk	rt||_tjWYSd}~XYnXdS)zs
    SecureTransport read callback. This is called by ST to request that data
    be returned from the socket.
    Nr	timed out)_connection_refsgetrerrSSLInternalsocket
gettimeoutrZ
wait_for_readerrorerrnoEAGAINctypesc_charZfrom_address	recv_intoerrSSLClosedGraceful
ECONNRESETEPIPEerrSSLClosedAborterrSSLWouldBlock	Exception
_exception)

connection_iddata_bufferdata_length_pointerwrapped_socketbase_socketZrequested_lengthtimeoutr&Z
read_countZ	remainingbufferZ
chunk_sizeerrr_read_callbacksP




r;c
Csrd}z(t|}|dkr"tjWS|j}|d}t||}|}d}d}	zV|	|kr|dksf|dkrt	||st
tjd|
|}
|	|
7}	||
d}qNWnltj
k
r}zJ|j}|dk	r|tjkr|	|d<|tjks|tjkrtjWY
WSW5d}~XYnX|	|d<|	|kr*tjWSWdStk
rl}z|dk	rR||_tjWYSd}~XYnXdS)zx
    SecureTransport write callback. This is called by ST to request that data
    actually be sent on the network.
    Nrr )r!r"rr#r$r)	string_atr%rZwait_for_writer&r'r(sendr-r.r/r0r1r2)r3r4r5r6r7Zbytes_to_writedatar8r&sentZ
chunk_sentr:rrr_write_callbacksD



r@c@seZdZdZddZejddZddZdd	Z	d
dZ
dd
ZddZddZ
d*ddZddZddZddZddZddZdd Zd+d"d#Zd$d%Zd&d'Zd(d)ZdS),
WrappedSocketz
    API-compatibility wrapper for Python's OpenSSL wrapped socket object.

    Note: _makefile_refs, _drop(), and _reuse() are needed for the garbage
    collector of PyPy.
    cCsL||_d|_d|_d|_d|_d|_d|_d|_|j|_	|j
ddSNrF)r$context_makefile_refs_closedr2	_keychain
_keychain_dir_client_cert_chainr%_timeout
settimeout)selfr$rrr__init__HszWrappedSocket.__init__ccs4d|_dV|jdk	r0|jd}|_||dS)a]
        A context manager that can be used to wrap calls that do I/O from
        SecureTransport. If any of the I/O callbacks hit an exception, this
        context manager will correctly propagate the exception after the fact.
        This avoids silently swallowing those exceptions.

        It also correctly forces the socket closed.
        N)r2close)rKZ	exceptionrrr_raise_on_errorZs

zWrappedSocket._raise_on_errorcCs2tjttt}t|j|tt}t|dS)a4
        Sets up the allowed ciphers. By default this matches the set in
        util.ssl_.DEFAULT_CIPHERS, at least as supported by macOS. This is done
        custom and doesn't allow changing at this time, mostly because parsing
        OpenSSL cipher strings is going to be a freaking nightmare.
        N)rZSSLCipherSuitelen
CIPHER_SUITESZSSLSetEnabledCiphersrCr	)rKciphersresultrrr_set_ciphersoszWrappedSocket._set_ciphersc		Cs|sdStj|r2t|d}|}W5QRXd}t}zt	|}t
|jt
|}t||srtdt||}t|t|d}t|t}t|t
|}t|W5|rt||dk	rt|Xtjtjf}|j|kr
td|jdS)z
        Called when we have set custom validation. We do this in two cases:
        first, when cert validation is entirely disabled; and second, when
        using a custom trust DB.
        NrbzFailed to copy trust referenceTz)certificate verify failed, error code: %d)ospathisfileopenreadrSecTrustRefr	CFReleaser
SSLCopyPeerTrustrCr)byrefr	sslSSLErrorZSecTrustSetAnchorCertificatesZ!SecTrustSetAnchorCertificatesOnlyZSecTrustResultTypeZSecTrustEvaluaterZkSecTrustResultUnspecifiedZkSecTrustResultProceedvalue)	rKverifytrust_bundlefZ
cert_arraytrustrRZtrust_resultZ	successesrrr_custom_validate|s>

zWrappedSocket._custom_validatec		Cstdtjtj|_t|jtt}	t	|	t
0t|d}
|
tkrT|
dd}
q>|t|
<W5QRXt
|j|
}	t	|	|rt|ts|d}t|j|t|}	t	|	|t|j|}	t	|	t|j|}	|	dkr|tjkrt|jtj}	t	|	|r|dk	r0t|jtjd}	t	|	|rlt\|_|_t|j|||_t|j|j}	t	|	| ht!|j}	|	tj"krt#$dn>|	tj%kr|&||W5QRqlnt	|	W5QRqW5QRXqldS)z
        Actually performs the TLS handshake. This is run automatically by
        wrapped socket, and shouldn't be needed in user code.
        Nirzutf-8rTzhandshake timed out)'rZSSLCreateContextrZkSSLClientSideZkSSLStreamTyperCZ
SSLSetIOFuncs_read_callback_pointer_write_callback_pointerr	_connection_ref_lockidr!ZSSLSetConnection
isinstancebytesencodeZSSLSetPeerDomainNamerOrSZSSLSetProtocolVersionMinZSSLSetProtocolVersionMaxkTLSProtocolMaxSupportedkTLSProtocol12ZSSLSetSessionOptionZ"kSSLSessionOptionBreakOnServerAuthrrFrGrrHZSSLSetCertificaterNZSSLHandshaker0r$r8ZerrSSLServerAuthCompletedre)rKserver_hostnamerarbZmin_versionZmax_versionZclient_certZ
client_keyZclient_key_passphraserRZhandlerrr	handshakes|


zWrappedSocket.handshakecCs
|jSN)r$filenorKrrrrrszWrappedSocket.filenocCs*|jdkr|jd8_|jr&|dS)Nrr)rDrErMrsrrr_decref_socketioss
zWrappedSocket._decref_socketioscCs&t|}|||}|d|}|Srq)r)Zcreate_string_bufferr+)rKZbufsizr9Z
bytes_readr>rrrrecv s
zWrappedSocket.recvNc	Cs|jr
dS|dkrt|}tj||}td}|t|j	||t
|}W5QRX|tjkr|j
dkrtdn"|tjtjfkr|nt||j
S)Nrzrecv timed out)rErOr)r*Zfrom_bufferc_size_trNrZSSLReadrCr]rr0r`r$r8r,ZerrSSLClosedNoNotifyrMr	)rKr9nbytesprocessed_bytesrRrrrr+&s.




zWrappedSocket.recv_intocCs
||_dSrqrI)rKr8rrrrJQszWrappedSocket.settimeoutcCs|jSrqryrsrrrr%TszWrappedSocket.gettimeoutc	Cshtd}|"t|j|t|t|}W5QRX|tj	krZ|j
dkrZtdnt
||j
S)Nrzsend timed out)r)rvrNrZSSLWriterCrOr]rr0r`r$r8r	)rKr>rxrRrrrr=Ws

zWrappedSocket.sendcCs4d}|t|kr0||||t}||7}qdSNr)rOr=SSL_WRITE_BLOCKSIZE)rKr>Z
total_sentr?rrrsendallhszWrappedSocket.sendallc	Cs$|t|jW5QRXdSrq)rNrZSSLCloserCrsrrrshutdownns
zWrappedSocket.shutdowncCs|jdkrd|_|jr(t|jd|_|jr@t|jd|_|jrvt|jt|jt	
|jd|_|_|j
S|jd8_dS)NrT)rDrErCrr[rHrFrZSecKeychainDeleteshutilZrmtreerGr$rMrsrrrrMrs

zWrappedSocket.closeFc
	Cs|stdt}d}d}zxt|jt|}t	||sFWTdSt
|}|s\W>dSt|d}t|}t
|}t|}	t|	|}W5|rt||rt|X|S)Nz2SecureTransport only supports dumping binary certsr)
ValueErrorrrZrr[r\rCr)r]r	ZSecTrustGetCertificateCountZSecTrustGetCertificateAtIndexZSecCertificateCopyDataZCFDataGetLengthZCFDataGetBytePtrr<)
rKZbinary_formrdZcertdataZ	der_bytesrRZ
cert_countZleafZdata_lengthr4rrrgetpeercerts.




zWrappedSocket.getpeercertcCst}t|jt|}t||jtj	kr4dS|jtj
krDdS|jtjkrTdS|jtjkrddS|jtj
krtdS|jtjkrdStd|dS)NzTLSv1.3zTLSv1.2zTLSv1.1ZTLSv1ZSSLv3ZSSLv2zUnknown TLS version: %r)rZSSLProtocolZSSLGetNegotiatedProtocolVersionrCr)r]r	r`rZkTLSProtocol13rnkTLSProtocol11
kTLSProtocol1
kSSLProtocol3
kSSLProtocol2r^r_)rKprotocolrRrrrversions&zWrappedSocket.versioncCs|jd7_dSNr)rDrsrrr_reuseszWrappedSocket._reusecCs&|jdkr|n|jd8_dSr)rDrMrsrrr_drops

zWrappedSocket._drop)N)F)__name__
__module____qualname____doc__rL
contextlibcontextmanagerrNrSrerprrrtrur+rJr%r=r|r}rMrrrrrrrrrA@s*

9a
+
:rAcCs|jd7_t|||ddS)NrT)rM)rDr
)rKmodebufsizerrrmakefilesrrcOsd}t|||f||Srzr)rKr	bufferingargskwargsrrrrsc@seZdZdZddZeddZejddZeddZejd	dZed
dZ	e	jddZ	d
dZ
ddZddZdddZ
dddZdddZdS)rz
    I am a wrapper class for the SecureTransport library, to translate the
    interface of the standard library ``SSLContext`` object to calls into
    SecureTransport.
    cCs8t|\|_|_d|_d|_d|_d|_d|_d|_dSrB)	_protocol_to_min_max_min_version_max_version_options_verify
_trust_bundle_client_cert_client_key_client_key_passphrase)rKrrrrrLszSecureTransportContext.__init__cCsdS)
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        Trrsrrrcheck_hostnamesz%SecureTransportContext.check_hostnamecCsdS)rNrrKr`rrrr	scCs|jSrqrrsrrroptionsszSecureTransportContext.optionscCs
||_dSrqrrrrrrscCs|jrtjStjSrq)rr^
CERT_REQUIREDZ	CERT_NONErsrrrverify_mode sz"SecureTransportContext.verify_modecCs|tjkrdnd|_dS)NTF)r^rrrrrrr$scCsdSrqrrsrrrset_default_verify_paths(s
z/SecureTransportContext.set_default_verify_pathscCs|Srq)rrsrrrload_default_certs4sz)SecureTransportContext.load_default_certscCs|tjjkrtddS)Nz5SecureTransport doesn't support custom cipher strings)rrZDEFAULT_CIPHERSr)rKrQrrrset_ciphers7sz"SecureTransportContext.set_ciphersNcCs|dk	rtd|p||_dS)Nz1SecureTransport does not support cert directories)rr)rKZcafileZcapathZcadatarrrload_verify_locations<sz,SecureTransportContext.load_verify_locationscCs||_||_||_dSrq)rrZ_client_cert_passphrase)rKZcertfileZkeyfileZpasswordrrrload_cert_chainCsz&SecureTransportContext.load_cert_chainFTc
Cs2t|}|||j|j|j|j|j|j|j|Srq)	rArprrrrrrr)rKZsockZserver_sideZdo_handshake_on_connectZsuppress_ragged_eofsror6rrrwrap_socketHs
z"SecureTransportContext.wrap_socket)NNN)NN)FTTN)rrrrrLpropertyrsetterrrrrrrrrrrrrrs0	


	




r)r)rN)`rZ
__future__rrr)r'Zos.pathrUr~r$r^Z	threadingweakrefrZ_securetransport.bindingsrrrZ_securetransport.low_levelr	r
rrr
ImportErrorZpackages.backports.makefiler__all__rrrrrZWeakValueDictionaryr!ZLockrhr{Z'TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384Z'TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256Z%TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384Z%TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256Z-TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256Z+TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256Z#TLS_DHE_RSA_WITH_AES_256_GCM_SHA384Z#TLS_DHE_RSA_WITH_AES_128_GCM_SHA256Z'TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384Z$TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHAZ'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256Z$TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHAZ%TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384Z"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHAZ%TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256Z"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_256_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_256_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_128_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_128_CBC_SHAZTLS_AES_256_GCM_SHA384ZTLS_AES_128_GCM_SHA256ZTLS_RSA_WITH_AES_256_GCM_SHA384ZTLS_RSA_WITH_AES_128_GCM_SHA256ZTLS_AES_128_CCM_8_SHA256ZTLS_AES_128_CCM_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHA256ZTLS_RSA_WITH_AES_128_CBC_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHAZTLS_RSA_WITH_AES_128_CBC_SHArPZPROTOCOL_TLSrrmrhasattrrrrrrrrrnrrrr;r@ZSSLReadFuncrfZSSLWriteFuncrgobjectrArrrrrrs3%




76

!
site-packages/pip/_vendor/urllib3/contrib/__pycache__/pyopenssl.cpython-38.opt-1.pyc000064400000035125151464452710024376 0ustar00U

.e8@@s\dZddlmZddlZddlmZddlmZ	ddl
mZzddlm
Z
Wn$ek
rpGdd	d	eZ
YnXdd
lmZmZddlmZzddlmZWn$ek
rdZd
dlmZYnXddlZddlZd
dlmZddlZd
dlmZddgZ dZ!ej"ej#j$ej%ej#j&iZ'e(edr@e(ej#dr@ej#j)e'ej*<e(edrhe(ej#drhej#j+e'ej,<e(edre(ej#drej#j-e'ej.<ej/ej#j0ej1ej#j2ej3ej#j2ej#j4iZ5e6dde57DZ8dZ9ej!Z:ej;je?Z@ddZAddZBdd ZCd!d"ZDd#d$ZEGd%d&d&eFZGer6d.d(d)ZHneZHeHeG_HGd*d+d+eFZId,d-ZJdS)/ab
SSL with SNI_-support for Python 2. Follow these instructions if you would
like to verify SSL certificates in Python 2. Note, the default libraries do
*not* do certificate checking; you need to do additional work to validate
certificates yourself.

This needs the following packages installed:

* pyOpenSSL (tested with 16.0.0)
* cryptography (minimum 1.3.4, from pyopenssl)
* idna (minimum 2.0, from cryptography)

However, pyopenssl depends on cryptography, which depends on idna, so while we
use all three directly here we end up having relatively few packages required.

You can install them with the following command:

    pip install pyopenssl cryptography idna

To activate certificate checking, call
:func:`~urllib3.contrib.pyopenssl.inject_into_urllib3` from your Python code
before you begin making HTTP requests. This can be done in a ``sitecustomize``
module, or at any other time before your application begins using ``urllib3``,
like this::

    try:
        import urllib3.contrib.pyopenssl
        urllib3.contrib.pyopenssl.inject_into_urllib3()
    except ImportError:
        pass

Now you can use :mod:`urllib3` as you normally would, and it will support SNI
when the required modules are installed.

Activating this module also has the positive side effect of disabling SSL/TLS
compression in Python 2 (see `CRIME attack`_).

If you want to configure the default list of supported cipher suites, you can
set the ``urllib3.contrib.pyopenssl.DEFAULT_SSL_CIPHER_LIST`` variable.

.. _sni: https://en.wikipedia.org/wiki/Server_Name_Indication
.. _crime attack: https://en.wikipedia.org/wiki/CRIME_(security_exploit)
)absolute_importN)x509)backend)_Certificate)UnsupportedExtensionc@seZdZdS)rN)__name__
__module____qualname__r
r
I/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/pyopenssl.pyr7sr)timeouterror)BytesIO)_fileobject)backport_makefile)six)utilinject_into_urllib3extract_from_urllib3TPROTOCOL_SSLv3SSLv3_METHODPROTOCOL_TLSv1_1TLSv1_1_METHODPROTOCOL_TLSv1_2TLSv1_2_METHODccs|]\}}||fVqdSNr
).0kvr
r
r	gsr i@cCs4ttt_ttj_tt_ttj_dt_dtj_dS)z7Monkey-patch urllib3 with PyOpenSSL-backed SSL-support.TN)_validate_dependencies_metPyOpenSSLContextr
SSLContextssl_HAS_SNIIS_PYOPENSSLr
r
r
rrsscCs.tt_ttj_tt_ttj_dt_dtj_dS)z4Undo monkey-patching by :func:`inject_into_urllib3`.FN)orig_util_SSLContextrr#r$orig_util_HAS_SNIr%r&r
r
r
rrscCsRddlm}t|dddkr$tdddlm}|}t|dddkrNtddS)	z{
    Verifies that PyOpenSSL's package-level dependencies have been met.
    Throws `ImportError` if they are not met.
    r)
Extensionsget_extension_for_classNzX'cryptography' module missing required functionality.  Try upgrading to v1.3.4 or newer.)X509_x509zS'pyOpenSSL' module missing required functionality. Try upgrading to v0.14 or newer.)Zcryptography.x509.extensionsr)getattrImportErrorZOpenSSL.cryptor+)r)r+rr
r
rr!sr!cCs@dd}d|kr|S||}|dkr(dStjdkr<|d}|S)a%
    Converts a dNSName SubjectAlternativeName field to the form used by the
    standard library on the given Python version.

    Cryptography produces a dNSName as a unicode string that was idna-decoded
    from ASCII bytes. We need to idna-encode that string to get it back, and
    then on Python 3 we also need to convert to unicode via UTF-8 (the stdlib
    uses PyUnicode_FromStringAndSize on it, which decodes via UTF-8).

    If the name cannot be idna-encoded then we return None signalling that
    the name given should be skipped.
    cSsvddlm}zJdD]8}||r|t|d}|d||WSq||WS|jjk
rpYdSXdS)z
        Borrowed wholesale from the Python Cryptography Project. It turns out
        that we can't just safely call `idna.encode`: it can explode for
        wildcard names. This avoids that problem.
        r)idna)z*..Nascii)Zpip._vendorr/
startswithlenencodeZcoreZ	IDNAError)namer/prefixr
r
ridna_encodes
z'_dnsname_to_stdlib..idna_encode:N)rutf-8)sysversion_infodecode)r5r7r
r
r_dnsname_to_stdlibs

r>cCst|dr|}ntt|j}z|jtjj	}WnZtj
k
rNgYStjttj
tfk
r}ztd|gWYSd}~XYnXddtt|tjD}|dd|tjD|S)zU
    Given an PyOpenSSL certificate, provides all the subject alternative names.
    to_cryptographyzA problem was encountered with the certificate that prevented urllib3 from finding the SubjectAlternativeName field. This can affect certificate validation. The error was %sNcSsg|]}|dk	rd|fqS)NZDNSr
rr5r
r
r
sz%get_subj_alt_name..css|]}dt|fVqdS)z
IP AddressN)strr@r
r
rr sz$get_subj_alt_name..)hasattrr?ropenssl_backendr,
extensionsr*rZSubjectAlternativeNamevalueZExtensionNotFoundZDuplicateExtensionrZUnsupportedGeneralNameTypeUnicodeErrorlogZwarningmapr>Zget_values_for_typeZDNSNameextendZ	IPAddress)Z	peer_certZcertZextenamesr
r
rget_subj_alt_names2

	

rMc@seZdZdZd!ddZddZddZd	d
ZddZd
dZ	ddZ
ddZddZddZ
d"ddZddZddZddZd S)#
WrappedSocketzAPI-compatibility wrapper for Python OpenSSL's Connection-class.

    Note: _makefile_refs, _drop() and _reuse() are needed for the garbage
    collector of pypy.
    TcCs"||_||_||_d|_d|_dSNrF)
connectionsocketsuppress_ragged_eofs_makefile_refs_closed)selfrPrQrRr
r
r__init__
s
zWrappedSocket.__init__cCs
|jSr)rQfilenorUr
r
rrWszWrappedSocket.filenocCs*|jdkr|jd8_|jr&|dS)Nr)rSrTcloserXr
r
r_decref_socketioss
zWrappedSocket._decref_socketiosc
Osz|jj||}Wntjjk
r^}z*|jrB|jdkrBWYdStt|W5d}~XYntjj	k
r|j
tjjkrYdSYn~tjjk
rt
|j|jstdn|j||YSYn:tjjk
r}ztd|W5d}~XYnX|SdS)NzUnexpected EOFThe read operation timed outread error: %r)rPrecvOpenSSLSSLSysCallErrorrRargsSocketErrorrBZeroReturnErrorget_shutdownRECEIVED_SHUTDOWN
WantReadErrorr
wait_for_readrQ
gettimeoutrErrorsslSSLError)rUrekwargsdatarKr
r
rras"

 zWrappedSocket.recvc
Os
z|jj||WStjjk
r\}z*|jr@|jdkr@WYdStt|W5d}~XYntjj	k
r|j
tjjkrYdSYnztjjk
rt
|j|jstdn|j||YSYn6tjjk
r}ztd|W5d}~XYnXdS)Nr\rr_r`)rP	recv_intorbrcrdrRrerfrBrgrhrirjrrkrQrlrrmrnro)rUrerprKr
r
rrr7s 

zWrappedSocket.recv_intocCs|j|Sr)rQ
settimeout)rUrr
r
rrsNszWrappedSocket.settimeoutc
Cs|z|j|WStjjk
rDt|j|jsttjs*d|jfdd|j|pJ|dS)Nr:csSrr
)_passwordr
rr^z2PyOpenSSLContext.load_cert_chain..)rZuse_certificate_chain_filerrZbinary_typer4Z
set_passwd_cbZuse_privatekey_file)rUZcertfileZkeyfilerr
rrload_cert_chains
z PyOpenSSLContext.load_cert_chainFTc
Cstj|j|}t|tjr&|d}|dk	r8|||	z|
Wqtjjk
rt
||sxtdYq@Yqtjjk
r}ztd|W5d}~XYqXqq@t||S)Nr:zselect timed outzbad handshake: %r)rbrcZ
Connectionrrrrr4Zset_tlsext_host_nameZset_connect_stateZdo_handshakerjrrkrlrrmrnrorN)rUZsockZserver_sideZdo_handshake_on_connectrRZserver_hostnamecnxrKr
r
rwrap_sockets 

 zPyOpenSSLContext.wrap_socket)NNN)NN)FTTN)rrr	rrVpropertyrsetterrrrrrrr
r
r
rr"s&




	
r"cCs|dkSrvr
)rrZerr_noZ	err_depthZreturn_coder
r
rrsr)r])KrZ
__future__rZOpenSSL.SSLrbZcryptographyrZ$cryptography.hazmat.backends.opensslrrDZ)cryptography.hazmat.backends.openssl.x509rZcryptography.x509rr.	ExceptionrQrr
rfiorrZpackages.backports.makefilerZloggingrnZpackagesrr;r__all__r%ZPROTOCOL_TLSrcZ
SSLv23_METHODZPROTOCOL_TLSv1ZTLSv1_METHODrrCrrrrrrZ	CERT_NONEZVERIFY_NONEZ
CERT_OPTIONALZVERIFY_PEERZ
CERT_REQUIREDZVERIFY_FAIL_IF_NO_PEER_CERTrdictitemsrrwr(r$r#r'Z	getLoggerrrHrrr!r>rMobjectrNrr"rr
r
r
rsz+

+6Wsite-packages/pip/_vendor/urllib3/contrib/__pycache__/securetransport.cpython-38.pyc000064400000046605151464452710024653 0ustar00U

.e@sdZddlmZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
mZddlmZmZmZddlmZmZmZmZzdd	l	mZWn$ek
rdZdd
lmZYnXddgZd
ZejZejj Z!e"Z#e$Z%dZ&ej'ej(ej)ej*ej+ej,ej-ej.ej/ej0ej1ej2ej3ej4ej5ej6ej7ej8ej9ej:ej;ejej?ej@ejAejBejCejDgZEejFejGejHfiZIeJe
drejKejKfeIe
jL<eJe
drejMejMfeIe
jN<eJe
drejGejGfeIe
jO<eJe
drejPejPfeIe
jQ<eJe
drejRejRfeIe
jS<ddZTddZUddZVddZWeXeVZYeZeWZ[Gddde\Z]ernd#ddZ^n
d$d dZ^e^e]_^Gd!d"d"e\Z_dS)%a	
SecureTranport support for urllib3 via ctypes.

This makes platform-native TLS available to urllib3 users on macOS without the
use of a compiler. This is an important feature because the Python Package
Index is moving to become a TLSv1.2-or-higher server, and the default OpenSSL
that ships with macOS is not capable of doing TLSv1.2. The only way to resolve
this is to give macOS users an alternative solution to the problem, and that
solution is to use SecureTransport.

We use ctypes here because this solution must not require a compiler. That's
because pip is not allowed to require a compiler either.

This is not intended to be a seriously long-term solution to this problem.
The hope is that PEP 543 will eventually solve this issue for us, at which
point we can retire this contrib module. But in the short term, we need to
solve the impending tire fire that is Python on Mac without this kind of
contrib module. So...here we are.

To use this module, simply import and inject it::

    import urllib3.contrib.securetransport
    urllib3.contrib.securetransport.inject_into_urllib3()

Happy TLSing!

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

    Copyright (c) 2015-2016 Will Bond 

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
)absolute_importN)util)Security
SecurityConstCoreFoundation)_assert_no_error_cert_array_from_pem_temporary_keychain_load_client_cert_chain)_fileobjectbackport_makefileinject_into_urllib3extract_from_urllib3Ti@PROTOCOL_SSLv2PROTOCOL_SSLv3PROTOCOL_TLSv1PROTOCOL_TLSv1_1PROTOCOL_TLSv1_2cCs.tt_ttj_tt_ttj_dt_dtj_dS)zG
    Monkey-patch urllib3 with SecureTransport-backed SSL-support.
    TN)SecureTransportContextr
SSLContextssl_HAS_SNIIS_SECURETRANSPORTrrO/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/securetransport.pyrscCs.tt_ttj_tt_ttj_dt_dtj_dS)z>
    Undo monkey-patching by :func:`inject_into_urllib3`.
    FN)orig_util_SSLContextrrrorig_util_HAS_SNIrrrrrrrsc

Csd}zJt|}|dkr"tjWS|j}|d}|}d}d}z|||kr|dksZ|dkrtt||sttt	j
d||}	tj|	
||}
||
|	}||7}|sB|stjWWSqqBWnttjk
r0}zR|j	}|dk	r |t	j
kr ||d<|t	jks|t	jkrtjWY
WSW5d}~XYnX||d<||krLtjWSWdStk
r}z|dk	rt||_tjWYSd}~XYnXdS)zs
    SecureTransport read callback. This is called by ST to request that data
    be returned from the socket.
    Nr	timed out)_connection_refsgetrerrSSLInternalsocket
gettimeoutrZ
wait_for_readerrorerrnoEAGAINctypesc_charZfrom_address	recv_intoerrSSLClosedGraceful
ECONNRESETEPIPEerrSSLClosedAborterrSSLWouldBlock	Exception
_exception)

connection_iddata_bufferdata_length_pointerwrapped_socketbase_socketZrequested_lengthtimeoutr&Z
read_countZ	remainingbufferZ
chunk_sizeerrr_read_callbacksP




r;c
Csrd}z(t|}|dkr"tjWS|j}|d}t||}|}d}d}	zV|	|kr|dksf|dkrt	||st
tjd|
|}
|	|
7}	||
d}qNWnltj
k
r}zJ|j}|dk	r|tjkr|	|d<|tjks|tjkrtjWY
WSW5d}~XYnX|	|d<|	|kr*tjWSWdStk
rl}z|dk	rR||_tjWYSd}~XYnXdS)zx
    SecureTransport write callback. This is called by ST to request that data
    actually be sent on the network.
    Nrr )r!r"rr#r$r)	string_atr%rZwait_for_writer&r'r(sendr-r.r/r0r1r2)r3r4r5r6r7Zbytes_to_writedatar8r&sentZ
chunk_sentr:rrr_write_callbacksD



r@c@seZdZdZddZejddZddZdd	Z	d
dZ
dd
ZddZddZ
d*ddZddZddZddZddZddZdd Zd+d"d#Zd$d%Zd&d'Zd(d)ZdS),
WrappedSocketz
    API-compatibility wrapper for Python's OpenSSL wrapped socket object.

    Note: _makefile_refs, _drop(), and _reuse() are needed for the garbage
    collector of PyPy.
    cCsL||_d|_d|_d|_d|_d|_d|_d|_|j|_	|j
ddSNrF)r$context_makefile_refs_closedr2	_keychain
_keychain_dir_client_cert_chainr%_timeout
settimeout)selfr$rrr__init__HszWrappedSocket.__init__ccs4d|_dV|jdk	r0|jd}|_||dS)a]
        A context manager that can be used to wrap calls that do I/O from
        SecureTransport. If any of the I/O callbacks hit an exception, this
        context manager will correctly propagate the exception after the fact.
        This avoids silently swallowing those exceptions.

        It also correctly forces the socket closed.
        N)r2close)rKZ	exceptionrrr_raise_on_errorZs

zWrappedSocket._raise_on_errorcCs2tjttt}t|j|tt}t|dS)a4
        Sets up the allowed ciphers. By default this matches the set in
        util.ssl_.DEFAULT_CIPHERS, at least as supported by macOS. This is done
        custom and doesn't allow changing at this time, mostly because parsing
        OpenSSL cipher strings is going to be a freaking nightmare.
        N)rZSSLCipherSuitelen
CIPHER_SUITESZSSLSetEnabledCiphersrCr	)rKciphersresultrrr_set_ciphersoszWrappedSocket._set_ciphersc		Cs|sdStj|r2t|d}|}W5QRXd}t}zt	|}t
|jt
|}t||srtdt||}t|t|d}t|t}t|t
|}t|W5|rt||dk	rt|Xtjtjf}|j|kr
td|jdS)z
        Called when we have set custom validation. We do this in two cases:
        first, when cert validation is entirely disabled; and second, when
        using a custom trust DB.
        NrbzFailed to copy trust referenceTz)certificate verify failed, error code: %d)ospathisfileopenreadrSecTrustRefr	CFReleaser
SSLCopyPeerTrustrCr)byrefr	sslSSLErrorZSecTrustSetAnchorCertificatesZ!SecTrustSetAnchorCertificatesOnlyZSecTrustResultTypeZSecTrustEvaluaterZkSecTrustResultUnspecifiedZkSecTrustResultProceedvalue)	rKverifytrust_bundlefZ
cert_arraytrustrRZtrust_resultZ	successesrrr_custom_validate|s>

zWrappedSocket._custom_validatec		Cstdtjtj|_t|jtt}	t	|	t
0t|d}
|
tkrT|
dd}
q>|t|
<W5QRXt
|j|
}	t	|	|rt|ts|d}t|j|t|}	t	|	|t|j|}	t	|	t|j|}	|	dkr|tjkrt|jtj}	t	|	|r|dk	r0t|jtjd}	t	|	|rlt\|_|_t|j|||_t|j|j}	t	|	| ht!|j}	|	tj"krt#$dn>|	tj%kr|&||W5QRqlnt	|	W5QRqW5QRXqldS)z
        Actually performs the TLS handshake. This is run automatically by
        wrapped socket, and shouldn't be needed in user code.
        Nirzutf-8rTzhandshake timed out)'rZSSLCreateContextrZkSSLClientSideZkSSLStreamTyperCZ
SSLSetIOFuncs_read_callback_pointer_write_callback_pointerr	_connection_ref_lockidr!ZSSLSetConnection
isinstancebytesencodeZSSLSetPeerDomainNamerOrSZSSLSetProtocolVersionMinZSSLSetProtocolVersionMaxkTLSProtocolMaxSupportedkTLSProtocol12ZSSLSetSessionOptionZ"kSSLSessionOptionBreakOnServerAuthrrFrGrrHZSSLSetCertificaterNZSSLHandshaker0r$r8ZerrSSLServerAuthCompletedre)rKserver_hostnamerarbZmin_versionZmax_versionZclient_certZ
client_keyZclient_key_passphraserRZhandlerrr	handshakes|


zWrappedSocket.handshakecCs
|jSN)r$filenorKrrrrrszWrappedSocket.filenocCs*|jdkr|jd8_|jr&|dS)Nrr)rDrErMrsrrr_decref_socketioss
zWrappedSocket._decref_socketioscCs&t|}|||}|d|}|Srq)r)Zcreate_string_bufferr+)rKZbufsizr9Z
bytes_readr>rrrrecv s
zWrappedSocket.recvNc	Cs|jr
dS|dkrt|}tj||}td}|t|j	||t
|}W5QRX|tjkr|j
dkrtdn"|tjtjfkr|nt||j
S)Nrzrecv timed out)rErOr)r*Zfrom_bufferc_size_trNrZSSLReadrCr]rr0r`r$r8r,ZerrSSLClosedNoNotifyrMr	)rKr9nbytesprocessed_bytesrRrrrr+&s.




zWrappedSocket.recv_intocCs
||_dSrqrI)rKr8rrrrJQszWrappedSocket.settimeoutcCs|jSrqryrsrrrr%TszWrappedSocket.gettimeoutc	Cshtd}|"t|j|t|t|}W5QRX|tj	krZ|j
dkrZtdnt
||j
S)Nrzsend timed out)r)rvrNrZSSLWriterCrOr]rr0r`r$r8r	)rKr>rxrRrrrr=Ws

zWrappedSocket.sendcCs4d}|t|kr0||||t}||7}qdSNr)rOr=SSL_WRITE_BLOCKSIZE)rKr>Z
total_sentr?rrrsendallhszWrappedSocket.sendallc	Cs$|t|jW5QRXdSrq)rNrZSSLCloserCrsrrrshutdownns
zWrappedSocket.shutdowncCs|jdkrd|_|jr(t|jd|_|jr@t|jd|_|jrvt|jt|jt	
|jd|_|_|j
S|jd8_dS)NrT)rDrErCrr[rHrFrZSecKeychainDeleteshutilZrmtreerGr$rMrsrrrrMrs

zWrappedSocket.closeFc
	Cs|stdt}d}d}zt|jt|}t	||sFWddSt
|}|s\WNdSt|d}|sptt
|}|stt|}t|}	t|	|}W5|rt||rt|X|S)Nz2SecureTransport only supports dumping binary certsr)
ValueErrorrrZrr[r\rCr)r]r	ZSecTrustGetCertificateCountZSecTrustGetCertificateAtIndexAssertionErrorZSecCertificateCopyDataZCFDataGetLengthZCFDataGetBytePtrr<)
rKZbinary_formrdZcertdataZ	der_bytesrRZ
cert_countZleafZdata_lengthr4rrrgetpeercerts2




zWrappedSocket.getpeercertcCst}t|jt|}t||jtj	kr4dS|jtj
krDdS|jtjkrTdS|jtjkrddS|jtj
krtdS|jtjkrdStd|dS)NzTLSv1.3zTLSv1.2zTLSv1.1ZTLSv1ZSSLv3ZSSLv2zUnknown TLS version: %r)rZSSLProtocolZSSLGetNegotiatedProtocolVersionrCr)r]r	r`rZkTLSProtocol13rnkTLSProtocol11
kTLSProtocol1
kSSLProtocol3
kSSLProtocol2r^r_)rKprotocolrRrrrversions&zWrappedSocket.versioncCs|jd7_dSNr)rDrsrrr_reuseszWrappedSocket._reusecCs&|jdkr|n|jd8_dSr)rDrMrsrrr_drops

zWrappedSocket._drop)N)F)__name__
__module____qualname____doc__rL
contextlibcontextmanagerrNrSrerprrrtrur+rJr%r=r|r}rMrrrrrrrrrA@s*

9a
+
:rAcCs|jd7_t|||ddS)NrT)rM)rDr
)rKmodebufsizerrrmakefilesrrcOsd}t|||f||Srzr)rKr	bufferingargskwargsrrrrsc@seZdZdZddZeddZejddZeddZejd	dZed
dZ	e	jddZ	d
dZ
ddZddZdddZ
dddZdddZdS)rz
    I am a wrapper class for the SecureTransport library, to translate the
    interface of the standard library ``SSLContext`` object to calls into
    SecureTransport.
    cCs8t|\|_|_d|_d|_d|_d|_d|_d|_dSrB)	_protocol_to_min_max_min_version_max_version_options_verify
_trust_bundle_client_cert_client_key_client_key_passphrase)rKrrrrrLszSecureTransportContext.__init__cCsdS)
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        Trrsrrrcheck_hostnamesz%SecureTransportContext.check_hostnamecCsdS)rNrrKr`rrrr	scCs|jSrqrrsrrroptionsszSecureTransportContext.optionscCs
||_dSrqrrrrrrscCs|jrtjStjSrq)rr^
CERT_REQUIREDZ	CERT_NONErsrrrverify_mode sz"SecureTransportContext.verify_modecCs|tjkrdnd|_dS)NTF)r^rrrrrrr$scCsdSrqrrsrrrset_default_verify_paths(s
z/SecureTransportContext.set_default_verify_pathscCs|Srq)rrsrrrload_default_certs4sz)SecureTransportContext.load_default_certscCs|tjjkrtddS)Nz5SecureTransport doesn't support custom cipher strings)rrZDEFAULT_CIPHERSr)rKrQrrrset_ciphers7sz"SecureTransportContext.set_ciphersNcCs|dk	rtd|p||_dS)Nz1SecureTransport does not support cert directories)rr)rKZcafileZcapathZcadatarrrload_verify_locations<sz,SecureTransportContext.load_verify_locationscCs||_||_||_dSrq)rrZ_client_cert_passphrase)rKZcertfileZkeyfileZpasswordrrrload_cert_chainCsz&SecureTransportContext.load_cert_chainFTc
CsJ|rt|st|stt|}|||j|j|j|j|j|j|j	|Srq)
rrArprrrrrrr)rKZsockZserver_sideZdo_handshake_on_connectZsuppress_ragged_eofsror6rrrwrap_socketHs
z"SecureTransportContext.wrap_socket)NNN)NN)FTTN)rrrrrLpropertyrsetterrrrrrrrrrrrrrs0	


	




r)r)rN)`rZ
__future__rrr)r'Zos.pathrUr~r$r^Z	threadingweakrefrZ_securetransport.bindingsrrrZ_securetransport.low_levelr	r
rrr
ImportErrorZpackages.backports.makefiler__all__rrrrrZWeakValueDictionaryr!ZLockrhr{Z'TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384Z'TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256Z%TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384Z%TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256Z-TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256Z+TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256Z#TLS_DHE_RSA_WITH_AES_256_GCM_SHA384Z#TLS_DHE_RSA_WITH_AES_128_GCM_SHA256Z'TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384Z$TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHAZ'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256Z$TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHAZ%TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384Z"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHAZ%TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256Z"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_256_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_256_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_128_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_128_CBC_SHAZTLS_AES_256_GCM_SHA384ZTLS_AES_128_GCM_SHA256ZTLS_RSA_WITH_AES_256_GCM_SHA384ZTLS_RSA_WITH_AES_128_GCM_SHA256ZTLS_AES_128_CCM_8_SHA256ZTLS_AES_128_CCM_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHA256ZTLS_RSA_WITH_AES_128_CBC_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHAZTLS_RSA_WITH_AES_128_CBC_SHArPZPROTOCOL_TLSrrmrhasattrrrrrrrrrnrrrr;r@ZSSLReadFuncrfZSSLWriteFuncrgobjectrArrrrrrs3%




76

!
site-packages/pip/_vendor/urllib3/contrib/__pycache__/_appengine_environ.cpython-38.pyc000064400000002073151464452710025244 0ustar00U

.e@s8dZddlZddZddZddZd	d
ZddZdS)
zB
This module provides means to detect the App Engine environment.
NcCstptptSN)is_local_appengineis_prod_appengineis_prod_appengine_mvmsrrR/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_appengine_environ.pyis_appenginesrcCstotSr)rrrrrris_appengine_sandboxsr	cCsdtjkodtjdkS)NAPPENGINE_RUNTIMEzDevelopment/SERVER_SOFTWARE)osenvironrrrrrs
rcCs dtjkodtjdkotS)Nr
zGoogle App Engine/r)rr
rrrrrrs

rcCstjdddkS)NZGAE_VMFtrue)rr
getrrrrrsr)__doc__rrr	rrrrrrrssite-packages/pip/_vendor/urllib3/contrib/__pycache__/__init__.cpython-38.pyc000064400000000243151464452710023133 0ustar00U

.e@sdS)NrrrH/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/__init__.pysite-packages/pip/_vendor/urllib3/contrib/__pycache__/_appengine_environ.cpython-38.opt-1.pyc000064400000002073151464452710026203 0ustar00U

.e@s8dZddlZddZddZddZd	d
ZddZdS)
zB
This module provides means to detect the App Engine environment.
NcCstptptSN)is_local_appengineis_prod_appengineis_prod_appengine_mvmsrrR/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_appengine_environ.pyis_appenginesrcCstotSr)rrrrrris_appengine_sandboxsr	cCsdtjkodtjdkS)NAPPENGINE_RUNTIMEzDevelopment/SERVER_SOFTWARE)osenvironrrrrrs
rcCs dtjkodtjdkotS)Nr
zGoogle App Engine/r)rr
rrrrrrs

rcCstjdddkS)NZGAE_VMFtrue)rr
getrrrrrsr)__doc__rrr	rrrrrrrssite-packages/pip/_vendor/urllib3/contrib/__pycache__/ntlmpool.cpython-38.pyc000064400000006260151464452710023245 0ustar00U

.ei@s\dZddlmZddlmZddlmZddlmZddlm	Z	ee
ZGdd	d	eZd
S)z
NTLM authenticating pool, contributed by erikcederstran

Issue #10, see: http://code.google.com/p/urllib3/issues/detail?id=10
)absolute_import)	getLogger)ntlm)HTTPSConnectionPool)HTTPSConnectioncs:eZdZdZdZfddZddZdfd
d	ZZS)
NTLMConnectionPoolzQ
    Implements an NTLM authentication version of an urllib3 connection pool
    ZhttpscsLtt|j||||_||_|dd}|d|_|d|_||_	dS)z
        authurl is a random URL on the server that is protected by NTLM.
        user is the Windows user, probably in the DOMAIN\username format.
        pw is the password for the user.
        \rN)
superr__init__authurlrawusersplitupperdomainuserpw)selfrrr
argskwargsZ
user_parts	__class__H/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/ntlmpool.pyrs
zNTLMConnectionPool.__init__c
Cs|jd7_td|j|j|jddi}d}d}t|j|jd}dt|j	||<td	||
d
|jd||}t|
}td|j|jtd|td
|dd|_||d}d}|D] }	|	dddkr|	dd}q|dkrtd|||ft|\}
}t|
|j|j|j|}d|||<td	||
d
|jd||}td|j|jtdt|
td
|dd|jdkr|jdkrtdtd|j|jfd|_td|S)Nr
z3Starting NTLM HTTPS connection no. %d: https://%s%s
Connection
Keep-AliveZ
Authorizationzwww-authenticate)hostportzNTLM %szRequest headers: %sZGETzResponse status: %s %szResponse headers: %szResponse data: %s [...]dz, zNTLM z!Unexpected %s response header: %siz3Server rejected request: wrong username or passwordzWrong server response: %s %szConnection established)Znum_connectionslogdebugrr
rrrZcreate_NTLM_NEGOTIATE_MESSAGErZrequestZgetresponsedictZ
getheadersZstatusreasonreadfpr	ExceptionZparse_NTLM_CHALLENGE_MESSAGEZ create_NTLM_AUTHENTICATE_MESSAGErrr)
rheadersZ
req_headerZresp_headerZconnresZreshdrZauth_header_valuesZauth_header_valuesZServerChallengeZNegotiateFlagsZauth_msgrrr	_new_conn'sr


zNTLMConnectionPool._new_connNTc	s0|dkri}d|d<tt||||||||S)Nrr)rrurlopen)rmethodZurlZbodyr)ZretriesZredirectZassert_same_hostrrrr.ls

zNTLMConnectionPool.urlopen)NNr-TT)	__name__
__module____qualname____doc__Zschemerr,r.
__classcell__rrrrrsIrN)
r3Z
__future__rZloggingrrrZpackages.six.moves.http_clientrr0r"rrrrrssite-packages/pip/_vendor/urllib3/contrib/__pycache__/appengine.cpython-38.pyc000064400000020312151464452710023341 0ustar00U

.e2,@sdZddlmZddlZddlZddlZddlmZddlm	Z	m
Z
mZmZm
Z
mZddlmZddlmZdd	lmZdd
lmZddlmZzdd
lmZWnek
rdZYnXeeZGddde
ZGddde	Z GdddeZ!ej"Z"ej#Z#ej$Z$ej%Z%ej&Z&dS)a[
This module provides a pool manager that uses Google App Engine's
`URLFetch Service `_.

Example usage::

    from pip._vendor.urllib3 import PoolManager
    from pip._vendor.urllib3.contrib.appengine import AppEngineManager, is_appengine_sandbox

    if is_appengine_sandbox():
        # AppEngineManager uses AppEngine's URLFetch API behind the scenes
        http = AppEngineManager()
    else:
        # PoolManager uses a socket-level API behind the scenes
        http = PoolManager()

    r = http.request('GET', 'https://google.com/')

There are `limitations `_ to the URLFetch service and it may not be
the best choice for your application. There are three options for using
urllib3 on Google App Engine:

1. You can use :class:`AppEngineManager` with URLFetch. URLFetch is
   cost-effective in many circumstances as long as your usage is within the
   limitations.
2. You can use a normal :class:`~urllib3.PoolManager` by enabling sockets.
   Sockets also have `limitations and restrictions
   `_ and have a lower free quota than URLFetch.
   To use sockets, be sure to specify the following in your ``app.yaml``::

        env_variables:
            GAE_USE_SOCKETS_HTTPLIB : 'true'

3. If you are using `App Engine Flexible
`_, you can use the standard
:class:`PoolManager` without any configuration or special environment variables.
)absolute_importN)urljoin)	HTTPErrorHTTPWarning
MaxRetryError
ProtocolErrorTimeoutErrorSSLError)RequestMethods)HTTPResponse)Timeout)Retry)_appengine_environ)urlfetchc@seZdZdS)AppEnginePlatformWarningN__name__
__module____qualname__rrI/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/appengine.pyrGsrc@seZdZdS)AppEnginePlatformErrorNrrrrrrKsrc@sXeZdZdZdddZddZdd	Zddddejfd
dZ	dd
Z
ddZddZdS)AppEngineManagera
    Connection manager for Google App Engine sandbox applications.

    This manager uses the URLFetch service directly instead of using the
    emulated httplib, and is subject to URLFetch limitations as described in
    the App Engine documentation `here
    `_.

    Notably it will raise an :class:`AppEnginePlatformError` if:
        * URLFetch is not available.
        * If you attempt to use this on App Engine Flexible, as full socket
          support is available.
        * If a request size is more than 10 megabytes.
        * If a response size is more than 32 megabtyes.
        * If you use an unsupported request method such as OPTIONS.

    Beyond those cases, it will raise normal urllib3 errors.
    NTcCsNtstdtrtdtdtt||||_||_	|pFt
j|_dS)Nz.URLFetch is not available in this environment.zUse normal urllib3.PoolManager instead of AppEngineManageron Managed VMs, as using URLFetch is not necessary in this environment.zurllib3 is using URLFetch on Google App Engine sandbox instead of sockets. To use sockets directly instead of URLFetch see https://urllib3.readthedocs.io/en/latest/reference/urllib3.contrib.html.)
rris_prod_appengine_mvmswarningswarnrr__init__validate_certificateurlfetch_retriesrZDEFAULTretries)selfheadersr!rr rrrrcs zAppEngineManager.__init__cCs|S)Nr)r"rrr	__enter__szAppEngineManager.__enter__cCsdS)NFr)r"exc_typeZexc_valZexc_tbrrr__exit__szAppEngineManager.__exit__c
Ks|||}zF|o |jdko |j}	tj||||p2id|jo<|	|||jd}
WnBtjk
r}zt	||W5d}~XYntj
k
r}z"dt|krtd|t
|W5d}~XYntjk
r}z&dt|krt|||dt
|W5d}~XYntjk
r6}ztd|W5d}~XYn`tjk
rb}zt|W5d}~XYn4tjk
r}ztd	||W5d}~XYnX|j|
fd
|i|}|o|}
|
rv|jr|jrt||dn|jdkrd
}z|j||||d}Wn.tk
r2|jr*t||d|YSX||td||
t||
}|j||||f|||d|St|d}| ||j|r|j||||d}td||!||j||f|||||d|S|S)NrF)Zpayloadmethodr#Zallow_truncatedfollow_redirectsZdeadlinerz	too largezOURLFetch request too large, URLFetch only supports requests up to 10mb in size.zToo many redirects)reasonzPURLFetch response too large, URLFetch only supportsresponses up to 32mb in size.z$URLFetch does not support method: %sr!ztoo many redirectsi/ZGET)responseZ_poolzRedirecting %s -> %s)r!redirecttimeoutzRetry-Afterz	Retry: %s)bodyr#r!r+r,)"_get_retriesr+totalrZfetchr _get_absolute_timeoutrZDeadlineExceededErrorr	ZInvalidURLErrorstrrrZ
DownloadErrorrZResponseTooLargeErrorZSSLCertificateErrorr
ZInvalidMethodError#_urlfetch_response_to_http_responseZget_redirect_locationZraise_on_redirectstatusZ	incrementZsleep_for_retrylogdebugrurlopenboolZ	getheaderZis_retrysleep)r"r'Zurlr-r#r!r+r,response_kwr(r*eZ
http_responseZredirect_locationZredirect_urlZhas_retry_afterrrrr6s





zAppEngineManager.urlopencKstr"|jd}|dkr"|jd=|jd}|dkrZ|d}|dd||jd<tft|j	|j
|j|jd|}tft|j	|j|j|d|S)Nzcontent-encodingZdeflateztransfer-encodingZchunked,)r-msgr#r3)r-r#r3original_response)is_prod_appenginer#getsplitremovejoinrioBytesIOZcontentZ
header_msgZstatus_code)r"Z
urlfetch_respr9Zcontent_encodingZtransfer_encodingZ	encodingsr=rrrr2s2




z4AppEngineManager._urlfetch_response_to_http_responsecCsB|tjkrdSt|tr>|jdk	s,|jdk	r8tdt|jS|S)NzdURLFetch does not support granular timeout settings, reverting to total or default URLFetch timeout.)	r
DEFAULT_TIMEOUT
isinstanceZ_readZ_connectrrrr/)r"r,rrrr0 s

z&AppEngineManager._get_absolute_timeoutcCs>t|tstj|||jd}|js.|js.|jr:tdt	|S)N)r+defaultzhURLFetch only supports total retries and does not recognize connect, read, or redirect retry parameters.)
rFrZfrom_intr!Zconnectreadr+rrr)r"r!r+rrrr.-s
zAppEngineManager._get_retries)NNTT)
rrr__doc__rr$r&r
rEr6r2r0r.rrrrrOs"
 
r$
r)'rIZ
__future__rrCZloggingrZpackages.six.moves.urllib.parser
exceptionsrrrrr	r
Zrequestrr*rZutil.timeoutr
Z
util.retryrrZgoogle.appengine.apirImportErrorZ	getLoggerrr4rrrZis_appengineZis_appengine_sandboxZis_local_appenginer>rrrrrs2( 	

osite-packages/pip/_vendor/urllib3/contrib/__pycache__/appengine.cpython-38.opt-1.pyc000064400000020312151464452710024300 0ustar00U

.e2,@sdZddlmZddlZddlZddlZddlmZddlm	Z	m
Z
mZmZm
Z
mZddlmZddlmZdd	lmZdd
lmZddlmZzdd
lmZWnek
rdZYnXeeZGddde
ZGddde	Z GdddeZ!ej"Z"ej#Z#ej$Z$ej%Z%ej&Z&dS)a[
This module provides a pool manager that uses Google App Engine's
`URLFetch Service `_.

Example usage::

    from pip._vendor.urllib3 import PoolManager
    from pip._vendor.urllib3.contrib.appengine import AppEngineManager, is_appengine_sandbox

    if is_appengine_sandbox():
        # AppEngineManager uses AppEngine's URLFetch API behind the scenes
        http = AppEngineManager()
    else:
        # PoolManager uses a socket-level API behind the scenes
        http = PoolManager()

    r = http.request('GET', 'https://google.com/')

There are `limitations `_ to the URLFetch service and it may not be
the best choice for your application. There are three options for using
urllib3 on Google App Engine:

1. You can use :class:`AppEngineManager` with URLFetch. URLFetch is
   cost-effective in many circumstances as long as your usage is within the
   limitations.
2. You can use a normal :class:`~urllib3.PoolManager` by enabling sockets.
   Sockets also have `limitations and restrictions
   `_ and have a lower free quota than URLFetch.
   To use sockets, be sure to specify the following in your ``app.yaml``::

        env_variables:
            GAE_USE_SOCKETS_HTTPLIB : 'true'

3. If you are using `App Engine Flexible
`_, you can use the standard
:class:`PoolManager` without any configuration or special environment variables.
)absolute_importN)urljoin)	HTTPErrorHTTPWarning
MaxRetryError
ProtocolErrorTimeoutErrorSSLError)RequestMethods)HTTPResponse)Timeout)Retry)_appengine_environ)urlfetchc@seZdZdS)AppEnginePlatformWarningN__name__
__module____qualname__rrI/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/appengine.pyrGsrc@seZdZdS)AppEnginePlatformErrorNrrrrrrKsrc@sXeZdZdZdddZddZdd	Zddddejfd
dZ	dd
Z
ddZddZdS)AppEngineManagera
    Connection manager for Google App Engine sandbox applications.

    This manager uses the URLFetch service directly instead of using the
    emulated httplib, and is subject to URLFetch limitations as described in
    the App Engine documentation `here
    `_.

    Notably it will raise an :class:`AppEnginePlatformError` if:
        * URLFetch is not available.
        * If you attempt to use this on App Engine Flexible, as full socket
          support is available.
        * If a request size is more than 10 megabytes.
        * If a response size is more than 32 megabtyes.
        * If you use an unsupported request method such as OPTIONS.

    Beyond those cases, it will raise normal urllib3 errors.
    NTcCsNtstdtrtdtdtt||||_||_	|pFt
j|_dS)Nz.URLFetch is not available in this environment.zUse normal urllib3.PoolManager instead of AppEngineManageron Managed VMs, as using URLFetch is not necessary in this environment.zurllib3 is using URLFetch on Google App Engine sandbox instead of sockets. To use sockets directly instead of URLFetch see https://urllib3.readthedocs.io/en/latest/reference/urllib3.contrib.html.)
rris_prod_appengine_mvmswarningswarnrr__init__validate_certificateurlfetch_retriesrZDEFAULTretries)selfheadersr!rr rrrrcs zAppEngineManager.__init__cCs|S)Nr)r"rrr	__enter__szAppEngineManager.__enter__cCsdS)NFr)r"exc_typeZexc_valZexc_tbrrr__exit__szAppEngineManager.__exit__c
Ks|||}zF|o |jdko |j}	tj||||p2id|jo<|	|||jd}
WnBtjk
r}zt	||W5d}~XYntj
k
r}z"dt|krtd|t
|W5d}~XYntjk
r}z&dt|krt|||dt
|W5d}~XYntjk
r6}ztd|W5d}~XYn`tjk
rb}zt|W5d}~XYn4tjk
r}ztd	||W5d}~XYnX|j|
fd
|i|}|o|}
|
rv|jr|jrt||dn|jdkrd
}z|j||||d}Wn.tk
r2|jr*t||d|YSX||td||
t||
}|j||||f|||d|St|d}| ||j|r|j||||d}td||!||j||f|||||d|S|S)NrF)Zpayloadmethodr#Zallow_truncatedfollow_redirectsZdeadlinerz	too largezOURLFetch request too large, URLFetch only supports requests up to 10mb in size.zToo many redirects)reasonzPURLFetch response too large, URLFetch only supportsresponses up to 32mb in size.z$URLFetch does not support method: %sr!ztoo many redirectsi/ZGET)responseZ_poolzRedirecting %s -> %s)r!redirecttimeoutzRetry-Afterz	Retry: %s)bodyr#r!r+r,)"_get_retriesr+totalrZfetchr _get_absolute_timeoutrZDeadlineExceededErrorr	ZInvalidURLErrorstrrrZ
DownloadErrorrZResponseTooLargeErrorZSSLCertificateErrorr
ZInvalidMethodError#_urlfetch_response_to_http_responseZget_redirect_locationZraise_on_redirectstatusZ	incrementZsleep_for_retrylogdebugrurlopenboolZ	getheaderZis_retrysleep)r"r'Zurlr-r#r!r+r,response_kwr(r*eZ
http_responseZredirect_locationZredirect_urlZhas_retry_afterrrrr6s





zAppEngineManager.urlopencKstr"|jd}|dkr"|jd=|jd}|dkrZ|d}|dd||jd<tft|j	|j
|j|jd|}tft|j	|j|j|d|S)Nzcontent-encodingZdeflateztransfer-encodingZchunked,)r-msgr#r3)r-r#r3original_response)is_prod_appenginer#getsplitremovejoinrioBytesIOZcontentZ
header_msgZstatus_code)r"Z
urlfetch_respr9Zcontent_encodingZtransfer_encodingZ	encodingsr=rrrr2s2




z4AppEngineManager._urlfetch_response_to_http_responsecCsB|tjkrdSt|tr>|jdk	s,|jdk	r8tdt|jS|S)NzdURLFetch does not support granular timeout settings, reverting to total or default URLFetch timeout.)	r
DEFAULT_TIMEOUT
isinstanceZ_readZ_connectrrrr/)r"r,rrrr0 s

z&AppEngineManager._get_absolute_timeoutcCs>t|tstj|||jd}|js.|js.|jr:tdt	|S)N)r+defaultzhURLFetch only supports total retries and does not recognize connect, read, or redirect retry parameters.)
rFrZfrom_intr!Zconnectreadr+rrr)r"r!r+rrrr.-s
zAppEngineManager._get_retries)NNTT)
rrr__doc__rr$r&r
rEr6r2r0r.rrrrrOs"
 
r$
r)'rIZ
__future__rrCZloggingrZpackages.six.moves.urllib.parser
exceptionsrrrrr	r
Zrequestrr*rZutil.timeoutr
Z
util.retryrrZgoogle.appengine.apirImportErrorZ	getLoggerrr4rrrZis_appengineZis_appengine_sandboxZis_local_appenginer>rrrrrs2( 	

osite-packages/pip/_vendor/urllib3/contrib/__pycache__/ntlmpool.cpython-38.opt-1.pyc000064400000006260151464452710024204 0ustar00U

.ei@s\dZddlmZddlmZddlmZddlmZddlm	Z	ee
ZGdd	d	eZd
S)z
NTLM authenticating pool, contributed by erikcederstran

Issue #10, see: http://code.google.com/p/urllib3/issues/detail?id=10
)absolute_import)	getLogger)ntlm)HTTPSConnectionPool)HTTPSConnectioncs:eZdZdZdZfddZddZdfd
d	ZZS)
NTLMConnectionPoolzQ
    Implements an NTLM authentication version of an urllib3 connection pool
    ZhttpscsLtt|j||||_||_|dd}|d|_|d|_||_	dS)z
        authurl is a random URL on the server that is protected by NTLM.
        user is the Windows user, probably in the DOMAIN\username format.
        pw is the password for the user.
        \rN)
superr__init__authurlrawusersplitupperdomainuserpw)selfrrr
argskwargsZ
user_parts	__class__H/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/ntlmpool.pyrs
zNTLMConnectionPool.__init__c
Cs|jd7_td|j|j|jddi}d}d}t|j|jd}dt|j	||<td	||
d
|jd||}t|
}td|j|jtd|td
|dd|_||d}d}|D] }	|	dddkr|	dd}q|dkrtd|||ft|\}
}t|
|j|j|j|}d|||<td	||
d
|jd||}td|j|jtdt|
td
|dd|jdkr|jdkrtdtd|j|jfd|_td|S)Nr
z3Starting NTLM HTTPS connection no. %d: https://%s%s
Connection
Keep-AliveZ
Authorizationzwww-authenticate)hostportzNTLM %szRequest headers: %sZGETzResponse status: %s %szResponse headers: %szResponse data: %s [...]dz, zNTLM z!Unexpected %s response header: %siz3Server rejected request: wrong username or passwordzWrong server response: %s %szConnection established)Znum_connectionslogdebugrr
rrrZcreate_NTLM_NEGOTIATE_MESSAGErZrequestZgetresponsedictZ
getheadersZstatusreasonreadfpr	ExceptionZparse_NTLM_CHALLENGE_MESSAGEZ create_NTLM_AUTHENTICATE_MESSAGErrr)
rheadersZ
req_headerZresp_headerZconnresZreshdrZauth_header_valuesZauth_header_valuesZServerChallengeZNegotiateFlagsZauth_msgrrr	_new_conn'sr


zNTLMConnectionPool._new_connNTc	s0|dkri}d|d<tt||||||||S)Nrr)rrurlopen)rmethodZurlZbodyr)ZretriesZredirectZassert_same_hostrrrr.ls

zNTLMConnectionPool.urlopen)NNr-TT)	__name__
__module____qualname____doc__Zschemerr,r.
__classcell__rrrrrsIrN)
r3Z
__future__rZloggingrrrZpackages.six.moves.http_clientrr0r"rrrrrssite-packages/pip/_vendor/urllib3/contrib/__pycache__/pyopenssl.cpython-38.pyc000064400000035125151464452710023437 0ustar00U

.e8@@s\dZddlmZddlZddlmZddlmZ	ddl
mZzddlm
Z
Wn$ek
rpGdd	d	eZ
YnXdd
lmZmZddlmZzddlmZWn$ek
rdZd
dlmZYnXddlZddlZd
dlmZddlZd
dlmZddgZ dZ!ej"ej#j$ej%ej#j&iZ'e(edr@e(ej#dr@ej#j)e'ej*<e(edrhe(ej#drhej#j+e'ej,<e(edre(ej#drej#j-e'ej.<ej/ej#j0ej1ej#j2ej3ej#j2ej#j4iZ5e6dde57DZ8dZ9ej!Z:ej;je?Z@ddZAddZBdd ZCd!d"ZDd#d$ZEGd%d&d&eFZGer6d.d(d)ZHneZHeHeG_HGd*d+d+eFZId,d-ZJdS)/ab
SSL with SNI_-support for Python 2. Follow these instructions if you would
like to verify SSL certificates in Python 2. Note, the default libraries do
*not* do certificate checking; you need to do additional work to validate
certificates yourself.

This needs the following packages installed:

* pyOpenSSL (tested with 16.0.0)
* cryptography (minimum 1.3.4, from pyopenssl)
* idna (minimum 2.0, from cryptography)

However, pyopenssl depends on cryptography, which depends on idna, so while we
use all three directly here we end up having relatively few packages required.

You can install them with the following command:

    pip install pyopenssl cryptography idna

To activate certificate checking, call
:func:`~urllib3.contrib.pyopenssl.inject_into_urllib3` from your Python code
before you begin making HTTP requests. This can be done in a ``sitecustomize``
module, or at any other time before your application begins using ``urllib3``,
like this::

    try:
        import urllib3.contrib.pyopenssl
        urllib3.contrib.pyopenssl.inject_into_urllib3()
    except ImportError:
        pass

Now you can use :mod:`urllib3` as you normally would, and it will support SNI
when the required modules are installed.

Activating this module also has the positive side effect of disabling SSL/TLS
compression in Python 2 (see `CRIME attack`_).

If you want to configure the default list of supported cipher suites, you can
set the ``urllib3.contrib.pyopenssl.DEFAULT_SSL_CIPHER_LIST`` variable.

.. _sni: https://en.wikipedia.org/wiki/Server_Name_Indication
.. _crime attack: https://en.wikipedia.org/wiki/CRIME_(security_exploit)
)absolute_importN)x509)backend)_Certificate)UnsupportedExtensionc@seZdZdS)rN)__name__
__module____qualname__r
r
I/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/pyopenssl.pyr7sr)timeouterror)BytesIO)_fileobject)backport_makefile)six)utilinject_into_urllib3extract_from_urllib3TPROTOCOL_SSLv3SSLv3_METHODPROTOCOL_TLSv1_1TLSv1_1_METHODPROTOCOL_TLSv1_2TLSv1_2_METHODccs|]\}}||fVqdSNr
).0kvr
r
r	gsr i@cCs4ttt_ttj_tt_ttj_dt_dtj_dS)z7Monkey-patch urllib3 with PyOpenSSL-backed SSL-support.TN)_validate_dependencies_metPyOpenSSLContextr
SSLContextssl_HAS_SNIIS_PYOPENSSLr
r
r
rrsscCs.tt_ttj_tt_ttj_dt_dtj_dS)z4Undo monkey-patching by :func:`inject_into_urllib3`.FN)orig_util_SSLContextrr#r$orig_util_HAS_SNIr%r&r
r
r
rrscCsRddlm}t|dddkr$tdddlm}|}t|dddkrNtddS)	z{
    Verifies that PyOpenSSL's package-level dependencies have been met.
    Throws `ImportError` if they are not met.
    r)
Extensionsget_extension_for_classNzX'cryptography' module missing required functionality.  Try upgrading to v1.3.4 or newer.)X509_x509zS'pyOpenSSL' module missing required functionality. Try upgrading to v0.14 or newer.)Zcryptography.x509.extensionsr)getattrImportErrorZOpenSSL.cryptor+)r)r+rr
r
rr!sr!cCs@dd}d|kr|S||}|dkr(dStjdkr<|d}|S)a%
    Converts a dNSName SubjectAlternativeName field to the form used by the
    standard library on the given Python version.

    Cryptography produces a dNSName as a unicode string that was idna-decoded
    from ASCII bytes. We need to idna-encode that string to get it back, and
    then on Python 3 we also need to convert to unicode via UTF-8 (the stdlib
    uses PyUnicode_FromStringAndSize on it, which decodes via UTF-8).

    If the name cannot be idna-encoded then we return None signalling that
    the name given should be skipped.
    cSsvddlm}zJdD]8}||r|t|d}|d||WSq||WS|jjk
rpYdSXdS)z
        Borrowed wholesale from the Python Cryptography Project. It turns out
        that we can't just safely call `idna.encode`: it can explode for
        wildcard names. This avoids that problem.
        r)idna)z*..Nascii)Zpip._vendorr/
startswithlenencodeZcoreZ	IDNAError)namer/prefixr
r
ridna_encodes
z'_dnsname_to_stdlib..idna_encode:N)rutf-8)sysversion_infodecode)r5r7r
r
r_dnsname_to_stdlibs

r>cCst|dr|}ntt|j}z|jtjj	}WnZtj
k
rNgYStjttj
tfk
r}ztd|gWYSd}~XYnXddtt|tjD}|dd|tjD|S)zU
    Given an PyOpenSSL certificate, provides all the subject alternative names.
    to_cryptographyzA problem was encountered with the certificate that prevented urllib3 from finding the SubjectAlternativeName field. This can affect certificate validation. The error was %sNcSsg|]}|dk	rd|fqS)NZDNSr
rr5r
r
r
sz%get_subj_alt_name..css|]}dt|fVqdS)z
IP AddressN)strr@r
r
rr sz$get_subj_alt_name..)hasattrr?ropenssl_backendr,
extensionsr*rZSubjectAlternativeNamevalueZExtensionNotFoundZDuplicateExtensionrZUnsupportedGeneralNameTypeUnicodeErrorlogZwarningmapr>Zget_values_for_typeZDNSNameextendZ	IPAddress)Z	peer_certZcertZextenamesr
r
rget_subj_alt_names2

	

rMc@seZdZdZd!ddZddZddZd	d
ZddZd
dZ	ddZ
ddZddZddZ
d"ddZddZddZddZd S)#
WrappedSocketzAPI-compatibility wrapper for Python OpenSSL's Connection-class.

    Note: _makefile_refs, _drop() and _reuse() are needed for the garbage
    collector of pypy.
    TcCs"||_||_||_d|_d|_dSNrF)
connectionsocketsuppress_ragged_eofs_makefile_refs_closed)selfrPrQrRr
r
r__init__
s
zWrappedSocket.__init__cCs
|jSr)rQfilenorUr
r
rrWszWrappedSocket.filenocCs*|jdkr|jd8_|jr&|dS)Nr)rSrTcloserXr
r
r_decref_socketioss
zWrappedSocket._decref_socketiosc
Osz|jj||}Wntjjk
r^}z*|jrB|jdkrBWYdStt|W5d}~XYntjj	k
r|j
tjjkrYdSYn~tjjk
rt
|j|jstdn|j||YSYn:tjjk
r}ztd|W5d}~XYnX|SdS)NzUnexpected EOFThe read operation timed outread error: %r)rPrecvOpenSSLSSLSysCallErrorrRargsSocketErrorrBZeroReturnErrorget_shutdownRECEIVED_SHUTDOWN
WantReadErrorr
wait_for_readrQ
gettimeoutrErrorsslSSLError)rUrekwargsdatarKr
r
rras"

 zWrappedSocket.recvc
Os
z|jj||WStjjk
r\}z*|jr@|jdkr@WYdStt|W5d}~XYntjj	k
r|j
tjjkrYdSYnztjjk
rt
|j|jstdn|j||YSYn6tjjk
r}ztd|W5d}~XYnXdS)Nr\rr_r`)rP	recv_intorbrcrdrRrerfrBrgrhrirjrrkrQrlrrmrnro)rUrerprKr
r
rrr7s 

zWrappedSocket.recv_intocCs|j|Sr)rQ
settimeout)rUrr
r
rrsNszWrappedSocket.settimeoutc
Cs|z|j|WStjjk
rDt|j|jsttjs*d|jfdd|j|pJ|dS)Nr:csSrr
)_passwordr
rr^z2PyOpenSSLContext.load_cert_chain..)rZuse_certificate_chain_filerrZbinary_typer4Z
set_passwd_cbZuse_privatekey_file)rUZcertfileZkeyfilerr
rrload_cert_chains
z PyOpenSSLContext.load_cert_chainFTc
Cstj|j|}t|tjr&|d}|dk	r8|||	z|
Wqtjjk
rt
||sxtdYq@Yqtjjk
r}ztd|W5d}~XYqXqq@t||S)Nr:zselect timed outzbad handshake: %r)rbrcZ
Connectionrrrrr4Zset_tlsext_host_nameZset_connect_stateZdo_handshakerjrrkrlrrmrnrorN)rUZsockZserver_sideZdo_handshake_on_connectrRZserver_hostnamecnxrKr
r
rwrap_sockets 

 zPyOpenSSLContext.wrap_socket)NNN)NN)FTTN)rrr	rrVpropertyrsetterrrrrrrr
r
r
rr"s&




	
r"cCs|dkSrvr
)rrZerr_noZ	err_depthZreturn_coder
r
rrsr)r])KrZ
__future__rZOpenSSL.SSLrbZcryptographyrZ$cryptography.hazmat.backends.opensslrrDZ)cryptography.hazmat.backends.openssl.x509rZcryptography.x509rr.	ExceptionrQrr
rfiorrZpackages.backports.makefilerZloggingrnZpackagesrr;r__all__r%ZPROTOCOL_TLSrcZ
SSLv23_METHODZPROTOCOL_TLSv1ZTLSv1_METHODrrCrrrrrrZ	CERT_NONEZVERIFY_NONEZ
CERT_OPTIONALZVERIFY_PEERZ
CERT_REQUIREDZVERIFY_FAIL_IF_NO_PEER_CERTrdictitemsrrwr(r$r#r'Z	getLoggerrrHrrr!r>rMobjectrNrr"rr
r
r
rsz+

+6Wsite-packages/pip/_vendor/urllib3/contrib/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000243151464452710024072 0ustar00U

.e@sdS)NrrrH/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/__init__.pysite-packages/pip/_vendor/urllib3/contrib/__pycache__/socks.cpython-38.pyc000064400000012653151464452710022526 0ustar00U

.e|@s(dZddlmZzddlZWn6ek
rRddlZddlmZedeYnXddl	m
ZmZ
ddlmZmZdd	lmZmZdd
lmZmZddlmZddlmZzddlZWnek
rdZYnXGd
ddeZGdddeeZGdddeZGdddeZGdddeZdS)a_
This module contains provisional support for SOCKS proxies from within
urllib3. This module supports SOCKS4, SOCKS4A (an extension of SOCKS4), and
SOCKS5. To enable its functionality, either install PySocks or install this
module with the ``socks`` extra.

The SOCKS implementation supports the full range of urllib3 features. It also
supports the following SOCKS features:

- SOCKS4A (``proxy_url='socks4a://...``)
- SOCKS4 (``proxy_url='socks4://...``)
- SOCKS5 with remote DNS (``proxy_url='socks5h://...``)
- SOCKS5 with local DNS (``proxy_url='socks5://...``)
- Usernames and passwords for the SOCKS proxy

 .. note::
    It is recommended to use ``socks5h://`` or ``socks4a://`` schemes in
    your ``proxy_url`` to ensure that DNS resolution is done from the remote
    server instead of client-side when connecting to a domain name.

SOCKS4 supports IPv4 and domain names with the SOCKS4A extension. SOCKS5
supports IPv4, IPv6, and domain names.

When connecting to a SOCKS4 proxy the ``username`` portion of the ``proxy_url``
will be sent as the ``userid`` section of the SOCKS request::

    proxy_url="socks4a://@proxy-host"

When connecting to a SOCKS5 proxy the ``username`` and ``password`` portion
of the ``proxy_url`` will be sent as the username/password to authenticate
with the proxy::

    proxy_url="socks5h://:@proxy-host"

)absolute_importN)DependencyWarningzSOCKS support in urllib3 requires the installation of optional dependencies: specifically, PySocks.  For more information, see https://urllib3.readthedocs.io/en/latest/contrib.html#socks-proxies)errortimeout)HTTPConnectionHTTPSConnection)HTTPConnectionPoolHTTPSConnectionPool)ConnectTimeoutErrorNewConnectionError)PoolManager)	parse_urlcs(eZdZdZfddZddZZS)SOCKSConnectionzG
    A plain-text HTTP connection that connects via a SOCKS proxy.
    cs"|d|_tt|j||dS)N_socks_options)poprsuperr__init__)selfargskwargs	__class__E/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/socks.pyrJszSOCKSConnection.__init__c
CsHi}|jr|j|d<|jr$|j|d<zTtj|j|jff|jd|jd|jd|jd|jd|jd|jd	|}Wntk
rt	|d
|j|jfYntj
k
r}zP|jr|j}t|trt	|d
|j|jfnt
|d|nt
|d|W5d}~XYn2tk
rB}zt
|d|W5d}~XYnX|S)
zA
        Establish a new connection via the SOCKS proxy.
        source_addresssocket_options
socks_version
proxy_host
proxy_portusernamepasswordrdns)Z
proxy_typeZ
proxy_addrrZproxy_usernameZproxy_passwordZ
proxy_rdnsrz0Connection to %s timed out. (connect timeout=%s)z(Failed to establish a new connection: %sN)rrsocksZcreate_connectionhostportrr
SocketTimeoutrZ
ProxyErrorZ
socket_err
isinstancerSocketError)rZextra_kwZconnerrrr	_new_connNsd


	



zSOCKSConnection._new_conn)__name__
__module____qualname____doc__rr*
__classcell__rrrrrEsrc@seZdZdS)SOCKSHTTPSConnectionN)r+r,r-rrrrr0sr0c@seZdZeZdS)SOCKSHTTPConnectionPoolN)r+r,r-r
ConnectionClsrrrrr1sr1c@seZdZeZdS)SOCKSHTTPSConnectionPoolN)r+r,r-r0r2rrrrr3sr3cs,eZdZdZeedZdfdd	ZZS)SOCKSProxyManagerzh
    A version of the urllib3 ProxyManager that routes connections via the
    defined SOCKS proxy.
    )ZhttpZhttpsN
cst|}|dkrB|dkrB|jdk	rB|jd}t|dkrB|\}}|jdkrXtj}	d}
nN|jdkrntj}	d}
n8|jdkrtj}	d}
n"|jdkrtj}	d}
ntd	|||_	|	|j
|j|||
d
}||d<tt
|j||f|t
j|_dS)N:rZsocks5FZsocks5hTZsocks4Zsocks4az)Unable to determine SOCKS version from %s)rrrr r!r"r)rZauthsplitlenZschemer#ZPROXY_TYPE_SOCKS5ZPROXY_TYPE_SOCKS4
ValueError	proxy_urlr$r%rr4rpool_classes_by_scheme)rr:r r!Z	num_poolsZheadersZconnection_pool_kwZparsedr7rr"Z
socks_optionsrrrrsD	




zSOCKSProxyManager.__init__)NNr5N)	r+r,r-r.r1r3r;rr/rrrrr4sr4) r.Z
__future__rr#ImportErrorwarnings
exceptionsrwarnZsocketrr(rr&Z
connectionrrZconnectionpoolr	r
rrZpoolmanagerr
Zutil.urlrZsslrr0r1r3r4rrrrs4#
Hsite-packages/pip/_vendor/urllib3/contrib/__pycache__/socks.cpython-38.opt-1.pyc000064400000012653151464452710023465 0ustar00U

.e|@s(dZddlmZzddlZWn6ek
rRddlZddlmZedeYnXddl	m
ZmZ
ddlmZmZdd	lmZmZdd
lmZmZddlmZddlmZzddlZWnek
rdZYnXGd
ddeZGdddeeZGdddeZGdddeZGdddeZdS)a_
This module contains provisional support for SOCKS proxies from within
urllib3. This module supports SOCKS4, SOCKS4A (an extension of SOCKS4), and
SOCKS5. To enable its functionality, either install PySocks or install this
module with the ``socks`` extra.

The SOCKS implementation supports the full range of urllib3 features. It also
supports the following SOCKS features:

- SOCKS4A (``proxy_url='socks4a://...``)
- SOCKS4 (``proxy_url='socks4://...``)
- SOCKS5 with remote DNS (``proxy_url='socks5h://...``)
- SOCKS5 with local DNS (``proxy_url='socks5://...``)
- Usernames and passwords for the SOCKS proxy

 .. note::
    It is recommended to use ``socks5h://`` or ``socks4a://`` schemes in
    your ``proxy_url`` to ensure that DNS resolution is done from the remote
    server instead of client-side when connecting to a domain name.

SOCKS4 supports IPv4 and domain names with the SOCKS4A extension. SOCKS5
supports IPv4, IPv6, and domain names.

When connecting to a SOCKS4 proxy the ``username`` portion of the ``proxy_url``
will be sent as the ``userid`` section of the SOCKS request::

    proxy_url="socks4a://@proxy-host"

When connecting to a SOCKS5 proxy the ``username`` and ``password`` portion
of the ``proxy_url`` will be sent as the username/password to authenticate
with the proxy::

    proxy_url="socks5h://:@proxy-host"

)absolute_importN)DependencyWarningzSOCKS support in urllib3 requires the installation of optional dependencies: specifically, PySocks.  For more information, see https://urllib3.readthedocs.io/en/latest/contrib.html#socks-proxies)errortimeout)HTTPConnectionHTTPSConnection)HTTPConnectionPoolHTTPSConnectionPool)ConnectTimeoutErrorNewConnectionError)PoolManager)	parse_urlcs(eZdZdZfddZddZZS)SOCKSConnectionzG
    A plain-text HTTP connection that connects via a SOCKS proxy.
    cs"|d|_tt|j||dS)N_socks_options)poprsuperr__init__)selfargskwargs	__class__E/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/socks.pyrJszSOCKSConnection.__init__c
CsHi}|jr|j|d<|jr$|j|d<zTtj|j|jff|jd|jd|jd|jd|jd|jd|jd	|}Wntk
rt	|d
|j|jfYntj
k
r}zP|jr|j}t|trt	|d
|j|jfnt
|d|nt
|d|W5d}~XYn2tk
rB}zt
|d|W5d}~XYnX|S)
zA
        Establish a new connection via the SOCKS proxy.
        source_addresssocket_options
socks_version
proxy_host
proxy_portusernamepasswordrdns)Z
proxy_typeZ
proxy_addrrZproxy_usernameZproxy_passwordZ
proxy_rdnsrz0Connection to %s timed out. (connect timeout=%s)z(Failed to establish a new connection: %sN)rrsocksZcreate_connectionhostportrr
SocketTimeoutrZ
ProxyErrorZ
socket_err
isinstancerSocketError)rZextra_kwZconnerrrr	_new_connNsd


	



zSOCKSConnection._new_conn)__name__
__module____qualname____doc__rr*
__classcell__rrrrrEsrc@seZdZdS)SOCKSHTTPSConnectionN)r+r,r-rrrrr0sr0c@seZdZeZdS)SOCKSHTTPConnectionPoolN)r+r,r-r
ConnectionClsrrrrr1sr1c@seZdZeZdS)SOCKSHTTPSConnectionPoolN)r+r,r-r0r2rrrrr3sr3cs,eZdZdZeedZdfdd	ZZS)SOCKSProxyManagerzh
    A version of the urllib3 ProxyManager that routes connections via the
    defined SOCKS proxy.
    )ZhttpZhttpsN
cst|}|dkrB|dkrB|jdk	rB|jd}t|dkrB|\}}|jdkrXtj}	d}
nN|jdkrntj}	d}
n8|jdkrtj}	d}
n"|jdkrtj}	d}
ntd	|||_	|	|j
|j|||
d
}||d<tt
|j||f|t
j|_dS)N:rZsocks5FZsocks5hTZsocks4Zsocks4az)Unable to determine SOCKS version from %s)rrrr r!r"r)rZauthsplitlenZschemer#ZPROXY_TYPE_SOCKS5ZPROXY_TYPE_SOCKS4
ValueError	proxy_urlr$r%rr4rpool_classes_by_scheme)rr:r r!Z	num_poolsZheadersZconnection_pool_kwZparsedr7rr"Z
socks_optionsrrrrsD	




zSOCKSProxyManager.__init__)NNr5N)	r+r,r-r.r1r3r;rr/rrrrr4sr4) r.Z
__future__rr#ImportErrorwarnings
exceptionsrwarnZsocketrr(rr&Z
connectionrrZconnectionpoolr	r
rrZpoolmanagerr
Zutil.urlrZsslrr0r1r3r4rrrrs4#
Hsite-packages/pip/_vendor/urllib3/contrib/_securetransport/bindings.py000064400000040645151464452710022317 0ustar00"""
This module uses ctypes to bind a whole bunch of functions and constants from
SecureTransport. The goal here is to provide the low-level API to
SecureTransport. These are essentially the C-level functions and constants, and
they're pretty gross to work with.

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

    Copyright (c) 2015-2016 Will Bond 

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
"""
from __future__ import absolute_import

import platform
from ctypes.util import find_library
from ctypes import (
    c_void_p,
    c_int32,
    c_char_p,
    c_size_t,
    c_byte,
    c_uint32,
    c_ulong,
    c_long,
    c_bool,
)
from ctypes import CDLL, POINTER, CFUNCTYPE


security_path = find_library("Security")
if not security_path:
    raise ImportError("The library Security could not be found")


core_foundation_path = find_library("CoreFoundation")
if not core_foundation_path:
    raise ImportError("The library CoreFoundation could not be found")


version = platform.mac_ver()[0]
version_info = tuple(map(int, version.split(".")))
if version_info < (10, 8):
    raise OSError(
        "Only OS X 10.8 and newer are supported, not %s.%s"
        % (version_info[0], version_info[1])
    )

Security = CDLL(security_path, use_errno=True)
CoreFoundation = CDLL(core_foundation_path, use_errno=True)

Boolean = c_bool
CFIndex = c_long
CFStringEncoding = c_uint32
CFData = c_void_p
CFString = c_void_p
CFArray = c_void_p
CFMutableArray = c_void_p
CFDictionary = c_void_p
CFError = c_void_p
CFType = c_void_p
CFTypeID = c_ulong

CFTypeRef = POINTER(CFType)
CFAllocatorRef = c_void_p

OSStatus = c_int32

CFDataRef = POINTER(CFData)
CFStringRef = POINTER(CFString)
CFArrayRef = POINTER(CFArray)
CFMutableArrayRef = POINTER(CFMutableArray)
CFDictionaryRef = POINTER(CFDictionary)
CFArrayCallBacks = c_void_p
CFDictionaryKeyCallBacks = c_void_p
CFDictionaryValueCallBacks = c_void_p

SecCertificateRef = POINTER(c_void_p)
SecExternalFormat = c_uint32
SecExternalItemType = c_uint32
SecIdentityRef = POINTER(c_void_p)
SecItemImportExportFlags = c_uint32
SecItemImportExportKeyParameters = c_void_p
SecKeychainRef = POINTER(c_void_p)
SSLProtocol = c_uint32
SSLCipherSuite = c_uint32
SSLContextRef = POINTER(c_void_p)
SecTrustRef = POINTER(c_void_p)
SSLConnectionRef = c_uint32
SecTrustResultType = c_uint32
SecTrustOptionFlags = c_uint32
SSLProtocolSide = c_uint32
SSLConnectionType = c_uint32
SSLSessionOption = c_uint32


try:
    Security.SecItemImport.argtypes = [
        CFDataRef,
        CFStringRef,
        POINTER(SecExternalFormat),
        POINTER(SecExternalItemType),
        SecItemImportExportFlags,
        POINTER(SecItemImportExportKeyParameters),
        SecKeychainRef,
        POINTER(CFArrayRef),
    ]
    Security.SecItemImport.restype = OSStatus

    Security.SecCertificateGetTypeID.argtypes = []
    Security.SecCertificateGetTypeID.restype = CFTypeID

    Security.SecIdentityGetTypeID.argtypes = []
    Security.SecIdentityGetTypeID.restype = CFTypeID

    Security.SecKeyGetTypeID.argtypes = []
    Security.SecKeyGetTypeID.restype = CFTypeID

    Security.SecCertificateCreateWithData.argtypes = [CFAllocatorRef, CFDataRef]
    Security.SecCertificateCreateWithData.restype = SecCertificateRef

    Security.SecCertificateCopyData.argtypes = [SecCertificateRef]
    Security.SecCertificateCopyData.restype = CFDataRef

    Security.SecCopyErrorMessageString.argtypes = [OSStatus, c_void_p]
    Security.SecCopyErrorMessageString.restype = CFStringRef

    Security.SecIdentityCreateWithCertificate.argtypes = [
        CFTypeRef,
        SecCertificateRef,
        POINTER(SecIdentityRef),
    ]
    Security.SecIdentityCreateWithCertificate.restype = OSStatus

    Security.SecKeychainCreate.argtypes = [
        c_char_p,
        c_uint32,
        c_void_p,
        Boolean,
        c_void_p,
        POINTER(SecKeychainRef),
    ]
    Security.SecKeychainCreate.restype = OSStatus

    Security.SecKeychainDelete.argtypes = [SecKeychainRef]
    Security.SecKeychainDelete.restype = OSStatus

    Security.SecPKCS12Import.argtypes = [
        CFDataRef,
        CFDictionaryRef,
        POINTER(CFArrayRef),
    ]
    Security.SecPKCS12Import.restype = OSStatus

    SSLReadFunc = CFUNCTYPE(OSStatus, SSLConnectionRef, c_void_p, POINTER(c_size_t))
    SSLWriteFunc = CFUNCTYPE(
        OSStatus, SSLConnectionRef, POINTER(c_byte), POINTER(c_size_t)
    )

    Security.SSLSetIOFuncs.argtypes = [SSLContextRef, SSLReadFunc, SSLWriteFunc]
    Security.SSLSetIOFuncs.restype = OSStatus

    Security.SSLSetPeerID.argtypes = [SSLContextRef, c_char_p, c_size_t]
    Security.SSLSetPeerID.restype = OSStatus

    Security.SSLSetCertificate.argtypes = [SSLContextRef, CFArrayRef]
    Security.SSLSetCertificate.restype = OSStatus

    Security.SSLSetCertificateAuthorities.argtypes = [SSLContextRef, CFTypeRef, Boolean]
    Security.SSLSetCertificateAuthorities.restype = OSStatus

    Security.SSLSetConnection.argtypes = [SSLContextRef, SSLConnectionRef]
    Security.SSLSetConnection.restype = OSStatus

    Security.SSLSetPeerDomainName.argtypes = [SSLContextRef, c_char_p, c_size_t]
    Security.SSLSetPeerDomainName.restype = OSStatus

    Security.SSLHandshake.argtypes = [SSLContextRef]
    Security.SSLHandshake.restype = OSStatus

    Security.SSLRead.argtypes = [SSLContextRef, c_char_p, c_size_t, POINTER(c_size_t)]
    Security.SSLRead.restype = OSStatus

    Security.SSLWrite.argtypes = [SSLContextRef, c_char_p, c_size_t, POINTER(c_size_t)]
    Security.SSLWrite.restype = OSStatus

    Security.SSLClose.argtypes = [SSLContextRef]
    Security.SSLClose.restype = OSStatus

    Security.SSLGetNumberSupportedCiphers.argtypes = [SSLContextRef, POINTER(c_size_t)]
    Security.SSLGetNumberSupportedCiphers.restype = OSStatus

    Security.SSLGetSupportedCiphers.argtypes = [
        SSLContextRef,
        POINTER(SSLCipherSuite),
        POINTER(c_size_t),
    ]
    Security.SSLGetSupportedCiphers.restype = OSStatus

    Security.SSLSetEnabledCiphers.argtypes = [
        SSLContextRef,
        POINTER(SSLCipherSuite),
        c_size_t,
    ]
    Security.SSLSetEnabledCiphers.restype = OSStatus

    Security.SSLGetNumberEnabledCiphers.argtype = [SSLContextRef, POINTER(c_size_t)]
    Security.SSLGetNumberEnabledCiphers.restype = OSStatus

    Security.SSLGetEnabledCiphers.argtypes = [
        SSLContextRef,
        POINTER(SSLCipherSuite),
        POINTER(c_size_t),
    ]
    Security.SSLGetEnabledCiphers.restype = OSStatus

    Security.SSLGetNegotiatedCipher.argtypes = [SSLContextRef, POINTER(SSLCipherSuite)]
    Security.SSLGetNegotiatedCipher.restype = OSStatus

    Security.SSLGetNegotiatedProtocolVersion.argtypes = [
        SSLContextRef,
        POINTER(SSLProtocol),
    ]
    Security.SSLGetNegotiatedProtocolVersion.restype = OSStatus

    Security.SSLCopyPeerTrust.argtypes = [SSLContextRef, POINTER(SecTrustRef)]
    Security.SSLCopyPeerTrust.restype = OSStatus

    Security.SecTrustSetAnchorCertificates.argtypes = [SecTrustRef, CFArrayRef]
    Security.SecTrustSetAnchorCertificates.restype = OSStatus

    Security.SecTrustSetAnchorCertificatesOnly.argstypes = [SecTrustRef, Boolean]
    Security.SecTrustSetAnchorCertificatesOnly.restype = OSStatus

    Security.SecTrustEvaluate.argtypes = [SecTrustRef, POINTER(SecTrustResultType)]
    Security.SecTrustEvaluate.restype = OSStatus

    Security.SecTrustGetCertificateCount.argtypes = [SecTrustRef]
    Security.SecTrustGetCertificateCount.restype = CFIndex

    Security.SecTrustGetCertificateAtIndex.argtypes = [SecTrustRef, CFIndex]
    Security.SecTrustGetCertificateAtIndex.restype = SecCertificateRef

    Security.SSLCreateContext.argtypes = [
        CFAllocatorRef,
        SSLProtocolSide,
        SSLConnectionType,
    ]
    Security.SSLCreateContext.restype = SSLContextRef

    Security.SSLSetSessionOption.argtypes = [SSLContextRef, SSLSessionOption, Boolean]
    Security.SSLSetSessionOption.restype = OSStatus

    Security.SSLSetProtocolVersionMin.argtypes = [SSLContextRef, SSLProtocol]
    Security.SSLSetProtocolVersionMin.restype = OSStatus

    Security.SSLSetProtocolVersionMax.argtypes = [SSLContextRef, SSLProtocol]
    Security.SSLSetProtocolVersionMax.restype = OSStatus

    Security.SecCopyErrorMessageString.argtypes = [OSStatus, c_void_p]
    Security.SecCopyErrorMessageString.restype = CFStringRef

    Security.SSLReadFunc = SSLReadFunc
    Security.SSLWriteFunc = SSLWriteFunc
    Security.SSLContextRef = SSLContextRef
    Security.SSLProtocol = SSLProtocol
    Security.SSLCipherSuite = SSLCipherSuite
    Security.SecIdentityRef = SecIdentityRef
    Security.SecKeychainRef = SecKeychainRef
    Security.SecTrustRef = SecTrustRef
    Security.SecTrustResultType = SecTrustResultType
    Security.SecExternalFormat = SecExternalFormat
    Security.OSStatus = OSStatus

    Security.kSecImportExportPassphrase = CFStringRef.in_dll(
        Security, "kSecImportExportPassphrase"
    )
    Security.kSecImportItemIdentity = CFStringRef.in_dll(
        Security, "kSecImportItemIdentity"
    )

    # CoreFoundation time!
    CoreFoundation.CFRetain.argtypes = [CFTypeRef]
    CoreFoundation.CFRetain.restype = CFTypeRef

    CoreFoundation.CFRelease.argtypes = [CFTypeRef]
    CoreFoundation.CFRelease.restype = None

    CoreFoundation.CFGetTypeID.argtypes = [CFTypeRef]
    CoreFoundation.CFGetTypeID.restype = CFTypeID

    CoreFoundation.CFStringCreateWithCString.argtypes = [
        CFAllocatorRef,
        c_char_p,
        CFStringEncoding,
    ]
    CoreFoundation.CFStringCreateWithCString.restype = CFStringRef

    CoreFoundation.CFStringGetCStringPtr.argtypes = [CFStringRef, CFStringEncoding]
    CoreFoundation.CFStringGetCStringPtr.restype = c_char_p

    CoreFoundation.CFStringGetCString.argtypes = [
        CFStringRef,
        c_char_p,
        CFIndex,
        CFStringEncoding,
    ]
    CoreFoundation.CFStringGetCString.restype = c_bool

    CoreFoundation.CFDataCreate.argtypes = [CFAllocatorRef, c_char_p, CFIndex]
    CoreFoundation.CFDataCreate.restype = CFDataRef

    CoreFoundation.CFDataGetLength.argtypes = [CFDataRef]
    CoreFoundation.CFDataGetLength.restype = CFIndex

    CoreFoundation.CFDataGetBytePtr.argtypes = [CFDataRef]
    CoreFoundation.CFDataGetBytePtr.restype = c_void_p

    CoreFoundation.CFDictionaryCreate.argtypes = [
        CFAllocatorRef,
        POINTER(CFTypeRef),
        POINTER(CFTypeRef),
        CFIndex,
        CFDictionaryKeyCallBacks,
        CFDictionaryValueCallBacks,
    ]
    CoreFoundation.CFDictionaryCreate.restype = CFDictionaryRef

    CoreFoundation.CFDictionaryGetValue.argtypes = [CFDictionaryRef, CFTypeRef]
    CoreFoundation.CFDictionaryGetValue.restype = CFTypeRef

    CoreFoundation.CFArrayCreate.argtypes = [
        CFAllocatorRef,
        POINTER(CFTypeRef),
        CFIndex,
        CFArrayCallBacks,
    ]
    CoreFoundation.CFArrayCreate.restype = CFArrayRef

    CoreFoundation.CFArrayCreateMutable.argtypes = [
        CFAllocatorRef,
        CFIndex,
        CFArrayCallBacks,
    ]
    CoreFoundation.CFArrayCreateMutable.restype = CFMutableArrayRef

    CoreFoundation.CFArrayAppendValue.argtypes = [CFMutableArrayRef, c_void_p]
    CoreFoundation.CFArrayAppendValue.restype = None

    CoreFoundation.CFArrayGetCount.argtypes = [CFArrayRef]
    CoreFoundation.CFArrayGetCount.restype = CFIndex

    CoreFoundation.CFArrayGetValueAtIndex.argtypes = [CFArrayRef, CFIndex]
    CoreFoundation.CFArrayGetValueAtIndex.restype = c_void_p

    CoreFoundation.kCFAllocatorDefault = CFAllocatorRef.in_dll(
        CoreFoundation, "kCFAllocatorDefault"
    )
    CoreFoundation.kCFTypeArrayCallBacks = c_void_p.in_dll(
        CoreFoundation, "kCFTypeArrayCallBacks"
    )
    CoreFoundation.kCFTypeDictionaryKeyCallBacks = c_void_p.in_dll(
        CoreFoundation, "kCFTypeDictionaryKeyCallBacks"
    )
    CoreFoundation.kCFTypeDictionaryValueCallBacks = c_void_p.in_dll(
        CoreFoundation, "kCFTypeDictionaryValueCallBacks"
    )

    CoreFoundation.CFTypeRef = CFTypeRef
    CoreFoundation.CFArrayRef = CFArrayRef
    CoreFoundation.CFStringRef = CFStringRef
    CoreFoundation.CFDictionaryRef = CFDictionaryRef

except (AttributeError):
    raise ImportError("Error initializing ctypes")


class CFConst(object):
    """
    A class object that acts as essentially a namespace for CoreFoundation
    constants.
    """

    kCFStringEncodingUTF8 = CFStringEncoding(0x08000100)


class SecurityConst(object):
    """
    A class object that acts as essentially a namespace for Security constants.
    """

    kSSLSessionOptionBreakOnServerAuth = 0

    kSSLProtocol2 = 1
    kSSLProtocol3 = 2
    kTLSProtocol1 = 4
    kTLSProtocol11 = 7
    kTLSProtocol12 = 8
    kTLSProtocol13 = 10
    kTLSProtocolMaxSupported = 999

    kSSLClientSide = 1
    kSSLStreamType = 0

    kSecFormatPEMSequence = 10

    kSecTrustResultInvalid = 0
    kSecTrustResultProceed = 1
    # This gap is present on purpose: this was kSecTrustResultConfirm, which
    # is deprecated.
    kSecTrustResultDeny = 3
    kSecTrustResultUnspecified = 4
    kSecTrustResultRecoverableTrustFailure = 5
    kSecTrustResultFatalTrustFailure = 6
    kSecTrustResultOtherError = 7

    errSSLProtocol = -9800
    errSSLWouldBlock = -9803
    errSSLClosedGraceful = -9805
    errSSLClosedNoNotify = -9816
    errSSLClosedAbort = -9806

    errSSLXCertChainInvalid = -9807
    errSSLCrypto = -9809
    errSSLInternal = -9810
    errSSLCertExpired = -9814
    errSSLCertNotYetValid = -9815
    errSSLUnknownRootCert = -9812
    errSSLNoRootCert = -9813
    errSSLHostNameMismatch = -9843
    errSSLPeerHandshakeFail = -9824
    errSSLPeerUserCancelled = -9839
    errSSLWeakPeerEphemeralDHKey = -9850
    errSSLServerAuthCompleted = -9841
    errSSLRecordOverflow = -9847

    errSecVerifyFailed = -67808
    errSecNoTrustSettings = -25263
    errSecItemNotFound = -25300
    errSecInvalidTrustSettings = -25262

    # Cipher suites. We only pick the ones our default cipher string allows.
    # Source: https://developer.apple.com/documentation/security/1550981-ssl_cipher_suite_values
    TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 = 0xC02C
    TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 = 0xC030
    TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 = 0xC02B
    TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 = 0xC02F
    TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 = 0xCCA9
    TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 = 0xCCA8
    TLS_DHE_RSA_WITH_AES_256_GCM_SHA384 = 0x009F
    TLS_DHE_RSA_WITH_AES_128_GCM_SHA256 = 0x009E
    TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384 = 0xC024
    TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384 = 0xC028
    TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA = 0xC00A
    TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA = 0xC014
    TLS_DHE_RSA_WITH_AES_256_CBC_SHA256 = 0x006B
    TLS_DHE_RSA_WITH_AES_256_CBC_SHA = 0x0039
    TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256 = 0xC023
    TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 = 0xC027
    TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA = 0xC009
    TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA = 0xC013
    TLS_DHE_RSA_WITH_AES_128_CBC_SHA256 = 0x0067
    TLS_DHE_RSA_WITH_AES_128_CBC_SHA = 0x0033
    TLS_RSA_WITH_AES_256_GCM_SHA384 = 0x009D
    TLS_RSA_WITH_AES_128_GCM_SHA256 = 0x009C
    TLS_RSA_WITH_AES_256_CBC_SHA256 = 0x003D
    TLS_RSA_WITH_AES_128_CBC_SHA256 = 0x003C
    TLS_RSA_WITH_AES_256_CBC_SHA = 0x0035
    TLS_RSA_WITH_AES_128_CBC_SHA = 0x002F
    TLS_AES_128_GCM_SHA256 = 0x1301
    TLS_AES_256_GCM_SHA384 = 0x1302
    TLS_AES_128_CCM_8_SHA256 = 0x1305
    TLS_AES_128_CCM_SHA256 = 0x1304
site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/__init__.cpython-38.pyc000064400000000264151464452710026540 0ustar00U

.e@sdS)NrrrY/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_securetransport/__init__.pypip/_vendor/urllib3/contrib/_securetransport/__pycache__/low_level.cpython-38.opt-1.pyc000064400000016540151464452710027655 0ustar00site-packagesU

.e.@sdZddlZddlZddlZddlZddlZddlZddlZddlm	Z	m
Z
mZedej
ZddZdd	Zd
dZddd
ZddZddZddZddZddZddZdS)a
Low-level helpers for the SecureTransport bindings.

These are Python functions that are not directly related to the high-level APIs
but are necessary to get them to work. They include a whole bunch of low-level
CoreFoundation messing about and memory management. The concerns in this module
are almost entirely about trying to avoid memory leaks and providing
appropriate and useful assistance to the higher-level code.
N)SecurityCoreFoundationCFConsts;-----BEGIN CERTIFICATE-----
(.*?)
-----END CERTIFICATE-----cCsttj|t|S)zv
    Given a bytestring, create a CFData object from it. This CFData object must
    be CFReleased by the caller.
    )rCFDataCreatekCFAllocatorDefaultlen)Z
bytestringr	Z/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_securetransport/low_level.py_cf_data_from_bytess
rcCsZt|}dd|D}dd|D}tj||}tj||}ttj|||tjtjS)zK
    Given a list of Python tuples, create an associated CFDictionary.
    css|]}|dVqdS)rNr	.0tr	r	r
	,sz-_cf_dictionary_from_tuples..css|]}|dVqdS)rNr	rr	r	r
r-s)rr	CFTypeRefZCFDictionaryCreaterZkCFTypeDictionaryKeyCallBacksZkCFTypeDictionaryValueCallBacks)ZtuplesZdictionary_sizekeysvaluesZcf_keysZ	cf_valuesr	r	r
_cf_dictionary_from_tuples%srcCsnt|ttj}t|tj}|dkrXtd}t	||dtj}|sRt
d|j}|dk	rj|d}|S)z
    Creates a Unicode string from a CFString object. Used entirely for error
    reporting.

    Yes, it annoys me quite a lot that this function is this complex.
    Niz'Error copying C string from CFStringRefutf-8)
ctypescastZPOINTERZc_void_prZCFStringGetCStringPtrrZkCFStringEncodingUTF8Zcreate_string_bufferZCFStringGetCStringOSErrorvaluedecode)rZvalue_as_void_pstringbufferresultr	r	r
_cf_string_to_unicode;s&

rcCs\|dkrdSt|d}t|}t||dks:|dkrBd|}|dkrPtj}||dS)z[
    Checks the return code and throws an exception if there is an error to
    report
    rNzOSStatus %s)rZSecCopyErrorMessageStringrr	CFReleasesslSSLError)errorZexception_classZcf_error_stringoutputr	r	r
_assert_no_errorTs
r$cCs|dd}ddt|D}|s.tdttjdt	tj
}|sTtdz`|D]V}t|}|sttdt
tj|}t||stdt||t|qZWntk
rt|YnX|S)	z
    Given a bundle of certs in PEM format, turns them into a CFArray of certs
    that can be used to validate a cert chain.
    s

cSsg|]}t|dqS)r)base64Z	b64decodegroup)r
matchr	r	r

qsz(_cert_array_from_pem..zNo root certificates specifiedrzUnable to allocate memory!zUnable to build cert object!)replace
_PEM_CERTS_REfinditerr r!rCFArrayCreateMutablerrbyrefkCFTypeArrayCallBacksrrZSecCertificateCreateWithDatarCFArrayAppendValue	Exception)Z
pem_bundleZ	der_certsZ
cert_arrayZ	der_bytesZcertdataZcertr	r	r
_cert_array_from_pemis<





r2cCst}t||kS)z=
    Returns True if a given CFTypeRef is a certificate.
    )rZSecCertificateGetTypeIDrCFGetTypeIDitemZexpectedr	r	r
_is_certsr6cCst}t||kS)z;
    Returns True if a given CFTypeRef is an identity.
    )rZSecIdentityGetTypeIDrr3r4r	r	r
_is_identitysr7c
Cstd}t|ddd}t|dd}t}tj||	d}t
}t
|t
||ddt|}t|||fS)a
    This function creates a temporary Mac keychain that we can use to work with
    credentials. This keychain uses a one-time password and a temporary file to
    store the data. We expect to have one keychain per socket. The returned
    SecKeychainRef must be freed by the caller, including calling
    SecKeychainDelete.

    Returns a tuple of the SecKeychainRef and the path to the temporary
    directory that contains it.
    (NrF)osurandomr&Z	b16encodertempfileZmkdtemppathjoinencoderZSecKeychainRefZSecKeychainCreaterrr.r$)Zrandom_bytesfilenameZpasswordZ
tempdirectoryZ
keychain_pathkeychainstatusr	r	r
_temporary_keychains 
rCcCsg}g}d}t|d}|}W5QRXzttj|t|}t}t	|ddddd|t
|}t|t
|}	t|	D]X}
t||
}t
|tj}t|rt|||qt|rt|||qW5|rt|t|X||fS)z
    Given a single file, loads all the trust objects from it into arrays and
    the keychain.
    Returns a tuple of lists: the first list is a list of identities, the
    second a list of certs.
    Nrbr)openreadrrrrrZ
CFArrayRefrZ
SecItemImportrr.r$ZCFArrayGetCountrangeZCFArrayGetValueAtIndexrrr6ZCFRetainappendr7)rAr=certificates
identitiesZresult_arrayfZraw_filedataZfiledatarZresult_countindexr5r	r	r
_load_items_from_filesJ




rMc
Gsg}g}dd|D}z|D]&}t||\}}||||q|st}t||dt	
|}	t|	||t|
dttjdt	
tj}
t||D]}t|
|q|
WSt||D]}t|qXdS)z
    Load certificates and maybe keys from a number of files. Has the end goal
    of returning a CFArray containing one SecIdentityRef, and then zero or more
    SecCertificateRef objects, suitable for use as a client certificate trust
    chain.
    css|]}|r|VqdS)Nr	)r
r=r	r	r
r$sz*_load_client_cert_chain..rN)	itertoolschainrrrMextendrZSecIdentityRefZ SecIdentityCreateWithCertificaterr.r$rHpopr-rr/r0)rApathsrIrJobjZ	file_pathZnew_identitiesZ	new_certsZnew_identityrBZtrust_chainr5r	r	r
_load_client_cert_chains8 


rT)N)__doc__r&rrNrer:r r<ZbindingsrrrcompileDOTALLr+rrrr$r2r6r7rCrMrTr	r	r	r
s,	

-#7site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/bindings.cpython-38.opt-1.pyc000064400000023727151464452710027546 0ustar00U

.eA	@sdZddlmZddlZddlmZddlmZmZm	Z	m
Z
mZmZm
Z
mZmZddlmZmZmZedZesxeded	Zesed
edZeeeedZedkred
ededfeeddZeeddZeZ eZ!eZ"eZ#eZ$eZ%eZ&eZ'eZ(eZ)e
Z*ee)Z+eZ,eZ-ee#Z.ee$Z/ee%Z0ee&Z1ee'Z2eZ3eZ4eZ5eeZ6eZ7eZ8eeZ9eZ:eZ;eeZeeZ?eeZ@eZAeZBeZCeZDeZEeZFze.e/ee7ee8e:ee;eee
geja_He-eja_Ie?ee>e
gejb_He-ejb_Ie?ee
gejc_de-ejc_Ie?ee>ee
geje_He-eje_Ie?ee>gejf_He-ejf_Ie?ee=gejg_He-ejg_Ie?ee@gejh_He-ejh_Ie@e0geji_He-eji_Ie@e gejj_ke-ejj_Ie@eeBgejl_He-ejl_Ie@gejm_He!ejm_Ie@e!gejn_He6ejn_Ie,eDeEgejo_He?ejo_Ie?eFe gejp_He-ejp_Ie?e=gejq_He-ejq_Ie?e=gejr_He-ejr_Ie-egejO_He/ejO_IeTe_TeUe_Ue?e_?e=e_=e>e_>e9e_9e

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
)absolute_importN)find_library)	c_void_pc_int32c_char_pc_size_tc_bytec_uint32c_ulongc_longc_bool)CDLLPOINTER	CFUNCTYPESecurityz'The library Security could not be foundCoreFoundationz-The library CoreFoundation could not be found.)
z1Only OS X 10.8 and newer are supported, not %s.%sT)Z	use_errnokSecImportExportPassphrasekSecImportItemIdentitykCFAllocatorDefaultkCFTypeArrayCallBackskCFTypeDictionaryKeyCallBackskCFTypeDictionaryValueCallBackszError initializing ctypesc@seZdZdZedZdS)CFConstz_
    A class object that acts as essentially a namespace for CoreFoundation
    constants.
    iN)__name__
__module____qualname____doc__CFStringEncodingZkCFStringEncodingUTF8r"r"Y/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_securetransport/bindings.pyrsrc@s(eZdZdZdZdZdZdZdZdZ	dZ
d	ZdZdZ
dZdZdZd
ZdZdZdZdZd
ZdZdZdZdZdZdZdZdZdZdZ dZ!dZ"dZ#dZ$dZ%dZ&dZ'dZ(d Z)d!Z*d"Z+d#Z,d$Z-d%Z.d&Z/d'Z0d(Z1d)Z2d*Z3d+Z4d,Z5d-Z6d.Z7d/Z8d0Z9d1Z:d2Z;d3Zd6Z?d7Z@d8ZAd9ZBd:ZCd;ZDdZGd?ZHd@ZIdAS)B
SecurityConstzU
    A class object that acts as essentially a namespace for Security constants.
    rrrriiiiiiiiiiiiiiiiiiii iQi,iRi,i0i+i/iii$i(i
ik9i#i'i	ig3=<5/iiiiN)Jrrrr Z"kSSLSessionOptionBreakOnServerAuthZ
kSSLProtocol2Z
kSSLProtocol3Z
kTLSProtocol1ZkTLSProtocol11ZkTLSProtocol12ZkTLSProtocol13ZkTLSProtocolMaxSupportedZkSSLClientSideZkSSLStreamTypeZkSecFormatPEMSequenceZkSecTrustResultInvalidZkSecTrustResultProceedZkSecTrustResultDenyZkSecTrustResultUnspecifiedZ&kSecTrustResultRecoverableTrustFailureZ kSecTrustResultFatalTrustFailureZkSecTrustResultOtherErrorZerrSSLProtocolZerrSSLWouldBlockZerrSSLClosedGracefulZerrSSLClosedNoNotifyZerrSSLClosedAbortZerrSSLXCertChainInvalidZerrSSLCryptoZerrSSLInternalZerrSSLCertExpiredZerrSSLCertNotYetValidZerrSSLUnknownRootCertZerrSSLNoRootCertZerrSSLHostNameMismatchZerrSSLPeerHandshakeFailZerrSSLPeerUserCancelledZerrSSLWeakPeerEphemeralDHKeyZerrSSLServerAuthCompletedZerrSSLRecordOverflowZerrSecVerifyFailedZerrSecNoTrustSettingsZerrSecItemNotFoundZerrSecInvalidTrustSettingsZ'TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384Z%TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384Z'TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256Z%TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256Z-TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256Z+TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256Z#TLS_DHE_RSA_WITH_AES_256_GCM_SHA384Z#TLS_DHE_RSA_WITH_AES_128_GCM_SHA256Z'TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384Z%TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384Z$TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHAZ"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_256_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_256_CBC_SHAZ'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256Z%TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256Z$TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHAZ"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_128_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_128_CBC_SHAZTLS_RSA_WITH_AES_256_GCM_SHA384ZTLS_RSA_WITH_AES_128_GCM_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHA256ZTLS_RSA_WITH_AES_128_CBC_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHAZTLS_RSA_WITH_AES_128_CBC_SHAZTLS_AES_128_GCM_SHA256ZTLS_AES_256_GCM_SHA384ZTLS_AES_128_CCM_8_SHA256ZTLS_AES_128_CCM_SHA256r"r"r"r#r$sr$)r Z
__future__rplatformZctypes.utilrZctypesrrrrrr	r
rrr
rrZ
security_pathImportErrorZcore_foundation_pathZmac_verversiontuplemapintsplitversion_infoOSErrorrrZBooleanZCFIndexr!ZCFDataZCFStringZCFArrayZCFMutableArrayZCFDictionaryZCFErrorZCFTypeZCFTypeIDZ	CFTypeRefZCFAllocatorRefZOSStatusZ	CFDataRefZCFStringRefZ
CFArrayRefZCFMutableArrayRefZCFDictionaryRefZCFArrayCallBacksZCFDictionaryKeyCallBacksZCFDictionaryValueCallBacksZSecCertificateRefZSecExternalFormatZSecExternalItemTypeZSecIdentityRefZSecItemImportExportFlagsZ SecItemImportExportKeyParametersZSecKeychainRefZSSLProtocolZSSLCipherSuiteZ
SSLContextRefZSecTrustRefZSSLConnectionRefZSecTrustResultTypeZSecTrustOptionFlagsZSSLProtocolSideZSSLConnectionTypeZSSLSessionOptionZ
SecItemImportZargtypesZrestypeZSecCertificateGetTypeIDZSecIdentityGetTypeIDZSecKeyGetTypeIDZSecCertificateCreateWithDataZSecCertificateCopyDataZSecCopyErrorMessageStringZ SecIdentityCreateWithCertificateZSecKeychainCreateZSecKeychainDeleteZSecPKCS12ImportZSSLReadFuncZSSLWriteFuncZ
SSLSetIOFuncsZSSLSetPeerIDZSSLSetCertificateZSSLSetCertificateAuthoritiesZSSLSetConnectionZSSLSetPeerDomainNameZSSLHandshakeZSSLReadZSSLWriteZSSLCloseZSSLGetNumberSupportedCiphersZSSLGetSupportedCiphersZSSLSetEnabledCiphersZSSLGetNumberEnabledCiphersZargtypeZSSLGetEnabledCiphersZSSLGetNegotiatedCipherZSSLGetNegotiatedProtocolVersionZSSLCopyPeerTrustZSecTrustSetAnchorCertificatesZ!SecTrustSetAnchorCertificatesOnlyZ	argstypesZSecTrustEvaluateZSecTrustGetCertificateCountZSecTrustGetCertificateAtIndexZSSLCreateContextZSSLSetSessionOptionZSSLSetProtocolVersionMinZSSLSetProtocolVersionMaxZin_dllrrZCFRetainZ	CFReleaseZCFGetTypeIDZCFStringCreateWithCStringZCFStringGetCStringPtrZCFStringGetCStringZCFDataCreateZCFDataGetLengthZCFDataGetBytePtrZCFDictionaryCreateZCFDictionaryGetValueZ
CFArrayCreateZCFArrayCreateMutableZCFArrayAppendValueZCFArrayGetCountZCFArrayGetValueAtIndexrrrrAttributeErrorobjectrr$r"r"r"r#s&,












	site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/low_level.cpython-38.pyc000064400000016540151464452710026775 0ustar00U

.e.@sdZddlZddlZddlZddlZddlZddlZddlZddlm	Z	m
Z
mZedej
ZddZdd	Zd
dZddd
ZddZddZddZddZddZddZdS)a
Low-level helpers for the SecureTransport bindings.

These are Python functions that are not directly related to the high-level APIs
but are necessary to get them to work. They include a whole bunch of low-level
CoreFoundation messing about and memory management. The concerns in this module
are almost entirely about trying to avoid memory leaks and providing
appropriate and useful assistance to the higher-level code.
N)SecurityCoreFoundationCFConsts;-----BEGIN CERTIFICATE-----
(.*?)
-----END CERTIFICATE-----cCsttj|t|S)zv
    Given a bytestring, create a CFData object from it. This CFData object must
    be CFReleased by the caller.
    )rCFDataCreatekCFAllocatorDefaultlen)Z
bytestringr	Z/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_securetransport/low_level.py_cf_data_from_bytess
rcCsZt|}dd|D}dd|D}tj||}tj||}ttj|||tjtjS)zK
    Given a list of Python tuples, create an associated CFDictionary.
    css|]}|dVqdS)rNr	.0tr	r	r
	,sz-_cf_dictionary_from_tuples..css|]}|dVqdS)rNr	rr	r	r
r-s)rr	CFTypeRefZCFDictionaryCreaterZkCFTypeDictionaryKeyCallBacksZkCFTypeDictionaryValueCallBacks)ZtuplesZdictionary_sizekeysvaluesZcf_keysZ	cf_valuesr	r	r
_cf_dictionary_from_tuples%srcCsnt|ttj}t|tj}|dkrXtd}t	||dtj}|sRt
d|j}|dk	rj|d}|S)z
    Creates a Unicode string from a CFString object. Used entirely for error
    reporting.

    Yes, it annoys me quite a lot that this function is this complex.
    Niz'Error copying C string from CFStringRefutf-8)
ctypescastZPOINTERZc_void_prZCFStringGetCStringPtrrZkCFStringEncodingUTF8Zcreate_string_bufferZCFStringGetCStringOSErrorvaluedecode)rZvalue_as_void_pstringbufferresultr	r	r
_cf_string_to_unicode;s&

rcCs\|dkrdSt|d}t|}t||dks:|dkrBd|}|dkrPtj}||dS)z[
    Checks the return code and throws an exception if there is an error to
    report
    rNzOSStatus %s)rZSecCopyErrorMessageStringrr	CFReleasesslSSLError)errorZexception_classZcf_error_stringoutputr	r	r
_assert_no_errorTs
r$cCs|dd}ddt|D}|s.tdttjdt	tj
}|sTtdz`|D]V}t|}|sttdt
tj|}t||stdt||t|qZWntk
rt|YnX|S)	z
    Given a bundle of certs in PEM format, turns them into a CFArray of certs
    that can be used to validate a cert chain.
    s

cSsg|]}t|dqS)r)base64Z	b64decodegroup)r
matchr	r	r

qsz(_cert_array_from_pem..zNo root certificates specifiedrzUnable to allocate memory!zUnable to build cert object!)replace
_PEM_CERTS_REfinditerr r!rCFArrayCreateMutablerrbyrefkCFTypeArrayCallBacksrrZSecCertificateCreateWithDatarCFArrayAppendValue	Exception)Z
pem_bundleZ	der_certsZ
cert_arrayZ	der_bytesZcertdataZcertr	r	r
_cert_array_from_pemis<





r2cCst}t||kS)z=
    Returns True if a given CFTypeRef is a certificate.
    )rZSecCertificateGetTypeIDrCFGetTypeIDitemZexpectedr	r	r
_is_certsr6cCst}t||kS)z;
    Returns True if a given CFTypeRef is an identity.
    )rZSecIdentityGetTypeIDrr3r4r	r	r
_is_identitysr7c
Cstd}t|ddd}t|dd}t}tj||	d}t
}t
|t
||ddt|}t|||fS)a
    This function creates a temporary Mac keychain that we can use to work with
    credentials. This keychain uses a one-time password and a temporary file to
    store the data. We expect to have one keychain per socket. The returned
    SecKeychainRef must be freed by the caller, including calling
    SecKeychainDelete.

    Returns a tuple of the SecKeychainRef and the path to the temporary
    directory that contains it.
    (NrF)osurandomr&Z	b16encodertempfileZmkdtemppathjoinencoderZSecKeychainRefZSecKeychainCreaterrr.r$)Zrandom_bytesfilenameZpasswordZ
tempdirectoryZ
keychain_pathkeychainstatusr	r	r
_temporary_keychains 
rCcCsg}g}d}t|d}|}W5QRXzttj|t|}t}t	|ddddd|t
|}t|t
|}	t|	D]X}
t||
}t
|tj}t|rt|||qt|rt|||qW5|rt|t|X||fS)z
    Given a single file, loads all the trust objects from it into arrays and
    the keychain.
    Returns a tuple of lists: the first list is a list of identities, the
    second a list of certs.
    Nrbr)openreadrrrrrZ
CFArrayRefrZ
SecItemImportrr.r$ZCFArrayGetCountrangeZCFArrayGetValueAtIndexrrr6ZCFRetainappendr7)rAr=certificates
identitiesZresult_arrayfZraw_filedataZfiledatarZresult_countindexr5r	r	r
_load_items_from_filesJ




rMc
Gsg}g}dd|D}z|D]&}t||\}}||||q|st}t||dt	
|}	t|	||t|
dttjdt	
tj}
t||D]}t|
|q|
WSt||D]}t|qXdS)z
    Load certificates and maybe keys from a number of files. Has the end goal
    of returning a CFArray containing one SecIdentityRef, and then zero or more
    SecCertificateRef objects, suitable for use as a client certificate trust
    chain.
    css|]}|r|VqdS)Nr	)r
r=r	r	r
r$sz*_load_client_cert_chain..rN)	itertoolschainrrrMextendrZSecIdentityRefZ SecIdentityCreateWithCertificaterr.r$rHpopr-rr/r0)rApathsrIrJobjZ	file_pathZnew_identitiesZ	new_certsZnew_identityrBZtrust_chainr5r	r	r
_load_client_cert_chains8 


rT)N)__doc__r&rrNrer:r r<ZbindingsrrrcompileDOTALLr+rrrr$r2r6r7rCrMrTr	r	r	r
s,	

-#7site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000264151464452710027477 0ustar00U

.e@sdS)NrrrY/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_securetransport/__init__.pysite-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/bindings.cpython-38.pyc000064400000023727151464452710026607 0ustar00U

.eA	@sdZddlmZddlZddlmZddlmZmZm	Z	m
Z
mZmZm
Z
mZmZddlmZmZmZedZesxeded	Zesed
edZeeeedZedkred
ededfeeddZeeddZeZ eZ!eZ"eZ#eZ$eZ%eZ&eZ'eZ(eZ)e
Z*ee)Z+eZ,eZ-ee#Z.ee$Z/ee%Z0ee&Z1ee'Z2eZ3eZ4eZ5eeZ6eZ7eZ8eeZ9eZ:eZ;eeZeeZ?eeZ@eZAeZBeZCeZDeZEeZFze.e/ee7ee8e:ee;eee
geja_He-eja_Ie?ee>e
gejb_He-ejb_Ie?ee
gejc_de-ejc_Ie?ee>ee
geje_He-eje_Ie?ee>gejf_He-ejf_Ie?ee=gejg_He-ejg_Ie?ee@gejh_He-ejh_Ie@e0geji_He-eji_Ie@e gejj_ke-ejj_Ie@eeBgejl_He-ejl_Ie@gejm_He!ejm_Ie@e!gejn_He6ejn_Ie,eDeEgejo_He?ejo_Ie?eFe gejp_He-ejp_Ie?e=gejq_He-ejq_Ie?e=gejr_He-ejr_Ie-egejO_He/ejO_IeTe_TeUe_Ue?e_?e=e_=e>e_>e9e_9e

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
)absolute_importN)find_library)	c_void_pc_int32c_char_pc_size_tc_bytec_uint32c_ulongc_longc_bool)CDLLPOINTER	CFUNCTYPESecurityz'The library Security could not be foundCoreFoundationz-The library CoreFoundation could not be found.)
z1Only OS X 10.8 and newer are supported, not %s.%sT)Z	use_errnokSecImportExportPassphrasekSecImportItemIdentitykCFAllocatorDefaultkCFTypeArrayCallBackskCFTypeDictionaryKeyCallBackskCFTypeDictionaryValueCallBackszError initializing ctypesc@seZdZdZedZdS)CFConstz_
    A class object that acts as essentially a namespace for CoreFoundation
    constants.
    iN)__name__
__module____qualname____doc__CFStringEncodingZkCFStringEncodingUTF8r"r"Y/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/contrib/_securetransport/bindings.pyrsrc@s(eZdZdZdZdZdZdZdZdZ	dZ
d	ZdZdZ
dZdZdZd
ZdZdZdZdZd
ZdZdZdZdZdZdZdZdZdZdZ dZ!dZ"dZ#dZ$dZ%dZ&dZ'dZ(d Z)d!Z*d"Z+d#Z,d$Z-d%Z.d&Z/d'Z0d(Z1d)Z2d*Z3d+Z4d,Z5d-Z6d.Z7d/Z8d0Z9d1Z:d2Z;d3Zd6Z?d7Z@d8ZAd9ZBd:ZCd;ZDdZGd?ZHd@ZIdAS)B
SecurityConstzU
    A class object that acts as essentially a namespace for Security constants.
    rrrriiiiiiiiiiiiiiiiiiii iQi,iRi,i0i+i/iii$i(i
ik9i#i'i	ig3=<5/iiiiN)Jrrrr Z"kSSLSessionOptionBreakOnServerAuthZ
kSSLProtocol2Z
kSSLProtocol3Z
kTLSProtocol1ZkTLSProtocol11ZkTLSProtocol12ZkTLSProtocol13ZkTLSProtocolMaxSupportedZkSSLClientSideZkSSLStreamTypeZkSecFormatPEMSequenceZkSecTrustResultInvalidZkSecTrustResultProceedZkSecTrustResultDenyZkSecTrustResultUnspecifiedZ&kSecTrustResultRecoverableTrustFailureZ kSecTrustResultFatalTrustFailureZkSecTrustResultOtherErrorZerrSSLProtocolZerrSSLWouldBlockZerrSSLClosedGracefulZerrSSLClosedNoNotifyZerrSSLClosedAbortZerrSSLXCertChainInvalidZerrSSLCryptoZerrSSLInternalZerrSSLCertExpiredZerrSSLCertNotYetValidZerrSSLUnknownRootCertZerrSSLNoRootCertZerrSSLHostNameMismatchZerrSSLPeerHandshakeFailZerrSSLPeerUserCancelledZerrSSLWeakPeerEphemeralDHKeyZerrSSLServerAuthCompletedZerrSSLRecordOverflowZerrSecVerifyFailedZerrSecNoTrustSettingsZerrSecItemNotFoundZerrSecInvalidTrustSettingsZ'TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384Z%TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384Z'TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256Z%TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256Z-TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256Z+TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256Z#TLS_DHE_RSA_WITH_AES_256_GCM_SHA384Z#TLS_DHE_RSA_WITH_AES_128_GCM_SHA256Z'TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384Z%TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384Z$TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHAZ"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_256_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_256_CBC_SHAZ'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256Z%TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256Z$TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHAZ"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_128_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_128_CBC_SHAZTLS_RSA_WITH_AES_256_GCM_SHA384ZTLS_RSA_WITH_AES_128_GCM_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHA256ZTLS_RSA_WITH_AES_128_CBC_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHAZTLS_RSA_WITH_AES_128_CBC_SHAZTLS_AES_128_GCM_SHA256ZTLS_AES_256_GCM_SHA384ZTLS_AES_128_CCM_8_SHA256ZTLS_AES_128_CCM_SHA256r"r"r"r#r$sr$)r Z
__future__rplatformZctypes.utilrZctypesrrrrrr	r
rrr
rrZ
security_pathImportErrorZcore_foundation_pathZmac_verversiontuplemapintsplitversion_infoOSErrorrrZBooleanZCFIndexr!ZCFDataZCFStringZCFArrayZCFMutableArrayZCFDictionaryZCFErrorZCFTypeZCFTypeIDZ	CFTypeRefZCFAllocatorRefZOSStatusZ	CFDataRefZCFStringRefZ
CFArrayRefZCFMutableArrayRefZCFDictionaryRefZCFArrayCallBacksZCFDictionaryKeyCallBacksZCFDictionaryValueCallBacksZSecCertificateRefZSecExternalFormatZSecExternalItemTypeZSecIdentityRefZSecItemImportExportFlagsZ SecItemImportExportKeyParametersZSecKeychainRefZSSLProtocolZSSLCipherSuiteZ
SSLContextRefZSecTrustRefZSSLConnectionRefZSecTrustResultTypeZSecTrustOptionFlagsZSSLProtocolSideZSSLConnectionTypeZSSLSessionOptionZ
SecItemImportZargtypesZrestypeZSecCertificateGetTypeIDZSecIdentityGetTypeIDZSecKeyGetTypeIDZSecCertificateCreateWithDataZSecCertificateCopyDataZSecCopyErrorMessageStringZ SecIdentityCreateWithCertificateZSecKeychainCreateZSecKeychainDeleteZSecPKCS12ImportZSSLReadFuncZSSLWriteFuncZ
SSLSetIOFuncsZSSLSetPeerIDZSSLSetCertificateZSSLSetCertificateAuthoritiesZSSLSetConnectionZSSLSetPeerDomainNameZSSLHandshakeZSSLReadZSSLWriteZSSLCloseZSSLGetNumberSupportedCiphersZSSLGetSupportedCiphersZSSLSetEnabledCiphersZSSLGetNumberEnabledCiphersZargtypeZSSLGetEnabledCiphersZSSLGetNegotiatedCipherZSSLGetNegotiatedProtocolVersionZSSLCopyPeerTrustZSecTrustSetAnchorCertificatesZ!SecTrustSetAnchorCertificatesOnlyZ	argstypesZSecTrustEvaluateZSecTrustGetCertificateCountZSecTrustGetCertificateAtIndexZSSLCreateContextZSSLSetSessionOptionZSSLSetProtocolVersionMinZSSLSetProtocolVersionMaxZin_dllrrZCFRetainZ	CFReleaseZCFGetTypeIDZCFStringCreateWithCStringZCFStringGetCStringPtrZCFStringGetCStringZCFDataCreateZCFDataGetLengthZCFDataGetBytePtrZCFDictionaryCreateZCFDictionaryGetValueZ
CFArrayCreateZCFArrayCreateMutableZCFArrayAppendValueZCFArrayGetCountZCFArrayGetValueAtIndexrrrrAttributeErrorobjectrr$r"r"r"r#s&,












	site-packages/pip/_vendor/urllib3/contrib/_securetransport/low_level.py000064400000027264151464452710022514 0ustar00"""
Low-level helpers for the SecureTransport bindings.

These are Python functions that are not directly related to the high-level APIs
but are necessary to get them to work. They include a whole bunch of low-level
CoreFoundation messing about and memory management. The concerns in this module
are almost entirely about trying to avoid memory leaks and providing
appropriate and useful assistance to the higher-level code.
"""
import base64
import ctypes
import itertools
import re
import os
import ssl
import tempfile

from .bindings import Security, CoreFoundation, CFConst


# This regular expression is used to grab PEM data out of a PEM bundle.
_PEM_CERTS_RE = re.compile(
    b"-----BEGIN CERTIFICATE-----\n(.*?)\n-----END CERTIFICATE-----", re.DOTALL
)


def _cf_data_from_bytes(bytestring):
    """
    Given a bytestring, create a CFData object from it. This CFData object must
    be CFReleased by the caller.
    """
    return CoreFoundation.CFDataCreate(
        CoreFoundation.kCFAllocatorDefault, bytestring, len(bytestring)
    )


def _cf_dictionary_from_tuples(tuples):
    """
    Given a list of Python tuples, create an associated CFDictionary.
    """
    dictionary_size = len(tuples)

    # We need to get the dictionary keys and values out in the same order.
    keys = (t[0] for t in tuples)
    values = (t[1] for t in tuples)
    cf_keys = (CoreFoundation.CFTypeRef * dictionary_size)(*keys)
    cf_values = (CoreFoundation.CFTypeRef * dictionary_size)(*values)

    return CoreFoundation.CFDictionaryCreate(
        CoreFoundation.kCFAllocatorDefault,
        cf_keys,
        cf_values,
        dictionary_size,
        CoreFoundation.kCFTypeDictionaryKeyCallBacks,
        CoreFoundation.kCFTypeDictionaryValueCallBacks,
    )


def _cf_string_to_unicode(value):
    """
    Creates a Unicode string from a CFString object. Used entirely for error
    reporting.

    Yes, it annoys me quite a lot that this function is this complex.
    """
    value_as_void_p = ctypes.cast(value, ctypes.POINTER(ctypes.c_void_p))

    string = CoreFoundation.CFStringGetCStringPtr(
        value_as_void_p, CFConst.kCFStringEncodingUTF8
    )
    if string is None:
        buffer = ctypes.create_string_buffer(1024)
        result = CoreFoundation.CFStringGetCString(
            value_as_void_p, buffer, 1024, CFConst.kCFStringEncodingUTF8
        )
        if not result:
            raise OSError("Error copying C string from CFStringRef")
        string = buffer.value
    if string is not None:
        string = string.decode("utf-8")
    return string


def _assert_no_error(error, exception_class=None):
    """
    Checks the return code and throws an exception if there is an error to
    report
    """
    if error == 0:
        return

    cf_error_string = Security.SecCopyErrorMessageString(error, None)
    output = _cf_string_to_unicode(cf_error_string)
    CoreFoundation.CFRelease(cf_error_string)

    if output is None or output == u"":
        output = u"OSStatus %s" % error

    if exception_class is None:
        exception_class = ssl.SSLError

    raise exception_class(output)


def _cert_array_from_pem(pem_bundle):
    """
    Given a bundle of certs in PEM format, turns them into a CFArray of certs
    that can be used to validate a cert chain.
    """
    # Normalize the PEM bundle's line endings.
    pem_bundle = pem_bundle.replace(b"\r\n", b"\n")

    der_certs = [
        base64.b64decode(match.group(1)) for match in _PEM_CERTS_RE.finditer(pem_bundle)
    ]
    if not der_certs:
        raise ssl.SSLError("No root certificates specified")

    cert_array = CoreFoundation.CFArrayCreateMutable(
        CoreFoundation.kCFAllocatorDefault,
        0,
        ctypes.byref(CoreFoundation.kCFTypeArrayCallBacks),
    )
    if not cert_array:
        raise ssl.SSLError("Unable to allocate memory!")

    try:
        for der_bytes in der_certs:
            certdata = _cf_data_from_bytes(der_bytes)
            if not certdata:
                raise ssl.SSLError("Unable to allocate memory!")
            cert = Security.SecCertificateCreateWithData(
                CoreFoundation.kCFAllocatorDefault, certdata
            )
            CoreFoundation.CFRelease(certdata)
            if not cert:
                raise ssl.SSLError("Unable to build cert object!")

            CoreFoundation.CFArrayAppendValue(cert_array, cert)
            CoreFoundation.CFRelease(cert)
    except Exception:
        # We need to free the array before the exception bubbles further.
        # We only want to do that if an error occurs: otherwise, the caller
        # should free.
        CoreFoundation.CFRelease(cert_array)

    return cert_array


def _is_cert(item):
    """
    Returns True if a given CFTypeRef is a certificate.
    """
    expected = Security.SecCertificateGetTypeID()
    return CoreFoundation.CFGetTypeID(item) == expected


def _is_identity(item):
    """
    Returns True if a given CFTypeRef is an identity.
    """
    expected = Security.SecIdentityGetTypeID()
    return CoreFoundation.CFGetTypeID(item) == expected


def _temporary_keychain():
    """
    This function creates a temporary Mac keychain that we can use to work with
    credentials. This keychain uses a one-time password and a temporary file to
    store the data. We expect to have one keychain per socket. The returned
    SecKeychainRef must be freed by the caller, including calling
    SecKeychainDelete.

    Returns a tuple of the SecKeychainRef and the path to the temporary
    directory that contains it.
    """
    # Unfortunately, SecKeychainCreate requires a path to a keychain. This
    # means we cannot use mkstemp to use a generic temporary file. Instead,
    # we're going to create a temporary directory and a filename to use there.
    # This filename will be 8 random bytes expanded into base64. We also need
    # some random bytes to password-protect the keychain we're creating, so we
    # ask for 40 random bytes.
    random_bytes = os.urandom(40)
    filename = base64.b16encode(random_bytes[:8]).decode("utf-8")
    password = base64.b16encode(random_bytes[8:])  # Must be valid UTF-8
    tempdirectory = tempfile.mkdtemp()

    keychain_path = os.path.join(tempdirectory, filename).encode("utf-8")

    # We now want to create the keychain itself.
    keychain = Security.SecKeychainRef()
    status = Security.SecKeychainCreate(
        keychain_path, len(password), password, False, None, ctypes.byref(keychain)
    )
    _assert_no_error(status)

    # Having created the keychain, we want to pass it off to the caller.
    return keychain, tempdirectory


def _load_items_from_file(keychain, path):
    """
    Given a single file, loads all the trust objects from it into arrays and
    the keychain.
    Returns a tuple of lists: the first list is a list of identities, the
    second a list of certs.
    """
    certificates = []
    identities = []
    result_array = None

    with open(path, "rb") as f:
        raw_filedata = f.read()

    try:
        filedata = CoreFoundation.CFDataCreate(
            CoreFoundation.kCFAllocatorDefault, raw_filedata, len(raw_filedata)
        )
        result_array = CoreFoundation.CFArrayRef()
        result = Security.SecItemImport(
            filedata,  # cert data
            None,  # Filename, leaving it out for now
            None,  # What the type of the file is, we don't care
            None,  # what's in the file, we don't care
            0,  # import flags
            None,  # key params, can include passphrase in the future
            keychain,  # The keychain to insert into
            ctypes.byref(result_array),  # Results
        )
        _assert_no_error(result)

        # A CFArray is not very useful to us as an intermediary
        # representation, so we are going to extract the objects we want
        # and then free the array. We don't need to keep hold of keys: the
        # keychain already has them!
        result_count = CoreFoundation.CFArrayGetCount(result_array)
        for index in range(result_count):
            item = CoreFoundation.CFArrayGetValueAtIndex(result_array, index)
            item = ctypes.cast(item, CoreFoundation.CFTypeRef)

            if _is_cert(item):
                CoreFoundation.CFRetain(item)
                certificates.append(item)
            elif _is_identity(item):
                CoreFoundation.CFRetain(item)
                identities.append(item)
    finally:
        if result_array:
            CoreFoundation.CFRelease(result_array)

        CoreFoundation.CFRelease(filedata)

    return (identities, certificates)


def _load_client_cert_chain(keychain, *paths):
    """
    Load certificates and maybe keys from a number of files. Has the end goal
    of returning a CFArray containing one SecIdentityRef, and then zero or more
    SecCertificateRef objects, suitable for use as a client certificate trust
    chain.
    """
    # Ok, the strategy.
    #
    # This relies on knowing that macOS will not give you a SecIdentityRef
    # unless you have imported a key into a keychain. This is a somewhat
    # artificial limitation of macOS (for example, it doesn't necessarily
    # affect iOS), but there is nothing inside Security.framework that lets you
    # get a SecIdentityRef without having a key in a keychain.
    #
    # So the policy here is we take all the files and iterate them in order.
    # Each one will use SecItemImport to have one or more objects loaded from
    # it. We will also point at a keychain that macOS can use to work with the
    # private key.
    #
    # Once we have all the objects, we'll check what we actually have. If we
    # already have a SecIdentityRef in hand, fab: we'll use that. Otherwise,
    # we'll take the first certificate (which we assume to be our leaf) and
    # ask the keychain to give us a SecIdentityRef with that cert's associated
    # key.
    #
    # We'll then return a CFArray containing the trust chain: one
    # SecIdentityRef and then zero-or-more SecCertificateRef objects. The
    # responsibility for freeing this CFArray will be with the caller. This
    # CFArray must remain alive for the entire connection, so in practice it
    # will be stored with a single SSLSocket, along with the reference to the
    # keychain.
    certificates = []
    identities = []

    # Filter out bad paths.
    paths = (path for path in paths if path)

    try:
        for file_path in paths:
            new_identities, new_certs = _load_items_from_file(keychain, file_path)
            identities.extend(new_identities)
            certificates.extend(new_certs)

        # Ok, we have everything. The question is: do we have an identity? If
        # not, we want to grab one from the first cert we have.
        if not identities:
            new_identity = Security.SecIdentityRef()
            status = Security.SecIdentityCreateWithCertificate(
                keychain, certificates[0], ctypes.byref(new_identity)
            )
            _assert_no_error(status)
            identities.append(new_identity)

            # We now want to release the original certificate, as we no longer
            # need it.
            CoreFoundation.CFRelease(certificates.pop(0))

        # We now need to build a new CFArray that holds the trust chain.
        trust_chain = CoreFoundation.CFArrayCreateMutable(
            CoreFoundation.kCFAllocatorDefault,
            0,
            ctypes.byref(CoreFoundation.kCFTypeArrayCallBacks),
        )
        for item in itertools.chain(identities, certificates):
            # ArrayAppendValue does a CFRetain on the item. That's fine,
            # because the finally block will release our other refs to them.
            CoreFoundation.CFArrayAppendValue(trust_chain, item)

        return trust_chain
    finally:
        for obj in itertools.chain(identities, certificates):
            CoreFoundation.CFRelease(obj)
site-packages/pip/_vendor/urllib3/contrib/_securetransport/__init__.py000064400000000000151464452710022236 0ustar00site-packages/pip/_vendor/urllib3/contrib/socks.py000064400000015574151464452710016245 0ustar00# -*- coding: utf-8 -*-
"""
This module contains provisional support for SOCKS proxies from within
urllib3. This module supports SOCKS4, SOCKS4A (an extension of SOCKS4), and
SOCKS5. To enable its functionality, either install PySocks or install this
module with the ``socks`` extra.

The SOCKS implementation supports the full range of urllib3 features. It also
supports the following SOCKS features:

- SOCKS4A (``proxy_url='socks4a://...``)
- SOCKS4 (``proxy_url='socks4://...``)
- SOCKS5 with remote DNS (``proxy_url='socks5h://...``)
- SOCKS5 with local DNS (``proxy_url='socks5://...``)
- Usernames and passwords for the SOCKS proxy

 .. note::
    It is recommended to use ``socks5h://`` or ``socks4a://`` schemes in
    your ``proxy_url`` to ensure that DNS resolution is done from the remote
    server instead of client-side when connecting to a domain name.

SOCKS4 supports IPv4 and domain names with the SOCKS4A extension. SOCKS5
supports IPv4, IPv6, and domain names.

When connecting to a SOCKS4 proxy the ``username`` portion of the ``proxy_url``
will be sent as the ``userid`` section of the SOCKS request::

    proxy_url="socks4a://@proxy-host"

When connecting to a SOCKS5 proxy the ``username`` and ``password`` portion
of the ``proxy_url`` will be sent as the username/password to authenticate
with the proxy::

    proxy_url="socks5h://:@proxy-host"

"""
from __future__ import absolute_import

try:
    import socks
except ImportError:
    import warnings
    from ..exceptions import DependencyWarning

    warnings.warn(
        (
            "SOCKS support in urllib3 requires the installation of optional "
            "dependencies: specifically, PySocks.  For more information, see "
            "https://urllib3.readthedocs.io/en/latest/contrib.html#socks-proxies"
        ),
        DependencyWarning,
    )
    raise

from socket import error as SocketError, timeout as SocketTimeout

from ..connection import HTTPConnection, HTTPSConnection
from ..connectionpool import HTTPConnectionPool, HTTPSConnectionPool
from ..exceptions import ConnectTimeoutError, NewConnectionError
from ..poolmanager import PoolManager
from ..util.url import parse_url

try:
    import ssl
except ImportError:
    ssl = None


class SOCKSConnection(HTTPConnection):
    """
    A plain-text HTTP connection that connects via a SOCKS proxy.
    """

    def __init__(self, *args, **kwargs):
        self._socks_options = kwargs.pop("_socks_options")
        super(SOCKSConnection, self).__init__(*args, **kwargs)

    def _new_conn(self):
        """
        Establish a new connection via the SOCKS proxy.
        """
        extra_kw = {}
        if self.source_address:
            extra_kw["source_address"] = self.source_address

        if self.socket_options:
            extra_kw["socket_options"] = self.socket_options

        try:
            conn = socks.create_connection(
                (self.host, self.port),
                proxy_type=self._socks_options["socks_version"],
                proxy_addr=self._socks_options["proxy_host"],
                proxy_port=self._socks_options["proxy_port"],
                proxy_username=self._socks_options["username"],
                proxy_password=self._socks_options["password"],
                proxy_rdns=self._socks_options["rdns"],
                timeout=self.timeout,
                **extra_kw
            )

        except SocketTimeout:
            raise ConnectTimeoutError(
                self,
                "Connection to %s timed out. (connect timeout=%s)"
                % (self.host, self.timeout),
            )

        except socks.ProxyError as e:
            # This is fragile as hell, but it seems to be the only way to raise
            # useful errors here.
            if e.socket_err:
                error = e.socket_err
                if isinstance(error, SocketTimeout):
                    raise ConnectTimeoutError(
                        self,
                        "Connection to %s timed out. (connect timeout=%s)"
                        % (self.host, self.timeout),
                    )
                else:
                    raise NewConnectionError(
                        self, "Failed to establish a new connection: %s" % error
                    )
            else:
                raise NewConnectionError(
                    self, "Failed to establish a new connection: %s" % e
                )

        except SocketError as e:  # Defensive: PySocks should catch all these.
            raise NewConnectionError(
                self, "Failed to establish a new connection: %s" % e
            )

        return conn


# We don't need to duplicate the Verified/Unverified distinction from
# urllib3/connection.py here because the HTTPSConnection will already have been
# correctly set to either the Verified or Unverified form by that module. This
# means the SOCKSHTTPSConnection will automatically be the correct type.
class SOCKSHTTPSConnection(SOCKSConnection, HTTPSConnection):
    pass


class SOCKSHTTPConnectionPool(HTTPConnectionPool):
    ConnectionCls = SOCKSConnection


class SOCKSHTTPSConnectionPool(HTTPSConnectionPool):
    ConnectionCls = SOCKSHTTPSConnection


class SOCKSProxyManager(PoolManager):
    """
    A version of the urllib3 ProxyManager that routes connections via the
    defined SOCKS proxy.
    """

    pool_classes_by_scheme = {
        "http": SOCKSHTTPConnectionPool,
        "https": SOCKSHTTPSConnectionPool,
    }

    def __init__(
        self,
        proxy_url,
        username=None,
        password=None,
        num_pools=10,
        headers=None,
        **connection_pool_kw
    ):
        parsed = parse_url(proxy_url)

        if username is None and password is None and parsed.auth is not None:
            split = parsed.auth.split(":")
            if len(split) == 2:
                username, password = split
        if parsed.scheme == "socks5":
            socks_version = socks.PROXY_TYPE_SOCKS5
            rdns = False
        elif parsed.scheme == "socks5h":
            socks_version = socks.PROXY_TYPE_SOCKS5
            rdns = True
        elif parsed.scheme == "socks4":
            socks_version = socks.PROXY_TYPE_SOCKS4
            rdns = False
        elif parsed.scheme == "socks4a":
            socks_version = socks.PROXY_TYPE_SOCKS4
            rdns = True
        else:
            raise ValueError("Unable to determine SOCKS version from %s" % proxy_url)

        self.proxy_url = proxy_url

        socks_options = {
            "socks_version": socks_version,
            "proxy_host": parsed.host,
            "proxy_port": parsed.port,
            "username": username,
            "password": password,
            "rdns": rdns,
        }
        connection_pool_kw["_socks_options"] = socks_options

        super(SOCKSProxyManager, self).__init__(
            num_pools, headers, **connection_pool_kw
        )

        self.pool_classes_by_scheme = SOCKSProxyManager.pool_classes_by_scheme
site-packages/pip/_vendor/urllib3/contrib/__init__.py000064400000000000151464452710016634 0ustar00site-packages/pip/_vendor/urllib3/exceptions.py000064400000014722151464452710015636 0ustar00from __future__ import absolute_import
from .packages.six.moves.http_client import IncompleteRead as httplib_IncompleteRead

# Base Exceptions


class HTTPError(Exception):
    "Base exception used by this module."
    pass


class HTTPWarning(Warning):
    "Base warning used by this module."
    pass


class PoolError(HTTPError):
    "Base exception for errors caused within a pool."

    def __init__(self, pool, message):
        self.pool = pool
        HTTPError.__init__(self, "%s: %s" % (pool, message))

    def __reduce__(self):
        # For pickling purposes.
        return self.__class__, (None, None)


class RequestError(PoolError):
    "Base exception for PoolErrors that have associated URLs."

    def __init__(self, pool, url, message):
        self.url = url
        PoolError.__init__(self, pool, message)

    def __reduce__(self):
        # For pickling purposes.
        return self.__class__, (None, self.url, None)


class SSLError(HTTPError):
    "Raised when SSL certificate fails in an HTTPS connection."
    pass


class ProxyError(HTTPError):
    "Raised when the connection to a proxy fails."
    pass


class DecodeError(HTTPError):
    "Raised when automatic decoding based on Content-Type fails."
    pass


class ProtocolError(HTTPError):
    "Raised when something unexpected happens mid-request/response."
    pass


#: Renamed to ProtocolError but aliased for backwards compatibility.
ConnectionError = ProtocolError


# Leaf Exceptions


class MaxRetryError(RequestError):
    """Raised when the maximum number of retries is exceeded.

    :param pool: The connection pool
    :type pool: :class:`~urllib3.connectionpool.HTTPConnectionPool`
    :param string url: The requested Url
    :param exceptions.Exception reason: The underlying error

    """

    def __init__(self, pool, url, reason=None):
        self.reason = reason

        message = "Max retries exceeded with url: %s (Caused by %r)" % (url, reason)

        RequestError.__init__(self, pool, url, message)


class HostChangedError(RequestError):
    "Raised when an existing pool gets a request for a foreign host."

    def __init__(self, pool, url, retries=3):
        message = "Tried to open a foreign host with url: %s" % url
        RequestError.__init__(self, pool, url, message)
        self.retries = retries


class TimeoutStateError(HTTPError):
    """ Raised when passing an invalid state to a timeout """

    pass


class TimeoutError(HTTPError):
    """ Raised when a socket timeout error occurs.

    Catching this error will catch both :exc:`ReadTimeoutErrors
    ` and :exc:`ConnectTimeoutErrors `.
    """

    pass


class ReadTimeoutError(TimeoutError, RequestError):
    "Raised when a socket timeout occurs while receiving data from a server"
    pass


# This timeout error does not have a URL attached and needs to inherit from the
# base HTTPError
class ConnectTimeoutError(TimeoutError):
    "Raised when a socket timeout occurs while connecting to a server"
    pass


class NewConnectionError(ConnectTimeoutError, PoolError):
    "Raised when we fail to establish a new connection. Usually ECONNREFUSED."
    pass


class EmptyPoolError(PoolError):
    "Raised when a pool runs out of connections and no more are allowed."
    pass


class ClosedPoolError(PoolError):
    "Raised when a request enters a pool after the pool has been closed."
    pass


class LocationValueError(ValueError, HTTPError):
    "Raised when there is something wrong with a given URL input."
    pass


class LocationParseError(LocationValueError):
    "Raised when get_host or similar fails to parse the URL input."

    def __init__(self, location):
        message = "Failed to parse: %s" % location
        HTTPError.__init__(self, message)

        self.location = location


class ResponseError(HTTPError):
    "Used as a container for an error reason supplied in a MaxRetryError."
    GENERIC_ERROR = "too many error responses"
    SPECIFIC_ERROR = "too many {status_code} error responses"


class SecurityWarning(HTTPWarning):
    "Warned when performing security reducing actions"
    pass


class SubjectAltNameWarning(SecurityWarning):
    "Warned when connecting to a host with a certificate missing a SAN."
    pass


class InsecureRequestWarning(SecurityWarning):
    "Warned when making an unverified HTTPS request."
    pass


class SystemTimeWarning(SecurityWarning):
    "Warned when system time is suspected to be wrong"
    pass


class InsecurePlatformWarning(SecurityWarning):
    "Warned when certain SSL configuration is not available on a platform."
    pass


class SNIMissingWarning(HTTPWarning):
    "Warned when making a HTTPS request without SNI available."
    pass


class DependencyWarning(HTTPWarning):
    """
    Warned when an attempt is made to import a module with missing optional
    dependencies.
    """

    pass


class ResponseNotChunked(ProtocolError, ValueError):
    "Response needs to be chunked in order to read it as chunks."
    pass


class BodyNotHttplibCompatible(HTTPError):
    """
    Body should be httplib.HTTPResponse like (have an fp attribute which
    returns raw chunks) for read_chunked().
    """

    pass


class IncompleteRead(HTTPError, httplib_IncompleteRead):
    """
    Response length doesn't match expected Content-Length

    Subclass of http_client.IncompleteRead to allow int value
    for `partial` to avoid creating large objects on streamed
    reads.
    """

    def __init__(self, partial, expected):
        super(IncompleteRead, self).__init__(partial, expected)

    def __repr__(self):
        return "IncompleteRead(%i bytes read, " "%i more expected)" % (
            self.partial,
            self.expected,
        )


class InvalidHeader(HTTPError):
    "The header provided was somehow invalid."
    pass


class ProxySchemeUnknown(AssertionError, ValueError):
    "ProxyManager does not support the supplied scheme"
    # TODO(t-8ch): Stop inheriting from AssertionError in v2.0.

    def __init__(self, scheme):
        message = "Not supported proxy scheme %s" % scheme
        super(ProxySchemeUnknown, self).__init__(message)


class HeaderParsingError(HTTPError):
    "Raised by assert_header_parsing, but we convert it to a log.warning statement."

    def __init__(self, defects, unparsed_data):
        message = "%s, unparsed data: %r" % (defects or "Unknown", unparsed_data)
        super(HeaderParsingError, self).__init__(message)


class UnrewindableBodyError(HTTPError):
    "urllib3 encountered an error when trying to rewind a body"
    pass
site-packages/pip/_vendor/urllib3/poolmanager.py000064400000041235151464452710015760 0ustar00from __future__ import absolute_import
import collections
import functools
import logging

from ._collections import RecentlyUsedContainer
from .connectionpool import HTTPConnectionPool, HTTPSConnectionPool
from .connectionpool import port_by_scheme
from .exceptions import LocationValueError, MaxRetryError, ProxySchemeUnknown
from .packages import six
from .packages.six.moves.urllib.parse import urljoin
from .request import RequestMethods
from .util.url import parse_url
from .util.retry import Retry


__all__ = ["PoolManager", "ProxyManager", "proxy_from_url"]


log = logging.getLogger(__name__)

SSL_KEYWORDS = (
    "key_file",
    "cert_file",
    "cert_reqs",
    "ca_certs",
    "ssl_version",
    "ca_cert_dir",
    "ssl_context",
    "key_password",
)

# All known keyword arguments that could be provided to the pool manager, its
# pools, or the underlying connections. This is used to construct a pool key.
_key_fields = (
    "key_scheme",  # str
    "key_host",  # str
    "key_port",  # int
    "key_timeout",  # int or float or Timeout
    "key_retries",  # int or Retry
    "key_strict",  # bool
    "key_block",  # bool
    "key_source_address",  # str
    "key_key_file",  # str
    "key_key_password",  # str
    "key_cert_file",  # str
    "key_cert_reqs",  # str
    "key_ca_certs",  # str
    "key_ssl_version",  # str
    "key_ca_cert_dir",  # str
    "key_ssl_context",  # instance of ssl.SSLContext or urllib3.util.ssl_.SSLContext
    "key_maxsize",  # int
    "key_headers",  # dict
    "key__proxy",  # parsed proxy url
    "key__proxy_headers",  # dict
    "key_socket_options",  # list of (level (int), optname (int), value (int or str)) tuples
    "key__socks_options",  # dict
    "key_assert_hostname",  # bool or string
    "key_assert_fingerprint",  # str
    "key_server_hostname",  # str
)

#: The namedtuple class used to construct keys for the connection pool.
#: All custom key schemes should include the fields in this key at a minimum.
PoolKey = collections.namedtuple("PoolKey", _key_fields)


def _default_key_normalizer(key_class, request_context):
    """
    Create a pool key out of a request context dictionary.

    According to RFC 3986, both the scheme and host are case-insensitive.
    Therefore, this function normalizes both before constructing the pool
    key for an HTTPS request. If you wish to change this behaviour, provide
    alternate callables to ``key_fn_by_scheme``.

    :param key_class:
        The class to use when constructing the key. This should be a namedtuple
        with the ``scheme`` and ``host`` keys at a minimum.
    :type  key_class: namedtuple
    :param request_context:
        A dictionary-like object that contain the context for a request.
    :type  request_context: dict

    :return: A namedtuple that can be used as a connection pool key.
    :rtype:  PoolKey
    """
    # Since we mutate the dictionary, make a copy first
    context = request_context.copy()
    context["scheme"] = context["scheme"].lower()
    context["host"] = context["host"].lower()

    # These are both dictionaries and need to be transformed into frozensets
    for key in ("headers", "_proxy_headers", "_socks_options"):
        if key in context and context[key] is not None:
            context[key] = frozenset(context[key].items())

    # The socket_options key may be a list and needs to be transformed into a
    # tuple.
    socket_opts = context.get("socket_options")
    if socket_opts is not None:
        context["socket_options"] = tuple(socket_opts)

    # Map the kwargs to the names in the namedtuple - this is necessary since
    # namedtuples can't have fields starting with '_'.
    for key in list(context.keys()):
        context["key_" + key] = context.pop(key)

    # Default to ``None`` for keys missing from the context
    for field in key_class._fields:
        if field not in context:
            context[field] = None

    return key_class(**context)


#: A dictionary that maps a scheme to a callable that creates a pool key.
#: This can be used to alter the way pool keys are constructed, if desired.
#: Each PoolManager makes a copy of this dictionary so they can be configured
#: globally here, or individually on the instance.
key_fn_by_scheme = {
    "http": functools.partial(_default_key_normalizer, PoolKey),
    "https": functools.partial(_default_key_normalizer, PoolKey),
}

pool_classes_by_scheme = {"http": HTTPConnectionPool, "https": HTTPSConnectionPool}


class PoolManager(RequestMethods):
    """
    Allows for arbitrary requests while transparently keeping track of
    necessary connection pools for you.

    :param num_pools:
        Number of connection pools to cache before discarding the least
        recently used pool.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param \\**connection_pool_kw:
        Additional parameters are used to create fresh
        :class:`urllib3.connectionpool.ConnectionPool` instances.

    Example::

        >>> manager = PoolManager(num_pools=2)
        >>> r = manager.request('GET', 'http://google.com/')
        >>> r = manager.request('GET', 'http://google.com/mail')
        >>> r = manager.request('GET', 'http://yahoo.com/')
        >>> len(manager.pools)
        2

    """

    proxy = None

    def __init__(self, num_pools=10, headers=None, **connection_pool_kw):
        RequestMethods.__init__(self, headers)
        self.connection_pool_kw = connection_pool_kw
        self.pools = RecentlyUsedContainer(num_pools, dispose_func=lambda p: p.close())

        # Locally set the pool classes and keys so other PoolManagers can
        # override them.
        self.pool_classes_by_scheme = pool_classes_by_scheme
        self.key_fn_by_scheme = key_fn_by_scheme.copy()

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        self.clear()
        # Return False to re-raise any potential exceptions
        return False

    def _new_pool(self, scheme, host, port, request_context=None):
        """
        Create a new :class:`ConnectionPool` based on host, port, scheme, and
        any additional pool keyword arguments.

        If ``request_context`` is provided, it is provided as keyword arguments
        to the pool class used. This method is used to actually create the
        connection pools handed out by :meth:`connection_from_url` and
        companion methods. It is intended to be overridden for customization.
        """
        pool_cls = self.pool_classes_by_scheme[scheme]
        if request_context is None:
            request_context = self.connection_pool_kw.copy()

        # Although the context has everything necessary to create the pool,
        # this function has historically only used the scheme, host, and port
        # in the positional args. When an API change is acceptable these can
        # be removed.
        for key in ("scheme", "host", "port"):
            request_context.pop(key, None)

        if scheme == "http":
            for kw in SSL_KEYWORDS:
                request_context.pop(kw, None)

        return pool_cls(host, port, **request_context)

    def clear(self):
        """
        Empty our store of pools and direct them all to close.

        This will not affect in-flight connections, but they will not be
        re-used after completion.
        """
        self.pools.clear()

    def connection_from_host(self, host, port=None, scheme="http", pool_kwargs=None):
        """
        Get a :class:`ConnectionPool` based on the host, port, and scheme.

        If ``port`` isn't given, it will be derived from the ``scheme`` using
        ``urllib3.connectionpool.port_by_scheme``. If ``pool_kwargs`` is
        provided, it is merged with the instance's ``connection_pool_kw``
        variable and used to create the new connection pool, if one is
        needed.
        """

        if not host:
            raise LocationValueError("No host specified.")

        request_context = self._merge_pool_kwargs(pool_kwargs)
        request_context["scheme"] = scheme or "http"
        if not port:
            port = port_by_scheme.get(request_context["scheme"].lower(), 80)
        request_context["port"] = port
        request_context["host"] = host

        return self.connection_from_context(request_context)

    def connection_from_context(self, request_context):
        """
        Get a :class:`ConnectionPool` based on the request context.

        ``request_context`` must at least contain the ``scheme`` key and its
        value must be a key in ``key_fn_by_scheme`` instance variable.
        """
        scheme = request_context["scheme"].lower()
        pool_key_constructor = self.key_fn_by_scheme[scheme]
        pool_key = pool_key_constructor(request_context)

        return self.connection_from_pool_key(pool_key, request_context=request_context)

    def connection_from_pool_key(self, pool_key, request_context=None):
        """
        Get a :class:`ConnectionPool` based on the provided pool key.

        ``pool_key`` should be a namedtuple that only contains immutable
        objects. At a minimum it must have the ``scheme``, ``host``, and
        ``port`` fields.
        """
        with self.pools.lock:
            # If the scheme, host, or port doesn't match existing open
            # connections, open a new ConnectionPool.
            pool = self.pools.get(pool_key)
            if pool:
                return pool

            # Make a fresh ConnectionPool of the desired type
            scheme = request_context["scheme"]
            host = request_context["host"]
            port = request_context["port"]
            pool = self._new_pool(scheme, host, port, request_context=request_context)
            self.pools[pool_key] = pool

        return pool

    def connection_from_url(self, url, pool_kwargs=None):
        """
        Similar to :func:`urllib3.connectionpool.connection_from_url`.

        If ``pool_kwargs`` is not provided and a new pool needs to be
        constructed, ``self.connection_pool_kw`` is used to initialize
        the :class:`urllib3.connectionpool.ConnectionPool`. If ``pool_kwargs``
        is provided, it is used instead. Note that if a new pool does not
        need to be created for the request, the provided ``pool_kwargs`` are
        not used.
        """
        u = parse_url(url)
        return self.connection_from_host(
            u.host, port=u.port, scheme=u.scheme, pool_kwargs=pool_kwargs
        )

    def _merge_pool_kwargs(self, override):
        """
        Merge a dictionary of override values for self.connection_pool_kw.

        This does not modify self.connection_pool_kw and returns a new dict.
        Any keys in the override dictionary with a value of ``None`` are
        removed from the merged dictionary.
        """
        base_pool_kwargs = self.connection_pool_kw.copy()
        if override:
            for key, value in override.items():
                if value is None:
                    try:
                        del base_pool_kwargs[key]
                    except KeyError:
                        pass
                else:
                    base_pool_kwargs[key] = value
        return base_pool_kwargs

    def urlopen(self, method, url, redirect=True, **kw):
        """
        Same as :meth:`urllib3.connectionpool.HTTPConnectionPool.urlopen`
        with custom cross-host redirect logic and only sends the request-uri
        portion of the ``url``.

        The given ``url`` parameter must be absolute, such that an appropriate
        :class:`urllib3.connectionpool.ConnectionPool` can be chosen for it.
        """
        u = parse_url(url)
        conn = self.connection_from_host(u.host, port=u.port, scheme=u.scheme)

        kw["assert_same_host"] = False
        kw["redirect"] = False

        if "headers" not in kw:
            kw["headers"] = self.headers.copy()

        if self.proxy is not None and u.scheme == "http":
            response = conn.urlopen(method, url, **kw)
        else:
            response = conn.urlopen(method, u.request_uri, **kw)

        redirect_location = redirect and response.get_redirect_location()
        if not redirect_location:
            return response

        # Support relative URLs for redirecting.
        redirect_location = urljoin(url, redirect_location)

        # RFC 7231, Section 6.4.4
        if response.status == 303:
            method = "GET"

        retries = kw.get("retries")
        if not isinstance(retries, Retry):
            retries = Retry.from_int(retries, redirect=redirect)

        # Strip headers marked as unsafe to forward to the redirected location.
        # Check remove_headers_on_redirect to avoid a potential network call within
        # conn.is_same_host() which may use socket.gethostbyname() in the future.
        if retries.remove_headers_on_redirect and not conn.is_same_host(
            redirect_location
        ):
            headers = list(six.iterkeys(kw["headers"]))
            for header in headers:
                if header.lower() in retries.remove_headers_on_redirect:
                    kw["headers"].pop(header, None)

        try:
            retries = retries.increment(method, url, response=response, _pool=conn)
        except MaxRetryError:
            if retries.raise_on_redirect:
                raise
            return response

        kw["retries"] = retries
        kw["redirect"] = redirect

        log.info("Redirecting %s -> %s", url, redirect_location)
        return self.urlopen(method, redirect_location, **kw)


class ProxyManager(PoolManager):
    """
    Behaves just like :class:`PoolManager`, but sends all requests through
    the defined proxy, using the CONNECT method for HTTPS URLs.

    :param proxy_url:
        The URL of the proxy to be used.

    :param proxy_headers:
        A dictionary containing headers that will be sent to the proxy. In case
        of HTTP they are being sent with each request, while in the
        HTTPS/CONNECT case they are sent only once. Could be used for proxy
        authentication.

    Example:
        >>> proxy = urllib3.ProxyManager('http://localhost:3128/')
        >>> r1 = proxy.request('GET', 'http://google.com/')
        >>> r2 = proxy.request('GET', 'http://httpbin.org/')
        >>> len(proxy.pools)
        1
        >>> r3 = proxy.request('GET', 'https://httpbin.org/')
        >>> r4 = proxy.request('GET', 'https://twitter.com/')
        >>> len(proxy.pools)
        3

    """

    def __init__(
        self,
        proxy_url,
        num_pools=10,
        headers=None,
        proxy_headers=None,
        **connection_pool_kw
    ):

        if isinstance(proxy_url, HTTPConnectionPool):
            proxy_url = "%s://%s:%i" % (
                proxy_url.scheme,
                proxy_url.host,
                proxy_url.port,
            )
        proxy = parse_url(proxy_url)
        if not proxy.port:
            port = port_by_scheme.get(proxy.scheme, 80)
            proxy = proxy._replace(port=port)

        if proxy.scheme not in ("http", "https"):
            raise ProxySchemeUnknown(proxy.scheme)

        self.proxy = proxy
        self.proxy_headers = proxy_headers or {}

        connection_pool_kw["_proxy"] = self.proxy
        connection_pool_kw["_proxy_headers"] = self.proxy_headers

        super(ProxyManager, self).__init__(num_pools, headers, **connection_pool_kw)

    def connection_from_host(self, host, port=None, scheme="http", pool_kwargs=None):
        if scheme == "https":
            return super(ProxyManager, self).connection_from_host(
                host, port, scheme, pool_kwargs=pool_kwargs
            )

        return super(ProxyManager, self).connection_from_host(
            self.proxy.host, self.proxy.port, self.proxy.scheme, pool_kwargs=pool_kwargs
        )

    def _set_proxy_headers(self, url, headers=None):
        """
        Sets headers needed by proxies: specifically, the Accept and Host
        headers. Only sets headers not provided by the user.
        """
        headers_ = {"Accept": "*/*"}

        netloc = parse_url(url).netloc
        if netloc:
            headers_["Host"] = netloc

        if headers:
            headers_.update(headers)
        return headers_

    def urlopen(self, method, url, redirect=True, **kw):
        "Same as HTTP(S)ConnectionPool.urlopen, ``url`` must be absolute."
        u = parse_url(url)

        if u.scheme == "http":
            # For proxied HTTPS requests, httplib sets the necessary headers
            # on the CONNECT to the proxy. For HTTP, we'll definitely
            # need to set 'Host' at the very least.
            headers = kw.get("headers", self.headers)
            kw["headers"] = self._set_proxy_headers(url, headers)

        return super(ProxyManager, self).urlopen(method, url, redirect=redirect, **kw)


def proxy_from_url(url, **kw):
    return ProxyManager(proxy_url=url, **kw)
pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/_implementation.cpython-38.opt-1.pyc000064400000006347151464452710031444 0ustar00site-packagesU

.e\@sxdZddlZddlZzddlmZWnek
r<dZYnXdZGdddeZddd	Z	d
dZ
dd
ZddZdS)zJThe match_hostname() function from Python 3.3.3, essential when using SSL.N)	ipaddressz3.5.0.1c@seZdZdS)CertificateErrorN)__name__
__module____qualname__rrc/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.pyrsrc
Csg}|sdS|d}|d}|dd}|d}||krLtdt||s`||kS|dkrt|dn>|d	s|d	r|t|n|t|	d
d|D]}|t|qt
dd
|dtj}	|	
|S)zhMatching according to RFC 6125, section 6.4.3

    http://tools.ietf.org/html/rfc6125#section-6.4.3
    F.rr	N*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)splitcountrreprlowerappend
startswithreescapereplacecompilejoin
IGNORECASEmatch)
ZdnhostnameZ
max_wildcardsZpatspartsZleftmostZ	remainderZ	wildcardsZfragZpatrrr_dnsname_matchs,


rcCs&t|tr"tjdkr"t|ddd}|S)N)asciistrict)encodingerrors)
isinstancestrsysversion_infoZunicode)objrrr_to_unicodeOsr&cCstt|}||kS)zExact matching of IP addresses.

    RFC 6125 explicitly doesn't define an algorithm for this
    (section 1.7.2 - "Out of Scope").
    )r
ip_addressr&rstrip)Zipnamehost_ipZiprrr_ipaddress_matchUsr*cCs|stdztt|}WnPtk
r6d}Yn:tk
rLd}Yn$tk
rntdkrhd}nYnXg}|dd}|D]^\}}|dkr|dkrt||rdS||q|dkr|dk	rt	||rdS||q|s0|ddD]8}|D].\}}|dkrt||r"dS||qqt
|d	kr\td
|dt
t|fn,t
|d	krtd||d
fntddS)a)Verify that *cert* (in decoded format as returned by
    SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
    rules are followed, but IP addresses are not accepted for *hostname*.

    CertificateError is raised on failure. On success, the function
    returns nothing.
    ztempty or no certificate, match_hostname needs a SSL socket or SSL context with either CERT_OPTIONAL or CERT_REQUIREDNZsubjectAltNamerZDNSz
IP AddressZsubjectZ
commonNamer	z&hostname %r doesn't match either of %sz, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found)
ValueErrorrr'r&UnicodeErrorAttributeErrorgetrrr*lenrrmapr)Zcertrr)ZdnsnamesZsankeyvaluesubrrrmatch_hostnameasZr4)r	)
__doc__rr#Zpip._vendorrImportError__version__r+rrr&r*r4rrrrs

6site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/__init__.cpython-38.pyc000064400000001006151464452710027122 0ustar00U

.e@sddlZz&ejdkredddlmZmZWnNek
r|zddlmZmZWn$ek
rvddlmZmZYnXYnXdZdS)N)zFallback to vendored code)CertificateErrormatch_hostname)	sysversion_infoImportErrorZsslrrZbackports.ssl_match_hostnameZ_implementation__all__rr\/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.pys
pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001006151464452710030002 0ustar00site-packagesU

.e@sddlZz&ejdkredddlmZmZWnNek
r|zddlmZmZWn$ek
rvddlmZmZYnXYnXdZdS)N)zFallback to vendored code)CertificateErrormatch_hostname)	sysversion_infoImportErrorZsslrrZbackports.ssl_match_hostnameZ_implementation__all__rr\/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.pys
pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/_implementation.cpython-38.pyc000064400000006347151464452710030505 0ustar00site-packagesU

.e\@sxdZddlZddlZzddlmZWnek
r<dZYnXdZGdddeZddd	Z	d
dZ
dd
ZddZdS)zJThe match_hostname() function from Python 3.3.3, essential when using SSL.N)	ipaddressz3.5.0.1c@seZdZdS)CertificateErrorN)__name__
__module____qualname__rrc/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.pyrsrc
Csg}|sdS|d}|d}|dd}|d}||krLtdt||s`||kS|dkrt|dn>|d	s|d	r|t|n|t|	d
d|D]}|t|qt
dd
|dtj}	|	
|S)zhMatching according to RFC 6125, section 6.4.3

    http://tools.ietf.org/html/rfc6125#section-6.4.3
    F.rr	N*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)splitcountrreprlowerappend
startswithreescapereplacecompilejoin
IGNORECASEmatch)
ZdnhostnameZ
max_wildcardsZpatspartsZleftmostZ	remainderZ	wildcardsZfragZpatrrr_dnsname_matchs,


rcCs&t|tr"tjdkr"t|ddd}|S)N)asciistrict)encodingerrors)
isinstancestrsysversion_infoZunicode)objrrr_to_unicodeOsr&cCstt|}||kS)zExact matching of IP addresses.

    RFC 6125 explicitly doesn't define an algorithm for this
    (section 1.7.2 - "Out of Scope").
    )r
ip_addressr&rstrip)Zipnamehost_ipZiprrr_ipaddress_matchUsr*cCs|stdztt|}WnPtk
r6d}Yn:tk
rLd}Yn$tk
rntdkrhd}nYnXg}|dd}|D]^\}}|dkr|dkrt||rdS||q|dkr|dk	rt	||rdS||q|s0|ddD]8}|D].\}}|dkrt||r"dS||qqt
|d	kr\td
|dt
t|fn,t
|d	krtd||d
fntddS)a)Verify that *cert* (in decoded format as returned by
    SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
    rules are followed, but IP addresses are not accepted for *hostname*.

    CertificateError is raised on failure. On success, the function
    returns nothing.
    ztempty or no certificate, match_hostname needs a SSL socket or SSL context with either CERT_OPTIONAL or CERT_REQUIREDNZsubjectAltNamerZDNSz
IP AddressZsubjectZ
commonNamer	z&hostname %r doesn't match either of %sz, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found)
ValueErrorrr'r&UnicodeErrorAttributeErrorgetrrr*lenrrmapr)Zcertrr)ZdnsnamesZsankeyvaluesubrrrmatch_hostnameasZr4)r	)
__doc__rr#Zpip._vendorrImportError__version__r+rrr&r*r4rrrrs

6site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.py000064400000013134151464452710024266 0ustar00"""The match_hostname() function from Python 3.3.3, essential when using SSL."""

# Note: This file is under the PSF license as the code comes from the python
# stdlib.   http://docs.python.org/3/license.html

import re
import sys

# ipaddress has been backported to 2.6+ in pypi.  If it is installed on the
# system, use it to handle IPAddress ServerAltnames (this was added in
# python-3.5) otherwise only do DNS matching.  This allows
# backports.ssl_match_hostname to continue to be used in Python 2.7.
try:
    from pip._vendor import ipaddress
except ImportError:
    ipaddress = None

__version__ = "3.5.0.1"


class CertificateError(ValueError):
    pass


def _dnsname_match(dn, hostname, max_wildcards=1):
    """Matching according to RFC 6125, section 6.4.3

    http://tools.ietf.org/html/rfc6125#section-6.4.3
    """
    pats = []
    if not dn:
        return False

    # Ported from python3-syntax:
    # leftmost, *remainder = dn.split(r'.')
    parts = dn.split(r".")
    leftmost = parts[0]
    remainder = parts[1:]

    wildcards = leftmost.count("*")
    if wildcards > max_wildcards:
        # Issue #17980: avoid denials of service by refusing more
        # than one wildcard per fragment.  A survey of established
        # policy among SSL implementations showed it to be a
        # reasonable choice.
        raise CertificateError(
            "too many wildcards in certificate DNS name: " + repr(dn)
        )

    # speed up common case w/o wildcards
    if not wildcards:
        return dn.lower() == hostname.lower()

    # RFC 6125, section 6.4.3, subitem 1.
    # The client SHOULD NOT attempt to match a presented identifier in which
    # the wildcard character comprises a label other than the left-most label.
    if leftmost == "*":
        # When '*' is a fragment by itself, it matches a non-empty dotless
        # fragment.
        pats.append("[^.]+")
    elif leftmost.startswith("xn--") or hostname.startswith("xn--"):
        # RFC 6125, section 6.4.3, subitem 3.
        # The client SHOULD NOT attempt to match a presented identifier
        # where the wildcard character is embedded within an A-label or
        # U-label of an internationalized domain name.
        pats.append(re.escape(leftmost))
    else:
        # Otherwise, '*' matches any dotless string, e.g. www*
        pats.append(re.escape(leftmost).replace(r"\*", "[^.]*"))

    # add the remaining fragments, ignore any wildcards
    for frag in remainder:
        pats.append(re.escape(frag))

    pat = re.compile(r"\A" + r"\.".join(pats) + r"\Z", re.IGNORECASE)
    return pat.match(hostname)


def _to_unicode(obj):
    if isinstance(obj, str) and sys.version_info < (3,):
        obj = unicode(obj, encoding="ascii", errors="strict")
    return obj


def _ipaddress_match(ipname, host_ip):
    """Exact matching of IP addresses.

    RFC 6125 explicitly doesn't define an algorithm for this
    (section 1.7.2 - "Out of Scope").
    """
    # OpenSSL may add a trailing newline to a subjectAltName's IP address
    # Divergence from upstream: ipaddress can't handle byte str
    ip = ipaddress.ip_address(_to_unicode(ipname).rstrip())
    return ip == host_ip


def match_hostname(cert, hostname):
    """Verify that *cert* (in decoded format as returned by
    SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
    rules are followed, but IP addresses are not accepted for *hostname*.

    CertificateError is raised on failure. On success, the function
    returns nothing.
    """
    if not cert:
        raise ValueError(
            "empty or no certificate, match_hostname needs a "
            "SSL socket or SSL context with either "
            "CERT_OPTIONAL or CERT_REQUIRED"
        )
    try:
        # Divergence from upstream: ipaddress can't handle byte str
        host_ip = ipaddress.ip_address(_to_unicode(hostname))
    except ValueError:
        # Not an IP address (common case)
        host_ip = None
    except UnicodeError:
        # Divergence from upstream: Have to deal with ipaddress not taking
        # byte strings.  addresses should be all ascii, so we consider it not
        # an ipaddress in this case
        host_ip = None
    except AttributeError:
        # Divergence from upstream: Make ipaddress library optional
        if ipaddress is None:
            host_ip = None
        else:
            raise
    dnsnames = []
    san = cert.get("subjectAltName", ())
    for key, value in san:
        if key == "DNS":
            if host_ip is None and _dnsname_match(value, hostname):
                return
            dnsnames.append(value)
        elif key == "IP Address":
            if host_ip is not None and _ipaddress_match(value, host_ip):
                return
            dnsnames.append(value)
    if not dnsnames:
        # The subject is only checked when there is no dNSName entry
        # in subjectAltName
        for sub in cert.get("subject", ()):
            for key, value in sub:
                # XXX according to RFC 2818, the most specific Common Name
                # must be used.
                if key == "commonName":
                    if _dnsname_match(value, hostname):
                        return
                    dnsnames.append(value)
    if len(dnsnames) > 1:
        raise CertificateError(
            "hostname %r "
            "doesn't match either of %s" % (hostname, ", ".join(map(repr, dnsnames)))
        )
    elif len(dnsnames) == 1:
        raise CertificateError(
            "hostname %r " "doesn't match %r" % (hostname, dnsnames[0])
        )
    else:
        raise CertificateError(
            "no appropriate commonName or " "subjectAltName fields were found"
        )
site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.py000064400000001260151464452710022636 0ustar00import sys

try:
    # Our match_hostname function is the same as 3.5's, so we only want to
    # import the match_hostname function if it's at least that good.
    if sys.version_info < (3, 5):
        raise ImportError("Fallback to vendored code")

    from ssl import CertificateError, match_hostname
except ImportError:
    try:
        # Backport of the function from a pypi module
        from backports.ssl_match_hostname import CertificateError, match_hostname
    except ImportError:
        # Our vendored copy
        from ._implementation import CertificateError, match_hostname

# Not needed, but documenting what we provide.
__all__ = ("CertificateError", "match_hostname")
site-packages/pip/_vendor/urllib3/packages/__pycache__/__init__.cpython-38.pyc000064400000000425151464452710023253 0ustar00U

.el@s ddlmZddlmZdZdS))absolute_import)ssl_match_hostnameN)Z
__future__rr__all__rrI/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/__init__.pyssite-packages/pip/_vendor/urllib3/packages/__pycache__/six.cpython-38.opt-1.pyc000064400000063571151464452710023271 0ustar00U

.eC@sdZddlmZddlZddlZddlZddlZddlZdZdZ	ej
ddkZej
ddkZej
dddkZ
erefZefZefZeZeZejZn~efZeefZeejfZeZeZejd	red
ZnHGdddeZ ze!e Wne"k
red
ZYn
Xed
Z[ ddZ#ddZ$GdddeZ%Gddde%Z&Gdddej'Z(Gddde%Z)GdddeZ*e*e+Z,Gddde(Z-e)dddd e)d!d"d#d$d!e)d%d"d"d&d%e)d'd(d#d)d'e)d*d(d+e)d,d"d#d-d,e)d.d/d/d0d.e)d1d/d/d.d1e)d2d3d4e)d5d(d#d6d5e)d7d(e
rd8nd9d:e)d;d(d<e)d=d>d?d@e)d d de)dAdAdBe)dCdCdBe)dDdDdBe)d6d(d#d6d5e)dEd"d#dFdEe)dGd"d"dHdGe&d#d(e&dIdJe&dKdLe&dMdNdOe&dPdQdPe&dRdSdTe&dUdVdWe&dXdYdZe&d[d\d]e&d^d_d`e&dadbdce&dddedfe&dgdhdie&djdkdle&dmdndoe&dpdpdqe&drdrdqe&dsdsdqe&dtdtdue&dvdwe&dxdye&dzd{e&d|d}d|e&d~de&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&de+dde&de+dde&de+de+de&ddde&ddde&dddg@Z.ejdkrfe.e&ddg7Z.e.D]2Z/e0e-e/j1e/e2e/e&rje,3e/de/j1qj[/e.e-_.e-e+dZ4e,3e4dGddde(Z5e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d@dde)ddde)ddde)ddde)ddddd΃e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)dddgZ6e6D]Z/e0e5e/j1e/q[/e6e5_.e,3e5e+dddڡGdd܄de(Z7e)ddde)ddde)dddgZ8e8D]Z/e0e7e/j1e/q>[/e8e7_.e,3e7e+dddGddde(Z9e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d	ddg#Z:e:D]Z/e0e9e/j1e/q[/e:e9_.e,3e9e+d
ddGd
dde(Z;e)ddde)ddde)ddde)dddgZe>D]Z/e0e=e/j1e/q[/e>e=_.e,3e=e+dddGdddej'Z?e,3e?e+ddd d!Z@d"d#ZAe	rd$ZBd%ZCd&ZDd'ZEd(ZFd)ZGn$d*ZBd+ZCd,ZDd-ZEd.ZFd/ZGzeHZIWn"eJk
	rd0d1ZIYnXeIZHzeKZKWn"eJk

rd2d3ZKYnXe
rFd4d5ZLejMZNd6d7ZOeZPn>d8d5ZLd9d:ZNd;d7ZOGd<d=d=eZPeKZKe#eLd>eQeBZReQeCZSeQeDZTeQeEZUeQeFZVeQeGZWer(d?d@ZXdAdBZYdCdDZZdEdFZ[e\dGZ]e\dHZ^e\dIZ_nTdJd@ZXdKdBZYdLdDZZdMdFZ[e\dNZ]e\dOZ^e\dPZ_e#eXdQe#eYdRe#eZdSe#e[dTerFdUdVZ`dWdXZaebZcddldZdededYjfZg[dehdZiejjZkelZmddlnZnenjoZoenjpZp[ndZZqej
d[d[kr8d\Zrd]Zsnd^Zrd_Zsnjd`dVZ`dadXZaecZcebZgdbdcZidddeZketejuevZmddloZoeojoZoZpdfZqd\Zrd]Zse#e`dge#eadhdidZZwdjd^Zxdkd_Zye
reze4j{dlZ|ddmdnZ}nddodpZ|e|dqej
dddrk
rNe|dsn.ej
dddrk
rpe|dtndudvZ~eze4j{dwdZedk
rdxdyZej
dddzk
reZd{dyZe#e}d|ej
dddkrejejfd}d~ZnejZddZddZdddZdddZdddZddZgZe+Zeddk	r|ge_ejreejD]4\ZZeej+dkrej1e+kreje=qĐq[[eje,dS(z6Utilities for writing code that runs on Python 2 and 3)absolute_importNz'Benjamin Peterson z1.12.0)rjavaic@seZdZddZdS)XcCsdS)NlselfrrD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/six.py__len__=sz	X.__len__N)__name__
__module____qualname__rrrrrr<srlcCs
||_dS)z Add documentation to a function.N)__doc__)funcdocrrr_add_docKsrcCst|tj|S)z7Import module, returning the module after the last dot.)
__import__sysmodulesnamerrr_import_modulePsrc@seZdZddZddZdS)
_LazyDescrcCs
||_dSNrr
rrrr__init__Wsz_LazyDescr.__init__cCsB|}t||j|zt|j|jWntk
r<YnX|Sr)_resolvesetattrrdelattr	__class__AttributeError)r
objtpresultrrr__get__Zsz_LazyDescr.__get__N)r
rrrr&rrrrrVsrcs.eZdZdfdd	ZddZddZZS)	MovedModuleNcs2tt||tr(|dkr |}||_n||_dSr)superr'rPY3mod)r
roldnewr!rrrgszMovedModule.__init__cCs
t|jSr)rr*r	rrrrpszMovedModule._resolvecCs"|}t||}t||||Sr)rgetattrr)r
attr_modulevaluerrr__getattr__ss
zMovedModule.__getattr__)N)r
rrrrr2
__classcell__rrr-rr'fs	r'cs(eZdZfddZddZgZZS)_LazyModulecstt|||jj|_dSr)r(r4rr!rrr-rrr{sz_LazyModule.__init__cCs ddg}|dd|jD7}|S)Nrr
cSsg|]
}|jqSrr).0r/rrr
sz'_LazyModule.__dir__..)_moved_attributes)r
Zattrsrrr__dir__sz_LazyModule.__dir__)r
rrrr8r7r3rrr-rr4zsr4cs&eZdZdfdd	ZddZZS)MovedAttributeNcsdtt||trH|dkr |}||_|dkr@|dkr<|}n|}||_n||_|dkrZ|}||_dSr)r(r9rr)r*r/)r
rZold_modZnew_modZold_attrZnew_attrr-rrrszMovedAttribute.__init__cCst|j}t||jSr)rr*r.r/)r
modulerrrrs
zMovedAttribute._resolve)NN)r
rrrrr3rrr-rr9sr9c@sVeZdZdZddZddZddZdd	d
ZddZd
dZ	ddZ
ddZeZdS)_SixMetaPathImporterz
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    cCs||_i|_dSr)r
known_modules)r
Zsix_module_namerrrrsz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdSN.r<r)r
r*Z	fullnamesfullnamerrr_add_modulesz _SixMetaPathImporter._add_modulecCs|j|jd|Sr=r?r
r@rrr_get_modulesz _SixMetaPathImporter._get_moduleNcCs||jkr|SdSr)r<)r
r@pathrrrfind_modules
z _SixMetaPathImporter.find_modulecCs2z|j|WStk
r,td|YnXdS)Nz!This loader does not know module )r<KeyErrorImportErrorrBrrrZ__get_modulesz!_SixMetaPathImporter.__get_modulecCsTztj|WStk
r YnX||}t|tr@|}n||_|tj|<|Sr)rrrF _SixMetaPathImporter__get_module
isinstancer'r
__loader__)r
r@r*rrrload_modules



z _SixMetaPathImporter.load_modulecCst||dS)z
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        __path__)hasattrrHrBrrr
is_packagesz_SixMetaPathImporter.is_packagecCs||dS)z;Return None

        Required, if is_package is implementedN)rHrBrrrget_codes
z_SixMetaPathImporter.get_code)N)
r
rrrrrArCrErHrKrNrO
get_sourcerrrrr;s
	r;c@seZdZdZgZdS)_MovedItemszLazy loading of moved objectsN)r
rrrrLrrrrrQsrQZ	cStringIOioStringIOfilter	itertoolsbuiltinsZifilterfilterfalseZifilterfalseinputZ__builtin__Z	raw_inputinternrmapimapgetcwdosZgetcwdugetcwdbZ	getoutputZcommands
subprocessrangeZxrangeZ
reload_module	importlibZimpreloadreduce	functoolsZshlex_quoteZpipesZshlexZquoteUserDictcollectionsUserList
UserStringzipZizipzip_longestZizip_longestZconfigparserZConfigParsercopyregZcopy_regZdbm_gnuZgdbmzdbm.gnuZ
_dummy_threadZdummy_threadZhttp_cookiejarZ	cookielibzhttp.cookiejarZhttp_cookiesZCookiezhttp.cookiesZ
html_entitiesZhtmlentitydefsz
html.entitiesZhtml_parserZ
HTMLParserzhtml.parserZhttp_clientZhttplibzhttp.clientZemail_mime_basezemail.MIMEBasezemail.mime.baseZemail_mime_imagezemail.MIMEImagezemail.mime.imageZemail_mime_multipartzemail.MIMEMultipartzemail.mime.multipartZemail_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipartZemail_mime_textzemail.MIMETextzemail.mime.textZBaseHTTPServerzhttp.serverZ
CGIHTTPServerZSimpleHTTPServerZcPicklepickleZqueueZQueuereprlibreprZsocketserverZSocketServer_threadthreadZtkinterZTkinterZtkinter_dialogZDialogztkinter.dialogZtkinter_filedialogZ
FileDialogztkinter.filedialogZtkinter_scrolledtextZScrolledTextztkinter.scrolledtextZtkinter_simpledialogZSimpleDialogztkinter.simpledialogZtkinter_tixZTixztkinter.tixZtkinter_ttkZttkztkinter.ttkZtkinter_constantsZTkconstantsztkinter.constantsZtkinter_dndZTkdndztkinter.dndZtkinter_colorchooserZtkColorChooserztkinter.colorchooserZtkinter_commondialogZtkCommonDialogztkinter.commondialogZtkinter_tkfiledialogZtkFileDialogZtkinter_fontZtkFontztkinter.fontZtkinter_messageboxZtkMessageBoxztkinter.messageboxZtkinter_tksimpledialogZtkSimpleDialogZurllib_parsez.moves.urllib_parsezurllib.parseZurllib_errorz.moves.urllib_errorzurllib.errorZurllibz
.moves.urllibZurllib_robotparserrobotparserzurllib.robotparserZ
xmlrpc_clientZ	xmlrpclibz
xmlrpc.clientZ
xmlrpc_serverZSimpleXMLRPCServerz
xmlrpc.serverZwin32winreg_winregzmoves.z.movesmovesc@seZdZdZdS)Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseNr
rrrrrrrruDsruZParseResultZurlparseZSplitResultZparse_qsZ	parse_qslZ	urldefragZurljoinZurlsplitZ
urlunparseZ
urlunsplitZ
quote_plusZunquoteZunquote_plusZunquote_to_bytesZ	urlencodeZ
splitqueryZsplittagZ	splituserZ
splitvalueZ
uses_fragmentZuses_netlocZuses_paramsZ
uses_queryZ
uses_relativemoves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNrvrrrrrxssrxZURLErrorZurllib2Z	HTTPErrorZContentTooShortErrorz.moves.urllib.errormoves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNrvrrrrrzsrzZurlopenzurllib.requestZinstall_openerZbuild_openerZpathname2urlZurl2pathnameZ
getproxiesZRequestZOpenerDirectorZHTTPDefaultErrorHandlerZHTTPRedirectHandlerZHTTPCookieProcessorZProxyHandlerZBaseHandlerZHTTPPasswordMgrZHTTPPasswordMgrWithDefaultRealmZAbstractBasicAuthHandlerZHTTPBasicAuthHandlerZProxyBasicAuthHandlerZAbstractDigestAuthHandlerZHTTPDigestAuthHandlerZProxyDigestAuthHandlerZHTTPHandlerZHTTPSHandlerZFileHandlerZ
FTPHandlerZCacheFTPHandlerZUnknownHandlerZHTTPErrorProcessorZurlretrieveZ
urlcleanupZ	URLopenerZFancyURLopenerZproxy_bypassZparse_http_listZparse_keqv_listz.moves.urllib.requestmoves.urllib_requestzmoves.urllib.requestc@seZdZdZdS) Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNrvrrrrr|sr|Zaddbasezurllib.responseZaddclosehookZaddinfoZ
addinfourlz.moves.urllib.responsemoves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNrvrrrrr~sr~ZRobotFileParserz.moves.urllib.robotparsermoves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZedZedZedZ	edZ
edZddZd	S)
Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacerwryr{r}rcCsdddddgS)Nparseerrorrequestresponserqrr	rrrr8szModule_six_moves_urllib.__dir__N)
r
rrrrL	_importerrCrrrrrqr8rrrrrs




rzmoves.urllibcCstt|j|dS)zAdd an item to six.moves.N)rrQr)Zmoverrradd_movesrcCsXztt|WnDtk
rRztj|=Wn"tk
rLtd|fYnXYnXdS)zRemove item from six.moves.zno such move, %rN)r rQr"rt__dict__rFrrrrremove_move	sr__func____self____closure____code____defaults____globals__im_funcZim_selfZfunc_closureZ	func_codeZ
func_defaultsZfunc_globalscCs|Sr)next)itrrradvance_iterator*srcCstddt|jDS)Ncss|]}d|jkVqdS)__call__N)r)r5klassrrr	6szcallable..)anytype__mro__)r#rrrcallable5srcCs|SrrZunboundrrrget_unbound_function;srcCs|Srrrclsrrrcreate_unbound_method@srcCs|jSr)rrrrrrFscCst|||jSr)types
MethodTyper!)rr#rrrcreate_bound_methodIsrcCst|d|Sr)rrrrrrrLsc@seZdZddZdS)IteratorcCst||Sr)r__next__r	rrrrPsz
Iterator.nextN)r
rrrrrrrrOsrz3Get the function out of a possibly unbound functioncKst|jf|Sr)iterkeysdkwrrriterkeyscsrcKst|jf|Sr)rvaluesrrrr
itervaluesfsrcKst|jf|Sr)ritemsrrrr	iteritemsisrcKst|jf|Sr)rZlistsrrrr	iterlistslsrrrrcKs|jf|Sr)rrrrrrvscKs|jf|Sr)rrrrrryscKs|jf|Sr)rrrrrr|scKs|jf|Sr)rrrrrrsviewkeys
viewvalues	viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs
|dS)Nzlatin-1)encodesrrrbsrcCs|Srrrrrrusrz>BassertCountEqualZassertRaisesRegexpZassertRegexpMatchesassertRaisesRegexassertRegexcCs|SrrrrrrrscCst|dddS)Nz\\z\\\\Zunicode_escape)unicodereplacerrrrrscCst|dS)Nrord)Zbsrrrbyte2intsrcCst||Srr)Zbufirrr
indexbytessrZassertItemsEqualzByte literalzText literalcOst|t||Sr)r._assertCountEqualr
argskwargsrrrrscOst|t||Sr)r._assertRaisesRegexrrrrrscOst|t||Sr)r._assertRegexrrrrrsexeccCs:z*|dkr|}|j|k	r$|||W5d}d}XdSr)
__traceback__with_traceback)r$r1tbrrrreraises

rcCsB|dkr*td}|j}|dkr&|j}~n|dkr6|}tddS)zExecute code in a namespace.Nrzexec _code_ in _globs_, _locs_)r	_getframe	f_globalsf_localsr)Z_code_Z_globs_Z_locs_framerrrexec_s
rzedef reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
)rrzdef raise_from(value, from_value):
    try:
        if from_value is None:
            raise value
        raise value from from_value
    finally:
        value = None
zrdef raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
cCs|dSrr)r1Z
from_valuerrr
raise_fromsrprintc
s.|dtjdkrdSfdd}d}|dd}|dk	r`t|trNd}nt|ts`td|d	d}|dk	rt|trd}nt|tstd
|rtd|s|D]}t|trd}qq|rtd}td
}nd}d
}|dkr|}|dkr|}t|D] \}	}|	r||||q||dS)z4The new-style print function for Python 2.4 and 2.5.fileNcsdt|tst|}ttrVt|trVjdk	rVtdd}|dkrHd}|j|}|dS)Nerrorsstrict)	rI
basestringstrrrencodingr.rwrite)datarfprrr s
zprint_..writeFsepTzsep must be None or a stringendzend must be None or a stringz$invalid keyword arguments to print()
 )poprstdoutrIrr	TypeError	enumerate)
rrrZwant_unicoderrargnewlineZspacerrrrprint_sL





r)rrcOs<|dtj}|dd}t|||r8|dk	r8|dS)NrflushF)getrrr_printr)rrrrrrrrWs

zReraise an exception.csfdd}|S)Ncst|}|_|Sr)rdwraps__wrapped__)fassignedupdatedwrappedrrwrapperhszwraps..wrapperr)rrrrrrrrcsrcs&Gfdddt}t|ddiS)z%Create a base class with a metaclass.cs,eZdZfddZefddZdS)z!with_metaclass..metaclasscs||Srr)rr
this_basesrbasesmetarr__new__zsz)with_metaclass..metaclass.__new__cs|Sr)__prepare__)rrrrrrr}sz-with_metaclass..metaclass.__prepare__N)r
rrrclassmethodrrrrr	metaclassysrZtemporary_classr)rr)rrrrrrwith_metaclasstsrcsfdd}|S)z6Class decorator for creating a class with a metaclass.cs||j}|d}|dk	r@t|tr,|g}|D]}||q0|dd|ddt|drl|j|d<|j|j	|S)N	__slots__r__weakref__r)
rcopyrrIrrrMrr
	__bases__)rZ	orig_varsslotsZ	slots_varrrrrs




zadd_metaclass..wrapperr)rrrrr
add_metaclasssrutf-8rcCs8t|tr|||St|tr$|Stdt|dS)zCoerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    not expecting type '%s'N)rI	text_typerbinary_typerrrrrrrr
ensure_binarys


r	cCsXt|ttfstdt|tr:t|tr:|||}ntrTt|trT|||}|S)zCoerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    r)	rIrrrrPY2rr)decoderrrr
ensure_strsrcCs8t|tr|||St|tr$|Stdt|dS)zCoerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    rN)rIrrrrrrrrrensure_texts


r
cCs2tr.d|jkrtd|j|j|_dd|_|S)a
    A decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    __str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|dS)Nr)__unicode__rr	rrrz-python_2_unicode_compatible..)r
r
ValueErrorr
rr)rrrrpython_2_unicode_compatibles

r__spec__)N)NN)rr)rr)rr)rZ
__future__rrdrUoperatorrr
__author____version__version_infor
r)ZPY34rZstring_typesintZ
integer_typesrZclass_typesrbytesrmaxsizeZMAXSIZErZlongZ	ClassTyperplatform
startswithobjectrlen
OverflowErrorrrrr'
ModuleTyper4r9r;r
rrQr7r/rrrIrArtruZ_urllib_parse_moved_attributesrxZ_urllib_error_moved_attributesrzZ _urllib_request_moved_attributesr|Z!_urllib_response_moved_attributesr~Z$_urllib_robotparser_moved_attributesrrrZ
_meth_funcZ
_meth_selfZ
_func_closureZ
_func_codeZ_func_defaultsZ
_func_globalsrr	NameErrorrrrrrr
attrgetterZget_method_functionZget_method_selfZget_function_closureZget_function_codeZget_function_defaultsZget_function_globalsrrrrmethodcallerrrrrrchrZunichrstructStructpackZint2byte
itemgetterrgetitemrrZ	iterbytesrRrSBytesIOrrrpartialr[rrrrr.rVrrrrrWRAPPER_ASSIGNMENTSWRAPPER_UPDATESrrrr	rr
rrL__package__globalsrrsubmodule_search_locations	meta_pathrrZimporterappendrrrrsN

@







































K





















































%












:
site-packages/pip/_vendor/urllib3/packages/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000425151464452710024212 0ustar00U

.el@s ddlmZddlmZdZdS))absolute_import)ssl_match_hostnameN)Z
__future__rr__all__rrI/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/__init__.pyssite-packages/pip/_vendor/urllib3/packages/__pycache__/six.cpython-38.pyc000064400000063571151464452710022332 0ustar00U

.eC@sdZddlmZddlZddlZddlZddlZddlZdZdZ	ej
ddkZej
ddkZej
dddkZ
erefZefZefZeZeZejZn~efZeefZeejfZeZeZejd	red
ZnHGdddeZ ze!e Wne"k
red
ZYn
Xed
Z[ ddZ#ddZ$GdddeZ%Gddde%Z&Gdddej'Z(Gddde%Z)GdddeZ*e*e+Z,Gddde(Z-e)dddd e)d!d"d#d$d!e)d%d"d"d&d%e)d'd(d#d)d'e)d*d(d+e)d,d"d#d-d,e)d.d/d/d0d.e)d1d/d/d.d1e)d2d3d4e)d5d(d#d6d5e)d7d(e
rd8nd9d:e)d;d(d<e)d=d>d?d@e)d d de)dAdAdBe)dCdCdBe)dDdDdBe)d6d(d#d6d5e)dEd"d#dFdEe)dGd"d"dHdGe&d#d(e&dIdJe&dKdLe&dMdNdOe&dPdQdPe&dRdSdTe&dUdVdWe&dXdYdZe&d[d\d]e&d^d_d`e&dadbdce&dddedfe&dgdhdie&djdkdle&dmdndoe&dpdpdqe&drdrdqe&dsdsdqe&dtdtdue&dvdwe&dxdye&dzd{e&d|d}d|e&d~de&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&de+dde&de+dde&de+de+de&ddde&ddde&dddg@Z.ejdkrfe.e&ddg7Z.e.D]2Z/e0e-e/j1e/e2e/e&rje,3e/de/j1qj[/e.e-_.e-e+dZ4e,3e4dGddde(Z5e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d@dde)ddde)ddde)ddde)ddddd΃e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)dddgZ6e6D]Z/e0e5e/j1e/q[/e6e5_.e,3e5e+dddڡGdd܄de(Z7e)ddde)ddde)dddgZ8e8D]Z/e0e7e/j1e/q>[/e8e7_.e,3e7e+dddGddde(Z9e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d	ddg#Z:e:D]Z/e0e9e/j1e/q[/e:e9_.e,3e9e+d
ddGd
dde(Z;e)ddde)ddde)ddde)dddgZe>D]Z/e0e=e/j1e/q[/e>e=_.e,3e=e+dddGdddej'Z?e,3e?e+ddd d!Z@d"d#ZAe	rd$ZBd%ZCd&ZDd'ZEd(ZFd)ZGn$d*ZBd+ZCd,ZDd-ZEd.ZFd/ZGzeHZIWn"eJk
	rd0d1ZIYnXeIZHzeKZKWn"eJk

rd2d3ZKYnXe
rFd4d5ZLejMZNd6d7ZOeZPn>d8d5ZLd9d:ZNd;d7ZOGd<d=d=eZPeKZKe#eLd>eQeBZReQeCZSeQeDZTeQeEZUeQeFZVeQeGZWer(d?d@ZXdAdBZYdCdDZZdEdFZ[e\dGZ]e\dHZ^e\dIZ_nTdJd@ZXdKdBZYdLdDZZdMdFZ[e\dNZ]e\dOZ^e\dPZ_e#eXdQe#eYdRe#eZdSe#e[dTerFdUdVZ`dWdXZaebZcddldZdededYjfZg[dehdZiejjZkelZmddlnZnenjoZoenjpZp[ndZZqej
d[d[kr8d\Zrd]Zsnd^Zrd_Zsnjd`dVZ`dadXZaecZcebZgdbdcZidddeZketejuevZmddloZoeojoZoZpdfZqd\Zrd]Zse#e`dge#eadhdidZZwdjd^Zxdkd_Zye
reze4j{dlZ|ddmdnZ}nddodpZ|e|dqej
dddrk
rNe|dsn.ej
dddrk
rpe|dtndudvZ~eze4j{dwdZedk
rdxdyZej
dddzk
reZd{dyZe#e}d|ej
dddkrejejfd}d~ZnejZddZddZdddZdddZdddZddZgZe+Zeddk	r|ge_ejreejD]4\ZZeej+dkrej1e+kreje=qĐq[[eje,dS(z6Utilities for writing code that runs on Python 2 and 3)absolute_importNz'Benjamin Peterson z1.12.0)rjavaic@seZdZddZdS)XcCsdS)NlselfrrD/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/six.py__len__=sz	X.__len__N)__name__
__module____qualname__rrrrrr<srlcCs
||_dS)z Add documentation to a function.N)__doc__)funcdocrrr_add_docKsrcCst|tj|S)z7Import module, returning the module after the last dot.)
__import__sysmodulesnamerrr_import_modulePsrc@seZdZddZddZdS)
_LazyDescrcCs
||_dSNrr
rrrr__init__Wsz_LazyDescr.__init__cCsB|}t||j|zt|j|jWntk
r<YnX|Sr)_resolvesetattrrdelattr	__class__AttributeError)r
objtpresultrrr__get__Zsz_LazyDescr.__get__N)r
rrrr&rrrrrVsrcs.eZdZdfdd	ZddZddZZS)	MovedModuleNcs2tt||tr(|dkr |}||_n||_dSr)superr'rPY3mod)r
roldnewr!rrrgszMovedModule.__init__cCs
t|jSr)rr*r	rrrrpszMovedModule._resolvecCs"|}t||}t||||Sr)rgetattrr)r
attr_modulevaluerrr__getattr__ss
zMovedModule.__getattr__)N)r
rrrrr2
__classcell__rrr-rr'fs	r'cs(eZdZfddZddZgZZS)_LazyModulecstt|||jj|_dSr)r(r4rr!rrr-rrr{sz_LazyModule.__init__cCs ddg}|dd|jD7}|S)Nrr
cSsg|]
}|jqSrr).0r/rrr
sz'_LazyModule.__dir__..)_moved_attributes)r
Zattrsrrr__dir__sz_LazyModule.__dir__)r
rrrr8r7r3rrr-rr4zsr4cs&eZdZdfdd	ZddZZS)MovedAttributeNcsdtt||trH|dkr |}||_|dkr@|dkr<|}n|}||_n||_|dkrZ|}||_dSr)r(r9rr)r*r/)r
rZold_modZnew_modZold_attrZnew_attrr-rrrszMovedAttribute.__init__cCst|j}t||jSr)rr*r.r/)r
modulerrrrs
zMovedAttribute._resolve)NN)r
rrrrr3rrr-rr9sr9c@sVeZdZdZddZddZddZdd	d
ZddZd
dZ	ddZ
ddZeZdS)_SixMetaPathImporterz
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    cCs||_i|_dSr)r
known_modules)r
Zsix_module_namerrrrsz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdSN.r<r)r
r*Z	fullnamesfullnamerrr_add_modulesz _SixMetaPathImporter._add_modulecCs|j|jd|Sr=r?r
r@rrr_get_modulesz _SixMetaPathImporter._get_moduleNcCs||jkr|SdSr)r<)r
r@pathrrrfind_modules
z _SixMetaPathImporter.find_modulecCs2z|j|WStk
r,td|YnXdS)Nz!This loader does not know module )r<KeyErrorImportErrorrBrrrZ__get_modulesz!_SixMetaPathImporter.__get_modulecCsTztj|WStk
r YnX||}t|tr@|}n||_|tj|<|Sr)rrrF _SixMetaPathImporter__get_module
isinstancer'r
__loader__)r
r@r*rrrload_modules



z _SixMetaPathImporter.load_modulecCst||dS)z
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        __path__)hasattrrHrBrrr
is_packagesz_SixMetaPathImporter.is_packagecCs||dS)z;Return None

        Required, if is_package is implementedN)rHrBrrrget_codes
z_SixMetaPathImporter.get_code)N)
r
rrrrrArCrErHrKrNrO
get_sourcerrrrr;s
	r;c@seZdZdZgZdS)_MovedItemszLazy loading of moved objectsN)r
rrrrLrrrrrQsrQZ	cStringIOioStringIOfilter	itertoolsbuiltinsZifilterfilterfalseZifilterfalseinputZ__builtin__Z	raw_inputinternrmapimapgetcwdosZgetcwdugetcwdbZ	getoutputZcommands
subprocessrangeZxrangeZ
reload_module	importlibZimpreloadreduce	functoolsZshlex_quoteZpipesZshlexZquoteUserDictcollectionsUserList
UserStringzipZizipzip_longestZizip_longestZconfigparserZConfigParsercopyregZcopy_regZdbm_gnuZgdbmzdbm.gnuZ
_dummy_threadZdummy_threadZhttp_cookiejarZ	cookielibzhttp.cookiejarZhttp_cookiesZCookiezhttp.cookiesZ
html_entitiesZhtmlentitydefsz
html.entitiesZhtml_parserZ
HTMLParserzhtml.parserZhttp_clientZhttplibzhttp.clientZemail_mime_basezemail.MIMEBasezemail.mime.baseZemail_mime_imagezemail.MIMEImagezemail.mime.imageZemail_mime_multipartzemail.MIMEMultipartzemail.mime.multipartZemail_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipartZemail_mime_textzemail.MIMETextzemail.mime.textZBaseHTTPServerzhttp.serverZ
CGIHTTPServerZSimpleHTTPServerZcPicklepickleZqueueZQueuereprlibreprZsocketserverZSocketServer_threadthreadZtkinterZTkinterZtkinter_dialogZDialogztkinter.dialogZtkinter_filedialogZ
FileDialogztkinter.filedialogZtkinter_scrolledtextZScrolledTextztkinter.scrolledtextZtkinter_simpledialogZSimpleDialogztkinter.simpledialogZtkinter_tixZTixztkinter.tixZtkinter_ttkZttkztkinter.ttkZtkinter_constantsZTkconstantsztkinter.constantsZtkinter_dndZTkdndztkinter.dndZtkinter_colorchooserZtkColorChooserztkinter.colorchooserZtkinter_commondialogZtkCommonDialogztkinter.commondialogZtkinter_tkfiledialogZtkFileDialogZtkinter_fontZtkFontztkinter.fontZtkinter_messageboxZtkMessageBoxztkinter.messageboxZtkinter_tksimpledialogZtkSimpleDialogZurllib_parsez.moves.urllib_parsezurllib.parseZurllib_errorz.moves.urllib_errorzurllib.errorZurllibz
.moves.urllibZurllib_robotparserrobotparserzurllib.robotparserZ
xmlrpc_clientZ	xmlrpclibz
xmlrpc.clientZ
xmlrpc_serverZSimpleXMLRPCServerz
xmlrpc.serverZwin32winreg_winregzmoves.z.movesmovesc@seZdZdZdS)Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseNr
rrrrrrrruDsruZParseResultZurlparseZSplitResultZparse_qsZ	parse_qslZ	urldefragZurljoinZurlsplitZ
urlunparseZ
urlunsplitZ
quote_plusZunquoteZunquote_plusZunquote_to_bytesZ	urlencodeZ
splitqueryZsplittagZ	splituserZ
splitvalueZ
uses_fragmentZuses_netlocZuses_paramsZ
uses_queryZ
uses_relativemoves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNrvrrrrrxssrxZURLErrorZurllib2Z	HTTPErrorZContentTooShortErrorz.moves.urllib.errormoves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNrvrrrrrzsrzZurlopenzurllib.requestZinstall_openerZbuild_openerZpathname2urlZurl2pathnameZ
getproxiesZRequestZOpenerDirectorZHTTPDefaultErrorHandlerZHTTPRedirectHandlerZHTTPCookieProcessorZProxyHandlerZBaseHandlerZHTTPPasswordMgrZHTTPPasswordMgrWithDefaultRealmZAbstractBasicAuthHandlerZHTTPBasicAuthHandlerZProxyBasicAuthHandlerZAbstractDigestAuthHandlerZHTTPDigestAuthHandlerZProxyDigestAuthHandlerZHTTPHandlerZHTTPSHandlerZFileHandlerZ
FTPHandlerZCacheFTPHandlerZUnknownHandlerZHTTPErrorProcessorZurlretrieveZ
urlcleanupZ	URLopenerZFancyURLopenerZproxy_bypassZparse_http_listZparse_keqv_listz.moves.urllib.requestmoves.urllib_requestzmoves.urllib.requestc@seZdZdZdS) Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNrvrrrrr|sr|Zaddbasezurllib.responseZaddclosehookZaddinfoZ
addinfourlz.moves.urllib.responsemoves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNrvrrrrr~sr~ZRobotFileParserz.moves.urllib.robotparsermoves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZedZedZedZ	edZ
edZddZd	S)
Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacerwryr{r}rcCsdddddgS)Nparseerrorrequestresponserqrr	rrrr8szModule_six_moves_urllib.__dir__N)
r
rrrrL	_importerrCrrrrrqr8rrrrrs




rzmoves.urllibcCstt|j|dS)zAdd an item to six.moves.N)rrQr)Zmoverrradd_movesrcCsXztt|WnDtk
rRztj|=Wn"tk
rLtd|fYnXYnXdS)zRemove item from six.moves.zno such move, %rN)r rQr"rt__dict__rFrrrrremove_move	sr__func____self____closure____code____defaults____globals__im_funcZim_selfZfunc_closureZ	func_codeZ
func_defaultsZfunc_globalscCs|Sr)next)itrrradvance_iterator*srcCstddt|jDS)Ncss|]}d|jkVqdS)__call__N)r)r5klassrrr	6szcallable..)anytype__mro__)r#rrrcallable5srcCs|SrrZunboundrrrget_unbound_function;srcCs|Srrrclsrrrcreate_unbound_method@srcCs|jSr)rrrrrrFscCst|||jSr)types
MethodTyper!)rr#rrrcreate_bound_methodIsrcCst|d|Sr)rrrrrrrLsc@seZdZddZdS)IteratorcCst||Sr)r__next__r	rrrrPsz
Iterator.nextN)r
rrrrrrrrOsrz3Get the function out of a possibly unbound functioncKst|jf|Sr)iterkeysdkwrrriterkeyscsrcKst|jf|Sr)rvaluesrrrr
itervaluesfsrcKst|jf|Sr)ritemsrrrr	iteritemsisrcKst|jf|Sr)rZlistsrrrr	iterlistslsrrrrcKs|jf|Sr)rrrrrrvscKs|jf|Sr)rrrrrryscKs|jf|Sr)rrrrrr|scKs|jf|Sr)rrrrrrsviewkeys
viewvalues	viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs
|dS)Nzlatin-1)encodesrrrbsrcCs|Srrrrrrusrz>BassertCountEqualZassertRaisesRegexpZassertRegexpMatchesassertRaisesRegexassertRegexcCs|SrrrrrrrscCst|dddS)Nz\\z\\\\Zunicode_escape)unicodereplacerrrrrscCst|dS)Nrord)Zbsrrrbyte2intsrcCst||Srr)Zbufirrr
indexbytessrZassertItemsEqualzByte literalzText literalcOst|t||Sr)r._assertCountEqualr
argskwargsrrrrscOst|t||Sr)r._assertRaisesRegexrrrrrscOst|t||Sr)r._assertRegexrrrrrsexeccCs:z*|dkr|}|j|k	r$|||W5d}d}XdSr)
__traceback__with_traceback)r$r1tbrrrreraises

rcCsB|dkr*td}|j}|dkr&|j}~n|dkr6|}tddS)zExecute code in a namespace.Nrzexec _code_ in _globs_, _locs_)r	_getframe	f_globalsf_localsr)Z_code_Z_globs_Z_locs_framerrrexec_s
rzedef reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
)rrzdef raise_from(value, from_value):
    try:
        if from_value is None:
            raise value
        raise value from from_value
    finally:
        value = None
zrdef raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
cCs|dSrr)r1Z
from_valuerrr
raise_fromsrprintc
s.|dtjdkrdSfdd}d}|dd}|dk	r`t|trNd}nt|ts`td|d	d}|dk	rt|trd}nt|tstd
|rtd|s|D]}t|trd}qq|rtd}td
}nd}d
}|dkr|}|dkr|}t|D] \}	}|	r||||q||dS)z4The new-style print function for Python 2.4 and 2.5.fileNcsdt|tst|}ttrVt|trVjdk	rVtdd}|dkrHd}|j|}|dS)Nerrorsstrict)	rI
basestringstrrrencodingr.rwrite)datarfprrr s
zprint_..writeFsepTzsep must be None or a stringendzend must be None or a stringz$invalid keyword arguments to print()
 )poprstdoutrIrr	TypeError	enumerate)
rrrZwant_unicoderrargnewlineZspacerrrrprint_sL





r)rrcOs<|dtj}|dd}t|||r8|dk	r8|dS)NrflushF)getrrr_printr)rrrrrrrrWs

zReraise an exception.csfdd}|S)Ncst|}|_|Sr)rdwraps__wrapped__)fassignedupdatedwrappedrrwrapperhszwraps..wrapperr)rrrrrrrrcsrcs&Gfdddt}t|ddiS)z%Create a base class with a metaclass.cs,eZdZfddZefddZdS)z!with_metaclass..metaclasscs||Srr)rr
this_basesrbasesmetarr__new__zsz)with_metaclass..metaclass.__new__cs|Sr)__prepare__)rrrrrrr}sz-with_metaclass..metaclass.__prepare__N)r
rrrclassmethodrrrrr	metaclassysrZtemporary_classr)rr)rrrrrrwith_metaclasstsrcsfdd}|S)z6Class decorator for creating a class with a metaclass.cs||j}|d}|dk	r@t|tr,|g}|D]}||q0|dd|ddt|drl|j|d<|j|j	|S)N	__slots__r__weakref__r)
rcopyrrIrrrMrr
	__bases__)rZ	orig_varsslotsZ	slots_varrrrrs




zadd_metaclass..wrapperr)rrrrr
add_metaclasssrutf-8rcCs8t|tr|||St|tr$|Stdt|dS)zCoerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    not expecting type '%s'N)rI	text_typerbinary_typerrrrrrrr
ensure_binarys


r	cCsXt|ttfstdt|tr:t|tr:|||}ntrTt|trT|||}|S)zCoerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    r)	rIrrrrPY2rr)decoderrrr
ensure_strsrcCs8t|tr|||St|tr$|Stdt|dS)zCoerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    rN)rIrrrrrrrrrensure_texts


r
cCs2tr.d|jkrtd|j|j|_dd|_|S)a
    A decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    __str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|dS)Nr)__unicode__rr	rrrz-python_2_unicode_compatible..)r
r
ValueErrorr
rr)rrrrpython_2_unicode_compatibles

r__spec__)N)NN)rr)rr)rr)rZ
__future__rrdrUoperatorrr
__author____version__version_infor
r)ZPY34rZstring_typesintZ
integer_typesrZclass_typesrbytesrmaxsizeZMAXSIZErZlongZ	ClassTyperplatform
startswithobjectrlen
OverflowErrorrrrr'
ModuleTyper4r9r;r
rrQr7r/rrrIrArtruZ_urllib_parse_moved_attributesrxZ_urllib_error_moved_attributesrzZ _urllib_request_moved_attributesr|Z!_urllib_response_moved_attributesr~Z$_urllib_robotparser_moved_attributesrrrZ
_meth_funcZ
_meth_selfZ
_func_closureZ
_func_codeZ_func_defaultsZ
_func_globalsrr	NameErrorrrrrrr
attrgetterZget_method_functionZget_method_selfZget_function_closureZget_function_codeZget_function_defaultsZget_function_globalsrrrrmethodcallerrrrrrchrZunichrstructStructpackZint2byte
itemgetterrgetitemrrZ	iterbytesrRrSBytesIOrrrpartialr[rrrrr.rVrrrrrWRAPPER_ASSIGNMENTSWRAPPER_UPDATESrrrr	rr
rrL__package__globalsrrsubmodule_search_locations	meta_pathrrZimporterappendrrrrsN

@







































K





















































%












:
site-packages/pip/_vendor/urllib3/packages/six.py000064400000077430151464452710016043 0ustar00# Copyright (c) 2010-2019 Benjamin Peterson
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.

"""Utilities for writing code that runs on Python 2 and 3"""

from __future__ import absolute_import

import functools
import itertools
import operator
import sys
import types

__author__ = "Benjamin Peterson "
__version__ = "1.12.0"


# Useful for very coarse version differentiation.
PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3
PY34 = sys.version_info[0:2] >= (3, 4)

if PY3:
    string_types = (str,)
    integer_types = (int,)
    class_types = (type,)
    text_type = str
    binary_type = bytes

    MAXSIZE = sys.maxsize
else:
    string_types = (basestring,)
    integer_types = (int, long)
    class_types = (type, types.ClassType)
    text_type = unicode
    binary_type = str

    if sys.platform.startswith("java"):
        # Jython always uses 32 bits.
        MAXSIZE = int((1 << 31) - 1)
    else:
        # It's possible to have sizeof(long) != sizeof(Py_ssize_t).
        class X(object):
            def __len__(self):
                return 1 << 31

        try:
            len(X())
        except OverflowError:
            # 32-bit
            MAXSIZE = int((1 << 31) - 1)
        else:
            # 64-bit
            MAXSIZE = int((1 << 63) - 1)
        del X


def _add_doc(func, doc):
    """Add documentation to a function."""
    func.__doc__ = doc


def _import_module(name):
    """Import module, returning the module after the last dot."""
    __import__(name)
    return sys.modules[name]


class _LazyDescr(object):
    def __init__(self, name):
        self.name = name

    def __get__(self, obj, tp):
        result = self._resolve()
        setattr(obj, self.name, result)  # Invokes __set__.
        try:
            # This is a bit ugly, but it avoids running this again by
            # removing this descriptor.
            delattr(obj.__class__, self.name)
        except AttributeError:
            pass
        return result


class MovedModule(_LazyDescr):
    def __init__(self, name, old, new=None):
        super(MovedModule, self).__init__(name)
        if PY3:
            if new is None:
                new = name
            self.mod = new
        else:
            self.mod = old

    def _resolve(self):
        return _import_module(self.mod)

    def __getattr__(self, attr):
        _module = self._resolve()
        value = getattr(_module, attr)
        setattr(self, attr, value)
        return value


class _LazyModule(types.ModuleType):
    def __init__(self, name):
        super(_LazyModule, self).__init__(name)
        self.__doc__ = self.__class__.__doc__

    def __dir__(self):
        attrs = ["__doc__", "__name__"]
        attrs += [attr.name for attr in self._moved_attributes]
        return attrs

    # Subclasses should override this
    _moved_attributes = []


class MovedAttribute(_LazyDescr):
    def __init__(self, name, old_mod, new_mod, old_attr=None, new_attr=None):
        super(MovedAttribute, self).__init__(name)
        if PY3:
            if new_mod is None:
                new_mod = name
            self.mod = new_mod
            if new_attr is None:
                if old_attr is None:
                    new_attr = name
                else:
                    new_attr = old_attr
            self.attr = new_attr
        else:
            self.mod = old_mod
            if old_attr is None:
                old_attr = name
            self.attr = old_attr

    def _resolve(self):
        module = _import_module(self.mod)
        return getattr(module, self.attr)


class _SixMetaPathImporter(object):

    """
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    """

    def __init__(self, six_module_name):
        self.name = six_module_name
        self.known_modules = {}

    def _add_module(self, mod, *fullnames):
        for fullname in fullnames:
            self.known_modules[self.name + "." + fullname] = mod

    def _get_module(self, fullname):
        return self.known_modules[self.name + "." + fullname]

    def find_module(self, fullname, path=None):
        if fullname in self.known_modules:
            return self
        return None

    def __get_module(self, fullname):
        try:
            return self.known_modules[fullname]
        except KeyError:
            raise ImportError("This loader does not know module " + fullname)

    def load_module(self, fullname):
        try:
            # in case of a reload
            return sys.modules[fullname]
        except KeyError:
            pass
        mod = self.__get_module(fullname)
        if isinstance(mod, MovedModule):
            mod = mod._resolve()
        else:
            mod.__loader__ = self
        sys.modules[fullname] = mod
        return mod

    def is_package(self, fullname):
        """
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        """
        return hasattr(self.__get_module(fullname), "__path__")

    def get_code(self, fullname):
        """Return None

        Required, if is_package is implemented"""
        self.__get_module(fullname)  # eventually raises ImportError
        return None

    get_source = get_code  # same as get_code


_importer = _SixMetaPathImporter(__name__)


class _MovedItems(_LazyModule):

    """Lazy loading of moved objects"""

    __path__ = []  # mark as package


_moved_attributes = [
    MovedAttribute("cStringIO", "cStringIO", "io", "StringIO"),
    MovedAttribute("filter", "itertools", "builtins", "ifilter", "filter"),
    MovedAttribute(
        "filterfalse", "itertools", "itertools", "ifilterfalse", "filterfalse"
    ),
    MovedAttribute("input", "__builtin__", "builtins", "raw_input", "input"),
    MovedAttribute("intern", "__builtin__", "sys"),
    MovedAttribute("map", "itertools", "builtins", "imap", "map"),
    MovedAttribute("getcwd", "os", "os", "getcwdu", "getcwd"),
    MovedAttribute("getcwdb", "os", "os", "getcwd", "getcwdb"),
    MovedAttribute("getoutput", "commands", "subprocess"),
    MovedAttribute("range", "__builtin__", "builtins", "xrange", "range"),
    MovedAttribute(
        "reload_module", "__builtin__", "importlib" if PY34 else "imp", "reload"
    ),
    MovedAttribute("reduce", "__builtin__", "functools"),
    MovedAttribute("shlex_quote", "pipes", "shlex", "quote"),
    MovedAttribute("StringIO", "StringIO", "io"),
    MovedAttribute("UserDict", "UserDict", "collections"),
    MovedAttribute("UserList", "UserList", "collections"),
    MovedAttribute("UserString", "UserString", "collections"),
    MovedAttribute("xrange", "__builtin__", "builtins", "xrange", "range"),
    MovedAttribute("zip", "itertools", "builtins", "izip", "zip"),
    MovedAttribute(
        "zip_longest", "itertools", "itertools", "izip_longest", "zip_longest"
    ),
    MovedModule("builtins", "__builtin__"),
    MovedModule("configparser", "ConfigParser"),
    MovedModule("copyreg", "copy_reg"),
    MovedModule("dbm_gnu", "gdbm", "dbm.gnu"),
    MovedModule("_dummy_thread", "dummy_thread", "_dummy_thread"),
    MovedModule("http_cookiejar", "cookielib", "http.cookiejar"),
    MovedModule("http_cookies", "Cookie", "http.cookies"),
    MovedModule("html_entities", "htmlentitydefs", "html.entities"),
    MovedModule("html_parser", "HTMLParser", "html.parser"),
    MovedModule("http_client", "httplib", "http.client"),
    MovedModule("email_mime_base", "email.MIMEBase", "email.mime.base"),
    MovedModule("email_mime_image", "email.MIMEImage", "email.mime.image"),
    MovedModule("email_mime_multipart", "email.MIMEMultipart", "email.mime.multipart"),
    MovedModule(
        "email_mime_nonmultipart", "email.MIMENonMultipart", "email.mime.nonmultipart"
    ),
    MovedModule("email_mime_text", "email.MIMEText", "email.mime.text"),
    MovedModule("BaseHTTPServer", "BaseHTTPServer", "http.server"),
    MovedModule("CGIHTTPServer", "CGIHTTPServer", "http.server"),
    MovedModule("SimpleHTTPServer", "SimpleHTTPServer", "http.server"),
    MovedModule("cPickle", "cPickle", "pickle"),
    MovedModule("queue", "Queue"),
    MovedModule("reprlib", "repr"),
    MovedModule("socketserver", "SocketServer"),
    MovedModule("_thread", "thread", "_thread"),
    MovedModule("tkinter", "Tkinter"),
    MovedModule("tkinter_dialog", "Dialog", "tkinter.dialog"),
    MovedModule("tkinter_filedialog", "FileDialog", "tkinter.filedialog"),
    MovedModule("tkinter_scrolledtext", "ScrolledText", "tkinter.scrolledtext"),
    MovedModule("tkinter_simpledialog", "SimpleDialog", "tkinter.simpledialog"),
    MovedModule("tkinter_tix", "Tix", "tkinter.tix"),
    MovedModule("tkinter_ttk", "ttk", "tkinter.ttk"),
    MovedModule("tkinter_constants", "Tkconstants", "tkinter.constants"),
    MovedModule("tkinter_dnd", "Tkdnd", "tkinter.dnd"),
    MovedModule("tkinter_colorchooser", "tkColorChooser", "tkinter.colorchooser"),
    MovedModule("tkinter_commondialog", "tkCommonDialog", "tkinter.commondialog"),
    MovedModule("tkinter_tkfiledialog", "tkFileDialog", "tkinter.filedialog"),
    MovedModule("tkinter_font", "tkFont", "tkinter.font"),
    MovedModule("tkinter_messagebox", "tkMessageBox", "tkinter.messagebox"),
    MovedModule("tkinter_tksimpledialog", "tkSimpleDialog", "tkinter.simpledialog"),
    MovedModule("urllib_parse", __name__ + ".moves.urllib_parse", "urllib.parse"),
    MovedModule("urllib_error", __name__ + ".moves.urllib_error", "urllib.error"),
    MovedModule("urllib", __name__ + ".moves.urllib", __name__ + ".moves.urllib"),
    MovedModule("urllib_robotparser", "robotparser", "urllib.robotparser"),
    MovedModule("xmlrpc_client", "xmlrpclib", "xmlrpc.client"),
    MovedModule("xmlrpc_server", "SimpleXMLRPCServer", "xmlrpc.server"),
]
# Add windows specific modules.
if sys.platform == "win32":
    _moved_attributes += [MovedModule("winreg", "_winreg")]

for attr in _moved_attributes:
    setattr(_MovedItems, attr.name, attr)
    if isinstance(attr, MovedModule):
        _importer._add_module(attr, "moves." + attr.name)
del attr

_MovedItems._moved_attributes = _moved_attributes

moves = _MovedItems(__name__ + ".moves")
_importer._add_module(moves, "moves")


class Module_six_moves_urllib_parse(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_parse"""


_urllib_parse_moved_attributes = [
    MovedAttribute("ParseResult", "urlparse", "urllib.parse"),
    MovedAttribute("SplitResult", "urlparse", "urllib.parse"),
    MovedAttribute("parse_qs", "urlparse", "urllib.parse"),
    MovedAttribute("parse_qsl", "urlparse", "urllib.parse"),
    MovedAttribute("urldefrag", "urlparse", "urllib.parse"),
    MovedAttribute("urljoin", "urlparse", "urllib.parse"),
    MovedAttribute("urlparse", "urlparse", "urllib.parse"),
    MovedAttribute("urlsplit", "urlparse", "urllib.parse"),
    MovedAttribute("urlunparse", "urlparse", "urllib.parse"),
    MovedAttribute("urlunsplit", "urlparse", "urllib.parse"),
    MovedAttribute("quote", "urllib", "urllib.parse"),
    MovedAttribute("quote_plus", "urllib", "urllib.parse"),
    MovedAttribute("unquote", "urllib", "urllib.parse"),
    MovedAttribute("unquote_plus", "urllib", "urllib.parse"),
    MovedAttribute(
        "unquote_to_bytes", "urllib", "urllib.parse", "unquote", "unquote_to_bytes"
    ),
    MovedAttribute("urlencode", "urllib", "urllib.parse"),
    MovedAttribute("splitquery", "urllib", "urllib.parse"),
    MovedAttribute("splittag", "urllib", "urllib.parse"),
    MovedAttribute("splituser", "urllib", "urllib.parse"),
    MovedAttribute("splitvalue", "urllib", "urllib.parse"),
    MovedAttribute("uses_fragment", "urlparse", "urllib.parse"),
    MovedAttribute("uses_netloc", "urlparse", "urllib.parse"),
    MovedAttribute("uses_params", "urlparse", "urllib.parse"),
    MovedAttribute("uses_query", "urlparse", "urllib.parse"),
    MovedAttribute("uses_relative", "urlparse", "urllib.parse"),
]
for attr in _urllib_parse_moved_attributes:
    setattr(Module_six_moves_urllib_parse, attr.name, attr)
del attr

Module_six_moves_urllib_parse._moved_attributes = _urllib_parse_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_parse(__name__ + ".moves.urllib_parse"),
    "moves.urllib_parse",
    "moves.urllib.parse",
)


class Module_six_moves_urllib_error(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_error"""


_urllib_error_moved_attributes = [
    MovedAttribute("URLError", "urllib2", "urllib.error"),
    MovedAttribute("HTTPError", "urllib2", "urllib.error"),
    MovedAttribute("ContentTooShortError", "urllib", "urllib.error"),
]
for attr in _urllib_error_moved_attributes:
    setattr(Module_six_moves_urllib_error, attr.name, attr)
del attr

Module_six_moves_urllib_error._moved_attributes = _urllib_error_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_error(__name__ + ".moves.urllib.error"),
    "moves.urllib_error",
    "moves.urllib.error",
)


class Module_six_moves_urllib_request(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_request"""


_urllib_request_moved_attributes = [
    MovedAttribute("urlopen", "urllib2", "urllib.request"),
    MovedAttribute("install_opener", "urllib2", "urllib.request"),
    MovedAttribute("build_opener", "urllib2", "urllib.request"),
    MovedAttribute("pathname2url", "urllib", "urllib.request"),
    MovedAttribute("url2pathname", "urllib", "urllib.request"),
    MovedAttribute("getproxies", "urllib", "urllib.request"),
    MovedAttribute("Request", "urllib2", "urllib.request"),
    MovedAttribute("OpenerDirector", "urllib2", "urllib.request"),
    MovedAttribute("HTTPDefaultErrorHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPRedirectHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPCookieProcessor", "urllib2", "urllib.request"),
    MovedAttribute("ProxyHandler", "urllib2", "urllib.request"),
    MovedAttribute("BaseHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPPasswordMgr", "urllib2", "urllib.request"),
    MovedAttribute("HTTPPasswordMgrWithDefaultRealm", "urllib2", "urllib.request"),
    MovedAttribute("AbstractBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("ProxyBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("AbstractDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("ProxyDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPSHandler", "urllib2", "urllib.request"),
    MovedAttribute("FileHandler", "urllib2", "urllib.request"),
    MovedAttribute("FTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("CacheFTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("UnknownHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPErrorProcessor", "urllib2", "urllib.request"),
    MovedAttribute("urlretrieve", "urllib", "urllib.request"),
    MovedAttribute("urlcleanup", "urllib", "urllib.request"),
    MovedAttribute("URLopener", "urllib", "urllib.request"),
    MovedAttribute("FancyURLopener", "urllib", "urllib.request"),
    MovedAttribute("proxy_bypass", "urllib", "urllib.request"),
    MovedAttribute("parse_http_list", "urllib2", "urllib.request"),
    MovedAttribute("parse_keqv_list", "urllib2", "urllib.request"),
]
for attr in _urllib_request_moved_attributes:
    setattr(Module_six_moves_urllib_request, attr.name, attr)
del attr

Module_six_moves_urllib_request._moved_attributes = _urllib_request_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_request(__name__ + ".moves.urllib.request"),
    "moves.urllib_request",
    "moves.urllib.request",
)


class Module_six_moves_urllib_response(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_response"""


_urllib_response_moved_attributes = [
    MovedAttribute("addbase", "urllib", "urllib.response"),
    MovedAttribute("addclosehook", "urllib", "urllib.response"),
    MovedAttribute("addinfo", "urllib", "urllib.response"),
    MovedAttribute("addinfourl", "urllib", "urllib.response"),
]
for attr in _urllib_response_moved_attributes:
    setattr(Module_six_moves_urllib_response, attr.name, attr)
del attr

Module_six_moves_urllib_response._moved_attributes = _urllib_response_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_response(__name__ + ".moves.urllib.response"),
    "moves.urllib_response",
    "moves.urllib.response",
)


class Module_six_moves_urllib_robotparser(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_robotparser"""


_urllib_robotparser_moved_attributes = [
    MovedAttribute("RobotFileParser", "robotparser", "urllib.robotparser")
]
for attr in _urllib_robotparser_moved_attributes:
    setattr(Module_six_moves_urllib_robotparser, attr.name, attr)
del attr

Module_six_moves_urllib_robotparser._moved_attributes = (
    _urllib_robotparser_moved_attributes
)

_importer._add_module(
    Module_six_moves_urllib_robotparser(__name__ + ".moves.urllib.robotparser"),
    "moves.urllib_robotparser",
    "moves.urllib.robotparser",
)


class Module_six_moves_urllib(types.ModuleType):

    """Create a six.moves.urllib namespace that resembles the Python 3 namespace"""

    __path__ = []  # mark as package
    parse = _importer._get_module("moves.urllib_parse")
    error = _importer._get_module("moves.urllib_error")
    request = _importer._get_module("moves.urllib_request")
    response = _importer._get_module("moves.urllib_response")
    robotparser = _importer._get_module("moves.urllib_robotparser")

    def __dir__(self):
        return ["parse", "error", "request", "response", "robotparser"]


_importer._add_module(
    Module_six_moves_urllib(__name__ + ".moves.urllib"), "moves.urllib"
)


def add_move(move):
    """Add an item to six.moves."""
    setattr(_MovedItems, move.name, move)


def remove_move(name):
    """Remove item from six.moves."""
    try:
        delattr(_MovedItems, name)
    except AttributeError:
        try:
            del moves.__dict__[name]
        except KeyError:
            raise AttributeError("no such move, %r" % (name,))


if PY3:
    _meth_func = "__func__"
    _meth_self = "__self__"

    _func_closure = "__closure__"
    _func_code = "__code__"
    _func_defaults = "__defaults__"
    _func_globals = "__globals__"
else:
    _meth_func = "im_func"
    _meth_self = "im_self"

    _func_closure = "func_closure"
    _func_code = "func_code"
    _func_defaults = "func_defaults"
    _func_globals = "func_globals"


try:
    advance_iterator = next
except NameError:

    def advance_iterator(it):
        return it.next()


next = advance_iterator


try:
    callable = callable
except NameError:

    def callable(obj):
        return any("__call__" in klass.__dict__ for klass in type(obj).__mro__)


if PY3:

    def get_unbound_function(unbound):
        return unbound

    create_bound_method = types.MethodType

    def create_unbound_method(func, cls):
        return func

    Iterator = object
else:

    def get_unbound_function(unbound):
        return unbound.im_func

    def create_bound_method(func, obj):
        return types.MethodType(func, obj, obj.__class__)

    def create_unbound_method(func, cls):
        return types.MethodType(func, None, cls)

    class Iterator(object):
        def next(self):
            return type(self).__next__(self)

    callable = callable
_add_doc(
    get_unbound_function, """Get the function out of a possibly unbound function"""
)


get_method_function = operator.attrgetter(_meth_func)
get_method_self = operator.attrgetter(_meth_self)
get_function_closure = operator.attrgetter(_func_closure)
get_function_code = operator.attrgetter(_func_code)
get_function_defaults = operator.attrgetter(_func_defaults)
get_function_globals = operator.attrgetter(_func_globals)


if PY3:

    def iterkeys(d, **kw):
        return iter(d.keys(**kw))

    def itervalues(d, **kw):
        return iter(d.values(**kw))

    def iteritems(d, **kw):
        return iter(d.items(**kw))

    def iterlists(d, **kw):
        return iter(d.lists(**kw))

    viewkeys = operator.methodcaller("keys")

    viewvalues = operator.methodcaller("values")

    viewitems = operator.methodcaller("items")
else:

    def iterkeys(d, **kw):
        return d.iterkeys(**kw)

    def itervalues(d, **kw):
        return d.itervalues(**kw)

    def iteritems(d, **kw):
        return d.iteritems(**kw)

    def iterlists(d, **kw):
        return d.iterlists(**kw)

    viewkeys = operator.methodcaller("viewkeys")

    viewvalues = operator.methodcaller("viewvalues")

    viewitems = operator.methodcaller("viewitems")

_add_doc(iterkeys, "Return an iterator over the keys of a dictionary.")
_add_doc(itervalues, "Return an iterator over the values of a dictionary.")
_add_doc(iteritems, "Return an iterator over the (key, value) pairs of a dictionary.")
_add_doc(
    iterlists, "Return an iterator over the (key, [values]) pairs of a dictionary."
)


if PY3:

    def b(s):
        return s.encode("latin-1")

    def u(s):
        return s

    unichr = chr
    import struct

    int2byte = struct.Struct(">B").pack
    del struct
    byte2int = operator.itemgetter(0)
    indexbytes = operator.getitem
    iterbytes = iter
    import io

    StringIO = io.StringIO
    BytesIO = io.BytesIO
    del io
    _assertCountEqual = "assertCountEqual"
    if sys.version_info[1] <= 1:
        _assertRaisesRegex = "assertRaisesRegexp"
        _assertRegex = "assertRegexpMatches"
    else:
        _assertRaisesRegex = "assertRaisesRegex"
        _assertRegex = "assertRegex"
else:

    def b(s):
        return s

    # Workaround for standalone backslash

    def u(s):
        return unicode(s.replace(r"\\", r"\\\\"), "unicode_escape")

    unichr = unichr
    int2byte = chr

    def byte2int(bs):
        return ord(bs[0])

    def indexbytes(buf, i):
        return ord(buf[i])

    iterbytes = functools.partial(itertools.imap, ord)
    import StringIO

    StringIO = BytesIO = StringIO.StringIO
    _assertCountEqual = "assertItemsEqual"
    _assertRaisesRegex = "assertRaisesRegexp"
    _assertRegex = "assertRegexpMatches"
_add_doc(b, """Byte literal""")
_add_doc(u, """Text literal""")


def assertCountEqual(self, *args, **kwargs):
    return getattr(self, _assertCountEqual)(*args, **kwargs)


def assertRaisesRegex(self, *args, **kwargs):
    return getattr(self, _assertRaisesRegex)(*args, **kwargs)


def assertRegex(self, *args, **kwargs):
    return getattr(self, _assertRegex)(*args, **kwargs)


if PY3:
    exec_ = getattr(moves.builtins, "exec")

    def reraise(tp, value, tb=None):
        try:
            if value is None:
                value = tp()
            if value.__traceback__ is not tb:
                raise value.with_traceback(tb)
            raise value
        finally:
            value = None
            tb = None


else:

    def exec_(_code_, _globs_=None, _locs_=None):
        """Execute code in a namespace."""
        if _globs_ is None:
            frame = sys._getframe(1)
            _globs_ = frame.f_globals
            if _locs_ is None:
                _locs_ = frame.f_locals
            del frame
        elif _locs_ is None:
            _locs_ = _globs_
        exec("""exec _code_ in _globs_, _locs_""")

    exec_(
        """def reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
"""
    )


if sys.version_info[:2] == (3, 2):
    exec_(
        """def raise_from(value, from_value):
    try:
        if from_value is None:
            raise value
        raise value from from_value
    finally:
        value = None
"""
    )
elif sys.version_info[:2] > (3, 2):
    exec_(
        """def raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
"""
    )
else:

    def raise_from(value, from_value):
        raise value


print_ = getattr(moves.builtins, "print", None)
if print_ is None:

    def print_(*args, **kwargs):
        """The new-style print function for Python 2.4 and 2.5."""
        fp = kwargs.pop("file", sys.stdout)
        if fp is None:
            return

        def write(data):
            if not isinstance(data, basestring):
                data = str(data)
            # If the file has an encoding, encode unicode with it.
            if (
                isinstance(fp, file)
                and isinstance(data, unicode)
                and fp.encoding is not None
            ):
                errors = getattr(fp, "errors", None)
                if errors is None:
                    errors = "strict"
                data = data.encode(fp.encoding, errors)
            fp.write(data)

        want_unicode = False
        sep = kwargs.pop("sep", None)
        if sep is not None:
            if isinstance(sep, unicode):
                want_unicode = True
            elif not isinstance(sep, str):
                raise TypeError("sep must be None or a string")
        end = kwargs.pop("end", None)
        if end is not None:
            if isinstance(end, unicode):
                want_unicode = True
            elif not isinstance(end, str):
                raise TypeError("end must be None or a string")
        if kwargs:
            raise TypeError("invalid keyword arguments to print()")
        if not want_unicode:
            for arg in args:
                if isinstance(arg, unicode):
                    want_unicode = True
                    break
        if want_unicode:
            newline = unicode("\n")
            space = unicode(" ")
        else:
            newline = "\n"
            space = " "
        if sep is None:
            sep = space
        if end is None:
            end = newline
        for i, arg in enumerate(args):
            if i:
                write(sep)
            write(arg)
        write(end)


if sys.version_info[:2] < (3, 3):
    _print = print_

    def print_(*args, **kwargs):
        fp = kwargs.get("file", sys.stdout)
        flush = kwargs.pop("flush", False)
        _print(*args, **kwargs)
        if flush and fp is not None:
            fp.flush()


_add_doc(reraise, """Reraise an exception.""")

if sys.version_info[0:2] < (3, 4):

    def wraps(
        wrapped,
        assigned=functools.WRAPPER_ASSIGNMENTS,
        updated=functools.WRAPPER_UPDATES,
    ):
        def wrapper(f):
            f = functools.wraps(wrapped, assigned, updated)(f)
            f.__wrapped__ = wrapped
            return f

        return wrapper


else:
    wraps = functools.wraps


def with_metaclass(meta, *bases):
    """Create a base class with a metaclass."""
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(type):
        def __new__(cls, name, this_bases, d):
            return meta(name, bases, d)

        @classmethod
        def __prepare__(cls, name, this_bases):
            return meta.__prepare__(name, bases)

    return type.__new__(metaclass, "temporary_class", (), {})


def add_metaclass(metaclass):
    """Class decorator for creating a class with a metaclass."""

    def wrapper(cls):
        orig_vars = cls.__dict__.copy()
        slots = orig_vars.get("__slots__")
        if slots is not None:
            if isinstance(slots, str):
                slots = [slots]
            for slots_var in slots:
                orig_vars.pop(slots_var)
        orig_vars.pop("__dict__", None)
        orig_vars.pop("__weakref__", None)
        if hasattr(cls, "__qualname__"):
            orig_vars["__qualname__"] = cls.__qualname__
        return metaclass(cls.__name__, cls.__bases__, orig_vars)

    return wrapper


def ensure_binary(s, encoding="utf-8", errors="strict"):
    """Coerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    """
    if isinstance(s, text_type):
        return s.encode(encoding, errors)
    elif isinstance(s, binary_type):
        return s
    else:
        raise TypeError("not expecting type '%s'" % type(s))


def ensure_str(s, encoding="utf-8", errors="strict"):
    """Coerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    """
    if not isinstance(s, (text_type, binary_type)):
        raise TypeError("not expecting type '%s'" % type(s))
    if PY2 and isinstance(s, text_type):
        s = s.encode(encoding, errors)
    elif PY3 and isinstance(s, binary_type):
        s = s.decode(encoding, errors)
    return s


def ensure_text(s, encoding="utf-8", errors="strict"):
    """Coerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    """
    if isinstance(s, binary_type):
        return s.decode(encoding, errors)
    elif isinstance(s, text_type):
        return s
    else:
        raise TypeError("not expecting type '%s'" % type(s))


def python_2_unicode_compatible(klass):
    """
    A decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    """
    if PY2:
        if "__str__" not in klass.__dict__:
            raise ValueError(
                "@python_2_unicode_compatible cannot be applied "
                "to %s because it doesn't define __str__()." % klass.__name__
            )
        klass.__unicode__ = klass.__str__
        klass.__str__ = lambda self: self.__unicode__().encode("utf-8")
    return klass


# Complete the moves implementation.
# This code is at the end of this module to speed up module loading.
# Turn this module into a package.
__path__ = []  # required for PEP 302 and PEP 451
__package__ = __name__  # see PEP 366 @ReservedAssignment
if globals().get("__spec__") is not None:
    __spec__.submodule_search_locations = []  # PEP 451 @UndefinedVariable
# Remove other six meta path importers, since they cause problems. This can
# happen if six is removed from sys.modules and then reloaded. (Setuptools does
# this for some reason.)
if sys.meta_path:
    for i, importer in enumerate(sys.meta_path):
        # Here's some real nastiness: Another "instance" of the six module might
        # be floating around. Therefore, we can't use isinstance() to check for
        # the six meta path importer, since the other six instance will have
        # inserted an importer with different class.
        if (
            type(importer).__name__ == "_SixMetaPathImporter"
            and importer.name == __name__
        ):
            del sys.meta_path[i]
            break
    del i, importer
# Finally, add the importer to the meta path import hook.
sys.meta_path.append(_importer)
site-packages/pip/_vendor/urllib3/packages/backports/__pycache__/makefile.cpython-38.pyc000064400000002376151464452710025270 0ustar00U

.e@s&dZddlZddlmZdddZdS)z
backports.makefile
~~~~~~~~~~~~~~~~~~

Backports the Python 3 ``socket.makefile`` method for use with anything that
wants to create a "fake" socket object.
N)SocketIOrc
Cst|dddhks td|fd|k}d|kp4|}|sB|sBtd|k}d}	|rZ|	d7}	|rf|	d7}	t||	}
|jd7_|dkrd}|d	krtj}|d	kr|std
|
S|r|rt|
|
|}n&|rt|
|}n|stt	|
|}|r|St
||||}||_|S)z:
    Backport of ``socket.makefile`` from Python 3.5.
    rwbz&invalid mode %r (only r, w, b allowed)Nrz!unbuffered streams must be binary)set
ValueErrorAssertionErrorrZ_makefile_refsioDEFAULT_BUFFER_SIZEBufferedRWPairBufferedReaderBufferedWriter
TextIOWrappermode)
selfr	bufferingencodingerrorsnewlineZwritingZreadingZbinaryZrawmoderawbuffertextrS/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/backports/makefile.pybackport_makefiles@
r)rNNNN)__doc__rZsocketrrrrrrssite-packages/pip/_vendor/urllib3/packages/backports/__pycache__/__init__.cpython-38.pyc000064400000000256151464452710025245 0ustar00U

.e@sdS)NrrrS/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/backports/__init__.pysite-packages/pip/_vendor/urllib3/packages/backports/__pycache__/makefile.cpython-38.opt-1.pyc000064400000002326151464452710026222 0ustar00U

.e@s&dZddlZddlmZdddZdS)z
backports.makefile
~~~~~~~~~~~~~~~~~~

Backports the Python 3 ``socket.makefile`` method for use with anything that
wants to create a "fake" socket object.
N)SocketIOrc
Cst|dddhks td|fd|k}d|kp4|}d|k}d}	|rN|	d7}	|rZ|	d7}	t||	}
|jd7_|dkr~d}|d	krtj}|d	kr|std
|
S|r|rt|
|
|}n|rt|
|}nt|
|}|r|St	||||}||_
|S)z:
    Backport of ``socket.makefile`` from Python 3.5.
    rwbz&invalid mode %r (only r, w, b allowed)Nrz!unbuffered streams must be binary)set
ValueErrorrZ_makefile_refsioDEFAULT_BUFFER_SIZEBufferedRWPairBufferedReaderBufferedWriter
TextIOWrappermode)
selfr	bufferingencodingerrorsnewlineZwritingZreadingZbinaryZrawmoderawbuffertextrS/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/backports/makefile.pybackport_makefiles<
r)rNNNN)__doc__rZsocketrrrrrrssite-packages/pip/_vendor/urllib3/packages/backports/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000256151464452710026204 0ustar00U

.e@sdS)NrrrS/usr/lib/python3.8/site-packages/pip/_vendor/urllib3/packages/backports/__init__.pysite-packages/pip/_vendor/urllib3/packages/backports/makefile.py000064400000002612151464452710020773 0ustar00# -*- coding: utf-8 -*-
"""
backports.makefile
~~~~~~~~~~~~~~~~~~

Backports the Python 3 ``socket.makefile`` method for use with anything that
wants to create a "fake" socket object.
"""
import io

from socket import SocketIO


def backport_makefile(
    self, mode="r", buffering=None, encoding=None, errors=None, newline=None
):
    """
    Backport of ``socket.makefile`` from Python 3.5.
    """
    if not set(mode) <= {"r", "w", "b"}:
        raise ValueError("invalid mode %r (only r, w, b allowed)" % (mode,))
    writing = "w" in mode
    reading = "r" in mode or not writing
    assert reading or writing
    binary = "b" in mode
    rawmode = ""
    if reading:
        rawmode += "r"
    if writing:
        rawmode += "w"
    raw = SocketIO(self, rawmode)
    self._makefile_refs += 1
    if buffering is None:
        buffering = -1
    if buffering < 0:
        buffering = io.DEFAULT_BUFFER_SIZE
    if buffering == 0:
        if not binary:
            raise ValueError("unbuffered streams must be binary")
        return raw
    if reading and writing:
        buffer = io.BufferedRWPair(raw, raw, buffering)
    elif reading:
        buffer = io.BufferedReader(raw, buffering)
    else:
        assert writing
        buffer = io.BufferedWriter(raw, buffering)
    if binary:
        return buffer
    text = io.TextIOWrapper(buffer, encoding, errors, newline)
    text.mode = mode
    return text
site-packages/pip/_vendor/urllib3/packages/backports/__init__.py000064400000000000151464452710020742 0ustar00site-packages/pip/_vendor/urllib3/packages/__init__.py000064400000000154151464452710016764 0ustar00from __future__ import absolute_import

from . import ssl_match_hostname

__all__ = ("ssl_match_hostname",)
site-packages/pip/_vendor/urllib3/__init__.py000064400000005173151464452710015214 0ustar00"""
urllib3 - Thread-safe connection pooling and re-using.
"""
from __future__ import absolute_import
import warnings

from .connectionpool import HTTPConnectionPool, HTTPSConnectionPool, connection_from_url

from . import exceptions
from .filepost import encode_multipart_formdata
from .poolmanager import PoolManager, ProxyManager, proxy_from_url
from .response import HTTPResponse
from .util.request import make_headers
from .util.url import get_host
from .util.timeout import Timeout
from .util.retry import Retry


# Set default logging handler to avoid "No handler found" warnings.
import logging
from logging import NullHandler

__author__ = "Andrey Petrov (andrey.petrov@shazow.net)"
__license__ = "MIT"
__version__ = "1.25.6"

__all__ = (
    "HTTPConnectionPool",
    "HTTPSConnectionPool",
    "PoolManager",
    "ProxyManager",
    "HTTPResponse",
    "Retry",
    "Timeout",
    "add_stderr_logger",
    "connection_from_url",
    "disable_warnings",
    "encode_multipart_formdata",
    "get_host",
    "make_headers",
    "proxy_from_url",
)

logging.getLogger(__name__).addHandler(NullHandler())


def add_stderr_logger(level=logging.DEBUG):
    """
    Helper for quickly adding a StreamHandler to the logger. Useful for
    debugging.

    Returns the handler after adding it.
    """
    # This method needs to be in this __init__.py to get the __name__ correct
    # even if urllib3 is vendored within another package.
    logger = logging.getLogger(__name__)
    handler = logging.StreamHandler()
    handler.setFormatter(logging.Formatter("%(asctime)s %(levelname)s %(message)s"))
    logger.addHandler(handler)
    logger.setLevel(level)
    logger.debug("Added a stderr logging handler to logger: %s", __name__)
    return handler


# ... Clean up.
del NullHandler


# All warning filters *must* be appended unless you're really certain that they
# shouldn't be: otherwise, it's very hard for users to use most Python
# mechanisms to silence them.
# SecurityWarning's always go off by default.
warnings.simplefilter("always", exceptions.SecurityWarning, append=True)
# SubjectAltNameWarning's should go off once per host
warnings.simplefilter("default", exceptions.SubjectAltNameWarning, append=True)
# InsecurePlatformWarning's don't vary between requests, so we keep it default.
warnings.simplefilter("default", exceptions.InsecurePlatformWarning, append=True)
# SNIMissingWarnings should go off only once.
warnings.simplefilter("default", exceptions.SNIMissingWarning, append=True)


def disable_warnings(category=exceptions.HTTPWarning):
    """
    Helper for quickly disabling all urllib3 warnings.
    """
    warnings.simplefilter("ignore", category)
site-packages/pip/_vendor/cachecontrol/cache.py000064400000001445151464452710015606 0ustar00"""
The cache object API for implementing caches. The default is a thread
safe in-memory dictionary.
"""
from threading import Lock


class BaseCache(object):

    def get(self, key):
        raise NotImplementedError()

    def set(self, key, value):
        raise NotImplementedError()

    def delete(self, key):
        raise NotImplementedError()

    def close(self):
        pass


class DictCache(BaseCache):

    def __init__(self, init_dict=None):
        self.lock = Lock()
        self.data = init_dict or {}

    def get(self, key):
        return self.data.get(key, None)

    def set(self, key, value):
        with self.lock:
            self.data.update({key: value})

    def delete(self, key):
        with self.lock:
            if key in self.data:
                self.data.pop(key)
site-packages/pip/_vendor/cachecontrol/_cmd.py000064400000002417151464452710015445 0ustar00import logging

from pip._vendor import requests

from pip._vendor.cachecontrol.adapter import CacheControlAdapter
from pip._vendor.cachecontrol.cache import DictCache
from pip._vendor.cachecontrol.controller import logger

from argparse import ArgumentParser


def setup_logging():
    logger.setLevel(logging.DEBUG)
    handler = logging.StreamHandler()
    logger.addHandler(handler)


def get_session():
    adapter = CacheControlAdapter(
        DictCache(), cache_etags=True, serializer=None, heuristic=None
    )
    sess = requests.Session()
    sess.mount("http://", adapter)
    sess.mount("https://", adapter)

    sess.cache_controller = adapter.controller
    return sess


def get_args():
    parser = ArgumentParser()
    parser.add_argument("url", help="The URL to try and cache")
    return parser.parse_args()


def main(args=None):
    args = get_args()
    sess = get_session()

    # Make a request to get a response
    resp = sess.get(args.url)

    # Turn on logging
    setup_logging()

    # try setting the cache
    sess.cache_controller.cache_response(resp.request, resp.raw)

    # Now try to get it
    if sess.cache_controller.cached_request(resp.request):
        print("Cached!")
    else:
        print("Not cached :(")


if __name__ == "__main__":
    main()
site-packages/pip/_vendor/cachecontrol/serialize.py000064400000015452151464452710016535 0ustar00import base64
import io
import json
import zlib

from pip._vendor import msgpack
from pip._vendor.requests.structures import CaseInsensitiveDict

from .compat import HTTPResponse, pickle, text_type


def _b64_decode_bytes(b):
    return base64.b64decode(b.encode("ascii"))


def _b64_decode_str(s):
    return _b64_decode_bytes(s).decode("utf8")


class Serializer(object):

    def dumps(self, request, response, body=None):
        response_headers = CaseInsensitiveDict(response.headers)

        if body is None:
            body = response.read(decode_content=False)

            # NOTE: 99% sure this is dead code. I'm only leaving it
            #       here b/c I don't have a test yet to prove
            #       it. Basically, before using
            #       `cachecontrol.filewrapper.CallbackFileWrapper`,
            #       this made an effort to reset the file handle. The
            #       `CallbackFileWrapper` short circuits this code by
            #       setting the body as the content is consumed, the
            #       result being a `body` argument is *always* passed
            #       into cache_response, and in turn,
            #       `Serializer.dump`.
            response._fp = io.BytesIO(body)

        # NOTE: This is all a bit weird, but it's really important that on
        #       Python 2.x these objects are unicode and not str, even when
        #       they contain only ascii. The problem here is that msgpack
        #       understands the difference between unicode and bytes and we
        #       have it set to differentiate between them, however Python 2
        #       doesn't know the difference. Forcing these to unicode will be
        #       enough to have msgpack know the difference.
        data = {
            u"response": {
                u"body": body,
                u"headers": dict(
                    (text_type(k), text_type(v)) for k, v in response.headers.items()
                ),
                u"status": response.status,
                u"version": response.version,
                u"reason": text_type(response.reason),
                u"strict": response.strict,
                u"decode_content": response.decode_content,
            }
        }

        # Construct our vary headers
        data[u"vary"] = {}
        if u"vary" in response_headers:
            varied_headers = response_headers[u"vary"].split(",")
            for header in varied_headers:
                header = text_type(header).strip()
                header_value = request.headers.get(header, None)
                if header_value is not None:
                    header_value = text_type(header_value)
                data[u"vary"][header] = header_value

        return b",".join([b"cc=4", msgpack.dumps(data, use_bin_type=True)])

    def loads(self, request, data):
        # Short circuit if we've been given an empty set of data
        if not data:
            return

        # Determine what version of the serializer the data was serialized
        # with
        try:
            ver, data = data.split(b",", 1)
        except ValueError:
            ver = b"cc=0"

        # Make sure that our "ver" is actually a version and isn't a false
        # positive from a , being in the data stream.
        if ver[:3] != b"cc=":
            data = ver + data
            ver = b"cc=0"

        # Get the version number out of the cc=N
        ver = ver.split(b"=", 1)[-1].decode("ascii")

        # Dispatch to the actual load method for the given version
        try:
            return getattr(self, "_loads_v{}".format(ver))(request, data)

        except AttributeError:
            # This is a version we don't have a loads function for, so we'll
            # just treat it as a miss and return None
            return

    def prepare_response(self, request, cached):
        """Verify our vary headers match and construct a real urllib3
        HTTPResponse object.
        """
        # Special case the '*' Vary value as it means we cannot actually
        # determine if the cached response is suitable for this request.
        if "*" in cached.get("vary", {}):
            return

        # Ensure that the Vary headers for the cached response match our
        # request
        for header, value in cached.get("vary", {}).items():
            if request.headers.get(header, None) != value:
                return

        body_raw = cached["response"].pop("body")

        headers = CaseInsensitiveDict(data=cached["response"]["headers"])
        if headers.get("transfer-encoding", "") == "chunked":
            headers.pop("transfer-encoding")

        cached["response"]["headers"] = headers

        try:
            body = io.BytesIO(body_raw)
        except TypeError:
            # This can happen if cachecontrol serialized to v1 format (pickle)
            # using Python 2. A Python 2 str(byte string) will be unpickled as
            # a Python 3 str (unicode string), which will cause the above to
            # fail with:
            #
            #     TypeError: 'str' does not support the buffer interface
            body = io.BytesIO(body_raw.encode("utf8"))

        return HTTPResponse(body=body, preload_content=False, **cached["response"])

    def _loads_v0(self, request, data):
        # The original legacy cache data. This doesn't contain enough
        # information to construct everything we need, so we'll treat this as
        # a miss.
        return

    def _loads_v1(self, request, data):
        try:
            cached = pickle.loads(data)
        except ValueError:
            return

        return self.prepare_response(request, cached)

    def _loads_v2(self, request, data):
        try:
            cached = json.loads(zlib.decompress(data).decode("utf8"))
        except (ValueError, zlib.error):
            return

        # We need to decode the items that we've base64 encoded
        cached["response"]["body"] = _b64_decode_bytes(cached["response"]["body"])
        cached["response"]["headers"] = dict(
            (_b64_decode_str(k), _b64_decode_str(v))
            for k, v in cached["response"]["headers"].items()
        )
        cached["response"]["reason"] = _b64_decode_str(cached["response"]["reason"])
        cached["vary"] = dict(
            (_b64_decode_str(k), _b64_decode_str(v) if v is not None else v)
            for k, v in cached["vary"].items()
        )

        return self.prepare_response(request, cached)

    def _loads_v3(self, request, data):
        # Due to Python 2 encoding issues, it's impossible to know for sure
        # exactly how to load v3 entries, thus we'll treat these as a miss so
        # that they get rewritten out as v4 entries.
        return

    def _loads_v4(self, request, data):
        try:
            cached = msgpack.loads(data, encoding="utf-8")
        except ValueError:
            return

        return self.prepare_response(request, cached)
site-packages/pip/_vendor/cachecontrol/controller.py000064400000032602151464452710016725 0ustar00"""
The httplib2 algorithms ported for use with requests.
"""
import logging
import re
import calendar
import time
from email.utils import parsedate_tz

from pip._vendor.requests.structures import CaseInsensitiveDict

from .cache import DictCache
from .serialize import Serializer


logger = logging.getLogger(__name__)

URI = re.compile(r"^(([^:/?#]+):)?(//([^/?#]*))?([^?#]*)(\?([^#]*))?(#(.*))?")


def parse_uri(uri):
    """Parses a URI using the regex given in Appendix B of RFC 3986.

        (scheme, authority, path, query, fragment) = parse_uri(uri)
    """
    groups = URI.match(uri).groups()
    return (groups[1], groups[3], groups[4], groups[6], groups[8])


class CacheController(object):
    """An interface to see if request should cached or not.
    """

    def __init__(
        self, cache=None, cache_etags=True, serializer=None, status_codes=None
    ):
        self.cache = cache or DictCache()
        self.cache_etags = cache_etags
        self.serializer = serializer or Serializer()
        self.cacheable_status_codes = status_codes or (200, 203, 300, 301)

    @classmethod
    def _urlnorm(cls, uri):
        """Normalize the URL to create a safe key for the cache"""
        (scheme, authority, path, query, fragment) = parse_uri(uri)
        if not scheme or not authority:
            raise Exception("Only absolute URIs are allowed. uri = %s" % uri)

        scheme = scheme.lower()
        authority = authority.lower()

        if not path:
            path = "/"

        # Could do syntax based normalization of the URI before
        # computing the digest. See Section 6.2.2 of Std 66.
        request_uri = query and "?".join([path, query]) or path
        defrag_uri = scheme + "://" + authority + request_uri

        return defrag_uri

    @classmethod
    def cache_url(cls, uri):
        return cls._urlnorm(uri)

    def parse_cache_control(self, headers):
        known_directives = {
            # https://tools.ietf.org/html/rfc7234#section-5.2
            "max-age": (int, True),
            "max-stale": (int, False),
            "min-fresh": (int, True),
            "no-cache": (None, False),
            "no-store": (None, False),
            "no-transform": (None, False),
            "only-if-cached": (None, False),
            "must-revalidate": (None, False),
            "public": (None, False),
            "private": (None, False),
            "proxy-revalidate": (None, False),
            "s-maxage": (int, True),
        }

        cc_headers = headers.get("cache-control", headers.get("Cache-Control", ""))

        retval = {}

        for cc_directive in cc_headers.split(","):
            if not cc_directive.strip():
                continue

            parts = cc_directive.split("=", 1)
            directive = parts[0].strip()

            try:
                typ, required = known_directives[directive]
            except KeyError:
                logger.debug("Ignoring unknown cache-control directive: %s", directive)
                continue

            if not typ or not required:
                retval[directive] = None
            if typ:
                try:
                    retval[directive] = typ(parts[1].strip())
                except IndexError:
                    if required:
                        logger.debug(
                            "Missing value for cache-control " "directive: %s",
                            directive,
                        )
                except ValueError:
                    logger.debug(
                        "Invalid value for cache-control directive " "%s, must be %s",
                        directive,
                        typ.__name__,
                    )

        return retval

    def cached_request(self, request):
        """
        Return a cached response if it exists in the cache, otherwise
        return False.
        """
        cache_url = self.cache_url(request.url)
        logger.debug('Looking up "%s" in the cache', cache_url)
        cc = self.parse_cache_control(request.headers)

        # Bail out if the request insists on fresh data
        if "no-cache" in cc:
            logger.debug('Request header has "no-cache", cache bypassed')
            return False

        if "max-age" in cc and cc["max-age"] == 0:
            logger.debug('Request header has "max_age" as 0, cache bypassed')
            return False

        # Request allows serving from the cache, let's see if we find something
        cache_data = self.cache.get(cache_url)
        if cache_data is None:
            logger.debug("No cache entry available")
            return False

        # Check whether it can be deserialized
        resp = self.serializer.loads(request, cache_data)
        if not resp:
            logger.warning("Cache entry deserialization failed, entry ignored")
            return False

        # If we have a cached 301, return it immediately. We don't
        # need to test our response for other headers b/c it is
        # intrinsically "cacheable" as it is Permanent.
        # See:
        #   https://tools.ietf.org/html/rfc7231#section-6.4.2
        #
        # Client can try to refresh the value by repeating the request
        # with cache busting headers as usual (ie no-cache).
        if resp.status == 301:
            msg = (
                'Returning cached "301 Moved Permanently" response '
                "(ignoring date and etag information)"
            )
            logger.debug(msg)
            return resp

        headers = CaseInsensitiveDict(resp.headers)
        if not headers or "date" not in headers:
            if "etag" not in headers:
                # Without date or etag, the cached response can never be used
                # and should be deleted.
                logger.debug("Purging cached response: no date or etag")
                self.cache.delete(cache_url)
            logger.debug("Ignoring cached response: no date")
            return False

        now = time.time()
        date = calendar.timegm(parsedate_tz(headers["date"]))
        current_age = max(0, now - date)
        logger.debug("Current age based on date: %i", current_age)

        # TODO: There is an assumption that the result will be a
        #       urllib3 response object. This may not be best since we
        #       could probably avoid instantiating or constructing the
        #       response until we know we need it.
        resp_cc = self.parse_cache_control(headers)

        # determine freshness
        freshness_lifetime = 0

        # Check the max-age pragma in the cache control header
        if "max-age" in resp_cc:
            freshness_lifetime = resp_cc["max-age"]
            logger.debug("Freshness lifetime from max-age: %i", freshness_lifetime)

        # If there isn't a max-age, check for an expires header
        elif "expires" in headers:
            expires = parsedate_tz(headers["expires"])
            if expires is not None:
                expire_time = calendar.timegm(expires) - date
                freshness_lifetime = max(0, expire_time)
                logger.debug("Freshness lifetime from expires: %i", freshness_lifetime)

        # Determine if we are setting freshness limit in the
        # request. Note, this overrides what was in the response.
        if "max-age" in cc:
            freshness_lifetime = cc["max-age"]
            logger.debug(
                "Freshness lifetime from request max-age: %i", freshness_lifetime
            )

        if "min-fresh" in cc:
            min_fresh = cc["min-fresh"]
            # adjust our current age by our min fresh
            current_age += min_fresh
            logger.debug("Adjusted current age from min-fresh: %i", current_age)

        # Return entry if it is fresh enough
        if freshness_lifetime > current_age:
            logger.debug('The response is "fresh", returning cached response')
            logger.debug("%i > %i", freshness_lifetime, current_age)
            return resp

        # we're not fresh. If we don't have an Etag, clear it out
        if "etag" not in headers:
            logger.debug('The cached response is "stale" with no etag, purging')
            self.cache.delete(cache_url)

        # return the original handler
        return False

    def conditional_headers(self, request):
        cache_url = self.cache_url(request.url)
        resp = self.serializer.loads(request, self.cache.get(cache_url))
        new_headers = {}

        if resp:
            headers = CaseInsensitiveDict(resp.headers)

            if "etag" in headers:
                new_headers["If-None-Match"] = headers["ETag"]

            if "last-modified" in headers:
                new_headers["If-Modified-Since"] = headers["Last-Modified"]

        return new_headers

    def cache_response(self, request, response, body=None, status_codes=None):
        """
        Algorithm for caching requests.

        This assumes a requests Response object.
        """
        # From httplib2: Don't cache 206's since we aren't going to
        #                handle byte range requests
        cacheable_status_codes = status_codes or self.cacheable_status_codes
        if response.status not in cacheable_status_codes:
            logger.debug(
                "Status code %s not in %s", response.status, cacheable_status_codes
            )
            return

        response_headers = CaseInsensitiveDict(response.headers)

        # If we've been given a body, our response has a Content-Length, that
        # Content-Length is valid then we can check to see if the body we've
        # been given matches the expected size, and if it doesn't we'll just
        # skip trying to cache it.
        if (
            body is not None
            and "content-length" in response_headers
            and response_headers["content-length"].isdigit()
            and int(response_headers["content-length"]) != len(body)
        ):
            return

        cc_req = self.parse_cache_control(request.headers)
        cc = self.parse_cache_control(response_headers)

        cache_url = self.cache_url(request.url)
        logger.debug('Updating cache with response from "%s"', cache_url)

        # Delete it from the cache if we happen to have it stored there
        no_store = False
        if "no-store" in cc:
            no_store = True
            logger.debug('Response header has "no-store"')
        if "no-store" in cc_req:
            no_store = True
            logger.debug('Request header has "no-store"')
        if no_store and self.cache.get(cache_url):
            logger.debug('Purging existing cache entry to honor "no-store"')
            self.cache.delete(cache_url)
        if no_store:
            return

        # If we've been given an etag, then keep the response
        if self.cache_etags and "etag" in response_headers:
            logger.debug("Caching due to etag")
            self.cache.set(
                cache_url, self.serializer.dumps(request, response, body=body)
            )

        # Add to the cache any 301s. We do this before looking that
        # the Date headers.
        elif response.status == 301:
            logger.debug("Caching permanant redirect")
            self.cache.set(cache_url, self.serializer.dumps(request, response))

        # Add to the cache if the response headers demand it. If there
        # is no date header then we can't do anything about expiring
        # the cache.
        elif "date" in response_headers:
            # cache when there is a max-age > 0
            if "max-age" in cc and cc["max-age"] > 0:
                logger.debug("Caching b/c date exists and max-age > 0")
                self.cache.set(
                    cache_url, self.serializer.dumps(request, response, body=body)
                )

            # If the request can expire, it means we should cache it
            # in the meantime.
            elif "expires" in response_headers:
                if response_headers["expires"]:
                    logger.debug("Caching b/c of expires header")
                    self.cache.set(
                        cache_url, self.serializer.dumps(request, response, body=body)
                    )

    def update_cached_response(self, request, response):
        """On a 304 we will get a new set of headers that we want to
        update our cached value with, assuming we have one.

        This should only ever be called when we've sent an ETag and
        gotten a 304 as the response.
        """
        cache_url = self.cache_url(request.url)

        cached_response = self.serializer.loads(request, self.cache.get(cache_url))

        if not cached_response:
            # we didn't have a cached response
            return response

        # Lets update our headers with the headers from the new request:
        # http://tools.ietf.org/html/draft-ietf-httpbis-p4-conditional-26#section-4.1
        #
        # The server isn't supposed to send headers that would make
        # the cached body invalid. But... just in case, we'll be sure
        # to strip out ones we know that might be problmatic due to
        # typical assumptions.
        excluded_headers = ["content-length"]

        cached_response.headers.update(
            dict(
                (k, v)
                for k, v in response.headers.items()
                if k.lower() not in excluded_headers
            )
        )

        # we want a 200 b/c we have content via the cache
        cached_response.status = 200

        # update our cache
        self.cache.set(cache_url, self.serializer.dumps(request, cached_response))

        return cached_response
site-packages/pip/_vendor/cachecontrol/caches/file_cache.py000064400000010071151464452710020026 0ustar00import hashlib
import os
from textwrap import dedent

from ..cache import BaseCache
from ..controller import CacheController

try:
    FileNotFoundError
except NameError:
    # py2.X
    FileNotFoundError = (IOError, OSError)


def _secure_open_write(filename, fmode):
    # We only want to write to this file, so open it in write only mode
    flags = os.O_WRONLY

    # os.O_CREAT | os.O_EXCL will fail if the file already exists, so we only
    #  will open *new* files.
    # We specify this because we want to ensure that the mode we pass is the
    # mode of the file.
    flags |= os.O_CREAT | os.O_EXCL

    # Do not follow symlinks to prevent someone from making a symlink that
    # we follow and insecurely open a cache file.
    if hasattr(os, "O_NOFOLLOW"):
        flags |= os.O_NOFOLLOW

    # On Windows we'll mark this file as binary
    if hasattr(os, "O_BINARY"):
        flags |= os.O_BINARY

    # Before we open our file, we want to delete any existing file that is
    # there
    try:
        os.remove(filename)
    except (IOError, OSError):
        # The file must not exist already, so we can just skip ahead to opening
        pass

    # Open our file, the use of os.O_CREAT | os.O_EXCL will ensure that if a
    # race condition happens between the os.remove and this line, that an
    # error will be raised. Because we utilize a lockfile this should only
    # happen if someone is attempting to attack us.
    fd = os.open(filename, flags, fmode)
    try:
        return os.fdopen(fd, "wb")

    except:
        # An error occurred wrapping our FD in a file object
        os.close(fd)
        raise


class FileCache(BaseCache):

    def __init__(
        self,
        directory,
        forever=False,
        filemode=0o0600,
        dirmode=0o0700,
        use_dir_lock=None,
        lock_class=None,
    ):

        if use_dir_lock is not None and lock_class is not None:
            raise ValueError("Cannot use use_dir_lock and lock_class together")

        try:
            from lockfile import LockFile
            from lockfile.mkdirlockfile import MkdirLockFile
        except ImportError:
            notice = dedent(
                """
            NOTE: In order to use the FileCache you must have
            lockfile installed. You can install it via pip:
              pip install lockfile
            """
            )
            raise ImportError(notice)

        else:
            if use_dir_lock:
                lock_class = MkdirLockFile

            elif lock_class is None:
                lock_class = LockFile

        self.directory = directory
        self.forever = forever
        self.filemode = filemode
        self.dirmode = dirmode
        self.lock_class = lock_class

    @staticmethod
    def encode(x):
        return hashlib.sha224(x.encode()).hexdigest()

    def _fn(self, name):
        # NOTE: This method should not change as some may depend on it.
        #       See: https://github.com/ionrock/cachecontrol/issues/63
        hashed = self.encode(name)
        parts = list(hashed[:5]) + [hashed]
        return os.path.join(self.directory, *parts)

    def get(self, key):
        name = self._fn(key)
        try:
            with open(name, "rb") as fh:
                return fh.read()

        except FileNotFoundError:
            return None

    def set(self, key, value):
        name = self._fn(key)

        # Make sure the directory exists
        try:
            os.makedirs(os.path.dirname(name), self.dirmode)
        except (IOError, OSError):
            pass

        with self.lock_class(name) as lock:
            # Write our actual file
            with _secure_open_write(lock.path, self.filemode) as fh:
                fh.write(value)

    def delete(self, key):
        name = self._fn(key)
        if not self.forever:
            try:
                os.remove(name)
            except FileNotFoundError:
                pass


def url_to_file_path(url, filecache):
    """Return the file cache path based on the URL.

    This does not ensure the file exists!
    """
    key = CacheController.cache_url(url)
    return filecache._fn(key)
site-packages/pip/_vendor/cachecontrol/caches/__pycache__/redis_cache.cpython-38.pyc000064400000002771151464452710024513 0ustar00U

.eX@s8ddlmZddlmZddlmZGdddeZdS))division)datetime)	BaseCachec@s>eZdZddZddZdddZdd	Zd
dZdd
ZdS)
RedisCachecCs
||_dSN)conn)selfrr	O/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/caches/redis_cache.py__init__	szRedisCache.__init__cCs|j|Sr)rgetrkeyr	r	r
rszRedisCache.getNcCs<|s|j||n$|t}|j|t||dSr)rsetrZutcnowZsetexintZ
total_seconds)rrvalueZexpiresr	r	r
rszRedisCache.setcCs|j|dSr)rdeleter
r	r	r
rszRedisCache.deletecCs |jD]}|j|q
dS)zIHelper for clearing all the keys in a database. Use with
        caution!N)rkeysrr
r	r	r
clearszRedisCache.clearcCsdS)z?Redis uses connection pooling, no need to close the connection.Nr	)rr	r	r
closeszRedisCache.close)N)	__name__
__module____qualname__rrrrrrr	r	r	r
rs
rN)Z
__future__rrZpip._vendor.cachecontrol.cacherrr	r	r	r
ssite-packages/pip/_vendor/cachecontrol/caches/__pycache__/__init__.cpython-38.pyc000064400000000401151464452710024005 0ustar00U

.eV@sddlmZddlmZdS))	FileCache)
RedisCacheN)Z
file_cacherZredis_cacherrrL/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/caches/__init__.pyssite-packages/pip/_vendor/cachecontrol/caches/__pycache__/redis_cache.cpython-38.opt-1.pyc000064400000002771151464452710025452 0ustar00U

.eX@s8ddlmZddlmZddlmZGdddeZdS))division)datetime)	BaseCachec@s>eZdZddZddZdddZdd	Zd
dZdd
ZdS)
RedisCachecCs
||_dSN)conn)selfrr	O/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/caches/redis_cache.py__init__	szRedisCache.__init__cCs|j|Sr)rgetrkeyr	r	r
rszRedisCache.getNcCs<|s|j||n$|t}|j|t||dSr)rsetrZutcnowZsetexintZ
total_seconds)rrvalueZexpiresr	r	r
rszRedisCache.setcCs|j|dSr)rdeleter
r	r	r
rszRedisCache.deletecCs |jD]}|j|q
dS)zIHelper for clearing all the keys in a database. Use with
        caution!N)rkeysrr
r	r	r
clearszRedisCache.clearcCsdS)z?Redis uses connection pooling, no need to close the connection.Nr	)rr	r	r
closeszRedisCache.close)N)	__name__
__module____qualname__rrrrrrr	r	r	r
rs
rN)Z
__future__rrZpip._vendor.cachecontrol.cacherrr	r	r	r
ssite-packages/pip/_vendor/cachecontrol/caches/__pycache__/file_cache.cpython-38.opt-1.pyc000064400000006237151464452710025264 0ustar00U

.e9@s~ddlZddlZddlmZddlmZddlmZzeWne	k
rXe
efZYnXddZGdd	d	eZ
d
dZdS)N)dedent)	BaseCache)CacheControllerc	Cstj}|tjtjBO}ttdr*|tjO}ttdr>|tjO}zt|Wntt	fk
rdYnXt
|||}zt|dWSt|YnXdS)N
O_NOFOLLOWO_BINARYwb)
osO_WRONLYO_CREATO_EXCLhasattrrrremoveIOErrorOSErroropenfdopenclose)filenameZfmodeflagsfdrN/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/caches/file_cache.py_secure_open_writes 




rc@sBeZdZdddZeddZd	d
ZddZd
dZddZ	dS)	FileCacheFNc
Cs|dk	r|dk	rtdzddlm}ddlm}Wn$tk
rXtd}	t|	YnX|rd|}n|dkrp|}||_||_||_	||_
||_dS)Nz/Cannot use use_dir_lock and lock_class togetherr)LockFile)
MkdirLockFilez
            NOTE: In order to use the FileCache you must have
            lockfile installed. You can install it via pip:
              pip install lockfile
            )
ValueErrorZlockfilerZlockfile.mkdirlockfilerImportErrorr	directoryforeverfilemodedirmode
lock_class)
selfr!r"r#r$Zuse_dir_lockr%rrZnoticerrr__init__:s&
zFileCache.__init__cCst|SN)hashlibZsha224encodeZ	hexdigest)xrrrr*aszFileCache.encodecCs4||}t|dd|g}tjj|jf|S)N)r*listr	pathjoinr!)r&nameZhashedpartsrrr_fnes
z
FileCache._fnc
CsR||}z,t|d}|W5QRWSQRXWntk
rLYdSXdS)Nrb)r2rreadFileNotFoundError)r&keyr0fhrrrgetls
 z
FileCache.getc
Cs|||}zttj||jWnttfk
r<YnX||*}t	|j|j
}||W5QRXW5QRXdSr()r2r	makedirsr.dirnamer$rrr%rr#write)r&r6valuer0lockr7rrrsetus
z
FileCache.setcCs8||}|js4zt|Wntk
r2YnXdSr()r2r"r	rr5)r&r6r0rrrdeletes
zFileCache.delete)FrrNN)
__name__
__module____qualname__r'staticmethodr*r2r8r>r?rrrrr8s
'
	rcCst|}||S)z\Return the file cache path based on the URL.

    This does not ensure the file exists!
    )rZ	cache_urlr2)ZurlZ	filecacher6rrrurl_to_file_paths
rD)r)r	textwraprcacherZ
controllerrr5	NameErrorrrrrrDrrrrs)Tsite-packages/pip/_vendor/cachecontrol/caches/__pycache__/file_cache.cpython-38.pyc000064400000006237151464452710024325 0ustar00U

.e9@s~ddlZddlZddlmZddlmZddlmZzeWne	k
rXe
efZYnXddZGdd	d	eZ
d
dZdS)N)dedent)	BaseCache)CacheControllerc	Cstj}|tjtjBO}ttdr*|tjO}ttdr>|tjO}zt|Wntt	fk
rdYnXt
|||}zt|dWSt|YnXdS)N
O_NOFOLLOWO_BINARYwb)
osO_WRONLYO_CREATO_EXCLhasattrrrremoveIOErrorOSErroropenfdopenclose)filenameZfmodeflagsfdrN/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/caches/file_cache.py_secure_open_writes 




rc@sBeZdZdddZeddZd	d
ZddZd
dZddZ	dS)	FileCacheFNc
Cs|dk	r|dk	rtdzddlm}ddlm}Wn$tk
rXtd}	t|	YnX|rd|}n|dkrp|}||_||_||_	||_
||_dS)Nz/Cannot use use_dir_lock and lock_class togetherr)LockFile)
MkdirLockFilez
            NOTE: In order to use the FileCache you must have
            lockfile installed. You can install it via pip:
              pip install lockfile
            )
ValueErrorZlockfilerZlockfile.mkdirlockfilerImportErrorr	directoryforeverfilemodedirmode
lock_class)
selfr!r"r#r$Zuse_dir_lockr%rrZnoticerrr__init__:s&
zFileCache.__init__cCst|SN)hashlibZsha224encodeZ	hexdigest)xrrrr*aszFileCache.encodecCs4||}t|dd|g}tjj|jf|S)N)r*listr	pathjoinr!)r&nameZhashedpartsrrr_fnes
z
FileCache._fnc
CsR||}z,t|d}|W5QRWSQRXWntk
rLYdSXdS)Nrb)r2rreadFileNotFoundError)r&keyr0fhrrrgetls
 z
FileCache.getc
Cs|||}zttj||jWnttfk
r<YnX||*}t	|j|j
}||W5QRXW5QRXdSr()r2r	makedirsr.dirnamer$rrr%rr#write)r&r6valuer0lockr7rrrsetus
z
FileCache.setcCs8||}|js4zt|Wntk
r2YnXdSr()r2r"r	rr5)r&r6r0rrrdeletes
zFileCache.delete)FrrNN)
__name__
__module____qualname__r'staticmethodr*r2r8r>r?rrrrr8s
'
	rcCst|}||S)z\Return the file cache path based on the URL.

    This does not ensure the file exists!
    )rZ	cache_urlr2)ZurlZ	filecacher6rrrurl_to_file_paths
rD)r)r	textwraprcacherZ
controllerrr5	NameErrorrrrrrDrrrrs)Tsite-packages/pip/_vendor/cachecontrol/caches/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000401151464452710024744 0ustar00U

.eV@sddlmZddlmZdS))	FileCache)
RedisCacheN)Z
file_cacherZredis_cacherrrL/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/caches/__init__.pyssite-packages/pip/_vendor/cachecontrol/caches/redis_cache.py000064400000001530151464452710020215 0ustar00from __future__ import division

from datetime import datetime
from pip._vendor.cachecontrol.cache import BaseCache


class RedisCache(BaseCache):

    def __init__(self, conn):
        self.conn = conn

    def get(self, key):
        return self.conn.get(key)

    def set(self, key, value, expires=None):
        if not expires:
            self.conn.set(key, value)
        else:
            expires = expires - datetime.utcnow()
            self.conn.setex(key, int(expires.total_seconds()), value)

    def delete(self, key):
        self.conn.delete(key)

    def clear(self):
        """Helper for clearing all the keys in a database. Use with
        caution!"""
        for key in self.conn.keys():
            self.conn.delete(key)

    def close(self):
        """Redis uses connection pooling, no need to close the connection."""
        pass
site-packages/pip/_vendor/cachecontrol/caches/__init__.py000064400000000126151464452710017523 0ustar00from .file_cache import FileCache  # noqa
from .redis_cache import RedisCache  # noqa
site-packages/pip/_vendor/cachecontrol/heuristics.py000064400000007746151464452710016737 0ustar00import calendar
import time

from email.utils import formatdate, parsedate, parsedate_tz

from datetime import datetime, timedelta

TIME_FMT = "%a, %d %b %Y %H:%M:%S GMT"


def expire_after(delta, date=None):
    date = date or datetime.utcnow()
    return date + delta


def datetime_to_header(dt):
    return formatdate(calendar.timegm(dt.timetuple()))


class BaseHeuristic(object):

    def warning(self, response):
        """
        Return a valid 1xx warning header value describing the cache
        adjustments.

        The response is provided too allow warnings like 113
        http://tools.ietf.org/html/rfc7234#section-5.5.4 where we need
        to explicitly say response is over 24 hours old.
        """
        return '110 - "Response is Stale"'

    def update_headers(self, response):
        """Update the response headers with any new headers.

        NOTE: This SHOULD always include some Warning header to
              signify that the response was cached by the client, not
              by way of the provided headers.
        """
        return {}

    def apply(self, response):
        updated_headers = self.update_headers(response)

        if updated_headers:
            response.headers.update(updated_headers)
            warning_header_value = self.warning(response)
            if warning_header_value is not None:
                response.headers.update({"Warning": warning_header_value})

        return response


class OneDayCache(BaseHeuristic):
    """
    Cache the response by providing an expires 1 day in the
    future.
    """

    def update_headers(self, response):
        headers = {}

        if "expires" not in response.headers:
            date = parsedate(response.headers["date"])
            expires = expire_after(timedelta(days=1), date=datetime(*date[:6]))
            headers["expires"] = datetime_to_header(expires)
            headers["cache-control"] = "public"
        return headers


class ExpiresAfter(BaseHeuristic):
    """
    Cache **all** requests for a defined time period.
    """

    def __init__(self, **kw):
        self.delta = timedelta(**kw)

    def update_headers(self, response):
        expires = expire_after(self.delta)
        return {"expires": datetime_to_header(expires), "cache-control": "public"}

    def warning(self, response):
        tmpl = "110 - Automatically cached for %s. Response might be stale"
        return tmpl % self.delta


class LastModified(BaseHeuristic):
    """
    If there is no Expires header already, fall back on Last-Modified
    using the heuristic from
    http://tools.ietf.org/html/rfc7234#section-4.2.2
    to calculate a reasonable value.

    Firefox also does something like this per
    https://developer.mozilla.org/en-US/docs/Web/HTTP/Caching_FAQ
    http://lxr.mozilla.org/mozilla-release/source/netwerk/protocol/http/nsHttpResponseHead.cpp#397
    Unlike mozilla we limit this to 24-hr.
    """
    cacheable_by_default_statuses = {
        200, 203, 204, 206, 300, 301, 404, 405, 410, 414, 501
    }

    def update_headers(self, resp):
        headers = resp.headers

        if "expires" in headers:
            return {}

        if "cache-control" in headers and headers["cache-control"] != "public":
            return {}

        if resp.status not in self.cacheable_by_default_statuses:
            return {}

        if "date" not in headers or "last-modified" not in headers:
            return {}

        date = calendar.timegm(parsedate_tz(headers["date"]))
        last_modified = parsedate(headers["last-modified"])
        if date is None or last_modified is None:
            return {}

        now = time.time()
        current_age = max(0, now - date)
        delta = date - calendar.timegm(last_modified)
        freshness_lifetime = max(0, min(delta / 10, 24 * 3600))
        if freshness_lifetime <= current_age:
            return {}

        expires = date + freshness_lifetime
        return {"expires": time.strftime(TIME_FMT, time.gmtime(expires))}

    def warning(self, resp):
        return None
site-packages/pip/_vendor/cachecontrol/__pycache__/cache.cpython-38.opt-1.pyc000064400000003322151464452710023027 0ustar00U

.e%@s4dZddlmZGdddeZGdddeZdS)zb
The cache object API for implementing caches. The default is a thread
safe in-memory dictionary.
)Lockc@s,eZdZddZddZddZddZd	S)
	BaseCachecCs
tdSNNotImplementedErrorselfkeyr
B/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/cache.pyget
sz
BaseCache.getcCs
tdSrrrr	valuer
r
rset
sz
BaseCache.setcCs
tdSrrrr
r
rdeleteszBaseCache.deletecCsdSrr
)rr
r
rcloseszBaseCache.closeN)__name__
__module____qualname__rrrrr
r
r
rrsrc@s.eZdZd
ddZddZddZdd	ZdS)	DictCacheNcCst|_|pi|_dSr)rlockdata)rZ	init_dictr
r
r__init__szDictCache.__init__cCs|j|dSr)rrrr
r
rrsz
DictCache.getc	Cs&|j|j||iW5QRXdSr)rrupdater
r
r
rr sz
DictCache.setc	Cs,|j||jkr|j|W5QRXdSr)rrpoprr
r
rr$s
zDictCache.delete)N)rrrrrrrr
r
r
rrs
rN)__doc__Z	threadingrobjectrrr
r
r
rssite-packages/pip/_vendor/cachecontrol/__pycache__/heuristics.cpython-38.pyc000064400000011132151464452710023205 0ustar00U

.e@sddlZddlZddlmZmZmZddlmZmZdZdddZ	ddZ
Gd	d
d
eZGdddeZ
Gd
ddeZGdddeZdS)N)
formatdate	parsedateparsedate_tz)datetime	timedeltaz%a, %d %b %Y %H:%M:%S GMTcCs|p
t}||SN)rZutcnow)deltadater
G/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/heuristics.pyexpire_aftersrcCstt|Sr)rcalendartimegmZ	timetuple)Zdtr
r
rdatetime_to_headersrc@s$eZdZddZddZddZdS)
BaseHeuristiccCsdS)a!
        Return a valid 1xx warning header value describing the cache
        adjustments.

        The response is provided too allow warnings like 113
        http://tools.ietf.org/html/rfc7234#section-5.5.4 where we need
        to explicitly say response is over 24 hours old.
        z110 - "Response is Stale"r
selfresponser
r
rwarnings	zBaseHeuristic.warningcCsiS)zUpdate the response headers with any new headers.

        NOTE: This SHOULD always include some Warning header to
              signify that the response was cached by the client, not
              by way of the provided headers.
        r
rr
r
rupdate_headers!szBaseHeuristic.update_headerscCs@||}|r<|j|||}|dk	r<|jd|i|S)NWarning)rheadersupdater)rrZupdated_headersZwarning_header_valuer
r
rapply*s

zBaseHeuristic.applyN)__name__
__module____qualname__rrrr
r
r
rrs	rc@seZdZdZddZdS)OneDayCachezM
    Cache the response by providing an expires 1 day in the
    future.
    cCsRi}d|jkrNt|jd}ttddt|ddd}t||d<d|d<|S)	Nexpiresr	)Zdays)r	public
cache-control)rrrrrr)rrrr	rr
r
rr<s
zOneDayCache.update_headersN)rrr__doc__rr
r
r
rr6src@s(eZdZdZddZddZddZdS)	ExpiresAfterz;
    Cache **all** requests for a defined time period.
    cKstf||_dSr)rr)rkwr
r
r__init__LszExpiresAfter.__init__cCst|j}t|ddS)Nr!)rr")rrr)rrrr
r
rrOs
zExpiresAfter.update_headerscCsd}||jS)Nz:110 - Automatically cached for %s. Response might be stale)r)rrZtmplr
r
rrSszExpiresAfter.warningN)rrrr#r&rrr
r
r
rr$Gsr$c@s:eZdZdZdddddddd	d
ddhZd
dZddZdS)LastModifieda
    If there is no Expires header already, fall back on Last-Modified
    using the heuristic from
    http://tools.ietf.org/html/rfc7234#section-4.2.2
    to calculate a reasonable value.

    Firefox also does something like this per
    https://developer.mozilla.org/en-US/docs/Web/HTTP/Caching_FAQ
    http://lxr.mozilla.org/mozilla-release/source/netwerk/protocol/http/nsHttpResponseHead.cpp#397
    Unlike mozilla we limit this to 24-hr.
    i,i-iiiiic
Cs|j}d|kriSd|kr*|ddkr*iS|j|jkr:iSd|ksJd|krNiStt|d}t|d}|dks||dkriSt}td||}|t|}tdt	|dd}||kriS||}	dt
tt|	iS)	Nrr"r!r	z
last-modifiedr
iQ)
rZstatuscacheable_by_default_statusesr
rrrtimemaxminstrftimeTIME_FMTgmtime)
rresprr	Z
last_modifiedZnowZcurrent_agerZfreshness_lifetimerr
r
rrhs*zLastModified.update_headerscCsdSrr
)rr4r
r
rrszLastModified.warningN)rrrr#r-rrr
r
r
rr'Xsr')N)r
r.Zemail.utilsrrrrrr2rrobjectrrr$r'r
r
r
rs
"site-packages/pip/_vendor/cachecontrol/__pycache__/compat.cpython-38.pyc000064400000001320151464452710022304 0ustar00U

.e@szddlmZWn ek
r0ddlmZYnXzddlZWnek
rZddlZYnXzddlmZWn ek
rddlmZYnXzddl	m
Z
Wn ek
rddlm
Z
YnXzeZ
Wnek
reZ
YnXdS))urljoinN)HTTPResponse)is_fp_closed)Zurllib.parserImportErrorZurlparseZcPicklepickleZ.pip._vendor.requests.packages.urllib3.responserZpip._vendor.urllib3.responseZ*pip._vendor.requests.packages.urllib3.utilrZpip._vendor.urllib3.utilZunicodeZ	text_type	NameErrorstrr	r	C/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/compat.pys&site-packages/pip/_vendor/cachecontrol/__pycache__/filewrapper.cpython-38.pyc000064400000004133151464452710023346 0ustar00U

.e	@s ddlmZGdddeZdS))BytesIOc@sBeZdZdZddZddZddZdd	ZdddZd
dZ	d
S)CallbackFileWrapperav
    Small wrapper around a fp object which will tee everything read into a
    buffer, and when that file is closed it will execute a callback with the
    contents of that buffer.

    All attributes are proxied to the underlying file object.

    This class uses members with a double underscore (__) leading prefix so as
    not to accidentally shadow an attribute.
    cCst|_||_||_dSN)r_CallbackFileWrapper__buf_CallbackFileWrapper__fp_CallbackFileWrapper__callback)selffpcallbackrH/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/filewrapper.py__init__szCallbackFileWrapper.__init__cCs|d}t||S)Nr)__getattribute__getattr)rnamer	rrr__getattr__s	
zCallbackFileWrapper.__getattr__cCsHz|jjdkWStk
r"YnXz
|jjWStk
rBYnXdS)NF)rr	AttributeErrorclosedrrrrZ__is_fp_closed!s
z"CallbackFileWrapper.__is_fp_closedcCs |jr||jd|_dSr)rrgetvaluerrrr_close2szCallbackFileWrapper._closeNcCs,|j|}|j||r(||Sr)rreadrwrite"_CallbackFileWrapper__is_fp_closedrrZamtdatarrrr=s
zCallbackFileWrapper.readcCs@|j|}|dkr |dkr |S|j||r<||S)Ns
)r
_safe_readrrrrrrrrrEszCallbackFileWrapper._safe_read)N)
__name__
__module____qualname____doc__r
rrrrrrrrrrs
rN)iorobjectrrrrrssite-packages/pip/_vendor/cachecontrol/__pycache__/controller.cpython-38.pyc000064400000016754151464452710023225 0ustar00U

.e5@sdZddlZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZe
eZedZd	d
ZGdddeZdS)
z7
The httplib2 algorithms ported for use with requests.
N)parsedate_tz)CaseInsensitiveDict)	DictCache)
Serializerz9^(([^:/?#]+):)?(//([^/?#]*))?([^?#]*)(\?([^#]*))?(#(.*))?cCs0t|}|d|d|d|d|dfS)zParses a URI using the regex given in Appendix B of RFC 3986.

        (scheme, authority, path, query, fragment) = parse_uri(uri)
    r)URImatchgroups)urir
rG/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/controller.py	parse_urisrc@s\eZdZdZdddZeddZedd	Zd
dZdd
Z	ddZ
dddZddZdS)CacheControllerz9An interface to see if request should cached or not.
    NTcCs,|pt|_||_|pt|_|p$d|_dS)N)i,-)rcachecache_etagsr
serializercacheable_status_codes)selfrrrstatus_codesrrr__init__"szCacheController.__init__c	Csht|\}}}}}|r|s&td||}|}|s>d}|rPd||gpR|}|d||}|S)z4Normalize the URL to create a safe key for the cachez(Only absolute URIs are allowed. uri = %s/?z://)r	Exceptionlowerjoin)	clsrZschemeZ	authoritypathZqueryZfragmentZrequest_uriZ
defrag_urirrr_urlnorm*szCacheController._urlnormcCs
||SN)r$)r"rrrr	cache_url>szCacheController.cache_urlc

Cs(tdftdftdfddddddddtdfd}|d|dd}i}|dD]}|s^qP|d	d
}|d}z||\}}	Wn$tk
rtd|YqPYnX|r|	sd||<|rPz||d
||<WqPtk
r|	rtd
|YqPtk
r td||j	YqPXqP|S)NTF)NF)max-agez	max-stale	min-freshno-cacheno-storezno-transformzonly-if-cachedzmust-revalidateZpublicZprivatezproxy-revalidatezs-maxagez
cache-controlz
Cache-Control,=rrz,Ignoring unknown cache-control directive: %sz-Missing value for cache-control directive: %sz8Invalid value for cache-control directive %s, must be %s)
intgetsplitstripKeyErrorloggerdebug
IndexError
ValueError__name__)
rheadersZknown_directivesZ
cc_headersZretvalZcc_directivepartsZ	directivetypZrequiredrrrparse_cache_controlBsV
z#CacheController.parse_cache_controlcCs0||j}td|||j}d|kr:tddSd|kr\|ddkr\tddS|j|}|dkr~td	dS|j	||}|st
d
dS|jdkrd}t||St|j}|rd
|krd|krtd|j
|tddSt}tt|d
}	td||	}
td|
||}d}d|kr^|d}td|nDd|krt|d}
|
dk	rt|
|	}td|}td|d|kr|d}td|d|kr|d}|
|7}
td|
||
krtdtd||
|Sd|kr,td|j
|dS)ze
        Return a cached response if it exists in the cache, otherwise
        return False.
        zLooking up "%s" in the cacher)z-Request header has "no-cache", cache bypassedFr'rz1Request header has "max_age" as 0, cache bypassedNzNo cache entry availablez1Cache entry deserialization failed, entry ignoredrzVReturning cached "301 Moved Permanently" response (ignoring date and etag information)dateetagz(Purging cached response: no date or etagz!Ignoring cached response: no datezCurrent age based on date: %iz#Freshness lifetime from max-age: %iexpiresz#Freshness lifetime from expires: %iz+Freshness lifetime from request max-age: %ir(z'Adjusted current age from min-fresh: %iz2The response is "fresh", returning cached responsez%i > %iz4The cached response is "stale" with no etag, purging)r&urlr3r4r;r8rr/rloadsZwarningstatusrdeletetimecalendarZtimegmrmax)rrequestr&ccZ
cache_datarespmsgr8Znowr<Zcurrent_ageZresp_ccZfreshness_lifetimer>Zexpire_timeZ	min_freshrrrcached_requestxs|




















zCacheController.cached_requestcCs`||j}|j||j|}i}|r\t|j}d|krH|d|d<d|kr\|d|d<|S)Nr=ZETagz
If-None-Matchz
last-modifiedz
Last-ModifiedzIf-Modified-Since)r&r?rr@rr/rr8)rrFr&rHZnew_headersr8rrrconditional_headerss
z#CacheController.conditional_headerscCs|p|j}|j|kr(td|j|dSt|j}|dk	rfd|krf|drft|dt|krfdS|	|j}|	|}|
|j}	td|	d}
d|krd}
tdd|krd}
td	|
r|j
|	rtd
|j|	|
rdS|jr,d|kr,td|j|	|jj|||d
n|jdkr\td|j|	|j||nd|krd|kr|ddkrtd|j|	|jj|||d
n:d|kr|drtd|j|	|jj|||d
dS)zc
        Algorithm for caching requests.

        This assumes a requests Response object.
        zStatus code %s not in %sNcontent-lengthz&Updating cache with response from "%s"Fr*TzResponse header has "no-store"zRequest header has "no-store"z0Purging existing cache entry to honor "no-store"r=zCaching due to etag)bodyrzCaching permanant redirectr<r'rz'Caching b/c date exists and max-age > 0r>zCaching b/c of expires header)rrAr3r4rr8isdigitr.lenr;r&r?rr/rBrsetrdumps)rrFresponserMrrZresponse_headersZcc_reqrGr&Zno_storerrrcache_responsesr














zCacheController.cache_responsecsv||j}|j||j|}|s*|Sdg|jtfdd|j	Dd|_
|j||j|||S)zOn a 304 we will get a new set of headers that we want to
        update our cached value with, assuming we have one.

        This should only ever be called when we've sent an ETag and
        gotten a 304 as the response.
        rLc3s&|]\}}|kr||fVqdSr%)r ).0kvZexcluded_headersrr	bsz9CacheController.update_cached_response..r)
r&r?rr@rr/r8updatedictitemsrArPrQ)rrFrRr&Zcached_responserrWrupdate_cached_responseIs		z&CacheController.update_cached_response)NTNN)NN)
r7
__module____qualname____doc__rclassmethodr$r&r;rJrKrSr\rrrrrs


6o
Rr)r_ZloggingrerDrCZemail.utilsrZpip._vendor.requests.structuresrrrZ	serializerZ	getLoggerr7r3compilerrobjectrrrrrs

	site-packages/pip/_vendor/cachecontrol/__pycache__/serialize.cpython-38.pyc000064400000010163151464452710023015 0ustar00U

.e*@spddlZddlZddlZddlZddlmZddlmZddlm	Z	m
Z
mZddZdd	Z
Gd
ddeZdS)N)msgpack)CaseInsensitiveDict)HTTPResponsepickle	text_typecCst|dS)Nascii)base64Z	b64decodeencode)brF/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/serialize.py_b64_decode_bytessrcCst|dS)Nutf8)rdecode)srrr
_b64_decode_strsrc@sNeZdZdddZddZddZdd	Zd
dZdd
ZddZ	ddZ
dS)
SerializerNc		Cst|j}|dkr*|jdd}t||_d|tdd|jD|j|j	t
|j|j|j
di}i|d<d|kr|dd}|D]:}t
|}|j|d}|dk	rt
|}||d|<qd	d
tj|ddgS)
NF)decode_contentresponsecss"|]\}}t|t|fVqdSN)r.0kvrrr
	2sz#Serializer.dumps..)bodyheadersstatusversionreasonstrictrvary,,scc=4T)Zuse_bin_type)rrreadioBytesIOZ_fpdictitemsrrrr r!rsplitstripgetjoinrdumps)	selfrequestrrZresponse_headersdataZvaried_headersheaderZheader_valuerrr
r.s4

zSerializer.dumpscCs|sdSz|dd\}}Wntk
r4d}YnX|dddkrR||}d}|dddd}zt|d	|||WStk
rYdSXdS)
Nr$rscc=0scc==rz
_loads_v{})r*
ValueErrorrgetattrformatAttributeError)r/r0r1Zverrrr
loadsJs
zSerializer.loadscCsd|dikrdS|diD] \}}|j|d|kr$dSq$|dd}t|ddd}|dd	d
kr|d||dd<zt|}Wn$tk
rt|d}YnXt	f|dd
|dS)z`Verify our vary headers match and construct a real urllib3
        HTTPResponse object.
        *r"Nrrr)r1ztransfer-encodingZchunkedrF)rZpreload_content)
r,r)rpoprr&r'	TypeErrorr
r)r/r0cachedr2valueZbody_rawrrrrr
prepare_responsehs
zSerializer.prepare_responsecCsdSrrr/r0r1rrr
	_loads_v0szSerializer._loads_v0cCs2zt|}Wntk
r$YdSX|||Sr)rr:r6rAr/r0r1r?rrr
	_loads_v1s
zSerializer._loads_v1c	Csztt|d}Wnttjfk
r6YdSXt|dd|dd<tdd|dd	D|dd<t
|dd|dd<tdd|d		D|d	<|||S)
Nrrrcss"|]\}}t|t|fVqdSrrrrrr
rsz'Serializer._loads_v2..rr css.|]&\}}t||dk	r t|n|fVqdSrrFrrrr
rsr")jsonr:zlib
decompressrr6errorrr(r)rrArDrrr
	_loads_v2s
zSerializer._loads_v2cCsdSrrrBrrr
	_loads_v3szSerializer._loads_v3cCs6ztj|dd}Wntk
r(YdSX|||S)Nzutf-8)encoding)rr:r6rArDrrr
	_loads_v4s
zSerializer._loads_v4)N)__name__
__module____qualname__r.r:rArCrErKrLrNrrrr
rs
4$r)r	r&rGrHZpip._vendorrZpip._vendor.requests.structuresrcompatrrrrrobjectrrrrr
ssite-packages/pip/_vendor/cachecontrol/__pycache__/__init__.cpython-38.pyc000064400000001001151464452710022554 0ustar00U

.e.@s8dZdZdZdZddlmZddlmZddlm	Z	dS)	zbCacheControl import Interface.

Make it easy to import from cachecontrol without long namespaces.
zEric Larsonzeric@ionrock.orgz0.12.5)CacheControl)CacheControlAdapter)CacheControllerN)
__doc__
__author__Z	__email____version__wrapperrZadapterrZ
controllerrr	r	E/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/__init__.pyssite-packages/pip/_vendor/cachecontrol/__pycache__/_cmd.cpython-38.pyc000064400000003004151464452710021724 0ustar00U

.e@sxddlZddlmZddlmZddlmZddlmZddl	m
Z
ddZd	d
ZddZ
dd
dZedkrtedS)N)requests)CacheControlAdapter)	DictCache)logger)ArgumentParsercCs"ttjt}t|dS)N)rZsetLevelloggingDEBUGZ
StreamHandlerZ
addHandler)Zhandlerr	A/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/_cmd.py
setup_loggingsrcCs>ttdddd}t}|d||d||j|_|S)NT)Zcache_etagsZ
serializerZ	heuristiczhttp://zhttps://)rrrZSessionZmountZ
controllercache_controller)Zadaptersessr	r	r
get_sessionsrcCst}|jddd|S)NurlzThe URL to try and cache)help)radd_argument
parse_args)parserr	r	r
get_argssrcCsTt}t}||j}t|j|j|j|j	|jrHt
dnt
ddS)NzCached!z
Not cached :()rrgetrrrZcache_responseZrequestrawZcached_requestprint)argsr
Zrespr	r	r
main$s
r__main__)N)rZpip._vendorrZ pip._vendor.cachecontrol.adapterrZpip._vendor.cachecontrol.cacherZ#pip._vendor.cachecontrol.controllerrargparserrrrr__name__r	r	r	r
s
site-packages/pip/_vendor/cachecontrol/__pycache__/heuristics.cpython-38.opt-1.pyc000064400000011132151464452710024144 0ustar00U

.e@sddlZddlZddlmZmZmZddlmZmZdZdddZ	ddZ
Gd	d
d
eZGdddeZ
Gd
ddeZGdddeZdS)N)
formatdate	parsedateparsedate_tz)datetime	timedeltaz%a, %d %b %Y %H:%M:%S GMTcCs|p
t}||SN)rZutcnow)deltadater
G/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/heuristics.pyexpire_aftersrcCstt|Sr)rcalendartimegmZ	timetuple)Zdtr
r
rdatetime_to_headersrc@s$eZdZddZddZddZdS)
BaseHeuristiccCsdS)a!
        Return a valid 1xx warning header value describing the cache
        adjustments.

        The response is provided too allow warnings like 113
        http://tools.ietf.org/html/rfc7234#section-5.5.4 where we need
        to explicitly say response is over 24 hours old.
        z110 - "Response is Stale"r
selfresponser
r
rwarnings	zBaseHeuristic.warningcCsiS)zUpdate the response headers with any new headers.

        NOTE: This SHOULD always include some Warning header to
              signify that the response was cached by the client, not
              by way of the provided headers.
        r
rr
r
rupdate_headers!szBaseHeuristic.update_headerscCs@||}|r<|j|||}|dk	r<|jd|i|S)NWarning)rheadersupdater)rrZupdated_headersZwarning_header_valuer
r
rapply*s

zBaseHeuristic.applyN)__name__
__module____qualname__rrrr
r
r
rrs	rc@seZdZdZddZdS)OneDayCachezM
    Cache the response by providing an expires 1 day in the
    future.
    cCsRi}d|jkrNt|jd}ttddt|ddd}t||d<d|d<|S)	Nexpiresr	)Zdays)r	public
cache-control)rrrrrr)rrrr	rr
r
rr<s
zOneDayCache.update_headersN)rrr__doc__rr
r
r
rr6src@s(eZdZdZddZddZddZdS)	ExpiresAfterz;
    Cache **all** requests for a defined time period.
    cKstf||_dSr)rr)rkwr
r
r__init__LszExpiresAfter.__init__cCst|j}t|ddS)Nr!)rr")rrr)rrrr
r
rrOs
zExpiresAfter.update_headerscCsd}||jS)Nz:110 - Automatically cached for %s. Response might be stale)r)rrZtmplr
r
rrSszExpiresAfter.warningN)rrrr#r&rrr
r
r
rr$Gsr$c@s:eZdZdZdddddddd	d
ddhZd
dZddZdS)LastModifieda
    If there is no Expires header already, fall back on Last-Modified
    using the heuristic from
    http://tools.ietf.org/html/rfc7234#section-4.2.2
    to calculate a reasonable value.

    Firefox also does something like this per
    https://developer.mozilla.org/en-US/docs/Web/HTTP/Caching_FAQ
    http://lxr.mozilla.org/mozilla-release/source/netwerk/protocol/http/nsHttpResponseHead.cpp#397
    Unlike mozilla we limit this to 24-hr.
    i,i-iiiiic
Cs|j}d|kriSd|kr*|ddkr*iS|j|jkr:iSd|ksJd|krNiStt|d}t|d}|dks||dkriSt}td||}|t|}tdt	|dd}||kriS||}	dt
tt|	iS)	Nrr"r!r	z
last-modifiedr
iQ)
rZstatuscacheable_by_default_statusesr
rrrtimemaxminstrftimeTIME_FMTgmtime)
rresprr	Z
last_modifiedZnowZcurrent_agerZfreshness_lifetimerr
r
rrhs*zLastModified.update_headerscCsdSrr
)rr4r
r
rrszLastModified.warningN)rrrr#r-rrr
r
r
rr'Xsr')N)r
r.Zemail.utilsrrrrrr2rrobjectrrr$r'r
r
r
rs
"site-packages/pip/_vendor/cachecontrol/__pycache__/filewrapper.cpython-38.opt-1.pyc000064400000004133151464452710024305 0ustar00U

.e	@s ddlmZGdddeZdS))BytesIOc@sBeZdZdZddZddZddZdd	ZdddZd
dZ	d
S)CallbackFileWrapperav
    Small wrapper around a fp object which will tee everything read into a
    buffer, and when that file is closed it will execute a callback with the
    contents of that buffer.

    All attributes are proxied to the underlying file object.

    This class uses members with a double underscore (__) leading prefix so as
    not to accidentally shadow an attribute.
    cCst|_||_||_dSN)r_CallbackFileWrapper__buf_CallbackFileWrapper__fp_CallbackFileWrapper__callback)selffpcallbackrH/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/filewrapper.py__init__szCallbackFileWrapper.__init__cCs|d}t||S)Nr)__getattribute__getattr)rnamer	rrr__getattr__s	
zCallbackFileWrapper.__getattr__cCsHz|jjdkWStk
r"YnXz
|jjWStk
rBYnXdS)NF)rr	AttributeErrorclosedrrrrZ__is_fp_closed!s
z"CallbackFileWrapper.__is_fp_closedcCs |jr||jd|_dSr)rrgetvaluerrrr_close2szCallbackFileWrapper._closeNcCs,|j|}|j||r(||Sr)rreadrwrite"_CallbackFileWrapper__is_fp_closedrrZamtdatarrrr=s
zCallbackFileWrapper.readcCs@|j|}|dkr |dkr |S|j||r<||S)Ns
)r
_safe_readrrrrrrrrrEszCallbackFileWrapper._safe_read)N)
__name__
__module____qualname____doc__r
rrrrrrrrrrs
rN)iorobjectrrrrrssite-packages/pip/_vendor/cachecontrol/__pycache__/wrapper.cpython-38.pyc000064400000001165151464452710022510 0ustar00U

.e@s&ddlmZddlmZdddZdS))CacheControlAdapter)	DictCacheNTc	CsB|pt}|pt}|||||||d}|d||d||S)N)cache_etags
serializer	heuristiccontroller_classcacheable_methodszhttp://zhttps://)rrZmount)	ZsesscacherrrrZ
adapter_classradapterrD/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/wrapper.pyCacheControls
r
)NTNNNNN)r
rr	rr
rrrrssite-packages/pip/_vendor/cachecontrol/__pycache__/controller.cpython-38.opt-1.pyc000064400000016754151464452710024164 0ustar00U

.e5@sdZddlZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZe
eZedZd	d
ZGdddeZdS)
z7
The httplib2 algorithms ported for use with requests.
N)parsedate_tz)CaseInsensitiveDict)	DictCache)
Serializerz9^(([^:/?#]+):)?(//([^/?#]*))?([^?#]*)(\?([^#]*))?(#(.*))?cCs0t|}|d|d|d|d|dfS)zParses a URI using the regex given in Appendix B of RFC 3986.

        (scheme, authority, path, query, fragment) = parse_uri(uri)
    r)URImatchgroups)urir
rG/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/controller.py	parse_urisrc@s\eZdZdZdddZeddZedd	Zd
dZdd
Z	ddZ
dddZddZdS)CacheControllerz9An interface to see if request should cached or not.
    NTcCs,|pt|_||_|pt|_|p$d|_dS)N)i,-)rcachecache_etagsr
serializercacheable_status_codes)selfrrrstatus_codesrrr__init__"szCacheController.__init__c	Csht|\}}}}}|r|s&td||}|}|s>d}|rPd||gpR|}|d||}|S)z4Normalize the URL to create a safe key for the cachez(Only absolute URIs are allowed. uri = %s/?z://)r	Exceptionlowerjoin)	clsrZschemeZ	authoritypathZqueryZfragmentZrequest_uriZ
defrag_urirrr_urlnorm*szCacheController._urlnormcCs
||SN)r$)r"rrrr	cache_url>szCacheController.cache_urlc

Cs(tdftdftdfddddddddtdfd}|d|dd}i}|dD]}|s^qP|d	d
}|d}z||\}}	Wn$tk
rtd|YqPYnX|r|	sd||<|rPz||d
||<WqPtk
r|	rtd
|YqPtk
r td||j	YqPXqP|S)NTF)NF)max-agez	max-stale	min-freshno-cacheno-storezno-transformzonly-if-cachedzmust-revalidateZpublicZprivatezproxy-revalidatezs-maxagez
cache-controlz
Cache-Control,=rrz,Ignoring unknown cache-control directive: %sz-Missing value for cache-control directive: %sz8Invalid value for cache-control directive %s, must be %s)
intgetsplitstripKeyErrorloggerdebug
IndexError
ValueError__name__)
rheadersZknown_directivesZ
cc_headersZretvalZcc_directivepartsZ	directivetypZrequiredrrrparse_cache_controlBsV
z#CacheController.parse_cache_controlcCs0||j}td|||j}d|kr:tddSd|kr\|ddkr\tddS|j|}|dkr~td	dS|j	||}|st
d
dS|jdkrd}t||St|j}|rd
|krd|krtd|j
|tddSt}tt|d
}	td||	}
td|
||}d}d|kr^|d}td|nDd|krt|d}
|
dk	rt|
|	}td|}td|d|kr|d}td|d|kr|d}|
|7}
td|
||
krtdtd||
|Sd|kr,td|j
|dS)ze
        Return a cached response if it exists in the cache, otherwise
        return False.
        zLooking up "%s" in the cacher)z-Request header has "no-cache", cache bypassedFr'rz1Request header has "max_age" as 0, cache bypassedNzNo cache entry availablez1Cache entry deserialization failed, entry ignoredrzVReturning cached "301 Moved Permanently" response (ignoring date and etag information)dateetagz(Purging cached response: no date or etagz!Ignoring cached response: no datezCurrent age based on date: %iz#Freshness lifetime from max-age: %iexpiresz#Freshness lifetime from expires: %iz+Freshness lifetime from request max-age: %ir(z'Adjusted current age from min-fresh: %iz2The response is "fresh", returning cached responsez%i > %iz4The cached response is "stale" with no etag, purging)r&urlr3r4r;r8rr/rloadsZwarningstatusrdeletetimecalendarZtimegmrmax)rrequestr&ccZ
cache_datarespmsgr8Znowr<Zcurrent_ageZresp_ccZfreshness_lifetimer>Zexpire_timeZ	min_freshrrrcached_requestxs|




















zCacheController.cached_requestcCs`||j}|j||j|}i}|r\t|j}d|krH|d|d<d|kr\|d|d<|S)Nr=ZETagz
If-None-Matchz
last-modifiedz
Last-ModifiedzIf-Modified-Since)r&r?rr@rr/rr8)rrFr&rHZnew_headersr8rrrconditional_headerss
z#CacheController.conditional_headerscCs|p|j}|j|kr(td|j|dSt|j}|dk	rfd|krf|drft|dt|krfdS|	|j}|	|}|
|j}	td|	d}
d|krd}
tdd|krd}
td	|
r|j
|	rtd
|j|	|
rdS|jr,d|kr,td|j|	|jj|||d
n|jdkr\td|j|	|j||nd|krd|kr|ddkrtd|j|	|jj|||d
n:d|kr|drtd|j|	|jj|||d
dS)zc
        Algorithm for caching requests.

        This assumes a requests Response object.
        zStatus code %s not in %sNcontent-lengthz&Updating cache with response from "%s"Fr*TzResponse header has "no-store"zRequest header has "no-store"z0Purging existing cache entry to honor "no-store"r=zCaching due to etag)bodyrzCaching permanant redirectr<r'rz'Caching b/c date exists and max-age > 0r>zCaching b/c of expires header)rrAr3r4rr8isdigitr.lenr;r&r?rr/rBrsetrdumps)rrFresponserMrrZresponse_headersZcc_reqrGr&Zno_storerrrcache_responsesr














zCacheController.cache_responsecsv||j}|j||j|}|s*|Sdg|jtfdd|j	Dd|_
|j||j|||S)zOn a 304 we will get a new set of headers that we want to
        update our cached value with, assuming we have one.

        This should only ever be called when we've sent an ETag and
        gotten a 304 as the response.
        rLc3s&|]\}}|kr||fVqdSr%)r ).0kvZexcluded_headersrr	bsz9CacheController.update_cached_response..r)
r&r?rr@rr/r8updatedictitemsrArPrQ)rrFrRr&Zcached_responserrWrupdate_cached_responseIs		z&CacheController.update_cached_response)NTNN)NN)
r7
__module____qualname____doc__rclassmethodr$r&r;rJrKrSr\rrrrrs


6o
Rr)r_ZloggingrerDrCZemail.utilsrZpip._vendor.requests.structuresrrrZ	serializerZ	getLoggerr7r3compilerrobjectrrrrrs

	site-packages/pip/_vendor/cachecontrol/__pycache__/cache.cpython-38.pyc000064400000003322151464452710022070 0ustar00U

.e%@s4dZddlmZGdddeZGdddeZdS)zb
The cache object API for implementing caches. The default is a thread
safe in-memory dictionary.
)Lockc@s,eZdZddZddZddZddZd	S)
	BaseCachecCs
tdSNNotImplementedErrorselfkeyr
B/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/cache.pyget
sz
BaseCache.getcCs
tdSrrrr	valuer
r
rset
sz
BaseCache.setcCs
tdSrrrr
r
rdeleteszBaseCache.deletecCsdSrr
)rr
r
rcloseszBaseCache.closeN)__name__
__module____qualname__rrrrr
r
r
rrsrc@s.eZdZd
ddZddZddZdd	ZdS)	DictCacheNcCst|_|pi|_dSr)rlockdata)rZ	init_dictr
r
r__init__szDictCache.__init__cCs|j|dSr)rrrr
r
rrsz
DictCache.getc	Cs&|j|j||iW5QRXdSr)rrupdater
r
r
rr sz
DictCache.setc	Cs,|j||jkr|j|W5QRXdSr)rrpoprr
r
rr$s
zDictCache.delete)N)rrrrrrrr
r
r
rrs
rN)__doc__Z	threadingrobjectrrr
r
r
rssite-packages/pip/_vendor/cachecontrol/__pycache__/wrapper.cpython-38.opt-1.pyc000064400000001165151464452710023447 0ustar00U

.e@s&ddlmZddlmZdddZdS))CacheControlAdapter)	DictCacheNTc	CsB|pt}|pt}|||||||d}|d||d||S)N)cache_etags
serializer	heuristiccontroller_classcacheable_methodszhttp://zhttps://)rrZmount)	ZsesscacherrrrZ
adapter_classradapterrD/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/wrapper.pyCacheControls
r
)NTNNNNN)r
rr	rr
rrrrssite-packages/pip/_vendor/cachecontrol/__pycache__/__init__.cpython-38.opt-1.pyc000064400000001001151464452710023513 0ustar00U

.e.@s8dZdZdZdZddlmZddlmZddlm	Z	dS)	zbCacheControl import Interface.

Make it easy to import from cachecontrol without long namespaces.
zEric Larsonzeric@ionrock.orgz0.12.5)CacheControl)CacheControlAdapter)CacheControllerN)
__doc__
__author__Z	__email____version__wrapperrZadapterrZ
controllerrr	r	E/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/__init__.pyssite-packages/pip/_vendor/cachecontrol/__pycache__/compat.cpython-38.opt-1.pyc000064400000001320151464452710023243 0ustar00U

.e@szddlmZWn ek
r0ddlmZYnXzddlZWnek
rZddlZYnXzddlmZWn ek
rddlmZYnXzddl	m
Z
Wn ek
rddlm
Z
YnXzeZ
Wnek
reZ
YnXdS))urljoinN)HTTPResponse)is_fp_closed)Zurllib.parserImportErrorZurlparseZcPicklepickleZ.pip._vendor.requests.packages.urllib3.responserZpip._vendor.urllib3.responseZ*pip._vendor.requests.packages.urllib3.utilrZpip._vendor.urllib3.utilZunicodeZ	text_type	NameErrorstrr	r	C/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/compat.pys&site-packages/pip/_vendor/cachecontrol/__pycache__/serialize.cpython-38.opt-1.pyc000064400000010163151464452710023754 0ustar00U

.e*@spddlZddlZddlZddlZddlmZddlmZddlm	Z	m
Z
mZddZdd	Z
Gd
ddeZdS)N)msgpack)CaseInsensitiveDict)HTTPResponsepickle	text_typecCst|dS)Nascii)base64Z	b64decodeencode)brF/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/serialize.py_b64_decode_bytessrcCst|dS)Nutf8)rdecode)srrr
_b64_decode_strsrc@sNeZdZdddZddZddZdd	Zd
dZdd
ZddZ	ddZ
dS)
SerializerNc		Cst|j}|dkr*|jdd}t||_d|tdd|jD|j|j	t
|j|j|j
di}i|d<d|kr|dd}|D]:}t
|}|j|d}|dk	rt
|}||d|<qd	d
tj|ddgS)
NF)decode_contentresponsecss"|]\}}t|t|fVqdSN)r.0kvrrr
	2sz#Serializer.dumps..)bodyheadersstatusversionreasonstrictrvary,,scc=4T)Zuse_bin_type)rrreadioBytesIOZ_fpdictitemsrrrr r!rsplitstripgetjoinrdumps)	selfrequestrrZresponse_headersdataZvaried_headersheaderZheader_valuerrr
r.s4

zSerializer.dumpscCs|sdSz|dd\}}Wntk
r4d}YnX|dddkrR||}d}|dddd}zt|d	|||WStk
rYdSXdS)
Nr$rscc=0scc==rz
_loads_v{})r*
ValueErrorrgetattrformatAttributeError)r/r0r1Zverrrr
loadsJs
zSerializer.loadscCsd|dikrdS|diD] \}}|j|d|kr$dSq$|dd}t|ddd}|dd	d
kr|d||dd<zt|}Wn$tk
rt|d}YnXt	f|dd
|dS)z`Verify our vary headers match and construct a real urllib3
        HTTPResponse object.
        *r"Nrrr)r1ztransfer-encodingZchunkedrF)rZpreload_content)
r,r)rpoprr&r'	TypeErrorr
r)r/r0cachedr2valueZbody_rawrrrrr
prepare_responsehs
zSerializer.prepare_responsecCsdSrrr/r0r1rrr
	_loads_v0szSerializer._loads_v0cCs2zt|}Wntk
r$YdSX|||Sr)rr:r6rAr/r0r1r?rrr
	_loads_v1s
zSerializer._loads_v1c	Csztt|d}Wnttjfk
r6YdSXt|dd|dd<tdd|dd	D|dd<t
|dd|dd<tdd|d		D|d	<|||S)
Nrrrcss"|]\}}t|t|fVqdSrrrrrr
rsz'Serializer._loads_v2..rr css.|]&\}}t||dk	r t|n|fVqdSrrFrrrr
rsr")jsonr:zlib
decompressrr6errorrr(r)rrArDrrr
	_loads_v2s
zSerializer._loads_v2cCsdSrrrBrrr
	_loads_v3szSerializer._loads_v3cCs6ztj|dd}Wntk
r(YdSX|||S)Nzutf-8)encoding)rr:r6rArDrrr
	_loads_v4s
zSerializer._loads_v4)N)__name__
__module____qualname__r.r:rArCrErKrLrNrrrr
rs
4$r)r	r&rGrHZpip._vendorrZpip._vendor.requests.structuresrcompatrrrrrobjectrrrrr
ssite-packages/pip/_vendor/cachecontrol/__pycache__/adapter.cpython-38.opt-1.pyc000064400000005737151464452710023420 0ustar00U

.e@s\ddlZddlZddlZddlmZddlmZddlmZddl	m
Z
GdddeZdS)	N)HTTPAdapter)CacheController)	DictCache)CallbackFileWrappercsNeZdZddhZdfdd	Zdfdd	Zdfd
d	Zfdd
ZZS)CacheControlAdapterZPUTZDELETENTc
sLtt|j|||pt|_||_|p*d|_|p4t}	|	|j||d|_dS)N)ZGET)cache_etags
serializer)	superr__init__rcache	heuristiccacheable_methodsr
controller)
selfrrZcontroller_classr	r
rargskwZcontroller_factory	__class__D/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/adapter.pyrs
zCacheControlAdapter.__init__cs|p|j}|j|krhz|j|}Wntjk
r>d}YnX|rT|j||ddS|j|j	|t
t|j|f|}|S)z
        Send a request. Use the request information to see if it
        exists in the cache and cache the response if we need to and can.
        NT)
from_cache)
rmethodrZcached_requestzliberrorbuild_responseZheadersupdateZconditional_headersr
rsend)rrequestrr	cacheablecached_responseresprrrr$s


zCacheControlAdapter.sendFc
s|p|j}|s|j|kr|jr*|j|}|jdkrh|j||}||k	rNd}|jdd||}n\|jdkr|j	||nBt
|jt
|jj	|||_|jr|jfdd}t|||_tt|||}|j|jkr|jr|j|j}	|j|	||_|S)z
        Build a response by making a request or using the cache.

        This will end up calling send and returning a potentially
        cached response
        i0TF)Zdecode_contenti-cs|jdkr|jdS)Nr)Z
chunk_left_fpZ_closerZsuper_update_chunk_lengthrr_update_chunk_lengthns
z@CacheControlAdapter.build_response.._update_chunk_length)rrr
ZapplyZstatusrZupdate_cached_responsereadZrelease_connZcache_responserr"	functoolspartialZchunkedr%types
MethodTyper
rrinvalidating_methodsok	cache_urlZurlrdeleter)
rrZresponserrrr r%r!r-rr$rr9sJ	


z"CacheControlAdapter.build_responsecs|jtt|dS)N)rcloser
rr#rrrr/s
zCacheControlAdapter.close)NTNNNN)N)FN)	__name__
__module____qualname__r+rrrr/
__classcell__rrrrrsJr)r)r'rZpip._vendor.requests.adaptersrrrrrZfilewrapperrrrrrrssite-packages/pip/_vendor/cachecontrol/__pycache__/_cmd.cpython-38.opt-1.pyc000064400000003004151464452710022663 0ustar00U

.e@sxddlZddlmZddlmZddlmZddlmZddl	m
Z
ddZd	d
ZddZ
dd
dZedkrtedS)N)requests)CacheControlAdapter)	DictCache)logger)ArgumentParsercCs"ttjt}t|dS)N)rZsetLevelloggingDEBUGZ
StreamHandlerZ
addHandler)Zhandlerr	A/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/_cmd.py
setup_loggingsrcCs>ttdddd}t}|d||d||j|_|S)NT)Zcache_etagsZ
serializerZ	heuristiczhttp://zhttps://)rrrZSessionZmountZ
controllercache_controller)Zadaptersessr	r	r
get_sessionsrcCst}|jddd|S)NurlzThe URL to try and cache)help)radd_argument
parse_args)parserr	r	r
get_argssrcCsTt}t}||j}t|j|j|j|j	|jrHt
dnt
ddS)NzCached!z
Not cached :()rrgetrrrZcache_responseZrequestrawZcached_requestprint)argsr
Zrespr	r	r
main$s
r__main__)N)rZpip._vendorrZ pip._vendor.cachecontrol.adapterrZpip._vendor.cachecontrol.cacherZ#pip._vendor.cachecontrol.controllerrargparserrrrr__name__r	r	r	r
s
site-packages/pip/_vendor/cachecontrol/__pycache__/adapter.cpython-38.pyc000064400000005737151464452710022461 0ustar00U

.e@s\ddlZddlZddlZddlmZddlmZddlmZddl	m
Z
GdddeZdS)	N)HTTPAdapter)CacheController)	DictCache)CallbackFileWrappercsNeZdZddhZdfdd	Zdfdd	Zdfd
d	Zfdd
ZZS)CacheControlAdapterZPUTZDELETENTc
sLtt|j|||pt|_||_|p*d|_|p4t}	|	|j||d|_dS)N)ZGET)cache_etags
serializer)	superr__init__rcache	heuristiccacheable_methodsr
controller)
selfrrZcontroller_classr	r
rargskwZcontroller_factory	__class__D/usr/lib/python3.8/site-packages/pip/_vendor/cachecontrol/adapter.pyrs
zCacheControlAdapter.__init__cs|p|j}|j|krhz|j|}Wntjk
r>d}YnX|rT|j||ddS|j|j	|t
t|j|f|}|S)z
        Send a request. Use the request information to see if it
        exists in the cache and cache the response if we need to and can.
        NT)
from_cache)
rmethodrZcached_requestzliberrorbuild_responseZheadersupdateZconditional_headersr
rsend)rrequestrr	cacheablecached_responseresprrrr$s


zCacheControlAdapter.sendFc
s|p|j}|s|j|kr|jr*|j|}|jdkrh|j||}||k	rNd}|jdd||}n\|jdkr|j	||nBt
|jt
|jj	|||_|jr|jfdd}t|||_tt|||}|j|jkr|jr|j|j}	|j|	||_|S)z
        Build a response by making a request or using the cache.

        This will end up calling send and returning a potentially
        cached response
        i0TF)Zdecode_contenti-cs|jdkr|jdS)Nr)Z
chunk_left_fpZ_closerZsuper_update_chunk_lengthrr_update_chunk_lengthns
z@CacheControlAdapter.build_response.._update_chunk_length)rrr
ZapplyZstatusrZupdate_cached_responsereadZrelease_connZcache_responserr"	functoolspartialZchunkedr%types
MethodTyper
rrinvalidating_methodsok	cache_urlZurlrdeleter)
rrZresponserrrr r%r!r-rr$rr9sJ	


z"CacheControlAdapter.build_responsecs|jtt|dS)N)rcloser
rr#rrrr/s
zCacheControlAdapter.close)NTNNNN)N)FN)	__name__
__module____qualname__r+rrrr/
__classcell__rrrrrsJr)r)r'rZpip._vendor.requests.adaptersrrrrrZfilewrapperrrrrrrssite-packages/pip/_vendor/cachecontrol/wrapper.py000064400000001237151464452710016222 0ustar00from .adapter import CacheControlAdapter
from .cache import DictCache


def CacheControl(
    sess,
    cache=None,
    cache_etags=True,
    serializer=None,
    heuristic=None,
    controller_class=None,
    adapter_class=None,
    cacheable_methods=None,
):

    cache = cache or DictCache()
    adapter_class = adapter_class or CacheControlAdapter
    adapter = adapter_class(
        cache,
        cache_etags=cache_etags,
        serializer=serializer,
        heuristic=heuristic,
        controller_class=controller_class,
        cacheable_methods=cacheable_methods,
    )
    sess.mount("http://", adapter)
    sess.mount("https://", adapter)

    return sess
site-packages/pip/_vendor/cachecontrol/adapter.py000064400000011377151464452710016170 0ustar00import types
import functools
import zlib

from pip._vendor.requests.adapters import HTTPAdapter

from .controller import CacheController
from .cache import DictCache
from .filewrapper import CallbackFileWrapper


class CacheControlAdapter(HTTPAdapter):
    invalidating_methods = {"PUT", "DELETE"}

    def __init__(
        self,
        cache=None,
        cache_etags=True,
        controller_class=None,
        serializer=None,
        heuristic=None,
        cacheable_methods=None,
        *args,
        **kw
    ):
        super(CacheControlAdapter, self).__init__(*args, **kw)
        self.cache = cache or DictCache()
        self.heuristic = heuristic
        self.cacheable_methods = cacheable_methods or ("GET",)

        controller_factory = controller_class or CacheController
        self.controller = controller_factory(
            self.cache, cache_etags=cache_etags, serializer=serializer
        )

    def send(self, request, cacheable_methods=None, **kw):
        """
        Send a request. Use the request information to see if it
        exists in the cache and cache the response if we need to and can.
        """
        cacheable = cacheable_methods or self.cacheable_methods
        if request.method in cacheable:
            try:
                cached_response = self.controller.cached_request(request)
            except zlib.error:
                cached_response = None
            if cached_response:
                return self.build_response(request, cached_response, from_cache=True)

            # check for etags and add headers if appropriate
            request.headers.update(self.controller.conditional_headers(request))

        resp = super(CacheControlAdapter, self).send(request, **kw)

        return resp

    def build_response(
        self, request, response, from_cache=False, cacheable_methods=None
    ):
        """
        Build a response by making a request or using the cache.

        This will end up calling send and returning a potentially
        cached response
        """
        cacheable = cacheable_methods or self.cacheable_methods
        if not from_cache and request.method in cacheable:
            # Check for any heuristics that might update headers
            # before trying to cache.
            if self.heuristic:
                response = self.heuristic.apply(response)

            # apply any expiration heuristics
            if response.status == 304:
                # We must have sent an ETag request. This could mean
                # that we've been expired already or that we simply
                # have an etag. In either case, we want to try and
                # update the cache if that is the case.
                cached_response = self.controller.update_cached_response(
                    request, response
                )

                if cached_response is not response:
                    from_cache = True

                # We are done with the server response, read a
                # possible response body (compliant servers will
                # not return one, but we cannot be 100% sure) and
                # release the connection back to the pool.
                response.read(decode_content=False)
                response.release_conn()

                response = cached_response

            # We always cache the 301 responses
            elif response.status == 301:
                self.controller.cache_response(request, response)
            else:
                # Wrap the response file with a wrapper that will cache the
                #   response when the stream has been consumed.
                response._fp = CallbackFileWrapper(
                    response._fp,
                    functools.partial(
                        self.controller.cache_response, request, response
                    ),
                )
                if response.chunked:
                    super_update_chunk_length = response._update_chunk_length

                    def _update_chunk_length(self):
                        super_update_chunk_length()
                        if self.chunk_left == 0:
                            self._fp._close()

                    response._update_chunk_length = types.MethodType(
                        _update_chunk_length, response
                    )

        resp = super(CacheControlAdapter, self).build_response(request, response)

        # See if we should invalidate the cache.
        if request.method in self.invalidating_methods and resp.ok:
            cache_url = self.controller.cache_url(request.url)
            self.cache.delete(cache_url)

        # Give the request a from_cache attr to let people use it
        resp.from_cache = from_cache

        return resp

    def close(self):
        self.cache.close()
        super(CacheControlAdapter, self).close()
site-packages/pip/_vendor/cachecontrol/filewrapper.py000064400000004745151464452710017071 0ustar00from io import BytesIO


class CallbackFileWrapper(object):
    """
    Small wrapper around a fp object which will tee everything read into a
    buffer, and when that file is closed it will execute a callback with the
    contents of that buffer.

    All attributes are proxied to the underlying file object.

    This class uses members with a double underscore (__) leading prefix so as
    not to accidentally shadow an attribute.
    """

    def __init__(self, fp, callback):
        self.__buf = BytesIO()
        self.__fp = fp
        self.__callback = callback

    def __getattr__(self, name):
        # The vaguaries of garbage collection means that self.__fp is
        # not always set.  By using __getattribute__ and the private
        # name[0] allows looking up the attribute value and raising an
        # AttributeError when it doesn't exist. This stop thigns from
        # infinitely recursing calls to getattr in the case where
        # self.__fp hasn't been set.
        #
        # [0] https://docs.python.org/2/reference/expressions.html#atom-identifiers
        fp = self.__getattribute__("_CallbackFileWrapper__fp")
        return getattr(fp, name)

    def __is_fp_closed(self):
        try:
            return self.__fp.fp is None

        except AttributeError:
            pass

        try:
            return self.__fp.closed

        except AttributeError:
            pass

        # We just don't cache it then.
        # TODO: Add some logging here...
        return False

    def _close(self):
        if self.__callback:
            self.__callback(self.__buf.getvalue())

        # We assign this to None here, because otherwise we can get into
        # really tricky problems where the CPython interpreter dead locks
        # because the callback is holding a reference to something which
        # has a __del__ method. Setting this to None breaks the cycle
        # and allows the garbage collector to do it's thing normally.
        self.__callback = None

    def read(self, amt=None):
        data = self.__fp.read(amt)
        self.__buf.write(data)
        if self.__is_fp_closed():
            self._close()

        return data

    def _safe_read(self, amt):
        data = self.__fp._safe_read(amt)
        if amt == 2 and data == b"\r\n":
            # urllib executes this read to toss the CRLF at the end
            # of the chunk.
            return data

        self.__buf.write(data)
        if self.__is_fp_closed():
            self._close()

        return data
site-packages/pip/_vendor/cachecontrol/compat.py000064400000001267151464452710016030 0ustar00try:
    from urllib.parse import urljoin
except ImportError:
    from urlparse import urljoin


try:
    import cPickle as pickle
except ImportError:
    import pickle


# Handle the case where the requests module has been patched to not have
# urllib3 bundled as part of its source.
try:
    from pip._vendor.requests.packages.urllib3.response import HTTPResponse
except ImportError:
    from pip._vendor.urllib3.response import HTTPResponse

try:
    from pip._vendor.requests.packages.urllib3.util import is_fp_closed
except ImportError:
    from pip._vendor.urllib3.util import is_fp_closed

# Replicate some six behaviour
try:
    text_type = unicode
except NameError:
    text_type = str
site-packages/pip/_vendor/cachecontrol/__init__.py000064400000000456151464452710016303 0ustar00"""CacheControl import Interface.

Make it easy to import from cachecontrol without long namespaces.
"""
__author__ = "Eric Larson"
__email__ = "eric@ionrock.org"
__version__ = "0.12.5"

from .wrapper import CacheControl
from .adapter import CacheControlAdapter
from .controller import CacheController
site-packages/pip/_vendor/distlib/resources.py000064400000025016151464452710015543 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from __future__ import unicode_literals

import bisect
import io
import logging
import os
import pkgutil
import shutil
import sys
import types
import zipimport

from . import DistlibException
from .util import cached_property, get_cache_base, path_to_cache_dir, Cache

logger = logging.getLogger(__name__)


cache = None    # created when needed


class ResourceCache(Cache):
    def __init__(self, base=None):
        if base is None:
            # Use native string to avoid issues on 2.x: see Python #20140.
            base = os.path.join(get_cache_base(), str('resource-cache'))
        super(ResourceCache, self).__init__(base)

    def is_stale(self, resource, path):
        """
        Is the cache stale for the given resource?

        :param resource: The :class:`Resource` being cached.
        :param path: The path of the resource in the cache.
        :return: True if the cache is stale.
        """
        # Cache invalidation is a hard problem :-)
        return True

    def get(self, resource):
        """
        Get a resource into the cache,

        :param resource: A :class:`Resource` instance.
        :return: The pathname of the resource in the cache.
        """
        prefix, path = resource.finder.get_cache_info(resource)
        if prefix is None:
            result = path
        else:
            result = os.path.join(self.base, self.prefix_to_dir(prefix), path)
            dirname = os.path.dirname(result)
            if not os.path.isdir(dirname):
                os.makedirs(dirname)
            if not os.path.exists(result):
                stale = True
            else:
                stale = self.is_stale(resource, path)
            if stale:
                # write the bytes of the resource to the cache location
                with open(result, 'wb') as f:
                    f.write(resource.bytes)
        return result


class ResourceBase(object):
    def __init__(self, finder, name):
        self.finder = finder
        self.name = name


class Resource(ResourceBase):
    """
    A class representing an in-package resource, such as a data file. This is
    not normally instantiated by user code, but rather by a
    :class:`ResourceFinder` which manages the resource.
    """
    is_container = False        # Backwards compatibility

    def as_stream(self):
        """
        Get the resource as a stream.

        This is not a property to make it obvious that it returns a new stream
        each time.
        """
        return self.finder.get_stream(self)

    @cached_property
    def file_path(self):
        global cache
        if cache is None:
            cache = ResourceCache()
        return cache.get(self)

    @cached_property
    def bytes(self):
        return self.finder.get_bytes(self)

    @cached_property
    def size(self):
        return self.finder.get_size(self)


class ResourceContainer(ResourceBase):
    is_container = True     # Backwards compatibility

    @cached_property
    def resources(self):
        return self.finder.get_resources(self)


class ResourceFinder(object):
    """
    Resource finder for file system resources.
    """

    if sys.platform.startswith('java'):
        skipped_extensions = ('.pyc', '.pyo', '.class')
    else:
        skipped_extensions = ('.pyc', '.pyo')

    def __init__(self, module):
        self.module = module
        self.loader = getattr(module, '__loader__', None)
        self.base = os.path.dirname(getattr(module, '__file__', ''))

    def _adjust_path(self, path):
        return os.path.realpath(path)

    def _make_path(self, resource_name):
        # Issue #50: need to preserve type of path on Python 2.x
        # like os.path._get_sep
        if isinstance(resource_name, bytes):    # should only happen on 2.x
            sep = b'/'
        else:
            sep = '/'
        parts = resource_name.split(sep)
        parts.insert(0, self.base)
        result = os.path.join(*parts)
        return self._adjust_path(result)

    def _find(self, path):
        return os.path.exists(path)

    def get_cache_info(self, resource):
        return None, resource.path

    def find(self, resource_name):
        path = self._make_path(resource_name)
        if not self._find(path):
            result = None
        else:
            if self._is_directory(path):
                result = ResourceContainer(self, resource_name)
            else:
                result = Resource(self, resource_name)
            result.path = path
        return result

    def get_stream(self, resource):
        return open(resource.path, 'rb')

    def get_bytes(self, resource):
        with open(resource.path, 'rb') as f:
            return f.read()

    def get_size(self, resource):
        return os.path.getsize(resource.path)

    def get_resources(self, resource):
        def allowed(f):
            return (f != '__pycache__' and not
                    f.endswith(self.skipped_extensions))
        return set([f for f in os.listdir(resource.path) if allowed(f)])

    def is_container(self, resource):
        return self._is_directory(resource.path)

    _is_directory = staticmethod(os.path.isdir)

    def iterator(self, resource_name):
        resource = self.find(resource_name)
        if resource is not None:
            todo = [resource]
            while todo:
                resource = todo.pop(0)
                yield resource
                if resource.is_container:
                    rname = resource.name
                    for name in resource.resources:
                        if not rname:
                            new_name = name
                        else:
                            new_name = '/'.join([rname, name])
                        child = self.find(new_name)
                        if child.is_container:
                            todo.append(child)
                        else:
                            yield child


class ZipResourceFinder(ResourceFinder):
    """
    Resource finder for resources in .zip files.
    """
    def __init__(self, module):
        super(ZipResourceFinder, self).__init__(module)
        archive = self.loader.archive
        self.prefix_len = 1 + len(archive)
        # PyPy doesn't have a _files attr on zipimporter, and you can't set one
        if hasattr(self.loader, '_files'):
            self._files = self.loader._files
        else:
            self._files = zipimport._zip_directory_cache[archive]
        self.index = sorted(self._files)

    def _adjust_path(self, path):
        return path

    def _find(self, path):
        path = path[self.prefix_len:]
        if path in self._files:
            result = True
        else:
            if path and path[-1] != os.sep:
                path = path + os.sep
            i = bisect.bisect(self.index, path)
            try:
                result = self.index[i].startswith(path)
            except IndexError:
                result = False
        if not result:
            logger.debug('_find failed: %r %r', path, self.loader.prefix)
        else:
            logger.debug('_find worked: %r %r', path, self.loader.prefix)
        return result

    def get_cache_info(self, resource):
        prefix = self.loader.archive
        path = resource.path[1 + len(prefix):]
        return prefix, path

    def get_bytes(self, resource):
        return self.loader.get_data(resource.path)

    def get_stream(self, resource):
        return io.BytesIO(self.get_bytes(resource))

    def get_size(self, resource):
        path = resource.path[self.prefix_len:]
        return self._files[path][3]

    def get_resources(self, resource):
        path = resource.path[self.prefix_len:]
        if path and path[-1] != os.sep:
            path += os.sep
        plen = len(path)
        result = set()
        i = bisect.bisect(self.index, path)
        while i < len(self.index):
            if not self.index[i].startswith(path):
                break
            s = self.index[i][plen:]
            result.add(s.split(os.sep, 1)[0])   # only immediate children
            i += 1
        return result

    def _is_directory(self, path):
        path = path[self.prefix_len:]
        if path and path[-1] != os.sep:
            path += os.sep
        i = bisect.bisect(self.index, path)
        try:
            result = self.index[i].startswith(path)
        except IndexError:
            result = False
        return result

_finder_registry = {
    type(None): ResourceFinder,
    zipimport.zipimporter: ZipResourceFinder
}

try:
    # In Python 3.6, _frozen_importlib -> _frozen_importlib_external
    try:
        import _frozen_importlib_external as _fi
    except ImportError:
        import _frozen_importlib as _fi
    _finder_registry[_fi.SourceFileLoader] = ResourceFinder
    _finder_registry[_fi.FileFinder] = ResourceFinder
    del _fi
except (ImportError, AttributeError):
    pass


def register_finder(loader, finder_maker):
    _finder_registry[type(loader)] = finder_maker

_finder_cache = {}


def finder(package):
    """
    Return a resource finder for a package.
    :param package: The name of the package.
    :return: A :class:`ResourceFinder` instance for the package.
    """
    if package in _finder_cache:
        result = _finder_cache[package]
    else:
        if package not in sys.modules:
            __import__(package)
        module = sys.modules[package]
        path = getattr(module, '__path__', None)
        if path is None:
            raise DistlibException('You cannot get a finder for a module, '
                                   'only for a package')
        loader = getattr(module, '__loader__', None)
        finder_maker = _finder_registry.get(type(loader))
        if finder_maker is None:
            raise DistlibException('Unable to locate finder for %r' % package)
        result = finder_maker(module)
        _finder_cache[package] = result
    return result


_dummy_module = types.ModuleType(str('__dummy__'))


def finder_for_path(path):
    """
    Return a resource finder for a path, which should represent a container.

    :param path: The path.
    :return: A :class:`ResourceFinder` instance for the path.
    """
    result = None
    # calls any path hooks, gets importer into cache
    pkgutil.get_importer(path)
    loader = sys.path_importer_cache.get(path)
    finder = _finder_registry.get(type(loader))
    if finder:
        module = _dummy_module
        module.__file__ = os.path.join(path, '')
        module.__loader__ = loader
        result = finder(module)
    return result
site-packages/pip/_vendor/distlib/_backport/misc.py000064400000001713151464452710016426 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Backports for individual classes and functions."""

import os
import sys

__all__ = ['cache_from_source', 'callable', 'fsencode']


try:
    from imp import cache_from_source
except ImportError:
    def cache_from_source(py_file, debug=__debug__):
        ext = debug and 'c' or 'o'
        return py_file + ext


try:
    callable = callable
except NameError:
    from collections import Callable

    def callable(obj):
        return isinstance(obj, Callable)


try:
    fsencode = os.fsencode
except AttributeError:
    def fsencode(filename):
        if isinstance(filename, bytes):
            return filename
        elif isinstance(filename, str):
            return filename.encode(sys.getfilesystemencoding())
        else:
            raise TypeError("expect bytes or str, not %s" %
                            type(filename).__name__)
site-packages/pip/_vendor/distlib/_backport/sysconfig.cfg000064400000005071151464452710017607 0ustar00[posix_prefix]
# Configuration directories.  Some of these come straight out of the
# configure script.  They are for implementing the other variables, not to
# be used directly in [resource_locations].
confdir = /etc
datadir = /usr/share
libdir = /usr/lib
statedir = /var
# User resource directory
local = ~/.local/{distribution.name}

stdlib = {base}/lib/python{py_version_short}
platstdlib = {platbase}/lib/python{py_version_short}
purelib = {base}/lib/python{py_version_short}/site-packages
platlib = {platbase}/lib/python{py_version_short}/site-packages
include = {base}/include/python{py_version_short}{abiflags}
platinclude = {platbase}/include/python{py_version_short}{abiflags}
data = {base}

[posix_home]
stdlib = {base}/lib/python
platstdlib = {base}/lib/python
purelib = {base}/lib/python
platlib = {base}/lib/python
include = {base}/include/python
platinclude = {base}/include/python
scripts = {base}/bin
data = {base}

[nt]
stdlib = {base}/Lib
platstdlib = {base}/Lib
purelib = {base}/Lib/site-packages
platlib = {base}/Lib/site-packages
include = {base}/Include
platinclude = {base}/Include
scripts = {base}/Scripts
data = {base}

[os2]
stdlib = {base}/Lib
platstdlib = {base}/Lib
purelib = {base}/Lib/site-packages
platlib = {base}/Lib/site-packages
include = {base}/Include
platinclude = {base}/Include
scripts = {base}/Scripts
data = {base}

[os2_home]
stdlib = {userbase}/lib/python{py_version_short}
platstdlib = {userbase}/lib/python{py_version_short}
purelib = {userbase}/lib/python{py_version_short}/site-packages
platlib = {userbase}/lib/python{py_version_short}/site-packages
include = {userbase}/include/python{py_version_short}
scripts = {userbase}/bin
data = {userbase}

[nt_user]
stdlib = {userbase}/Python{py_version_nodot}
platstdlib = {userbase}/Python{py_version_nodot}
purelib = {userbase}/Python{py_version_nodot}/site-packages
platlib = {userbase}/Python{py_version_nodot}/site-packages
include = {userbase}/Python{py_version_nodot}/Include
scripts = {userbase}/Scripts
data = {userbase}

[posix_user]
stdlib = {userbase}/lib/python{py_version_short}
platstdlib = {userbase}/lib/python{py_version_short}
purelib = {userbase}/lib/python{py_version_short}/site-packages
platlib = {userbase}/lib/python{py_version_short}/site-packages
include = {userbase}/include/python{py_version_short}
scripts = {userbase}/bin
data = {userbase}

[osx_framework_user]
stdlib = {userbase}/lib/python
platstdlib = {userbase}/lib/python
purelib = {userbase}/lib/python/site-packages
platlib = {userbase}/lib/python/site-packages
include = {userbase}/include
scripts = {userbase}/bin
data = {userbase}
site-packages/pip/_vendor/distlib/_backport/__pycache__/__init__.cpython-38.pyc000064400000000677151464452710023530 0ustar00U

.e@sdZdS)aModules copied from Python 3 standard libraries, for internal use only.

Individual classes and functions are found in d2._backport.misc.  Intended
usage is to always import things missing from 3.1 from that module: the
built-in/stdlib objects will be used if found.
N)__doc__rrJ/usr/lib/python3.8/site-packages/pip/_vendor/distlib/_backport/__init__.pysite-packages/pip/_vendor/distlib/_backport/__pycache__/sysconfig.cpython-38.opt-1.pyc000064400000036666151464452710024743 0ustar00U

.eTi@sdZddlZddlZddlZddlZddlmZmZzddlZWne	k
r\ddl
ZYnXdddddd	d
ddd
dgZddZej
rejeej
ZneeZejdkrdeddkreejeeZejdkrdeddkreejeeeZejdkrBdeddkrBeejeeeZddZeZdaddZeZedZddZejdZ ejdd Z!e de d!Z"ej#ej$Z%ej#ej&Z'da(dZ)d"d#Z*d$d%Z+d&d'Z,d(d)Z-d*d+Z.d,d-Z/dEd.d/Z0d0dZ1d1d2Z2d3d4Z3dFd5dZ4d6dZ5d7d
Z6d8d	Z7e.dd9fd:d
Z8e.dd9fd;dZ9ddZdBdCZ?e@dDkre?dS)Gz-Access to Python's configuration information.N)pardirrealpathget_config_h_filenameget_config_varget_config_varsget_makefile_filenameget_pathget_path_names	get_pathsget_platformget_python_versionget_scheme_namesparse_config_hcCs(z
t|WStk
r"|YSXdSN)rOSError)pathrK/usr/lib/python3.8/site-packages/pip/_vendor/distlib/_backport/sysconfig.py_safe_realpath"s
rntZpcbuildiz\pc\viz\pcbuild\amd64icCs,dD]"}tjtjtd|rdSqdS)N)z
Setup.distzSetup.localZModulesTF)osrisfilejoin
_PROJECT_BASE)fnrrris_python_build:srFc	Csts~ddlm}tddd}||}|d}|}t|W5QRXt	rzdD] }t
|dd	t
|d
dqXdadS)
N)finder.rz
sysconfig.cfg)posix_prefixZ
posix_homeincludez{srcdir}/Includeplatincludez{projectbase}/.T)	_cfg_readZ	resourcesr__name__rsplitfindZ	as_stream_SCHEMESZreadfp
_PYTHON_BUILDset)rZbackport_packageZ_finderZ_cfgfilesschemerrr_ensure_cfg_readDs

r,z\{([^{]*?)\}c
st|dr|d}nt}|}|D]8}|dkr._replacer)r,Zhas_sectionitemstuplesectionsZ
has_optionr)Zremove_sectiondict	_VAR_REPLsub)configr-r8ZsectionZoptionvaluer5rr3r_expand_globalsYs$

r>rcsfdd}t||S)zIn the string `path`, replace tokens like {some.thing} with the
    corresponding value from the map `local_vars`.

    If there is no corresponding value, leave the token unchanged.
    cs8|d}|kr|S|tjkr.tj|S|dSr.)r0renvironr1
local_varsrrr5s


z_subst_vars.._replacerr:r;)rrBr5rrAr_subst_varssrDcCs0|}|D]\}}||kr"q|||<qdSr)keysr6)Ztarget_dictZ
other_dictZtarget_keyskeyr=rrr_extend_dicts
rGcCs`i}|dkri}t|tt|D]4\}}tjdkrDtj|}tjt	||||<q&|S)N)posixr)
rGrr'r6rr2r
expandusernormpathrD)r+varsresrFr=rrr_expand_varss
rMcsfdd}t||S)Ncs$|d}|kr|S|dSr.r/r1rKrrr5s
zformat_value.._replacerrC)r=rKr5rrNrformat_valuesrOcCstjdkrdStjS)NrHr )rr2rrrr_get_default_schemes
rPcCstjdd}dd}tjdkrBtjdp.d}|r8|S||dStjdkr|td	}|r||r`|S|dd
|dtjddS|r|S|dd
SdS)NPYTHONUSERBASEcWstjtjj|Sr)rrrIr)argsrrrjoinusersz_getuserbase..joinuserrAPPDATA~PythondarwinZPYTHONFRAMEWORKLibraryz%d.%drz.local)rr@getr2sysplatformrversion_info)env_baserSbaseZ	frameworkrrr_getuserbases$



r_c	Cstd}td}td}|dkr*i}i}i}tj|ddd}|}W5QRX|D]}	|	dsZ|	d	krvqZ||	}
|
rZ|
d
d\}}|}|	dd	}
d
|
kr|||<qZzt
|}Wn$tk
r|	dd
||<YqZX|||<qZt|
}d}t|dkrt|D]}||}||p>||}
|
dk	r|
d
}d}||krpt||}n||krd}nx|tjkrtj|}n`||kr|dr|dd|krd	}n$d||krd}nt|d|}nd	||<}|r||
d}|d|
||}d
|kr:|||<n~zt
|}Wn"tk
rh|||<Yn
X|||<|||dr|dd|kr|dd}||kr|||<n|||<||qq|D]"\}}t|tr|||<q|||S)zParse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    z"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)z\$\(([A-Za-z][A-Za-z0-9_]*)\)z\${([A-Za-z][A-Za-z0-9_]*)}Nzutf-8surrogateescape)encodingerrors#rrz$$$)CFLAGSLDFLAGSZCPPFLAGSrTFZPY_r?)recompilecodecsopen	readlines
startswithstripmatchr0replaceint
ValueErrorlistrElenr7searchstrrr@endstartremover6
isinstanceupdate)filenamerKZ_variable_rxZ_findvar1_rxZ_findvar2_rxZdoneZnotdoneflineslinemnvZtmpvr4Zrenamed_variablesr2r=founditemZafterkrrr_parse_makefiles	
















rcCsDtrtjtdSttdr,dttjf}nd}tjt	d|dS)z Return the path of the Makefile.ZMakefileabiflagszconfig-%s%sr<stdlib)
r(rrrrhasattrrZ_PY_VERSION_SHORTrr)Zconfig_dir_namerrrrMs
c
Cst}zt||WnJtk
r^}z,d|}t|drF|d|j}t|W5d}~XYnXt}z"t|}t||W5QRXWnJtk
r}z,d|}t|dr|d|j}t|W5d}~XYnXtr|d|d<dS)z7Initialize the module as appropriate for POSIX systems.z.invalid Python installation: unable to open %sstrerrorz (%s)N	BLDSHAREDZLDSHARED)	rrIOErrorrrrrkrr()rKZmakefileemsgZconfig_hr}rrr_init_posixXs&


rcCsVtd|d<td|d<td|d<d|d<d	|d
<t|d<tjttj|d<d
S)z+Initialize the module as appropriate for NTrZLIBDESTZ
platstdlibZ
BINLIBDESTr!Z	INCLUDEPYz.pydZSOz.exeZEXEZVERSIONZBINDIRN)r_PY_VERSION_SHORT_NO_DOTrrdirnamerrZ
executablerNrrr_init_non_posixtsrcCs|dkri}td}td}|}|s.q||}|rx|dd\}}zt|}Wntk
rlYnX|||<q ||}|r d||d<q |S)zParse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    Nz"#define ([A-Z][A-Za-z0-9_]+) (.*)
z&/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/
rrr)rhrireadlineror0rqrr)fprKZ	define_rxZundef_rxrrrrrrrrs&




cCs:tr$tjdkrtjtd}q,t}ntd}tj|dS)zReturn the path of pyconfig.h.rZPCr"z
pyconfig.h)r(rr2rrrr)Zinc_dirrrrrs
cCstttS)z,Return a tuple containing the schemes names.)r7sortedr'r8rrrrr
scCs
tdS)z*Return a tuple containing the paths names.r )r'Zoptionsrrrrr	sTcCs&t|rt||Stt|SdS)zReturn a mapping containing an install scheme.

    ``scheme`` is the install scheme name. If not provided, it will
    return the default scheme for the current platform.
    N)r,rMr9r'r6)r+rKexpandrrrr
s
cCst||||S)z[Return a path corresponding to the scheme.

    ``scheme`` is the install scheme name.
    )r
)r2r+rKrrrrrscGstdkrziattd<ttd<ttd<ttd<tdtdtd<ttd	<ttd
<ttd<ztjtd<Wntk
rd
td<YnXt	j
dkrttt	j
dkrtttj
dkrttd<dtkrttd<nttdtd<tr\t	j
dkr\t}zt	}Wntk
rd}YnXt	jtds\||kr\t	j|td}t	j|td<tjdkrzt	d}t|dd}|dkrdD]2}t|}tdd|}tdd|}|t|<qndt	jkrt	jd}dD]0}t|}tdd|}|d|}|t|<qtdd
}	td|	}
|
dk	rz|
 d}t	j!|szdD]$}t|}tdd|}|t|<qT|rg}|D]}
|"t|
q|StSdS)ayWith no arguments, return a dictionary of all configuration
    variables relevant for the current platform.

    On Unix, this means every variable defined in Python's installed Makefile;
    On Windows and Mac OS it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    Nprefixexec_prefixZ
py_versionZpy_version_shortrrZpy_version_nodotr^ZplatbaseZprojectbaserrd)rZos2rHz2.6userbasesrcdirrWr)rgZ
BASECFLAGSrfZ	PY_CFLAGSrz
-arch\s+\w+\s z-isysroot [^ 	]*Z	ARCHFLAGSrfz-isysroot\s+(\S+)rz-isysroot\s+\S+(\s|$))#_CONFIG_VARS_PREFIX_EXEC_PREFIX_PY_VERSIONrrrZrAttributeErrorrr2rrversionr_rr(getcwdrrisabsrrJr[unamerqsplitrhr;r@rYrur0existsappend)rRr^cwdrZkernel_versionZ
major_versionrFflagsZarchrfrZsdkZvalsr2rrrrs










cCst|S)zReturn the value of a single variable using the dictionary returned by
    'get_config_vars()'.

    Equivalent to get_config_vars().get(name)
    )rrY)r2rrrrRscCs^tjdkrnd}tj|}|dkr(tjStjd|}tj|t||}|dkr\dS|dkrhdStjStjd	ksttd
stjSt	\}}}}}|
dd}|
d
d}|
dd}|dddkrd||fS|dddkr&|ddkrPd}dt|dd|ddf}n*|dddkrDd||fS|dddkrdd|||fS|ddd krd }t
d!}	|	|}
|
rP|
}n|ddd"krPt}|d#}|}
ztd$}Wntk
rYnJXztd%|}
W5|X|
dk	r2d&|
d'd&dd}
|s<|
}|rP|}d(}|
d&d)krd*td+dkrd,}td+}td-|}ttt|}t|d'kr|d}n^|d.krd,}nN|d/krd0}n>|d1krd2}n.|d3krd4}n|d5krd6}ntd7|fn<|d8kr0tjd9krPd:}n |d;krPtjd9krLd<}nd=}d>|||fS)?aReturn a string that identifies the current platform.

    This is used mainly to distinguish platform-specific build directories and
    platform-specific built distributions.  Typically includes the OS name
    and version and the architecture (as supplied by 'os.uname()'),
    although the exact information included depends on the OS; eg. for IRIX
    the architecture isn't particularly important (IRIX only runs on SGI
    hardware), but for Linux the kernel version isn't particularly
    important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u
       irix-5.3
       irix64-6.2

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win-ia64 (64bit Windows on Itanium)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.
    rz bit ()Zamd64z	win-amd64Zitaniumzwin-ia64rHr/rdr_-NZlinuxz%s-%sZsunosr5Zsolarisz%d.%sr?rZirixZaixz%s-%s.%scygwinz[\d.]+rWZMACOSX_DEPLOYMENT_TARGETz0/System/Library/CoreServices/SystemVersion.plistz=ProductUserVisibleVersion\s*(.*?)rrZmacosxz10.4.z-archrfZfatz
-arch\s+(\S+))i386ppc)rx86_64Zintel)rrrZfat3)ppc64rZfat64)rrrrZ	universalz%Don't know machine value for archs=%rrlr)ZPowerPCZPower_Macintoshrrz%s-%s-%s) rr2rZrr&r[rtlowerrrrprqrhriror0rrYrkrcloserureadrrrnfindallr7rr)rrmaxsize)rijZlookZosnameZhostreleasermachineZrel_rerZcfgvarsZmacverZ
macreleaser}ZcflagsZarchsrrrr[s
$












cCstSr)rrrrrrscCsFtt|D]0\}\}}|dkr0td|td||fqdS)Nrz%s: z
	%s = "%s")	enumeraterr6print)titledataindexrFr=rrr_print_dictsrcCsRtdttdttdtttdtttdtdS)z*Display all information sysconfig detains.zPlatform: "%s"zPython version: "%s"z!Current installation scheme: "%s"ZPathsZ	VariablesN)rrrrPrr
rrrrr_mainsr__main__)N)N)A__doc__rjrrhrZZos.pathrrZconfigparserImportErrorZConfigParser__all__rrrrrrr2rrrr(r#r,ZRawConfigParserr'rir:r>rrrrrrJrrrrrZ
_USER_BASErDrGrMrOrPr_rrrrrrr
r	r
rrrrrrrr$rrrrs""
#
	
v

	#
site-packages/pip/_vendor/distlib/_backport/__pycache__/misc.cpython-38.pyc000064400000002042151464452710022710 0ustar00U

.e@sdZddlZddlZdddgZzddlmZWnek
rLdddZYnXzeZWn(ek
r~dd	l	m
Z
d
dZYnXz
ejZWnek
rddZYnXdS)
z/Backports for individual classes and functions.Ncache_from_sourcecallablefsencode)rTcCs|rdp
d}||S)Nco)Zpy_filedebugZextrrF/usr/lib/python3.8/site-packages/pip/_vendor/distlib/_backport/misc.pyrs)CallablecCs
t|tS)N)
isinstancer
)objrrr	rscCs<t|tr|St|tr&|tStdt|jdS)Nzexpect bytes or str, not %s)	rbytesstrencodesysgetfilesystemencoding	TypeErrortype__name__)filenamerrr	r"s

)T)
__doc__osr__all__ZimprImportErrorr	NameErrorcollectionsr
rAttributeErrorrrrr	s 

site-packages/pip/_vendor/distlib/_backport/__pycache__/sysconfig.cpython-38.pyc000064400000036752151464452710024000 0ustar00U

.eTi@sdZddlZddlZddlZddlZddlmZmZzddlZWne	k
r\ddl
ZYnXdddddd	d
ddd
dgZddZej
rejeej
ZneeZejdkrdeddkreejeeZejdkrdeddkreejeeeZejdkrBdeddkrBeejeeeZddZeZdaddZeZedZddZejdZ ejdd Z!e de d!Z"ej#ej$Z%ej#ej&Z'da(dZ)d"d#Z*d$d%Z+d&d'Z,d(d)Z-d*d+Z.d,d-Z/dEd.d/Z0d0dZ1d1d2Z2d3d4Z3dFd5dZ4d6dZ5d7d
Z6d8d	Z7e.dd9fd:d
Z8e.dd9fd;dZ9ddZdBdCZ?e@dDkre?dS)Gz-Access to Python's configuration information.N)pardirrealpathget_config_h_filenameget_config_varget_config_varsget_makefile_filenameget_pathget_path_names	get_pathsget_platformget_python_versionget_scheme_namesparse_config_hcCs(z
t|WStk
r"|YSXdSN)rOSError)pathrK/usr/lib/python3.8/site-packages/pip/_vendor/distlib/_backport/sysconfig.py_safe_realpath"s
rntZpcbuildiz\pc\viz\pcbuild\amd64icCs,dD]"}tjtjtd|rdSqdS)N)z
Setup.distzSetup.localZModulesTF)osrisfilejoin
_PROJECT_BASE)fnrrris_python_build:srFc	Cstsddlm}tddd}||}|d}|s>td|}t	|W5QRXt
rdD] }t|d	d
t|ddqdd
adS)N)finder.rz
sysconfig.cfgzsysconfig.cfg exists)posix_prefixZ
posix_homeincludez{srcdir}/Includeplatincludez{projectbase}/.T)	_cfg_readZ	resourcesr__name__rsplitfindAssertionErrorZ	as_stream_SCHEMESZreadfp
_PYTHON_BUILDset)rZbackport_packageZ_finderZ_cfgfilesschemerrr_ensure_cfg_readDs

r-z\{([^{]*?)\}c
st|dr|d}nt}|}|D]8}|dkr._replacer)r-Zhas_sectionitemstuplesectionsZ
has_optionr*Zremove_sectiondict	_VAR_REPLsub)configr.r9ZsectionZoptionvaluer6rr4r_expand_globalsYs$

r?rcsfdd}t||S)zIn the string `path`, replace tokens like {some.thing} with the
    corresponding value from the map `local_vars`.

    If there is no corresponding value, leave the token unchanged.
    cs8|d}|kr|S|tjkr.tj|S|dSr/)r1renvironr2
local_varsrrr6s


z_subst_vars.._replacerr;r<)rrCr6rrBr_subst_varssrEcCs0|}|D]\}}||kr"q|||<qdSr)keysr7)Ztarget_dictZ
other_dictZtarget_keyskeyr>rrr_extend_dicts
rHcCs`i}|dkri}t|tt|D]4\}}tjdkrDtj|}tjt	||||<q&|S)N)posixr)
rHrr(r7rr3r
expandusernormpathrE)r,varsresrGr>rrr_expand_varss
rNcsfdd}t||S)Ncs$|d}|kr|S|dSr/r0r2rLrrr6s
zformat_value.._replacerrD)r>rLr6rrOrformat_valuesrPcCstjdkrdStjS)NrIr )rr3rrrr_get_default_schemes
rQcCstjdd}dd}tjdkrBtjdp.d}|r8|S||dStjdkr|td	}|r||r`|S|dd
|dtjddS|r|S|dd
SdS)NPYTHONUSERBASEcWstjtjj|Sr)rrrJr)argsrrrjoinusersz_getuserbase..joinuserrAPPDATA~PythondarwinZPYTHONFRAMEWORKLibraryz%d.%drz.local)rrAgetr3sysplatformrversion_info)env_baserTbaseZ	frameworkrrr_getuserbases$



r`c	Cstd}td}td}|dkr*i}i}i}tj|ddd}|}W5QRX|D]}	|	dsZ|	d	krvqZ||	}
|
rZ|
d
d\}}|}|	dd	}
d
|
kr|||<qZzt
|}Wn$tk
r|	dd
||<YqZX|||<qZt|
}d}t|dkrt|D]}||}||p>||}
|
dk	r|
d
}d}||krpt||}n||krd}nx|tjkrtj|}n`||kr|dr|dd|krd	}n$d||krd}nt|d|}nd	||<}|r||
d}|d|
||}d
|kr:|||<n~zt
|}Wn"tk
rh|||<Yn
X|||<|||dr|dd|kr|dd}||kr|||<n|||<||qq|D]"\}}t|tr|||<q|||S)zParse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    z"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)z\$\(([A-Za-z][A-Za-z0-9_]*)\)z\${([A-Za-z][A-Za-z0-9_]*)}Nzutf-8surrogateescape)encodingerrors#rrz$$$)CFLAGSLDFLAGSZCPPFLAGSrTFZPY_r@)recompilecodecsopen	readlines
startswithstripmatchr1replaceint
ValueErrorlistrFlenr8searchstrrrAendstartremover7
isinstanceupdate)filenamerLZ_variable_rxZ_findvar1_rxZ_findvar2_rxZdoneZnotdoneflineslinemnvZtmpvr5Zrenamed_variablesr3r>founditemZafterkrrr_parse_makefiles	
















rcCsDtrtjtdSttdr,dttjf}nd}tjt	d|dS)z Return the path of the Makefile.ZMakefileabiflagszconfig-%s%sr=stdlib)
r)rrrrhasattrr[_PY_VERSION_SHORTrr)Zconfig_dir_namerrrrMs
c
Cst}zt||WnJtk
r^}z,d|}t|drF|d|j}t|W5d}~XYnXt}z"t|}t||W5QRXWnJtk
r}z,d|}t|dr|d|j}t|W5d}~XYnXtr|d|d<dS)z7Initialize the module as appropriate for POSIX systems.z.invalid Python installation: unable to open %sstrerrorz (%s)N	BLDSHAREDZLDSHARED)	rrIOErrorrrrrlrr))rLZmakefileemsgZconfig_hr~rrr_init_posixXs&


rcCsVtd|d<td|d<td|d<d|d<d	|d
<t|d<tjttj|d<d
S)z+Initialize the module as appropriate for NTrZLIBDESTZ
platstdlibZ
BINLIBDESTr!Z	INCLUDEPYz.pydZSOz.exeZEXEZVERSIONZBINDIRN)r_PY_VERSION_SHORT_NO_DOTrrdirnamerr[
executablerOrrr_init_non_posixtsrcCs|dkri}td}td}|}|s.q||}|rx|dd\}}zt|}Wntk
rlYnX|||<q ||}|r d||d<q |S)zParse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    Nz"#define ([A-Z][A-Za-z0-9_]+) (.*)
z&/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/
rrr)rirjreadlinerpr1rrrs)fprLZ	define_rxZundef_rxrrrrrrrrs&




cCs:tr$tjdkrtjtd}q,t}ntd}tj|dS)zReturn the path of pyconfig.h.rZPCr"z
pyconfig.h)r)rr3rrrr)Zinc_dirrrrrs
cCstttS)z,Return a tuple containing the schemes names.)r8sortedr(r9rrrrr
scCs
tdS)z*Return a tuple containing the paths names.r )r(Zoptionsrrrrr	sTcCs&t|rt||Stt|SdS)zReturn a mapping containing an install scheme.

    ``scheme`` is the install scheme name. If not provided, it will
    return the default scheme for the current platform.
    N)r-rNr:r(r7)r,rLexpandrrrr
s
cCst||||S)z[Return a path corresponding to the scheme.

    ``scheme`` is the install scheme name.
    )r
)r3r,rLrrrrrscGstdkrziattd<ttd<ttd<ttd<tdtdtd<ttd	<ttd
<ttd<ztjtd<Wntk
rd
td<YnXt	j
dkrttt	j
dkrtttj
dkrttd<dtkrttd<nttdtd<tr\t	j
dkr\t}zt	}Wntk
rd}YnXt	jtds\||kr\t	j|td}t	j|td<tjdkrzt	d}t|dd}|dkrdD]2}t|}tdd|}tdd|}|t|<qndt	jkrt	jd}dD]0}t|}tdd|}|d|}|t|<qtdd
}	td|	}
|
dk	rz|
 d}t	j!|szdD]$}t|}tdd|}|t|<qT|rg}|D]}
|"t|
q|StSdS)ayWith no arguments, return a dictionary of all configuration
    variables relevant for the current platform.

    On Unix, this means every variable defined in Python's installed Makefile;
    On Windows and Mac OS it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    Nprefixexec_prefixZ
py_versionZpy_version_shortrrZpy_version_nodotr_ZplatbaseZprojectbaserre)rZos2rIz2.6userbasesrcdirrXr)rhZ
BASECFLAGSrgZ	PY_CFLAGSrz
-arch\s+\w+\s z-isysroot [^ 	]*Z	ARCHFLAGSrgz-isysroot\s+(\S+)rz-isysroot\s+\S+(\s|$))#_CONFIG_VARS_PREFIX_EXEC_PREFIX_PY_VERSIONrrr[rAttributeErrorrr3rrversionr`rr)getcwdrrisabsrrKr\unamerrsplitrir<rArZrvr1existsappend)rSr_cwdrZkernel_versionZ
major_versionrGflagsZarchrgrZsdkZvalsr3rrrrs










cCst|S)zReturn the value of a single variable using the dictionary returned by
    'get_config_vars()'.

    Equivalent to get_config_vars().get(name)
    )rrZ)r3rrrrRscCs^tjdkrnd}tj|}|dkr(tjStjd|}tj|t||}|dkr\dS|dkrhdStjStjd	ksttd
stjSt	\}}}}}|
dd}|
d
d}|
dd}|dddkrd||fS|dddkr&|ddkrPd}dt|dd|ddf}n*|dddkrDd||fS|dddkrdd|||fS|ddd krd }t
d!}	|	|}
|
rP|
}n|ddd"krPt}|d#}|}
ztd$}Wntk
rYnJXztd%|}
W5|X|
dk	r2d&|
d'd&dd}
|s<|
}|rP|}d(}|
d&d)krd*td+dkrd,}td+}td-|}ttt|}t|d'kr|d}n^|d.krd,}nN|d/krd0}n>|d1krd2}n.|d3krd4}n|d5krd6}ntd7|fn<|d8kr0tjd9krPd:}n |d;krPtjd9krLd<}nd=}d>|||fS)?aReturn a string that identifies the current platform.

    This is used mainly to distinguish platform-specific build directories and
    platform-specific built distributions.  Typically includes the OS name
    and version and the architecture (as supplied by 'os.uname()'),
    although the exact information included depends on the OS; eg. for IRIX
    the architecture isn't particularly important (IRIX only runs on SGI
    hardware), but for Linux the kernel version isn't particularly
    important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u
       irix-5.3
       irix64-6.2

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win-ia64 (64bit Windows on Itanium)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.
    rz bit ()Zamd64z	win-amd64Zitaniumzwin-ia64rIr/rer_-NZlinuxz%s-%sZsunosr5Zsolarisz%d.%sr@rZirixZaixz%s-%s.%scygwinz[\d.]+rXZMACOSX_DEPLOYMENT_TARGETz0/System/Library/CoreServices/SystemVersion.plistz=ProductUserVisibleVersion\s*(.*?)rrZmacosxz10.4.z-archrgZfatz
-arch\s+(\S+))i386ppc)rx86_64Zintel)rrrZfat3)ppc64rZfat64)rrrrZ	universalz%Don't know machine value for archs=%rrlr)ZPowerPCZPower_Macintoshrrz%s-%s-%s) rr3r[rr&r\rulowerrrrqrrrirjrpr1rrZrlrcloservreadrrrofindallr8rr*rsmaxsize)rijZlookZosnameZhostreleasermachineZrel_rerZcfgvarsZmacverZ
macreleaser~ZcflagsZarchsrrrr[s
$












cCstSr)rrrrrrscCsFtt|D]0\}\}}|dkr0td|td||fqdS)Nrz%s: z
	%s = "%s")	enumeraterr7print)titledataindexrGr>rrr_print_dictsrcCsRtdttdttdtttdtttdtdS)z*Display all information sysconfig detains.zPlatform: "%s"zPython version: "%s"z!Current installation scheme: "%s"ZPathsZ	VariablesN)rrrrQrr
rrrrr_mainsr__main__)N)N)A__doc__rkrrir[Zos.pathrrZconfigparserImportErrorZConfigParser__all__rrrrrrr3rrrr)r#r-ZRawConfigParserr(rjr;r?rrrrrrKrrrrrZ
_USER_BASErErHrNrPrQr`rrrrrrr
r	r
rrrrrrrr$rrrrs""
#
	
v

	#
site-packages/pip/_vendor/distlib/_backport/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000677151464452710024467 0ustar00U

.e@sdZdS)aModules copied from Python 3 standard libraries, for internal use only.

Individual classes and functions are found in d2._backport.misc.  Intended
usage is to always import things missing from 3.1 from that module: the
built-in/stdlib objects will be used if found.
N)__doc__rrJ/usr/lib/python3.8/site-packages/pip/_vendor/distlib/_backport/__init__.pysite-packages/pip/_vendor/distlib/_backport/__pycache__/misc.cpython-38.opt-1.pyc000064400000002042151464452710023647 0ustar00U

.e@sdZddlZddlZdddgZzddlmZWnek
rLdddZYnXzeZWn(ek
r~dd	l	m
Z
d
dZYnXz
ejZWnek
rddZYnXdS)
z/Backports for individual classes and functions.Ncache_from_sourcecallablefsencode)rFcCs|rdp
d}||S)Nco)Zpy_filedebugZextrrF/usr/lib/python3.8/site-packages/pip/_vendor/distlib/_backport/misc.pyrs)CallablecCs
t|tS)N)
isinstancer
)objrrr	rscCs<t|tr|St|tr&|tStdt|jdS)Nzexpect bytes or str, not %s)	rbytesstrencodesysgetfilesystemencoding	TypeErrortype__name__)filenamerrr	r"s

)F)
__doc__osr__all__ZimprImportErrorr	NameErrorcollectionsr
rAttributeErrorrrrr	s 

site-packages/pip/_vendor/distlib/_backport/sysconfig.py000064400000064524151464452710017510 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Access to Python's configuration information."""

import codecs
import os
import re
import sys
from os.path import pardir, realpath
try:
    import configparser
except ImportError:
    import ConfigParser as configparser


__all__ = [
    'get_config_h_filename',
    'get_config_var',
    'get_config_vars',
    'get_makefile_filename',
    'get_path',
    'get_path_names',
    'get_paths',
    'get_platform',
    'get_python_version',
    'get_scheme_names',
    'parse_config_h',
]


def _safe_realpath(path):
    try:
        return realpath(path)
    except OSError:
        return path


if sys.executable:
    _PROJECT_BASE = os.path.dirname(_safe_realpath(sys.executable))
else:
    # sys.executable can be empty if argv[0] has been changed and Python is
    # unable to retrieve the real program name
    _PROJECT_BASE = _safe_realpath(os.getcwd())

if os.name == "nt" and "pcbuild" in _PROJECT_BASE[-8:].lower():
    _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir))
# PC/VS7.1
if os.name == "nt" and "\\pc\\v" in _PROJECT_BASE[-10:].lower():
    _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir, pardir))
# PC/AMD64
if os.name == "nt" and "\\pcbuild\\amd64" in _PROJECT_BASE[-14:].lower():
    _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir, pardir))


def is_python_build():
    for fn in ("Setup.dist", "Setup.local"):
        if os.path.isfile(os.path.join(_PROJECT_BASE, "Modules", fn)):
            return True
    return False

_PYTHON_BUILD = is_python_build()

_cfg_read = False

def _ensure_cfg_read():
    global _cfg_read
    if not _cfg_read:
        from ..resources import finder
        backport_package = __name__.rsplit('.', 1)[0]
        _finder = finder(backport_package)
        _cfgfile = _finder.find('sysconfig.cfg')
        assert _cfgfile, 'sysconfig.cfg exists'
        with _cfgfile.as_stream() as s:
            _SCHEMES.readfp(s)
        if _PYTHON_BUILD:
            for scheme in ('posix_prefix', 'posix_home'):
                _SCHEMES.set(scheme, 'include', '{srcdir}/Include')
                _SCHEMES.set(scheme, 'platinclude', '{projectbase}/.')

        _cfg_read = True


_SCHEMES = configparser.RawConfigParser()
_VAR_REPL = re.compile(r'\{([^{]*?)\}')

def _expand_globals(config):
    _ensure_cfg_read()
    if config.has_section('globals'):
        globals = config.items('globals')
    else:
        globals = tuple()

    sections = config.sections()
    for section in sections:
        if section == 'globals':
            continue
        for option, value in globals:
            if config.has_option(section, option):
                continue
            config.set(section, option, value)
    config.remove_section('globals')

    # now expanding local variables defined in the cfg file
    #
    for section in config.sections():
        variables = dict(config.items(section))

        def _replacer(matchobj):
            name = matchobj.group(1)
            if name in variables:
                return variables[name]
            return matchobj.group(0)

        for option, value in config.items(section):
            config.set(section, option, _VAR_REPL.sub(_replacer, value))

#_expand_globals(_SCHEMES)

 # FIXME don't rely on sys.version here, its format is an implementation detail
 # of CPython, use sys.version_info or sys.hexversion
_PY_VERSION = sys.version.split()[0]
_PY_VERSION_SHORT = sys.version[:3]
_PY_VERSION_SHORT_NO_DOT = _PY_VERSION[0] + _PY_VERSION[2]
_PREFIX = os.path.normpath(sys.prefix)
_EXEC_PREFIX = os.path.normpath(sys.exec_prefix)
_CONFIG_VARS = None
_USER_BASE = None


def _subst_vars(path, local_vars):
    """In the string `path`, replace tokens like {some.thing} with the
    corresponding value from the map `local_vars`.

    If there is no corresponding value, leave the token unchanged.
    """
    def _replacer(matchobj):
        name = matchobj.group(1)
        if name in local_vars:
            return local_vars[name]
        elif name in os.environ:
            return os.environ[name]
        return matchobj.group(0)
    return _VAR_REPL.sub(_replacer, path)


def _extend_dict(target_dict, other_dict):
    target_keys = target_dict.keys()
    for key, value in other_dict.items():
        if key in target_keys:
            continue
        target_dict[key] = value


def _expand_vars(scheme, vars):
    res = {}
    if vars is None:
        vars = {}
    _extend_dict(vars, get_config_vars())

    for key, value in _SCHEMES.items(scheme):
        if os.name in ('posix', 'nt'):
            value = os.path.expanduser(value)
        res[key] = os.path.normpath(_subst_vars(value, vars))
    return res


def format_value(value, vars):
    def _replacer(matchobj):
        name = matchobj.group(1)
        if name in vars:
            return vars[name]
        return matchobj.group(0)
    return _VAR_REPL.sub(_replacer, value)


def _get_default_scheme():
    if os.name == 'posix':
        # the default scheme for posix is posix_prefix
        return 'posix_prefix'
    return os.name


def _getuserbase():
    env_base = os.environ.get("PYTHONUSERBASE", None)

    def joinuser(*args):
        return os.path.expanduser(os.path.join(*args))

    # what about 'os2emx', 'riscos' ?
    if os.name == "nt":
        base = os.environ.get("APPDATA") or "~"
        if env_base:
            return env_base
        else:
            return joinuser(base, "Python")

    if sys.platform == "darwin":
        framework = get_config_var("PYTHONFRAMEWORK")
        if framework:
            if env_base:
                return env_base
            else:
                return joinuser("~", "Library", framework, "%d.%d" %
                                sys.version_info[:2])

    if env_base:
        return env_base
    else:
        return joinuser("~", ".local")


def _parse_makefile(filename, vars=None):
    """Parse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    """
    # Regexes needed for parsing Makefile (and similar syntaxes,
    # like old-style Setup files).
    _variable_rx = re.compile(r"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)")
    _findvar1_rx = re.compile(r"\$\(([A-Za-z][A-Za-z0-9_]*)\)")
    _findvar2_rx = re.compile(r"\${([A-Za-z][A-Za-z0-9_]*)}")

    if vars is None:
        vars = {}
    done = {}
    notdone = {}

    with codecs.open(filename, encoding='utf-8', errors="surrogateescape") as f:
        lines = f.readlines()

    for line in lines:
        if line.startswith('#') or line.strip() == '':
            continue
        m = _variable_rx.match(line)
        if m:
            n, v = m.group(1, 2)
            v = v.strip()
            # `$$' is a literal `$' in make
            tmpv = v.replace('$$', '')

            if "$" in tmpv:
                notdone[n] = v
            else:
                try:
                    v = int(v)
                except ValueError:
                    # insert literal `$'
                    done[n] = v.replace('$$', '$')
                else:
                    done[n] = v

    # do variable interpolation here
    variables = list(notdone.keys())

    # Variables with a 'PY_' prefix in the makefile. These need to
    # be made available without that prefix through sysconfig.
    # Special care is needed to ensure that variable expansion works, even
    # if the expansion uses the name without a prefix.
    renamed_variables = ('CFLAGS', 'LDFLAGS', 'CPPFLAGS')

    while len(variables) > 0:
        for name in tuple(variables):
            value = notdone[name]
            m = _findvar1_rx.search(value) or _findvar2_rx.search(value)
            if m is not None:
                n = m.group(1)
                found = True
                if n in done:
                    item = str(done[n])
                elif n in notdone:
                    # get it on a subsequent round
                    found = False
                elif n in os.environ:
                    # do it like make: fall back to environment
                    item = os.environ[n]

                elif n in renamed_variables:
                    if (name.startswith('PY_') and
                        name[3:] in renamed_variables):
                        item = ""

                    elif 'PY_' + n in notdone:
                        found = False

                    else:
                        item = str(done['PY_' + n])

                else:
                    done[n] = item = ""

                if found:
                    after = value[m.end():]
                    value = value[:m.start()] + item + after
                    if "$" in after:
                        notdone[name] = value
                    else:
                        try:
                            value = int(value)
                        except ValueError:
                            done[name] = value.strip()
                        else:
                            done[name] = value
                        variables.remove(name)

                        if (name.startswith('PY_') and
                            name[3:] in renamed_variables):

                            name = name[3:]
                            if name not in done:
                                done[name] = value

            else:
                # bogus variable reference (e.g. "prefix=$/opt/python");
                # just drop it since we can't deal
                done[name] = value
                variables.remove(name)

    # strip spurious spaces
    for k, v in done.items():
        if isinstance(v, str):
            done[k] = v.strip()

    # save the results in the global dictionary
    vars.update(done)
    return vars


def get_makefile_filename():
    """Return the path of the Makefile."""
    if _PYTHON_BUILD:
        return os.path.join(_PROJECT_BASE, "Makefile")
    if hasattr(sys, 'abiflags'):
        config_dir_name = 'config-%s%s' % (_PY_VERSION_SHORT, sys.abiflags)
    else:
        config_dir_name = 'config'
    return os.path.join(get_path('stdlib'), config_dir_name, 'Makefile')


def _init_posix(vars):
    """Initialize the module as appropriate for POSIX systems."""
    # load the installed Makefile:
    makefile = get_makefile_filename()
    try:
        _parse_makefile(makefile, vars)
    except IOError as e:
        msg = "invalid Python installation: unable to open %s" % makefile
        if hasattr(e, "strerror"):
            msg = msg + " (%s)" % e.strerror
        raise IOError(msg)
    # load the installed pyconfig.h:
    config_h = get_config_h_filename()
    try:
        with open(config_h) as f:
            parse_config_h(f, vars)
    except IOError as e:
        msg = "invalid Python installation: unable to open %s" % config_h
        if hasattr(e, "strerror"):
            msg = msg + " (%s)" % e.strerror
        raise IOError(msg)
    # On AIX, there are wrong paths to the linker scripts in the Makefile
    # -- these paths are relative to the Python source, but when installed
    # the scripts are in another directory.
    if _PYTHON_BUILD:
        vars['LDSHARED'] = vars['BLDSHARED']


def _init_non_posix(vars):
    """Initialize the module as appropriate for NT"""
    # set basic install directories
    vars['LIBDEST'] = get_path('stdlib')
    vars['BINLIBDEST'] = get_path('platstdlib')
    vars['INCLUDEPY'] = get_path('include')
    vars['SO'] = '.pyd'
    vars['EXE'] = '.exe'
    vars['VERSION'] = _PY_VERSION_SHORT_NO_DOT
    vars['BINDIR'] = os.path.dirname(_safe_realpath(sys.executable))

#
# public APIs
#


def parse_config_h(fp, vars=None):
    """Parse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    """
    if vars is None:
        vars = {}
    define_rx = re.compile("#define ([A-Z][A-Za-z0-9_]+) (.*)\n")
    undef_rx = re.compile("/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/\n")

    while True:
        line = fp.readline()
        if not line:
            break
        m = define_rx.match(line)
        if m:
            n, v = m.group(1, 2)
            try:
                v = int(v)
            except ValueError:
                pass
            vars[n] = v
        else:
            m = undef_rx.match(line)
            if m:
                vars[m.group(1)] = 0
    return vars


def get_config_h_filename():
    """Return the path of pyconfig.h."""
    if _PYTHON_BUILD:
        if os.name == "nt":
            inc_dir = os.path.join(_PROJECT_BASE, "PC")
        else:
            inc_dir = _PROJECT_BASE
    else:
        inc_dir = get_path('platinclude')
    return os.path.join(inc_dir, 'pyconfig.h')


def get_scheme_names():
    """Return a tuple containing the schemes names."""
    return tuple(sorted(_SCHEMES.sections()))


def get_path_names():
    """Return a tuple containing the paths names."""
    # xxx see if we want a static list
    return _SCHEMES.options('posix_prefix')


def get_paths(scheme=_get_default_scheme(), vars=None, expand=True):
    """Return a mapping containing an install scheme.

    ``scheme`` is the install scheme name. If not provided, it will
    return the default scheme for the current platform.
    """
    _ensure_cfg_read()
    if expand:
        return _expand_vars(scheme, vars)
    else:
        return dict(_SCHEMES.items(scheme))


def get_path(name, scheme=_get_default_scheme(), vars=None, expand=True):
    """Return a path corresponding to the scheme.

    ``scheme`` is the install scheme name.
    """
    return get_paths(scheme, vars, expand)[name]


def get_config_vars(*args):
    """With no arguments, return a dictionary of all configuration
    variables relevant for the current platform.

    On Unix, this means every variable defined in Python's installed Makefile;
    On Windows and Mac OS it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    """
    global _CONFIG_VARS
    if _CONFIG_VARS is None:
        _CONFIG_VARS = {}
        # Normalized versions of prefix and exec_prefix are handy to have;
        # in fact, these are the standard versions used most places in the
        # distutils2 module.
        _CONFIG_VARS['prefix'] = _PREFIX
        _CONFIG_VARS['exec_prefix'] = _EXEC_PREFIX
        _CONFIG_VARS['py_version'] = _PY_VERSION
        _CONFIG_VARS['py_version_short'] = _PY_VERSION_SHORT
        _CONFIG_VARS['py_version_nodot'] = _PY_VERSION[0] + _PY_VERSION[2]
        _CONFIG_VARS['base'] = _PREFIX
        _CONFIG_VARS['platbase'] = _EXEC_PREFIX
        _CONFIG_VARS['projectbase'] = _PROJECT_BASE
        try:
            _CONFIG_VARS['abiflags'] = sys.abiflags
        except AttributeError:
            # sys.abiflags may not be defined on all platforms.
            _CONFIG_VARS['abiflags'] = ''

        if os.name in ('nt', 'os2'):
            _init_non_posix(_CONFIG_VARS)
        if os.name == 'posix':
            _init_posix(_CONFIG_VARS)
        # Setting 'userbase' is done below the call to the
        # init function to enable using 'get_config_var' in
        # the init-function.
        if sys.version >= '2.6':
            _CONFIG_VARS['userbase'] = _getuserbase()

        if 'srcdir' not in _CONFIG_VARS:
            _CONFIG_VARS['srcdir'] = _PROJECT_BASE
        else:
            _CONFIG_VARS['srcdir'] = _safe_realpath(_CONFIG_VARS['srcdir'])

        # Convert srcdir into an absolute path if it appears necessary.
        # Normally it is relative to the build directory.  However, during
        # testing, for example, we might be running a non-installed python
        # from a different directory.
        if _PYTHON_BUILD and os.name == "posix":
            base = _PROJECT_BASE
            try:
                cwd = os.getcwd()
            except OSError:
                cwd = None
            if (not os.path.isabs(_CONFIG_VARS['srcdir']) and
                base != cwd):
                # srcdir is relative and we are not in the same directory
                # as the executable. Assume executable is in the build
                # directory and make srcdir absolute.
                srcdir = os.path.join(base, _CONFIG_VARS['srcdir'])
                _CONFIG_VARS['srcdir'] = os.path.normpath(srcdir)

        if sys.platform == 'darwin':
            kernel_version = os.uname()[2]  # Kernel version (8.4.3)
            major_version = int(kernel_version.split('.')[0])

            if major_version < 8:
                # On Mac OS X before 10.4, check if -arch and -isysroot
                # are in CFLAGS or LDFLAGS and remove them if they are.
                # This is needed when building extensions on a 10.3 system
                # using a universal build of python.
                for key in ('LDFLAGS', 'BASECFLAGS',
                        # a number of derived variables. These need to be
                        # patched up as well.
                        'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'):
                    flags = _CONFIG_VARS[key]
                    flags = re.sub(r'-arch\s+\w+\s', ' ', flags)
                    flags = re.sub('-isysroot [^ \t]*', ' ', flags)
                    _CONFIG_VARS[key] = flags
            else:
                # Allow the user to override the architecture flags using
                # an environment variable.
                # NOTE: This name was introduced by Apple in OSX 10.5 and
                # is used by several scripting languages distributed with
                # that OS release.
                if 'ARCHFLAGS' in os.environ:
                    arch = os.environ['ARCHFLAGS']
                    for key in ('LDFLAGS', 'BASECFLAGS',
                        # a number of derived variables. These need to be
                        # patched up as well.
                        'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'):

                        flags = _CONFIG_VARS[key]
                        flags = re.sub(r'-arch\s+\w+\s', ' ', flags)
                        flags = flags + ' ' + arch
                        _CONFIG_VARS[key] = flags

                # If we're on OSX 10.5 or later and the user tries to
                # compiles an extension using an SDK that is not present
                # on the current machine it is better to not use an SDK
                # than to fail.
                #
                # The major usecase for this is users using a Python.org
                # binary installer  on OSX 10.6: that installer uses
                # the 10.4u SDK, but that SDK is not installed by default
                # when you install Xcode.
                #
                CFLAGS = _CONFIG_VARS.get('CFLAGS', '')
                m = re.search(r'-isysroot\s+(\S+)', CFLAGS)
                if m is not None:
                    sdk = m.group(1)
                    if not os.path.exists(sdk):
                        for key in ('LDFLAGS', 'BASECFLAGS',
                             # a number of derived variables. These need to be
                             # patched up as well.
                            'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'):

                            flags = _CONFIG_VARS[key]
                            flags = re.sub(r'-isysroot\s+\S+(\s|$)', ' ', flags)
                            _CONFIG_VARS[key] = flags

    if args:
        vals = []
        for name in args:
            vals.append(_CONFIG_VARS.get(name))
        return vals
    else:
        return _CONFIG_VARS


def get_config_var(name):
    """Return the value of a single variable using the dictionary returned by
    'get_config_vars()'.

    Equivalent to get_config_vars().get(name)
    """
    return get_config_vars().get(name)


def get_platform():
    """Return a string that identifies the current platform.

    This is used mainly to distinguish platform-specific build directories and
    platform-specific built distributions.  Typically includes the OS name
    and version and the architecture (as supplied by 'os.uname()'),
    although the exact information included depends on the OS; eg. for IRIX
    the architecture isn't particularly important (IRIX only runs on SGI
    hardware), but for Linux the kernel version isn't particularly
    important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u
       irix-5.3
       irix64-6.2

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win-ia64 (64bit Windows on Itanium)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.
    """
    if os.name == 'nt':
        # sniff sys.version for architecture.
        prefix = " bit ("
        i = sys.version.find(prefix)
        if i == -1:
            return sys.platform
        j = sys.version.find(")", i)
        look = sys.version[i+len(prefix):j].lower()
        if look == 'amd64':
            return 'win-amd64'
        if look == 'itanium':
            return 'win-ia64'
        return sys.platform

    if os.name != "posix" or not hasattr(os, 'uname'):
        # XXX what about the architecture? NT is Intel or Alpha,
        # Mac OS is M68k or PPC, etc.
        return sys.platform

    # Try to distinguish various flavours of Unix
    osname, host, release, version, machine = os.uname()

    # Convert the OS name to lowercase, remove '/' characters
    # (to accommodate BSD/OS), and translate spaces (for "Power Macintosh")
    osname = osname.lower().replace('/', '')
    machine = machine.replace(' ', '_')
    machine = machine.replace('/', '-')

    if osname[:5] == "linux":
        # At least on Linux/Intel, 'machine' is the processor --
        # i386, etc.
        # XXX what about Alpha, SPARC, etc?
        return  "%s-%s" % (osname, machine)
    elif osname[:5] == "sunos":
        if release[0] >= "5":           # SunOS 5 == Solaris 2
            osname = "solaris"
            release = "%d.%s" % (int(release[0]) - 3, release[2:])
        # fall through to standard osname-release-machine representation
    elif osname[:4] == "irix":              # could be "irix64"!
        return "%s-%s" % (osname, release)
    elif osname[:3] == "aix":
        return "%s-%s.%s" % (osname, version, release)
    elif osname[:6] == "cygwin":
        osname = "cygwin"
        rel_re = re.compile(r'[\d.]+')
        m = rel_re.match(release)
        if m:
            release = m.group()
    elif osname[:6] == "darwin":
        #
        # For our purposes, we'll assume that the system version from
        # distutils' perspective is what MACOSX_DEPLOYMENT_TARGET is set
        # to. This makes the compatibility story a bit more sane because the
        # machine is going to compile and link as if it were
        # MACOSX_DEPLOYMENT_TARGET.
        cfgvars = get_config_vars()
        macver = cfgvars.get('MACOSX_DEPLOYMENT_TARGET')

        if True:
            # Always calculate the release of the running machine,
            # needed to determine if we can build fat binaries or not.

            macrelease = macver
            # Get the system version. Reading this plist is a documented
            # way to get the system version (see the documentation for
            # the Gestalt Manager)
            try:
                f = open('/System/Library/CoreServices/SystemVersion.plist')
            except IOError:
                # We're on a plain darwin box, fall back to the default
                # behaviour.
                pass
            else:
                try:
                    m = re.search(r'ProductUserVisibleVersion\s*'
                                  r'(.*?)', f.read())
                finally:
                    f.close()
                if m is not None:
                    macrelease = '.'.join(m.group(1).split('.')[:2])
                # else: fall back to the default behaviour

        if not macver:
            macver = macrelease

        if macver:
            release = macver
            osname = "macosx"

            if ((macrelease + '.') >= '10.4.' and
                '-arch' in get_config_vars().get('CFLAGS', '').strip()):
                # The universal build will build fat binaries, but not on
                # systems before 10.4
                #
                # Try to detect 4-way universal builds, those have machine-type
                # 'universal' instead of 'fat'.

                machine = 'fat'
                cflags = get_config_vars().get('CFLAGS')

                archs = re.findall(r'-arch\s+(\S+)', cflags)
                archs = tuple(sorted(set(archs)))

                if len(archs) == 1:
                    machine = archs[0]
                elif archs == ('i386', 'ppc'):
                    machine = 'fat'
                elif archs == ('i386', 'x86_64'):
                    machine = 'intel'
                elif archs == ('i386', 'ppc', 'x86_64'):
                    machine = 'fat3'
                elif archs == ('ppc64', 'x86_64'):
                    machine = 'fat64'
                elif archs == ('i386', 'ppc', 'ppc64', 'x86_64'):
                    machine = 'universal'
                else:
                    raise ValueError(
                       "Don't know machine value for archs=%r" % (archs,))

            elif machine == 'i386':
                # On OSX the machine type returned by uname is always the
                # 32-bit variant, even if the executable architecture is
                # the 64-bit variant
                if sys.maxsize >= 2**32:
                    machine = 'x86_64'

            elif machine in ('PowerPC', 'Power_Macintosh'):
                # Pick a sane name for the PPC architecture.
                # See 'i386' case
                if sys.maxsize >= 2**32:
                    machine = 'ppc64'
                else:
                    machine = 'ppc'

    return "%s-%s-%s" % (osname, release, machine)


def get_python_version():
    return _PY_VERSION_SHORT


def _print_dict(title, data):
    for index, (key, value) in enumerate(sorted(data.items())):
        if index == 0:
            print('%s: ' % (title))
        print('\t%s = "%s"' % (key, value))


def _main():
    """Display all information sysconfig detains."""
    print('Platform: "%s"' % get_platform())
    print('Python version: "%s"' % get_python_version())
    print('Current installation scheme: "%s"' % _get_default_scheme())
    print()
    _print_dict('Paths', get_paths())
    print()
    _print_dict('Variables', get_config_vars())


if __name__ == '__main__':
    _main()
site-packages/pip/_vendor/distlib/_backport/__init__.py000064400000000422151464452710017226 0ustar00"""Modules copied from Python 3 standard libraries, for internal use only.

Individual classes and functions are found in d2._backport.misc.  Intended
usage is to always import things missing from 3.1 from that module: the
built-in/stdlib objects will be used if found.
"""
site-packages/pip/_vendor/distlib/manifest.py000064400000034733151464452710015345 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2013 Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""
Class representing the list of files in a distribution.

Equivalent to distutils.filelist, but fixes some problems.
"""
import fnmatch
import logging
import os
import re
import sys

from . import DistlibException
from .compat import fsdecode
from .util import convert_path


__all__ = ['Manifest']

logger = logging.getLogger(__name__)

# a \ followed by some spaces + EOL
_COLLAPSE_PATTERN = re.compile('\\\\w*\n', re.M)
_COMMENTED_LINE = re.compile('#.*?(?=\n)|\n(?=$)', re.M | re.S)

#
# Due to the different results returned by fnmatch.translate, we need
# to do slightly different processing for Python 2.7 and 3.2 ... this needed
# to be brought in for Python 3.6 onwards.
#
_PYTHON_VERSION = sys.version_info[:2]

class Manifest(object):
    """A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.
    """

    def __init__(self, base=None):
        """
        Initialise an instance.

        :param base: The base directory to explore under.
        """
        self.base = os.path.abspath(os.path.normpath(base or os.getcwd()))
        self.prefix = self.base + os.sep
        self.allfiles = None
        self.files = set()

    #
    # Public API
    #

    def findall(self):
        """Find all files under the base and set ``allfiles`` to the absolute
        pathnames of files found.
        """
        from stat import S_ISREG, S_ISDIR, S_ISLNK

        self.allfiles = allfiles = []
        root = self.base
        stack = [root]
        pop = stack.pop
        push = stack.append

        while stack:
            root = pop()
            names = os.listdir(root)

            for name in names:
                fullname = os.path.join(root, name)

                # Avoid excess stat calls -- just one will do, thank you!
                stat = os.stat(fullname)
                mode = stat.st_mode
                if S_ISREG(mode):
                    allfiles.append(fsdecode(fullname))
                elif S_ISDIR(mode) and not S_ISLNK(mode):
                    push(fullname)

    def add(self, item):
        """
        Add a file to the manifest.

        :param item: The pathname to add. This can be relative to the base.
        """
        if not item.startswith(self.prefix):
            item = os.path.join(self.base, item)
        self.files.add(os.path.normpath(item))

    def add_many(self, items):
        """
        Add a list of files to the manifest.

        :param items: The pathnames to add. These can be relative to the base.
        """
        for item in items:
            self.add(item)

    def sorted(self, wantdirs=False):
        """
        Return sorted files in directory order
        """

        def add_dir(dirs, d):
            dirs.add(d)
            logger.debug('add_dir added %s', d)
            if d != self.base:
                parent, _ = os.path.split(d)
                assert parent not in ('', '/')
                add_dir(dirs, parent)

        result = set(self.files)    # make a copy!
        if wantdirs:
            dirs = set()
            for f in result:
                add_dir(dirs, os.path.dirname(f))
            result |= dirs
        return [os.path.join(*path_tuple) for path_tuple in
                sorted(os.path.split(path) for path in result)]

    def clear(self):
        """Clear all collected files."""
        self.files = set()
        self.allfiles = []

    def process_directive(self, directive):
        """
        Process a directive which either adds some files from ``allfiles`` to
        ``files``, or removes some files from ``files``.

        :param directive: The directive to process. This should be in a format
                     compatible with distutils ``MANIFEST.in`` files:

                     http://docs.python.org/distutils/sourcedist.html#commands
        """
        # Parse the line: split it up, make sure the right number of words
        # is there, and return the relevant words.  'action' is always
        # defined: it's the first word of the line.  Which of the other
        # three are defined depends on the action; it'll be either
        # patterns, (dir and patterns), or (dirpattern).
        action, patterns, thedir, dirpattern = self._parse_directive(directive)

        # OK, now we know that the action is valid and we have the
        # right number of words on the line for that action -- so we
        # can proceed with minimal error-checking.
        if action == 'include':
            for pattern in patterns:
                if not self._include_pattern(pattern, anchor=True):
                    logger.warning('no files found matching %r', pattern)

        elif action == 'exclude':
            for pattern in patterns:
                found = self._exclude_pattern(pattern, anchor=True)
                #if not found:
                #    logger.warning('no previously-included files '
                #                   'found matching %r', pattern)

        elif action == 'global-include':
            for pattern in patterns:
                if not self._include_pattern(pattern, anchor=False):
                    logger.warning('no files found matching %r '
                                   'anywhere in distribution', pattern)

        elif action == 'global-exclude':
            for pattern in patterns:
                found = self._exclude_pattern(pattern, anchor=False)
                #if not found:
                #    logger.warning('no previously-included files '
                #                   'matching %r found anywhere in '
                #                   'distribution', pattern)

        elif action == 'recursive-include':
            for pattern in patterns:
                if not self._include_pattern(pattern, prefix=thedir):
                    logger.warning('no files found matching %r '
                                   'under directory %r', pattern, thedir)

        elif action == 'recursive-exclude':
            for pattern in patterns:
                found = self._exclude_pattern(pattern, prefix=thedir)
                #if not found:
                #    logger.warning('no previously-included files '
                #                   'matching %r found under directory %r',
                #                   pattern, thedir)

        elif action == 'graft':
            if not self._include_pattern(None, prefix=dirpattern):
                logger.warning('no directories found matching %r',
                               dirpattern)

        elif action == 'prune':
            if not self._exclude_pattern(None, prefix=dirpattern):
                logger.warning('no previously-included directories found '
                               'matching %r', dirpattern)
        else:   # pragma: no cover
            # This should never happen, as it should be caught in
            # _parse_template_line
            raise DistlibException(
                'invalid action %r' % action)

    #
    # Private API
    #

    def _parse_directive(self, directive):
        """
        Validate a directive.
        :param directive: The directive to validate.
        :return: A tuple of action, patterns, thedir, dir_patterns
        """
        words = directive.split()
        if len(words) == 1 and words[0] not in ('include', 'exclude',
                                                'global-include',
                                                'global-exclude',
                                                'recursive-include',
                                                'recursive-exclude',
                                                'graft', 'prune'):
            # no action given, let's use the default 'include'
            words.insert(0, 'include')

        action = words[0]
        patterns = thedir = dir_pattern = None

        if action in ('include', 'exclude',
                      'global-include', 'global-exclude'):
            if len(words) < 2:
                raise DistlibException(
                    '%r expects   ...' % action)

            patterns = [convert_path(word) for word in words[1:]]

        elif action in ('recursive-include', 'recursive-exclude'):
            if len(words) < 3:
                raise DistlibException(
                    '%r expects    ...' % action)

            thedir = convert_path(words[1])
            patterns = [convert_path(word) for word in words[2:]]

        elif action in ('graft', 'prune'):
            if len(words) != 2:
                raise DistlibException(
                    '%r expects a single ' % action)

            dir_pattern = convert_path(words[1])

        else:
            raise DistlibException('unknown action %r' % action)

        return action, patterns, thedir, dir_pattern

    def _include_pattern(self, pattern, anchor=True, prefix=None,
                         is_regex=False):
        """Select strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.

        Patterns are not quite the same as implemented by the 'fnmatch'
        module: '*' and '?'  match non-special characters, where "special"
        is platform-dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found.
        """
        # XXX docstring lying about what the special chars are?
        found = False
        pattern_re = self._translate_pattern(pattern, anchor, prefix, is_regex)

        # delayed loading of allfiles list
        if self.allfiles is None:
            self.findall()

        for name in self.allfiles:
            if pattern_re.search(name):
                self.files.add(name)
                found = True
        return found

    def _exclude_pattern(self, pattern, anchor=True, prefix=None,
                         is_regex=False):
        """Remove strings (presumably filenames) from 'files' that match
        'pattern'.

        Other parameters are the same as for 'include_pattern()', above.
        The list 'self.files' is modified in place. Return True if files are
        found.

        This API is public to allow e.g. exclusion of SCM subdirs, e.g. when
        packaging source distributions
        """
        found = False
        pattern_re = self._translate_pattern(pattern, anchor, prefix, is_regex)
        for f in list(self.files):
            if pattern_re.search(f):
                self.files.remove(f)
                found = True
        return found

    def _translate_pattern(self, pattern, anchor=True, prefix=None,
                           is_regex=False):
        """Translate a shell-like wildcard pattern to a compiled regular
        expression.

        Return the compiled regex.  If 'is_regex' true,
        then 'pattern' is directly compiled to a regex (if it's a string)
        or just returned as-is (assumes it's a regex object).
        """
        if is_regex:
            if isinstance(pattern, str):
                return re.compile(pattern)
            else:
                return pattern

        if _PYTHON_VERSION > (3, 2):
            # ditch start and end characters
            start, _, end = self._glob_to_re('_').partition('_')

        if pattern:
            pattern_re = self._glob_to_re(pattern)
            if _PYTHON_VERSION > (3, 2):
                assert pattern_re.startswith(start) and pattern_re.endswith(end)
        else:
            pattern_re = ''

        base = re.escape(os.path.join(self.base, ''))
        if prefix is not None:
            # ditch end of pattern character
            if _PYTHON_VERSION <= (3, 2):
                empty_pattern = self._glob_to_re('')
                prefix_re = self._glob_to_re(prefix)[:-len(empty_pattern)]
            else:
                prefix_re = self._glob_to_re(prefix)
                assert prefix_re.startswith(start) and prefix_re.endswith(end)
                prefix_re = prefix_re[len(start): len(prefix_re) - len(end)]
            sep = os.sep
            if os.sep == '\\':
                sep = r'\\'
            if _PYTHON_VERSION <= (3, 2):
                pattern_re = '^' + base + sep.join((prefix_re,
                                                    '.*' + pattern_re))
            else:
                pattern_re = pattern_re[len(start): len(pattern_re) - len(end)]
                pattern_re = r'%s%s%s%s.*%s%s' % (start, base, prefix_re, sep,
                                                  pattern_re, end)
        else:  # no prefix -- respect anchor flag
            if anchor:
                if _PYTHON_VERSION <= (3, 2):
                    pattern_re = '^' + base + pattern_re
                else:
                    pattern_re = r'%s%s%s' % (start, base, pattern_re[len(start):])

        return re.compile(pattern_re)

    def _glob_to_re(self, pattern):
        """Translate a shell-like glob pattern to a regular expression.

        Return a string containing the regex.  Differs from
        'fnmatch.translate()' in that '*' does not match "special characters"
        (which are platform-specific).
        """
        pattern_re = fnmatch.translate(pattern)

        # '?' and '*' in the glob pattern become '.' and '.*' in the RE, which
        # IMHO is wrong -- '?' and '*' aren't supposed to match slash in Unix,
        # and by extension they shouldn't match such "special characters" under
        # any OS.  So change all non-escaped dots in the RE to match any
        # character except the special characters (currently: just os.sep).
        sep = os.sep
        if os.sep == '\\':
            # we're using a regex to manipulate a regex, so we need
            # to escape the backslash twice
            sep = r'\\\\'
        escaped = r'\1[^%s]' % sep
        pattern_re = re.sub(r'((? 0

    def parse(self, s):
        raise NotImplementedError('please implement in a subclass')

    def _check_compatible(self, other):
        if type(self) != type(other):
            raise TypeError('cannot compare %r and %r' % (self, other))

    def __eq__(self, other):
        self._check_compatible(other)
        return self._parts == other._parts

    def __ne__(self, other):
        return not self.__eq__(other)

    def __lt__(self, other):
        self._check_compatible(other)
        return self._parts < other._parts

    def __gt__(self, other):
        return not (self.__lt__(other) or self.__eq__(other))

    def __le__(self, other):
        return self.__lt__(other) or self.__eq__(other)

    def __ge__(self, other):
        return self.__gt__(other) or self.__eq__(other)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    def __hash__(self):
        return hash(self._parts)

    def __repr__(self):
        return "%s('%s')" % (self.__class__.__name__, self._string)

    def __str__(self):
        return self._string

    @property
    def is_prerelease(self):
        raise NotImplementedError('Please implement in subclasses.')


class Matcher(object):
    version_class = None

    # value is either a callable or the name of a method
    _operators = {
        '<': lambda v, c, p: v < c,
        '>': lambda v, c, p: v > c,
        '<=': lambda v, c, p: v == c or v < c,
        '>=': lambda v, c, p: v == c or v > c,
        '==': lambda v, c, p: v == c,
        '===': lambda v, c, p: v == c,
        # by default, compatible => >=.
        '~=': lambda v, c, p: v == c or v > c,
        '!=': lambda v, c, p: v != c,
    }

    # this is a method only to support alternative implementations
    # via overriding
    def parse_requirement(self, s):
        return parse_requirement(s)

    def __init__(self, s):
        if self.version_class is None:
            raise ValueError('Please specify a version class')
        self._string = s = s.strip()
        r = self.parse_requirement(s)
        if not r:
            raise ValueError('Not valid: %r' % s)
        self.name = r.name
        self.key = self.name.lower()    # for case-insensitive comparisons
        clist = []
        if r.constraints:
            # import pdb; pdb.set_trace()
            for op, s in r.constraints:
                if s.endswith('.*'):
                    if op not in ('==', '!='):
                        raise ValueError('\'.*\' not allowed for '
                                         '%r constraints' % op)
                    # Could be a partial version (e.g. for '2.*') which
                    # won't parse as a version, so keep it as a string
                    vn, prefix = s[:-2], True
                    # Just to check that vn is a valid version
                    self.version_class(vn)
                else:
                    # Should parse as a version, so we can create an
                    # instance for the comparison
                    vn, prefix = self.version_class(s), False
                clist.append((op, vn, prefix))
        self._parts = tuple(clist)

    def match(self, version):
        """
        Check if the provided version matches the constraints.

        :param version: The version to match against this instance.
        :type version: String or :class:`Version` instance.
        """
        if isinstance(version, string_types):
            version = self.version_class(version)
        for operator, constraint, prefix in self._parts:
            f = self._operators.get(operator)
            if isinstance(f, string_types):
                f = getattr(self, f)
            if not f:
                msg = ('%r not implemented '
                       'for %s' % (operator, self.__class__.__name__))
                raise NotImplementedError(msg)
            if not f(version, constraint, prefix):
                return False
        return True

    @property
    def exact_version(self):
        result = None
        if len(self._parts) == 1 and self._parts[0][0] in ('==', '==='):
            result = self._parts[0][1]
        return result

    def _check_compatible(self, other):
        if type(self) != type(other) or self.name != other.name:
            raise TypeError('cannot compare %s and %s' % (self, other))

    def __eq__(self, other):
        self._check_compatible(other)
        return self.key == other.key and self._parts == other._parts

    def __ne__(self, other):
        return not self.__eq__(other)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    def __hash__(self):
        return hash(self.key) + hash(self._parts)

    def __repr__(self):
        return "%s(%r)" % (self.__class__.__name__, self._string)

    def __str__(self):
        return self._string


PEP440_VERSION_RE = re.compile(r'^v?(\d+!)?(\d+(\.\d+)*)((a|b|c|rc)(\d+))?'
                               r'(\.(post)(\d+))?(\.(dev)(\d+))?'
                               r'(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$')


def _pep_440_key(s):
    s = s.strip()
    m = PEP440_VERSION_RE.match(s)
    if not m:
        raise UnsupportedVersionError('Not a valid version: %s' % s)
    groups = m.groups()
    nums = tuple(int(v) for v in groups[1].split('.'))
    while len(nums) > 1 and nums[-1] == 0:
        nums = nums[:-1]

    if not groups[0]:
        epoch = 0
    else:
        epoch = int(groups[0])
    pre = groups[4:6]
    post = groups[7:9]
    dev = groups[10:12]
    local = groups[13]
    if pre == (None, None):
        pre = ()
    else:
        pre = pre[0], int(pre[1])
    if post == (None, None):
        post = ()
    else:
        post = post[0], int(post[1])
    if dev == (None, None):
        dev = ()
    else:
        dev = dev[0], int(dev[1])
    if local is None:
        local = ()
    else:
        parts = []
        for part in local.split('.'):
            # to ensure that numeric compares as > lexicographic, avoid
            # comparing them directly, but encode a tuple which ensures
            # correct sorting
            if part.isdigit():
                part = (1, int(part))
            else:
                part = (0, part)
            parts.append(part)
        local = tuple(parts)
    if not pre:
        # either before pre-release, or final release and after
        if not post and dev:
            # before pre-release
            pre = ('a', -1)     # to sort before a0
        else:
            pre = ('z',)        # to sort after all pre-releases
    # now look at the state of post and dev.
    if not post:
        post = ('_',)   # sort before 'a'
    if not dev:
        dev = ('final',)

    #print('%s -> %s' % (s, m.groups()))
    return epoch, nums, pre, post, dev, local


_normalized_key = _pep_440_key


class NormalizedVersion(Version):
    """A rational version.

    Good:
        1.2         # equivalent to "1.2.0"
        1.2.0
        1.2a1
        1.2.3a2
        1.2.3b1
        1.2.3c1
        1.2.3.4
        TODO: fill this out

    Bad:
        1           # minimum two numbers
        1.2a        # release level must have a release serial
        1.2.3b
    """
    def parse(self, s):
        result = _normalized_key(s)
        # _normalized_key loses trailing zeroes in the release
        # clause, since that's needed to ensure that X.Y == X.Y.0 == X.Y.0.0
        # However, PEP 440 prefix matching needs it: for example,
        # (~= 1.4.5.0) matches differently to (~= 1.4.5.0.0).
        m = PEP440_VERSION_RE.match(s)      # must succeed
        groups = m.groups()
        self._release_clause = tuple(int(v) for v in groups[1].split('.'))
        return result

    PREREL_TAGS = set(['a', 'b', 'c', 'rc', 'dev'])

    @property
    def is_prerelease(self):
        return any(t[0] in self.PREREL_TAGS for t in self._parts if t)


def _match_prefix(x, y):
    x = str(x)
    y = str(y)
    if x == y:
        return True
    if not x.startswith(y):
        return False
    n = len(y)
    return x[n] == '.'


class NormalizedMatcher(Matcher):
    version_class = NormalizedVersion

    # value is either a callable or the name of a method
    _operators = {
        '~=': '_match_compatible',
        '<': '_match_lt',
        '>': '_match_gt',
        '<=': '_match_le',
        '>=': '_match_ge',
        '==': '_match_eq',
        '===': '_match_arbitrary',
        '!=': '_match_ne',
    }

    def _adjust_local(self, version, constraint, prefix):
        if prefix:
            strip_local = '+' not in constraint and version._parts[-1]
        else:
            # both constraint and version are
            # NormalizedVersion instances.
            # If constraint does not have a local component,
            # ensure the version doesn't, either.
            strip_local = not constraint._parts[-1] and version._parts[-1]
        if strip_local:
            s = version._string.split('+', 1)[0]
            version = self.version_class(s)
        return version, constraint

    def _match_lt(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if version >= constraint:
            return False
        release_clause = constraint._release_clause
        pfx = '.'.join([str(i) for i in release_clause])
        return not _match_prefix(version, pfx)

    def _match_gt(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if version <= constraint:
            return False
        release_clause = constraint._release_clause
        pfx = '.'.join([str(i) for i in release_clause])
        return not _match_prefix(version, pfx)

    def _match_le(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        return version <= constraint

    def _match_ge(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        return version >= constraint

    def _match_eq(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if not prefix:
            result = (version == constraint)
        else:
            result = _match_prefix(version, constraint)
        return result

    def _match_arbitrary(self, version, constraint, prefix):
        return str(version) == str(constraint)

    def _match_ne(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if not prefix:
            result = (version != constraint)
        else:
            result = not _match_prefix(version, constraint)
        return result

    def _match_compatible(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if version == constraint:
            return True
        if version < constraint:
            return False
#        if not prefix:
#            return True
        release_clause = constraint._release_clause
        if len(release_clause) > 1:
            release_clause = release_clause[:-1]
        pfx = '.'.join([str(i) for i in release_clause])
        return _match_prefix(version, pfx)

_REPLACEMENTS = (
    (re.compile('[.+-]$'), ''),                     # remove trailing puncts
    (re.compile(r'^[.](\d)'), r'0.\1'),             # .N -> 0.N at start
    (re.compile('^[.-]'), ''),                      # remove leading puncts
    (re.compile(r'^\((.*)\)$'), r'\1'),             # remove parentheses
    (re.compile(r'^v(ersion)?\s*(\d+)'), r'\2'),    # remove leading v(ersion)
    (re.compile(r'^r(ev)?\s*(\d+)'), r'\2'),        # remove leading v(ersion)
    (re.compile('[.]{2,}'), '.'),                   # multiple runs of '.'
    (re.compile(r'\b(alfa|apha)\b'), 'alpha'),      # misspelt alpha
    (re.compile(r'\b(pre-alpha|prealpha)\b'),
                'pre.alpha'),                       # standardise
    (re.compile(r'\(beta\)$'), 'beta'),             # remove parentheses
)

_SUFFIX_REPLACEMENTS = (
    (re.compile('^[:~._+-]+'), ''),                   # remove leading puncts
    (re.compile('[,*")([\\]]'), ''),                  # remove unwanted chars
    (re.compile('[~:+_ -]'), '.'),                    # replace illegal chars
    (re.compile('[.]{2,}'), '.'),                   # multiple runs of '.'
    (re.compile(r'\.$'), ''),                       # trailing '.'
)

_NUMERIC_PREFIX = re.compile(r'(\d+(\.\d+)*)')


def _suggest_semantic_version(s):
    """
    Try to suggest a semantic form for a version for which
    _suggest_normalized_version couldn't come up with anything.
    """
    result = s.strip().lower()
    for pat, repl in _REPLACEMENTS:
        result = pat.sub(repl, result)
    if not result:
        result = '0.0.0'

    # Now look for numeric prefix, and separate it out from
    # the rest.
    #import pdb; pdb.set_trace()
    m = _NUMERIC_PREFIX.match(result)
    if not m:
        prefix = '0.0.0'
        suffix = result
    else:
        prefix = m.groups()[0].split('.')
        prefix = [int(i) for i in prefix]
        while len(prefix) < 3:
            prefix.append(0)
        if len(prefix) == 3:
            suffix = result[m.end():]
        else:
            suffix = '.'.join([str(i) for i in prefix[3:]]) + result[m.end():]
            prefix = prefix[:3]
        prefix = '.'.join([str(i) for i in prefix])
        suffix = suffix.strip()
    if suffix:
        #import pdb; pdb.set_trace()
        # massage the suffix.
        for pat, repl in _SUFFIX_REPLACEMENTS:
            suffix = pat.sub(repl, suffix)

    if not suffix:
        result = prefix
    else:
        sep = '-' if 'dev' in suffix else '+'
        result = prefix + sep + suffix
    if not is_semver(result):
        result = None
    return result


def _suggest_normalized_version(s):
    """Suggest a normalized version close to the given version string.

    If you have a version string that isn't rational (i.e. NormalizedVersion
    doesn't like it) then you might be able to get an equivalent (or close)
    rational version from this function.

    This does a number of simple normalizations to the given string, based
    on observation of versions currently in use on PyPI. Given a dump of
    those version during PyCon 2009, 4287 of them:
    - 2312 (53.93%) match NormalizedVersion without change
      with the automatic suggestion
    - 3474 (81.04%) match when using this suggestion method

    @param s {str} An irrational version string.
    @returns A rational version string, or None, if couldn't determine one.
    """
    try:
        _normalized_key(s)
        return s   # already rational
    except UnsupportedVersionError:
        pass

    rs = s.lower()

    # part of this could use maketrans
    for orig, repl in (('-alpha', 'a'), ('-beta', 'b'), ('alpha', 'a'),
                       ('beta', 'b'), ('rc', 'c'), ('-final', ''),
                       ('-pre', 'c'),
                       ('-release', ''), ('.release', ''), ('-stable', ''),
                       ('+', '.'), ('_', '.'), (' ', ''), ('.final', ''),
                       ('final', '')):
        rs = rs.replace(orig, repl)

    # if something ends with dev or pre, we add a 0
    rs = re.sub(r"pre$", r"pre0", rs)
    rs = re.sub(r"dev$", r"dev0", rs)

    # if we have something like "b-2" or "a.2" at the end of the
    # version, that is probably beta, alpha, etc
    # let's remove the dash or dot
    rs = re.sub(r"([abc]|rc)[\-\.](\d+)$", r"\1\2", rs)

    # 1.0-dev-r371 -> 1.0.dev371
    # 0.1-dev-r79 -> 0.1.dev79
    rs = re.sub(r"[\-\.](dev)[\-\.]?r?(\d+)$", r".\1\2", rs)

    # Clean: 2.0.a.3, 2.0.b1, 0.9.0~c1
    rs = re.sub(r"[.~]?([abc])\.?", r"\1", rs)

    # Clean: v0.3, v1.0
    if rs.startswith('v'):
        rs = rs[1:]

    # Clean leading '0's on numbers.
    #TODO: unintended side-effect on, e.g., "2003.05.09"
    # PyPI stats: 77 (~2%) better
    rs = re.sub(r"\b0+(\d+)(?!\d)", r"\1", rs)

    # Clean a/b/c with no version. E.g. "1.0a" -> "1.0a0". Setuptools infers
    # zero.
    # PyPI stats: 245 (7.56%) better
    rs = re.sub(r"(\d+[abc])$", r"\g<1>0", rs)

    # the 'dev-rNNN' tag is a dev tag
    rs = re.sub(r"\.?(dev-r|dev\.r)\.?(\d+)$", r".dev\2", rs)

    # clean the - when used as a pre delimiter
    rs = re.sub(r"-(a|b|c)(\d+)$", r"\1\2", rs)

    # a terminal "dev" or "devel" can be changed into ".dev0"
    rs = re.sub(r"[\.\-](dev|devel)$", r".dev0", rs)

    # a terminal "dev" can be changed into ".dev0"
    rs = re.sub(r"(?![\.\-])dev$", r".dev0", rs)

    # a terminal "final" or "stable" can be removed
    rs = re.sub(r"(final|stable)$", "", rs)

    # The 'r' and the '-' tags are post release tags
    #   0.4a1.r10       ->  0.4a1.post10
    #   0.9.33-17222    ->  0.9.33.post17222
    #   0.9.33-r17222   ->  0.9.33.post17222
    rs = re.sub(r"\.?(r|-|-r)\.?(\d+)$", r".post\2", rs)

    # Clean 'r' instead of 'dev' usage:
    #   0.9.33+r17222   ->  0.9.33.dev17222
    #   1.0dev123       ->  1.0.dev123
    #   1.0.git123      ->  1.0.dev123
    #   1.0.bzr123      ->  1.0.dev123
    #   0.1a0dev.123    ->  0.1a0.dev123
    # PyPI stats:  ~150 (~4%) better
    rs = re.sub(r"\.?(dev|git|bzr)\.?(\d+)$", r".dev\2", rs)

    # Clean '.pre' (normalized from '-pre' above) instead of 'c' usage:
    #   0.2.pre1        ->  0.2c1
    #   0.2-c1         ->  0.2c1
    #   1.0preview123   ->  1.0c123
    # PyPI stats: ~21 (0.62%) better
    rs = re.sub(r"\.?(pre|preview|-c)(\d+)$", r"c\g<2>", rs)

    # Tcl/Tk uses "px" for their post release markers
    rs = re.sub(r"p(\d+)$", r".post\1", rs)

    try:
        _normalized_key(rs)
    except UnsupportedVersionError:
        rs = None
    return rs

#
#   Legacy version processing (distribute-compatible)
#

_VERSION_PART = re.compile(r'([a-z]+|\d+|[\.-])', re.I)
_VERSION_REPLACE = {
    'pre': 'c',
    'preview': 'c',
    '-': 'final-',
    'rc': 'c',
    'dev': '@',
    '': None,
    '.': None,
}


def _legacy_key(s):
    def get_parts(s):
        result = []
        for p in _VERSION_PART.split(s.lower()):
            p = _VERSION_REPLACE.get(p, p)
            if p:
                if '0' <= p[:1] <= '9':
                    p = p.zfill(8)
                else:
                    p = '*' + p
                result.append(p)
        result.append('*final')
        return result

    result = []
    for p in get_parts(s):
        if p.startswith('*'):
            if p < '*final':
                while result and result[-1] == '*final-':
                    result.pop()
            while result and result[-1] == '00000000':
                result.pop()
        result.append(p)
    return tuple(result)


class LegacyVersion(Version):
    def parse(self, s):
        return _legacy_key(s)

    @property
    def is_prerelease(self):
        result = False
        for x in self._parts:
            if (isinstance(x, string_types) and x.startswith('*') and
                x < '*final'):
                result = True
                break
        return result


class LegacyMatcher(Matcher):
    version_class = LegacyVersion

    _operators = dict(Matcher._operators)
    _operators['~='] = '_match_compatible'

    numeric_re = re.compile(r'^(\d+(\.\d+)*)')

    def _match_compatible(self, version, constraint, prefix):
        if version < constraint:
            return False
        m = self.numeric_re.match(str(constraint))
        if not m:
            logger.warning('Cannot compute compatible match for version %s '
                           ' and constraint %s', version, constraint)
            return True
        s = m.groups()[0]
        if '.' in s:
            s = s.rsplit('.', 1)[0]
        return _match_prefix(version, s)

#
#   Semantic versioning
#

_SEMVER_RE = re.compile(r'^(\d+)\.(\d+)\.(\d+)'
                        r'(-[a-z0-9]+(\.[a-z0-9-]+)*)?'
                        r'(\+[a-z0-9]+(\.[a-z0-9-]+)*)?$', re.I)


def is_semver(s):
    return _SEMVER_RE.match(s)


def _semantic_key(s):
    def make_tuple(s, absent):
        if s is None:
            result = (absent,)
        else:
            parts = s[1:].split('.')
            # We can't compare ints and strings on Python 3, so fudge it
            # by zero-filling numeric values so simulate a numeric comparison
            result = tuple([p.zfill(8) if p.isdigit() else p for p in parts])
        return result

    m = is_semver(s)
    if not m:
        raise UnsupportedVersionError(s)
    groups = m.groups()
    major, minor, patch = [int(i) for i in groups[:3]]
    # choose the '|' and '*' so that versions sort correctly
    pre, build = make_tuple(groups[3], '|'), make_tuple(groups[5], '*')
    return (major, minor, patch), pre, build


class SemanticVersion(Version):
    def parse(self, s):
        return _semantic_key(s)

    @property
    def is_prerelease(self):
        return self._parts[1][0] != '|'


class SemanticMatcher(Matcher):
    version_class = SemanticVersion


class VersionScheme(object):
    def __init__(self, key, matcher, suggester=None):
        self.key = key
        self.matcher = matcher
        self.suggester = suggester

    def is_valid_version(self, s):
        try:
            self.matcher.version_class(s)
            result = True
        except UnsupportedVersionError:
            result = False
        return result

    def is_valid_matcher(self, s):
        try:
            self.matcher(s)
            result = True
        except UnsupportedVersionError:
            result = False
        return result

    def is_valid_constraint_list(self, s):
        """
        Used for processing some metadata fields
        """
        return self.is_valid_matcher('dummy_name (%s)' % s)

    def suggest(self, s):
        if self.suggester is None:
            result = None
        else:
            result = self.suggester(s)
        return result

_SCHEMES = {
    'normalized': VersionScheme(_normalized_key, NormalizedMatcher,
                                _suggest_normalized_version),
    'legacy': VersionScheme(_legacy_key, LegacyMatcher, lambda self, s: s),
    'semantic': VersionScheme(_semantic_key, SemanticMatcher,
                              _suggest_semantic_version),
}

_SCHEMES['default'] = _SCHEMES['normalized']


def get_scheme(name):
    if name not in _SCHEMES:
        raise ValueError('unknown scheme name: %r' % name)
    return _SCHEMES[name]
site-packages/pip/_vendor/distlib/scripts.py000064400000037642151464452710015230 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2015 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from io import BytesIO
import logging
import os
import re
import struct
import sys

from .compat import sysconfig, detect_encoding, ZipFile
from .resources import finder
from .util import (FileOperator, get_export_entry, convert_path,
                   get_executable, in_venv)

logger = logging.getLogger(__name__)

_DEFAULT_MANIFEST = '''


 

 
 
 
 
 
 
 
 
'''.strip()

# check if Python is called on the first line with this expression
FIRST_LINE_RE = re.compile(b'^#!.*pythonw?[0-9.]*([ \t].*)?$')
SCRIPT_TEMPLATE = r'''# -*- coding: utf-8 -*-
import re
import sys
from %(module)s import %(import_name)s
if __name__ == '__main__':
    sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
    sys.exit(%(func)s())
'''


def _enquote_executable(executable):
    if ' ' in executable:
        # make sure we quote only the executable in case of env
        # for example /usr/bin/env "/dir with spaces/bin/jython"
        # instead of "/usr/bin/env /dir with spaces/bin/jython"
        # otherwise whole
        if executable.startswith('/usr/bin/env '):
            env, _executable = executable.split(' ', 1)
            if ' ' in _executable and not _executable.startswith('"'):
                executable = '%s "%s"' % (env, _executable)
        else:
            if not executable.startswith('"'):
                executable = '"%s"' % executable
    return executable


class ScriptMaker(object):
    """
    A class to copy or create scripts from source scripts or callable
    specifications.
    """
    script_template = SCRIPT_TEMPLATE

    executable = None  # for shebangs

    def __init__(self, source_dir, target_dir, add_launchers=True,
                 dry_run=False, fileop=None):
        self.source_dir = source_dir
        self.target_dir = target_dir
        self.add_launchers = add_launchers
        self.force = False
        self.clobber = False
        # It only makes sense to set mode bits on POSIX.
        self.set_mode = (os.name == 'posix') or (os.name == 'java' and
                                                 os._name == 'posix')
        self.variants = set(('', 'X.Y'))
        self._fileop = fileop or FileOperator(dry_run)

        self._is_nt = os.name == 'nt' or (
            os.name == 'java' and os._name == 'nt')

    def _get_alternate_executable(self, executable, options):
        if options.get('gui', False) and self._is_nt:  # pragma: no cover
            dn, fn = os.path.split(executable)
            fn = fn.replace('python', 'pythonw')
            executable = os.path.join(dn, fn)
        return executable

    if sys.platform.startswith('java'):  # pragma: no cover
        def _is_shell(self, executable):
            """
            Determine if the specified executable is a script
            (contains a #! line)
            """
            try:
                with open(executable) as fp:
                    return fp.read(2) == '#!'
            except (OSError, IOError):
                logger.warning('Failed to open %s', executable)
                return False

        def _fix_jython_executable(self, executable):
            if self._is_shell(executable):
                # Workaround for Jython is not needed on Linux systems.
                import java

                if java.lang.System.getProperty('os.name') == 'Linux':
                    return executable
            elif executable.lower().endswith('jython.exe'):
                # Use wrapper exe for Jython on Windows
                return executable
            return '/usr/bin/env %s' % executable

    def _build_shebang(self, executable, post_interp):
        """
        Build a shebang line. In the simple case (on Windows, or a shebang line
        which is not too long or contains spaces) use a simple formulation for
        the shebang. Otherwise, use /bin/sh as the executable, with a contrived
        shebang which allows the script to run either under Python or sh, using
        suitable quoting. Thanks to Harald Nordgren for his input.

        See also: http://www.in-ulm.de/~mascheck/various/shebang/#length
                  https://hg.mozilla.org/mozilla-central/file/tip/mach
        """
        if os.name != 'posix':
            simple_shebang = True
        else:
            # Add 3 for '#!' prefix and newline suffix.
            shebang_length = len(executable) + len(post_interp) + 3
            if sys.platform == 'darwin':
                max_shebang_length = 512
            else:
                max_shebang_length = 127
            simple_shebang = ((b' ' not in executable) and
                              (shebang_length <= max_shebang_length))

        if simple_shebang:
            result = b'#!' + executable + post_interp + b'\n'
        else:
            result = b'#!/bin/sh\n'
            result += b"'''exec' " + executable + post_interp + b' "$0" "$@"\n'
            result += b"' '''"
        return result

    def _get_shebang(self, encoding, post_interp=b'', options=None):
        enquote = True
        if self.executable:
            executable = self.executable
            enquote = False     # assume this will be taken care of
        elif not sysconfig.is_python_build():
            executable = get_executable()
        elif in_venv():  # pragma: no cover
            executable = os.path.join(sysconfig.get_path('scripts'),
                            'python%s' % sysconfig.get_config_var('EXE'))
        else:  # pragma: no cover
            executable = os.path.join(
                sysconfig.get_config_var('BINDIR'),
               'python%s%s' % (sysconfig.get_config_var('VERSION'),
                               sysconfig.get_config_var('EXE')))
        if options:
            executable = self._get_alternate_executable(executable, options)

        if sys.platform.startswith('java'):  # pragma: no cover
            executable = self._fix_jython_executable(executable)
        # Normalise case for Windows
        executable = os.path.normcase(executable)
        # If the user didn't specify an executable, it may be necessary to
        # cater for executable paths with spaces (not uncommon on Windows)
        if enquote:
            executable = _enquote_executable(executable)
        # Issue #51: don't use fsencode, since we later try to
        # check that the shebang is decodable using utf-8.
        executable = executable.encode('utf-8')
        # in case of IronPython, play safe and enable frames support
        if (sys.platform == 'cli' and '-X:Frames' not in post_interp
            and '-X:FullFrames' not in post_interp):  # pragma: no cover
            post_interp += b' -X:Frames'
        shebang = self._build_shebang(executable, post_interp)
        # Python parser starts to read a script using UTF-8 until
        # it gets a #coding:xxx cookie. The shebang has to be the
        # first line of a file, the #coding:xxx cookie cannot be
        # written before. So the shebang has to be decodable from
        # UTF-8.
        try:
            shebang.decode('utf-8')
        except UnicodeDecodeError:  # pragma: no cover
            raise ValueError(
                'The shebang (%r) is not decodable from utf-8' % shebang)
        # If the script is encoded to a custom encoding (use a
        # #coding:xxx cookie), the shebang has to be decodable from
        # the script encoding too.
        if encoding != 'utf-8':
            try:
                shebang.decode(encoding)
            except UnicodeDecodeError:  # pragma: no cover
                raise ValueError(
                    'The shebang (%r) is not decodable '
                    'from the script encoding (%r)' % (shebang, encoding))
        return shebang

    def _get_script_text(self, entry):
        return self.script_template % dict(module=entry.prefix,
                                           import_name=entry.suffix.split('.')[0],
                                           func=entry.suffix)

    manifest = _DEFAULT_MANIFEST

    def get_manifest(self, exename):
        base = os.path.basename(exename)
        return self.manifest % base

    def _write_script(self, names, shebang, script_bytes, filenames, ext):
        use_launcher = self.add_launchers and self._is_nt
        linesep = os.linesep.encode('utf-8')
        if not shebang.endswith(linesep):
            shebang += linesep
        if not use_launcher:
            script_bytes = shebang + script_bytes
        else:  # pragma: no cover
            if ext == 'py':
                launcher = self._get_launcher('t')
            else:
                launcher = self._get_launcher('w')
            stream = BytesIO()
            with ZipFile(stream, 'w') as zf:
                zf.writestr('__main__.py', script_bytes)
            zip_data = stream.getvalue()
            script_bytes = launcher + shebang + zip_data
        for name in names:
            outname = os.path.join(self.target_dir, name)
            if use_launcher:  # pragma: no cover
                n, e = os.path.splitext(outname)
                if e.startswith('.py'):
                    outname = n
                outname = '%s.exe' % outname
                try:
                    self._fileop.write_binary_file(outname, script_bytes)
                except Exception:
                    # Failed writing an executable - it might be in use.
                    logger.warning('Failed to write executable - trying to '
                                   'use .deleteme logic')
                    dfname = '%s.deleteme' % outname
                    if os.path.exists(dfname):
                        os.remove(dfname)       # Not allowed to fail here
                    os.rename(outname, dfname)  # nor here
                    self._fileop.write_binary_file(outname, script_bytes)
                    logger.debug('Able to replace executable using '
                                 '.deleteme logic')
                    try:
                        os.remove(dfname)
                    except Exception:
                        pass    # still in use - ignore error
            else:
                if self._is_nt and not outname.endswith('.' + ext):  # pragma: no cover
                    outname = '%s.%s' % (outname, ext)
                if os.path.exists(outname) and not self.clobber:
                    logger.warning('Skipping existing file %s', outname)
                    continue
                self._fileop.write_binary_file(outname, script_bytes)
                if self.set_mode:
                    self._fileop.set_executable_mode([outname])
            filenames.append(outname)

    def _make_script(self, entry, filenames, options=None):
        post_interp = b''
        if options:
            args = options.get('interpreter_args', [])
            if args:
                args = ' %s' % ' '.join(args)
                post_interp = args.encode('utf-8')
        shebang = self._get_shebang('utf-8', post_interp, options=options)
        script = self._get_script_text(entry).encode('utf-8')
        name = entry.name
        scriptnames = set()
        if '' in self.variants:
            scriptnames.add(name)
        if 'X' in self.variants:
            scriptnames.add('%s%s' % (name, sys.version[0]))
        if 'X.Y' in self.variants:
            scriptnames.add('%s-%s' % (name, sys.version[:3]))
        if options and options.get('gui', False):
            ext = 'pyw'
        else:
            ext = 'py'
        self._write_script(scriptnames, shebang, script, filenames, ext)

    def _copy_script(self, script, filenames):
        adjust = False
        script = os.path.join(self.source_dir, convert_path(script))
        outname = os.path.join(self.target_dir, os.path.basename(script))
        if not self.force and not self._fileop.newer(script, outname):
            logger.debug('not copying %s (up-to-date)', script)
            return

        # Always open the file, but ignore failures in dry-run mode --
        # that way, we'll get accurate feedback if we can read the
        # script.
        try:
            f = open(script, 'rb')
        except IOError:  # pragma: no cover
            if not self.dry_run:
                raise
            f = None
        else:
            first_line = f.readline()
            if not first_line:  # pragma: no cover
                logger.warning('%s: %s is an empty file (skipping)',
                               self.get_command_name(),  script)
                return

            match = FIRST_LINE_RE.match(first_line.replace(b'\r\n', b'\n'))
            if match:
                adjust = True
                post_interp = match.group(1) or b''

        if not adjust:
            if f:
                f.close()
            self._fileop.copy_file(script, outname)
            if self.set_mode:
                self._fileop.set_executable_mode([outname])
            filenames.append(outname)
        else:
            logger.info('copying and adjusting %s -> %s', script,
                        self.target_dir)
            if not self._fileop.dry_run:
                encoding, lines = detect_encoding(f.readline)
                f.seek(0)
                shebang = self._get_shebang(encoding, post_interp)
                if b'pythonw' in first_line:  # pragma: no cover
                    ext = 'pyw'
                else:
                    ext = 'py'
                n = os.path.basename(outname)
                self._write_script([n], shebang, f.read(), filenames, ext)
            if f:
                f.close()

    @property
    def dry_run(self):
        return self._fileop.dry_run

    @dry_run.setter
    def dry_run(self, value):
        self._fileop.dry_run = value

    if os.name == 'nt' or (os.name == 'java' and os._name == 'nt'):  # pragma: no cover
        # Executable launcher support.
        # Launchers are from https://bitbucket.org/vinay.sajip/simple_launcher/

        def _get_launcher(self, kind):
            if struct.calcsize('P') == 8:   # 64-bit
                bits = '64'
            else:
                bits = '32'
            name = '%s%s.exe' % (kind, bits)
            # Issue 31: don't hardcode an absolute package name, but
            # determine it relative to the current package
            distlib_package = __name__.rsplit('.', 1)[0]
            result = finder(distlib_package).find(name).bytes
            return result

    # Public API follows

    def make(self, specification, options=None):
        """
        Make a script.

        :param specification: The specification, which is either a valid export
                              entry specification (to make a script from a
                              callable) or a filename (to make a script by
                              copying from a source location).
        :param options: A dictionary of options controlling script generation.
        :return: A list of all absolute pathnames written to.
        """
        filenames = []
        entry = get_export_entry(specification)
        if entry is None:
            self._copy_script(specification, filenames)
        else:
            self._make_script(entry, filenames, options=options)
        return filenames

    def make_multiple(self, specifications, options=None):
        """
        Take a list of specifications and make scripts from them,
        :param specifications: A list of specifications.
        :return: A list of all absolute pathnames written to,
        """
        filenames = []
        for specification in specifications:
            filenames.extend(self.make(specification, options))
        return filenames
site-packages/pip/_vendor/distlib/database.py000064400000143525151464452710015303 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2017 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""PEP 376 implementation."""

from __future__ import unicode_literals

import base64
import codecs
import contextlib
import hashlib
import logging
import os
import posixpath
import sys
import zipimport

from . import DistlibException, resources
from .compat import StringIO
from .version import get_scheme, UnsupportedVersionError
from .metadata import (Metadata, METADATA_FILENAME, WHEEL_METADATA_FILENAME,
                       LEGACY_METADATA_FILENAME)
from .util import (parse_requirement, cached_property, parse_name_and_version,
                   read_exports, write_exports, CSVReader, CSVWriter)


__all__ = ['Distribution', 'BaseInstalledDistribution',
           'InstalledDistribution', 'EggInfoDistribution',
           'DistributionPath']


logger = logging.getLogger(__name__)

EXPORTS_FILENAME = 'pydist-exports.json'
COMMANDS_FILENAME = 'pydist-commands.json'

DIST_FILES = ('INSTALLER', METADATA_FILENAME, 'RECORD', 'REQUESTED',
              'RESOURCES', EXPORTS_FILENAME, 'SHARED')

DISTINFO_EXT = '.dist-info'


class _Cache(object):
    """
    A simple cache mapping names and .dist-info paths to distributions
    """
    def __init__(self):
        """
        Initialise an instance. There is normally one for each DistributionPath.
        """
        self.name = {}
        self.path = {}
        self.generated = False

    def clear(self):
        """
        Clear the cache, setting it to its initial state.
        """
        self.name.clear()
        self.path.clear()
        self.generated = False

    def add(self, dist):
        """
        Add a distribution to the cache.
        :param dist: The distribution to add.
        """
        if dist.path not in self.path:
            self.path[dist.path] = dist
            self.name.setdefault(dist.key, []).append(dist)


class DistributionPath(object):
    """
    Represents a set of distributions installed on a path (typically sys.path).
    """
    def __init__(self, path=None, include_egg=False):
        """
        Create an instance from a path, optionally including legacy (distutils/
        setuptools/distribute) distributions.
        :param path: The path to use, as a list of directories. If not specified,
                     sys.path is used.
        :param include_egg: If True, this instance will look for and return legacy
                            distributions as well as those based on PEP 376.
        """
        if path is None:
            path = sys.path
        self.path = path
        self._include_dist = True
        self._include_egg = include_egg

        self._cache = _Cache()
        self._cache_egg = _Cache()
        self._cache_enabled = True
        self._scheme = get_scheme('default')

    def _get_cache_enabled(self):
        return self._cache_enabled

    def _set_cache_enabled(self, value):
        self._cache_enabled = value

    cache_enabled = property(_get_cache_enabled, _set_cache_enabled)

    def clear_cache(self):
        """
        Clears the internal cache.
        """
        self._cache.clear()
        self._cache_egg.clear()


    def _yield_distributions(self):
        """
        Yield .dist-info and/or .egg(-info) distributions.
        """
        # We need to check if we've seen some resources already, because on
        # some Linux systems (e.g. some Debian/Ubuntu variants) there are
        # symlinks which alias other files in the environment.
        seen = set()
        for path in self.path:
            finder = resources.finder_for_path(path)
            if finder is None:
                continue
            r = finder.find('')
            if not r or not r.is_container:
                continue
            rset = sorted(r.resources)
            for entry in rset:
                r = finder.find(entry)
                if not r or r.path in seen:
                    continue
                if self._include_dist and entry.endswith(DISTINFO_EXT):
                    possible_filenames = [METADATA_FILENAME,
                                          WHEEL_METADATA_FILENAME,
                                          LEGACY_METADATA_FILENAME]
                    for metadata_filename in possible_filenames:
                        metadata_path = posixpath.join(entry, metadata_filename)
                        pydist = finder.find(metadata_path)
                        if pydist:
                            break
                    else:
                        continue

                    with contextlib.closing(pydist.as_stream()) as stream:
                        metadata = Metadata(fileobj=stream, scheme='legacy')
                    logger.debug('Found %s', r.path)
                    seen.add(r.path)
                    yield new_dist_class(r.path, metadata=metadata,
                                         env=self)
                elif self._include_egg and entry.endswith(('.egg-info',
                                                          '.egg')):
                    logger.debug('Found %s', r.path)
                    seen.add(r.path)
                    yield old_dist_class(r.path, self)

    def _generate_cache(self):
        """
        Scan the path for distributions and populate the cache with
        those that are found.
        """
        gen_dist = not self._cache.generated
        gen_egg = self._include_egg and not self._cache_egg.generated
        if gen_dist or gen_egg:
            for dist in self._yield_distributions():
                if isinstance(dist, InstalledDistribution):
                    self._cache.add(dist)
                else:
                    self._cache_egg.add(dist)

            if gen_dist:
                self._cache.generated = True
            if gen_egg:
                self._cache_egg.generated = True

    @classmethod
    def distinfo_dirname(cls, name, version):
        """
        The *name* and *version* parameters are converted into their
        filename-escaped form, i.e. any ``'-'`` characters are replaced
        with ``'_'`` other than the one in ``'dist-info'`` and the one
        separating the name from the version number.

        :parameter name: is converted to a standard distribution name by replacing
                         any runs of non- alphanumeric characters with a single
                         ``'-'``.
        :type name: string
        :parameter version: is converted to a standard version string. Spaces
                            become dots, and all other non-alphanumeric characters
                            (except dots) become dashes, with runs of multiple
                            dashes condensed to a single dash.
        :type version: string
        :returns: directory name
        :rtype: string"""
        name = name.replace('-', '_')
        return '-'.join([name, version]) + DISTINFO_EXT

    def get_distributions(self):
        """
        Provides an iterator that looks for distributions and returns
        :class:`InstalledDistribution` or
        :class:`EggInfoDistribution` instances for each one of them.

        :rtype: iterator of :class:`InstalledDistribution` and
                :class:`EggInfoDistribution` instances
        """
        if not self._cache_enabled:
            for dist in self._yield_distributions():
                yield dist
        else:
            self._generate_cache()

            for dist in self._cache.path.values():
                yield dist

            if self._include_egg:
                for dist in self._cache_egg.path.values():
                    yield dist

    def get_distribution(self, name):
        """
        Looks for a named distribution on the path.

        This function only returns the first result found, as no more than one
        value is expected. If nothing is found, ``None`` is returned.

        :rtype: :class:`InstalledDistribution`, :class:`EggInfoDistribution`
                or ``None``
        """
        result = None
        name = name.lower()
        if not self._cache_enabled:
            for dist in self._yield_distributions():
                if dist.key == name:
                    result = dist
                    break
        else:
            self._generate_cache()

            if name in self._cache.name:
                result = self._cache.name[name][0]
            elif self._include_egg and name in self._cache_egg.name:
                result = self._cache_egg.name[name][0]
        return result

    def provides_distribution(self, name, version=None):
        """
        Iterates over all distributions to find which distributions provide *name*.
        If a *version* is provided, it will be used to filter the results.

        This function only returns the first result found, since no more than
        one values are expected. If the directory is not found, returns ``None``.

        :parameter version: a version specifier that indicates the version
                            required, conforming to the format in ``PEP-345``

        :type name: string
        :type version: string
        """
        matcher = None
        if version is not None:
            try:
                matcher = self._scheme.matcher('%s (%s)' % (name, version))
            except ValueError:
                raise DistlibException('invalid name or version: %r, %r' %
                                      (name, version))

        for dist in self.get_distributions():
            # We hit a problem on Travis where enum34 was installed and doesn't
            # have a provides attribute ...
            if not hasattr(dist, 'provides'):
                logger.debug('No "provides": %s', dist)
            else:
                provided = dist.provides

                for p in provided:
                    p_name, p_ver = parse_name_and_version(p)
                    if matcher is None:
                        if p_name == name:
                            yield dist
                            break
                    else:
                        if p_name == name and matcher.match(p_ver):
                            yield dist
                            break

    def get_file_path(self, name, relative_path):
        """
        Return the path to a resource file.
        """
        dist = self.get_distribution(name)
        if dist is None:
            raise LookupError('no distribution named %r found' % name)
        return dist.get_resource_path(relative_path)

    def get_exported_entries(self, category, name=None):
        """
        Return all of the exported entries in a particular category.

        :param category: The category to search for entries.
        :param name: If specified, only entries with that name are returned.
        """
        for dist in self.get_distributions():
            r = dist.exports
            if category in r:
                d = r[category]
                if name is not None:
                    if name in d:
                        yield d[name]
                else:
                    for v in d.values():
                        yield v


class Distribution(object):
    """
    A base class for distributions, whether installed or from indexes.
    Either way, it must have some metadata, so that's all that's needed
    for construction.
    """

    build_time_dependency = False
    """
    Set to True if it's known to be only a build-time dependency (i.e.
    not needed after installation).
    """

    requested = False
    """A boolean that indicates whether the ``REQUESTED`` metadata file is
    present (in other words, whether the package was installed by user
    request or it was installed as a dependency)."""

    def __init__(self, metadata):
        """
        Initialise an instance.
        :param metadata: The instance of :class:`Metadata` describing this
        distribution.
        """
        self.metadata = metadata
        self.name = metadata.name
        self.key = self.name.lower()    # for case-insensitive comparisons
        self.version = metadata.version
        self.locator = None
        self.digest = None
        self.extras = None      # additional features requested
        self.context = None     # environment marker overrides
        self.download_urls = set()
        self.digests = {}

    @property
    def source_url(self):
        """
        The source archive download URL for this distribution.
        """
        return self.metadata.source_url

    download_url = source_url   # Backward compatibility

    @property
    def name_and_version(self):
        """
        A utility property which displays the name and version in parentheses.
        """
        return '%s (%s)' % (self.name, self.version)

    @property
    def provides(self):
        """
        A set of distribution names and versions provided by this distribution.
        :return: A set of "name (version)" strings.
        """
        plist = self.metadata.provides
        s = '%s (%s)' % (self.name, self.version)
        if s not in plist:
            plist.append(s)
        return plist

    def _get_requirements(self, req_attr):
        md = self.metadata
        logger.debug('Getting requirements from metadata %r', md.todict())
        reqts = getattr(md, req_attr)
        return set(md.get_requirements(reqts, extras=self.extras,
                                       env=self.context))

    @property
    def run_requires(self):
        return self._get_requirements('run_requires')

    @property
    def meta_requires(self):
        return self._get_requirements('meta_requires')

    @property
    def build_requires(self):
        return self._get_requirements('build_requires')

    @property
    def test_requires(self):
        return self._get_requirements('test_requires')

    @property
    def dev_requires(self):
        return self._get_requirements('dev_requires')

    def matches_requirement(self, req):
        """
        Say if this instance matches (fulfills) a requirement.
        :param req: The requirement to match.
        :rtype req: str
        :return: True if it matches, else False.
        """
        # Requirement may contain extras - parse to lose those
        # from what's passed to the matcher
        r = parse_requirement(req)
        scheme = get_scheme(self.metadata.scheme)
        try:
            matcher = scheme.matcher(r.requirement)
        except UnsupportedVersionError:
            # XXX compat-mode if cannot read the version
            logger.warning('could not read version %r - using name only',
                           req)
            name = req.split()[0]
            matcher = scheme.matcher(name)

        name = matcher.key   # case-insensitive

        result = False
        for p in self.provides:
            p_name, p_ver = parse_name_and_version(p)
            if p_name != name:
                continue
            try:
                result = matcher.match(p_ver)
                break
            except UnsupportedVersionError:
                pass
        return result

    def __repr__(self):
        """
        Return a textual representation of this instance,
        """
        if self.source_url:
            suffix = ' [%s]' % self.source_url
        else:
            suffix = ''
        return '' % (self.name, self.version, suffix)

    def __eq__(self, other):
        """
        See if this distribution is the same as another.
        :param other: The distribution to compare with. To be equal to one
                      another. distributions must have the same type, name,
                      version and source_url.
        :return: True if it is the same, else False.
        """
        if type(other) is not type(self):
            result = False
        else:
            result = (self.name == other.name and
                      self.version == other.version and
                      self.source_url == other.source_url)
        return result

    def __hash__(self):
        """
        Compute hash in a way which matches the equality test.
        """
        return hash(self.name) + hash(self.version) + hash(self.source_url)


class BaseInstalledDistribution(Distribution):
    """
    This is the base class for installed distributions (whether PEP 376 or
    legacy).
    """

    hasher = None

    def __init__(self, metadata, path, env=None):
        """
        Initialise an instance.
        :param metadata: An instance of :class:`Metadata` which describes the
                         distribution. This will normally have been initialised
                         from a metadata file in the ``path``.
        :param path:     The path of the ``.dist-info`` or ``.egg-info``
                         directory for the distribution.
        :param env:      This is normally the :class:`DistributionPath`
                         instance where this distribution was found.
        """
        super(BaseInstalledDistribution, self).__init__(metadata)
        self.path = path
        self.dist_path = env

    def get_hash(self, data, hasher=None):
        """
        Get the hash of some data, using a particular hash algorithm, if
        specified.

        :param data: The data to be hashed.
        :type data: bytes
        :param hasher: The name of a hash implementation, supported by hashlib,
                       or ``None``. Examples of valid values are ``'sha1'``,
                       ``'sha224'``, ``'sha384'``, '``sha256'``, ``'md5'`` and
                       ``'sha512'``. If no hasher is specified, the ``hasher``
                       attribute of the :class:`InstalledDistribution` instance
                       is used. If the hasher is determined to be ``None``, MD5
                       is used as the hashing algorithm.
        :returns: The hash of the data. If a hasher was explicitly specified,
                  the returned hash will be prefixed with the specified hasher
                  followed by '='.
        :rtype: str
        """
        if hasher is None:
            hasher = self.hasher
        if hasher is None:
            hasher = hashlib.md5
            prefix = ''
        else:
            hasher = getattr(hashlib, hasher)
            prefix = '%s=' % self.hasher
        digest = hasher(data).digest()
        digest = base64.urlsafe_b64encode(digest).rstrip(b'=').decode('ascii')
        return '%s%s' % (prefix, digest)


class InstalledDistribution(BaseInstalledDistribution):
    """
    Created with the *path* of the ``.dist-info`` directory provided to the
    constructor. It reads the metadata contained in ``pydist.json`` when it is
    instantiated., or uses a passed in Metadata instance (useful for when
    dry-run mode is being used).
    """

    hasher = 'sha256'

    def __init__(self, path, metadata=None, env=None):
        self.modules = []
        self.finder = finder = resources.finder_for_path(path)
        if finder is None:
            raise ValueError('finder unavailable for %s' % path)
        if env and env._cache_enabled and path in env._cache.path:
            metadata = env._cache.path[path].metadata
        elif metadata is None:
            r = finder.find(METADATA_FILENAME)
            # Temporary - for Wheel 0.23 support
            if r is None:
                r = finder.find(WHEEL_METADATA_FILENAME)
            # Temporary - for legacy support
            if r is None:
                r = finder.find('METADATA')
            if r is None:
                raise ValueError('no %s found in %s' % (METADATA_FILENAME,
                                                        path))
            with contextlib.closing(r.as_stream()) as stream:
                metadata = Metadata(fileobj=stream, scheme='legacy')

        super(InstalledDistribution, self).__init__(metadata, path, env)

        if env and env._cache_enabled:
            env._cache.add(self)

        r = finder.find('REQUESTED')
        self.requested = r is not None
        p  = os.path.join(path, 'top_level.txt')
        if os.path.exists(p):
            with open(p, 'rb') as f:
                data = f.read()
            self.modules = data.splitlines()

    def __repr__(self):
        return '' % (
            self.name, self.version, self.path)

    def __str__(self):
        return "%s %s" % (self.name, self.version)

    def _get_records(self):
        """
        Get the list of installed files for the distribution
        :return: A list of tuples of path, hash and size. Note that hash and
                 size might be ``None`` for some entries. The path is exactly
                 as stored in the file (which is as in PEP 376).
        """
        results = []
        r = self.get_distinfo_resource('RECORD')
        with contextlib.closing(r.as_stream()) as stream:
            with CSVReader(stream=stream) as record_reader:
                # Base location is parent dir of .dist-info dir
                #base_location = os.path.dirname(self.path)
                #base_location = os.path.abspath(base_location)
                for row in record_reader:
                    missing = [None for i in range(len(row), 3)]
                    path, checksum, size = row + missing
                    #if not os.path.isabs(path):
                    #    path = path.replace('/', os.sep)
                    #    path = os.path.join(base_location, path)
                    results.append((path, checksum, size))
        return results

    @cached_property
    def exports(self):
        """
        Return the information exported by this distribution.
        :return: A dictionary of exports, mapping an export category to a dict
                 of :class:`ExportEntry` instances describing the individual
                 export entries, and keyed by name.
        """
        result = {}
        r = self.get_distinfo_resource(EXPORTS_FILENAME)
        if r:
            result = self.read_exports()
        return result

    def read_exports(self):
        """
        Read exports data from a file in .ini format.

        :return: A dictionary of exports, mapping an export category to a list
                 of :class:`ExportEntry` instances describing the individual
                 export entries.
        """
        result = {}
        r = self.get_distinfo_resource(EXPORTS_FILENAME)
        if r:
            with contextlib.closing(r.as_stream()) as stream:
                result = read_exports(stream)
        return result

    def write_exports(self, exports):
        """
        Write a dictionary of exports to a file in .ini format.
        :param exports: A dictionary of exports, mapping an export category to
                        a list of :class:`ExportEntry` instances describing the
                        individual export entries.
        """
        rf = self.get_distinfo_file(EXPORTS_FILENAME)
        with open(rf, 'w') as f:
            write_exports(exports, f)

    def get_resource_path(self, relative_path):
        """
        NOTE: This API may change in the future.

        Return the absolute path to a resource file with the given relative
        path.

        :param relative_path: The path, relative to .dist-info, of the resource
                              of interest.
        :return: The absolute path where the resource is to be found.
        """
        r = self.get_distinfo_resource('RESOURCES')
        with contextlib.closing(r.as_stream()) as stream:
            with CSVReader(stream=stream) as resources_reader:
                for relative, destination in resources_reader:
                    if relative == relative_path:
                        return destination
        raise KeyError('no resource file with relative path %r '
                       'is installed' % relative_path)

    def list_installed_files(self):
        """
        Iterates over the ``RECORD`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: iterator of (path, hash, size)
        """
        for result in self._get_records():
            yield result

    def write_installed_files(self, paths, prefix, dry_run=False):
        """
        Writes the ``RECORD`` file, using the ``paths`` iterable passed in. Any
        existing ``RECORD`` file is silently overwritten.

        prefix is used to determine when to write absolute paths.
        """
        prefix = os.path.join(prefix, '')
        base = os.path.dirname(self.path)
        base_under_prefix = base.startswith(prefix)
        base = os.path.join(base, '')
        record_path = self.get_distinfo_file('RECORD')
        logger.info('creating %s', record_path)
        if dry_run:
            return None
        with CSVWriter(record_path) as writer:
            for path in paths:
                if os.path.isdir(path) or path.endswith(('.pyc', '.pyo')):
                    # do not put size and hash, as in PEP-376
                    hash_value = size = ''
                else:
                    size = '%d' % os.path.getsize(path)
                    with open(path, 'rb') as fp:
                        hash_value = self.get_hash(fp.read())
                if path.startswith(base) or (base_under_prefix and
                                             path.startswith(prefix)):
                    path = os.path.relpath(path, base)
                writer.writerow((path, hash_value, size))

            # add the RECORD file itself
            if record_path.startswith(base):
                record_path = os.path.relpath(record_path, base)
            writer.writerow((record_path, '', ''))
        return record_path

    def check_installed_files(self):
        """
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        """
        mismatches = []
        base = os.path.dirname(self.path)
        record_path = self.get_distinfo_file('RECORD')
        for path, hash_value, size in self.list_installed_files():
            if not os.path.isabs(path):
                path = os.path.join(base, path)
            if path == record_path:
                continue
            if not os.path.exists(path):
                mismatches.append((path, 'exists', True, False))
            elif os.path.isfile(path):
                actual_size = str(os.path.getsize(path))
                if size and actual_size != size:
                    mismatches.append((path, 'size', size, actual_size))
                elif hash_value:
                    if '=' in hash_value:
                        hasher = hash_value.split('=', 1)[0]
                    else:
                        hasher = None

                    with open(path, 'rb') as f:
                        actual_hash = self.get_hash(f.read(), hasher)
                        if actual_hash != hash_value:
                            mismatches.append((path, 'hash', hash_value, actual_hash))
        return mismatches

    @cached_property
    def shared_locations(self):
        """
        A dictionary of shared locations whose keys are in the set 'prefix',
        'purelib', 'platlib', 'scripts', 'headers', 'data' and 'namespace'.
        The corresponding value is the absolute path of that category for
        this distribution, and takes into account any paths selected by the
        user at installation time (e.g. via command-line arguments). In the
        case of the 'namespace' key, this would be a list of absolute paths
        for the roots of namespace packages in this distribution.

        The first time this property is accessed, the relevant information is
        read from the SHARED file in the .dist-info directory.
        """
        result = {}
        shared_path = os.path.join(self.path, 'SHARED')
        if os.path.isfile(shared_path):
            with codecs.open(shared_path, 'r', encoding='utf-8') as f:
                lines = f.read().splitlines()
            for line in lines:
                key, value = line.split('=', 1)
                if key == 'namespace':
                    result.setdefault(key, []).append(value)
                else:
                    result[key] = value
        return result

    def write_shared_locations(self, paths, dry_run=False):
        """
        Write shared location information to the SHARED file in .dist-info.
        :param paths: A dictionary as described in the documentation for
        :meth:`shared_locations`.
        :param dry_run: If True, the action is logged but no file is actually
                        written.
        :return: The path of the file written to.
        """
        shared_path = os.path.join(self.path, 'SHARED')
        logger.info('creating %s', shared_path)
        if dry_run:
            return None
        lines = []
        for key in ('prefix', 'lib', 'headers', 'scripts', 'data'):
            path = paths[key]
            if os.path.isdir(paths[key]):
                lines.append('%s=%s' % (key,  path))
        for ns in paths.get('namespace', ()):
            lines.append('namespace=%s' % ns)

        with codecs.open(shared_path, 'w', encoding='utf-8') as f:
            f.write('\n'.join(lines))
        return shared_path

    def get_distinfo_resource(self, path):
        if path not in DIST_FILES:
            raise DistlibException('invalid path for a dist-info file: '
                                   '%r at %r' % (path, self.path))
        finder = resources.finder_for_path(self.path)
        if finder is None:
            raise DistlibException('Unable to get a finder for %s' % self.path)
        return finder.find(path)

    def get_distinfo_file(self, path):
        """
        Returns a path located under the ``.dist-info`` directory. Returns a
        string representing the path.

        :parameter path: a ``'/'``-separated path relative to the
                         ``.dist-info`` directory or an absolute path;
                         If *path* is an absolute path and doesn't start
                         with the ``.dist-info`` directory path,
                         a :class:`DistlibException` is raised
        :type path: str
        :rtype: str
        """
        # Check if it is an absolute path  # XXX use relpath, add tests
        if path.find(os.sep) >= 0:
            # it's an absolute path?
            distinfo_dirname, path = path.split(os.sep)[-2:]
            if distinfo_dirname != self.path.split(os.sep)[-1]:
                raise DistlibException(
                    'dist-info file %r does not belong to the %r %s '
                    'distribution' % (path, self.name, self.version))

        # The file must be relative
        if path not in DIST_FILES:
            raise DistlibException('invalid path for a dist-info file: '
                                   '%r at %r' % (path, self.path))

        return os.path.join(self.path, path)

    def list_distinfo_files(self):
        """
        Iterates over the ``RECORD`` entries and returns paths for each line if
        the path is pointing to a file located in the ``.dist-info`` directory
        or one of its subdirectories.

        :returns: iterator of paths
        """
        base = os.path.dirname(self.path)
        for path, checksum, size in self._get_records():
            # XXX add separator or use real relpath algo
            if not os.path.isabs(path):
                path = os.path.join(base, path)
            if path.startswith(self.path):
                yield path

    def __eq__(self, other):
        return (isinstance(other, InstalledDistribution) and
                self.path == other.path)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    __hash__ = object.__hash__


class EggInfoDistribution(BaseInstalledDistribution):
    """Created with the *path* of the ``.egg-info`` directory or file provided
    to the constructor. It reads the metadata contained in the file itself, or
    if the given path happens to be a directory, the metadata is read from the
    file ``PKG-INFO`` under that directory."""

    requested = True    # as we have no way of knowing, assume it was
    shared_locations = {}

    def __init__(self, path, env=None):
        def set_name_and_version(s, n, v):
            s.name = n
            s.key = n.lower()   # for case-insensitive comparisons
            s.version = v

        self.path = path
        self.dist_path = env
        if env and env._cache_enabled and path in env._cache_egg.path:
            metadata = env._cache_egg.path[path].metadata
            set_name_and_version(self, metadata.name, metadata.version)
        else:
            metadata = self._get_metadata(path)

            # Need to be set before caching
            set_name_and_version(self, metadata.name, metadata.version)

            if env and env._cache_enabled:
                env._cache_egg.add(self)
        super(EggInfoDistribution, self).__init__(metadata, path, env)

    def _get_metadata(self, path):
        requires = None

        def parse_requires_data(data):
            """Create a list of dependencies from a requires.txt file.

            *data*: the contents of a setuptools-produced requires.txt file.
            """
            reqs = []
            lines = data.splitlines()
            for line in lines:
                line = line.strip()
                if line.startswith('['):
                    logger.warning('Unexpected line: quitting requirement scan: %r',
                                   line)
                    break
                r = parse_requirement(line)
                if not r:
                    logger.warning('Not recognised as a requirement: %r', line)
                    continue
                if r.extras:
                    logger.warning('extra requirements in requires.txt are '
                                   'not supported')
                if not r.constraints:
                    reqs.append(r.name)
                else:
                    cons = ', '.join('%s%s' % c for c in r.constraints)
                    reqs.append('%s (%s)' % (r.name, cons))
            return reqs

        def parse_requires_path(req_path):
            """Create a list of dependencies from a requires.txt file.

            *req_path*: the path to a setuptools-produced requires.txt file.
            """

            reqs = []
            try:
                with codecs.open(req_path, 'r', 'utf-8') as fp:
                    reqs = parse_requires_data(fp.read())
            except IOError:
                pass
            return reqs

        tl_path = tl_data = None
        if path.endswith('.egg'):
            if os.path.isdir(path):
                p = os.path.join(path, 'EGG-INFO')
                meta_path = os.path.join(p, 'PKG-INFO')
                metadata = Metadata(path=meta_path, scheme='legacy')
                req_path = os.path.join(p, 'requires.txt')
                tl_path = os.path.join(p, 'top_level.txt')
                requires = parse_requires_path(req_path)
            else:
                # FIXME handle the case where zipfile is not available
                zipf = zipimport.zipimporter(path)
                fileobj = StringIO(
                    zipf.get_data('EGG-INFO/PKG-INFO').decode('utf8'))
                metadata = Metadata(fileobj=fileobj, scheme='legacy')
                try:
                    data = zipf.get_data('EGG-INFO/requires.txt')
                    tl_data = zipf.get_data('EGG-INFO/top_level.txt').decode('utf-8')
                    requires = parse_requires_data(data.decode('utf-8'))
                except IOError:
                    requires = None
        elif path.endswith('.egg-info'):
            if os.path.isdir(path):
                req_path = os.path.join(path, 'requires.txt')
                requires = parse_requires_path(req_path)
                path = os.path.join(path, 'PKG-INFO')
                tl_path = os.path.join(path, 'top_level.txt')
            metadata = Metadata(path=path, scheme='legacy')
        else:
            raise DistlibException('path must end with .egg-info or .egg, '
                                   'got %r' % path)

        if requires:
            metadata.add_requirements(requires)
        # look for top-level modules in top_level.txt, if present
        if tl_data is None:
            if tl_path is not None and os.path.exists(tl_path):
                with open(tl_path, 'rb') as f:
                    tl_data = f.read().decode('utf-8')
        if not tl_data:
            tl_data = []
        else:
            tl_data = tl_data.splitlines()
        self.modules = tl_data
        return metadata

    def __repr__(self):
        return '' % (
            self.name, self.version, self.path)

    def __str__(self):
        return "%s %s" % (self.name, self.version)

    def check_installed_files(self):
        """
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        """
        mismatches = []
        record_path = os.path.join(self.path, 'installed-files.txt')
        if os.path.exists(record_path):
            for path, _, _ in self.list_installed_files():
                if path == record_path:
                    continue
                if not os.path.exists(path):
                    mismatches.append((path, 'exists', True, False))
        return mismatches

    def list_installed_files(self):
        """
        Iterates over the ``installed-files.txt`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: a list of (path, hash, size)
        """

        def _md5(path):
            f = open(path, 'rb')
            try:
                content = f.read()
            finally:
                f.close()
            return hashlib.md5(content).hexdigest()

        def _size(path):
            return os.stat(path).st_size

        record_path = os.path.join(self.path, 'installed-files.txt')
        result = []
        if os.path.exists(record_path):
            with codecs.open(record_path, 'r', encoding='utf-8') as f:
                for line in f:
                    line = line.strip()
                    p = os.path.normpath(os.path.join(self.path, line))
                    # "./" is present as a marker between installed files
                    # and installation metadata files
                    if not os.path.exists(p):
                        logger.warning('Non-existent file: %s', p)
                        if p.endswith(('.pyc', '.pyo')):
                            continue
                        #otherwise fall through and fail
                    if not os.path.isdir(p):
                        result.append((p, _md5(p), _size(p)))
            result.append((record_path, None, None))
        return result

    def list_distinfo_files(self, absolute=False):
        """
        Iterates over the ``installed-files.txt`` entries and returns paths for
        each line if the path is pointing to a file located in the
        ``.egg-info`` directory or one of its subdirectories.

        :parameter absolute: If *absolute* is ``True``, each returned path is
                          transformed into a local absolute path. Otherwise the
                          raw value from ``installed-files.txt`` is returned.
        :type absolute: boolean
        :returns: iterator of paths
        """
        record_path = os.path.join(self.path, 'installed-files.txt')
        if os.path.exists(record_path):
            skip = True
            with codecs.open(record_path, 'r', encoding='utf-8') as f:
                for line in f:
                    line = line.strip()
                    if line == './':
                        skip = False
                        continue
                    if not skip:
                        p = os.path.normpath(os.path.join(self.path, line))
                        if p.startswith(self.path):
                            if absolute:
                                yield p
                            else:
                                yield line

    def __eq__(self, other):
        return (isinstance(other, EggInfoDistribution) and
                self.path == other.path)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    __hash__ = object.__hash__

new_dist_class = InstalledDistribution
old_dist_class = EggInfoDistribution


class DependencyGraph(object):
    """
    Represents a dependency graph between distributions.

    The dependency relationships are stored in an ``adjacency_list`` that maps
    distributions to a list of ``(other, label)`` tuples where  ``other``
    is a distribution and the edge is labeled with ``label`` (i.e. the version
    specifier, if such was provided). Also, for more efficient traversal, for
    every distribution ``x``, a list of predecessors is kept in
    ``reverse_list[x]``. An edge from distribution ``a`` to
    distribution ``b`` means that ``a`` depends on ``b``. If any missing
    dependencies are found, they are stored in ``missing``, which is a
    dictionary that maps distributions to a list of requirements that were not
    provided by any other distributions.
    """

    def __init__(self):
        self.adjacency_list = {}
        self.reverse_list = {}
        self.missing = {}

    def add_distribution(self, distribution):
        """Add the *distribution* to the graph.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        """
        self.adjacency_list[distribution] = []
        self.reverse_list[distribution] = []
        #self.missing[distribution] = []

    def add_edge(self, x, y, label=None):
        """Add an edge from distribution *x* to distribution *y* with the given
        *label*.

        :type x: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type y: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type label: ``str`` or ``None``
        """
        self.adjacency_list[x].append((y, label))
        # multiple edges are allowed, so be careful
        if x not in self.reverse_list[y]:
            self.reverse_list[y].append(x)

    def add_missing(self, distribution, requirement):
        """
        Add a missing *requirement* for the given *distribution*.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        :type requirement: ``str``
        """
        logger.debug('%s missing %r', distribution, requirement)
        self.missing.setdefault(distribution, []).append(requirement)

    def _repr_dist(self, dist):
        return '%s %s' % (dist.name, dist.version)

    def repr_node(self, dist, level=1):
        """Prints only a subgraph"""
        output = [self._repr_dist(dist)]
        for other, label in self.adjacency_list[dist]:
            dist = self._repr_dist(other)
            if label is not None:
                dist = '%s [%s]' % (dist, label)
            output.append('    ' * level + str(dist))
            suboutput = self.repr_node(other, level + 1)
            subs = suboutput.split('\n')
            output.extend(subs[1:])
        return '\n'.join(output)

    def to_dot(self, f, skip_disconnected=True):
        """Writes a DOT output for the graph to the provided file *f*.

        If *skip_disconnected* is set to ``True``, then all distributions
        that are not dependent on any other distribution are skipped.

        :type f: has to support ``file``-like operations
        :type skip_disconnected: ``bool``
        """
        disconnected = []

        f.write("digraph dependencies {\n")
        for dist, adjs in self.adjacency_list.items():
            if len(adjs) == 0 and not skip_disconnected:
                disconnected.append(dist)
            for other, label in adjs:
                if not label is None:
                    f.write('"%s" -> "%s" [label="%s"]\n' %
                            (dist.name, other.name, label))
                else:
                    f.write('"%s" -> "%s"\n' % (dist.name, other.name))
        if not skip_disconnected and len(disconnected) > 0:
            f.write('subgraph disconnected {\n')
            f.write('label = "Disconnected"\n')
            f.write('bgcolor = red\n')

            for dist in disconnected:
                f.write('"%s"' % dist.name)
                f.write('\n')
            f.write('}\n')
        f.write('}\n')

    def topological_sort(self):
        """
        Perform a topological sort of the graph.
        :return: A tuple, the first element of which is a topologically sorted
                 list of distributions, and the second element of which is a
                 list of distributions that cannot be sorted because they have
                 circular dependencies and so form a cycle.
        """
        result = []
        # Make a shallow copy of the adjacency list
        alist = {}
        for k, v in self.adjacency_list.items():
            alist[k] = v[:]
        while True:
            # See what we can remove in this run
            to_remove = []
            for k, v in list(alist.items())[:]:
                if not v:
                    to_remove.append(k)
                    del alist[k]
            if not to_remove:
                # What's left in alist (if anything) is a cycle.
                break
            # Remove from the adjacency list of others
            for k, v in alist.items():
                alist[k] = [(d, r) for d, r in v if d not in to_remove]
            logger.debug('Moving to result: %s',
                         ['%s (%s)' % (d.name, d.version) for d in to_remove])
            result.extend(to_remove)
        return result, list(alist.keys())

    def __repr__(self):
        """Representation of the graph"""
        output = []
        for dist, adjs in self.adjacency_list.items():
            output.append(self.repr_node(dist))
        return '\n'.join(output)


def make_graph(dists, scheme='default'):
    """Makes a dependency graph from the given distributions.

    :parameter dists: a list of distributions
    :type dists: list of :class:`distutils2.database.InstalledDistribution` and
                 :class:`distutils2.database.EggInfoDistribution` instances
    :rtype: a :class:`DependencyGraph` instance
    """
    scheme = get_scheme(scheme)
    graph = DependencyGraph()
    provided = {}  # maps names to lists of (version, dist) tuples

    # first, build the graph and find out what's provided
    for dist in dists:
        graph.add_distribution(dist)

        for p in dist.provides:
            name, version = parse_name_and_version(p)
            logger.debug('Add to provided: %s, %s, %s', name, version, dist)
            provided.setdefault(name, []).append((version, dist))

    # now make the edges
    for dist in dists:
        requires = (dist.run_requires | dist.meta_requires |
                    dist.build_requires | dist.dev_requires)
        for req in requires:
            try:
                matcher = scheme.matcher(req)
            except UnsupportedVersionError:
                # XXX compat-mode if cannot read the version
                logger.warning('could not read version %r - using name only',
                               req)
                name = req.split()[0]
                matcher = scheme.matcher(name)

            name = matcher.key   # case-insensitive

            matched = False
            if name in provided:
                for version, provider in provided[name]:
                    try:
                        match = matcher.match(version)
                    except UnsupportedVersionError:
                        match = False

                    if match:
                        graph.add_edge(dist, provider, req)
                        matched = True
                        break
            if not matched:
                graph.add_missing(dist, req)
    return graph


def get_dependent_dists(dists, dist):
    """Recursively generate a list of distributions from *dists* that are
    dependent on *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    """
    if dist not in dists:
        raise DistlibException('given distribution %r is not a member '
                               'of the list' % dist.name)
    graph = make_graph(dists)

    dep = [dist]  # dependent distributions
    todo = graph.reverse_list[dist]  # list of nodes we should inspect

    while todo:
        d = todo.pop()
        dep.append(d)
        for succ in graph.reverse_list[d]:
            if succ not in dep:
                todo.append(succ)

    dep.pop(0)  # remove dist from dep, was there to prevent infinite loops
    return dep


def get_required_dists(dists, dist):
    """Recursively generate a list of distributions from *dists* that are
    required by *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    """
    if dist not in dists:
        raise DistlibException('given distribution %r is not a member '
                               'of the list' % dist.name)
    graph = make_graph(dists)

    req = []  # required distributions
    todo = graph.adjacency_list[dist]  # list of nodes we should inspect

    while todo:
        d = todo.pop()[0]
        req.append(d)
        for pred in graph.adjacency_list[d]:
            if pred not in req:
                todo.append(pred)

    return req


def make_dist(name, version, **kwargs):
    """
    A convenience method for making a dist given just a name and version.
    """
    summary = kwargs.pop('summary', 'Placeholder for summary')
    md = Metadata(**kwargs)
    md.name = name
    md.version = version
    md.summary = summary or 'Placeholder for summary'
    return Distribution(md)
site-packages/pip/_vendor/distlib/__pycache__/locators.cpython-38.opt-1.pyc000064400000112527151464452710022610 0ustar00U

.e_@sDddlZddlmZddlZddlZddlZddlZddlZzddlZWne	k
rdddl
ZYnXddlZddlm
Z
ddlmZmZmZmZmZmZmZmZmZmZmZmZmZmZmZddlm Z m!Z!m"Z"ddl#m$Z$m%Z%ddl&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/dd	l0m1Z1m2Z2dd
l3m4Z4m5Z5e6e7Z8e9dZ:e9dej;Zd-ddZ?GdddeZ@GdddeAZBGdddeBZCGdddeBZDGdddeAZEGdddeBZFGdddeBZGGdd d eBZHGd!d"d"eBZIGd#d$d$eBZJeJeHeFd%d&d'd(d)ZKeKjLZLe9d*ZMGd+d,d,eAZNdS).N)BytesIO)DistlibException)urljoinurlparse
urlunparseurl2pathnamepathname2urlqueuequoteunescapestring_typesbuild_openerHTTPRedirectHandler	text_typeRequest	HTTPErrorURLError)DistributionDistributionPath	make_dist)MetadataMetadataInvalidError)	cached_propertyparse_credentialsensure_slashsplit_filenameget_project_dataparse_requirementparse_name_and_versionServerProxynormalize_name)
get_schemeUnsupportedVersionError)Wheel
is_compatiblez^(\w+)=([a-f0-9]+)z;\s*charset\s*=\s*(.*)\s*$ztext/html|application/x(ht)?mlzhttps://pypi.org/pypicCs6|dkrt}t|dd}z|WS|dXdS)z
    Return all distribution names known by an index.
    :param url: The URL of the index.
    :return: A list of all known distribution names.
    N@timeoutclose)
DEFAULT_INDEXr 
list_packages)urlclientr.@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/locators.pyget_all_distribution_names)sr0c@s$eZdZdZddZeZZZdS)RedirectHandlerzE
    A class to work around a bug in some Python 3.2.x releases.
    c	Csd}dD]}||kr||}q"q|dkr.dSt|}|jdkrnt||}t|drf|||n|||<t||||||S)N)locationZurireplace_header)rschemerZget_full_urlhasattrr4BaseRedirectHandlerhttp_error_302)	selfreqfpcodemsgheadersZnewurlkeyZurlpartsr.r.r/r8@s 

zRedirectHandler.http_error_302N)__name__
__module____qualname____doc__r8Zhttp_error_301Zhttp_error_303Zhttp_error_307r.r.r.r/r17sr1c@seZdZdZdZdZdZdZedZd)dd	Z	d
dZ
dd
ZddZddZ
ddZee
eZddZddZddZddZddZddZd d!Zd"d#Zd$d%Zd*d'd(ZdS)+LocatorzG
    A base class for locators - things that locate distributions.
    )z.tar.gzz.tar.bz2z.tarz.zipz.tgzz.tbz)z.eggz.exe.whl)z.pdfN)rEdefaultcCs,i|_||_tt|_d|_t|_dS)a^
        Initialise an instance.
        :param scheme: Because locators look for most recent versions, they
                       need to know the version scheme to use. This specifies
                       the current PEP-recommended scheme - use ``'legacy'``
                       if you need to support existing distributions on PyPI.
        N)	_cacher5rr1openermatcherr
Queueerrors)r9r5r.r.r/__init__fs
zLocator.__init__cCsVg}|jsRz|jd}||Wn|jjk
rDYqYnX|jq|S)z8
        Return any errors which have occurred.
        F)rKemptygetappendZEmpty	task_done)r9resulter.r.r/
get_errorsys

zLocator.get_errorscCs|dS)z>
        Clear any errors which may have been logged.
        N)rSr9r.r.r/clear_errorsszLocator.clear_errorscCs|jdSN)rGclearrTr.r.r/clear_cacheszLocator.clear_cachecCs|jSrV_schemerTr.r.r/_get_schemeszLocator._get_schemecCs
||_dSrVrY)r9valuer.r.r/_set_schemeszLocator._set_schemecCstddS)a=
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This should be implemented in subclasses.

        If called from a locate() request, self.matcher will be set to a
        matcher for the requirement to satisfy, otherwise it will be None.
         Please implement in the subclassNNotImplementedError)r9namer.r.r/_get_projects
zLocator._get_projectcCstddS)J
        Return all the distribution names known to this locator.
        r^Nr_rTr.r.r/get_distribution_namesszLocator.get_distribution_namescCsL|jdkr||}n2||jkr,|j|}n|||}||j|<|S)z
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This calls _get_project to do all the work, and just implements a caching layer on top.
        N)rGrbrU)r9rarQr.r.r/get_projects



zLocator.get_projectcCs^t|}t|j}d}|d}||j}|rBtt||j}|j	dkd|j
k||||fS)zu
        Give an url a score which can be used to choose preferred URLs
        for a given project release.
        TrEhttpszpypi.org)r	posixpathbasenamepathendswithdownloadable_extensionsr%r$
wheel_tagsr5netloc)r9r,trhZ
compatibleZis_wheelZis_downloadabler.r.r/	score_urls
zLocator.score_urlcCsR|}|rN||}||}||kr(|}||kr@td||ntd|||S)a{
        Choose one of two URLs where both are candidates for distribution
        archives for the same version of a distribution (for example,
        .tar.gz vs. zip).

        The current implementation favours https:// URLs over http://, archives
        from PyPI over those from other locations, wheel compatibility (if a
        wheel) and then the archive name.
        zNot replacing %r with %rzReplacing %r with %r)rologgerdebug)r9url1url2rQs1s2r.r.r/
prefer_urls


zLocator.prefer_urlcCs
t||S)zZ
        Attempt to split a filename in project name, version and Python version.
        )r)r9filenameproject_namer.r.r/rszLocator.split_filenamecCsdd}d}t|\}}}}}	}
|
drz~t	|}t
||jstd
|nX|dkrd}n||j|}|r|j|j
|jt|||||	dfd
dd|jDd}Wn0tk
r:}ztd|W5d}~XYnXn||jsZtd|nt|}}|jD]}||rn|dt|}|||}|std|nH|\}}}|r|||r|||t|||||	dfd}|r||d<qqn|r|r|
|d|<|S)a
        See if a URL is a candidate for a download URL for a project (the URL
        has typically been scraped from an HTML page).

        If it is, a dictionary is returned with keys "name", "version",
        "filename" and "url"; otherwise, None is returned.
        cSst|t|kSrV)r!)Zname1Zname2r.r.r/same_projectsz:Locator.convert_url_to_download_info..same_projectNzegg=z %s: version hint in fragment: %r)NN/rEzWheel not compatible: %sTr3z, cSs"g|]}dt|ddqS).N)joinlist).0vr.r.r/
sz8Locator.convert_url_to_download_info..)raversionrwr,python-versionzinvalid path for wheel: %szNot downloadable: %sz No match for project/version: %s)rarrwr,r	%s_digest)rlower
startswithrprqHASHER_HASHmatchgroupsrjr$r%rlrarrwrr~pyver	Exceptionwarningrkrgrhlenr)r9r,rxryrQr5rmriparamsqueryfragmalgodigestZorigpathwheelZincluderRrwZextrnrarrr.r.r/convert_url_to_download_infos

	 


z$Locator.convert_url_to_download_infocCs2d}dD]$}d|}||kr|||f}q.q|S)z
        Get a digest from a dictionary by looking at keys of the form
        'algo_digest'.

        Returns a 2-tuple (algo, digest) if found, else None. Currently
        looks only for SHA256, then MD5.
        N)Zsha256md5rr.)r9inforQrr?r.r.r/_get_digest1szLocator._get_digestc	Cs|d}|d}||kr,||}|j}nt|||jd}|j}|||_}|d}||d|<|j|dkr||j||_|d|t	
|||_|||<dS)z
        Update a result dictionary (the final result from _get_project) with a
        dictionary for a specific version, which typically holds information
        gleaned from a filename or URL for an archive for the distribution.
        rarr5r,digestsurlsN)popmetadatarr5rr
source_urlrv
setdefaultsetaddlocator)	r9rQrrardistmdrr,r.r.r/_update_version_dataAs

zLocator._update_version_dataFc	Csd}t|}|dkr td|t|j}||j|_}td|t|j	|
|j}t|dkr2g}|j
}	|D]z}
|
dkrqxzH||
std||
n*|s|	|
js||
ntd|
|jWqxtk
rtd||
YqxXqxt|d	krt||jd
}|r2td||d}||}|r|jrH|j|_|d
i|t|_i}|di}
|jD]}||
krv|
|||<qv||_d|_|S)a
        Find the most recent distribution which matches the given
        requirement.

        :param requirement: A requirement of the form 'foo (1.0)' or perhaps
                            'foo (>= 1.0, < 2.0, != 1.3)'
        :param prereleases: If ``True``, allow pre-release versions
                            to be located. Otherwise, pre-release versions
                            are not returned.
        :return: A :class:`Distribution` instance, or ``None`` if no such
                 distribution could be located.
        NzNot a valid requirement: %rzmatcher: %s (%s)r}rrz%s did not match %rz%skipping pre-release version %s of %szerror matching %s with %rr)r?zsorted list: %srzrr)rrr"r5rIrequirementrprqtyper@rerarZ
version_classrZ
is_prereleaserOrrsortedr?ZextrasrNr
download_urlsr)r9rprereleasesrQrr5rIversionsZslistZvclskrdZsdr,r.r.r/locateXsX




zLocator.locate)rF)F)r@rArBrCsource_extensionsbinary_extensionsexcluded_extensionsrlrkrLrSrUrXr[r]propertyr5rbrdrerorvrrrrrr.r.r.r/rDVs.

JrDcs0eZdZdZfddZddZddZZS)PyPIRPCLocatorz
    This locator uses XML-RPC to locate distributions. It therefore
    cannot be used with simple mirrors (that only mirror file content).
    cs*tt|jf|||_t|dd|_dS)z
        Initialise an instance.

        :param url: The URL to use for XML-RPC.
        :param kwargs: Passed to the superclass constructor.
        r&r'N)superrrLbase_urlr r-r9r,kwargs	__class__r.r/rLszPyPIRPCLocator.__init__cCst|jSrc)rr-r+rTr.r.r/rdsz%PyPIRPCLocator.get_distribution_namescCsiid}|j|d}|D]}|j||}|j||}t|jd}|d|_|d|_|d|_	|dg|_
|d|_t|}|r|d	}	|	d
|_
||	|_||_|||<|D]:}	|	d
}
||	}|d|t|
||d|
<qq|S)
NrTrrarlicensekeywordssummaryrr,rr)r-Zpackage_releasesZrelease_urlsZrelease_datarr5rarrNrrrrrrrrrrr)r9rarQrrrdatarrrr,rr.r.r/rbs0




zPyPIRPCLocator._get_projectr@rArBrCrLrdrb
__classcell__r.r.rr/rsrcs0eZdZdZfddZddZddZZS)PyPIJSONLocatorzw
    This locator uses PyPI's JSON interface. It's very limited in functionality
    and probably not worth using.
    cs tt|jf|t||_dSrV)rrrLrrrrr.r/rLszPyPIJSONLocator.__init__cCstddSrczNot available from this locatorNr_rTr.r.r/rdsz&PyPIJSONLocator.get_distribution_namesc
Cs
iid}t|jdt|}z|j|}|}t|}t	|j
d}|d}|d|_|d|_|
d|_|
dg|_|
d	|_t|}||_|d
}	|||j<|d
D]T}
|
d}|j|||
|j|<|d
|jt|||
|d|<q|d
D]\}}||jkr4qt	|j
d}
|j|
_||
_t|
}||_|||<|D]T}
|
d}|j|||
|j|<|d
|t|||
|d|<qhqWn@tk
r}z |jt|td|W5d}~XYnX|S)Nrz%s/jsonrrrarrrrrr,rZreleaseszJSON fetch failed: %s) rrrrHopenreaddecodejsonloadsrr5rarrNrrrrrrrrrrritemsrrKputrrp	exception)r9rarQr,resprrrrrrrZinfosZomdodistrRr.r.r/rbsT




	zPyPIJSONLocator._get_projectrr.r.rr/rsrc@s`eZdZdZedejejBejBZ	edejejBZ
ddZedejZe
ddZd	S)
Pagez4
    This class represents a scraped HTML page.
    z
(rel\s*=\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\s
]*))\s+)?
href\s*=\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\s
]*))
(\s+rel\s*=\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\s
]*)))?
z!]+)cCs4||_||_|_|j|j}|r0|d|_dS)zk
        Initialise an instance with the Unicode page contents and the URL they
        came from.
        rN)rrr,_basesearchgroup)r9rr,rr.r.r/rLs
z
Page.__init__z[^a-z0-9$&+,/:;=?@.#%_\\|-]cCsdd}t}|j|jD]}|d}|dpX|dpX|dpX|dpX|dpX|d	}|d
pp|dpp|d}t|j|}t|}|j	d
d|}|
||fqt|dddd}|S)z
        Return the URLs of all the links on a page together with information
        about their "rel" attribute, for determining which ones to treat as
        downloads and which ones to queue for further scraping.
        cSs,t|\}}}}}}t||t||||fS)zTidy up an URL.)rrr)r,r5rmrirrrr.r.r/clean-szPage.links..cleanr3Zrel1Zrel2Zrel3Zrel4Zrel5Zrel6rrrsZurl3cSsdt|dS)Nz%%%2xr)ordr)rr.r.r/;zPage.links..cSs|dS)Nrr.)rnr.r.r/r?rT)r?reverse)r_hreffinditerr	groupdictrrr	_clean_resubrr)r9rrQrrrelr,r.r.r/links&s$
z
Page.linksN)r@rArBrCrecompileISXrrrLrrrr.r.r.r/r	srcseZdZdZejdddddZdfdd		Zd
dZdd
Z	ddZ
edej
ZddZddZddZddZddZedZddZZS)SimpleScrapingLocatorz
    A locator which scrapes HTML pages to locate downloads for a distribution.
    This runs multiple threads to do the I/O; performance is at least as good
    as pip's PackageFinder, which works in an analogous fashion.
    cCstjttdS)N)Zfileobj)gzipZGzipFilerrrbr.r.r/rMrzSimpleScrapingLocator.cCs|SrVr.rr.r.r/rNr)ZdeflaterZnoneN
csltt|jf|t||_||_i|_t|_t	
|_t|_d|_
||_t|_t|_d|_dS)a
        Initialise an instance.
        :param url: The root URL to use for scraping.
        :param timeout: The timeout, in seconds, to be applied to requests.
                        This defaults to ``None`` (no timeout specified).
        :param num_workers: The number of worker threads you want to do I/O,
                            This defaults to 10.
        :param kwargs: Passed to the superclass.
        FN)rrrLrrr(_page_cacher_seenr
rJ	_to_fetch
_bad_hostsskip_externalsnum_workers	threadingRLock_lock_gplockplatform_check)r9r,r(rrrr.r/rLQs




zSimpleScrapingLocator.__init__cCsFg|_t|jD]0}tj|jd}|d||j|qdS)z
        Threads are created only when get_project is called, and terminate
        before it returns. They are there primarily to parallelise I/O (i.e.
        fetching web pages).
        )targetTN)	_threadsrangerrZThread_fetchZ	setDaemonstartrO)r9irnr.r.r/_prepare_threadsls
z&SimpleScrapingLocator._prepare_threadscCs6|jD]}|jdq|jD]}|qg|_dS)zu
        Tell all the threads to terminate (by sending a sentinel value) and
        wait for them to do so.
        N)rrrr~)r9rnr.r.r/
_wait_threadsys



z#SimpleScrapingLocator._wait_threadsc	Csiid}|jx||_||_t|jdt|}|j|j|	z&td||j
||j
W5|
X|`W5QRX|S)Nrz%s/zQueueing %s)rrQrxrrrrrWrrrrprqrrr~)r9rarQr,r.r.r/rbs



z"SimpleScrapingLocator._get_projectz<\b(linux_(i\d86|x86_64|arm\w+)|win(32|_amd64)|macosx_?\d+)\bcCs|j|S)zD
        Does an URL refer to a platform-specific download?
        )platform_dependentr)r9r,r.r.r/_is_platform_dependentsz,SimpleScrapingLocator._is_platform_dependentc	CsZ|jr||rd}n|||j}td|||rV|j||j|W5QRX|S)a%
        See if an URL is a suitable download for a project.

        If it is, register information in the result dictionary (for
        _get_project) about the specific version it's for.

        Note that the return value isn't actually used other than as a boolean
        value.
        Nzprocess_download: %s -> %s)	rrrrxrprqrrrQ)r9r,rr.r.r/_process_downloads
z'SimpleScrapingLocator._process_downloadc
Cst|\}}}}}}||j|j|jr2d}n||jrJ||jsJd}nd||js\d}nR|dkrjd}nD|dkrxd}n6||rd}n&|	ddd}	|	
dkrd}nd}td	|||||S)
z
        Determine whether a link URL from a referring page and with a
        particular "rel" attribute should be queued for scraping.
        F)ZhomepageZdownload)ZhttprfZftp:rrZ	localhostTz#should_queue: %s (%s) from %s -> %s)
rrjrrrrrrrsplitrrprq)
r9linkZreferrerrr5rmri_rQhostr.r.r/
_should_queues0

z#SimpleScrapingLocator._should_queuec
Cs|j}zz|r||}|dkr,WWq|jD]j\}}||jkr2zB|j|||s||||rt	
d|||j|Wq2tk
rYq2Xq2Wn2t
k
r}z|jt|W5d}~XYnXW5|jX|sqqdS)z
        Get a URL to fetch from the work queue, get the HTML page, examine its
        links for download candidates and candidates for further scraping.

        This is a handy method to run in a thread.
        NzQueueing %s from %s)rrNrPget_pagerrrrrrprqrrrrKr)r9r,pagerrrRr.r.r/rs,



&zSimpleScrapingLocator._fetchcCsXt|\}}}}}}|dkr:tjt|r:tt|d}||jkr`|j|}t	d||n|
ddd}d}||jkrt	d||nt|d	d
id}zzt	d||j
j||jd
}	t	d||	}
|
dd}t|r|	}|	}
|
d}|r"|j|}||
}
d}t|}|r@|d}z|
|}
Wn tk
rn|
d}
YnXt|
|}||j|<Wntk
r}z|jdkrtd||W5d}~XYnt k
r}z0td|||j!|j"|W5QRXW5d}~XYn2t#k
rB}ztd||W5d}~XYnXW5||j|<X|S)a
        Get the HTML for an URL, possibly from an in-memory cache.

        XXX TODO Note: this cache is never actually cleared. It's assumed that
        the data won't get stale over the lifetime of a locator instance (not
        necessarily true for the default_locator).
        filez
index.htmlzReturning %s from cache: %srrrNzSkipping %s due to bad host %szAccept-encodingZidentity)r>zFetching %sr'z
Fetched %szContent-Typer3zContent-Encodingzutf-8zlatin-1izFetch failed: %s: %s)$rosriisdirrrrrrprqrrrrHrr(rrNHTML_CONTENT_TYPErZgeturlrdecodersCHARSETrrrUnicodeErrorrrr<rrrrr)r9r,r5rmrirrQrr:rr>Zcontent_typeZ	final_urlrencodingdecoderrrRr.r.r/rsZ	







&$zSimpleScrapingLocator.get_pagez]*>([^<]+)[\w-]+)\s*\(\s*(==\s*)?(?P[^)]+)\)$c@sLeZdZdZdddZddZddZd	d
ZddZd
dZ	dddZ
dS)DependencyFinderz0
    Locate dependencies for distributions.
    NcCs|pt|_t|jj|_dS)zf
        Initialise an instance, using the specified locator
        to locate distributions.
        N)default_locatorrr"r5r'r.r.r/rL*s
zDependencyFinder.__init__cCsrtd||j}||j|<||j||jf<|jD]:}t|\}}td||||j	|t
||fq2dS)z
        Add a distribution to the finder. This will update internal information
        about who provides what.
        :param dist: The distribution to add.
        zadding distribution %szAdd to provided: %s, %s, %sN)rprqr?
dists_by_namedistsrprovidesrprovidedrrr)r9rraprr.r.r/add_distribution2s

z!DependencyFinder.add_distributioncCsxtd||j}|j|=|j||jf=|jD]D}t|\}}td||||j|}|	||f|s.|j|=q.dS)z
        Remove a distribution from the finder. This will update internal
        information about who provides what.
        :param dist: The distribution to remove.
        zremoving distribution %sz Remove from provided: %s, %s, %sN)
rprqr?r.r/rr0rr1remove)r9rrar2rsr.r.r/remove_distributionAs

z$DependencyFinder.remove_distributioncCsBz|j|}Wn,tk
r<|d}|j|}YnX|S)z
        Get a version matcher for a requirement.
        :param reqt: The requirement
        :type reqt: str
        :return: A version matcher (an instance of
                 :class:`distlib.version.Matcher`).
        r)r5rIr#r)r9reqtrIrar.r.r/get_matcherSszDependencyFinder.get_matcherc		Cst||}|j}t}|j}||krp||D]B\}}z||}Wntk
rZd}YnX|r,||qpq,|S)z
        Find the distributions which can fulfill a requirement.

        :param reqt: The requirement.
         :type reqt: str
        :return: A set of distribution which can fulfill the requirement.
        F)r8r?rr1rr#r)	r9r7rIrarQr1rproviderrr.r.r/find_providerscs


zDependencyFinder.find_providersc	Cs|j|}t}|D]$}||}||js||q|rZ|d||t|fd}n@|||j|=|D]}|j|t|qp|	|d}|S)a
        Attempt to replace one provider with another. This is typically used
        when resolving dependencies from multiple sources, e.g. A requires
        (B >= 1.0) while C requires (B >= 1.1).

        For successful replacement, ``provider`` must meet all the requirements
        which ``other`` fulfills.

        :param provider: The provider we are trying to replace with.
        :param other: The provider we're trying to replace.
        :param problems: If False is returned, this will contain what
                         problems prevented replacement. This is currently
                         a tuple of the literal string 'cantreplace',
                         ``provider``, ``other``  and the set of requirements
                         that ``provider`` couldn't fulfill.
        :return: True if we can replace ``other`` with ``provider``, else
                 False.
        ZcantreplaceFT)
reqtsrr8rrr	frozensetr6rr3)	r9r9otherproblemsZrlistZ	unmatchedr5rIrQr.r.r/try_to_replace{s$




zDependencyFinder.try_to_replaceFcCsi|_i|_i|_i|_t|p g}d|krH|d|tdddgO}t|trh|}}t	d|n4|j
j||d}}|dkrtd|t	d	|d
|_
t}t|g}t|g}|r|}|j}	|	|jkr||n"|j|	}
|
|kr|||
||j|jB}|j}t}
|r`||kr`dD]*}d|}||kr4|
t|d
|O}
q4||B|
B}|D].}||}|sFt	d||j
j||d}|dkr|s|j
j|d
d}|dkrt	d||d|fn^|j|j}}||f|jkr||||||krF||krF||t	d|j|D]R}|j}	|	|jkrx|j|t|n"|j|	}
|
|krJ|||
|qJqpqt|j}|D]&}||k|_|jrt	d|jqt	d|||fS)a
        Find a distribution and all distributions it depends on.

        :param requirement: The requirement specifying the distribution to
                            find, or a Distribution instance.
        :param meta_extras: A list of meta extras such as :test:, :build: and
                            so on.
        :param prereleases: If ``True``, allow pre-release versions to be
                            returned - otherwise, don't return prereleases
                            unless they're all that's available.

        Return a set of :class:`Distribution` instances and a set of
        problems.

        The distributions returned should be such that they have the
        :attr:`required` attribute set to ``True`` if they were
        from the ``requirement`` passed to ``find()``, and they have the
        :attr:`build_time_dependency` attribute set to ``True`` unless they
        are post-installation dependencies of the ``requirement``.

        The problems should be a tuple consisting of the string
        ``'unsatisfied'`` and the requirement which couldn't be satisfied
        by any distribution known to the locator.
        z:*:z:test:z:build:z:dev:zpassed %s as requirement)rNzUnable to locate %rz
located %sT)ZtestZbuildZdevz:%s:z%s_requireszNo providers found for %rzCannot satisfy %rZunsatisfiedzAdding %s to install_distsz#%s is a build-time dependency only.zfind done for %s)r1r/r.r;rr4
isinstancerrprqrrrZ	requestedrr?r3r?Zrun_requiresZ
meta_requiresZbuild_requiresgetattrr:rrZname_and_versionrvaluesZbuild_time_dependency)r9rZmeta_extrasrrrr>ZtodoZ
install_distsrar=ZireqtsZsreqtsZereqtsr?rRZ	all_reqtsrZ	providersr9nrr2r/r.r.r/finds

















zDependencyFinder.find)N)NF)r@rArBrCrLr3r6r8r:r?rDr.r.r.r/r,%s
(r,)N)OriorrZloggingr	rgrrImportErrorZdummy_threadingrr3rcompatrrrrr	r
rrr
rrr7rrrrZdatabaserrrrrrutilrrrrrrrr r!rr"r#rr$r%Z	getLoggerr@rprrrr
rr*r0r1objectrDrrrrrrr"r$r-rZNAME_VERSION_REr,r.r.r.r/s^D,



@0E:zA&[
site-packages/pip/_vendor/distlib/__pycache__/locators.cpython-38.pyc000064400000112601151464452710021642 0ustar00U

.e_@sDddlZddlmZddlZddlZddlZddlZddlZzddlZWne	k
rdddl
ZYnXddlZddlm
Z
ddlmZmZmZmZmZmZmZmZmZmZmZmZmZmZmZddlm Z m!Z!m"Z"ddl#m$Z$m%Z%ddl&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/dd	l0m1Z1m2Z2dd
l3m4Z4m5Z5e6e7Z8e9dZ:e9dej;Zd-ddZ?GdddeZ@GdddeAZBGdddeBZCGdddeBZDGdddeAZEGdddeBZFGdddeBZGGdd d eBZHGd!d"d"eBZIGd#d$d$eBZJeJeHeFd%d&d'd(d)ZKeKjLZLe9d*ZMGd+d,d,eAZNdS).N)BytesIO)DistlibException)urljoinurlparse
urlunparseurl2pathnamepathname2urlqueuequoteunescapestring_typesbuild_openerHTTPRedirectHandler	text_typeRequest	HTTPErrorURLError)DistributionDistributionPath	make_dist)MetadataMetadataInvalidError)	cached_propertyparse_credentialsensure_slashsplit_filenameget_project_dataparse_requirementparse_name_and_versionServerProxynormalize_name)
get_schemeUnsupportedVersionError)Wheel
is_compatiblez^(\w+)=([a-f0-9]+)z;\s*charset\s*=\s*(.*)\s*$ztext/html|application/x(ht)?mlzhttps://pypi.org/pypicCs6|dkrt}t|dd}z|WS|dXdS)z
    Return all distribution names known by an index.
    :param url: The URL of the index.
    :return: A list of all known distribution names.
    N@timeoutclose)
DEFAULT_INDEXr 
list_packages)urlclientr.@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/locators.pyget_all_distribution_names)sr0c@s$eZdZdZddZeZZZdS)RedirectHandlerzE
    A class to work around a bug in some Python 3.2.x releases.
    c	Csd}dD]}||kr||}q"q|dkr.dSt|}|jdkrnt||}t|drf|||n|||<t||||||S)N)locationZurireplace_header)rschemerZget_full_urlhasattrr4BaseRedirectHandlerhttp_error_302)	selfreqfpcodemsgheadersZnewurlkeyZurlpartsr.r.r/r8@s 

zRedirectHandler.http_error_302N)__name__
__module____qualname____doc__r8Zhttp_error_301Zhttp_error_303Zhttp_error_307r.r.r.r/r17sr1c@seZdZdZdZdZdZdZedZd)dd	Z	d
dZ
dd
ZddZddZ
ddZee
eZddZddZddZddZddZddZd d!Zd"d#Zd$d%Zd*d'd(ZdS)+LocatorzG
    A base class for locators - things that locate distributions.
    )z.tar.gzz.tar.bz2z.tarz.zipz.tgzz.tbz)z.eggz.exe.whl)z.pdfN)rEdefaultcCs,i|_||_tt|_d|_t|_dS)a^
        Initialise an instance.
        :param scheme: Because locators look for most recent versions, they
                       need to know the version scheme to use. This specifies
                       the current PEP-recommended scheme - use ``'legacy'``
                       if you need to support existing distributions on PyPI.
        N)	_cacher5rr1openermatcherr
Queueerrors)r9r5r.r.r/__init__fs
zLocator.__init__cCsVg}|jsRz|jd}||Wn|jjk
rDYqYnX|jq|S)z8
        Return any errors which have occurred.
        F)rKemptygetappendZEmpty	task_done)r9resulter.r.r/
get_errorsys

zLocator.get_errorscCs|dS)z>
        Clear any errors which may have been logged.
        N)rSr9r.r.r/clear_errorsszLocator.clear_errorscCs|jdSN)rGclearrTr.r.r/clear_cacheszLocator.clear_cachecCs|jSrV_schemerTr.r.r/_get_schemeszLocator._get_schemecCs
||_dSrVrY)r9valuer.r.r/_set_schemeszLocator._set_schemecCstddS)a=
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This should be implemented in subclasses.

        If called from a locate() request, self.matcher will be set to a
        matcher for the requirement to satisfy, otherwise it will be None.
         Please implement in the subclassNNotImplementedError)r9namer.r.r/_get_projects
zLocator._get_projectcCstddS)J
        Return all the distribution names known to this locator.
        r^Nr_rTr.r.r/get_distribution_namesszLocator.get_distribution_namescCsL|jdkr||}n2||jkr,|j|}n|||}||j|<|S)z
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This calls _get_project to do all the work, and just implements a caching layer on top.
        N)rGrbrU)r9rarQr.r.r/get_projects



zLocator.get_projectcCs^t|}t|j}d}|d}||j}|rBtt||j}|j	dkd|j
k||||fS)zu
        Give an url a score which can be used to choose preferred URLs
        for a given project release.
        TrEhttpszpypi.org)r	posixpathbasenamepathendswithdownloadable_extensionsr%r$
wheel_tagsr5netloc)r9r,trhZ
compatibleZis_wheelZis_downloadabler.r.r/	score_urls
zLocator.score_urlcCsR|}|rN||}||}||kr(|}||kr@td||ntd|||S)a{
        Choose one of two URLs where both are candidates for distribution
        archives for the same version of a distribution (for example,
        .tar.gz vs. zip).

        The current implementation favours https:// URLs over http://, archives
        from PyPI over those from other locations, wheel compatibility (if a
        wheel) and then the archive name.
        zNot replacing %r with %rzReplacing %r with %r)rologgerdebug)r9url1url2rQs1s2r.r.r/
prefer_urls


zLocator.prefer_urlcCs
t||S)zZ
        Attempt to split a filename in project name, version and Python version.
        )r)r9filenameproject_namer.r.r/rszLocator.split_filenamecCsdd}d}t|\}}}}}	}
|
drz~t	|}t
||jstd
|nX|dkrd}n||j|}|r|j|j
|jt|||||	dfd
dd|jDd}Wn0tk
r:}ztd|W5d}~XYnXn||jsZtd|nt|}}|jD]}||rn|dt|}|||}|std|nH|\}}}|r|||r|||t|||||	dfd}|r||d<qqn|r|r|
|d|<|S)a
        See if a URL is a candidate for a download URL for a project (the URL
        has typically been scraped from an HTML page).

        If it is, a dictionary is returned with keys "name", "version",
        "filename" and "url"; otherwise, None is returned.
        cSst|t|kSrV)r!)Zname1Zname2r.r.r/same_projectsz:Locator.convert_url_to_download_info..same_projectNzegg=z %s: version hint in fragment: %r)NN/rEzWheel not compatible: %sTr3z, cSs"g|]}dt|ddqS).N)joinlist).0vr.r.r/
sz8Locator.convert_url_to_download_info..)raversionrwr,python-versionzinvalid path for wheel: %szNot downloadable: %sz No match for project/version: %s)rarrwr,r	%s_digest)rlower
startswithrprqHASHER_HASHmatchgroupsrjr$r%rlrarrwrr~pyver	Exceptionwarningrkrgrhlenr)r9r,rxryrQr5rmriparamsqueryfragmalgodigestZorigpathwheelZincluderRrwZextrnrarrr.r.r/convert_url_to_download_infos

	 


z$Locator.convert_url_to_download_infocCs2d}dD]$}d|}||kr|||f}q.q|S)z
        Get a digest from a dictionary by looking at keys of the form
        'algo_digest'.

        Returns a 2-tuple (algo, digest) if found, else None. Currently
        looks only for SHA256, then MD5.
        N)Zsha256md5rr.)r9inforQrr?r.r.r/_get_digest1szLocator._get_digestc	Cs|d}|d}||kr,||}|j}nt|||jd}|j}|||_}|d}||d|<|j|dkr||j||_|d|t	
|||_|||<dS)z
        Update a result dictionary (the final result from _get_project) with a
        dictionary for a specific version, which typically holds information
        gleaned from a filename or URL for an archive for the distribution.
        rarr5r,digestsurlsN)popmetadatarr5rr
source_urlrv
setdefaultsetaddlocator)	r9rQrrardistmdrr,r.r.r/_update_version_dataAs

zLocator._update_version_dataFc	Csd}t|}|dkr td|t|j}||j|_}td|t|j	|
|j}t|dkr2g}|j
}	|D]z}
|
dkrqxzH||
std||
n*|s|	|
js||
ntd|
|jWqxtk
rtd||
YqxXqxt|d	krt||jd
}|r2td||d}||}|r|jrH|j|_|d
i|t|_i}|di}
|jD]}||
krv|
|||<qv||_d|_|S)a
        Find the most recent distribution which matches the given
        requirement.

        :param requirement: A requirement of the form 'foo (1.0)' or perhaps
                            'foo (>= 1.0, < 2.0, != 1.3)'
        :param prereleases: If ``True``, allow pre-release versions
                            to be located. Otherwise, pre-release versions
                            are not returned.
        :return: A :class:`Distribution` instance, or ``None`` if no such
                 distribution could be located.
        NzNot a valid requirement: %rzmatcher: %s (%s)r}rrz%s did not match %rz%skipping pre-release version %s of %szerror matching %s with %rr)r?zsorted list: %srzrr)rrr"r5rIrequirementrprqtyper@rerarZ
version_classrZ
is_prereleaserOrrsortedr?ZextrasrNr
download_urlsr)r9rprereleasesrQrr5rIversionsZslistZvclskrdZsdr,r.r.r/locateXsX




zLocator.locate)rF)F)r@rArBrCsource_extensionsbinary_extensionsexcluded_extensionsrlrkrLrSrUrXr[r]propertyr5rbrdrerorvrrrrrr.r.r.r/rDVs.

JrDcs0eZdZdZfddZddZddZZS)PyPIRPCLocatorz
    This locator uses XML-RPC to locate distributions. It therefore
    cannot be used with simple mirrors (that only mirror file content).
    cs*tt|jf|||_t|dd|_dS)z
        Initialise an instance.

        :param url: The URL to use for XML-RPC.
        :param kwargs: Passed to the superclass constructor.
        r&r'N)superrrLbase_urlr r-r9r,kwargs	__class__r.r/rLszPyPIRPCLocator.__init__cCst|jSrc)rr-r+rTr.r.r/rdsz%PyPIRPCLocator.get_distribution_namescCsiid}|j|d}|D]}|j||}|j||}t|jd}|d|_|d|_|d|_	|dg|_
|d|_t|}|r|d	}	|	d
|_
||	|_||_|||<|D]:}	|	d
}
||	}|d|t|
||d|
<qq|S)
NrTrrarlicensekeywordssummaryrr,rr)r-Zpackage_releasesZrelease_urlsZrelease_datarr5rarrNrrrrrrrrrrr)r9rarQrrrdatarrrr,rr.r.r/rbs0




zPyPIRPCLocator._get_projectr@rArBrCrLrdrb
__classcell__r.r.rr/rsrcs0eZdZdZfddZddZddZZS)PyPIJSONLocatorzw
    This locator uses PyPI's JSON interface. It's very limited in functionality
    and probably not worth using.
    cs tt|jf|t||_dSrV)rrrLrrrrr.r/rLszPyPIJSONLocator.__init__cCstddSrczNot available from this locatorNr_rTr.r.r/rdsz&PyPIJSONLocator.get_distribution_namesc
Cs
iid}t|jdt|}z|j|}|}t|}t	|j
d}|d}|d|_|d|_|
d|_|
dg|_|
d	|_t|}||_|d
}	|||j<|d
D]T}
|
d}|j|||
|j|<|d
|jt|||
|d|<q|d
D]\}}||jkr4qt	|j
d}
|j|
_||
_t|
}||_|||<|D]T}
|
d}|j|||
|j|<|d
|t|||
|d|<qhqWn@tk
r}z |jt|td|W5d}~XYnX|S)Nrz%s/jsonrrrarrrrrr,rZreleaseszJSON fetch failed: %s) rrrrHopenreaddecodejsonloadsrr5rarrNrrrrrrrrrrritemsrrKputrrp	exception)r9rarQr,resprrrrrrrZinfosZomdodistrRr.r.r/rbsT




	zPyPIJSONLocator._get_projectrr.r.rr/rsrc@s`eZdZdZedejejBejBZ	edejejBZ
ddZedejZe
ddZd	S)
Pagez4
    This class represents a scraped HTML page.
    z
(rel\s*=\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\s
]*))\s+)?
href\s*=\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\s
]*))
(\s+rel\s*=\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\s
]*)))?
z!]+)cCs4||_||_|_|j|j}|r0|d|_dS)zk
        Initialise an instance with the Unicode page contents and the URL they
        came from.
        rN)rrr,_basesearchgroup)r9rr,rr.r.r/rLs
z
Page.__init__z[^a-z0-9$&+,/:;=?@.#%_\\|-]cCsdd}t}|j|jD]}|d}|dpX|dpX|dpX|dpX|dpX|d	}|d
pp|dpp|d}t|j|}t|}|j	d
d|}|
||fqt|dddd}|S)z
        Return the URLs of all the links on a page together with information
        about their "rel" attribute, for determining which ones to treat as
        downloads and which ones to queue for further scraping.
        cSs,t|\}}}}}}t||t||||fS)zTidy up an URL.)rrr)r,r5rmrirrrr.r.r/clean-szPage.links..cleanr3Zrel1Zrel2Zrel3Zrel4Zrel5Zrel6rrrsZurl3cSsdt|dS)Nz%%%2xr)ordr)rr.r.r/;zPage.links..cSs|dS)Nrr.)rnr.r.r/r?rT)r?reverse)r_hreffinditerr	groupdictrrr	_clean_resubrr)r9rrQrrrelr,r.r.r/links&s$
z
Page.linksN)r@rArBrCrecompileISXrrrLrrrr.r.r.r/r	srcseZdZdZejdddddZdfdd		Zd
dZdd
Z	ddZ
edej
ZddZddZddZddZddZedZddZZS)SimpleScrapingLocatorz
    A locator which scrapes HTML pages to locate downloads for a distribution.
    This runs multiple threads to do the I/O; performance is at least as good
    as pip's PackageFinder, which works in an analogous fashion.
    cCstjttdS)N)Zfileobj)gzipZGzipFilerrrbr.r.r/rMrzSimpleScrapingLocator.cCs|SrVr.rr.r.r/rNr)ZdeflaterZnoneN
csltt|jf|t||_||_i|_t|_t	
|_t|_d|_
||_t|_t|_d|_dS)a
        Initialise an instance.
        :param url: The root URL to use for scraping.
        :param timeout: The timeout, in seconds, to be applied to requests.
                        This defaults to ``None`` (no timeout specified).
        :param num_workers: The number of worker threads you want to do I/O,
                            This defaults to 10.
        :param kwargs: Passed to the superclass.
        FN)rrrLrrr(_page_cacher_seenr
rJ	_to_fetch
_bad_hostsskip_externalsnum_workers	threadingRLock_lock_gplockplatform_check)r9r,r(rrrr.r/rLQs




zSimpleScrapingLocator.__init__cCsFg|_t|jD]0}tj|jd}|d||j|qdS)z
        Threads are created only when get_project is called, and terminate
        before it returns. They are there primarily to parallelise I/O (i.e.
        fetching web pages).
        )targetTN)	_threadsrangerrZThread_fetchZ	setDaemonstartrO)r9irnr.r.r/_prepare_threadsls
z&SimpleScrapingLocator._prepare_threadscCs6|jD]}|jdq|jD]}|qg|_dS)zu
        Tell all the threads to terminate (by sending a sentinel value) and
        wait for them to do so.
        N)rrrr~)r9rnr.r.r/
_wait_threadsys



z#SimpleScrapingLocator._wait_threadsc	Csiid}|jx||_||_t|jdt|}|j|j|	z&td||j
||j
W5|
X|`W5QRX|S)Nrz%s/zQueueing %s)rrQrxrrrrrWrrrrprqrrr~)r9rarQr,r.r.r/rbs



z"SimpleScrapingLocator._get_projectz<\b(linux_(i\d86|x86_64|arm\w+)|win(32|_amd64)|macosx_?\d+)\bcCs|j|S)zD
        Does an URL refer to a platform-specific download?
        )platform_dependentr)r9r,r.r.r/_is_platform_dependentsz,SimpleScrapingLocator._is_platform_dependentc	CsZ|jr||rd}n|||j}td|||rV|j||j|W5QRX|S)a%
        See if an URL is a suitable download for a project.

        If it is, register information in the result dictionary (for
        _get_project) about the specific version it's for.

        Note that the return value isn't actually used other than as a boolean
        value.
        Nzprocess_download: %s -> %s)	rrrrxrprqrrrQ)r9r,rr.r.r/_process_downloads
z'SimpleScrapingLocator._process_downloadc
Cst|\}}}}}}||j|j|jr2d}n||jrJ||jsJd}nd||js\d}nR|dkrjd}nD|dkrxd}n6||rd}n&|	ddd}	|	
dkrd}nd}td	|||||S)
z
        Determine whether a link URL from a referring page and with a
        particular "rel" attribute should be queued for scraping.
        F)ZhomepageZdownload)ZhttprfZftp:rrZ	localhostTz#should_queue: %s (%s) from %s -> %s)
rrjrrrrrrrsplitrrprq)
r9linkZreferrerrr5rmri_rQhostr.r.r/
_should_queues0

z#SimpleScrapingLocator._should_queuec
Cs|j}zz|r||}|dkr,WWq|jD]j\}}||jkr2zB|j|||s||||rt	
d|||j|Wq2tk
rYq2Xq2Wn2t
k
r}z|jt|W5d}~XYnXW5|jX|sqqdS)z
        Get a URL to fetch from the work queue, get the HTML page, examine its
        links for download candidates and candidates for further scraping.

        This is a handy method to run in a thread.
        NzQueueing %s from %s)rrNrPget_pagerrrrrrprqrrrrKr)r9r,pagerrrRr.r.r/rs,



&zSimpleScrapingLocator._fetchcCsXt|\}}}}}}|dkr:tjt|r:tt|d}||jkr`|j|}t	d||n|
ddd}d}||jkrt	d||nt|d	d
id}zzt	d||j
j||jd
}	t	d||	}
|
dd}t|r|	}|	}
|
d}|r"|j|}||
}
d}t|}|r@|d}z|
|}
Wn tk
rn|
d}
YnXt|
|}||j|<Wntk
r}z|jdkrtd||W5d}~XYnt k
r}z0td|||j!|j"|W5QRXW5d}~XYn2t#k
rB}ztd||W5d}~XYnXW5||j|<X|S)a
        Get the HTML for an URL, possibly from an in-memory cache.

        XXX TODO Note: this cache is never actually cleared. It's assumed that
        the data won't get stale over the lifetime of a locator instance (not
        necessarily true for the default_locator).
        filez
index.htmlzReturning %s from cache: %srrrNzSkipping %s due to bad host %szAccept-encodingZidentity)r>zFetching %sr'z
Fetched %szContent-Typer3zContent-Encodingzutf-8zlatin-1izFetch failed: %s: %s)$rosriisdirrrrrrprqrrrrHrr(rrNHTML_CONTENT_TYPErZgeturlrdecodersCHARSETrrrUnicodeErrorrrr<rrrrr)r9r,r5rmrirrQrr:rr>Zcontent_typeZ	final_urlrencodingdecoderrrRr.r.r/rsZ	







&$zSimpleScrapingLocator.get_pagez]*>([^<]+)[\w-]+)\s*\(\s*(==\s*)?(?P[^)]+)\)$c@sLeZdZdZdddZddZddZd	d
ZddZd
dZ	dddZ
dS)DependencyFinderz0
    Locate dependencies for distributions.
    NcCs|pt|_t|jj|_dS)zf
        Initialise an instance, using the specified locator
        to locate distributions.
        N)default_locatorrr"r5r)r.r.r/rL*s
zDependencyFinder.__init__cCsrtd||j}||j|<||j||jf<|jD]:}t|\}}td||||j	|t
||fq2dS)z
        Add a distribution to the finder. This will update internal information
        about who provides what.
        :param dist: The distribution to add.
        zadding distribution %szAdd to provided: %s, %s, %sN)rprqr?
dists_by_namedistsrprovidesrprovidedrrr)r9rraprr.r.r/add_distribution2s

z!DependencyFinder.add_distributioncCsxtd||j}|j|=|j||jf=|jD]D}t|\}}td||||j|}|	||f|s.|j|=q.dS)z
        Remove a distribution from the finder. This will update internal
        information about who provides what.
        :param dist: The distribution to remove.
        zremoving distribution %sz Remove from provided: %s, %s, %sN)
rprqr?r0r1rr2rr3remove)r9rrar4rsr.r.r/remove_distributionAs

z$DependencyFinder.remove_distributioncCsBz|j|}Wn,tk
r<|d}|j|}YnX|S)z
        Get a version matcher for a requirement.
        :param reqt: The requirement
        :type reqt: str
        :return: A version matcher (an instance of
                 :class:`distlib.version.Matcher`).
        r)r5rIr#r)r9reqtrIrar.r.r/get_matcherSszDependencyFinder.get_matcherc		Cst||}|j}t}|j}||krp||D]B\}}z||}Wntk
rZd}YnX|r,||qpq,|S)z
        Find the distributions which can fulfill a requirement.

        :param reqt: The requirement.
         :type reqt: str
        :return: A set of distribution which can fulfill the requirement.
        F)r:r?rr3rr#r)	r9r9rIrarQr3rproviderrr.r.r/find_providerscs


zDependencyFinder.find_providersc	Cs|j|}t}|D]$}||}||js||q|rZ|d||t|fd}n@|||j|=|D]}|j|t|qp|	|d}|S)a
        Attempt to replace one provider with another. This is typically used
        when resolving dependencies from multiple sources, e.g. A requires
        (B >= 1.0) while C requires (B >= 1.1).

        For successful replacement, ``provider`` must meet all the requirements
        which ``other`` fulfills.

        :param provider: The provider we are trying to replace with.
        :param other: The provider we're trying to replace.
        :param problems: If False is returned, this will contain what
                         problems prevented replacement. This is currently
                         a tuple of the literal string 'cantreplace',
                         ``provider``, ``other``  and the set of requirements
                         that ``provider`` couldn't fulfill.
        :return: True if we can replace ``other`` with ``provider``, else
                 False.
        ZcantreplaceFT)
reqtsrr:rrr	frozensetr8rr5)	r9r;otherproblemsZrlistZ	unmatchedr7rIrQr.r.r/try_to_replace{s$




zDependencyFinder.try_to_replaceFcCsi|_i|_i|_i|_t|p g}d|krH|d|tdddgO}t|trh|}}t	d|n4|j
j||d}}|dkrtd|t	d	|d
|_
t}t|g}t|g}|r|}|j}	|	|jkr||n"|j|	}
|
|kr|||
||j|jB}|j}t}
|r`||kr`dD]*}d|}||kr4|
t|d
|O}
q4||B|
B}|D].}||}|sFt	d||j
j||d}|dkr|s|j
j|d
d}|dkrt	d||d|fn^|j|j}}||f|jkr||||||krF||krF||t	d|j|D]R}|j}	|	|jkrx|j|t|n"|j|	}
|
|krJ|||
|qJqpqt|j}|D]&}||k|_|jrt	d|jqt	d|||fS)a
        Find a distribution and all distributions it depends on.

        :param requirement: The requirement specifying the distribution to
                            find, or a Distribution instance.
        :param meta_extras: A list of meta extras such as :test:, :build: and
                            so on.
        :param prereleases: If ``True``, allow pre-release versions to be
                            returned - otherwise, don't return prereleases
                            unless they're all that's available.

        Return a set of :class:`Distribution` instances and a set of
        problems.

        The distributions returned should be such that they have the
        :attr:`required` attribute set to ``True`` if they were
        from the ``requirement`` passed to ``find()``, and they have the
        :attr:`build_time_dependency` attribute set to ``True`` unless they
        are post-installation dependencies of the ``requirement``.

        The problems should be a tuple consisting of the string
        ``'unsatisfied'`` and the requirement which couldn't be satisfied
        by any distribution known to the locator.
        z:*:z:test:z:build:z:dev:zpassed %s as requirement)rNzUnable to locate %rz
located %sT)ZtestZbuildZdevz:%s:z%s_requireszNo providers found for %rzCannot satisfy %rZunsatisfiedzAdding %s to install_distsz#%s is a build-time dependency only.zfind done for %s)r3r1r0r=rr6r#rrprqrrrZ	requestedrr?r5rAZrun_requiresZ
meta_requiresZbuild_requiresgetattrr<rrZname_and_versionrvaluesZbuild_time_dependency)r9rZmeta_extrasrrrr@ZtodoZ
install_distsrar?ZireqtsZsreqtsZereqtsr?rRZ	all_reqtsrZ	providersr;nrr4r1r.r.r/finds

















zDependencyFinder.find)N)NF)r@rArBrCrLr5r8r:r<rArEr.r.r.r/r.%s
(r.)N)OriorrZloggingr	rgrrImportErrorZdummy_threadingrr3rcompatrrrrr	r
rrr
rrr7rrrrZdatabaserrrrrrutilrrrrrrrr r!rr"r#rr$r%Z	getLoggerr@rprrrr
rr*r0r1objectrDrrrrrrr"r&r/rZNAME_VERSION_REr.r.r.r.r/s^D,



@0E:zA&[
site-packages/pip/_vendor/distlib/__pycache__/version.cpython-38.opt-1.pyc000064400000047451151464452710022452 0ustar00U

.e_[@sfdZddlZddlZddlmZddlmZdddd	d
ddd
gZee	Z
GdddeZGddde
ZGddde
ZedZddZeZGdddeZddZGdddeZeddfeddfeddfedd fed!d"fed#d"fed$d%fed&d'fed(d)fed*d+ff
Zed,dfed-dfed.d%fed$d%fed/dffZed0Zd1d2Zd3d4Zed5ejZd6d6d7d6d8ddd9Zd:d;ZGdejZ"d?d@Z#dAdBZ$GdCd
d
eZ%GdDddeZ&GdEdFdFe
Z'e'eeee'ee!dGdHe'e$e&edIZ(e(dJe(dK<dLd
Z)dS)Mz~
Implementation of a flexible versioning scheme providing support for PEP-440,
setuptools-compatible and semantic versioning.
N)string_typesparse_requirementNormalizedVersionNormalizedMatcher
LegacyVersion
LegacyMatcherSemanticVersionSemanticMatcherUnsupportedVersionError
get_schemec@seZdZdZdS)rzThis is an unsupported version.N)__name__
__module____qualname____doc__rr?/usr/lib/python3.8/site-packages/pip/_vendor/distlib/version.pyrsc@sxeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZddZddZ
ddZeddZdS)VersioncCs"||_}|||_}dSN)strip_stringparse_parts)selfspartsrrr__init__szVersion.__init__cCstddS)Nzplease implement in a subclassNotImplementedErrorrrrrrr%sz
Version.parsecCs$t|t|kr td||fdS)Nzcannot compare %r and %r)type	TypeErrorrotherrrr_check_compatible(szVersion._check_compatiblecCs|||j|jkSrr%rr#rrr__eq__,s
zVersion.__eq__cCs||Srr'r#rrr__ne__0szVersion.__ne__cCs|||j|jkSrr&r#rrr__lt__3s
zVersion.__lt__cCs||p||Srr*r'r#rrr__gt__7szVersion.__gt__cCs||p||Srr+r#rrr__le__:szVersion.__le__cCs||p||Sr)r,r'r#rrr__ge__=szVersion.__ge__cCs
t|jSr)hashrrrrr__hash__AszVersion.__hash__cCsd|jj|jfS)Nz%s('%s')	__class__rrr0rrr__repr__DszVersion.__repr__cCs|jSrrr0rrr__str__GszVersion.__str__cCstddS)NzPlease implement in subclasses.rr0rrr
is_prereleaseJszVersion.is_prereleaseN)rrrrrr%r'r)r*r,r-r.r1r4r6propertyr7rrrrrsrc	@seZdZdZddddddddddddd	dd
ddZdd
ZddZddZeddZ	ddZ
ddZddZddZ
ddZddZdS) MatcherNcCs||kSrrvcprrrTzMatcher.cCs||kSrrr:rrrr>Ur?cCs||kp||kSrrr:rrrr>Vr?cCs||kp||kSrrr:rrrr>Wr?cCs||kSrrr:rrrr>Xr?cCs||kSrrr:rrrr>Yr?cCs||kp||kSrrr:rrrr>[r?cCs||kSrrr:rrrr>\r?)<><=>======~=!=cCst|Srrr rrrraszMatcher.parse_requirementcCs|jdkrtd||_}||}|s:td||j|_|j|_g}|jr|jD]d\}}|	dr|dkrtd||ddd}}||n||d}}|
|||fq^t||_dS)	NzPlease specify a version classz
Not valid: %rz.*)rDrGz#'.*' not allowed for %r constraintsTF)

version_class
ValueErrorrrrnamelowerkeyZconstraintsendswithappendtupler)rrrZclistopZvnprefixrrrrds*


zMatcher.__init__cCsxt|tr||}|jD]X\}}}|j|}t|trDt||}|s`d||jjf}t	|||||sdSqdS)z
        Check if the provided version matches the constraints.

        :param version: The version to match against this instance.
        :type version: String or :class:`Version` instance.
        z%r not implemented for %sFT)

isinstancerrIr
_operatorsgetgetattrr3rr)rversionoperator
constraintrSfmsgrrrmatchs




z
Matcher.matchcCs6d}t|jdkr2|jdddkr2|jdd}|S)Nrr)rDrE)lenr)rresultrrr
exact_versions zMatcher.exact_versioncCs0t|t|ks|j|jkr,td||fdS)Nzcannot compare %s and %s)r!rKr"r#rrrr%szMatcher._check_compatiblecCs"|||j|jko |j|jkSr)r%rMrr#rrrr's
zMatcher.__eq__cCs||Srr(r#rrrr)szMatcher.__ne__cCst|jt|jSr)r/rMrr0rrrr1szMatcher.__hash__cCsd|jj|jfS)Nz%s(%r)r2r0rrrr4szMatcher.__repr__cCs|jSrr5r0rrrr6szMatcher.__str__)rrrrIrUrrr]r8r`r%r'r)r1r4r6rrrrr9Os*
r9zk^v?(\d+!)?(\d+(\.\d+)*)((a|b|c|rc)(\d+))?(\.(post)(\d+))?(\.(dev)(\d+))?(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$cCs|}t|}|s"td||}tdd|ddD}t|dkrl|ddkrl|dd}qF|dszd}nt|d}|dd	}|d
d}|dd
}|d}|dkrd}n|dt|df}|dkrd}n|dt|df}|dkrd}n|dt|df}|dkr*d}nHg}	|dD]0}
|
	rTdt|
f}
nd|
f}
|	
|
q8t|	}|s|s|rd}nd}|sd}|sd}||||||fS)NzNot a valid version: %scss|]}t|VqdSrint.0r;rrr	sz_pep_440_key..r.r	

)NNr)arg)z)_)final)rPEP440_VERSION_REr]rgroupsrPsplitr^rbisdigitrO)rmrtZnumsZepochpreZpostdevZlocalrpartrrr_pep_440_keysT



r{c@s6eZdZdZddZedddddgZed	d
ZdS)raIA rational version.

    Good:
        1.2         # equivalent to "1.2.0"
        1.2.0
        1.2a1
        1.2.3a2
        1.2.3b1
        1.2.3c1
        1.2.3.4
        TODO: fill this out

    Bad:
        1           # minimum two numbers
        1.2a        # release level must have a release serial
        1.2.3b
    cCs<t|}t|}|}tdd|ddD|_|S)Ncss|]}t|VqdSrrarcrrrresz*NormalizedVersion.parse..rrf)_normalized_keyrsr]rtrPru_release_clause)rrr_rwrtrrrr
s

zNormalizedVersion.parserobr<rcrycstfddjDS)Nc3s |]}|r|djkVqdS)rN)PREREL_TAGS)rdtr0rrresz2NormalizedVersion.is_prerelease..)anyrr0rr0rr7szNormalizedVersion.is_prereleaseN)	rrrrrsetrr8r7rrrrrs
cCs>t|}t|}||krdS||s*dSt|}||dkS)NTFrf)str
startswithr^)xynrrr
_match_prefixs
rc	@sneZdZeZddddddddd	Zd
dZdd
ZddZddZ	ddZ
ddZddZddZ
ddZdS)r_match_compatible	_match_lt	_match_gt	_match_le	_match_ge	_match_eq_match_arbitrary	_match_ne)rFr@rArBrCrDrErGcCsV|rd|ko|jd}n|jdo,|jd}|rN|jddd}||}||fS)N+rgrr)rrrurI)rrXrZrSZstrip_localrrrr
_adjust_local6s
zNormalizedMatcher._adjust_localcCsD||||\}}||krdS|j}ddd|D}t||S)NFrfcSsg|]}t|qSrrrdirrr
Isz/NormalizedMatcher._match_lt..rr}joinrrrXrZrSZrelease_clauseZpfxrrrrDszNormalizedMatcher._match_ltcCsD||||\}}||krdS|j}ddd|D}t||S)NFrfcSsg|]}t|qSrrrrrrrQsz/NormalizedMatcher._match_gt..rrrrrrLszNormalizedMatcher._match_gtcCs||||\}}||kSrrrrXrZrSrrrrTszNormalizedMatcher._match_lecCs||||\}}||kSrrrrrrrXszNormalizedMatcher._match_gecCs.||||\}}|s ||k}n
t||}|SrrrrrXrZrSr_rrrr\s


zNormalizedMatcher._match_eqcCst|t|kSrrrrrrrdsz"NormalizedMatcher._match_arbitrarycCs0||||\}}|s ||k}nt||}|Srrrrrrrgs

zNormalizedMatcher._match_necCsf||||\}}||krdS||kr*dS|j}t|dkrH|dd}ddd|D}t||S)NTFrrgrfcSsg|]}t|qSrrrrrrrzsz7NormalizedMatcher._match_compatible..)rr}r^rrrrrrrosz#NormalizedMatcher._match_compatibleN)rrrrrIrUrrrrrrrrrrrrrr's&z[.+-]$z^[.](\d)z0.\1z^[.-]z
^\((.*)\)$\1z^v(ersion)?\s*(\d+)z\2z^r(ev)?\s*(\d+)z[.]{2,}rfz\b(alfa|apha)\balphaz\b(pre-alpha|prealpha)\bz	pre.alphaz	\(beta\)$betaz
^[:~._+-]+z
[,*")([\]]z[~:+_ -]z\.$z
(\d+(\.\d+)*)cCsL|}tD]\}}|||}q|s.d}t|}|sFd}|}n|dd}dd|D}t|dkr~|	dqft|dkr||
d}n8ddd|ddD||
d}|dd}dd	d|D}|}|rtD]\}}|||}q|s|}nd
|kr*dnd}|||}t
|sHd}|S)
z
    Try to suggest a semantic form for a version for which
    _suggest_normalized_version couldn't come up with anything.
    z0.0.0rrfcSsg|]}t|qSrrarrrrrsz-_suggest_semantic_version..NcSsg|]}t|qSrrrrrrrscSsg|]}t|qSrrrrrrrsry-r)rrL
_REPLACEMENTSsub_NUMERIC_PREFIXr]rtrur^rOendr_SUFFIX_REPLACEMENTS	is_semver)rr_ZpatreplrwrSsuffixseprrr_suggest_semantic_versions:
,
rcCshzt||WStk
r"YnX|}dD]\}}|||}q0tdd|}tdd|}tdd|}tdd	|}td
d|}|dr|d
d}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd |}zt|Wntk
rbd}YnX|S)!aSuggest a normalized version close to the given version string.

    If you have a version string that isn't rational (i.e. NormalizedVersion
    doesn't like it) then you might be able to get an equivalent (or close)
    rational version from this function.

    This does a number of simple normalizations to the given string, based
    on observation of versions currently in use on PyPI. Given a dump of
    those version during PyCon 2009, 4287 of them:
    - 2312 (53.93%) match NormalizedVersion without change
      with the automatic suggestion
    - 3474 (81.04%) match when using this suggestion method

    @param s {str} An irrational version string.
    @returns A rational version string, or None, if couldn't determine one.
    ))z-alpharo)z-betar~)rro)rr~)rr<)z-finalr)z-prer<)z-releaser)z.releaser)z-stabler)rrf)rqrf) r)z.finalr)rrrzpre$Zpre0zdev$Zdev0z([abc]|rc)[\-\.](\d+)$z\1\2z[\-\.](dev)[\-\.]?r?(\d+)$z.\1\2z[.~]?([abc])\.?rr;rNz\b0+(\d+)(?!\d)z(\d+[abc])$z\g<1>0z\.?(dev-r|dev\.r)\.?(\d+)$z.dev\2z-(a|b|c)(\d+)$z[\.\-](dev|devel)$z.dev0z(?![\.\-])dev$z(final|stable)$rz\.?(r|-|-r)\.?(\d+)$z.post\2z\.?(dev|git|bzr)\.?(\d+)$z\.?(pre|preview|-c)(\d+)$zc\g<2>zp(\d+)$z.post\1)r|rrLreplacererr)rZrsZorigrrrr_suggest_normalized_versions>
	
rz([a-z]+|\d+|[\.-])r<zfinal-@)rxZpreviewrrryrrfcCsrdd}g}||D]T}|dr^|dkrD|rD|ddkrD|q*|r^|ddkr^|qD||qt|S)NcSstg}t|D]R}t||}|rd|ddkrBdkrRnn|d}nd|}||q|d|S)N0r9**final)
_VERSION_PARTrurL_VERSION_REPLACErVzfillrO)rr_r=rrr	get_partsCs 
z_legacy_key..get_partsrrrgz*final-Z00000000)rpoprOrP)rrr_r=rrr_legacy_keyBs



rc@s eZdZddZeddZdS)rcCst|Sr)rr rrrr]szLegacyVersion.parsecCs8d}|jD](}t|tr
|dr
|dkr
d}q4q
|S)NFrrT)rrTrr)rr_rrrrr7`s
zLegacyVersion.is_prereleaseNrrrrr8r7rrrrr\sc@s4eZdZeZeejZded<e	dZ
ddZdS)r	rrFz^(\d+(\.\d+)*)cCs`||krdS|jt|}|s2td||dS|d}d|krV|ddd}t||S)NFzACannot compute compatible match for version %s  and constraint %sTrrfr)
numeric_rer]rloggerZwarningrtrsplitr)rrXrZrSrwrrrrrsszLegacyMatcher._match_compatibleN)rrrrrIdictr9rUrcompilerrrrrrr	ks


zN^(\d+)\.(\d+)\.(\d+)(-[a-z0-9]+(\.[a-z0-9-]+)*)?(\+[a-z0-9]+(\.[a-z0-9-]+)*)?$cCs
t|Sr)
_SEMVER_REr])rrrrrsrc	Csndd}t|}|st||}dd|ddD\}}}||dd||dd}}|||f||fS)	NcSs8|dkr|f}n$|ddd}tdd|D}|S)NrrfcSs"g|]}|r|dn|qS)r)rvr)rdr=rrrrsz5_semantic_key..make_tuple..)rurP)rZabsentr_rrrr
make_tuples
z!_semantic_key..make_tuplecSsg|]}t|qSrrarrrrrsz!_semantic_key..r|r)rrrt)	rrrwrtmajorminorZpatchrxZbuildrrr
_semantic_keys
rc@s eZdZddZeddZdS)r
cCst|Sr)rr rrrrszSemanticVersion.parsecCs|jdddkS)Nrrr)rr0rrrr7szSemanticVersion.is_prereleaseNrrrrrr
sc@seZdZeZdS)rN)rrrr
rIrrrrrsc@s6eZdZdddZddZddZdd	Zd
dZdS)

VersionSchemeNcCs||_||_||_dSr)rMmatcher	suggester)rrMrrrrrrszVersionScheme.__init__cCs2z|j|d}Wntk
r,d}YnX|SNTF)rrIrrrr_rrris_valid_versions
zVersionScheme.is_valid_versioncCs0z||d}Wntk
r*d}YnX|Sr)rrrrrris_valid_matchers

zVersionScheme.is_valid_matchercCs|d|S)z:
        Used for processing some metadata fields
        zdummy_name (%s))rr rrris_valid_constraint_listsz&VersionScheme.is_valid_constraint_listcCs|jdkrd}n
||}|Sr)rrrrrsuggests

zVersionScheme.suggest)N)rrrrrrrrrrrrrs

rcCs|Srrr rrrr>r?r>)
normalizedlegacyZsemanticrdefaultcCs|tkrtd|t|S)Nzunknown scheme name: %r)_SCHEMESrJ)rKrrrr
s)*rZloggingrcompatrutilr__all__Z	getLoggerrrrJrobjectrr9rrsr{r|rrrrrrrrIrrrrr	rrrr
rrrr
rrrrs
1d
=$W
.r	$site-packages/pip/_vendor/distlib/__pycache__/scripts.cpython-38.opt-1.pyc000064400000025002151464452710022440 0ustar00U

.e?@sddlmZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
ddlmZddl
mZmZmZmZmZeeZdZedZd	Zd
dZGdd
d
eZdS))BytesION)	sysconfigdetect_encodingZipFile)finder)FileOperatorget_export_entryconvert_pathget_executablein_venva


 

 
 
 
 
 
 
 
 
s^#!.*pythonw?[0-9.]*([ 	].*)?$z# -*- coding: utf-8 -*-
import re
import sys
from %(module)s import %(import_name)s
if __name__ == '__main__':
    sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
    sys.exit(%(func)s())
cCsXd|krT|drB|dd\}}d|krT|dsTd||f}n|dsTd|}|S)N z
/usr/bin/env r"z%s "%s"z"%s")
startswithsplit)
executableenvZ_executabler?/usr/lib/python3.8/site-packages/pip/_vendor/distlib/scripts.py_enquote_executable3s

rc@seZdZdZeZdZd'ddZddZe	j
d	rBd
dZdd
Z
ddZd(ddZddZeZddZddZd)ddZddZeddZejddZejd ksejd	krejd krd!d"Zd*d#d$Zd+d%d&ZdS),ScriptMakerz_
    A class to copy or create scripts from source scripts or callable
    specifications.
    NTFcCsz||_||_||_d|_d|_tjdkp:tjdko:tjdk|_t	d|_
|pRt||_tjdkprtjdkortjdk|_
dS)NFposixjava)X.Ynt)
source_dir
target_dir
add_launchersforceclobberosname_nameset_modesetvariantsr_fileop_is_nt)selfrrrdry_runZfileoprrr__init__Ls

zScriptMaker.__init__cCs@|ddr<|jr %srspythonwr|rd)!r!r/r1rr
rr`rr'Znewerr:ror6r9r*readliner;Zget_command_name
FIRST_LINE_REmatchr0groupcloseZ	copy_filer$rprqinforseekrUryr7)r)rrsZadjustrvfZ
first_linerrIrSlinesrTrtrwrrr_copy_script)sX



zScriptMaker._copy_scriptcCs|jjSr_r'r*)r)rrrr*]szScriptMaker.dry_runcCs||j_dSr_r)r)valuerrrr*asrcCsHtddkrd}nd}d||f}tddd}t||j}|S)	NPZ64Z32z%s%s.exerVrr)structcalcsize__name__rsplitrfindbytes)r)Zkindbitsr"Zdistlib_packagerJrrrrhiszScriptMaker._get_launchercCs6g}t|}|dkr"|||n|j|||d|S)a
        Make a script.

        :param specification: The specification, which is either a valid export
                              entry specification (to make a script from a
                              callable) or a filename (to make a script by
                              copying from a source location).
        :param options: A dictionary of options controlling script generation.
        :return: A list of all absolute pathnames written to.
        Nrz)r	rr)r)
specificationr2rsr]rrrmakewszScriptMaker.makecCs$g}|D]}||||q|S)z
        Take a list of specifications and make scripts from them,
        :param specifications: A list of specifications.
        :return: A list of all absolute pathnames written to,
        )extendr)r)Zspecificationsr2rsrrrr
make_multipleszScriptMaker.make_multiple)TFN)rLN)N)N)N) r
__module____qualname____doc__SCRIPT_TEMPLATErYrr+r4rGrHrr=r@rKrUr^_DEFAULT_MANIFESTrarcryrrpropertyr*setterr!r"r#rhrrrrrrrCs6

84
4


r)iorZloggingr!rerrGcompatrrrZ	resourcesrutilrr	r
rrZ	getLoggerrr:striprcompilerrrobjectrrrrrs


site-packages/pip/_vendor/distlib/__pycache__/manifest.cpython-38.opt-1.pyc000064400000023545151464452710022571 0ustar00U

.e9@sdZddlZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZdgZe
eZedejZed	ejejBZejdd
ZGdddeZdS)zu
Class representing the list of files in a distribution.

Equivalent to distutils.filelist, but fixes some problems.
N)DistlibException)fsdecodeconvert_pathManifestz\\w*
z#.*?(?=
)|
(?=$)c@szeZdZdZdddZddZddZd	d
Zddd
ZddZ	ddZ
ddZdddZd ddZ
d!ddZddZdS)"rz~A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.
    NcCs>tjtj|pt|_|jtj|_d|_t	|_
dS)zd
        Initialise an instance.

        :param base: The base directory to explore under.
        N)ospathabspathnormpathgetcwdbasesepprefixallfilessetfiles)selfrr@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/manifest.py__init__*szManifest.__init__cCsddlm}m}m}g|_}|j}|g}|j}|j}|r|}t	|}	|	D]R}
tj
||
}t|}|j}
||
r|t
|qN||
rN||
sN||qNq6dS)zmFind all files under the base and set ``allfiles`` to the absolute
        pathnames of files found.
        r)S_ISREGS_ISDIRS_ISLNKN)statrrrrrpopappendr	listdirr
joinst_moder)rrrrrrootstackrpushnamesnamefullnamermoderrrfindall9s"


zManifest.findallcCs4||jstj|j|}|jtj|dS)zz
        Add a file to the manifest.

        :param item: The pathname to add. This can be relative to the base.
        N)	
startswithrr	r
rrraddr)ritemrrrr*TszManifest.addcCs|D]}||qdS)z
        Add a list of files to the manifest.

        :param items: The pathnames to add. These can be relative to the base.
        N)r*)ritemsr+rrradd_many^szManifest.add_manyFcsbfddtj}|rFt}|D]}|tj|q&||O}ddtdd|DDS)z8
        Return sorted files in directory order
        cs>||td||jkr:tj|\}}||dS)Nzadd_dir added %s)r*loggerdebugrr	r
split)dirsdparent_add_dirrrrr6ls


z Manifest.sorted..add_dircSsg|]}tjj|qSr)r	r
r).0Z
path_tuplerrr
zsz#Manifest.sorted..css|]}tj|VqdS)N)r	r
r0)r7r
rrr	{sz"Manifest.sorted..)rrr	r
dirnamesorted)rZwantdirsresultr1frr5rr;gs
zManifest.sortedcCst|_g|_dS)zClear all collected files.N)rrr)rrrrclear}szManifest.clearcCsj||\}}}}|dkrB|D]}|j|ddstd|qn$|dkrf|D]}|j|dd}qNn|dkr|D]}|j|ddsrtd|qrn|d	kr|D]}|j|dd}qn|d
kr|D] }|j||dstd||qn|d
kr
|D]}|j||d}qn\|dkr2|jd|dsftd|n4|dkrZ|jd|dsftd|ntd|dS)av
        Process a directive which either adds some files from ``allfiles`` to
        ``files``, or removes some files from ``files``.

        :param directive: The directive to process. This should be in a format
                     compatible with distutils ``MANIFEST.in`` files:

                     http://docs.python.org/distutils/sourcedist.html#commands
        includeT)anchorzno files found matching %rexcludeglobal-includeFz3no files found matching %r anywhere in distributionglobal-excluderecursive-include)rz-no files found matching %r under directory %rrecursive-excludegraftNz no directories found matching %rprunez4no previously-included directories found matching %rzinvalid action %r)_parse_directive_include_patternr.Zwarning_exclude_patternr)r	directiveactionpatternsthedirZ
dirpatternpatternfoundrrrprocess_directivesP


zManifest.process_directivecCs|}t|dkr,|ddkr,|dd|d}d}}}|dkrxt|dkr`td|d	d
|ddD}n~|dkrt|dkrtd
|t|d}dd
|ddD}n:|dkrt|dkrtd|t|d}ntd|||||fS)z
        Validate a directive.
        :param directive: The directive to validate.
        :return: A tuple of action, patterns, thedir, dir_patterns
        rr)r?rArBrCrDrErFrGr?N)r?rArBrCrz$%r expects   ...cSsg|]}t|qSrrr7Zwordrrrr8sz-Manifest._parse_directive..)rDrEz*%r expects    ...cSsg|]}t|qSrrrRrrrr8s)rFrGz!%r expects a single zunknown action %r)r0leninsertrr)rrKZwordsrLrMrNZdir_patternrrrrHs4zManifest._parse_directiveTcCsPd}|||||}|jdkr&||jD]}||r,|j|d}q,|S)aSelect strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.

        Patterns are not quite the same as implemented by the 'fnmatch'
        module: '*' and '?'  match non-special characters, where "special"
        is platform-dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found.
        FNT)_translate_patternrr(searchrr*)rrOr@ris_regexrP
pattern_rer%rrrrIs


zManifest._include_patterncCsBd}|||||}t|jD]}||r|j|d}q|S)atRemove strings (presumably filenames) from 'files' that match
        'pattern'.

        Other parameters are the same as for 'include_pattern()', above.
        The list 'self.files' is modified in place. Return True if files are
        found.

        This API is public to allow e.g. exclusion of SCM subdirs, e.g. when
        packaging source distributions
        FT)rVlistrrWremove)rrOr@rrXrPrYr=rrrrJ)s
zManifest._exclude_patternc
Csv|rt|trt|S|Stdkr:|dd\}}}|rR||}tdkrVnd}ttj	
|jd}	|dk	r4tdkr|d}
||dt|
}n&||}|t|t|t|}tj
}tj
dkrd}tdkrd|	|
|d|f}n0|t|t|t|}d	||	||||f}n8|rltdkrRd|	|}nd
||	|t|df}t|S)aTranslate a shell-like wildcard pattern to a compiled regular
        expression.

        Return the compiled regex.  If 'is_regex' true,
        then 'pattern' is directly compiled to a regex (if it's a string)
        or just returned as-is (assumes it's a regex object).
        )rSrr4N\z\\^z.*z%s%s%s%s.*%s%sz%s%s%s)
isinstancestrrecompile_PYTHON_VERSION_glob_to_re	partitionescaper	r
rrrTr)
rrOr@rrXstartr4endrYrZ
empty_patternZ	prefix_rerrrrrV=sF	










zManifest._translate_patterncCs8t|}tj}tjdkrd}d|}td||}|S)zTranslate a shell-like glob pattern to a regular expression.

        Return a string containing the regex.  Differs from
        'fnmatch.translate()' in that '*' does not match "special characters"
        (which are platform-specific).
        r]z\\\\z\1[^%s]z((?rQrHrIrJrVrdrrrrr%s&

	
O/
)

7)roriZloggingr	rasysr\rcompatrutilr__all__Z	getLoggerrlr.rbMZ_COLLAPSE_PATTERNSZ_COMMENTED_LINEversion_inforcobjectrrrrrs
site-packages/pip/_vendor/distlib/__pycache__/index.cpython-38.pyc000064400000041737151464452710021136 0ustar00U

.eJR@sddlZddlZddlZddlZddlZddlZzddlmZWn ek
r`ddl	mZYnXddl
mZddlm
Z
mZmZmZmZmZddlmZmZmZeeZdZdZGd	d
d
eZdS)N)Thread)DistlibException)HTTPBasicAuthHandlerRequestHTTPPasswordMgrurlparsebuild_openerstring_types)cached_propertyzip_dirServerProxyzhttps://pypi.org/pypipypic@seZdZdZdZd*ddZddZdd	Zd
dZdd
Z	ddZ
ddZd+ddZd,ddZ
d-ddZd.ddZddZd/ddZd0d d!Zd1d"d#Zd$d%Zd&d'Zd2d(d)ZdS)3PackageIndexzc
    This class represents a package index compatible with PyPI, the Python
    Package Index.
    s.----------ThIs_Is_tHe_distlib_index_bouNdaRY_$Nc
Cs|pt|_|t|j\}}}}}}|s<|s<|s<|dkrJtd|jd|_d|_d|_d|_t	t
jdR}dD]F}	z,tj
|	dg||d}
|
dkr|	|_WqWqttk
rYqtXqtW5QRXdS)	z
        Initialise an instance.

        :param url: The URL of the index. If not specified, the URL for PyPI is
                    used.
        )ZhttpZhttpszinvalid repository: %sNw)gpgZgpg2z	--versionstdoutstderrr)
DEFAULT_INDEXurlread_configurationrrpassword_handlerssl_verifierrgpg_homeopenosdevnull
subprocessZ
check_callOSError)selfrZschemenetlocpathZparamsZqueryZfragZsinksrcr%=/usr/lib/python3.8/site-packages/pip/_vendor/distlib/index.py__init__$s(

zPackageIndex.__init__cCs&ddlm}ddlm}|}||S)zs
        Get the distutils command for interacting with PyPI configurations.
        :return: the command.
        r)Distribution)
PyPIRCCommand)Zdistutils.corer(Zdistutils.configr))r r(r)dr%r%r&_get_pypirc_commandAsz PackageIndex._get_pypirc_commandcCsR|}|j|_|}|d|_|d|_|dd|_|d|j|_dS)z
        Read the PyPI access configuration as supported by distutils, getting
        PyPI to do the actual work. This populates ``username``, ``password``,
        ``realm`` and ``url`` attributes from the configuration.
        usernamepasswordrealmr
repositoryN)r+rr/Z_read_pypircgetr,r-r.)r cZcfgr%r%r&rKszPackageIndex.read_configurationcCs$||}||j|jdS)z
        Save the PyPI access configuration. You must have set ``username`` and
        ``password`` attributes before calling this method.

        Again, distutils is used to do the actual work.
        N)check_credentialsr+Z
_store_pypircr,r-)r r1r%r%r&save_configurationZszPackageIndex.save_configurationcCs\|jdks|jdkrtdt}t|j\}}}}}}||j||j|jt||_	dS)zp
        Check that ``username`` and ``password`` have been set, and raise an
        exception if not.
        Nz!username and password must be set)
r,r-rrrrZadd_passwordr.rr)r Zpm_r!r%r%r&r2fszPackageIndex.check_credentialscCs\|||}d|d<||g}||}d|d<||g}||S)aq
        Register a distribution on PyPI, using the provided metadata.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the distribution to be
                         registered.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        Zverify:actionZsubmit)r2validatetodictencode_requestitemssend_request)r metadatar*requestZresponser%r%r&registerrs

zPackageIndex.registercCsF|}|sq:|d}||td||fq|dS)ar
        Thread runner for reading lines of from a subprocess into a buffer.

        :param name: The logical name of the stream (used for logging only).
        :param stream: The stream to read from. This will typically a pipe
                       connected to the output stream of a subprocess.
        :param outbuf: The list to append the read lines to.
        utf-8z%s: %sN)readlinedecoderstripappendloggerdebugclose)r namestreamZoutbufr#r%r%r&_readers

zPackageIndex._readerc	Cs|jdddg}|dkr|j}|r.|d|g|dk	rF|dddgt}tj|tj|d	}|d
dd|d
||gt	
dd|||fS)a
        Return a suitable command for signing a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The signing command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        --status-fd2--no-ttyN	--homedirz--batchz--passphrase-fd0z.ascz
--detach-signz--armorz--local-userz--outputinvoking: %s )rrextendtempfileZmkdtemprr"joinbasenamerCrD)r filenamesigner
sign_passwordkeystorecmdZtdZsfr%r%r&get_sign_commands"
zPackageIndex.get_sign_commandc	Cstjtjd}|dk	r tj|d<g}g}tj|f|}t|jd|j|fd}|t|jd|j|fd}||dk	r|j	||j
||||j
||fS)a
        Run a command in a child process , passing it any input data specified.

        :param cmd: The command to run.
        :param input_data: If specified, this must be a byte string containing
                           data to be sent to the child process.
        :return: A tuple consisting of the subprocess' exit code, a list of
                 lines read from the subprocess' ``stdout``, and a list of
                 lines read from the subprocess' ``stderr``.
        rNstdinr)targetargsr)rPIPEPopenrrHrstartrrZwriterEwaitrR
returncode)	r rXZ
input_datakwargsrrpt1t2r%r%r&run_commands&

zPackageIndex.run_commandc
CsD|||||\}}|||d\}}}	|dkr@td||S)aR
        Sign a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The absolute pathname of the file where the signature is
                 stored.
        r>rz&sign command failed with error code %s)rYrgencoder)
r rTrUrVrWrXsig_filer$rrr%r%r&	sign_files

zPackageIndex.sign_filesdistsourcec	Cs(|tj|s td|||}d}	|rZ|jsJt	dn|
||||}	t|d}
|
}W5QRXt
|}t
|}
|dd||||
ddtj||fg}|	rt|	d}
|
}W5QRX|d	tj|	|fttj|	|||}||S)
a
        Upload a release file to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the file to be uploaded.
        :param filename: The pathname of the file to be uploaded.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param filetype: The type of the file being uploaded. This is the
                        distutils command which produced that file, e.g.
                        ``sdist`` or ``bdist_wheel``.
        :param pyversion: The version of Python which the release relates
                          to. For code compatible with any Python, this would
                          be ``source``, otherwise it would be e.g. ``3.2``.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        z
not found: %sNz)no signing program available - not signedrbZfile_upload1)r5Zprotocol_versionfiletype	pyversion
md5_digest
sha256_digestcontentZ
gpg_signature)r2rr"existsrr6r7rrCZwarningrjrreadhashlibmd5	hexdigestZsha256updaterSrBshutilZrmtreedirnamer8r9r:)r r;rTrUrVrorprWr*rifZ	file_datarqrrfilesZsig_datar<r%r%r&upload_filesD
zPackageIndex.upload_filec
Cs|tj|s td|tj|d}tj|sFtd|||j|j	}}t
|}dd|fd|fg}d||fg}|||}	|
|	S)a2
        Upload documentation to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the documentation to be
                         uploaded.
        :param doc_dir: The pathname of the directory which contains the
                        documentation. This should be the directory that
                        contains the ``index.html`` for the documentation.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        znot a directory: %rz
index.htmlz
not found: %r)r5Z
doc_uploadrFversionrs)r2rr"isdirrrRrtr6rFrrgetvaluer8r:)
r r;Zdoc_dirfnrFrZzip_datafieldsr}r<r%r%r&upload_documentation(s 
z!PackageIndex.upload_documentationcCsT|jdddg}|dkr|j}|r.|d|g|d||gtdd||S)	a|
        Return a suitable command for verifying a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The verifying command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        rIrJrKNrLz--verifyrNrO)rrrPrCrDrR)r signature_filename
data_filenamerWrXr%r%r&get_verify_commandDszPackageIndex.get_verify_commandcCsH|jstd||||}||\}}}|dkr@td||dkS)a6
        Verify a signature for a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: True if the signature was verified, else False.
        z0verification unavailable because gpg unavailable)rrz(verify command failed with error code %sr)rrrrg)r rrrWrXr$rrr%r%r&verify_signature\szPackageIndex.verify_signaturec	Csl|dkrd}tdn6t|ttfr0|\}}nd}tt|}td|t|d}|t	|}z|}	d}
d}d}d}
d	|	krt|	d
}|r||
|
||
|
}|sq|t|7}|||r|||
d7}
|r||
|
|qW5|
XW5QRX|dkr0||kr0td||f|rh|}||kr\td
||||ftd|dS)a
        This is a convenience method for downloading a file from an URL.
        Normally, this will be a file from the index, though currently
        no check is made for this (i.e. a file can be downloaded from
        anywhere).

        The method is just like the :func:`urlretrieve` function in the
        standard library, except that it allows digest computation to be
        done during download and checking that the downloaded data
        matched any expected value.

        :param url: The URL of the file to be downloaded (assumed to be
                    available via an HTTP GET request).
        :param destfile: The pathname where the downloaded file is to be
                         saved.
        :param digest: If specified, this must be a (hasher, value)
                       tuple, where hasher is the algorithm used (e.g.
                       ``'md5'``) and ``value`` is the expected value.
        :param reporthook: The same as for :func:`urlretrieve` in the
                           standard library.
        NzNo digest specifiedrwzDigest specified: %swbi rzcontent-lengthzContent-Lengthrz1retrieval incomplete: got only %d out of %d bytesz.%s digest mismatch for %s: expected %s, got %szDigest verified: %s)rCrD
isinstancelisttuplegetattrrvrr:rrEinfointrulenr`ryrrx)r rZdestfileZdigestZ
reporthookZdigesterZhasherZdfpZsfpheadersZ	blocksizesizeruZblocknumblockactualr%r%r&
download_fileus^




zPackageIndex.download_filecCs:g}|jr||j|jr(||jt|}||S)z
        Send a standard library :class:`Request` to PyPI and return its
        response.

        :param req: The request to send.
        :return: The HTTP response from PyPI (a standard library HTTPResponse).
        )rrBrr	r)r ZreqZhandlersZopenerr%r%r&r:szPackageIndex.send_requestc
Csg}|j}|D]L\}}t|ttfs*|g}|D]*}|d|d|dd|dfq.q|D].\}}	}
|d|d||	fdd|
fq`|d|ddfd|}d|}|tt|d}
t	|j
||
S)	a&
        Encode fields and files for posting to an HTTP server.

        :param fields: The fields to send as a list of (fieldname, value)
                       tuples.
        :param files: The files to send as a list of (fieldname, filename,
                      file_bytes) tuple.
        s--z)Content-Disposition: form-data; name="%s"r>z8Content-Disposition: form-data; name="%s"; filename="%s"s
smultipart/form-data; boundary=)zContent-typezContent-length)boundaryrrrrPrhrRstrrrr)r rr}partsrkvaluesvkeyrTvalueZbodyZctrr%r%r&r8sD


zPackageIndex.encode_requestcCsFt|trd|i}t|jdd}z|||p.dWS|dXdS)NrFg@)ZtimeoutrEand)rr
r
rsearch)r ZtermsoperatorZ	rpc_proxyr%r%r&rs
zPackageIndex.search)N)N)N)N)NNrkrlN)N)N)NN)N)__name__
__module____qualname____doc__rr'r+rr3r2r=rHrYrgrjr~rrrrr:r8rr%r%r%r&rs6



#

9


M+r)rvZloggingrrzrrQZ	threadingrImportErrorZdummy_threadingrcompatrrrrr	r
utilrrr
Z	getLoggerrrCrZ
DEFAULT_REALMobjectrr%r%r%r&s  
site-packages/pip/_vendor/distlib/__pycache__/compat.cpython-38.pyc000064400000076653151464452710021317 0ustar00U

.e@sBddlmZddlZddlZddlZzddlZWnek
rHdZYnXejddkr~ddlmZe	fZ
eZddl
mZddlZddlZddlmZddlmZmZmZmZmZdd	lmZmZmZm Z m!Z!m"Z"m#Z#d
dZddl$Z$ddl$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-er&dd
l$m.Z.ddl/Z/ddl0Z0ddl1Z2ddl3m3Z3ddl4Z4e5Z5ddl6m7Z8ddl6m9Z:da;ddZ<nddl=mZe>fZ
e>Zddl=m?ZddlZddlZddlZddl@mZmZmZmd?Zod@dAZiYnXzddBlpmqZqWn"ek
rddBlrmqZqYnXejddCdDkr:e3jsZsnddElpmsZszddFl`mtZtWndek
rddGl`muZuzddHlvmwZxWn ek
rdedJdKZxYnXGdLdMdMeuZtYnXzddNlymzZzWnHek
rzddNl{mzZzWn ek
rdfdOdPZzYnXYnXzddQl`m|Z|Wnek
rzddRl}m~ZWn"ek
rjddRlm~ZYnXzddSlmZmZmZWnek
rYnXGdTdUdUeZ|YnXzddVlmZmZWnvek
r<emdWejZdXdYZGdZd[d[eZdgd\d]ZGd^d_d_eZGd`dadaeZGdbdcdceQZYnXdS)h)absolute_importN)StringIO)FileType)shutil)urlparse
urlunparseurljoinurlsplit
urlunsplit)urlretrievequoteunquoteurl2pathnamepathname2urlContentTooShortError	splittypecCst|tr|d}t|S)Nutf-8)
isinstanceunicodeencode_quote)sr>/usr/lib/python3.8/site-packages/pip/_vendor/distlib/compat.pyrs

r)	RequesturlopenURLError	HTTPErrorHTTPBasicAuthHandlerHTTPPasswordMgrHTTPHandlerHTTPRedirectHandlerbuild_opener)HTTPSHandler)
HTMLParser)ifilter)ifilterfalsecCs<tdkrddl}|dat|}|r4|ddSd|fS)zJsplituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.Nrz^(.*)@(.*)$r)	_userprogrecompilematchgroup)Zhostr+r-rrr	splituser4s

r/)
TextIOWrapper)	rr	r
r/rrrrr)
rr
rrrr r!r"r#r$)rrr)filterfalse)match_hostnameCertificateErrorc@seZdZdS)r3N)__name__
__module____qualname__rrrrr3^sr3c
Csg}|sdS|d}|d|dd}}|d}||krNtdt||sb||kS|dkrv|dn>|d	s|d	r|t|n|t|	d
d|D]}|t|qt
dd
|dtj}	|	
|S)zpMatching according to RFC 6125, section 6.4.3

        http://tools.ietf.org/html/rfc6125#section-6.4.3
        F.rrN*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)splitcountr3reprlowerappend
startswithr+escapereplacer,join
IGNORECASEr-)
ZdnhostnameZ
max_wildcardsZpatspartsZleftmostZ	remainderZ	wildcardsfragZpatrrr_dnsname_matchbs*


rFcCs|stdg}|dd}|D]*\}}|dkr t||r@dS||q |s|ddD]6}|D],\}}|dkrdt||rdS||qdq\t|dkrtd	|d
tt|fn*t|dkrtd||dfntd
dS)a=Verify that *cert* (in decoded format as returned by
        SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
        rules are followed, but IP addresses are not accepted for *hostname*.

        CertificateError is raised on failure. On success, the function
        returns nothing.
        ztempty or no certificate, match_hostname needs a SSL socket or SSL context with either CERT_OPTIONAL or CERT_REQUIREDZsubjectAltNamerZDNSNZsubjectZ
commonNamerz&hostname %r doesn't match either of %s, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found)	
ValueErrorgetrFr=lenr3rAmapr;)ZcertrCZdnsnamesZsankeyvaluesubrrrr2s2


r2)SimpleNamespacec@seZdZdZddZdS)	ContainerzR
        A generic container for when multiple values need to be returned
        cKs|j|dSN__dict__update)selfkwargsrrr__init__szContainer.__init__N)r4r5r6__doc__rWrrrrrPsrP)whichcs"dd}tjr&||r"SdS|dkr>tjdtj}|sFdS|tj}tj	dkrtj
|krt|dtj
tjddtj}tfd	d
|Drg}q‡fdd|D}ng}t
}|D]P}tj|}||kr|||D](}	tj||	}
||
|r|
SqqdS)
aKGiven a command, mode, and a PATH string, return the path which
        conforms to the given mode on the PATH, or None if there is no such
        file.

        `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result
        of os.environ.get("PATH"), or can be overridden with a custom search
        path.

        cSs&tj|o$t||o$tj|SrQ)ospathexistsaccessisdir)fnmoderrr
_access_checkszwhich.._access_checkNPATHZwin32rZPATHEXTc3s |]}|VqdSrQ)r<endswith.0extcmdrr	szwhich..csg|]}|qSrrrerhrr
szwhich..)rZr[dirnameenvironrIdefpathr9pathsepsysplatformcurdirinsertanysetnormcaseaddrA)rir`r[raZpathextfilesseendirZnormdirZthefilenamerrhrrYs8





rY)ZipFile	__enter__)
ZipExtFilec@s$eZdZddZddZddZdS)r~cCs|j|jdSrQrR)rUbaserrrrWszZipExtFile.__init__cCs|SrQrrUrrrr}szZipExtFile.__enter__cGs|dSrQcloserUexc_inforrr__exit__szZipExtFile.__exit__N)r4r5r6rWr}rrrrrr~sr~c@s$eZdZddZddZddZdS)r|cCs|SrQrrrrrr}"szZipFile.__enter__cGs|dSrQrrrrrr%szZipFile.__exit__cOstj|f||}t|SrQ)BaseZipFileopenr~)rUargsrVrrrrr)szZipFile.openN)r4r5r6r}rrrrrrr|!sr|)python_implementationcCs0dtjkrdStjdkrdStjdr,dSdS)z6Return a string identifying the Python implementation.ZPyPyjavaZJythonZ
IronPythonZCPython)rpversionrZr{r>rrrrr0s

r)	sysconfig)CallablecCs
t|tSrQ)rr)objrrrcallableDsrrmbcsstrictsurrogateescapecCs:t|tr|St|tr$|ttStdt|jdSNzexpect bytes or str, not %s)	rbytes	text_typer_fsencoding	_fserrors	TypeErrortyper4filenamerrrfsencodeXs

rcCs:t|tr|St|tr$|ttStdt|jdSr)	rrrdecoderrrrr4rrrrfsdecodeas

r)detect_encoding)BOM_UTF8lookupzcoding[:=]\s*([-\w.]+)cCsH|dddd}|dks*|dr.dS|dks@|drDd	S|S)
z(Imitates get_normal_name in tokenizer.c.N_-rzutf-8-)zlatin-1
iso-8859-1ziso-latin-1)zlatin-1-ziso-8859-1-ziso-latin-1-r)r<r@r>)orig_encencrrr_get_normal_namersrcszjjWntk
r$dYnXdd}d}fdd}fdd}|}|trpd|d	d}d
}|s||gfS||}|r||gfS|}|s||gfS||}|r|||gfS|||gfS)a?
        The detect_encoding() function is used to detect the encoding that should
        be used to decode a Python source file.  It requires one argument, readline,
        in the same way as the tokenize() generator.

        It will call readline a maximum of twice, and return the encoding used
        (as a string) and a list of any lines (left as bytes) it has read in.

        It detects the encoding from the presence of a utf-8 bom or an encoding
        cookie as specified in pep-0263.  If both a bom and a cookie are present,
        but disagree, a SyntaxError will be raised.  If the encoding cookie is an
        invalid charset, raise a SyntaxError.  Note that if a utf-8 bom is found,
        'utf-8-sig' is returned.

        If no encoding is specified, then the default of 'utf-8' will be returned.
        NFrcs$zWStk
rYdSXdS)N)
StopIterationr)readlinerrread_or_stopsz%detect_encoding..read_or_stopcsz|d}Wn4tk
rBd}dk	r6d|}t|YnXt|}|sVdSt|d}zt|}Wn:tk
rdkrd|}nd|}t|YnXr|j	dkr؈dkrd}n
d}t||d	7}|S)
Nrz'invalid or missing encoding declarationz{} for {!r}rzunknown encoding: zunknown encoding for {!r}: {}zencoding problem: utf-8z encoding problem for {!r}: utf-8z-sig)
rUnicodeDecodeErrorformatSyntaxError	cookie_refindallrrLookupErrorr{)lineline_stringmsgZmatchesencodingcodec)	bom_foundrrrfind_cookies8



z$detect_encoding..find_cookieTrz	utf-8-sig)__self__r{AttributeErrorr>r)rrdefaultrrfirstsecondr)rrrrr}s4
&


r)r?r))r)unescape)ChainMap)MutableMapping)recursive_repr...csfdd}|S)zm
            Decorator to make a repr function return fillvalue for a recursive
            call
            csLtfdd}td|_td|_td|_tdi|_|S)Nc	sBt|tf}|krS|z|}W5|X|SrQ)id	get_identrwdiscard)rUrLresult)	fillvaluerepr_running
user_functionrrwrappers
z=_recursive_repr..decorating_function..wrapperr5rXr4__annotations__)rugetattrr5rXr4r)rrr)rrrdecorating_functionsz,_recursive_repr..decorating_functionr)rrrrr_recursive_reprsrc@seZdZdZddZddZddZd'd	d
ZddZd
dZ	ddZ
ddZeddZ
eddZddZeZddZeddZddZdd Zd!d"Zd#d$Zd%d&ZdS)(ra A ChainMap groups multiple dicts (or other mappings) together
        to create a single, updateable view.

        The underlying mappings are stored in a list.  That list is public and can
        accessed or updated using the *maps* attribute.  There is no other state.

        Lookups search the underlying mappings successively until a key is found.
        In contrast, writes, updates, and deletions only operate on the first
        mapping.

        cGst|pig|_dS)zInitialize a ChainMap by setting *maps* to the given mappings.
            If no mappings are provided, a single empty dictionary is used.

            N)listmaps)rUrrrrrWszChainMap.__init__cCst|dSrQ)KeyErrorrUrLrrr__missing__szChainMap.__missing__c	Cs:|jD](}z||WStk
r,YqXq||SrQ)rrr)rUrLmappingrrr__getitem__s
zChainMap.__getitem__NcCs||kr||S|SrQrrUrLrrrrrI%szChainMap.getcCsttj|jSrQ)rJruunionrrrrr__len__(szChainMap.__len__cCsttj|jSrQ)iterrurrrrrr__iter__+szChainMap.__iter__cstfdd|jDS)Nc3s|]}|kVqdSrQr)rfmrLrrrj/sz(ChainMap.__contains__..rtrrrrr__contains__.szChainMap.__contains__cCs
t|jSrQrrrrr__bool__1szChainMap.__bool__cCsd|dtt|jS)Nz{0.__class__.__name__}({1})rG)rrArKr;rrrrr__repr__4szChainMap.__repr__cGs|tj|f|S)z?Create a ChainMap with a single dict created from the iterable.)dictfromkeys)clsiterablerrrrr9szChainMap.fromkeyscCs$|j|jdf|jddS)zHNew ChainMap or subclass with a new copy of maps[0] and refs to maps[1:]rrN)	__class__rcopyrrrrr>sz
ChainMap.copycCs|jif|jS)z;New ChainMap with a new dict followed by all previous maps.rrrrrr	new_childDszChainMap.new_childcCs|j|jddS)zNew ChainMap from maps[1:].rNrrrrrparentsHszChainMap.parentscCs||jd|<dS)Nr)r)rUrLrMrrr__setitem__MszChainMap.__setitem__cCs8z|jd|=Wn"tk
r2td|YnXdS)Nr(Key not found in the first mapping: {!r})rrrrrrr__delitem__PszChainMap.__delitem__cCs2z|jdWStk
r,tdYnXdS)zPRemove and return an item pair from maps[0]. Raise KeyError is maps[0] is empty.rz#No keys found in the first mapping.N)rpopitemrrrrrrVszChainMap.popitemcGs@z|jdj|f|WStk
r:td|YnXdS)zWRemove *key* from maps[0] and return its value. Raise KeyError if *key* not in maps[0].rrN)rpoprr)rUrLrrrrr]szChainMap.popcCs|jddS)z'Clear maps[0], leaving maps[1:] intact.rN)rclearrrrrrdszChainMap.clear)N)r4r5r6rXrWrrrIrrrrrrclassmethodrr__copy__rpropertyrrrrrrrrrrrs.



r)cache_from_sourcecCs0|dst|dkrd}|r$d}nd}||S)Nz.pyTco)rdAssertionError)r[debug_overridesuffixrrrrnsr)OrderedDict)r)KeysView
ValuesView	ItemsViewc@seZdZdZddZejfddZejfddZdd	Zd
dZ	dd
Z
d6ddZddZddZ
ddZddZddZddZddZeZeZefdd Zd7d"d#Zd8d$d%Zd&d'Zd(d)Zed9d*d+Zd,d-Zd.d/Zd0d1Zd2d3Z d4d5Z!d!S):rz)Dictionary that remembers insertion ordercOsnt|dkrtdt|z
|jWn6tk
r\g|_}||dg|dd<i|_YnX|j||dS)zInitialize an ordered dictionary.  Signature is the same as for
            regular dictionaries, but keyword arguments are not recommended
            because their insertion order is arbitrary.

            rz$expected at most 1 arguments, got %dN)rJr_OrderedDict__rootr_OrderedDict__map_OrderedDict__update)rUrkwdsrootrrrrWs

zOrderedDict.__init__cCsF||kr6|j}|d}|||g|d<|d<|j|<||||dS)z!od.__setitem__(i, y) <==> od[i]=yrrN)rr)rUrLrMZdict_setitemrlastrrrrs
 zOrderedDict.__setitem__cCs0||||j|\}}}||d<||d<dS)z od.__delitem__(y) <==> del od[y]rrN)rr)rUrLZdict_delitem	link_prev	link_nextrrrrs
zOrderedDict.__delitem__ccs.|j}|d}||k	r*|dV|d}qdS)zod.__iter__() <==> iter(od)rr)NrrUrZcurrrrrrs

zOrderedDict.__iter__ccs.|j}|d}||k	r*|dV|d}qdS)z#od.__reversed__() <==> reversed(od)rr)Nrr	rrr__reversed__s

zOrderedDict.__reversed__cCsdz@|jD]}|dd=q|j}||dg|dd<|jWntk
rTYnXt|dS)z.od.clear() -> None.  Remove all items from od.N)r
itervaluesrrrr)rUZnoderrrrrszOrderedDict.clearTcCs||std|j}|r8|d}|d}||d<||d<n |d}|d}||d<||d<|d}|j|=t||}||fS)zod.popitem() -> (k, v), return and remove a (key, value) pair.
            Pairs are returned in LIFO order if last is true or FIFO order if false.

            zdictionary is emptyrrr))rrrrr)rUrrlinkrrrLrMrrrrs 
zOrderedDict.popitemcCst|S)zod.keys() -> list of keys in od)rrrrrkeysszOrderedDict.keyscsfddDS)z#od.values() -> list of values in odcsg|]}|qSrrrfrLrrrrksz&OrderedDict.values..rrrrrvaluesszOrderedDict.valuescsfddDS)z.od.items() -> list of (key, value) pairs in odcsg|]}||fqSrrrrrrrksz%OrderedDict.items..rrrrritemsszOrderedDict.itemscCst|S)z0od.iterkeys() -> an iterator over the keys in od)rrrrriterkeysszOrderedDict.iterkeysccs|D]}||VqdS)z2od.itervalues -> an iterator over the values in odNrrUkrrrrszOrderedDict.itervaluesccs|D]}|||fVqdS)z=od.iteritems -> an iterator over the (key, value) items in odNrrrrr	iteritemsszOrderedDict.iteritemscOst|dkr tdt|fn|s,td|d}d}t|dkrL|d}t|trn|D]}||||<qZn None.  Update od from dict/iterable E and F.

            If E is a dict instance, does:           for k in E: od[k] = E[k]
            If E has a .keys() method, does:         for k in E.keys(): od[k] = E[k]
            Or if E is an iterable of items, does:   for k, v in E: od[k] = v
            In either case, this is followed by:     for k, v in F.items(): od[k] = v

            r)z8update() takes at most 2 positional arguments (%d given)z,update() takes at least 1 argument (0 given)rrrr
N)rJrrrhasattrr
r)rrrUotherrLrMrrrrTs(	


zOrderedDict.updatecCs0||kr||}||=|S||jkr,t||S)zod.pop(k[,d]) -> v, remove specified key and return the corresponding value.
            If key is not found, d is returned if given, otherwise KeyError is raised.

            )_OrderedDict__markerr)rUrLrrrrrr*s
zOrderedDict.popNcCs||kr||S|||<|S)zDod.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in odrrrrr
setdefault7szOrderedDict.setdefaultcCsf|si}t|tf}||kr"dSd||<z.|sBd|jjfWSd|jj|fWS||=XdS)zod.__repr__() <==> repr(od)rrz%s()z%s(%r)N)r
_get_identrr4r)rUZ
_repr_runningZcall_keyrrrr>szOrderedDict.__repr__csXfddD}t}ttD]}||dq(|rLj|f|fSj|ffS)z%Return state information for picklingcsg|]}||gqSrrrfrrrrrkNsz*OrderedDict.__reduce__..N)varsrrrr)rUrZ	inst_dictrrrr
__reduce__LszOrderedDict.__reduce__cCs
||S)z!od.copy() -> a shallow copy of od)rrrrrrVszOrderedDict.copycCs|}|D]}|||<q
|S)zOD.fromkeys(S[, v]) -> New ordered dictionary with keys from S
            and values equal to v (which defaults to None).

            r)rrrMdrLrrrrZs
zOrderedDict.fromkeyscCs6t|tr*t|t|ko(||kSt||S)zod.__eq__(y) <==> od==y.  Comparison to another OD is order-sensitive
            while comparison to a regular mapping is order-insensitive.

            )rrrJrr__eq__rUrrrrres
 zOrderedDict.__eq__cCs
||kSrQrrrrr__ne__nszOrderedDict.__ne__cCst|S)z@od.viewkeys() -> a set-like object providing a view on od's keys)rrrrrviewkeyssszOrderedDict.viewkeyscCst|S)z an object providing a view on od's values)rrrrr
viewvalueswszOrderedDict.viewvaluescCst|S)zBod.viewitems() -> a set-like object providing a view on od's items)rrrrr	viewitems{szOrderedDict.viewitems)T)N)N)N)"r4r5r6rXrWrrrrr
rrr
rrrrrrTrobjectrrrrrrrrrr r!r"r#rrrrrs:
	




	r)BaseConfiguratorvalid_identz^[a-z_][a-z0-9_]*$cCst|}|std|dS)Nz!Not a valid Python identifier: %rT)
IDENTIFIERr-rH)rrrrrr&s
r&c@s"eZdZdZddZdddZdS)ConvertingDictz A converting dictionary wrapper.cCsJt||}|j|}||k	rF|||<t|tttfkrF||_||_	|SrQ)
rrconfiguratorconvertrr(ConvertingListConvertingTupleparentrLrUrLrMrrrrrs
zConvertingDict.__getitem__NcCsLt|||}|j|}||k	rH|||<t|tttfkrH||_||_	|SrQ)
rrIr)r*rr(r+r,r-rLrUrLrrMrrrrrIs
zConvertingDict.get)N)r4r5r6rXrrIrrrrr(sr(cCsDt|||}|j|}||k	r@t|tttfkr@||_||_	|SrQ)
rrr)r*rr(r+r,r-rLr/rrrrs
rc@s"eZdZdZddZdddZdS)	r+zA converting list wrapper.cCsJt||}|j|}||k	rF|||<t|tttfkrF||_||_	|SrQ)
rrr)r*rr(r+r,r-rLr.rrrrs
zConvertingList.__getitem__cCs<t||}|j|}||k	r8t|tttfkr8||_|SrQ)	rrr)r*rr(r+r,r-)rUidxrMrrrrrs
zConvertingList.popN)r0)r4r5r6rXrrrrrrr+sr+c@seZdZdZddZdS)r,zA converting tuple wrapper.cCsBt||}|j|}||k	r>t|tttfkr>||_||_	|SrQ)
tuplerr)r*rr(r+r,r-rLr.rrrrs
zConvertingTuple.__getitem__N)r4r5r6rXrrrrrr,sr,c@seZdZdZedZedZedZedZ	edZ
ddd	Zee
Zd
dZdd
ZddZddZddZddZddZdS)r%zQ
        The configurator base class which defines some useful defaults.
        z%^(?P[a-z]+)://(?P.*)$z^\s*(\w+)\s*z^\.\s*(\w+)\s*z^\[\s*(\w+)\s*\]\s*z^\d+$ext_convertcfg_convert)rgZcfgcCst||_||j_dSrQ)r(configr))rUr5rrrrWs
zBaseConfigurator.__init__c		Cs|d}|d}z^||}|D]H}|d|7}zt||}Wq$tk
rj||t||}Yq$Xq$|WStk
rtdd\}}td||f}|||_	|_
|YnXdS)zl
            Resolve strings to objects using standard import and attribute
            syntax.
            r7rrNzCannot resolve %r: %s)r9rimporterrrImportErrorrprrH	__cause__
__traceback__)	rUrr{ZusedfoundrEetbvrrrresolves"



zBaseConfigurator.resolvecCs
||S)z*Default converter for the ext:// protocol.)r>rUrMrrrr3szBaseConfigurator.ext_convertcCs|}|j|}|dkr&td|n||d}|j|d}|r|j|}|rn||d}nd|j|}|r|d}|j|s||}n2zt	|}||}Wnt
k
r||}YnX|r||d}qHtd||fqH|S)z*Default converter for the cfg:// protocol.NzUnable to convert %rrzUnable to convert %r at %r)WORD_PATTERNr-rHendr5groupsDOT_PATTERN
INDEX_PATTERN
DIGIT_PATTERNintr)rUrMrestrrr1nrrrr4s4
zBaseConfigurator.cfg_convertcCst|ts$t|tr$t|}||_nt|tsHt|trHt|}||_nzt|tslt|trlt|}||_nVt|tr|j	
|}|r|}|d}|j
|d}|r|d}t||}||}|S)z
            Convert values to an appropriate type. dicts, lists and tuples are
            replaced by their converting alternatives. Strings are checked to
            see if they have a conversion format and are converted if they do.
            prefixNr)rr(rr)r+rr,r2string_typesCONVERT_PATTERNr-	groupdictvalue_convertersrIr)rUrMrrrIZ	converterrrrrr*2s,


zBaseConfigurator.convertcsnd}t|s||}dd}tfddD}|f|}|rj|D]\}}t|||qT|S)z1Configure an object with a user-supplied factory.z()r7Ncs g|]}t|r||fqSr)r&rr5rrrkUsz5BaseConfigurator.configure_custom..)rrr>rrsetattr)rUr5rZpropsrVrr{rMrrNrconfigure_customNs


z!BaseConfigurator.configure_customcCst|trt|}|S)z0Utility function which converts lists to tuples.)rrr2r?rrras_tuple\s
zBaseConfigurator.as_tupleN)r4r5r6rXr+r,rKr@rCrDrErMstaticmethod
__import__r6rWr>r3r4r*rPrQrrrrr%s"




"r%)r)r)N)N)Z
__future__rrZr+rpZsslr7version_inforZ
basestringrJrrtypesrZ	file_typeZ__builtin__builtinsZConfigParserZconfigparserZ	_backportrrr	r
rrZurllibr
rrrrrrrZurllib2rrrrr r!r"r#r$r%ZhttplibZ	xmlrpclibZQueueZqueuer&ZhtmlentitydefsZ	raw_input	itertoolsr'filterr(r1r*r/iostrr0Zurllib.parseZurllib.requestZurllib.errorZhttp.clientZclientZrequestZ
xmlrpc.clientZhtml.parserZ
html.entitiesZentitiesinputr2r3rHrFrOrPr$rYF_OKX_OKZzipfiler|rrr~ZBaseZipExtFilerqrrr	NameErrorcollectionsrrrrgetfilesystemencodingrrtokenizercodecsrrr,rrZhtmlr?Zcgirrrreprlibrrimportlib.utilrZimprthreadrrZdummy_threadZ_abcollrrrrZlogging.configr%r&Ir'r(rrr+r2r,rrrrs,
$,,0
2+A


		
[
b
w

site-packages/pip/_vendor/distlib/__pycache__/version.cpython-38.pyc000064400000047553151464452710021516 0ustar00U

.e_[@sfdZddlZddlZddlmZddlmZdddd	d
ddd
gZee	Z
GdddeZGddde
ZGddde
ZedZddZeZGdddeZddZGdddeZeddfeddfeddfedd fed!d"fed#d"fed$d%fed&d'fed(d)fed*d+ff
Zed,dfed-dfed.d%fed$d%fed/dffZed0Zd1d2Zd3d4Zed5ejZd6d6d7d6d8ddd9Zd:d;ZGdejZ"d?d@Z#dAdBZ$GdCd
d
eZ%GdDddeZ&GdEdFdFe
Z'e'eeee'ee!dGdHe'e$e&edIZ(e(dJe(dK<dLd
Z)dS)Mz~
Implementation of a flexible versioning scheme providing support for PEP-440,
setuptools-compatible and semantic versioning.
N)string_typesparse_requirementNormalizedVersionNormalizedMatcher
LegacyVersion
LegacyMatcherSemanticVersionSemanticMatcherUnsupportedVersionError
get_schemec@seZdZdZdS)rzThis is an unsupported version.N)__name__
__module____qualname____doc__rr?/usr/lib/python3.8/site-packages/pip/_vendor/distlib/version.pyrsc@sxeZdZddZddZddZddZd	d
ZddZd
dZ	ddZ
ddZddZddZ
ddZeddZdS)VersioncCs@||_}|||_}t|ts,tt|dksTzMatcher.cCs||kSr)rr>rrrrBUrCcCs||kp||kSr)rr>rrrrBVrCcCs||kp||kSr)rr>rrrrBWrCcCs||kSr)rr>rrrrBXrCcCs||kSr)rr>rrrrBYrCcCs||kp||kSr)rr>rrrrB[rCcCs||kSr)rr>rrrrB\rC)<><=>======~=!=cCst|Sr)rr#rrrraszMatcher.parse_requirementcCs|jdkrtd||_}||}|s:td||j|_|j|_g}|jr|jD]d\}}|	dr|dkrtd||ddd}}||n||d}}|
|||fq^t||_dS)	NzPlease specify a version classz
Not valid: %rz.*)rHrKz#'.*' not allowed for %r constraintsTF)

version_class
ValueErrorrrrnamelowerkeyZconstraintsendswithappendrr)rrrZclistopZvnprefixrrrr ds*


zMatcher.__init__cCsxt|tr||}|jD]X\}}}|j|}t|trDt||}|s`d||jjf}t	|||||sdSqdS)z
        Check if the provided version matches the constraints.

        :param version: The version to match against this instance.
        :type version: String or :class:`Version` instance.
        z%r not implemented for %sFT)
rrrMr
_operatorsgetgetattrr7rr")rversionoperator
constraintrVfmsgrrrmatchs




z
Matcher.matchcCs6d}t|jdkr2|jdddkr2|jdd}|S)Nrr)rHrI)rr)rresultrrr
exact_versions zMatcher.exact_versioncCs0t|t|ks|j|jkr,td||fdS)Nzcannot compare %s and %s)r$rOr%r&rrrr(szMatcher._check_compatiblecCs"|||j|jko |j|jkSr))r(rQrr&rrrr+s
zMatcher.__eq__cCs||Sr)r,r&rrrr-szMatcher.__ne__cCst|jt|jSr))r3rQrr4rrrr5szMatcher.__hash__cCsd|jj|jfS)Nz%s(%r)r6r4rrrr8szMatcher.__repr__cCs|jSr)r9r4rrrr:szMatcher.__str__)rrrrMrWrr r_r<rar(r+r-r5r8r:rrrrr=Os*
r=zk^v?(\d+!)?(\d+(\.\d+)*)((a|b|c|rc)(\d+))?(\.(post)(\d+))?(\.(dev)(\d+))?(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$cCs|}t|}|s"td||}tdd|ddD}t|dkrl|ddkrl|dd}qF|dszd}nt|d}|dd	}|d
d}|dd
}|d}|dkrd}n|dt|df}|dkrd}n|dt|df}|dkrd}n|dt|df}|dkr*d}nHg}	|dD]0}
|
	rTdt|
f}
nd|
f}
|	
|
q8t|	}|s|s|rd}nd}|sd}|sd}||||||fS)NzNot a valid version: %scss|]}t|VqdSr)int.0r?rrr	sz_pep_440_key..r.r	

)NNr)arh)z)_)final)rPEP440_VERSION_REr_rgroupsrsplitrrcisdigitrS)rmruZnumsZepochpreZpostdevZlocalrpartrrr_pep_440_keysT



r|c@s6eZdZdZddZedddddgZed	d
ZdS)raIA rational version.

    Good:
        1.2         # equivalent to "1.2.0"
        1.2.0
        1.2a1
        1.2.3a2
        1.2.3b1
        1.2.3c1
        1.2.3.4
        TODO: fill this out

    Bad:
        1           # minimum two numbers
        1.2a        # release level must have a release serial
        1.2.3b
    cCs<t|}t|}|}tdd|ddD|_|S)Ncss|]}t|VqdSr)rbrdrrrrfsz*NormalizedVersion.parse..rrg)_normalized_keyrtr_rurrv_release_clause)rrr`rxrurrrr
s

zNormalizedVersion.parserpbr@rcrzcstfddjDS)Nc3s |]}|r|djkVqdS)rN)PREREL_TAGS)retr4rrrfsz2NormalizedVersion.is_prerelease..)anyrr4rr4rr;szNormalizedVersion.is_prereleaseN)	rrrrrsetrr<r;rrrrrs
cCs>t|}t|}||krdS||s*dSt|}||dkS)NTFrg)str
startswithr)xynrrr
_match_prefixs
rc	@sneZdZeZddddddddd	Zd
dZdd
ZddZddZ	ddZ
ddZddZddZ
ddZdS)r_match_compatible	_match_lt	_match_gt	_match_le	_match_ge	_match_eq_match_arbitrary	_match_ne)rJrDrErFrGrHrIrKcCsV|rd|ko|jd}n|jdo,|jd}|rN|jddd}||}||fS)N+rhrr)rrrvrM)rrZr\rVZstrip_localrrrr
_adjust_local6s
zNormalizedMatcher._adjust_localcCsD||||\}}||krdS|j}ddd|D}t||S)NFrgcSsg|]}t|qSrrreirrr
Isz/NormalizedMatcher._match_lt..rr~joinrrrZr\rVZrelease_clauseZpfxrrrrDszNormalizedMatcher._match_ltcCsD||||\}}||krdS|j}ddd|D}t||S)NFrgcSsg|]}t|qSrrrrrrrQsz/NormalizedMatcher._match_gt..rrrrrrLszNormalizedMatcher._match_gtcCs||||\}}||kSr)rrrZr\rVrrrrTszNormalizedMatcher._match_lecCs||||\}}||kSr)rrrrrrXszNormalizedMatcher._match_gecCs.||||\}}|s ||k}n
t||}|Sr)rrrrZr\rVr`rrrr\s


zNormalizedMatcher._match_eqcCst|t|kSr)rrrrrrdsz"NormalizedMatcher._match_arbitrarycCs0||||\}}|s ||k}nt||}|Sr)rrrrrrgs

zNormalizedMatcher._match_necCsf||||\}}||krdS||kr*dS|j}t|dkrH|dd}ddd|D}t||S)NTFrrhrgcSsg|]}t|qSrrrrrrrzsz7NormalizedMatcher._match_compatible..)rr~rrrrrrrrosz#NormalizedMatcher._match_compatibleN)rrrrrMrWrrrrrrrrrrrrrr's&z[.+-]$z^[.](\d)z0.\1z^[.-]z
^\((.*)\)$\1z^v(ersion)?\s*(\d+)z\2z^r(ev)?\s*(\d+)z[.]{2,}rgz\b(alfa|apha)\balphaz\b(pre-alpha|prealpha)\bz	pre.alphaz	\(beta\)$betaz
^[:~._+-]+z
[,*")([\]]z[~:+_ -]z\.$z
(\d+(\.\d+)*)cCsL|}tD]\}}|||}q|s.d}t|}|sFd}|}n|dd}dd|D}t|dkr~|	dqft|dkr||
d}n8ddd|ddD||
d}|dd}dd	d|D}|}|rtD]\}}|||}q|s|}nd
|kr*dnd}|||}t
|sHd}|S)
z
    Try to suggest a semantic form for a version for which
    _suggest_normalized_version couldn't come up with anything.
    z0.0.0rrgcSsg|]}t|qSrrbrrrrrsz-_suggest_semantic_version..NcSsg|]}t|qSrrrrrrrscSsg|]}t|qSrrrrrrrsrz-r)rrP
_REPLACEMENTSsub_NUMERIC_PREFIXr_rurvrrSendr_SUFFIX_REPLACEMENTS	is_semver)rr`ZpatreplrxrVsuffixseprrr_suggest_semantic_versions:
,
rcCshzt||WStk
r"YnX|}dD]\}}|||}q0tdd|}tdd|}tdd|}tdd	|}td
d|}|dr|d
d}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd|}tdd |}zt|Wntk
rbd}YnX|S)!aSuggest a normalized version close to the given version string.

    If you have a version string that isn't rational (i.e. NormalizedVersion
    doesn't like it) then you might be able to get an equivalent (or close)
    rational version from this function.

    This does a number of simple normalizations to the given string, based
    on observation of versions currently in use on PyPI. Given a dump of
    those version during PyCon 2009, 4287 of them:
    - 2312 (53.93%) match NormalizedVersion without change
      with the automatic suggestion
    - 3474 (81.04%) match when using this suggestion method

    @param s {str} An irrational version string.
    @returns A rational version string, or None, if couldn't determine one.
    ))z-alpharp)z-betar)rrp)rr)rr@)z-finalr)z-prer@)z-releaser)z.releaser)z-stabler)rrg)rrrg) r)z.finalr)rsrzpre$Zpre0zdev$Zdev0z([abc]|rc)[\-\.](\d+)$z\1\2z[\-\.](dev)[\-\.]?r?(\d+)$z.\1\2z[.~]?([abc])\.?rr?rNz\b0+(\d+)(?!\d)z(\d+[abc])$z\g<1>0z\.?(dev-r|dev\.r)\.?(\d+)$z.dev\2z-(a|b|c)(\d+)$z[\.\-](dev|devel)$z.dev0z(?![\.\-])dev$z(final|stable)$rz\.?(r|-|-r)\.?(\d+)$z.post\2z\.?(dev|git|bzr)\.?(\d+)$z\.?(pre|preview|-c)(\d+)$zc\g<2>zp(\d+)$z.post\1)r}rrPreplacererr)rZrsZorigrrrr_suggest_normalized_versions>
	
rz([a-z]+|\d+|[\.-])r@zfinal-@)ryZpreviewrrrzrrgcCsrdd}g}||D]T}|dr^|dkrD|rD|ddkrD|q*|r^|ddkr^|qD||qt|S)NcSstg}t|D]R}t||}|rd|ddkrBdkrRnn|d}nd|}||q|d|S)N0r9**final)
_VERSION_PARTrvrP_VERSION_REPLACErXzfillrS)rr`rArrr	get_partsCs 
z_legacy_key..get_partsrrrhz*final-Z00000000)rpoprSr)rrr`rArrr_legacy_keyBs



rc@s eZdZddZeddZdS)rcCst|Sr))rr#rrrr]szLegacyVersion.parsecCs8d}|jD](}t|tr
|dr
|dkr
d}q4q
|S)NFrrT)rrrr)rr`rrrrr;`s
zLegacyVersion.is_prereleaseNrrrrr<r;rrrrr\sc@s4eZdZeZeejZded<e	dZ
ddZdS)r	rrJz^(\d+(\.\d+)*)cCs`||krdS|jt|}|s2td||dS|d}d|krV|ddd}t||S)NFzACannot compute compatible match for version %s  and constraint %sTrrgr)
numeric_rer_rloggerZwarningrursplitr)rrZr\rVrxrrrrrsszLegacyMatcher._match_compatibleN)rrrrrMdictr=rWrcompilerrrrrrr	ks


zN^(\d+)\.(\d+)\.(\d+)(-[a-z0-9]+(\.[a-z0-9-]+)*)?(\+[a-z0-9]+(\.[a-z0-9-]+)*)?$cCs
t|Sr))
_SEMVER_REr_)rrrrrsrc	Csndd}t|}|st||}dd|ddD\}}}||dd||dd}}|||f||fS)	NcSs8|dkr|f}n$|ddd}tdd|D}|S)NrrgcSs"g|]}|r|dn|qS)r)rwr)rerArrrrsz5_semantic_key..make_tuple..)rvr)rZabsentr`rrrr
make_tuples
z!_semantic_key..make_tuplecSsg|]}t|qSrrbrrrrrsz!_semantic_key..r|r)rrru)	rrrxrumajorminorZpatchryZbuildrrr
_semantic_keys
rc@s eZdZddZeddZdS)r
cCst|Sr))rr#rrrrszSemanticVersion.parsecCs|jdddkS)Nrrr)rr4rrrr;szSemanticVersion.is_prereleaseNrrrrrr
sc@seZdZeZdS)rN)rrrr
rMrrrrrsc@s6eZdZdddZddZddZdd	Zd
dZdS)

VersionSchemeNcCs||_||_||_dSr))rQmatcher	suggester)rrQrrrrrr szVersionScheme.__init__cCs2z|j|d}Wntk
r,d}YnX|SNTF)rrMrrrr`rrris_valid_versions
zVersionScheme.is_valid_versioncCs0z||d}Wntk
r*d}YnX|Sr)rrrrrris_valid_matchers

zVersionScheme.is_valid_matchercCs|d|S)z:
        Used for processing some metadata fields
        zdummy_name (%s))rr#rrris_valid_constraint_listsz&VersionScheme.is_valid_constraint_listcCs|jdkrd}n
||}|Sr))rrrrrsuggests

zVersionScheme.suggest)N)rrrr rrrrrrrrrs

rcCs|Sr)rr#rrrrBrCrB)
normalizedlegacyZsemanticrdefaultcCs|tkrtd|t|S)Nzunknown scheme name: %r)_SCHEMESrN)rOrrrr
s)*rZloggingrcompatrutilr__all__Z	getLoggerrrrNrobjectrr=rrtr|r}rrrrrrrrIrrrrr	rrrr
rrrr
rrrrs
1d
=$W
.r	$site-packages/pip/_vendor/distlib/__pycache__/resources.cpython-38.pyc000064400000025334151464452710022034 0ustar00U

.e*	@sddlmZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlmZddl
mZmZmZmZeeZdaGdddeZGdd	d	eZGd
ddeZGdd
d
eZGdddeZGdddeZedee
jeiZzFzddl Z!Wne"k
r$ddl#Z!YnXeee!j$<eee!j%<[!Wne"e&fk
rXYnXddZ'iZ(ddZ)e	*e+dZ,ddZ-dS))unicode_literalsN)DistlibException)cached_propertyget_cache_basepath_to_cache_dirCachecs.eZdZdfdd	ZddZddZZS)	
ResourceCacheNcs0|dkrtjttd}tt||dS)Nzresource-cache)ospathjoinrstrsuperr	__init__)selfbase	__class__A/usr/lib/python3.8/site-packages/pip/_vendor/distlib/resources.pyrszResourceCache.__init__cCsdS)z
        Is the cache stale for the given resource?

        :param resource: The :class:`Resource` being cached.
        :param path: The path of the resource in the cache.
        :return: True if the cache is stale.
        Trrresourcerrrris_stale#s	zResourceCache.is_stalec	Cs|j|\}}|dkr|}n~tj|j|||}tj|}tj|sXt	|tj
|sjd}n|||}|rt|d}|
|jW5QRX|S)z
        Get a resource into the cache,

        :param resource: A :class:`Resource` instance.
        :return: The pathname of the resource in the cache.
        NTwb)finderget_cache_infor
rrrZ
prefix_to_dirdirnameisdirmakedirsexistsropenwritebytes)rrprefixrresultrZstalefrrrget.s
zResourceCache.get)N)__name__
__module____qualname__rrr&
__classcell__rrrrr	sr	c@seZdZddZdS)ResourceBasecCs||_||_dSN)rname)rrr-rrrrIszResourceBase.__init__N)r'r(r)rrrrrr+Hsr+c@s@eZdZdZdZddZeddZeddZed	d
Z	dS)Resourcez
    A class representing an in-package resource, such as a data file. This is
    not normally instantiated by user code, but rather by a
    :class:`ResourceFinder` which manages the resource.
    FcCs|j|S)z
        Get the resource as a stream.

        This is not a property to make it obvious that it returns a new stream
        each time.
        )r
get_streamrrrr	as_streamVszResource.as_streamcCstdkrtat|Sr,)cacher	r&r0rrr	file_path_szResource.file_pathcCs|j|Sr,)r	get_bytesr0rrrr"fszResource.bytescCs|j|Sr,)rget_sizer0rrrsizejsz
Resource.sizeN)
r'r(r)__doc__is_containerr1rr3r"r6rrrrr.Ns	

r.c@seZdZdZeddZdS)ResourceContainerTcCs|j|Sr,)r
get_resourcesr0rrr	resourcesrszResourceContainer.resourcesN)r'r(r)r8rr;rrrrr9osr9c@seZdZdZejdrdZndZddZddZ	d	d
Z
ddZd
dZddZ
ddZddZddZddZddZeejjZddZdS)ResourceFinderz4
    Resource finder for file system resources.
    java).pyc.pyoz.class)r>r?cCs.||_t|dd|_tjt|dd|_dS)N
__loader____file__)modulegetattrloaderr
rrr)rrCrrrrszResourceFinder.__init__cCstj|Sr,)r
rrealpathrrrrr_adjust_pathszResourceFinder._adjust_pathcCsBt|trd}nd}||}|d|jtjj|}||S)N//r)	
isinstancer"splitinsertrr
rrrH)r
resource_nameseppartsr$rrr
_make_paths

zResourceFinder._make_pathcCstj|Sr,)r
rrrGrrr_findszResourceFinder._findcCs
d|jfSr,)rrrrrrrszResourceFinder.get_cache_infocCsD||}||sd}n&||r0t||}n
t||}||_|Sr,)rQrR
_is_directoryr9r.r)rrNrr$rrrfinds



zResourceFinder.findcCst|jdSNrb)r rrSrrrr/szResourceFinder.get_streamc
Cs,t|jd}|W5QRSQRXdSrV)r rread)rrr%rrrr4szResourceFinder.get_bytescCstj|jSr,)r
rgetsizerSrrrr5szResourceFinder.get_sizecs*fddtfddt|jDS)Ncs|dko|jS)N__pycache__)endswithskipped_extensions)r%r0rralloweds
z-ResourceFinder.get_resources..allowedcsg|]}|r|qSrr).0r%)r]rr
sz0ResourceFinder.get_resources..)setr
listdirrrSr)r]rrr:szResourceFinder.get_resourcescCs||jSr,)rTrrSrrrr8szResourceFinder.is_containerccs||}|dk	r|g}|r|d}|V|jr|j}|jD]>}|sL|}nd||g}||}|jrv||q>|Vq>qdS)NrrJ)rUpopr8r-r;rappend)rrNrZtodoZrnamer-new_nameZchildrrriterators 



zResourceFinder.iteratorN)r'r(r)r7sysplatform
startswithr\rrHrQrRrrUr/r4r5r:r8staticmethodr
rrrTrerrrrr<ws"r<cs`eZdZdZfddZddZddZdd	Zd
dZdd
Z	ddZ
ddZddZZ
S)ZipResourceFinderz6
    Resource finder for resources in .zip files.
    csZtt|||jj}dt||_t|jdr>|jj|_nt	j
||_t|j|_dS)Nr_files)
rrjrrEarchivelen
prefix_lenhasattrrk	zipimport_zip_directory_cachesortedindex)rrCrlrrrrszZipResourceFinder.__init__cCs|Sr,rrGrrrrHszZipResourceFinder._adjust_pathcCs||jd}||jkrd}nX|r:|dtjkr:|tj}t|j|}z|j||}Wntk
rtd}YnX|st	d||j
jnt	d||j
j|S)NTFz_find failed: %r %rz_find worked: %r %r)rnrkr
rObisectrsrh
IndexErrorloggerdebugrEr#)rrr$irrrrRs


zZipResourceFinder._findcCs&|jj}|jdt|d}||fS)Nr)rErlrrm)rrr#rrrrrsz ZipResourceFinder.get_cache_infocCs|j|jSr,)rEget_datarrSrrrr4szZipResourceFinder.get_bytescCst||Sr,)ioBytesIOr4rSrrrr/szZipResourceFinder.get_streamcCs|j|jd}|j|dS)N)rrnrkrrrrr5szZipResourceFinder.get_sizecCs|j|jd}|r,|dtjkr,|tj7}t|}t}t|j|}|t|jkr|j||shq|j||d}|	|
tjdd|d7}qH|S)Nrtrr)rrnr
rOrmr`rursrhaddrL)rrrZplenr$rysrrrr:s

zZipResourceFinder.get_resourcescCsj||jd}|r*|dtjkr*|tj7}t|j|}z|j||}Wntk
rdd}YnX|S)NrtF)rnr
rOrursrhrv)rrryr$rrrrTs

zZipResourceFinder._is_directory)r'r(r)r7rrHrRrr4r/r5r:rTr*rrrrrjsrjcCs|tt|<dSr,)_finder_registrytype)rEfinder_makerrrrregister_finder0srcCs|tkrt|}nv|tjkr$t|tj|}t|dd}|dkrJtdt|dd}tt|}|dkrxtd|||}|t|<|S)z
    Return a resource finder for a package.
    :param package: The name of the package.
    :return: A :class:`ResourceFinder` instance for the package.
    __path__Nz8You cannot get a finder for a module, only for a packager@zUnable to locate finder for %r)	
_finder_cacherfmodules
__import__rDrrr&r)packager$rCrrErrrrr6s


rZ	__dummy__cCsRd}t|tj|}tt|}|rNt}tj	
|d|_||_||}|S)z
    Return a resource finder for a path, which should represent a container.

    :param path: The path.
    :return: A :class:`ResourceFinder` instance for the path.
    NrB)
pkgutilZget_importerrfpath_importer_cacher&rr
_dummy_moduler
rrrAr@)rr$rErrCrrrfinder_for_pathRs
r).Z
__future__rrur{Zloggingr
rZshutilrftypesrprBrutilrrrrZ	getLoggerr'rwr2r	objectr+r.r9r<rjrzipimporterr_frozen_importlib_externalZ_fiImportError_frozen_importlibSourceFileLoader
FileFinderAttributeErrorrrr
ModuleTyper
rrrrrrsN
,!ZN

site-packages/pip/_vendor/distlib/__pycache__/resources.cpython-38.opt-1.pyc000064400000025334151464452710022773 0ustar00U

.e*	@sddlmZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlmZddl
mZmZmZmZeeZdaGdddeZGdd	d	eZGd
ddeZGdd
d
eZGdddeZGdddeZedee
jeiZzFzddl Z!Wne"k
r$ddl#Z!YnXeee!j$<eee!j%<[!Wne"e&fk
rXYnXddZ'iZ(ddZ)e	*e+dZ,ddZ-dS))unicode_literalsN)DistlibException)cached_propertyget_cache_basepath_to_cache_dirCachecs.eZdZdfdd	ZddZddZZS)	
ResourceCacheNcs0|dkrtjttd}tt||dS)Nzresource-cache)ospathjoinrstrsuperr	__init__)selfbase	__class__A/usr/lib/python3.8/site-packages/pip/_vendor/distlib/resources.pyrszResourceCache.__init__cCsdS)z
        Is the cache stale for the given resource?

        :param resource: The :class:`Resource` being cached.
        :param path: The path of the resource in the cache.
        :return: True if the cache is stale.
        Trrresourcerrrris_stale#s	zResourceCache.is_stalec	Cs|j|\}}|dkr|}n~tj|j|||}tj|}tj|sXt	|tj
|sjd}n|||}|rt|d}|
|jW5QRX|S)z
        Get a resource into the cache,

        :param resource: A :class:`Resource` instance.
        :return: The pathname of the resource in the cache.
        NTwb)finderget_cache_infor
rrrZ
prefix_to_dirdirnameisdirmakedirsexistsropenwritebytes)rrprefixrresultrZstalefrrrget.s
zResourceCache.get)N)__name__
__module____qualname__rrr&
__classcell__rrrrr	sr	c@seZdZddZdS)ResourceBasecCs||_||_dSN)rname)rrr-rrrrIszResourceBase.__init__N)r'r(r)rrrrrr+Hsr+c@s@eZdZdZdZddZeddZeddZed	d
Z	dS)Resourcez
    A class representing an in-package resource, such as a data file. This is
    not normally instantiated by user code, but rather by a
    :class:`ResourceFinder` which manages the resource.
    FcCs|j|S)z
        Get the resource as a stream.

        This is not a property to make it obvious that it returns a new stream
        each time.
        )r
get_streamrrrr	as_streamVszResource.as_streamcCstdkrtat|Sr,)cacher	r&r0rrr	file_path_szResource.file_pathcCs|j|Sr,)r	get_bytesr0rrrr"fszResource.bytescCs|j|Sr,)rget_sizer0rrrsizejsz
Resource.sizeN)
r'r(r)__doc__is_containerr1rr3r"r6rrrrr.Ns	

r.c@seZdZdZeddZdS)ResourceContainerTcCs|j|Sr,)r
get_resourcesr0rrr	resourcesrszResourceContainer.resourcesN)r'r(r)r8rr;rrrrr9osr9c@seZdZdZejdrdZndZddZddZ	d	d
Z
ddZd
dZddZ
ddZddZddZddZddZeejjZddZdS)ResourceFinderz4
    Resource finder for file system resources.
    java).pyc.pyoz.class)r>r?cCs.||_t|dd|_tjt|dd|_dS)N
__loader____file__)modulegetattrloaderr
rrr)rrCrrrrszResourceFinder.__init__cCstj|Sr,)r
rrealpathrrrrr_adjust_pathszResourceFinder._adjust_pathcCsBt|trd}nd}||}|d|jtjj|}||S)N//r)	
isinstancer"splitinsertrr
rrrH)r
resource_nameseppartsr$rrr
_make_paths

zResourceFinder._make_pathcCstj|Sr,)r
rrrGrrr_findszResourceFinder._findcCs
d|jfSr,)rrrrrrrszResourceFinder.get_cache_infocCsD||}||sd}n&||r0t||}n
t||}||_|Sr,)rQrR
_is_directoryr9r.r)rrNrr$rrrfinds



zResourceFinder.findcCst|jdSNrb)r rrSrrrr/szResourceFinder.get_streamc
Cs,t|jd}|W5QRSQRXdSrV)r rread)rrr%rrrr4szResourceFinder.get_bytescCstj|jSr,)r
rgetsizerSrrrr5szResourceFinder.get_sizecs*fddtfddt|jDS)Ncs|dko|jS)N__pycache__)endswithskipped_extensions)r%r0rralloweds
z-ResourceFinder.get_resources..allowedcsg|]}|r|qSrr).0r%)r]rr
sz0ResourceFinder.get_resources..)setr
listdirrrSr)r]rrr:szResourceFinder.get_resourcescCs||jSr,)rTrrSrrrr8szResourceFinder.is_containerccs||}|dk	r|g}|r|d}|V|jr|j}|jD]>}|sL|}nd||g}||}|jrv||q>|Vq>qdS)NrrJ)rUpopr8r-r;rappend)rrNrZtodoZrnamer-new_nameZchildrrriterators 



zResourceFinder.iteratorN)r'r(r)r7sysplatform
startswithr\rrHrQrRrrUr/r4r5r:r8staticmethodr
rrrTrerrrrr<ws"r<cs`eZdZdZfddZddZddZdd	Zd
dZdd
Z	ddZ
ddZddZZ
S)ZipResourceFinderz6
    Resource finder for resources in .zip files.
    csZtt|||jj}dt||_t|jdr>|jj|_nt	j
||_t|j|_dS)Nr_files)
rrjrrEarchivelen
prefix_lenhasattrrk	zipimport_zip_directory_cachesortedindex)rrCrlrrrrszZipResourceFinder.__init__cCs|Sr,rrGrrrrHszZipResourceFinder._adjust_pathcCs||jd}||jkrd}nX|r:|dtjkr:|tj}t|j|}z|j||}Wntk
rtd}YnX|st	d||j
jnt	d||j
j|S)NTFz_find failed: %r %rz_find worked: %r %r)rnrkr
rObisectrsrh
IndexErrorloggerdebugrEr#)rrr$irrrrRs


zZipResourceFinder._findcCs&|jj}|jdt|d}||fS)Nr)rErlrrm)rrr#rrrrrsz ZipResourceFinder.get_cache_infocCs|j|jSr,)rEget_datarrSrrrr4szZipResourceFinder.get_bytescCst||Sr,)ioBytesIOr4rSrrrr/szZipResourceFinder.get_streamcCs|j|jd}|j|dS)N)rrnrkrrrrr5szZipResourceFinder.get_sizecCs|j|jd}|r,|dtjkr,|tj7}t|}t}t|j|}|t|jkr|j||shq|j||d}|	|
tjdd|d7}qH|S)Nrtrr)rrnr
rOrmr`rursrhaddrL)rrrZplenr$rysrrrr:s

zZipResourceFinder.get_resourcescCsj||jd}|r*|dtjkr*|tj7}t|j|}z|j||}Wntk
rdd}YnX|S)NrtF)rnr
rOrursrhrv)rrryr$rrrrTs

zZipResourceFinder._is_directory)r'r(r)r7rrHrRrr4r/r5r:rTr*rrrrrjsrjcCs|tt|<dSr,)_finder_registrytype)rEfinder_makerrrrregister_finder0srcCs|tkrt|}nv|tjkr$t|tj|}t|dd}|dkrJtdt|dd}tt|}|dkrxtd|||}|t|<|S)z
    Return a resource finder for a package.
    :param package: The name of the package.
    :return: A :class:`ResourceFinder` instance for the package.
    __path__Nz8You cannot get a finder for a module, only for a packager@zUnable to locate finder for %r)	
_finder_cacherfmodules
__import__rDrrr&r)packager$rCrrErrrrr6s


rZ	__dummy__cCsRd}t|tj|}tt|}|rNt}tj	
|d|_||_||}|S)z
    Return a resource finder for a path, which should represent a container.

    :param path: The path.
    :return: A :class:`ResourceFinder` instance for the path.
    NrB)
pkgutilZget_importerrfpath_importer_cacher&rr
_dummy_moduler
rrrAr@)rr$rErrCrrrfinder_for_pathRs
r).Z
__future__rrur{Zloggingr
rZshutilrftypesrprBrutilrrrrZ	getLoggerr'rwr2r	objectr+r.r9r<rjrzipimporterr_frozen_importlib_externalZ_fiImportError_frozen_importlibSourceFileLoader
FileFinderAttributeErrorrrr
ModuleTyper
rrrrrrsN
,!ZN

site-packages/pip/_vendor/distlib/__pycache__/database.cpython-38.opt-1.pyc000064400000122124151464452710022520 0ustar00U

.eU@sdZddlmZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlm
Z
mZddlmZddlmZmZddlmZmZmZmZdd	lmZmZmZmZmZmZm Z d
ddd
dgZ!e"e#Z$dZ%dZ&deddde%dfZ'dZ(Gddde)Z*Gddde)Z+Gdd
d
e)Z,Gddde,Z-Gddde-Z.Gdd
d
e-Z/e.Z0e/Z1Gddde)Z2d)d!d"Z3d#d$Z4d%d&Z5d'd(Z6dS)*zPEP 376 implementation.)unicode_literalsN)DistlibException	resources)StringIO)
get_schemeUnsupportedVersionError)MetadataMETADATA_FILENAMEWHEEL_METADATA_FILENAMELEGACY_METADATA_FILENAME)parse_requirementcached_propertyparse_name_and_versionread_exports
write_exports	CSVReader	CSVWriterDistributionBaseInstalledDistributionInstalledDistributionEggInfoDistributionDistributionPathzpydist-exports.jsonzpydist-commands.jsonZ	INSTALLERRECORD	REQUESTED	RESOURCESSHAREDz
.dist-infoc@s(eZdZdZddZddZddZdS)	_CachezL
    A simple cache mapping names and .dist-info paths to distributions
    cCsi|_i|_d|_dS)zZ
        Initialise an instance. There is normally one for each DistributionPath.
        FN)namepath	generatedselfr#@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/database.py__init__1sz_Cache.__init__cCs|j|jd|_dS)zC
        Clear the cache, setting it to its initial state.
        FN)rclearrr r!r#r#r$r&9s

z_Cache.clearcCs2|j|jkr.||j|j<|j|jg|dS)z`
        Add a distribution to the cache.
        :param dist: The distribution to add.
        N)rr
setdefaultkeyappendr"distr#r#r$addAsz
_Cache.addN)__name__
__module____qualname____doc__r%r&r,r#r#r#r$r-src@seZdZdZdddZddZdd	ZeeeZd
dZ	dd
Z
ddZeddZ
ddZddZdddZddZdddZdS)rzU
    Represents a set of distributions installed on a path (typically sys.path).
    NFcCsD|dkrtj}||_d|_||_t|_t|_d|_td|_	dS)a
        Create an instance from a path, optionally including legacy (distutils/
        setuptools/distribute) distributions.
        :param path: The path to use, as a list of directories. If not specified,
                     sys.path is used.
        :param include_egg: If True, this instance will look for and return legacy
                            distributions as well as those based on PEP 376.
        NTdefault)
sysr
_include_dist_include_eggr_cache
_cache_egg_cache_enabledr_scheme)r"rZinclude_eggr#r#r$r%Os	zDistributionPath.__init__cCs|jSNr7r!r#r#r$_get_cache_enabledcsz#DistributionPath._get_cache_enabledcCs
||_dSr9r:)r"valuer#r#r$_set_cache_enabledfsz#DistributionPath._set_cache_enabledcCs|j|jdS)z,
        Clears the internal cache.
        N)r5r&r6r!r#r#r$clear_cacheks
zDistributionPath.clear_cachec
csDt}|jD]0}t|}|dkr&q|d}|r|js||jjD]
}|Vq0|jrZ|jjD]
}|VqNdS)a5
        Provides an iterator that looks for distributions and returns
        :class:`InstalledDistribution` or
        :class:`EggInfoDistribution` instances for each one of them.

        :rtype: iterator of :class:`InstalledDistribution` and
                :class:`EggInfoDistribution` instances
        N)r7r\r^r5rvaluesr4r6r*r#r#r$get_distributionss	
z"DistributionPath.get_distributionscCsd}|}|js4|D]}|j|kr|}q|qnH|||jjkrZ|jj|d}n"|jr|||jjkr||jj|d}|S)a=
        Looks for a named distribution on the path.

        This function only returns the first result found, as no more than one
        value is expected. If nothing is found, ``None`` is returned.

        :rtype: :class:`InstalledDistribution`, :class:`EggInfoDistribution`
                or ``None``
        Nr)	lowerr7r\r(r^r5rr4r6)r"rresultr+r#r#r$get_distributions

z!DistributionPath.get_distributionc	csd}|dk	rJz|jd||f}Wn$tk
rHtd||fYnX|D]p}t|dsntd|qR|j}|D]H}t	|\}}|dkr||kr|VqRqx||krx|
|rx|VqRqxqRdS)a
        Iterates over all distributions to find which distributions provide *name*.
        If a *version* is provided, it will be used to filter the results.

        This function only returns the first result found, since no more than
        one values are expected. If the directory is not found, returns ``None``.

        :parameter version: a version specifier that indicates the version
                            required, conforming to the format in ``PEP-345``

        :type name: string
        :type version: string
        N%s (%s)zinvalid name or version: %r, %rprovideszNo "provides": %s)r8matcher
ValueErrorrrfhasattrrSrTrkrmatch)	r"rrcrlr+providedpp_namep_verr#r#r$provides_distributions*
z&DistributionPath.provides_distributioncCs(||}|dkrtd|||S)z5
        Return the path to a resource file.
        Nzno distribution named %r found)riLookupErrorget_resource_path)r"r
relative_pathr+r#r#r$
get_file_path!s
zDistributionPath.get_file_pathccsX|D]J}|j}||kr||}|dk	r>||krR||Vq|D]
}|VqFqdS)z
        Return all of the exported entries in a particular category.

        :param category: The category to search for entries.
        :param name: If specified, only entries with that name are returned.
        N)rfexportsre)r"categoryrr+rYdvr#r#r$get_exported_entries*sz%DistributionPath.get_exported_entries)NF)N)N)r-r.r/r0r%r;r=propertyZ
cache_enabledr>r\r^classmethodrdrfrirtrxr}r#r#r#r$rKs

,

)	c@seZdZdZdZdZddZeddZeZ	eddZ
ed	d
ZddZed
dZ
eddZeddZeddZeddZddZddZddZddZdS) rz
    A base class for distributions, whether installed or from indexes.
    Either way, it must have some metadata, so that's all that's needed
    for construction.
    FcCsL||_|j|_|j|_|j|_d|_d|_d|_d|_t	|_
i|_dS)z
        Initialise an instance.
        :param metadata: The instance of :class:`Metadata` describing this
        distribution.
        N)rDrrgr(rcZlocatordigestextrascontextrHZ
download_urlsZdigests)r"rDr#r#r$r%OszDistribution.__init__cCs|jjS)zH
        The source archive download URL for this distribution.
        )rD
source_urlr!r#r#r$r`szDistribution.source_urlcCsd|j|jfS)zX
        A utility property which displays the name and version in parentheses.
        rjrrcr!r#r#r$name_and_versioniszDistribution.name_and_versioncCs.|jj}d|j|jf}||kr*|||S)z
        A set of distribution names and versions provided by this distribution.
        :return: A set of "name (version)" strings.
        rj)rDrkrrcr))r"Zplistsr#r#r$rkps

zDistribution.providescCs8|j}td|t||}t|j||j|jdS)Nz%Getting requirements from metadata %r)rrE)	rDrSrTZtodictgetattrrHZget_requirementsrr)r"Zreq_attrmdZreqtsr#r#r$_get_requirements|s
zDistribution._get_requirementscCs
|dS)Nrun_requiresrr!r#r#r$rszDistribution.run_requirescCs
|dS)N
meta_requiresrr!r#r#r$rszDistribution.meta_requirescCs
|dS)Nbuild_requiresrr!r#r#r$rszDistribution.build_requirescCs
|dS)N
test_requiresrr!r#r#r$rszDistribution.test_requirescCs
|dS)Ndev_requiresrr!r#r#r$rszDistribution.dev_requiresc
	Cst|}t|jj}z||j}Wn6tk
rZtd||	d}||}YnX|j
}d}|jD]D}t|\}}	||krqlz|
|	}WqWqltk
rYqlXql|S)z
        Say if this instance matches (fulfills) a requirement.
        :param req: The requirement to match.
        :rtype req: str
        :return: True if it matches, else False.
        +could not read version %r - using name onlyrF)r
rrDrCrlrequirementrrSwarningsplitr(rkrro)
r"reqrYrCrlrrhrqrrrsr#r#r$matches_requirements,	


z Distribution.matches_requirementcCs(|jrd|j}nd}d|j|j|fS)zC
        Return a textual representation of this instance,
        z [%s]r?z)rrrc)r"suffixr#r#r$__repr__szDistribution.__repr__cCs>t|t|k	rd}n$|j|jko8|j|jko8|j|jk}|S)a<
        See if this distribution is the same as another.
        :param other: The distribution to compare with. To be equal to one
                      another. distributions must have the same type, name,
                      version and source_url.
        :return: True if it is the same, else False.
        F)typerrcr)r"otherrhr#r#r$__eq__s

zDistribution.__eq__cCst|jt|jt|jS)zH
        Compute hash in a way which matches the equality test.
        )hashrrcrr!r#r#r$__hash__szDistribution.__hash__N)r-r.r/r0Zbuild_time_dependency	requestedr%r~rZdownload_urlrrkrrrrrrrrrrr#r#r#r$r=s4







"
cs0eZdZdZdZdfdd	ZdddZZS)	rz]
    This is the base class for installed distributions (whether PEP 376 or
    legacy).
    Ncs tt||||_||_dS)a
        Initialise an instance.
        :param metadata: An instance of :class:`Metadata` which describes the
                         distribution. This will normally have been initialised
                         from a metadata file in the ``path``.
        :param path:     The path of the ``.dist-info`` or ``.egg-info``
                         directory for the distribution.
        :param env:      This is normally the :class:`DistributionPath`
                         instance where this distribution was found.
        N)superrr%r	dist_path)r"rDrrE	__class__r#r$r%sz"BaseInstalledDistribution.__init__cCsd|dkr|j}|dkr"tj}d}ntt|}d|j}||}t|dd}d||fS)a
        Get the hash of some data, using a particular hash algorithm, if
        specified.

        :param data: The data to be hashed.
        :type data: bytes
        :param hasher: The name of a hash implementation, supported by hashlib,
                       or ``None``. Examples of valid values are ``'sha1'``,
                       ``'sha224'``, ``'sha384'``, '``sha256'``, ``'md5'`` and
                       ``'sha512'``. If no hasher is specified, the ``hasher``
                       attribute of the :class:`InstalledDistribution` instance
                       is used. If the hasher is determined to be ``None``, MD5
                       is used as the hashing algorithm.
        :returns: The hash of the data. If a hasher was explicitly specified,
                  the returned hash will be prefixed with the specified hasher
                  followed by '='.
        :rtype: str
        Nr?z%s==ascii%s%s)	hasherhashlibmd5rrbase64Zurlsafe_b64encoderstripdecode)r"datarprefixrr#r#r$get_hashs

z"BaseInstalledDistribution.get_hash)N)N)r-r.r/r0rr%r
__classcell__r#r#rr$rscseZdZdZdZd'fdd	ZddZdd	Zd
dZe	dd
Z
ddZddZddZ
ddZd(ddZddZe	ddZd)ddZdd Zd!d"Zd#d$Zd%d&ZejZZS)*ra
    Created with the *path* of the ``.dist-info`` directory provided to the
    constructor. It reads the metadata contained in ``pydist.json`` when it is
    instantiated., or uses a passed in Metadata instance (useful for when
    dry-run mode is being used).
    Zsha256Nc
	sJg|_t||_}|dkr*td||rP|jrP||jjkrP|jj|j}nt|dkr|	t
}|dkrt|	t}|dkr|	d}|dkrtdt
|ft
|}t|dd}W5QRXtt|||||r|jr|j||	d}|dk	|_tj|d}tj|rFt|d}|}	W5QRX|	|_dS)	Nzfinder unavailable for %sZMETADATAzno %s found in %sr@rAr
top_level.txtrb)modulesrrIrXrmr7r5rrDrJr
rrPrQrRr	rrr%r,rosrOexistsopenread
splitlines)
r"rrDrErXrYr[rqfrrr#r$r%s8





zInstalledDistribution.__init__cCsd|j|j|jfS)Nz#rrcrr!r#r#r$r=s
zInstalledDistribution.__repr__cCsd|j|jfSNz%s %srr!r#r#r$__str__AszInstalledDistribution.__str__c

Csg}|d}t|\}t|dF}|D]:}ddtt|dD}||\}}}	||||	fq.W5QRXW5QRX|S)a"
        Get the list of installed files for the distribution
        :return: A list of tuples of path, hash and size. Note that hash and
                 size might be ``None`` for some entries. The path is exactly
                 as stored in the file (which is as in PEP 376).
        rr[cSsg|]}dqSr9r#).0ir#r#r$
Ssz6InstalledDistribution._get_records..)get_distinfo_resourcerPrQrRrrangelenr))
r"resultsrYr[Z
record_readerrowmissingrchecksumsizer#r#r$_get_recordsDs
&z"InstalledDistribution._get_recordscCsi}|t}|r|}|S)a
        Return the information exported by this distribution.
        :return: A dictionary of exports, mapping an export category to a dict
                 of :class:`ExportEntry` instances describing the individual
                 export entries, and keyed by name.
        )rEXPORTS_FILENAMEr)r"rhrYr#r#r$ry[s

zInstalledDistribution.exportsc	Cs8i}|t}|r4t|}t|}W5QRX|S)z
        Read exports data from a file in .ini format.

        :return: A dictionary of exports, mapping an export category to a list
                 of :class:`ExportEntry` instances describing the individual
                 export entries.
        )rrrPrQrRr)r"rhrYr[r#r#r$ris
z"InstalledDistribution.read_exportsc	Cs.|t}t|d}t||W5QRXdS)a
        Write a dictionary of exports to a file in .ini format.
        :param exports: A dictionary of exports, mapping an export category to
                        a list of :class:`ExportEntry` instances describing the
                        individual export entries.
        wN)get_distinfo_filerrr)r"ryZrfrr#r#r$rxs
z#InstalledDistribution.write_exportscCs|d}t|R}t|d<}|D]0\}}||kr*|W5QRW5QRSq*W5QRXW5QRXtd|dS)aW
        NOTE: This API may change in the future.

        Return the absolute path to a resource file with the given relative
        path.

        :param relative_path: The path, relative to .dist-info, of the resource
                              of interest.
        :return: The absolute path where the resource is to be found.
        rrz3no resource file with relative path %r is installedN)rrPrQrRrKeyError)r"rwrYr[Zresources_readerrelativeZdestinationr#r#r$rvs
6z'InstalledDistribution.get_resource_pathccs|D]
}|VqdS)z
        Iterates over the ``RECORD`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: iterator of (path, hash, size)
        N)r)r"rhr#r#r$list_installed_filessz*InstalledDistribution.list_installed_filesFcCs(tj|d}tj|j}||}tj|d}|d}td||rRdSt|}|D]}tj	|sz|
drd}	}
n4dtj|}
t|d}|
|}	W5QRX||s|r||rtj||}|||	|
fq`||r
tj||}||ddfW5QRX|S)z
        Writes the ``RECORD`` file, using the ``paths`` iterable passed in. Any
        existing ``RECORD`` file is silently overwritten.

        prefix is used to determine when to write absolute paths.
        r?rcreating %sNz.pycz.pyoz%dr)rrrOdirname
startswithrrSinforisdirrLgetsizerrrrelpathZwriterow)r"pathsrdry_runbaseZbase_under_prefixrecord_pathwriterr
hash_valuerfpr#r#r$write_installed_filess0



z+InstalledDistribution.write_installed_filesc
Csg}tj|j}|d}|D]\}}}tj|sHtj||}||krRq$tj|sr||dddfq$tj	|r$t
tj|}|r||kr||d||fq$|r$d|kr|ddd}nd	}t
|d
2}	||	|}
|
|kr||d||
fW5QRXq$|S)
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        rrTFr=rrNrr)rrrrrisabsrOrr)isfilestrrrrrr)r"
mismatchesrrrrrZactual_sizerrZactual_hashr#r#r$check_installed_filess.	

z+InstalledDistribution.check_installed_filesc	Csi}tj|jd}tj|rtj|ddd}|}W5QRX|D]8}|dd\}}|dkr||	|g
|qL|||<qL|S)a
        A dictionary of shared locations whose keys are in the set 'prefix',
        'purelib', 'platlib', 'scripts', 'headers', 'data' and 'namespace'.
        The corresponding value is the absolute path of that category for
        this distribution, and takes into account any paths selected by the
        user at installation time (e.g. via command-line arguments). In the
        case of the 'namespace' key, this would be a list of absolute paths
        for the roots of namespace packages in this distribution.

        The first time this property is accessed, the relevant information is
        read from the SHARED file in the .dist-info directory.
        rrYutf-8encodingrr	namespace)rrrOrcodecsrrrrr'r))r"rhshared_pathrlinesliner(r<r#r#r$shared_locationss
z&InstalledDistribution.shared_locationsc		Cstj|jd}td||r$dSg}dD].}||}tj||r,|d||fq,|ddD]}|d|qhtj	|d	d
d}|
d|W5QRX|S)
aa
        Write shared location information to the SHARED file in .dist-info.
        :param paths: A dictionary as described in the documentation for
        :meth:`shared_locations`.
        :param dry_run: If True, the action is logged but no file is actually
                        written.
        :return: The path of the file written to.
        rrN)rlibZheadersZscriptsrz%s=%srr#znamespace=%srrr
)rrrOrSrrr)getrrwrite)	r"rrrrr(rnsrr#r#r$write_shared_locationss	z,InstalledDistribution.write_shared_locationscCsF|tkrtd||jft|j}|dkr.set_name_and_version)rrr7r6rDrrc
_get_metadatar,rrr%)r"rrErrDrr#r$r%bs

zEggInfoDistribution.__init__c	sd}ddfdd}d}}|drtj|rtj|d}tj|d}t|dd	}tj|d
}	tj|d}||	}npt|}
t|
	d
d
}t|dd}z,|
	d}|
	d
d}|
d}Wntk
rd}YnXnf|drPtj|rBtj|d
}	||	}tj|d}tj|d}t|dd	}ntd||rl|
||dkr|dk	rtj|rt|d}
|

d}W5QRX|sg}n|}||_|S)NcSsg}|}|D]}|}|dr6td|qt|}|sPtd|q|jr`td|jst||j	qd
dd|jD}|d|j	|fq|S)	zCreate a list of dependencies from a requires.txt file.

            *data*: the contents of a setuptools-produced requires.txt file.
            [z.Unexpected line: quitting requirement scan: %rz#Not recognised as a requirement: %rz4extra requirements in requires.txt are not supportedz, css|]}d|VqdS)rNr#)rcr#r#r$	szQEggInfoDistribution._get_metadata..parse_requires_data..rj)rstriprrSrr
rZconstraintsr)rrO)rreqsrrrYZconsr#r#r$parse_requires_datazs(

z>EggInfoDistribution._get_metadata..parse_requires_datac	sHg}z*t|dd}|}W5QRXWntk
rBYnX|S)zCreate a list of dependencies from a requires.txt file.

            *req_path*: the path to a setuptools-produced requires.txt file.
            rYr)rrrIOError)req_pathrrrr#r$parse_requires_pathsz>EggInfoDistribution._get_metadata..parse_requires_pathrGzEGG-INFOzPKG-INFOr@)rrCzrequires.txtrzEGG-INFO/PKG-INFOutf8rAzEGG-INFO/requires.txtzEGG-INFO/top_level.txtrrFz,path must end with .egg-info or .egg, got %rr)rLrrrrOr		zipimportzipimporterrget_datarrrZadd_requirementsrrrrr)r"rrequiresr	Ztl_pathZtl_datarq	meta_pathrDrZzipfrBrrr#rr$rwsX





z!EggInfoDistribution._get_metadatacCsd|j|j|jfS)Nz!rr!r#r#r$rs
zEggInfoDistribution.__repr__cCsd|j|jfSrrr!r#r#r$rszEggInfoDistribution.__str__cCs`g}tj|jd}tj|r\|D]2\}}}||kr._md5cSst|jSr9)rstatst_size)rr#r#r$_sizesz7EggInfoDistribution.list_installed_files.._sizerrYrrzNon-existent file: %srN)
rrrOrrrrnormpathrSrrLrr))r"rrrrhrrrqr#r#r$rs"
$z(EggInfoDistribution.list_installed_filesFc	cstj|jd}tj|rd}tj|ddd`}|D]T}|}|dkrPd}q6|s6tjtj|j|}||jr6|r|Vq6|Vq6W5QRXdS)	a
        Iterates over the ``installed-files.txt`` entries and returns paths for
        each line if the path is pointing to a file located in the
        ``.egg-info`` directory or one of its subdirectories.

        :parameter absolute: If *absolute* is ``True``, each returned path is
                          transformed into a local absolute path. Otherwise the
                          raw value from ``installed-files.txt`` is returned.
        :type absolute: boolean
        :returns: iterator of paths
        rTrYrrz./FN)	rrrOrrrrrr)r"Zabsoluterskiprrrqr#r#r$rsz'EggInfoDistribution.list_distinfo_filescCst|to|j|jkSr9)r]rrrr#r#r$r.s

zEggInfoDistribution.__eq__)N)F)r-r.r/r0rrr%rrrrrrrrrrr#r#rr$rYsZ&
c@s^eZdZdZddZddZdddZd	d
ZddZdddZ	dddZ
ddZddZdS)DependencyGrapha
    Represents a dependency graph between distributions.

    The dependency relationships are stored in an ``adjacency_list`` that maps
    distributions to a list of ``(other, label)`` tuples where  ``other``
    is a distribution and the edge is labeled with ``label`` (i.e. the version
    specifier, if such was provided). Also, for more efficient traversal, for
    every distribution ``x``, a list of predecessors is kept in
    ``reverse_list[x]``. An edge from distribution ``a`` to
    distribution ``b`` means that ``a`` depends on ``b``. If any missing
    dependencies are found, they are stored in ``missing``, which is a
    dictionary that maps distributions to a list of requirements that were not
    provided by any other distributions.
    cCsi|_i|_i|_dSr9)adjacency_listreverse_listrr!r#r#r$r%IszDependencyGraph.__init__cCsg|j|<g|j|<dS)zAdd the *distribution* to the graph.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        N)rr)r"distributionr#r#r$add_distributionNs
z DependencyGraph.add_distributionNcCs6|j|||f||j|kr2|j||dS)aAdd an edge from distribution *x* to distribution *y* with the given
        *label*.

        :type x: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type y: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type label: ``str`` or ``None``
        N)rr)r)r"xylabelr#r#r$add_edgeXs
zDependencyGraph.add_edgecCs&td|||j|g|dS)a
        Add a missing *requirement* for the given *distribution*.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        :type requirement: ``str``
        z
%s missing %rN)rSrTrr'r))r"rrr#r#r$add_missinggszDependencyGraph.add_missingcCsd|j|jfSrrr*r#r#r$
_repr_distrszDependencyGraph._repr_distrcCs||g}|j|D]h\}}||}|dk	r|d|j|jfq>q|st|dkr|d|d|d|D]}|d	|j|d
q|d|ddS)a9Writes a DOT output for the graph to the provided file *f*.

        If *skip_disconnected* is set to ``True``, then all distributions
        that are not dependent on any other distribution are skipped.

        :type f: has to support ``file``-like operations
        :type skip_disconnected: ``bool``
        zdigraph dependencies {
rNz"%s" -> "%s" [label="%s"]
z
"%s" -> "%s"
zsubgraph disconnected {
zlabel = "Disconnected"
zbgcolor = red
z"%s"rz}
)rritemsrr)r)r"rZskip_disconnectedZdisconnectedr+adjsrrr#r#r$to_dots(	





zDependencyGraph.to_dotcsg}i}|jD]\}}|dd||<qgt|ddD]\}}|sD|||=qDshq|D]\}}fdd|D||<qptdddD|q,|t|fS)aa
        Perform a topological sort of the graph.
        :return: A tuple, the first element of which is a topologically sorted
                 list of distributions, and the second element of which is a
                 list of distributions that cannot be sorted because they have
                 circular dependencies and so form a cycle.
        Ncs g|]\}}|kr||fqSr#r#)rr{rYZ	to_remover#r$rsz4DependencyGraph.topological_sort..zMoving to result: %scSsg|]}d|j|jfqS)rjr)rr{r#r#r$rs)rr'listr)rSrTr$keys)r"rhZalistkr|r#r*r$topological_sorts$
z DependencyGraph.topological_sortcCs2g}|jD]\}}|||qd|S)zRepresentation of the graphr)rr'r)r#rO)r"r&r+r(r#r#r$rszDependencyGraph.__repr__)N)r)T)
r-r.r/r0r%rr r!r"r#r)r.rr#r#r#r$r9s



 rr1cCsVt|}t}i}|D]L}|||jD]6}t|\}}td|||||g||fq*q|D]}|j	|j
B|jB|jB}|D]}	z|
|	}
Wn6tk
rtd|	|	d}|
|}
YnX|
j}d}||kr>||D]N\}}z|
|}
Wntk
rd}
YnX|
r||||	d}q>q|s|||	qqh|S)a6Makes a dependency graph from the given distributions.

    :parameter dists: a list of distributions
    :type dists: list of :class:`distutils2.database.InstalledDistribution` and
                 :class:`distutils2.database.EggInfoDistribution` instances
    :rtype: a :class:`DependencyGraph` instance
    zAdd to provided: %s, %s, %srrFT)rrrrkrrSrTr'r)rrrrrlrrrr(ror r!)distsrCgraphrpr+rqrrcrrrlZmatchedZproviderror#r#r$
make_graphsN




r1cCsv||krtd|jt|}|g}|j|}|rh|}|||j|D]}||krN||qNq.|d|S)zRecursively generate a list of distributions from *dists* that are
    dependent on *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    1given distribution %r is not a member of the listr)rrr1rpopr))r/r+r0Zdeptodor{Zsuccr#r#r$get_dependent_distss


r5cCsn||krtd|jt|}g}|j|}|rj|d}|||j|D]}||krP||qPq,|S)zRecursively generate a list of distributions from *dists* that are
    required by *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    r2r)rrr1rr3r))r/r+r0rr4r{Zpredr#r#r$get_required_distss

r6cKs4|dd}tf|}||_||_|p(d|_t|S)zO
    A convenience method for making a dist given just a name and version.
    summaryzPlaceholder for summary)r3r	rrcr7r)rrckwargsr7rr#r#r$	make_dist2s

r9)r1)7r0Z
__future__rrrrPrZloggingrrNr2rr?rrcompatrrcrrrDr	r
rrutilr
rrrrrr__all__Z	getLoggerr-rSrZCOMMANDS_FILENAMErrMrrrrrrrrUrVrr1r5r6r9r#r#r#r$s`$
s7J]
6site-packages/pip/_vendor/distlib/__pycache__/metadata.cpython-38.opt-1.pyc000064400000063655151464452710022551 0ustar00U

.e*@sdZddlmZddlZddlmZddlZddlZddlZddl	m
Z
mZddlm
Z
mZmZddlmZdd	lmZmZdd
lmZmZeeZGddde
ZGd
dde
ZGddde
ZGddde
ZdddgZdZ dZ!e"dZ#e"dZ$dZ%dZ&dZ'dZ(dZ)dZ*d Z+e*d!Z,d"Z-e.Z/e/0e%e/0e&e/0e(e/0e*e/0e,e"d#Z1d$d%Z2d&d'Z3d(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFZ4dGZ5dHZ6dIZ7dJZ8dKZ9dLZ:dMZ;e<Z=e"dNZ>dXdPdQZ?GdRdSdSe<Z@dTZAdUZBdVZCGdWdde<ZDdS)YzImplementation of the Metadata for Python packages PEPs.

Supports all metadata formats (1.0, 1.1, 1.2, and 2.0 experimental).
)unicode_literalsN)message_from_file)DistlibException__version__)StringIOstring_types	text_type)	interpret)extract_by_key
get_extras)
get_schemePEP440_VERSION_REc@seZdZdZdS)MetadataMissingErrorzA required metadata is missingN__name__
__module____qualname____doc__rr@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/metadata.pyrsrc@seZdZdZdS)MetadataConflictErrorz>Attempt to read or write metadata fields that are conflictual.Nrrrrrr src@seZdZdZdS) MetadataUnrecognizedVersionErrorz Unknown metadata version number.Nrrrrrr$src@seZdZdZdS)MetadataInvalidErrorzA metadata value is invalidNrrrrrr(srMetadataPKG_INFO_ENCODINGPKG_INFO_PREFERRED_VERSIONutf-81.1z

       \|	
        )Metadata-VersionNameVersionPlatformSummaryDescriptionKeywords	Home-pageAuthorAuthor-emailLicense)r r!r"r#Supported-Platformr$r%r&r'r(r)r*
ClassifierDownload-URL	ObsoletesProvidesRequires)r.r/r0r,r-)r r!r"r#r+r$r%r&r'r(r)
MaintainerMaintainer-emailr*r,r-Obsoletes-DistProject-URL
Provides-Dist
Requires-DistRequires-PythonRequires-External)r5r6r7r3r8r1r2r4)r r!r"r#r+r$r%r&r'r(r)r1r2r*r,r-r3r4r5r6r7r8Private-VersionObsoleted-BySetup-Requires-Dist	ExtensionProvides-Extra)r9r=r:r;r<)Description-Content-Typer0r/)r>z"extra\s*==\s*("([^"]+)"|'([^']+)')cCsL|dkrtS|dkrtS|dkr$tS|dkr4ttS|dkr@tSt|dS)N1.0r1.2)1.32.12.0)_241_FIELDS_314_FIELDS_345_FIELDS_566_FIELDS_426_FIELDSr)versionrrr_version2fieldlistpsrJc
CsBdd}g}|D]"\}}|gddfkr,q||qddddd	d
g}|D]}|tkrvd|krv|dtd||tkrd|kr|dtd||tkrd|kr|dtd
||tkrd|kr|dtd||tkrd
|kr|dkr|d
td||t	krLd	|krL|d	td|qLt
|dkrZ|dSt
|dkr|td|tdd|ko||t}d|ko||t
}d
|ko||t}d	|ko||t}	t|t|t|t|	dkrtd|s |s |s |	s t|kr tS|r*dS|r4dS|r>d
Sd	S)z5Detect the best version depending on the fields used.cSs|D]}||krdSqdS)NTFr)keysmarkersmarkerrrr_has_markersz"_best_version.._has_markerUNKNOWNNr?rr@rArCrBzRemoved 1.0 due to %szRemoved 1.1 due to %szRemoved 1.2 due to %szRemoved 1.3 due to %sr%zRemoved 2.1 due to %szRemoved 2.0 due to %srrz)Out of options - unknown metadata set: %szUnknown metadata setz,You used incompatible 1.1/1.2/2.0/2.1 fields)itemsappendrDremoveloggerdebugrErFrGrHlenr_314_MARKERS_345_MARKERS_566_MARKERS_426_MARKERSintr)
fieldsrNrKkeyvalueZpossible_versionsZis_1_1Zis_1_2Zis_2_1Zis_2_0rrr
_best_version~s`






&
r^r r!r"r#r+r$r%r&r'r(r)r1r2r*r,r-r3r5r6r;r7r8r0r/r.r4r9r:r<r=)metadata_versionnamerIplatformZsupported_platformsummarydescriptionkeywords	home_pageauthorauthor_email
maintainermaintainer_emaillicense
classifierdownload_urlobsoletes_dist
provides_dist
requires_distsetup_requires_distrequires_pythonrequires_externalrequiresprovides	obsoletesproject_urlZprivate_versionZobsoleted_by	extensionZprovides_extra)r6r3r5)r7)r")r#r,r.r0r/r3r5r6r8r4r+r;r=r<)r4)r&)r(r1r$r%z[^A-Za-z0-9.]+FcCs0|r$td|}td|dd}d||fS)zhReturn the distribution name with version.

    If for_filename is true, return a filename-escaped form.- .z%s-%s)	_FILESAFEsubreplace)r`rIZfor_filenamerrr_get_name_and_versionsr~c@s
eZdZdZd?ddZddZdd	Zd
dZdd
ZddZ	ddZ
ddZddZddZ
ddZd@ddZddZdd Zd!d"Zd#d$ZdAd%d&ZdBd'd(ZdCd)d*Zd+d,Zefd-d.ZdDd/d0ZdEd1d2Zd3d4Zd5d6Zd7d8Zd9d:Zd;d<Z d=d>Z!dS)FLegacyMetadataaaThe legacy metadata of a release.

    Supports versions 1.0, 1.1 and 1.2 (auto-detected). You can
    instantiate the class with one of these arguments (or none):
    - *path*, the path to a metadata file
    - *fileobj* give a file-like object with metadata as content
    - *mapping* is a dict-like object
    - *scheme* is a version scheme name
    NdefaultcCsz|||gddkrtdi|_g|_d|_||_|dk	rH||n.|dk	r\||n|dk	rv|||	dS)N'path, fileobj and mapping are exclusive)
count	TypeError_fieldsZrequires_filesZ
_dependenciesschemeread	read_fileupdateset_metadata_version)selfpathfileobjmappingrrrr__init__s
zLegacyMetadata.__init__cCst|j|jd<dSNr )r^rrrrrr"sz#LegacyMetadata.set_metadata_versioncCs|d||fdS)Nz%s: %s
)write)rrr`r]rrr_write_field%szLegacyMetadata._write_fieldcCs
||SN)getrr`rrr__getitem__(szLegacyMetadata.__getitem__cCs|||Sr)set)rr`r]rrr__setitem__+szLegacyMetadata.__setitem__cCs8||}z|j|=Wntk
r2t|YnXdSr)
_convert_namerKeyError)rr`
field_namerrr__delitem__.s

zLegacyMetadata.__delitem__cCs||jkp|||jkSr)rrrrrr__contains__5s
zLegacyMetadata.__contains__cCs(|tkr|S|dd}t||S)Nrx_)_ALL_FIELDSr}lower_ATTR2FIELDrrrrrr9szLegacyMetadata._convert_namecCs|tks|tkrgSdS)NrO)_LISTFIELDS_ELEMENTSFIELDrrrr_default_value?szLegacyMetadata._default_valuecCs&|jdkrtd|Std|SdS)Nr?r
)r__LINE_PREFIX_PRE_1_2r|_LINE_PREFIX_1_2rr]rrr_remove_line_prefixDs
z"LegacyMetadata._remove_line_prefixcCs|tkr||St|dSr)rAttributeErrorrrrr__getattr__JszLegacyMetadata.__getattr__FcCst|d|d|S)zhReturn the distribution name with version.

        If filesafe is true, return a filename-escaped form.r!r")r~)rZfilesaferrrget_fullnameUszLegacyMetadata.get_fullnamecCs||}|tkS)z+return True if name is a valid metadata key)rrrrrris_field[s
zLegacyMetadata.is_fieldcCs||}|tkSr)rrrrrris_multi_field`s
zLegacyMetadata.is_multi_fieldcCs.tj|ddd}z||W5|XdS)z*Read the metadata values from a file path.rrencodingN)codecsopencloser)rfilepathfprrrrdszLegacyMetadata.readcCst|}|d|jd<tD]p}||kr(q|tkrf||}|tkrX|dk	rXdd|D}|||q||}|dk	r|dkr|||qdS)z,Read the metadata values from a file object.zmetadata-versionr NcSsg|]}t|dqS,)tuplesplit.0r]rrr
ysz,LegacyMetadata.read_file..rO)rrrrZget_all_LISTTUPLEFIELDSr)rZfileobmsgfieldvaluesr]rrrrls
zLegacyMetadata.read_filecCs0tj|ddd}z|||W5|XdS)z&Write the metadata fields to filepath.wrrN)rrr
write_file)rrskip_unknownrrrrrszLegacyMetadata.writecCs|t|dD]}||}|r8|dgdgfkr8q|tkrV|||d|q|tkr|dkr|jdkr~|dd}n|dd}|g}|t	krd	d
|D}|D]}||||qqdS)z0Write the PKG-INFO format data to a file object.r rOrr%rrrz	
       |cSsg|]}d|qSrjoinrrrrrsz-LegacyMetadata.write_file..N)
rrJrrrrrr_r}r)rZ
fileobjectrrrr]rrrrs$

zLegacyMetadata.write_filecsvfdd}|sn@t|dr:|D]}||||q$n|D]\}}|||q>|rr|D]\}}|||q^dS)aSet metadata values from the given iterable `other` and kwargs.

        Behavior is like `dict.update`: If `other` has a ``keys`` method,
        they are looped over and ``self[key]`` is assigned ``other[key]``.
        Else, ``other`` is an iterable of ``(key, value)`` iterables.

        Keys that don't match a metadata field or that have an empty value are
        dropped.
        cs"|tkr|r||dSr)rrr)r\r]rrr_setsz#LegacyMetadata.update.._setrKN)hasattrrKrP)rotherkwargsrkvrrrrs

zLegacyMetadata.updatecCsh||}|tks|dkrNt|ttfsNt|trHdd|dD}qzg}n,|tkrzt|ttfszt|trv|g}ng}t	t
jr<|d}t|j
}|tkr|dk	r|D](}||ddstd	|||qnb|tkr|dk	r||s.rr!N;rz$'%s': '%s' is not valid (field '%s')z.'%s': '%s' is not a valid version (field '%s')r%)rr
isinstancelistrrrrrSZisEnabledForloggingZWARNINGr
r_PREDICATE_FIELDSis_valid_matcherwarning_VERSIONS_FIELDSis_valid_constraint_list_VERSION_FIELDSis_valid_version_UNICODEFIELDSrr)rr`r]Zproject_namerrrrrrsV






zLegacyMetadata.setcCs||}||jkr*|tkr&||}|S|tkr@|j|}|S|tkr|j|}|dkr^gSg}|D].}|tkr~||qf||d|dfqf|S|tkr|j|}t	|t
r|dS|j|S)zGet a metadata field.Nrrr)rr_MISSINGrrrrrQrrrr)rr`rr]resvalrrrrs.







zLegacyMetadata.getcs|gg}}dD]}||kr||q|rP|gkrPdd|}t|dD]}||krT||qT|ddkr||fSt|jfdd}t|ftjft	j
ffD]@\}}|D]2}	||	d	}
|
d	k	r||
s|d
|	|
fqq||fS)zkCheck if the metadata is compliant. If strict is True then raise if
        no Name or Version are provided)r!r"zmissing required metadata: %s, )r'r(r r@cs(|D]}|ddsdSqdS)NrrFT)rr)r]rrrrare_valid_constraints#sz3LegacyMetadata.check..are_valid_constraintsNzWrong value for '%s': %s)rrQrrr
rrrrrrr)rstrictmissingwarningsattrrrr[Z
controllerrr]rrrcheck	s8

zLegacyMetadata.checkcCs|d}i}|D]"\}}|r*||jkr||||<q|ddkrd}|D]B\}}|rb||jkrL|dkrx||||<qLdd||D||<qLn8|ddkrd	}|D]"\}}|r||jkr||||<q|S)
zReturn fields as a dict.

        Field names will be converted to use the underscore-lowercase style
        instead of hyphen-mixed case (i.e. home_page instead of Home-page).
        )
)r_r )r`r!)rIr")rbr$)rer')rfr()rgr))rjr*)rcr%)rdr&)rar#)classifiersr,)rlr-r r@))ror6)rqr7)rrr8)rnr5)rmr3)rvr4)rhr1)rir2rvcSsg|]}d|qSrr)rurrrrbsz)LegacyMetadata.todict..r))rtr/)rsr0)rur.)rr)rZskip_missingZmapping_1_0datar\rZmapping_1_2Zmapping_1_1rrrtodict5s&
zLegacyMetadata.todictcCs8|ddkr$dD]}||kr||=q|d|7<dS)Nr r)r.r0r/r6r)rrequirementsrrrradd_requirementsps
zLegacyMetadata.add_requirementscCstt|dSr)rrJrrrrrK{szLegacyMetadata.keysccs|D]
}|VqdSrrK)rr\rrr__iter__~szLegacyMetadata.__iter__csfddDS)Ncsg|]}|qSrrrr\rrrrsz)LegacyMetadata.values..rrrrrrszLegacyMetadata.valuescsfddDS)Ncsg|]}||fqSrrrrrrrsz(LegacyMetadata.items..rrrrrrPszLegacyMetadata.itemscCsd|jj|j|jfS)Nz
<%s %s %s>)	__class__rr`rIrrrr__repr__szLegacyMetadata.__repr__)NNNr)F)F)F)N)F)F)"rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrKrrrPrrrrrrs@




,
,
;rzpydist.jsonz
metadata.jsonZMETADATAc@seZdZdZedZedejZe	Z
edZdZde
Zdddd	Zd
ZdZedfedfe
dfedfdZd
ZdDddZedZdefZdefZdefdefeeedefeeeedefddd
Z[[ddZdEddZddZed d!Z ed"d#Z!e!j"d$d#Z!dFd%d&Z#ed'd(Z$ed)d*Z%e%j"d+d*Z%d,d-Z&d.d/Z'd0d1Z(d2d3Z)d4d5d6d7d8dd9Z*d:d;Z+dGd>d?Z,d@dAZ-dBdCZ.dS)Hrz
    The metadata of a release. This implementation uses 2.0 (JSON)
    metadata where possible. If not possible, it wraps a LegacyMetadata
    instance which handles the key-value metadata format.
    z
^\d+(\.\d+)*$z!^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$z	.{1,2047}rCzdistlib (%s)r)legacy)r`rIrbzqname version license summary description author author_email keywords platform home_page classifiers download_urlzwextras run_requires test_requires build_requires dev_requires provides meta_requires obsoleted_by supports_environments)r_r`rIrb)_legacy_datarNrc	Cs0|||gddkrtdd|_d|_||_|dk	rzz|||||_Wn*tk
rvt||d|_|YnXnd}|rt	|d}|
}W5QRXn|r|
}|dkr|j|jd|_ndt
|ts|d}zt||_||j|Wn0tk
r*tt||d|_|YnXdS)Nrr)rrrbr_	generatorr)rr)rrrrr_validate_mappingrrvalidaterrMETADATA_VERSION	GENERATORrr	decodejsonloads
ValueErrorr)rrrrrrfrrrrs@



zMetadata.__init__)r`rIrjrdrbr6r;r=r,)r-N)r N)
run_requiresbuild_requiresdev_requiresZ
test_requires
meta_requiresextrasmodules
namespacesexportscommandsrZ
source_urlr_c
CsXt|d}t|d}||kr||\}}|jr^|dkrP|dkrHdn|}n|j|}n|dkrjdn|}|dkr|j||}nt}|}|jd}	|	r
|dkr|	d|}nP|dkr|	d}	|	r|	||}n,|	d	}	|	s|jd	}	|	r
|	||}||krT|}n:||kr2t||}n"|jrH|j|}n|j|}|S)
Ncommon_keysmapped_keysr	rrrr
extensionsr	python.commandsrpython.detailspython.exports)object__getattribute__rrr)
rr\commonmappedlkZmakerresultr]sentineldrrrrsD




zMetadata.__getattribute__cCsH||jkrD|j|\}}|p |j|krD||}|sDtd||fdS)Nz.'%s' is an invalid value for the '%s' property)SYNTAX_VALIDATORSrmatchr)rr\r]rpattern
exclusionsmrrr_validate_value+s

zMetadata._validate_valuecCs*|||t|d}t|d}||kr||\}}|jrV|dkrJt||j|<nf|dkrj||j|<nR|jdi}|dkr||d<n2|dkr|di}|||<n|d	i}|||<nh||krt|||nP|d
krt|t	r|
}|r|}ng}|jr||j|<n
||j|<dS)Nr
rrr
r	rrrrrd)rrrrNotImplementedErrorr
setdefault__setattr__rrrr)rr\r]rrrrrrrrr!5s<




zMetadata.__setattr__cCst|j|jdSNT)r~r`rIrrrrname_and_version\szMetadata.name_and_versioncCsF|jr|jd}n|jdg}d|j|jf}||krB|||S)Nr5rtz%s (%s))rrr r`rIrQ)rrsrrrrt`s
zMetadata.providescCs |jr||jd<n
||jd<dS)Nr5rt)rrrrrrrtksc
Cs|jr|}ng}t|pg|j}|D]d}d|kr>d|kr>d}n8d|krLd}n|d|k}|rv|d}|rvt||}|r$||dq$dD]F}d|}	|	|kr||	|jd|g}||j|||dq|S)	a
        Base method to get dependencies, given a set of extras
        to satisfy and an optional environment context.
        :param reqts: A list of sometimes-wanted dependencies,
                      perhaps dependent on extras and environment.
        :param extras: A list of optional components being requested.
        :param env: An optional environment for marker evaluation.
        extraenvironmentTrs)ZbuildZdevZtestz:%s:z%s_requires)renv)	rrrrr
extendrRrget_requirements)
rreqtsrr'rrZincluderMr\errrr)rs2	



zMetadata.get_requirementscCs|jr|S|jSr)r_from_legacyrrrrr
dictionaryszMetadata.dictionarycCs|jrtnt|j|jSdSr)rrrrDEPENDENCY_KEYSrrrrdependenciesszMetadata.dependenciescCs|jrtn|j|dSr)rrrrrrrrr/sc	Cs|d|jkrtg}|jD]"\}}||kr$||kr$||q$|rbdd|}t||D]\}}||||qjdS)Nr_zMissing metadata items: %sr)	rrrMANDATORY_KEYSrPrQrrr)	rrrrr\rrrrrrrrszMetadata._validate_mappingcCsB|jr.|jd\}}|s|r>td||n||j|jdS)NTz#Metadata: missing: %s, warnings: %s)rrrSrrrr)rrrrrrrszMetadata.validatecCs(|jr|jdSt|j|j}|SdSr")rrrr
INDEX_KEYS)rrrrrrszMetadata.todictc
Cs|j|jd}|jd}dD]*}||kr|dkr8d}n|}||||<q|dg}|dgkrdg}||d<d	}|D]*\}}||krt||rtd
||ig||<qt|j|d<i}i}	|S)NrT)r`rIrjrbrcrkrkrr&rd))ror)rprrsrt)rrrrrrt)
rrZlmdrnkkwrKokrfrhrrrr,s,

zMetadata._from_legacyr!r"r*r$r%)r`rIrjrbrcrcCsdd}t}|j}|jD]\}}||kr||||<q||j|j}||j|j}|jrpt	|j|d<t	||d<t	||d<|S)NcSst}|D]|}|d}|d}|d}|D]V}|sF|sF||q.d}|rVd|}|rp|rld||f}n|}|d||fq.q
|S)Nr%r&rsr2z
extra == "%s"z(%s) and %sr)rraddr)entriesr*r+r%r'ZrlistrrMrrrprocess_entriess"

z,Metadata._to_legacy..process_entriesr=r6r;)
rrLEGACY_MAPPINGrPrrrrrsorted)rr8rZnmdr3r5Zr1Zr2rrr
_to_legacyszMetadata._to_legacyFTc	Cs||gddkrtd||r`|jr4|j}n|}|rP|j||dq|j||dn^|jrp|}n|j}|rt	j
||ddddn.t|dd}t	j
||ddddW5QRXdS)	Nrz)Exactly one of path and fileobj is needed)rTr)Zensure_asciiindentZ	sort_keysrr)
rrrrr;rrr,rrdumprr)rrrrrZ	legacy_mdrrrrrrs*
zMetadata.writecCs|jr|j|nr|jdg}d}|D]}d|kr*d|kr*|}qHq*|dkrfd|i}|d|n t|dt|B}t||d<dS)Nrr&r%rsr)rrrr insertrr:)rrralwaysentryZrsetrrrr3szMetadata.add_requirementscCs*|jpd}|jpd}d|jj|j||fS)Nz	(no name)z
no versionz<%s %s %s (%s)>)r`rIrrr_)rr`rIrrrrDs

zMetadata.__repr__)NNNr)N)NN)NNFT)/rrrrrecompileZMETADATA_VERSION_MATCHERIZNAME_MATCHERrZVERSION_MATCHERZSUMMARY_MATCHERrrrr0r1r.r	__slots__rrr
rZ	none_listdictZ	none_dictrrrr!propertyr#rtsetterr)r-r/rrrr,r9r;rrrrrrrrs


-+

'




*


		%
)F)ErZ
__future__rrZemailrrrrAr2rrcompatrrr	rLr
utilrrrIr
rZ	getLoggerrrSrrrr__all__rrrBrrrDrErVrFrWrHrYrGrXrrrZEXTRA_RErJr^rrrrrrrrrrr{r~rZMETADATA_FILENAMEZWHEEL_METADATA_FILENAMEZLEGACY_METADATA_FILENAMErrrrrs










H!



site-packages/pip/_vendor/distlib/__pycache__/__init__.cpython-38.pyc000064400000002004151464452710021546 0ustar00U

.eK@snddlZdZGdddeZzddlmZWn&ek
rRGdddejZYnXeeZ	e	
edS)Nz0.2.9.post0c@seZdZdS)DistlibExceptionN)__name__
__module____qualname__rr@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/__init__.pyrsr)NullHandlerc@s$eZdZddZddZddZdS)rcCsdSNrselfrecordrrrhandlezNullHandler.handlecCsdSr	rr
rrremitrzNullHandler.emitcCs
d|_dSr	)lock)rrrr
createLockrzNullHandler.createLockN)rrrr
rrrrrrrsr)Zlogging__version__	ExceptionrrImportErrorZHandlerZ	getLoggerrZloggerZ
addHandlerrrrrs
site-packages/pip/_vendor/distlib/__pycache__/util.cpython-38.pyc000064400000136475151464452710021010 0ustar00U

.e@sddlZddlmZddlZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZzddlZWnek
rdZYnXddlZddlZddlZddlZddlZzddlZWnek
rddlZYnXddlZddlmZddlmZmZmZmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0e
1e2Z3e
4dZ5e
4dZ6e
4d	Z7e
4d
Z8e
4dZ9e
4dZ:e
4d
Z;e
4dZddZ?ddZ@ddZAdddZBddZCddZDdd ZEejFd!d"ZGejFd#d$ZHejFdd&d'ZIGd(d)d)eJZKd*d+ZLGd,d-d-eJZMd.d/ZNGd0d1d1eJZOe
4d2e
jPZQd3d4ZRdd5d6ZSd7d8ZTd9d:ZUd;d<ZVd=d>ZWd?d@ZXe
4dAe
jYZZe
4dBZ[ddCdDZ\e
4dEZ]dFdGZ^dHdIZ_dJdKZ`dLZadMdNZbdOdPZcGdQdRdReJZdGdSdTdTeJZeGdUdVdVeJZfdWZgddYdZZhd[d\Zid]ZjGd^d_d_eJZke
4d`Zle
4daZme
4dbZndcddZdedfZoerddglmpZqmrZrmsZsGdhdidie$jtZtGdjdkdkeqZpGdldmdmepe'ZuejvddnZwewdokrGdpdqdqe$jxZxerGdrdsdse$jyZyGdtdudue%jzZzerBGdvdwdwe%j{Z{Gdxdydye%j|Z|dzd{Z}Gd|d}d}eJZ~Gd~dde~ZGddde~ZGddde(ZGdddeJZddZdS)N)deque)iglob)DistlibException)string_types	text_typeshutil	raw_inputStringIOcache_from_sourceurlopenurljoinhttplib	xmlrpclib	splittypeHTTPHandlerBaseConfiguratorvalid_ident	ContainerconfigparserURLErrorZipFilefsdecodeunquoteurlparsez^([\w\.-]+)\s*z^([\w\.*+-]+)\s*z^(<=?|>=?|={2,3}|[~!]=)\s*z*^((<=?)|(>=?)|={2,3}|[~!]=|in|not\s+in)\s*z^or\b\s*z	^and\b\s*z(\S+)\s*z(([\s\w\.{}()*+#:;,/?!~`@$%^&=|<>\[\]-]+)cs6ddfddfddfdd|S)	ae
    Parse a marker string and return a dictionary containing a marker expression.

    The dictionary will contain keys "op", "lhs" and "rhs" for non-terminals in
    the expression grammar, or strings. A string contained in quotes is to be
    interpreted as a literal string, and a string not contained in quotes is a
    variable (such as os_name).
    cSs.t|}|r,|d}||d}n|s:tdn|d}|dkrVtd|d|d}|dd}|g}|r|d|krqqt|d|kr|||dd}qtt|}|std|||d||d}qtd|}td|||d|}|dd	}||fS)	Nrzunexpected end of inputz'"zinvalid expression: %srzerror in string literal: %szunterminated string: %s)

IDENTIFIERmatchgroupsendSyntaxErrorreplaceappendSTRING_CHUNKjoinlstrip)	remainingmresultqZoqpartssr,.marker_varcs|rR|ddkrR|dd\}}|ddkr@td||dd}nZ|\}}|rt|}|srq|d}||d}|\}}|||d}q^|}||fS)Nr(r)unterminated parenthesis: %soplhsrhs)r%r 	MARKER_OPrrr)r&r(r4r'r3r5)markerr.r,r-marker_expres 
z!parse_marker..marker_exprcsR|\}}|rJt|}|s qJ||d}|\}}d||d}q||fS)Nandr2)ANDrrr&r4r'r5)r8r,r-
marker_andxs
z parse_marker..marker_andcsR|\}}|rJt|}|s qJ||d}|\}}d||d}q||fS)Norr2)ORrrr;)r<r,r-r7s
zparse_marker..markerr,)Z
marker_stringr,)r7r<r8r.r-parse_marker8s
	$r?cCs0|}|r|drdSt|}|s4td||d}||d}d}}}}|r:|ddkr:|dd}|dkrtd||d|}	||dd}g}|	r0t|	}|std	|	|	|d|	|d}	|	sq0|	dd
krtd|	|	dd}	q|s:d}|r|ddkr|dd}t
|}|sztd
||d}t|}
|
jr|
j
std|||d}ndd}|ddkr||\}}n|dd}|dkrtd||d|}	||dd}t|	r@||	\}}nXt|	}|s\td|	|d}
|	|d}	|	rtd|	d|
fg}|r|ddkrtd||dd}t|\}}|r|ddkrtd||s|}nd|ddd|Df}t||||||dS)z
    Parse a requirement passed in as a string. Return a Container
    whose attributes contain the various parts of the requirement.
    #Nzname expected: %sr[]rzunterminated extra: %szmalformed extra: %s,zcomma expected in extras: %s@zinvalid URI: %szInvalid URL: %scSst|}d}|rg}|d}||d}t|}|sLtd||d}|||f||d}|r|ddkrq|dd}t|}|std|q|sd}||fS)z|
                Return a list of operator, version tuples if any are
                specified, else None.
                Nrzinvalid version: %srCrinvalid constraint: %s)
COMPARE_OPrrrVERSION_IDENTIFIERr r"r%)Z
ver_remainingr'versionsr3vr,r,r-get_versionss*


z'parse_requirement..get_versionsr/r0r1rEz~=;zinvalid requirement: %szunexpected trailing data: %s%s %s, cSsg|]}d|qS)rLr,).0Zconr,r,r-
sz%parse_requirement..)nameextrasZconstraintsr7urlZrequirement)strip
startswithrrr rrfindr%r"	NON_SPACErschemenetlocrFrGr?r$r)reqr&r'ZdistnamerQZ	mark_exprrHuriir+trJ_rIZrsr,r,r-parse_requirements





r^cCsdd}i}|D]\}}}tj||}t|D]p}tj||}	t|	D]T}
|||
}|dkrn||dqJ|||
}|tjjdd}
|
d|||<qJq0q|S)z%Find destinations for resources filescSsD|tjjd}|tjjd}||s.t|t|ddSN/)r!ospathseprTAssertionErrorlenr%)rootrbr,r,r-get_rel_path
sz)get_resources_dests..get_rel_pathNr`)rarbr$rpopr!rcrstrip)Zresources_rootZrulesrgZdestinationsbasesuffixdestprefixZabs_baseZabs_globZabs_pathZ
resource_fileZrel_pathZrel_destr,r,r-get_resources_dests
s

rncCs(ttdrd}ntjttdtjk}|S)NZreal_prefixTbase_prefix)hasattrsysrmgetattrr(r,r,r-in_venv$s
rtcCs$tjtj}t|ts t|}|SN)rarbnormcaserq
executable
isinstancerrrsr,r,r-get_executable.s

rycCsN|}t|}|}|s|r|}|r|d}||kr6qJ|rd|||f}q|S)Nrz	%c: %s
%s)r	lower)promptZ
allowed_charsZerror_promptdefaultpr+cr,r,r-proceed>srcCs8t|tr|}i}|D]}||kr||||<q|Sru)rxrsplit)dkeysr(keyr,r,r-extract_by_keyNs
rcCsztjddkrtd|}|}t|}zlt|}|ddd}|D]B\}}|D]0\}}d||f}t	|}	|	dk	st
|	||<q`qP|WStk
r|ddYnXdd	}
t
}z|
||Wn<t
jk
r|t|}t|}|
||YnXi}|D]R}i||<}||D]4\}
}d|
|f}t	|}	|	dk	sft
|	||
<q<q"|S)
Nrutf-8
extensionszpython.exportsexportsz%s = %scSs$t|dr||n
||dS)N	read_file)rprZreadfp)cpstreamr,r,r-read_streamks
z!read_exports..read_stream)rqversion_infocodecs	getreaderreadr
jsonloaditemsget_export_entryrd	ExceptionseekrConfigParserZMissingSectionHeaderErrorclosetextwrapdedentZsections)rdataZjdatar(groupentrieskrIr+entryrrrrPvaluer,r,r-read_exportsWsD

rcCstjddkrtd|}t}|D]l\}}|||D]P}|j	dkr\|j
}nd|j
|j	f}|jrd|d|jf}|
||j|qFq,||dS)Nrrrz%s:%sz%s [%s]rM)rqrr	getwriterrrrZadd_sectionvaluesrkrmflagsr$setrPwrite)rrrrrIrr+r,r,r-
write_exportss

rc	cs$t}z
|VW5t|XdSru)tempfilemkdtemprrmtree)Ztdr,r,r-tempdirs
rc	cs.t}zt|dVW5t|XdSru)ragetcwdchdir)rcwdr,r,r-rs


rc	cs.t}zt|dVW5t|XdSru)socketZgetdefaulttimeoutZsetdefaulttimeout)ZsecondsZctor,r,r-socket_timeouts


rc@seZdZddZdddZdS)cached_propertycCs
||_dSru)func)selfrr,r,r-__init__szcached_property.__init__NcCs,|dkr|S||}t||jj||Sru)robject__setattr____name__)robjclsrr,r,r-__get__s

zcached_property.__get__)N)r
__module____qualname__rrr,r,r,r-rsrcCs~tjdkr|S|s|S|ddkr.td||ddkrFtd||d}tj|krh|tjqP|srtjStjj|S)aReturn 'pathname' as a name that will work on the native filesystem.

    The path is split on '/' and put back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    r`rzpath '%s' cannot be absolutezpath '%s' cannot end with '/')rarc
ValueErrorrcurdirremoverbr$)pathnamepathsr,r,r-convert_paths



rc@seZdZd$ddZddZddZdd	Zd%ddZd&ddZddZ	ddZ
ddZddZddZ
d'ddZddZddZd d!Zd"d#Zd
S)(FileOperatorFcCs||_t|_|dSru)dry_runrensured_init_record)rrr,r,r-rszFileOperator.__init__cCsd|_t|_t|_dSNF)recordr
files_writtendirs_createdrr,r,r-rszFileOperator._init_recordcCs|jr|j|dSru)rradd)rrbr,r,r-record_as_writtenszFileOperator.record_as_writtencCsHtj|s tdtj|tj|s0dSt|jt|jkS)aTell if the target is newer than the source.

        Returns true if 'source' exists and is more recently modified than
        'target', or if 'source' exists and 'target' doesn't.

        Returns false if both exist and 'target' is the same age or younger
        than 'source'. Raise PackagingFileError if 'source' does not exist.

        Note that this test is not very accurate: files created in the same
        second will have the same "age".
        zfile '%r' does not existT)rarbexistsrabspathstatst_mtime)rsourcetargetr,r,r-newers
zFileOperator.newerTcCs|tj|td|||jsd}|rdtj|rDd|}n tj|rdtj	|sdd|}|rtt
|dt|||
|dS)z8Copy a file respecting dry-run and force flags.
        zCopying %s to %sNz%s is a symlinkz%s is a non-regular filez which would be overwritten)
ensure_dirrarbdirnameloggerinforislinkrisfilerrZcopyfiler)rZinfileoutfilecheckmsgr,r,r-	copy_files
zFileOperator.copy_fileNcCstj|rt|tj|td|||jsv|dkrJt	|d}nt
j	|d|d}zt
||W5|X||dS)NzCopying stream %s to %swbwencoding)rarbisdirrdrrrrropenrrrZcopyfileobjr)rZinstreamrrZ	outstreamr,r,r-copy_streams
zFileOperator.copy_streamc	Cs\|tj||jsNtj|r.t|t|d}||W5QRX|	|dS)Nr)
rrarbrrrrrrr)rrbrfr,r,r-write_binary_file!s
zFileOperator.write_binary_filecCs||||dSru)rencode)rrbrrr,r,r-write_text_file*szFileOperator.write_text_filecCsntjdkstjdkrjtjdkrj|D]F}|jr:td|q"t|j|B|@}td||t||q"dS)Nposixjavazchanging mode of %szchanging mode of %s to %o)	rarP_namerrrrst_modechmod)rbitsmaskfilesrmoder,r,r-set_mode-szFileOperator.set_modecCs|dd|S)Nimi)r)r+rr,r,r-9zFileOperator.cCs|tj|}||jkrxtj|sx|j|tj|\}}||t	d||j
sft||jrx|j
|dS)NzCreating %s)rarbrrrrrrrrrmkdirrr)rrbrrr,r,r-r;s

zFileOperator.ensure_dirc	Cst||}td|||js|s0|||rX|s:d}n||sHt|t|d}i}|rvtt	drvt	j
j|d<t	j|||df||
||S)NzByte-compiling %s to %sPycInvalidationModeinvalidation_modeT)rrrrrrTrdrerp
py_compilerCHECKED_HASHcompiler)	rrboptimizeforcermZhashed_invalidationZdpathZdiagpathZcompile_kwargsr,r,r-byte_compileGs
zFileOperator.byte_compilecCstj|rtj|r^tj|s^td||js@t	||j
r||jkr|j|nPtj|rpd}nd}td|||jst||j
r||j
kr|j
|dS)NzRemoving directory tree at %slinkfilezRemoving %s %s)rarbrrrrdebugrrrrrrr)rrbr+r,r,r-ensure_removedXs"



zFileOperator.ensure_removedcCsDd}|s@tj|r$t|tj}q@tj|}||kr:q@|}q|Sr)rarbraccessW_OKr)rrbr(parentr,r,r-is_writablemszFileOperator.is_writablecCs"|js
t|j|jf}||S)zV
        Commit recorded changes, turn off recording, return
        changes.
        )rrdrrr)rr(r,r,r-commitys
zFileOperator.commitcCs|jst|jD]}tj|rt|qt|jdd}|D]F}t	|}|rz|dgks^t
tj||d}t|t|q>|
dS)NT)reverse__pycache__r)rlistrrarbrrsortedrlistdirrdr$rmdirr)rrdirsrflistZsdr,r,r-rollbacks

zFileOperator.rollback)F)T)N)FFNF)rrrrrrrrrrrrZset_executable_moderrrrr	rr,r,r,r-rs 


	

rcCs^|tjkrtj|}nt|}|dkr,|}n.|d}t||d}|D]}t||}qJ|S)N.r)rqmodules
__import__rrrrh)Zmodule_nameZdotted_pathmodr(r*r}r,r,r-resolves

rc@s6eZdZddZeddZddZddZej	Z	d	S)
ExportEntrycCs||_||_||_||_dSrurPrmrkr)rrPrmrkrr,r,r-rszExportEntry.__init__cCst|j|jSru)rrmrkrr,r,r-rszExportEntry.valuecCsd|j|j|j|jfS)Nzrrr,r,r-__repr__s
zExportEntry.__repr__cCsDt|tsd}n0|j|jko>|j|jko>|j|jko>|j|jk}|Sr)rxrrPrmrkr)rotherr(r,r,r-__eq__s



zExportEntry.__eq__N)
rrrrrrrrr__hash__r,r,r,r-rs

rz(?P(\w|[-.+])+)
                      \s*=\s*(?P(\w+)([:\.]\w+)*)
                      \s*(\[\s*(?P\w+(=\w+)?(,\s*\w+(=\w+)?)*)\s*\])?
                      c
Cst|}|s0d}d|ks"d|krtd|n|}|d}|d}|d}|dkrf|d}}n"|dkrztd||d\}}|d	}	|	dkrd|ksd|krtd|g}	nd
d|	dD}	t||||	}|S)
NrArBzInvalid specification '%s'rPcallable:rrrcSsg|]}|qSr,)rS)rNrr,r,r-rOsz$get_export_entry..rC)ENTRY_REsearchr	groupdictcountrr)
Z
specificationr'r(rrPrbZcolonsrmrkrr,r,r-rs8

rcCs|dkrd}tjdkr.dtjkr.tjd}ntjd}tj|rft|tj}|st	
d|n|dd\}}d|kr.|}n|dd\}}|rJt|}|rVt|}|||fS)NrDrr)rsplitrr)rXZusernameZpasswordrmr,r,r-parse_credentials$sr3cCstd}t||S)N)raumaskrsr,r,r-get_process_umask3s

r6cCs<d}d}t|D]\}}t|tsd}q,q|dk	s8t|S)NTF)	enumeraterxrrd)seqr(r[r+r,r,r-is_string_sequence8s
r9z3([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-([a-z0-9_.+-]+)z
-py(\d\.?\d?)cCsd}d}t|dd}t|}|r@|d}|d|}|rt|t|dkrtt	|d|}|r|
}|d|||dd|f}|dkrt|}|r|d|d|f}|S)zw
    Extract name, version, python version from a filename (no extension)

    Return name, version, pyver or None
    N -rz\br)rr!PYTHON_VERSIONr!rstartrererescaperPROJECT_NAME_AND_VERSION)filenameZproject_namer(Zpyverr'nr,r,r-split_filenameGs"


rCz-(?P[\w .-]+)\s*\(\s*(?P[^\s)]+)\)$cCs:t|}|std||}|d|dfS)z
    A utility method used to get name and version from a string.

    From e.g. a Provides-Dist value.

    :param p: A value in a form 'foo (1.0)'
    :return: The name and version as a tuple.
    z$Ill-formed name/version string: '%s'rPZver)NAME_VERSION_RErrr"rSrz)r}r'rr,r,r-parse_name_and_versioncs
	
rEcCst}t|pg}t|pg}d|kr8|d||O}|D]x}|dkrT||q<|dr|dd}||krtd|||kr||q<||krtd|||q<|S)N*r;rzundeclared extra: %s)rrrrTrr*)Z	requestedZ	availabler(rZunwantedr,r,r-
get_extrasrs&

rHc
Csi}zNt|}|}|d}|ds8td|ntd|}t	|}Wn0t
k
r}ztd||W5d}~XYnX|S)NzContent-Typezapplication/jsonz(Unexpected response for JSON request: %srz&Failed to get external data for %s: %s)rrgetrTrrrrrrr	exception)rRr(ZrespZheadersZctreaderer,r,r-_get_external_datas

 rMz'https://www.red-dove.com/pypi/projects/cCs*d|d|f}tt|}t|}|S)Nz%s/%s/project.jsonrupperr
_external_data_base_urlrM)rPrRr(r,r,r-get_project_datas
rQcCs(d|d||f}tt|}t|S)Nz%s/%s/package-%s.jsonrrN)rPversionrRr,r,r-get_package_datas
rSc@s(eZdZdZddZddZddZdS)	Cachez
    A class implementing a cache for resources that need to live in the file system
    e.g. shared libraries. This class was moved from resources to here because it
    could be used by other modules, e.g. the wheel module.
    cCsPtj|st|t|jd@dkr6td|tjtj	||_
dS)zu
        Initialise an instance.

        :param base: The base directory where the cache should be located.
        ?rzDirectory '%s' is not privateN)rarbrr+rrrr*rnormpathrj)rrjr,r,r-rs

zCache.__init__cCst|S)zN
        Converts a resource prefix to a directory name in the cache.
        )r/)rrmr,r,r-
prefix_to_dirszCache.prefix_to_dirc	Csg}t|jD]r}tj|j|}z>tj|s>tj|rJt|ntj|r`t	
|Wqtk
r||YqXq|S)z"
        Clear the cache.
        )
rarrjrbr$rrrrrrrr")rZnot_removedfnr,r,r-clearszCache.clearN)rrr__doc__rrWrYr,r,r,r-rTsrTc@s:eZdZdZddZdddZddZd	d
ZddZd
S)
EventMixinz1
    A very simple publish/subscribe system.
    cCs
i|_dSru)_subscribersrr,r,r-rszEventMixin.__init__TcCsD|j}||krt|g||<n"||}|r6||n
||dS)a`
        Add a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be added (and called when the
                           event is published).
        :param append: Whether to append or prepend the subscriber to an
                       existing subscriber list for the event.
        N)r\rr"
appendleft)revent
subscriberr"subsZsqr,r,r-rs
zEventMixin.addcCs,|j}||krtd||||dS)z
        Remove a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be removed.
        zNo subscribers: %rN)r\rr)rr^r_r`r,r,r-rszEventMixin.removecCst|j|dS)z
        Return an iterator for the subscribers for an event.
        :param event: The event to return subscribers for.
        r,)iterr\rI)rr^r,r,r-get_subscribersszEventMixin.get_subscribersc	Oslg}||D]F}z||f||}Wn"tk
rHtdd}YnX||qtd|||||S)a^
        Publish a event and return a list of values returned by its
        subscribers.

        :param event: The event to publish.
        :param args: The positional arguments to pass to the event's
                     subscribers.
        :param kwargs: The keyword arguments to pass to the event's
                       subscribers.
        z"Exception during event publicationNz/publish %s: args = %s, kwargs = %s, result = %s)rbrrrJr"r)rr^argskwargsr(r_rr,r,r-publishs

zEventMixin.publishN)T)	rrrrZrrrrbrer,r,r,r-r[s
r[c@s^eZdZddZddZdddZdd	Zd
dZdd
ZddZ	e
ddZe
ddZdS)	SequencercCsi|_i|_t|_dSru)_preds_succsr_nodesrr,r,r-r"szSequencer.__init__cCs|j|dSru)rirrnoder,r,r-add_node'szSequencer.add_nodeFcCs||jkr|j||rt|j|dD]}|||q,t|j|dD]}|||qPt|jD]\}}|sp|j|=qpt|jD]\}}|s|j|=qdS)Nr,)rirrrgrIrhrr)rrkZedgesr}r+rrIr,r,r-remove_node*s

zSequencer.remove_nodecCs<||kst|j|t||j|t|dSru)rdrg
setdefaultrrrh)rpredsuccr,r,r-r:sz
Sequencer.addcCs||kstz|j|}|j|}Wn tk
rDtd|YnXz||||Wn$tk
rtd||fYnXdS)Nz%r not a successor of anythingz%r not a successor of %r)rdrgrhKeyErrorrr)rrorppredsZsuccsr,r,r-r?s

zSequencer.removecCs||jkp||jkp||jkSru)rgrhri)rstepr,r,r-is_stepLszSequencer.is_stepcCs||std|g}g}t}|||r|d}||krb||kr||||q.|||||j|d}|	|q.t
|S)NzUnknown: %rrr,)rtrrr"rhrrrgrIextendreversed)rfinalr(Ztodoseenrsrrr,r,r-	get_stepsPs"





zSequencer.get_stepscsRdggiig|jfddD]}|kr8|q8S)Nrcsd|<d|<dd7<|z|}Wntk
rVg}YnX|D]J}|kr|t|||<q\|kr\t|||<q\||krg}}||||krqqt|}|dS)Nrr)r"rminrhtuple)rkZ
successorsZ	successorZconnected_componentZ	componentZgraphindexZ
index_counterZlowlinksr(stack
strongconnectr,r-rts,


z3Sequencer.strong_connections..strongconnect)rhrjr,r|r-strong_connectionsis"
zSequencer.strong_connectionscCsfdg}|jD]*}|j|}|D]}|d||fqq|jD]}|d|q>|dd|S)Nzdigraph G {z  %s -> %s;z  %s;}
)rgr"rir$)rr(rprrrorkr,r,r-dots



z
Sequencer.dotN)F)
rrrrrlrmrrrtrypropertyrrr,r,r,r-rf!s


2rf).tar.gz.tar.bz2.tar.zip.tgz.tbz.whlTcsZfdd}tjtd}|dkr|dr>d}nH|drRd}d}n4|drfd	}d
}n |drzd}d
}ntd|z|dkrt|d
}|r|}|D]}||qn*t	
||}|r|}|D]}||q|dkr*tj
ddkr*|D]"}	t|	jts|	jd|	_qdd}
|
|_|W5|rT|XdS)NcsRt|ts|d}tjtj|}|rB|tjkrNt	d|dS)Nrzpath outside destination: %r)
rxrdecoderarbrr$rTrcr)rbr}dest_dirZplenr,r-
check_paths


zunarchive..check_path)rrzip)rrZtgzzr:gz)rrZtbzzr:bz2rZtarrGzUnknown format for %rrrrc
SsDzt||WStjk
r>}ztt|W5d}~XYnXdS)z:Run tarfile.tar_fillter, but raise the expected ValueErrorN)tarfileZ
tar_filterZFilterErrorrstr)memberrbexcr,r,r-extraction_filtersz$unarchive..extraction_filter)rarbrrer0rrrZnamelistrrZgetnamesrqrZ
getmembersrxrPrrrZ
extractall)Zarchive_filenamerformatrrarchivernamesrPZtarinforr,rr-	unarchivesL





rc	Cs~t}t|}t|dZ}t|D]H\}}}|D]8}tj||}||d}	tj|	|}
|||
q4q&W5QRX|S)z*zip a directory tree into a BytesIO objectrN)	ioBytesIOrerrawalkrbr$r)Z	directoryr(ZdlenZzfrfrrrPZfullZrelrlr,r,r-zip_dirsr)rKMGTPc@sreZdZdZdddZddZdd	Zd
dZdd
Ze	ddZ
e	ddZddZe	ddZ
e	ddZdS)ProgressZUNKNOWNrdcCs<|dks||kst||_|_||_d|_d|_d|_dS)NrF)rdrzcurmaxstartedelapseddone)rZminvalZmaxvalr,r,r-r	szProgress.__init__cCsV|j|kst|jdks&||jks&t||_t}|jdkrF||_n||j|_dSru)rzrdrrtimerr)rZcurvalZnowr,r,r-updates
zProgress.updatecCs |dkst||j|dSNr)rdrr)rZincrr,r,r-	incrementszProgress.incrementcCs||j|Sru)rrzrr,r,r-r=szProgress.startcCs |jdk	r||jd|_dS)NT)rrrrr,r,r-stop#s
z
Progress.stopcCs|jdkr|jS|jSru)runknownrr,r,r-maximum(szProgress.maximumcCsD|jrd}n4|jdkrd}n$d|j|j|j|j}d|}|S)Nz100 %z ?? %gY@z%3d %%)rrrrz)rr(rIr,r,r-
percentage,s
zProgress.percentagecCs:|dkr|jdks|j|jkr$d}ntdt|}|S)Nrz??:??:??z%H:%M:%S)rrrzrstrftimegmtime)rZdurationr(r,r,r-format_duration7szProgress.format_durationcCs|jrd}|j}n^d}|jdkr&d}nJ|jdks<|j|jkrBd}n.t|j|j}||j|j}|d|j}d|||fS)NZDonezETA rrrz%s: %s)rrrrrzfloatr)rrmr\r,r,r-ETA@s
zProgress.ETAcCsL|jdkrd}n|j|j|j}tD]}|dkr6q@|d}q&d||fS)Nrgig@@z%d %sB/s)rrrzUNITS)rr(Zunitr,r,r-speedSs

zProgress.speedN)rr)rrrrrrrr=rrrrrrrr,r,r,r-rs




	
rz\{([^}]*)\}z[^/\\,{]\*\*|\*\*[^/\\,}]z^[^{]*\}|\{[^}]*$cCs<t|rd}t||t|r4d}t||t|S)zAExtended globbing function that supports ** and {opt1,opt2,opt3}.z7invalid glob %r: recursive glob "**" must be used alonez2invalid glob %r: mismatching set marker '{' or '}')_CHECK_RECURSIVE_GLOBr!r_CHECK_MISMATCH_SET_iglob)	path_globrr,r,r-rhs

rccst|d}t|dkrht|dks,t||\}}}|dD]$}td|||fD]
}|VqXq@nd|krt|D]
}|Vqxn~|dd\}}|dkrd}|dkrd}n|d}|d	}t	|D]4\}}}	tj
|}ttj
||D]
}
|
VqqdS)
NrrrCrz**rrFr`\)	RICH_GLOBrrerdrr$	std_iglobr%rarrbrV)rZrich_path_globrmrrkitemrbZradicaldirrrXr,r,r-rss*



r)HTTPSHandlermatch_hostnameCertificateErrorc@seZdZdZdZddZdS)HTTPSConnectionNTcCsRt|j|jf|j}t|ddr0||_|tt	dsp|j
rHt	j}nt	j}t	j
||j|j|t	j|j
d|_nxt	t	j}|jt	jO_|jr||j|ji}|j
rt	j|_|j|j
dtt	ddr|j|d<|j
|f||_|j
rN|jrNz$t|j|jtd|jWn0tk
rL|jtj|jYnXdS)	NZ_tunnel_hostF
SSLContext)	cert_reqsZssl_versionca_certs)ZcafileZHAS_SNIZserver_hostnamezHost verified: %s) rZcreate_connectionhostporttimeoutrrsockZ_tunnelrpsslrZ
CERT_REQUIREDZ	CERT_NONEZwrap_socketZkey_fileZ	cert_fileZPROTOCOL_SSLv23rZoptionsZOP_NO_SSLv2Zload_cert_chainZverify_modeZload_verify_locationscheck_domainrZgetpeercertrrrZshutdownZ	SHUT_RDWRr)rrrcontextrdr,r,r-connects@



zHTTPSConnection.connect)rrrrrrr,r,r,r-rsrc@s&eZdZd	ddZddZddZdS)
rTcCst|||_||_dSru)BaseHTTPSHandlerrrr)rrrr,r,r-rs
zHTTPSHandler.__init__cOs$t||}|jr |j|_|j|_|S)a
            This is called to create a connection instance. Normally you'd
            pass a connection class to do_open, but it doesn't actually check for
            a class, and just expects a callable. As long as we behave just as a
            constructor would have, we should be OK. If it ever changes so that
            we *must* pass a class, we'll create an UnsafeHTTPSConnection class
            which just sets check_domain to False in the class definition, and
            choose which one to pass to do_open.
            )rrr)rrcrdr(r,r,r-_conn_makers


zHTTPSHandler._conn_makerc
CsXz||j|WStk
rR}z$dt|jkr@td|jnW5d}~XYnXdS)Nzcertificate verify failedz*Unable to verify server certificate for %s)Zdo_openrrrreasonrr)rrYrLr,r,r-
https_openszHTTPSHandler.https_openN)T)rrrrrrr,r,r,r-rs
rc@seZdZddZdS)HTTPSOnlyHandlercCstd|dS)NzAUnexpected HTTP request on what should be a secure connection: %s)r)rrYr,r,r-	http_openszHTTPSOnlyHandler.http_openN)rrrrr,r,r,r-rsrrc@seZdZdddZdS)HTTPrNcKs&|dkrd}||j||f|dSr_setupZ_connection_classrrrrdr,r,r-rsz
HTTP.__init__)rNrrrrr,r,r,r-rsrc@seZdZdddZdS)HTTPSrNcKs&|dkrd}||j||f|dSrrrr,r,r-rszHTTPS.__init__)rNrr,r,r,r-rsrc@seZdZdddZddZdS)	TransportrcCs||_tj||dSru)rrrrrruse_datetimer,r,r-rszTransport.__init__cCs`||\}}}tdkr(t||jd}n4|jr<||jdkrR||_|t|f|_|jd}|S)Nr)rrr)
get_host_info	_ver_inforr_connection_extra_headersrZHTTPConnection)rrhehZx509r(r,r,r-make_connection
s
zTransport.make_connectionN)rrrrrrr,r,r,r-rs
rc@seZdZdddZddZdS)
SafeTransportrcCs||_tj||dSru)rrrrrr,r,r-rszSafeTransport.__init__cCsx||\}}}|si}|j|d<tdkr:t|df|}n:|jrN||jdkrj||_|tj|df|f|_|jd}|S)Nrrrr)rrrrrrrr)rrrrrdr(r,r,r-rs

zSafeTransport.make_connectionN)rrr,r,r,r-rs
rc@seZdZddZdS)ServerProxyc	Kst|dd|_}|dk	r^t|\}}|dd}|dkr@t}nt}|||d|d<}||_tjj	||f|dS)NrrrZhttps)r	transport)
rhrrrIrrrrrr)	rrZrdrrWr]rZtclsr\r,r,r-r,szServerProxy.__init__Nrr,r,r,r-r+srcKs6tjddkr|d7}nd|d<d|d<t||f|S)Nrrbrnewlinerr)rqrr)rXrrdr,r,r-	_csv_open@s

rc@s4eZdZedededdZddZddZd	S)
CSVBaserC"r)Z	delimiterZ	quotecharZlineterminatorcCs|Srur,rr,r,r-	__enter__RszCSVBase.__enter__cGs|jdSru)rr)rr&r,r,r-__exit__UszCSVBase.__exit__N)rrrrdefaultsrrr,r,r,r-rKsrc@s(eZdZddZddZddZeZdS)	CSVReadercKs\d|kr4|d}tjddkr,td|}||_nt|dd|_tj|jf|j|_dS)NrrrrrbrG)	rqrrrrrcsvrKr)rrdrr,r,r-rZszCSVReader.__init__cCs|Srur,rr,r,r-__iter__eszCSVReader.__iter__cCsFt|j}tjddkrBt|D] \}}t|ts |d||<q |SNrrr)nextrKrqrr7rxrr)rr(r[rr,r,r-rhs

zCSVReader.nextN)rrrrrr__next__r,r,r,r-rYsrc@seZdZddZddZdS)	CSVWritercKs$t|d|_tj|jf|j|_dS)Nr)rrrwriterr)rrXrdr,r,r-rsszCSVWriter.__init__cCsNtjddkr>g}|D]"}t|tr.|d}||q|}|j|dSr)rqrrxrrr"rwriterow)rrowrGrr,r,r-rws

zCSVWriter.writerowN)rrrrrr,r,r,r-rrsrcsHeZdZeejZded<dfdd	ZddZdd	Zd
dZ	Z
S)
Configuratorinc_convertZincNcs"tt|||pt|_dSru)superrrrarrj)rconfigrj	__class__r,r-rszConfigurator.__init__c
sfddd}t|s*|}dd}dd}|r\tfdd|D}fd	dD}t|}|||}|r|D]\}}	t|||	q|S)
Ncsvt|ttfr*t|fdd|D}nHt|trhd|krH|}qri}|D]}||||<qPn
|}|S)Ncsg|]}|qSr,r,)rNr[convertr,r-rOszBConfigurator.configure_custom..convert..())rxrr{typedictconfigure_customr)or(r)rrr,r-rs

z.Configurator.configure_custom..convertrrz[]r,csg|]}|qSr,r,)rNrrr,r-rOsz1Configurator.configure_custom..cs$g|]}t|r||fqSr,)r)rNr)rrr,r-rOs)rhrrr{r
rsetattr)
rrr~Zpropsrcrrdr(rBrIr,)rrrr-rs


zConfigurator.configure_customcCs4|j|}t|tr0d|kr0|||j|<}|S)Nr)rrxr
r)rrr(r,r,r-__getitem__s
zConfigurator.__getitem__c	CsFtj|stj|j|}tj|ddd}t|}W5QRX|S)z*Default converter for the inc:// protocol.rGrr)	rarbisabsr$rjrrrr)rrrr(r,r,r-rs
zConfigurator.inc_convert)N)rrrr
rZvalue_convertersrrrr
__classcell__r,r,rr-rs
rc@s*eZdZdZd
ddZddZdd	ZdS)SubprocessMixinzC
    Mixin for running subprocesses and capturing their output
    FNcCs||_||_dSru)verboseprogress)rrrr,r,r-rszSubprocessMixin.__init__cCsj|j}|j}|}|sq^|dk	r.|||q|s@tjdntj|dtjq|dS)z
        Read lines from a subprocess' output stream and either pass to a progress
        callable (if specified) or write progress information to sys.stderr.
        Nrr)	rrreadlinerqstderrrrflushr)rrrrrr+r,r,r-rKszSubprocessMixin.readercKstj|ftjtjd|}tj|j|jdfd}|tj|j|jdfd}||	|
|
|jdk	r|ddn|jrt
jd|S)N)stdoutrr)rrcrzdone.mainzdone.
)
subprocessPopenPIPE	threadingZThreadrKrr=rwaitr$rrrqr)rcmdrdr}t1t2r,r,r-run_commands"
zSubprocessMixin.run_command)FN)rrrrZrrKr!r,r,r,r-rs
rcCstdd|S)z,Normalize a python package name a la PEP 503z[-_.]+r;)r>subrz)rPr,r,r-normalize_namesr#)NN)r)N)N)NT)rcollectionsr
contextlibrZglobrrrrZloggingrarr>rrImportErrorrrqrrrrZdummy_threadingrrrcompatrrrr	r
rrr
rrrrrrrrrrrrrZ	getLoggerrrrrrGrFr6r>r:rVr#r?r^rnrtryrrrrcontextmanagerrrrrrrrrrVERBOSEr rr-r/r1r3r6r9Ir@r<rCrDrErHrMrPrQrSrTr[rfZARCHIVE_EXTENSIONSrrrrrrrrrrrrrrrrrrrrrrrrrrrr#r,r,r,r-s
\








Yy

	/

		8
)



,H
C]


*)	
7.site-packages/pip/_vendor/distlib/__pycache__/scripts.cpython-38.pyc000064400000025002151464452710021501 0ustar00U

.e?@sddlmZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
ddlmZddl
mZmZmZmZmZeeZdZedZd	Zd
dZGdd
d
eZdS))BytesION)	sysconfigdetect_encodingZipFile)finder)FileOperatorget_export_entryconvert_pathget_executablein_venva


 

 
 
 
 
 
 
 
 
s^#!.*pythonw?[0-9.]*([ 	].*)?$z# -*- coding: utf-8 -*-
import re
import sys
from %(module)s import %(import_name)s
if __name__ == '__main__':
    sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
    sys.exit(%(func)s())
cCsXd|krT|drB|dd\}}d|krT|dsTd||f}n|dsTd|}|S)N z
/usr/bin/env r"z%s "%s"z"%s")
startswithsplit)
executableenvZ_executabler?/usr/lib/python3.8/site-packages/pip/_vendor/distlib/scripts.py_enquote_executable3s

rc@seZdZdZeZdZd'ddZddZe	j
d	rBd
dZdd
Z
ddZd(ddZddZeZddZddZd)ddZddZeddZejddZejd ksejd	krejd krd!d"Zd*d#d$Zd+d%d&ZdS),ScriptMakerz_
    A class to copy or create scripts from source scripts or callable
    specifications.
    NTFcCsz||_||_||_d|_d|_tjdkp:tjdko:tjdk|_t	d|_
|pRt||_tjdkprtjdkortjdk|_
dS)NFposixjava)X.Ynt)
source_dir
target_dir
add_launchersforceclobberosname_nameset_modesetvariantsr_fileop_is_nt)selfrrrdry_runZfileoprrr__init__Ls

zScriptMaker.__init__cCs@|ddr<|jr %srspythonwr|rd)!r!r/r1rr
rr`rr'Znewerr:ror6r9r*readliner;Zget_command_name
FIRST_LINE_REmatchr0groupcloseZ	copy_filer$rprqinforseekrUryr7)r)rrsZadjustrvfZ
first_linerrIrSlinesrTrtrwrrr_copy_script)sX



zScriptMaker._copy_scriptcCs|jjSr_r'r*)r)rrrr*]szScriptMaker.dry_runcCs||j_dSr_r)r)valuerrrr*asrcCsHtddkrd}nd}d||f}tddd}t||j}|S)	NPZ64Z32z%s%s.exerVrr)structcalcsize__name__rsplitrfindbytes)r)Zkindbitsr"Zdistlib_packagerJrrrrhiszScriptMaker._get_launchercCs6g}t|}|dkr"|||n|j|||d|S)a
        Make a script.

        :param specification: The specification, which is either a valid export
                              entry specification (to make a script from a
                              callable) or a filename (to make a script by
                              copying from a source location).
        :param options: A dictionary of options controlling script generation.
        :return: A list of all absolute pathnames written to.
        Nrz)r	rr)r)
specificationr2rsr]rrrmakewszScriptMaker.makecCs$g}|D]}||||q|S)z
        Take a list of specifications and make scripts from them,
        :param specifications: A list of specifications.
        :return: A list of all absolute pathnames written to,
        )extendr)r)Zspecificationsr2rsrrrr
make_multipleszScriptMaker.make_multiple)TFN)rLN)N)N)N) r
__module____qualname____doc__SCRIPT_TEMPLATErYrr+r4rGrHrr=r@rKrUr^_DEFAULT_MANIFESTrarcryrrpropertyr*setterr!r"r#rhrrrrrrrCs6

84
4


r)iorZloggingr!rerrGcompatrrrZ	resourcesrutilrr	r
rrZ	getLoggerrr:striprcompilerrrobjectrrrrrs


site-packages/pip/_vendor/distlib/__pycache__/index.cpython-38.opt-1.pyc000064400000041737151464452710022075 0ustar00U

.eJR@sddlZddlZddlZddlZddlZddlZzddlmZWn ek
r`ddl	mZYnXddl
mZddlm
Z
mZmZmZmZmZddlmZmZmZeeZdZdZGd	d
d
eZdS)N)Thread)DistlibException)HTTPBasicAuthHandlerRequestHTTPPasswordMgrurlparsebuild_openerstring_types)cached_propertyzip_dirServerProxyzhttps://pypi.org/pypipypic@seZdZdZdZd*ddZddZdd	Zd
dZdd
Z	ddZ
ddZd+ddZd,ddZ
d-ddZd.ddZddZd/ddZd0d d!Zd1d"d#Zd$d%Zd&d'Zd2d(d)ZdS)3PackageIndexzc
    This class represents a package index compatible with PyPI, the Python
    Package Index.
    s.----------ThIs_Is_tHe_distlib_index_bouNdaRY_$Nc
Cs|pt|_|t|j\}}}}}}|s<|s<|s<|dkrJtd|jd|_d|_d|_d|_t	t
jdR}dD]F}	z,tj
|	dg||d}
|
dkr|	|_WqWqttk
rYqtXqtW5QRXdS)	z
        Initialise an instance.

        :param url: The URL of the index. If not specified, the URL for PyPI is
                    used.
        )ZhttpZhttpszinvalid repository: %sNw)gpgZgpg2z	--versionstdoutstderrr)
DEFAULT_INDEXurlread_configurationrrpassword_handlerssl_verifierrgpg_homeopenosdevnull
subprocessZ
check_callOSError)selfrZschemenetlocpathZparamsZqueryZfragZsinksrcr%=/usr/lib/python3.8/site-packages/pip/_vendor/distlib/index.py__init__$s(

zPackageIndex.__init__cCs&ddlm}ddlm}|}||S)zs
        Get the distutils command for interacting with PyPI configurations.
        :return: the command.
        r)Distribution)
PyPIRCCommand)Zdistutils.corer(Zdistutils.configr))r r(r)dr%r%r&_get_pypirc_commandAsz PackageIndex._get_pypirc_commandcCsR|}|j|_|}|d|_|d|_|dd|_|d|j|_dS)z
        Read the PyPI access configuration as supported by distutils, getting
        PyPI to do the actual work. This populates ``username``, ``password``,
        ``realm`` and ``url`` attributes from the configuration.
        usernamepasswordrealmr
repositoryN)r+rr/Z_read_pypircgetr,r-r.)r cZcfgr%r%r&rKszPackageIndex.read_configurationcCs$||}||j|jdS)z
        Save the PyPI access configuration. You must have set ``username`` and
        ``password`` attributes before calling this method.

        Again, distutils is used to do the actual work.
        N)check_credentialsr+Z
_store_pypircr,r-)r r1r%r%r&save_configurationZszPackageIndex.save_configurationcCs\|jdks|jdkrtdt}t|j\}}}}}}||j||j|jt||_	dS)zp
        Check that ``username`` and ``password`` have been set, and raise an
        exception if not.
        Nz!username and password must be set)
r,r-rrrrZadd_passwordr.rr)r Zpm_r!r%r%r&r2fszPackageIndex.check_credentialscCs\|||}d|d<||g}||}d|d<||g}||S)aq
        Register a distribution on PyPI, using the provided metadata.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the distribution to be
                         registered.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        Zverify:actionZsubmit)r2validatetodictencode_requestitemssend_request)r metadatar*requestZresponser%r%r&registerrs

zPackageIndex.registercCsF|}|sq:|d}||td||fq|dS)ar
        Thread runner for reading lines of from a subprocess into a buffer.

        :param name: The logical name of the stream (used for logging only).
        :param stream: The stream to read from. This will typically a pipe
                       connected to the output stream of a subprocess.
        :param outbuf: The list to append the read lines to.
        utf-8z%s: %sN)readlinedecoderstripappendloggerdebugclose)r namestreamZoutbufr#r%r%r&_readers

zPackageIndex._readerc	Cs|jdddg}|dkr|j}|r.|d|g|dk	rF|dddgt}tj|tj|d	}|d
dd|d
||gt	
dd|||fS)a
        Return a suitable command for signing a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The signing command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        --status-fd2--no-ttyN	--homedirz--batchz--passphrase-fd0z.ascz
--detach-signz--armorz--local-userz--outputinvoking: %s )rrextendtempfileZmkdtemprr"joinbasenamerCrD)r filenamesigner
sign_passwordkeystorecmdZtdZsfr%r%r&get_sign_commands"
zPackageIndex.get_sign_commandc	Cstjtjd}|dk	r tj|d<g}g}tj|f|}t|jd|j|fd}|t|jd|j|fd}||dk	r|j	||j
||||j
||fS)a
        Run a command in a child process , passing it any input data specified.

        :param cmd: The command to run.
        :param input_data: If specified, this must be a byte string containing
                           data to be sent to the child process.
        :return: A tuple consisting of the subprocess' exit code, a list of
                 lines read from the subprocess' ``stdout``, and a list of
                 lines read from the subprocess' ``stderr``.
        rNstdinr)targetargsr)rPIPEPopenrrHrstartrrZwriterEwaitrR
returncode)	r rXZ
input_datakwargsrrpt1t2r%r%r&run_commands&

zPackageIndex.run_commandc
CsD|||||\}}|||d\}}}	|dkr@td||S)aR
        Sign a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The absolute pathname of the file where the signature is
                 stored.
        r>rz&sign command failed with error code %s)rYrgencoder)
r rTrUrVrWrXsig_filer$rrr%r%r&	sign_files

zPackageIndex.sign_filesdistsourcec	Cs(|tj|s td|||}d}	|rZ|jsJt	dn|
||||}	t|d}
|
}W5QRXt
|}t
|}
|dd||||
ddtj||fg}|	rt|	d}
|
}W5QRX|d	tj|	|fttj|	|||}||S)
a
        Upload a release file to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the file to be uploaded.
        :param filename: The pathname of the file to be uploaded.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param filetype: The type of the file being uploaded. This is the
                        distutils command which produced that file, e.g.
                        ``sdist`` or ``bdist_wheel``.
        :param pyversion: The version of Python which the release relates
                          to. For code compatible with any Python, this would
                          be ``source``, otherwise it would be e.g. ``3.2``.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        z
not found: %sNz)no signing program available - not signedrbZfile_upload1)r5Zprotocol_versionfiletype	pyversion
md5_digest
sha256_digestcontentZ
gpg_signature)r2rr"existsrr6r7rrCZwarningrjrreadhashlibmd5	hexdigestZsha256updaterSrBshutilZrmtreedirnamer8r9r:)r r;rTrUrVrorprWr*rifZ	file_datarqrrfilesZsig_datar<r%r%r&upload_filesD
zPackageIndex.upload_filec
Cs|tj|s td|tj|d}tj|sFtd|||j|j	}}t
|}dd|fd|fg}d||fg}|||}	|
|	S)a2
        Upload documentation to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the documentation to be
                         uploaded.
        :param doc_dir: The pathname of the directory which contains the
                        documentation. This should be the directory that
                        contains the ``index.html`` for the documentation.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        znot a directory: %rz
index.htmlz
not found: %r)r5Z
doc_uploadrFversionrs)r2rr"isdirrrRrtr6rFrrgetvaluer8r:)
r r;Zdoc_dirfnrFrZzip_datafieldsr}r<r%r%r&upload_documentation(s 
z!PackageIndex.upload_documentationcCsT|jdddg}|dkr|j}|r.|d|g|d||gtdd||S)	a|
        Return a suitable command for verifying a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The verifying command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        rIrJrKNrLz--verifyrNrO)rrrPrCrDrR)r signature_filename
data_filenamerWrXr%r%r&get_verify_commandDszPackageIndex.get_verify_commandcCsH|jstd||||}||\}}}|dkr@td||dkS)a6
        Verify a signature for a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: True if the signature was verified, else False.
        z0verification unavailable because gpg unavailable)rrz(verify command failed with error code %sr)rrrrg)r rrrWrXr$rrr%r%r&verify_signature\szPackageIndex.verify_signaturec	Csl|dkrd}tdn6t|ttfr0|\}}nd}tt|}td|t|d}|t	|}z|}	d}
d}d}d}
d	|	krt|	d
}|r||
|
||
|
}|sq|t|7}|||r|||
d7}
|r||
|
|qW5|
XW5QRX|dkr0||kr0td||f|rh|}||kr\td
||||ftd|dS)a
        This is a convenience method for downloading a file from an URL.
        Normally, this will be a file from the index, though currently
        no check is made for this (i.e. a file can be downloaded from
        anywhere).

        The method is just like the :func:`urlretrieve` function in the
        standard library, except that it allows digest computation to be
        done during download and checking that the downloaded data
        matched any expected value.

        :param url: The URL of the file to be downloaded (assumed to be
                    available via an HTTP GET request).
        :param destfile: The pathname where the downloaded file is to be
                         saved.
        :param digest: If specified, this must be a (hasher, value)
                       tuple, where hasher is the algorithm used (e.g.
                       ``'md5'``) and ``value`` is the expected value.
        :param reporthook: The same as for :func:`urlretrieve` in the
                           standard library.
        NzNo digest specifiedrwzDigest specified: %swbi rzcontent-lengthzContent-Lengthrz1retrieval incomplete: got only %d out of %d bytesz.%s digest mismatch for %s: expected %s, got %szDigest verified: %s)rCrD
isinstancelisttuplegetattrrvrr:rrEinfointrulenr`ryrrx)r rZdestfileZdigestZ
reporthookZdigesterZhasherZdfpZsfpheadersZ	blocksizesizeruZblocknumblockactualr%r%r&
download_fileus^




zPackageIndex.download_filecCs:g}|jr||j|jr(||jt|}||S)z
        Send a standard library :class:`Request` to PyPI and return its
        response.

        :param req: The request to send.
        :return: The HTTP response from PyPI (a standard library HTTPResponse).
        )rrBrr	r)r ZreqZhandlersZopenerr%r%r&r:szPackageIndex.send_requestc
Csg}|j}|D]L\}}t|ttfs*|g}|D]*}|d|d|dd|dfq.q|D].\}}	}
|d|d||	fdd|
fq`|d|ddfd|}d|}|tt|d}
t	|j
||
S)	a&
        Encode fields and files for posting to an HTTP server.

        :param fields: The fields to send as a list of (fieldname, value)
                       tuples.
        :param files: The files to send as a list of (fieldname, filename,
                      file_bytes) tuple.
        s--z)Content-Disposition: form-data; name="%s"r>z8Content-Disposition: form-data; name="%s"; filename="%s"s
smultipart/form-data; boundary=)zContent-typezContent-length)boundaryrrrrPrhrRstrrrr)r rr}partsrkvaluesvkeyrTvalueZbodyZctrr%r%r&r8sD


zPackageIndex.encode_requestcCsFt|trd|i}t|jdd}z|||p.dWS|dXdS)NrFg@)ZtimeoutrEand)rr
r
rsearch)r ZtermsoperatorZ	rpc_proxyr%r%r&rs
zPackageIndex.search)N)N)N)N)NNrkrlN)N)N)NN)N)__name__
__module____qualname____doc__rr'r+rr3r2r=rHrYrgrjr~rrrrr:r8rr%r%r%r&rs6



#

9


M+r)rvZloggingrrzrrQZ	threadingrImportErrorZdummy_threadingrcompatrrrrr	r
utilrrr
Z	getLoggerrrCrZ
DEFAULT_REALMobjectrr%r%r%r&s  
site-packages/pip/_vendor/distlib/__pycache__/metadata.cpython-38.pyc000064400000063760151464452710021607 0ustar00U

.e*@sdZddlmZddlZddlmZddlZddlZddlZddl	m
Z
mZddlm
Z
mZmZddlmZdd	lmZmZdd
lmZmZeeZGddde
ZGd
dde
ZGddde
ZGddde
ZdddgZdZ dZ!e"dZ#e"dZ$dZ%dZ&dZ'dZ(dZ)dZ*d Z+e*d!Z,d"Z-e.Z/e/0e%e/0e&e/0e(e/0e*e/0e,e"d#Z1d$d%Z2d&d'Z3d(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFZ4dGZ5dHZ6dIZ7dJZ8dKZ9dLZ:dMZ;e<Z=e"dNZ>dXdPdQZ?GdRdSdSe<Z@dTZAdUZBdVZCGdWdde<ZDdS)YzImplementation of the Metadata for Python packages PEPs.

Supports all metadata formats (1.0, 1.1, 1.2, and 2.0 experimental).
)unicode_literalsN)message_from_file)DistlibException__version__)StringIOstring_types	text_type)	interpret)extract_by_key
get_extras)
get_schemePEP440_VERSION_REc@seZdZdZdS)MetadataMissingErrorzA required metadata is missingN__name__
__module____qualname____doc__rr@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/metadata.pyrsrc@seZdZdZdS)MetadataConflictErrorz>Attempt to read or write metadata fields that are conflictual.Nrrrrrr src@seZdZdZdS) MetadataUnrecognizedVersionErrorz Unknown metadata version number.Nrrrrrr$src@seZdZdZdS)MetadataInvalidErrorzA metadata value is invalidNrrrrrr(srMetadataPKG_INFO_ENCODINGPKG_INFO_PREFERRED_VERSIONutf-81.1z

       \|	
        )Metadata-VersionNameVersionPlatformSummaryDescriptionKeywords	Home-pageAuthorAuthor-emailLicense)r r!r"r#Supported-Platformr$r%r&r'r(r)r*
ClassifierDownload-URL	ObsoletesProvidesRequires)r.r/r0r,r-)r r!r"r#r+r$r%r&r'r(r)
MaintainerMaintainer-emailr*r,r-Obsoletes-DistProject-URL
Provides-Dist
Requires-DistRequires-PythonRequires-External)r5r6r7r3r8r1r2r4)r r!r"r#r+r$r%r&r'r(r)r1r2r*r,r-r3r4r5r6r7r8Private-VersionObsoleted-BySetup-Requires-Dist	ExtensionProvides-Extra)r9r=r:r;r<)Description-Content-Typer0r/)r>z"extra\s*==\s*("([^"]+)"|'([^']+)')cCsL|dkrtS|dkrtS|dkr$tS|dkr4ttS|dkr@tSt|dS)N1.0r1.2)1.32.12.0)_241_FIELDS_314_FIELDS_345_FIELDS_566_FIELDS_426_FIELDSr)versionrrr_version2fieldlistpsrJc
CsBdd}g}|D]"\}}|gddfkr,q||qddddd	d
g}|D]}|tkrvd|krv|dtd||tkrd|kr|dtd||tkrd|kr|dtd
||tkrd|kr|dtd||tkrd
|kr|dkr|d
td||t	krLd	|krL|d	td|qLt
|dkrZ|dSt
|dkr|td|tdd|ko||t}d|ko||t
}d
|ko||t}d	|ko||t}	t|t|t|t|	dkrtd|s |s |s |	s t|kr tS|r*dS|r4dS|r>d
Sd	S)z5Detect the best version depending on the fields used.cSs|D]}||krdSqdS)NTFr)keysmarkersmarkerrrr_has_markersz"_best_version.._has_markerUNKNOWNNr?rr@rArCrBzRemoved 1.0 due to %szRemoved 1.1 due to %szRemoved 1.2 due to %szRemoved 1.3 due to %sr%zRemoved 2.1 due to %szRemoved 2.0 due to %srrz)Out of options - unknown metadata set: %szUnknown metadata setz,You used incompatible 1.1/1.2/2.0/2.1 fields)itemsappendrDremoveloggerdebugrErFrGrHlenr_314_MARKERS_345_MARKERS_566_MARKERS_426_MARKERSintr)
fieldsrNrKkeyvalueZpossible_versionsZis_1_1Zis_1_2Zis_2_1Zis_2_0rrr
_best_version~s`






&
r^r r!r"r#r+r$r%r&r'r(r)r1r2r*r,r-r3r5r6r;r7r8r0r/r.r4r9r:r<r=)metadata_versionnamerIplatformZsupported_platformsummarydescriptionkeywords	home_pageauthorauthor_email
maintainermaintainer_emaillicense
classifierdownload_urlobsoletes_dist
provides_dist
requires_distsetup_requires_distrequires_pythonrequires_externalrequiresprovides	obsoletesproject_urlZprivate_versionZobsoleted_by	extensionZprovides_extra)r6r3r5)r7)r")r#r,r.r0r/r3r5r6r8r4r+r;r=r<)r4)r&)r(r1r$r%z[^A-Za-z0-9.]+FcCs0|r$td|}td|dd}d||fS)zhReturn the distribution name with version.

    If for_filename is true, return a filename-escaped form.- .z%s-%s)	_FILESAFEsubreplace)r`rIZfor_filenamerrr_get_name_and_versionsr~c@s
eZdZdZd?ddZddZdd	Zd
dZdd
ZddZ	ddZ
ddZddZddZ
ddZd@ddZddZdd Zd!d"Zd#d$ZdAd%d&ZdBd'd(ZdCd)d*Zd+d,Zefd-d.ZdDd/d0ZdEd1d2Zd3d4Zd5d6Zd7d8Zd9d:Zd;d<Z d=d>Z!dS)FLegacyMetadataaaThe legacy metadata of a release.

    Supports versions 1.0, 1.1 and 1.2 (auto-detected). You can
    instantiate the class with one of these arguments (or none):
    - *path*, the path to a metadata file
    - *fileobj* give a file-like object with metadata as content
    - *mapping* is a dict-like object
    - *scheme* is a version scheme name
    NdefaultcCsz|||gddkrtdi|_g|_d|_||_|dk	rH||n.|dk	r\||n|dk	rv|||	dS)N'path, fileobj and mapping are exclusive)
count	TypeError_fieldsZrequires_filesZ
_dependenciesschemeread	read_fileupdateset_metadata_version)selfpathfileobjmappingrrrr__init__s
zLegacyMetadata.__init__cCst|j|jd<dSNr )r^rrrrrr"sz#LegacyMetadata.set_metadata_versioncCs|d||fdS)Nz%s: %s
)write)rrr`r]rrr_write_field%szLegacyMetadata._write_fieldcCs
||SN)getrr`rrr__getitem__(szLegacyMetadata.__getitem__cCs|||Sr)set)rr`r]rrr__setitem__+szLegacyMetadata.__setitem__cCs8||}z|j|=Wntk
r2t|YnXdSr)
_convert_namerKeyError)rr`
field_namerrr__delitem__.s

zLegacyMetadata.__delitem__cCs||jkp|||jkSr)rrrrrr__contains__5s
zLegacyMetadata.__contains__cCs(|tkr|S|dd}t||S)Nrx_)_ALL_FIELDSr}lower_ATTR2FIELDrrrrrr9szLegacyMetadata._convert_namecCs|tks|tkrgSdS)NrO)_LISTFIELDS_ELEMENTSFIELDrrrr_default_value?szLegacyMetadata._default_valuecCs&|jdkrtd|Std|SdS)Nr?r
)r__LINE_PREFIX_PRE_1_2r|_LINE_PREFIX_1_2rr]rrr_remove_line_prefixDs
z"LegacyMetadata._remove_line_prefixcCs|tkr||St|dSr)rAttributeErrorrrrr__getattr__JszLegacyMetadata.__getattr__FcCst|d|d|S)zhReturn the distribution name with version.

        If filesafe is true, return a filename-escaped form.r!r")r~)rZfilesaferrrget_fullnameUszLegacyMetadata.get_fullnamecCs||}|tkS)z+return True if name is a valid metadata key)rrrrrris_field[s
zLegacyMetadata.is_fieldcCs||}|tkSr)rrrrrris_multi_field`s
zLegacyMetadata.is_multi_fieldcCs.tj|ddd}z||W5|XdS)z*Read the metadata values from a file path.rrencodingN)codecsopencloser)rfilepathfprrrrdszLegacyMetadata.readcCst|}|d|jd<tD]p}||kr(q|tkrf||}|tkrX|dk	rXdd|D}|||q||}|dk	r|dkr|||qdS)z,Read the metadata values from a file object.zmetadata-versionr NcSsg|]}t|dqS,)tuplesplit.0r]rrr
ysz,LegacyMetadata.read_file..rO)rrrrZget_all_LISTTUPLEFIELDSr)rZfileobmsgfieldvaluesr]rrrrls
zLegacyMetadata.read_filecCs0tj|ddd}z|||W5|XdS)z&Write the metadata fields to filepath.wrrN)rrr
write_file)rrskip_unknownrrrrrszLegacyMetadata.writecCs|t|dD]}||}|r8|dgdgfkr8q|tkrV|||d|q|tkr|dkr|jdkr~|dd}n|dd}|g}|t	krd	d
|D}|D]}||||qqdS)z0Write the PKG-INFO format data to a file object.r rOrr%rrrz	
       |cSsg|]}d|qSrjoinrrrrrsz-LegacyMetadata.write_file..N)
rrJrrrrrr_r}r)rZ
fileobjectrrrr]rrrrs$

zLegacyMetadata.write_filecsvfdd}|sn@t|dr:|D]}||||q$n|D]\}}|||q>|rr|D]\}}|||q^dS)aSet metadata values from the given iterable `other` and kwargs.

        Behavior is like `dict.update`: If `other` has a ``keys`` method,
        they are looped over and ``self[key]`` is assigned ``other[key]``.
        Else, ``other`` is an iterable of ``(key, value)`` iterables.

        Keys that don't match a metadata field or that have an empty value are
        dropped.
        cs"|tkr|r||dSr)rrr)r\r]rrr_setsz#LegacyMetadata.update.._setrKN)hasattrrKrP)rotherkwargsrkvrrrrs

zLegacyMetadata.updatecCsh||}|tks|dkrNt|ttfsNt|trHdd|dD}qzg}n,|tkrzt|ttfszt|trv|g}ng}t	t
jr<|d}t|j
}|tkr|dk	r|D](}||ddstd	|||qnb|tkr|dk	r||s.rr!N;rz$'%s': '%s' is not valid (field '%s')z.'%s': '%s' is not a valid version (field '%s')r%)rr
isinstancelistrrrrrSZisEnabledForloggingZWARNINGr
r_PREDICATE_FIELDSis_valid_matcherwarning_VERSIONS_FIELDSis_valid_constraint_list_VERSION_FIELDSis_valid_version_UNICODEFIELDSrr)rr`r]Zproject_namerrrrrrsV






zLegacyMetadata.setcCs||}||jkr*|tkr&||}|S|tkr@|j|}|S|tkr|j|}|dkr^gSg}|D].}|tkr~||qf||d|dfqf|S|tkr|j|}t	|t
r|dS|j|S)zGet a metadata field.Nrrr)rr_MISSINGrrrrrQrrrr)rr`rr]resvalrrrrs.







zLegacyMetadata.getcs|gg}}dD]}||kr||q|rP|gkrPdd|}t|dD]}||krT||qT|ddkr||fSt|jfdd}t|ftjft	j
ffD]@\}}|D]2}	||	d	}
|
d	k	r||
s|d
|	|
fqq||fS)zkCheck if the metadata is compliant. If strict is True then raise if
        no Name or Version are provided)r!r"zmissing required metadata: %s, )r'r(r r@cs(|D]}|ddsdSqdS)NrrFT)rr)r]rrrrare_valid_constraints#sz3LegacyMetadata.check..are_valid_constraintsNzWrong value for '%s': %s)rrQrrr
rrrrrrr)rstrictmissingwarningsattrrrr[Z
controllerrr]rrrcheck	s8

zLegacyMetadata.checkcCs|d}i}|D]"\}}|r*||jkr||||<q|ddkrd}|D]B\}}|rb||jkrL|dkrx||||<qLdd||D||<qLn8|ddkrd	}|D]"\}}|r||jkr||||<q|S)
zReturn fields as a dict.

        Field names will be converted to use the underscore-lowercase style
        instead of hyphen-mixed case (i.e. home_page instead of Home-page).
        )
)r_r )r`r!)rIr")rbr$)rer')rfr()rgr))rjr*)rcr%)rdr&)rar#)classifiersr,)rlr-r r@))ror6)rqr7)rrr8)rnr5)rmr3)rvr4)rhr1)rir2rvcSsg|]}d|qSrr)rurrrrbsz)LegacyMetadata.todict..r))rtr/)rsr0)rur.)rr)rZskip_missingZmapping_1_0datar\rZmapping_1_2Zmapping_1_1rrrtodict5s&
zLegacyMetadata.todictcCs8|ddkr$dD]}||kr||=q|d|7<dS)Nr r)r.r0r/r6r)rrequirementsrrrradd_requirementsps
zLegacyMetadata.add_requirementscCstt|dSr)rrJrrrrrK{szLegacyMetadata.keysccs|D]
}|VqdSrrK)rr\rrr__iter__~szLegacyMetadata.__iter__csfddDS)Ncsg|]}|qSrrrr\rrrrsz)LegacyMetadata.values..rrrrrrszLegacyMetadata.valuescsfddDS)Ncsg|]}||fqSrrrrrrrsz(LegacyMetadata.items..rrrrrrPszLegacyMetadata.itemscCsd|jj|j|jfS)Nz
<%s %s %s>)	__class__rr`rIrrrr__repr__szLegacyMetadata.__repr__)NNNr)F)F)F)N)F)F)"rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrKrrrPrrrrrrs@




,
,
;rzpydist.jsonz
metadata.jsonZMETADATAc@seZdZdZedZedejZe	Z
edZdZde
Zdddd	Zd
ZdZedfedfe
dfedfdZd
ZdDddZedZdefZdefZdefdefeeedefeeeedefddd
Z[[ddZdEddZddZed d!Z ed"d#Z!e!j"d$d#Z!dFd%d&Z#ed'd(Z$ed)d*Z%e%j"d+d*Z%d,d-Z&d.d/Z'd0d1Z(d2d3Z)d4d5d6d7d8dd9Z*d:d;Z+dGd>d?Z,d@dAZ-dBdCZ.dS)Hrz
    The metadata of a release. This implementation uses 2.0 (JSON)
    metadata where possible. If not possible, it wraps a LegacyMetadata
    instance which handles the key-value metadata format.
    z
^\d+(\.\d+)*$z!^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$z	.{1,2047}rCzdistlib (%s)r)legacy)r`rIrbzqname version license summary description author author_email keywords platform home_page classifiers download_urlzwextras run_requires test_requires build_requires dev_requires provides meta_requires obsoleted_by supports_environments)r_r`rIrb)_legacy_datarNrc	Cs0|||gddkrtdd|_d|_||_|dk	rzz|||||_Wn*tk
rvt||d|_|YnXnd}|rt	|d}|
}W5QRXn|r|
}|dkr|j|jd|_ndt
|ts|d}zt||_||j|Wn0tk
r*tt||d|_|YnXdS)Nrr)rrrbr_	generatorr)rr)rrrrr_validate_mappingrrvalidaterrMETADATA_VERSION	GENERATORrr	decodejsonloads
ValueErrorr)rrrrrrfrrrrs@



zMetadata.__init__)r`rIrjrdrbr6r;r=r,)r-N)r N)
run_requiresbuild_requiresdev_requiresZ
test_requires
meta_requiresextrasmodules
namespacesexportscommandsrZ
source_urlr_c
CsXt|d}t|d}||kr||\}}|jr^|dkrP|dkrHdn|}n|j|}n|dkrjdn|}|dkr|j||}nt}|}|jd}	|	r
|dkr|	d|}nP|dkr|	d}	|	r|	||}n,|	d	}	|	s|jd	}	|	r
|	||}||krT|}n:||kr2t||}n"|jrH|j|}n|j|}|S)
Ncommon_keysmapped_keysr	rrrr
extensionsr	python.commandsrpython.detailspython.exports)object__getattribute__rrr)
rr\commonmappedlkZmakerresultr]sentineldrrrrsD




zMetadata.__getattribute__cCsH||jkrD|j|\}}|p |j|krD||}|sDtd||fdS)Nz.'%s' is an invalid value for the '%s' property)SYNTAX_VALIDATORSrmatchr)rr\r]rpattern
exclusionsmrrr_validate_value+s

zMetadata._validate_valuecCs*|||t|d}t|d}||kr||\}}|jrV|dkrJt||j|<nf|dkrj||j|<nR|jdi}|dkr||d<n2|dkr|di}|||<n|d	i}|||<nh||krt|||nP|d
krt|t	r|
}|r|}ng}|jr||j|<n
||j|<dS)Nr
rrr
r	rrrrrd)rrrrNotImplementedErrorr
setdefault__setattr__rrrr)rr\r]rrrrrrrrr!5s<




zMetadata.__setattr__cCst|j|jdSNT)r~r`rIrrrrname_and_version\szMetadata.name_and_versioncCsF|jr|jd}n|jdg}d|j|jf}||krB|||S)Nr5rtz%s (%s))rrr r`rIrQ)rrsrrrrt`s
zMetadata.providescCs |jr||jd<n
||jd<dS)Nr5rt)rrrrrrrtksc
Cs|jr|}ng}t|pg|j}|D]d}d|kr>d|kr>d}n8d|krLd}n|d|k}|rv|d}|rvt||}|r$||dq$dD]F}d|}	|	|kr||	|jd|g}||j|||dq|S)	a
        Base method to get dependencies, given a set of extras
        to satisfy and an optional environment context.
        :param reqts: A list of sometimes-wanted dependencies,
                      perhaps dependent on extras and environment.
        :param extras: A list of optional components being requested.
        :param env: An optional environment for marker evaluation.
        extraenvironmentTrs)ZbuildZdevZtestz:%s:z%s_requires)renv)	rrrrr
extendrRrget_requirements)
rreqtsrr'rrZincluderMr\errrr)rs2	



zMetadata.get_requirementscCs|jr|S|jSr)r_from_legacyrrrrr
dictionaryszMetadata.dictionarycCs|jrtnt|j|jSdSr)rrrrDEPENDENCY_KEYSrrrrdependenciesszMetadata.dependenciescCs|jrtn|j|dSr)rrrrrrrrr/sc	Cs|d|jkrtg}|jD]"\}}||kr$||kr$||q$|rbdd|}t||D]\}}||||qjdS)Nr_zMissing metadata items: %sr)	rrrMANDATORY_KEYSrPrQrrr)	rrrrr\rrrrrrrrszMetadata._validate_mappingcCsB|jr.|jd\}}|s|r>td||n||j|jdS)NTz#Metadata: missing: %s, warnings: %s)rrrSrrrr)rrrrrrrszMetadata.validatecCs(|jr|jdSt|j|j}|SdSr")rrrr
INDEX_KEYS)rrrrrrszMetadata.todictc
Cs|jr|jrt|j|jd}|jd}dD]*}||kr.|dkrHd}n|}||||<q.|dg}|dgkrtg}||d<d	}|D]*\}}||kr||rd
||ig||<q|j|d<i}i}	|S)NrT)r`rIrjrbrcrkrkrr&rd))ror)rprrsrt)rrAssertionErrorrrrrrt)
rrZlmdrnkkwrKokrfrhrrrr,s.

zMetadata._from_legacyr!r"r*r$r%)r`rIrjrbrcrcCsdd}|jr|jrtt}|j}|jD]\}}||kr.||||<q.||j|j}||j|j	}|j
rt|j
|d<t||d<t||d<|S)NcSst}|D]|}|d}|d}|d}|D]V}|sF|sF||q.d}|rVd|}|rp|rld||f}n|}|d||fq.q
|S)Nr%r&rsr2z
extra == "%s"z(%s) and %sr)rraddr)entriesr*r+r%r'ZrlistrrMrrrprocess_entriess"

z,Metadata._to_legacy..process_entriesr=r6r;)rrr3rLEGACY_MAPPINGrPrrrrrsorted)rr9rZnmdr4r6Zr1Zr2rrr
_to_legacyszMetadata._to_legacyFTc	Cs||gddkrtd||r`|jr4|j}n|}|rP|j||dq|j||dn^|jrp|}n|j}|rt	j
||ddddn.t|dd}t	j
||ddddW5QRXdS)	Nrz)Exactly one of path and fileobj is needed)rTr)Zensure_asciiindentZ	sort_keysrr)
rrrrr<rrr,rrdumprr)rrrrrZ	legacy_mdrrrrrrs*
zMetadata.writecCs|jr|j|nr|jdg}d}|D]}d|kr*d|kr*|}qHq*|dkrfd|i}|d|n t|dt|B}t||d<dS)Nrr&r%rsr)rrrr insertrr;)rrralwaysentryZrsetrrrr3szMetadata.add_requirementscCs*|jpd}|jpd}d|jj|j||fS)Nz	(no name)z
no versionz<%s %s %s (%s)>)r`rIrrr_)rr`rIrrrrDs

zMetadata.__repr__)NNNr)N)NN)NNFT)/rrrrrecompileZMETADATA_VERSION_MATCHERIZNAME_MATCHERrZVERSION_MATCHERZSUMMARY_MATCHERrrrr0r1r.r	__slots__rrr
rZ	none_listdictZ	none_dictrrrr!propertyr#rtsetterr)r-r/rrrr,r:r<rrrrrrrrs


-+

'




*


		%
)F)ErZ
__future__rrZemailrrrrBr2rrcompatrrr	rLr
utilrrrIr
rZ	getLoggerrrSrrrr__all__rrrCrrrDrErVrFrWrHrYrGrXrrrZEXTRA_RErJr^rrrrrrrrrrr{r~rZMETADATA_FILENAMEZWHEEL_METADATA_FILENAMEZLEGACY_METADATA_FILENAMErrrrrs










H!



site-packages/pip/_vendor/distlib/__pycache__/database.cpython-38.pyc000064400000122124151464452710021561 0ustar00U

.eU@sdZddlmZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlm
Z
mZddlmZddlmZmZddlmZmZmZmZdd	lmZmZmZmZmZmZm Z d
ddd
dgZ!e"e#Z$dZ%dZ&deddde%dfZ'dZ(Gddde)Z*Gddde)Z+Gdd
d
e)Z,Gddde,Z-Gddde-Z.Gdd
d
e-Z/e.Z0e/Z1Gddde)Z2d)d!d"Z3d#d$Z4d%d&Z5d'd(Z6dS)*zPEP 376 implementation.)unicode_literalsN)DistlibException	resources)StringIO)
get_schemeUnsupportedVersionError)MetadataMETADATA_FILENAMEWHEEL_METADATA_FILENAMELEGACY_METADATA_FILENAME)parse_requirementcached_propertyparse_name_and_versionread_exports
write_exports	CSVReader	CSVWriterDistributionBaseInstalledDistributionInstalledDistributionEggInfoDistributionDistributionPathzpydist-exports.jsonzpydist-commands.jsonZ	INSTALLERRECORD	REQUESTED	RESOURCESSHAREDz
.dist-infoc@s(eZdZdZddZddZddZdS)	_CachezL
    A simple cache mapping names and .dist-info paths to distributions
    cCsi|_i|_d|_dS)zZ
        Initialise an instance. There is normally one for each DistributionPath.
        FN)namepath	generatedselfr#@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/database.py__init__1sz_Cache.__init__cCs|j|jd|_dS)zC
        Clear the cache, setting it to its initial state.
        FN)rclearrr r!r#r#r$r&9s

z_Cache.clearcCs2|j|jkr.||j|j<|j|jg|dS)z`
        Add a distribution to the cache.
        :param dist: The distribution to add.
        N)rr
setdefaultkeyappendr"distr#r#r$addAsz
_Cache.addN)__name__
__module____qualname____doc__r%r&r,r#r#r#r$r-src@seZdZdZdddZddZdd	ZeeeZd
dZ	dd
Z
ddZeddZ
ddZddZdddZddZdddZdS)rzU
    Represents a set of distributions installed on a path (typically sys.path).
    NFcCsD|dkrtj}||_d|_||_t|_t|_d|_td|_	dS)a
        Create an instance from a path, optionally including legacy (distutils/
        setuptools/distribute) distributions.
        :param path: The path to use, as a list of directories. If not specified,
                     sys.path is used.
        :param include_egg: If True, this instance will look for and return legacy
                            distributions as well as those based on PEP 376.
        NTdefault)
sysr
_include_dist_include_eggr_cache
_cache_egg_cache_enabledr_scheme)r"rZinclude_eggr#r#r$r%Os	zDistributionPath.__init__cCs|jSNr7r!r#r#r$_get_cache_enabledcsz#DistributionPath._get_cache_enabledcCs
||_dSr9r:)r"valuer#r#r$_set_cache_enabledfsz#DistributionPath._set_cache_enabledcCs|j|jdS)z,
        Clears the internal cache.
        N)r5r&r6r!r#r#r$clear_cacheks
zDistributionPath.clear_cachec
csDt}|jD]0}t|}|dkr&q|d}|r|js||jjD]
}|Vq0|jrZ|jjD]
}|VqNdS)a5
        Provides an iterator that looks for distributions and returns
        :class:`InstalledDistribution` or
        :class:`EggInfoDistribution` instances for each one of them.

        :rtype: iterator of :class:`InstalledDistribution` and
                :class:`EggInfoDistribution` instances
        N)r7r\r^r5rvaluesr4r6r*r#r#r$get_distributionss	
z"DistributionPath.get_distributionscCsd}|}|js4|D]}|j|kr|}q|qnH|||jjkrZ|jj|d}n"|jr|||jjkr||jj|d}|S)a=
        Looks for a named distribution on the path.

        This function only returns the first result found, as no more than one
        value is expected. If nothing is found, ``None`` is returned.

        :rtype: :class:`InstalledDistribution`, :class:`EggInfoDistribution`
                or ``None``
        Nr)	lowerr7r\r(r^r5rr4r6)r"rresultr+r#r#r$get_distributions

z!DistributionPath.get_distributionc	csd}|dk	rJz|jd||f}Wn$tk
rHtd||fYnX|D]p}t|dsntd|qR|j}|D]H}t	|\}}|dkr||kr|VqRqx||krx|
|rx|VqRqxqRdS)a
        Iterates over all distributions to find which distributions provide *name*.
        If a *version* is provided, it will be used to filter the results.

        This function only returns the first result found, since no more than
        one values are expected. If the directory is not found, returns ``None``.

        :parameter version: a version specifier that indicates the version
                            required, conforming to the format in ``PEP-345``

        :type name: string
        :type version: string
        N%s (%s)zinvalid name or version: %r, %rprovideszNo "provides": %s)r8matcher
ValueErrorrrfhasattrrSrTrkrmatch)	r"rrcrlr+providedpp_namep_verr#r#r$provides_distributions*
z&DistributionPath.provides_distributioncCs(||}|dkrtd|||S)z5
        Return the path to a resource file.
        Nzno distribution named %r found)riLookupErrorget_resource_path)r"r
relative_pathr+r#r#r$
get_file_path!s
zDistributionPath.get_file_pathccsX|D]J}|j}||kr||}|dk	r>||krR||Vq|D]
}|VqFqdS)z
        Return all of the exported entries in a particular category.

        :param category: The category to search for entries.
        :param name: If specified, only entries with that name are returned.
        N)rfexportsre)r"categoryrr+rYdvr#r#r$get_exported_entries*sz%DistributionPath.get_exported_entries)NF)N)N)r-r.r/r0r%r;r=propertyZ
cache_enabledr>r\r^classmethodrdrfrirtrxr}r#r#r#r$rKs

,

)	c@seZdZdZdZdZddZeddZeZ	eddZ
ed	d
ZddZed
dZ
eddZeddZeddZeddZddZddZddZddZdS) rz
    A base class for distributions, whether installed or from indexes.
    Either way, it must have some metadata, so that's all that's needed
    for construction.
    FcCsL||_|j|_|j|_|j|_d|_d|_d|_d|_t	|_
i|_dS)z
        Initialise an instance.
        :param metadata: The instance of :class:`Metadata` describing this
        distribution.
        N)rDrrgr(rcZlocatordigestextrascontextrHZ
download_urlsZdigests)r"rDr#r#r$r%OszDistribution.__init__cCs|jjS)zH
        The source archive download URL for this distribution.
        )rD
source_urlr!r#r#r$r`szDistribution.source_urlcCsd|j|jfS)zX
        A utility property which displays the name and version in parentheses.
        rjrrcr!r#r#r$name_and_versioniszDistribution.name_and_versioncCs.|jj}d|j|jf}||kr*|||S)z
        A set of distribution names and versions provided by this distribution.
        :return: A set of "name (version)" strings.
        rj)rDrkrrcr))r"Zplistsr#r#r$rkps

zDistribution.providescCs8|j}td|t||}t|j||j|jdS)Nz%Getting requirements from metadata %r)rrE)	rDrSrTZtodictgetattrrHZget_requirementsrr)r"Zreq_attrmdZreqtsr#r#r$_get_requirements|s
zDistribution._get_requirementscCs
|dS)Nrun_requiresrr!r#r#r$rszDistribution.run_requirescCs
|dS)N
meta_requiresrr!r#r#r$rszDistribution.meta_requirescCs
|dS)Nbuild_requiresrr!r#r#r$rszDistribution.build_requirescCs
|dS)N
test_requiresrr!r#r#r$rszDistribution.test_requirescCs
|dS)Ndev_requiresrr!r#r#r$rszDistribution.dev_requiresc
	Cst|}t|jj}z||j}Wn6tk
rZtd||	d}||}YnX|j
}d}|jD]D}t|\}}	||krqlz|
|	}WqWqltk
rYqlXql|S)z
        Say if this instance matches (fulfills) a requirement.
        :param req: The requirement to match.
        :rtype req: str
        :return: True if it matches, else False.
        +could not read version %r - using name onlyrF)r
rrDrCrlrequirementrrSwarningsplitr(rkrro)
r"reqrYrCrlrrhrqrrrsr#r#r$matches_requirements,	


z Distribution.matches_requirementcCs(|jrd|j}nd}d|j|j|fS)zC
        Return a textual representation of this instance,
        z [%s]r?z)rrrc)r"suffixr#r#r$__repr__szDistribution.__repr__cCs>t|t|k	rd}n$|j|jko8|j|jko8|j|jk}|S)a<
        See if this distribution is the same as another.
        :param other: The distribution to compare with. To be equal to one
                      another. distributions must have the same type, name,
                      version and source_url.
        :return: True if it is the same, else False.
        F)typerrcr)r"otherrhr#r#r$__eq__s

zDistribution.__eq__cCst|jt|jt|jS)zH
        Compute hash in a way which matches the equality test.
        )hashrrcrr!r#r#r$__hash__szDistribution.__hash__N)r-r.r/r0Zbuild_time_dependency	requestedr%r~rZdownload_urlrrkrrrrrrrrrrr#r#r#r$r=s4







"
cs0eZdZdZdZdfdd	ZdddZZS)	rz]
    This is the base class for installed distributions (whether PEP 376 or
    legacy).
    Ncs tt||||_||_dS)a
        Initialise an instance.
        :param metadata: An instance of :class:`Metadata` which describes the
                         distribution. This will normally have been initialised
                         from a metadata file in the ``path``.
        :param path:     The path of the ``.dist-info`` or ``.egg-info``
                         directory for the distribution.
        :param env:      This is normally the :class:`DistributionPath`
                         instance where this distribution was found.
        N)superrr%r	dist_path)r"rDrrE	__class__r#r$r%sz"BaseInstalledDistribution.__init__cCsd|dkr|j}|dkr"tj}d}ntt|}d|j}||}t|dd}d||fS)a
        Get the hash of some data, using a particular hash algorithm, if
        specified.

        :param data: The data to be hashed.
        :type data: bytes
        :param hasher: The name of a hash implementation, supported by hashlib,
                       or ``None``. Examples of valid values are ``'sha1'``,
                       ``'sha224'``, ``'sha384'``, '``sha256'``, ``'md5'`` and
                       ``'sha512'``. If no hasher is specified, the ``hasher``
                       attribute of the :class:`InstalledDistribution` instance
                       is used. If the hasher is determined to be ``None``, MD5
                       is used as the hashing algorithm.
        :returns: The hash of the data. If a hasher was explicitly specified,
                  the returned hash will be prefixed with the specified hasher
                  followed by '='.
        :rtype: str
        Nr?z%s==ascii%s%s)	hasherhashlibmd5rrbase64Zurlsafe_b64encoderstripdecode)r"datarprefixrr#r#r$get_hashs

z"BaseInstalledDistribution.get_hash)N)N)r-r.r/r0rr%r
__classcell__r#r#rr$rscseZdZdZdZd'fdd	ZddZdd	Zd
dZe	dd
Z
ddZddZddZ
ddZd(ddZddZe	ddZd)ddZdd Zd!d"Zd#d$Zd%d&ZejZZS)*ra
    Created with the *path* of the ``.dist-info`` directory provided to the
    constructor. It reads the metadata contained in ``pydist.json`` when it is
    instantiated., or uses a passed in Metadata instance (useful for when
    dry-run mode is being used).
    Zsha256Nc
	sJg|_t||_}|dkr*td||rP|jrP||jjkrP|jj|j}nt|dkr|	t
}|dkrt|	t}|dkr|	d}|dkrtdt
|ft
|}t|dd}W5QRXtt|||||r|jr|j||	d}|dk	|_tj|d}tj|rFt|d}|}	W5QRX|	|_dS)	Nzfinder unavailable for %sZMETADATAzno %s found in %sr@rAr
top_level.txtrb)modulesrrIrXrmr7r5rrDrJr
rrPrQrRr	rrr%r,rosrOexistsopenread
splitlines)
r"rrDrErXrYr[rqfrrr#r$r%s8





zInstalledDistribution.__init__cCsd|j|j|jfS)Nz#rrcrr!r#r#r$r=s
zInstalledDistribution.__repr__cCsd|j|jfSNz%s %srr!r#r#r$__str__AszInstalledDistribution.__str__c

Csg}|d}t|\}t|dF}|D]:}ddtt|dD}||\}}}	||||	fq.W5QRXW5QRX|S)a"
        Get the list of installed files for the distribution
        :return: A list of tuples of path, hash and size. Note that hash and
                 size might be ``None`` for some entries. The path is exactly
                 as stored in the file (which is as in PEP 376).
        rr[cSsg|]}dqSr9r#).0ir#r#r$
Ssz6InstalledDistribution._get_records..)get_distinfo_resourcerPrQrRrrangelenr))
r"resultsrYr[Z
record_readerrowmissingrchecksumsizer#r#r$_get_recordsDs
&z"InstalledDistribution._get_recordscCsi}|t}|r|}|S)a
        Return the information exported by this distribution.
        :return: A dictionary of exports, mapping an export category to a dict
                 of :class:`ExportEntry` instances describing the individual
                 export entries, and keyed by name.
        )rEXPORTS_FILENAMEr)r"rhrYr#r#r$ry[s

zInstalledDistribution.exportsc	Cs8i}|t}|r4t|}t|}W5QRX|S)z
        Read exports data from a file in .ini format.

        :return: A dictionary of exports, mapping an export category to a list
                 of :class:`ExportEntry` instances describing the individual
                 export entries.
        )rrrPrQrRr)r"rhrYr[r#r#r$ris
z"InstalledDistribution.read_exportsc	Cs.|t}t|d}t||W5QRXdS)a
        Write a dictionary of exports to a file in .ini format.
        :param exports: A dictionary of exports, mapping an export category to
                        a list of :class:`ExportEntry` instances describing the
                        individual export entries.
        wN)get_distinfo_filerrr)r"ryZrfrr#r#r$rxs
z#InstalledDistribution.write_exportscCs|d}t|R}t|d<}|D]0\}}||kr*|W5QRW5QRSq*W5QRXW5QRXtd|dS)aW
        NOTE: This API may change in the future.

        Return the absolute path to a resource file with the given relative
        path.

        :param relative_path: The path, relative to .dist-info, of the resource
                              of interest.
        :return: The absolute path where the resource is to be found.
        rrz3no resource file with relative path %r is installedN)rrPrQrRrKeyError)r"rwrYr[Zresources_readerrelativeZdestinationr#r#r$rvs
6z'InstalledDistribution.get_resource_pathccs|D]
}|VqdS)z
        Iterates over the ``RECORD`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: iterator of (path, hash, size)
        N)r)r"rhr#r#r$list_installed_filessz*InstalledDistribution.list_installed_filesFcCs(tj|d}tj|j}||}tj|d}|d}td||rRdSt|}|D]}tj	|sz|
drd}	}
n4dtj|}
t|d}|
|}	W5QRX||s|r||rtj||}|||	|
fq`||r
tj||}||ddfW5QRX|S)z
        Writes the ``RECORD`` file, using the ``paths`` iterable passed in. Any
        existing ``RECORD`` file is silently overwritten.

        prefix is used to determine when to write absolute paths.
        r?rcreating %sNz.pycz.pyoz%dr)rrrOdirname
startswithrrSinforisdirrLgetsizerrrrelpathZwriterow)r"pathsrdry_runbaseZbase_under_prefixrecord_pathwriterr
hash_valuerfpr#r#r$write_installed_filess0



z+InstalledDistribution.write_installed_filesc
Csg}tj|j}|d}|D]\}}}tj|sHtj||}||krRq$tj|sr||dddfq$tj	|r$t
tj|}|r||kr||d||fq$|r$d|kr|ddd}nd	}t
|d
2}	||	|}
|
|kr||d||
fW5QRXq$|S)
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        rrTFr=rrNrr)rrrrrisabsrOrr)isfilestrrrrrr)r"
mismatchesrrrrrZactual_sizerrZactual_hashr#r#r$check_installed_filess.	

z+InstalledDistribution.check_installed_filesc	Csi}tj|jd}tj|rtj|ddd}|}W5QRX|D]8}|dd\}}|dkr||	|g
|qL|||<qL|S)a
        A dictionary of shared locations whose keys are in the set 'prefix',
        'purelib', 'platlib', 'scripts', 'headers', 'data' and 'namespace'.
        The corresponding value is the absolute path of that category for
        this distribution, and takes into account any paths selected by the
        user at installation time (e.g. via command-line arguments). In the
        case of the 'namespace' key, this would be a list of absolute paths
        for the roots of namespace packages in this distribution.

        The first time this property is accessed, the relevant information is
        read from the SHARED file in the .dist-info directory.
        rrYutf-8encodingrr	namespace)rrrOrcodecsrrrrr'r))r"rhshared_pathrlinesliner(r<r#r#r$shared_locationss
z&InstalledDistribution.shared_locationsc		Cstj|jd}td||r$dSg}dD].}||}tj||r,|d||fq,|ddD]}|d|qhtj	|d	d
d}|
d|W5QRX|S)
aa
        Write shared location information to the SHARED file in .dist-info.
        :param paths: A dictionary as described in the documentation for
        :meth:`shared_locations`.
        :param dry_run: If True, the action is logged but no file is actually
                        written.
        :return: The path of the file written to.
        rrN)rlibZheadersZscriptsrz%s=%srr#znamespace=%srrr
)rrrOrSrrr)getrrwrite)	r"rrrrr(rnsrr#r#r$write_shared_locationss	z,InstalledDistribution.write_shared_locationscCsF|tkrtd||jft|j}|dkr.set_name_and_version)rrr7r6rDrrc
_get_metadatar,rrr%)r"rrErrDrr#r$r%bs

zEggInfoDistribution.__init__c	sd}ddfdd}d}}|drtj|rtj|d}tj|d}t|dd	}tj|d
}	tj|d}||	}npt|}
t|
	d
d
}t|dd}z,|
	d}|
	d
d}|
d}Wntk
rd}YnXnf|drPtj|rBtj|d
}	||	}tj|d}tj|d}t|dd	}ntd||rl|
||dkr|dk	rtj|rt|d}
|

d}W5QRX|sg}n|}||_|S)NcSsg}|}|D]}|}|dr6td|qt|}|sPtd|q|jr`td|jst||j	qd
dd|jD}|d|j	|fq|S)	zCreate a list of dependencies from a requires.txt file.

            *data*: the contents of a setuptools-produced requires.txt file.
            [z.Unexpected line: quitting requirement scan: %rz#Not recognised as a requirement: %rz4extra requirements in requires.txt are not supportedz, css|]}d|VqdS)rNr#)rcr#r#r$	szQEggInfoDistribution._get_metadata..parse_requires_data..rj)rstriprrSrr
rZconstraintsr)rrO)rreqsrrrYZconsr#r#r$parse_requires_datazs(

z>EggInfoDistribution._get_metadata..parse_requires_datac	sHg}z*t|dd}|}W5QRXWntk
rBYnX|S)zCreate a list of dependencies from a requires.txt file.

            *req_path*: the path to a setuptools-produced requires.txt file.
            rYr)rrrIOError)req_pathrrrr#r$parse_requires_pathsz>EggInfoDistribution._get_metadata..parse_requires_pathrGzEGG-INFOzPKG-INFOr@)rrCzrequires.txtrzEGG-INFO/PKG-INFOutf8rAzEGG-INFO/requires.txtzEGG-INFO/top_level.txtrrFz,path must end with .egg-info or .egg, got %rr)rLrrrrOr		zipimportzipimporterrget_datarrrZadd_requirementsrrrrr)r"rrequiresr	Ztl_pathZtl_datarq	meta_pathrDrZzipfrBrrr#rr$rwsX





z!EggInfoDistribution._get_metadatacCsd|j|j|jfS)Nz!rr!r#r#r$rs
zEggInfoDistribution.__repr__cCsd|j|jfSrrr!r#r#r$rszEggInfoDistribution.__str__cCs`g}tj|jd}tj|r\|D]2\}}}||kr._md5cSst|jSr9)rstatst_size)rr#r#r$_sizesz7EggInfoDistribution.list_installed_files.._sizerrYrrzNon-existent file: %srN)
rrrOrrrrnormpathrSrrLrr))r"rrrrhrrrqr#r#r$rs"
$z(EggInfoDistribution.list_installed_filesFc	cstj|jd}tj|rd}tj|ddd`}|D]T}|}|dkrPd}q6|s6tjtj|j|}||jr6|r|Vq6|Vq6W5QRXdS)	a
        Iterates over the ``installed-files.txt`` entries and returns paths for
        each line if the path is pointing to a file located in the
        ``.egg-info`` directory or one of its subdirectories.

        :parameter absolute: If *absolute* is ``True``, each returned path is
                          transformed into a local absolute path. Otherwise the
                          raw value from ``installed-files.txt`` is returned.
        :type absolute: boolean
        :returns: iterator of paths
        rTrYrrz./FN)	rrrOrrrrrr)r"Zabsoluterskiprrrqr#r#r$rsz'EggInfoDistribution.list_distinfo_filescCst|to|j|jkSr9)r]rrrr#r#r$r.s

zEggInfoDistribution.__eq__)N)F)r-r.r/r0rrr%rrrrrrrrrrr#r#rr$rYsZ&
c@s^eZdZdZddZddZdddZd	d
ZddZdddZ	dddZ
ddZddZdS)DependencyGrapha
    Represents a dependency graph between distributions.

    The dependency relationships are stored in an ``adjacency_list`` that maps
    distributions to a list of ``(other, label)`` tuples where  ``other``
    is a distribution and the edge is labeled with ``label`` (i.e. the version
    specifier, if such was provided). Also, for more efficient traversal, for
    every distribution ``x``, a list of predecessors is kept in
    ``reverse_list[x]``. An edge from distribution ``a`` to
    distribution ``b`` means that ``a`` depends on ``b``. If any missing
    dependencies are found, they are stored in ``missing``, which is a
    dictionary that maps distributions to a list of requirements that were not
    provided by any other distributions.
    cCsi|_i|_i|_dSr9)adjacency_listreverse_listrr!r#r#r$r%IszDependencyGraph.__init__cCsg|j|<g|j|<dS)zAdd the *distribution* to the graph.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        N)rr)r"distributionr#r#r$add_distributionNs
z DependencyGraph.add_distributionNcCs6|j|||f||j|kr2|j||dS)aAdd an edge from distribution *x* to distribution *y* with the given
        *label*.

        :type x: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type y: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type label: ``str`` or ``None``
        N)rr)r)r"xylabelr#r#r$add_edgeXs
zDependencyGraph.add_edgecCs&td|||j|g|dS)a
        Add a missing *requirement* for the given *distribution*.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        :type requirement: ``str``
        z
%s missing %rN)rSrTrr'r))r"rrr#r#r$add_missinggszDependencyGraph.add_missingcCsd|j|jfSrrr*r#r#r$
_repr_distrszDependencyGraph._repr_distrcCs||g}|j|D]h\}}||}|dk	r|d|j|jfq>q|st|dkr|d|d|d|D]}|d	|j|d
q|d|ddS)a9Writes a DOT output for the graph to the provided file *f*.

        If *skip_disconnected* is set to ``True``, then all distributions
        that are not dependent on any other distribution are skipped.

        :type f: has to support ``file``-like operations
        :type skip_disconnected: ``bool``
        zdigraph dependencies {
rNz"%s" -> "%s" [label="%s"]
z
"%s" -> "%s"
zsubgraph disconnected {
zlabel = "Disconnected"
zbgcolor = red
z"%s"rz}
)rritemsrr)r)r"rZskip_disconnectedZdisconnectedr+adjsrrr#r#r$to_dots(	





zDependencyGraph.to_dotcsg}i}|jD]\}}|dd||<qgt|ddD]\}}|sD|||=qDshq|D]\}}fdd|D||<qptdddD|q,|t|fS)aa
        Perform a topological sort of the graph.
        :return: A tuple, the first element of which is a topologically sorted
                 list of distributions, and the second element of which is a
                 list of distributions that cannot be sorted because they have
                 circular dependencies and so form a cycle.
        Ncs g|]\}}|kr||fqSr#r#)rr{rYZ	to_remover#r$rsz4DependencyGraph.topological_sort..zMoving to result: %scSsg|]}d|j|jfqS)rjr)rr{r#r#r$rs)rr'listr)rSrTr$keys)r"rhZalistkr|r#r*r$topological_sorts$
z DependencyGraph.topological_sortcCs2g}|jD]\}}|||qd|S)zRepresentation of the graphr)rr'r)r#rO)r"r&r+r(r#r#r$rszDependencyGraph.__repr__)N)r)T)
r-r.r/r0r%rr r!r"r#r)r.rr#r#r#r$r9s



 rr1cCsVt|}t}i}|D]L}|||jD]6}t|\}}td|||||g||fq*q|D]}|j	|j
B|jB|jB}|D]}	z|
|	}
Wn6tk
rtd|	|	d}|
|}
YnX|
j}d}||kr>||D]N\}}z|
|}
Wntk
rd}
YnX|
r||||	d}q>q|s|||	qqh|S)a6Makes a dependency graph from the given distributions.

    :parameter dists: a list of distributions
    :type dists: list of :class:`distutils2.database.InstalledDistribution` and
                 :class:`distutils2.database.EggInfoDistribution` instances
    :rtype: a :class:`DependencyGraph` instance
    zAdd to provided: %s, %s, %srrFT)rrrrkrrSrTr'r)rrrrrlrrrr(ror r!)distsrCgraphrpr+rqrrcrrrlZmatchedZproviderror#r#r$
make_graphsN




r1cCsv||krtd|jt|}|g}|j|}|rh|}|||j|D]}||krN||qNq.|d|S)zRecursively generate a list of distributions from *dists* that are
    dependent on *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    1given distribution %r is not a member of the listr)rrr1rpopr))r/r+r0Zdeptodor{Zsuccr#r#r$get_dependent_distss


r5cCsn||krtd|jt|}g}|j|}|rj|d}|||j|D]}||krP||qPq,|S)zRecursively generate a list of distributions from *dists* that are
    required by *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    r2r)rrr1rr3r))r/r+r0rr4r{Zpredr#r#r$get_required_distss

r6cKs4|dd}tf|}||_||_|p(d|_t|S)zO
    A convenience method for making a dist given just a name and version.
    summaryzPlaceholder for summary)r3r	rrcr7r)rrckwargsr7rr#r#r$	make_dist2s

r9)r1)7r0Z
__future__rrrrPrZloggingrrNr2rr?rrcompatrrcrrrDr	r
rrutilr
rrrrrr__all__Z	getLoggerr-rSrZCOMMANDS_FILENAMErrMrrrrrrrrUrVrr1r5r6r9r#r#r#r$s`$
s7J]
6site-packages/pip/_vendor/distlib/__pycache__/wheel.cpython-38.opt-1.pyc000064400000061544151464452710022070 0ustar00U

.e@sddlmZddlZddlZddlZddlZddlmZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlZddlZddlZddlmZmZddlmZmZmZmZmZddlmZddlm Z m!Z!m"Z"dd	l#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,dd
l-m.Z.m/Z/e0e1Z2da3e4edr8dZ5n*ej67d
rLdZ5nej6dkr^dZ5ndZ5e8dZ9e9sdej:ddZ9de9Z;e5e9Zdd>ddZ?e8dZ@e@re@7dre@>ddZ@nddZAeAZ@[AeBdejCejDBZEeBdejCejDBZFeBdZGeBd ZHd!ZId"ZJe
jKd#krBd$d%ZLnd&d%ZLGd'd(d(eMZNeNZOGd)d*d*eMZPd+d,ZQeQZR[Qd/d-d.ZSdS)0)unicode_literalsN)message_from_file)__version__DistlibException)	sysconfigZipFilefsdecode	text_typefilter)InstalledDistribution)MetadataMETADATA_FILENAMEWHEEL_METADATA_FILENAME)	FileOperatorconvert_path	CSVReader	CSVWriterCachecached_propertyget_cache_baseread_exportstempdir)NormalizedVersionUnsupportedVersionErrorZpypy_version_infoZppjavaZjyZcliZipcpZpy_version_nodotz%s%spy-_.ZSOABIzcpython-cCsRdtg}tdr|dtdr0|dtddkrH|dd	|S)
NrZPy_DEBUGdZ
WITH_PYMALLOCmZPy_UNICODE_SIZEu)
VER_SUFFIXrget_config_varappendjoin)partsr,=/usr/lib/python3.8/site-packages/pip/_vendor/distlib/wheel.py_derive_abi;s




r.zz
(?P[^-]+)
-(?P\d+[^-]*)
(-(?P\d+[^-]*))?
-(?P\w+\d+(\.\w+\d+)*)
-(?P\w+)
-(?P\w+(\.\w+)*)
\.whl$
z7
(?P[^-]+)
-(?P\d+[^-]*)
(-(?P\d+[^-]*))?$
s
\s*#![^\r\n]*s^(\s*#!("[^"]+"|\S+))\s+(.*)$s#!pythons	#!pythonw/cCs|SNr,or,r,r-]r3cCs|tjdS)Nr/)replaceossepr1r,r,r-r3_r4c@s6eZdZddZddZddZddd	Zd
dZdS)
MountercCsi|_i|_dSr0)
impure_wheelslibsselfr,r,r-__init__cszMounter.__init__cCs||j|<|j|dSr0)r9r:update)r<pathname
extensionsr,r,r-addgs
zMounter.addcCs0|j|}|D]\}}||jkr|j|=qdSr0)r9popr:)r<r?r@kvr,r,r-removeks
zMounter.removeNcCs||jkr|}nd}|Sr0)r:)r<fullnamepathresultr,r,r-find_moduleqs
zMounter.find_modulecCsj|tjkrtj|}nP||jkr,td|t||j|}||_|dd}t|dkrf|d|_	|S)Nzunable to find extension for %sr!rr)
sysmodulesr:ImportErrorimpZload_dynamic
__loader__rsplitlen__package__)r<rFrHr+r,r,r-load_modulexs


zMounter.load_module)N)__name__
__module____qualname__r=rArErIrRr,r,r,r-r8bs

r8c@seZdZdZdZdZd4ddZedd	Zed
dZ	edd
Z
eddZddZ
eddZddZd5ddZddZddZddZd6ddZd d!Zd"d#Zd$d%Zd&d'Zd(d)Zd*d+Zd7d,d-Zd.d/Zd0d1Zd8d2d3ZdS)9Wheelz@
    Class to build and install from Wheel files (PEP 427).
    rrZsha256NFcCs8||_||_d|_tg|_dg|_dg|_t|_	|dkrRd|_
d|_|j|_
nt|}|r|d}|d|_
|dd	d
|_|d|_|j|_
ntj|\}}t|}|std||rtj||_	||_
|d}|d|_
|d|_|d|_|d
d|_|dd|_|dd|_dS)zB
        Initialise an instance using a (valid) filename.
        r&noneanyNZdummyz0.1ZnmZvnr rZbnzInvalid name or filename: %rrr!ZbiZar)signZ
should_verifybuildverPYVERpyverabiarchr6getcwddirnamenameversionfilenameZ	_filenameNAME_VERSION_REmatch	groupdictr5rGsplitFILENAME_RErabspath)r<rdrZverifyr#inforar,r,r-r=sD











zWheel.__init__cCs^|jrd|j}nd}d|j}d|j}d|j}|jdd}d|j|||||fS)zJ
        Build and return a filename from the various components.
        rr&r!r z%s-%s%s-%s-%s-%s.whl)r[r*r]r^r_rcr5rb)r<r[r]r^r_rcr,r,r-rds
zWheel.filenamecCstj|j|j}tj|Sr0)r6rGr*rardisfile)r<rGr,r,r-existsszWheel.existsccs4|jD](}|jD]}|jD]}|||fVqqqdSr0)r]r^r_)r<r]r^r_r,r,r-tagss


z
Wheel.tagscCstj|j|j}d|j|jf}d|}td}t	|d}|
|}|ddd}tdd	|D}|d
krt
tdg}	nt
tg}	d}
|	D]f}zLt||}||,}
||
}t|d}
|
rW5QRWqW5QRXWqtk
rYqXq|
std
d|	W5QRX|
S)N%s-%s%s.dist-infoutf-8r
Wheel-Versionr!rcSsg|]}t|qSr,int.0ir,r,r-
sz"Wheel.metadata..rWZMETADATA)Zfileobjz8Invalid wheel, because metadata is missing: looked in %sz, )r6rGr*rardrbrccodecs	getreaderrget_wheel_metadatarhtuplerr	posixpathopenr
KeyError
ValueError)r<r?name_verinfo_dirwrapperzfwheel_metadatawvfile_versionZfnsrHfnmetadata_filenamebfwfr,r,r-metadatas6


zWheel.metadatac	CsXd|j|jf}d|}t|d}||}td|}t|}W5QRXt|S)NrprqWHEELrr)	rbrcrr*rr{r|rdict)r<rrrrrrmessager,r,r-r}szWheel.get_wheel_metadatac	Cs6tj|j|j}t|d}||}W5QRX|S)Nrs)r6rGr*rardrr})r<r?rrHr,r,r-rlsz
Wheel.infocCst|}|r||}|d|||d}}d|krBt}nt}t|}|rfd|d}nd}||}||}nT|d}|d}	|dks||	krd}
n|||dd	krd	}
nd}
t|
|}|S)
Nspythonw r4

rrs
)	
SHEBANG_RErfendlowerSHEBANG_PYTHONWSHEBANG_PYTHONSHEBANG_DETAIL_REgroupsfind)r<datar#rZshebangZdata_after_shebangZshebang_pythonargsZcrZlfZtermr,r,r-process_shebangs,




zWheel.process_shebangcCsh|dkr|j}ztt|}Wn tk
r<td|YnX||}t|d	d}||fS)NzUnsupported hash algorithm: %r=ascii)
	hash_kindgetattrhashlibAttributeErrorrdigestbase64Zurlsafe_b64encoderstripdecode)r<rrhasherrHr,r,r-get_hash"szWheel.get_hashc	Cs^t|}ttj||}||ddf|t|}|D]}||q@W5QRXdS)Nr&)	listto_posixr6rGrelpathr)sortrZwriterow)r<recordsZrecord_pathbasepwriterrowr,r,r-write_record-s
zWheel.write_recordc
Csg}|\}}tt|j}|D]P\}}	t|	d}
|
}W5QRXd||}tj|	}
|	|||
fqtj
|d}	|||	|ttj
|d}|	||	fdS)Nrbz%s=%sRECORD)
rrrrreadrr6rGgetsizer)r*rr)r<rllibdir
archive_pathsrdistinforraprfrrsizer,r,r-
write_records6szWheel.write_recordsc	CsFt|dtj.}|D]"\}}td|||||qW5QRXdS)NwzWrote %s to %s in wheel)rzipfileZZIP_DEFLATEDloggerdebugwrite)r<r?rrrrr,r,r-	build_zipFszWheel.build_zipc!sp|dkri}ttfdddd}|dkrFd}tg}tg}tg}nd}tg}d	g}d
g}|d||_|d||_|d
||_	|}	d|j
|jf}
d|
}d|
}g}
dD]}|krq|}tj
|rt|D]\}}}|D]}ttj
||}tj
||}ttj
|||}|
||f|dkr|dst|d}|}W5QRX||}t|d}||W5QRXqqq|	}d}t|D]\}}}||krt|D]8\}}t|}|drtj
||}||=qq|D]H}t|drqtj
||}ttj
||}|
||fqqt|}|D]B}|dkr^ttj
||}ttj
||}|
||fq^d|p|jdtd|g}|jD] \}}}|d|||fqtj
|d}t|d}|d|W5QRXttj
|d}|
||f|||f|	|
tj
|j |j!} |"| |
| S) z
        Build a wheel from files in specified paths, and use any specified tags
        when determining the name of the wheel.
        Ncs|kSr0r,r1pathsr,r-r3Tr4zWheel.build..)purelibplatlibrrZfalsetruerXrYr]r^r_rp%s.datarq)rZheadersscriptsr.exerwbz
.dist-info)z.pycz.pyo)rZ	INSTALLERZSHAREDrzWheel-Version: %d.%dzGenerator: distlib %szRoot-Is-Purelib: %sz
Tag: %s-%s-%srr
)#rrIMPVERABIARCHr\getr]r^r_rbrcr6rGisdirwalkr	r*rrr)endswithrrrr	enumeratelistdir
wheel_versionrrorrardr)!r<rrorZlibkeyZis_pureZ
default_pyverZdefault_abiZdefault_archrrdata_dirrrkeyrGrootdirsfilesrrrprrrrrydnrr]r^r_r?r,rr-buildLs




zWheel.buildcCs
|dS)zl
        Determine whether an archive entry should be skipped when verifying
        or installing.
        )r/z/RECORD.jws)r)r<arcnamer,r,r-
skip_entryszWheel.skip_entrycCKsf|j}|d}|dd}|dd}tj|j|j}d|j|jf}	d|	}
d|	}t	|t
}t	|d}
t	|d	}td
}t
|d}||
}||}t|}W5QRX|dd
d}tdd|D}||jkr|r||j||ddkr|d}n|d}i}||8}t|d"}|D]}|d}|||<q8W5QRXW5QRXt	|
d}t	|d}t	|
dd}t|d}d|_tj} g}!t}"|"|_d|_zz\|D]}#|#j}$t|$t r|$}%n
|$!d
}%|"|%rq||%}|dr4t#|#j$|dkr4t%d|%|dr|ddd\}&}'||$}|&}(W5QRX|'|(|&\})}*|*|'krt%d|$|r|%(||frt)*d |%q|%(|o|%+d!}+|%(|r
|%d"d\})},}-tj||,t,|-}.n$|%|
|fkrqtj|t,|%}.|+s
||$}|-||.W5QRX|!.|.|s|drt|.d#4}|&}(|'|(|&\})}/|/|*krt%d$|.W5QRX| r~|.+d%r~z|j/|.|d&}0|!.|0Wn$t0k
rt)j1d'dd(YnXnttj2t,|$}1tj|"|1}2||$}|-||2W5QRXtj|.\}3}1|3|_|3|1}4|4|4|!5|4q|rt)*d)d}5nnd}6|j6d}|d*krzt	|d+}7z||7}t7|}8W5QRXi}6d,D]l}9d-|9}:|:|8kri|6d.|9<};|8|:8D]6}|6d6i}?|>s|?r|dd}@tj>|@s*t?d7|@|_|>@D]*\}:}.zRoot-Is-Purelibrrrstreamrr&r)dry_runTNrsize mismatch for %s=digest mismatch for %szlib_only: skipping %srr/rzdigest mismatch on write for %sz.py)Zhashed_invalidationzByte-compilation failed)exc_infozlib_only: returning Nonez1.0zentry_points.txt)Zconsoleguiz
%s_scriptszwrap_%sz%s:%sz %szAUnable to read legacy script metadata, so cannot generate scriptsr@zpython.commandsz8Unable to read JSON metadata, so cannot generate scriptsZwrap_consoleZwrap_guizValid script path not specifiedz%s = %srlibprefixzinstallation failed.)Grrr6rGr*rardrbrcrrr{r|rrrrhr~rrrrecordrJdont_write_bytecodetempfileZmkdtempZ
source_dirZ
target_dirshutilZrmtreeinfolist
isinstancer
rrstr	file_sizerrr
startswithrrrrZcopy_streamr)Zbyte_compile	ExceptionZwarningbasenameZmakeZset_executable_modeextendrlrvaluesrsuffixflagsjsonloadrritemsrrZwrite_shared_locationsZwrite_installed_filesZ	exceptionZrollback)Cr<rZmakerkwargsrrrZbc_hashed_invalidationr?rrr
metadata_namewheel_metadata_namerecord_namerrbwfrrrrrrrreaderrrZdata_pfxZinfo_pfxZ
script_pfxZfileopZbcZoutfilesworkdirzinfor	u_arcnamekindvaluerr rZ	is_scriptwhererZoutfileZ	newdigestZpycrZworknamer	filenamesZdistZcommandsZepZepdatarrCr"rDsZconsole_scriptsZgui_scriptsZ
script_dirZscriptZoptionsr,r,r-installsT


 















z
Wheel.installcCs4tdkr0tjttdtjdd}t|atS)Nzdylib-cache)	cacher6rGr*rrrJrcr)r<rr,r,r-_get_dylib_cacheszWheel._get_dylib_cachec
Cshtj|j|j}d|j|jf}d|}t|d}t	d}g}t
|d}z||}||}	t
|	}
|}||}tj|j|}
tj|
st|
|
D]\}}tj|
t|}tj|sd}n6t|j}tj|}||}tj|j}||k}|r&|||
|||fqW5QRXWntk
rXYnXW5QRX|S)NrprqZ
EXTENSIONSrrrsT)r6rGr*rardrbrcrr{r|rrrrrZ
prefix_to_dirrrmakedirsrrrnstatst_mtimedatetimeZ
fromtimestampZgetinfoZ	date_timeextractr)r)r<r?rrrrrHrrrr@rrZ
cache_baserbrdestrZ	file_timerlZ
wheel_timer,r,r-_get_extensionss>




zWheel._get_extensionscCst|S)zM
        Determine if a wheel is compatible with the running system.
        )
is_compatibler;r,r,r-rszWheel.is_compatiblecCsdS)zP
        Determine if a wheel is asserted as mountable by its metadata.
        Tr,r;r,r,r-is_mountableszWheel.is_mountablecCstjtj|j|j}|s2d|}t||sJd|}t||t	jkrbt
d|nN|rtt	j|nt	j
d||}|rtt	jkrt	jtt||dS)Nz)Wheel %s not compatible with this Python.z$Wheel %s is marked as not mountable.z%s already in pathr)r6rGrjr*rardrrrrJrrr)insertr_hook	meta_pathrA)r<r)r?msgr@r,r,r-mounts"

zWheel.mountcCsrtjtj|j|j}|tjkr2td|n.rrr/..invalid entry in wheel: %rrrrr)r6rGr*rardrbrcrrr{r|rrrrhr~rrrr
rrrrrrr)r<r?rrrrrrrrrrrrrrrr	rrrrrr
rrr rr,r,r-rks\





zWheel.verifyc
Ksdd}dd}tj|j|j}d|j|jf}d|}t|d}	t}
t	|d|}i}|
D]h}
|
j}t|tr|}n
|
d	}||	krqhd
|krtd|||
|
tj|
t|}|||<qhW5QRX|||\}}||f|}|r|||\}}|r$||kr$||||dkrNtjd
d|
d\}}t|n*tj|shtd|tj||j}t|}tj|
|}||f}|||
|||||dkrt||W5QRX|S)a
        Update the contents of a wheel in a generic way. The modifier should
        be a callable which expects a dictionary argument: its keys are
        archive-entry paths, and its values are absolute filesystem paths
        where the contents the corresponding archive entries can be found. The
        modifier is free to change the contents of the files pointed to, add
        new entries and remove entries, before returning. This method will
        extract the entire contents of the wheel to a temporary location, call
        the modifier, and then use the passed (and possibly updated)
        dictionary to write a new wheel. If ``dest_dir`` is specified, the new
        wheel is written there -- otherwise, the original wheel is overwritten.

        The modifier should return True if it updated the wheel, else False.
        This method returns the same value the modifier returns.
        cSsHd}}d|tf}||kr$d|}||kr@||}t|dj}||fS)Nz%s/%sz%s/PKG-INFOrG)rr
rc)path_maprrcrGrr,r,r-get_versionKsz!Wheel.update..get_versioncSsd}z|t|}|d}|dkr*d|}nTdd||dddD}|dd7<d	|d|dd
d|Df}Wn tk
rtd|YnX|rt|d
}||_|	t
}|j||dtd||dS)Nrrz%s+1cSsg|]}t|qSr,ru)rxrr,r,r-rz]sz8Wheel.update..update_version..rr!rz%s+%scss|]}t|VqdSr0)rrwr,r,r-	`sz7Wheel.update..update_version..z0Cannot update non-compliant (PEP-440) version %rr')rGlegacyzVersion updated from %r to %r)rrrhr*rrrr
rcrrr)rcrGupdatedrDryr+Zmdr+r,r,r-update_versionUs.

 


z$Wheel.update..update_versionrprqrrsrrr%r&Nz.whlz
wheel-update-)rrdirzNot a directory: %r)r6rGr*rardrbrcrrrrrr
rrrrrZmkstempcloserrrrrrZcopyfile)r<ZmodifierZdest_dirrr)r-r?rrrr
rr(rrrrGZoriginal_versionr ZmodifiedZcurrent_versionfdnewpathrrrlr,r,r-r>:s\







zWheel.update)NFF)N)NN)F)N)rSrTrU__doc__rrr=propertyrdrnrorrr}rlrrrrrrrrrrrrr#r$rkr>r,r,r,r-rVs@
)



	

	
hg	"
8rVc	Csxtg}td}ttjddddD]}|d|t|gq$g}tD]*\}}}|	drN||
dddqN|tdkr|
dt|dg}tg}tjd	krtd
t}|r|\}	}}}
t|}|
g}|
dkr|d|
d
kr|d|
dkr |d|
dkr4|d|
dkrH|d|dkr|D]*}d|	|||f}
|
tkrV||
qV|d8}qH|D]0}|D]$}
|dt|df||
fqqt|D]L\}}|dt|fddf|dkr|dt|dfddfqt|D]L\}}|dd|fddf|dkr"|dd|dfddfq"t|S)zG
    Return (pyver, abi, arch) tuples compatible with this Python.
    rrrr&z.abir!rrXdarwinz(\w+)_(\d+)_(\d+)_(\w+)$)i386ppcZfat)r5r6x86_64Zfat3)ppc64r7Zfat64)r5r7intel)r5r7r9r6r8Z	universalz%s_%s_%s_%srYr)r'rangerJversion_infor)r*rrMZget_suffixesrrhrrrrplatformrerfrrv
IMP_PREFIXrset)ZversionsmajorminorZabisrr rHZarchesr#rbr_Zmatchesrfrr^ryrcr,r,r-compatible_tagss`












&
"
"rBcCs\t|tst|}d}|dkr"t}|D]0\}}}||jkr&||jkr&||jkr&d}qXq&|S)NFT)rrVCOMPATIBLE_TAGSr]r^r_)ZwheelrorHZverr^r_r,r,r-rs
r)N)TZ
__future__rrr{rZdistutils.utilZ	distutilsZemailrrrMrZloggingr6rr=rrJrrr&rrcompatrrr	r
rZdatabaserrr
rrutilrrrrrrrrrrcrrZ	getLoggerrSrrhasattrr>r<rr(r'r;r\rZget_platformr5rrr.compile
IGNORECASEVERBOSErirerrrrr7robjectr8r rVrBrCrr,r,r,r-s,


	





#>site-packages/pip/_vendor/distlib/__pycache__/__init__.cpython-38.opt-1.pyc000064400000002004151464452710022505 0ustar00U

.eK@snddlZdZGdddeZzddlmZWn&ek
rRGdddejZYnXeeZ	e	
edS)Nz0.2.9.post0c@seZdZdS)DistlibExceptionN)__name__
__module____qualname__rr@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/__init__.pyrsr)NullHandlerc@s$eZdZddZddZddZdS)rcCsdSNrselfrecordrrrhandlezNullHandler.handlecCsdSr	rr
rrremitrzNullHandler.emitcCs
d|_dSr	)lock)rrrr
createLockrzNullHandler.createLockN)rrrr
rrrrrrrsr)Zlogging__version__	ExceptionrrImportErrorZHandlerZ	getLoggerrZloggerZ
addHandlerrrrrs
site-packages/pip/_vendor/distlib/__pycache__/markers.cpython-38.opt-1.pyc000064400000010510151464452710022413 0ustar00U

.e#@sdZddlZddlZddlZddlZddlmZmZmZddl	m
Z
mZdgZddZ
Gd	d
d
eZddZeZ[eZdd
dZdS)zG
Parser for the environment markers micro-language defined in PEP 508.
N)python_implementationurlparsestring_types)in_venvparse_marker	interpretcCst|tr|sdS|ddkS)NFr'")
isinstancer)or?/usr/lib/python3.8/site-packages/pip/_vendor/distlib/markers.py_is_literalsrc
@sfeZdZdZddddddddddddd	dd
dddddd
ddddZddZdS)	Evaluatorz;
    This class is used to evaluate marker expessions.
    cCs||kSNrxyrrr
$zEvaluator.cCs||kSrrrrrr
r%rcCs||kp||kSrrrrrr
r&rcCs||kSrrrrrr
r'rcCs||kSrrrrrr
r(rcCs||kp||kSrrrrrr
r)rcCs||kSrrrrrr
r*rcCs||kp||kSrrrrrr
r+rcCs|o|Srrrrrr
r,rcCs|p|Srrrrrr
r-rcCs||kSrrrrrr
r.rcCs||kSrrrrrr
r/r)z==z===z~=z!=z>=andorinznot inc	Cst|trB|ddkr$|dd}q||kr8td|||}n|d}||jkr`td||d}|d	}t|drt|d	rtd
|||f|||}|||}|j|||}|S)z
        Evaluate a marker expression returned by the :func:`parse_requirement`
        function in the specified context.
        rr	rzunknown variable: %sopzop not implemented: %slhsrhszinvalid comparison: %s %s %s)r
rSyntaxError
operationsNotImplementedErrorrevaluate)	selfexprcontextresultrZelhsZerhsrrrrr
r"2s"


zEvaluator.evaluateN)__name__
__module____qualname____doc__r r"rrrr
rsrc
Csdd}ttdr(|tjj}tjj}nd}d}||tjttt	t
tttt
t
ddtjd}|S)NcSs<d|j|j|jf}|j}|dkr8||dt|j7}|S)Nz%s.%s.%sfinalr)majorminormicroreleaselevelstrserial)infoversionZkindrrr
format_full_versionNs
z,default_context..format_full_versionimplementation0)implementation_nameimplementation_versionZos_nameZplatform_machineZplatform_python_implementationZplatform_releaseZplatform_systemZplatform_versionZplatform_in_venvZpython_full_versionpython_versionZsys_platform)hasattrsysr5r3nameosplatformmachinerreleasesystemr0rr;)r4r:r9r&rrr
default_contextMs(

rDc
Cszt|\}}Wn2tk
rB}ztd||fW5d}~XYnX|rd|ddkrdtd||ftt}|rz||t||S)z
    Interpret a marker and return a result depending on environment.

    :param marker: The marker to interpret.
    :type marker: str
    :param execution_context: The context used for name lookup.
    :type execution_context: mapping
    z)Unable to interpret marker syntax: %s: %sNr#z*unexpected trailing data in marker: %s: %s)r	ExceptionrdictDEFAULT_CONTEXTupdate	evaluatorr")ZmarkerZexecution_contextr$rester%rrr
rqs	"
)N)r*r?r=r@recompatrrrutilrr__all__robjectrrDrHrJrrrrr
s/site-packages/pip/_vendor/distlib/__pycache__/compat.cpython-38.opt-1.pyc000064400000076604151464452710022252 0ustar00U

.e@sBddlmZddlZddlZddlZzddlZWnek
rHdZYnXejddkr~ddlmZe	fZ
eZddl
mZddlZddlZddlmZddlmZmZmZmZmZdd	lmZmZmZm Z m!Z!m"Z"m#Z#d
dZddl$Z$ddl$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-er&dd
l$m.Z.ddl/Z/ddl0Z0ddl1Z2ddl3m3Z3ddl4Z4e5Z5ddl6m7Z8ddl6m9Z:da;ddZ<nddl=mZe>fZ
e>Zddl=m?ZddlZddlZddlZddl@mZmZmZmd?Zod@dAZiYnXzddBlpmqZqWn"ek
rddBlrmqZqYnXejddCdDkr:e3jsZsnddElpmsZszddFl`mtZtWndek
rddGl`muZuzddHlvmwZxWn ek
rdedJdKZxYnXGdLdMdMeuZtYnXzddNlymzZzWnHek
rzddNl{mzZzWn ek
rdfdOdPZzYnXYnXzddQl`m|Z|Wnek
rzddRl}m~ZWn"ek
rjddRlm~ZYnXzddSlmZmZmZWnek
rYnXGdTdUdUeZ|YnXzddVlmZmZWnvek
r<emdWejZdXdYZGdZd[d[eZdgd\d]ZGd^d_d_eZGd`dadaeZGdbdcdceQZYnXdS)h)absolute_importN)StringIO)FileType)shutil)urlparse
urlunparseurljoinurlsplit
urlunsplit)urlretrievequoteunquoteurl2pathnamepathname2urlContentTooShortError	splittypecCst|tr|d}t|S)Nutf-8)
isinstanceunicodeencode_quote)sr>/usr/lib/python3.8/site-packages/pip/_vendor/distlib/compat.pyrs

r)	RequesturlopenURLError	HTTPErrorHTTPBasicAuthHandlerHTTPPasswordMgrHTTPHandlerHTTPRedirectHandlerbuild_opener)HTTPSHandler)
HTMLParser)ifilter)ifilterfalsecCs<tdkrddl}|dat|}|r4|ddSd|fS)zJsplituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'.Nrz^(.*)@(.*)$r)	_userprogrecompilematchgroup)Zhostr+r-rrr	splituser4s

r/)
TextIOWrapper)	rr	r
r/rrrrr)
rr
rrrr r!r"r#r$)rrr)filterfalse)match_hostnameCertificateErrorc@seZdZdS)r3N)__name__
__module____qualname__rrrrr3^sr3c
Csg}|sdS|d}|d|dd}}|d}||krNtdt||sb||kS|dkrv|dn>|d	s|d	r|t|n|t|	d
d|D]}|t|qt
dd
|dtj}	|	
|S)zpMatching according to RFC 6125, section 6.4.3

        http://tools.ietf.org/html/rfc6125#section-6.4.3
        F.rrN*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)splitcountr3reprlowerappend
startswithr+escapereplacer,join
IGNORECASEr-)
ZdnhostnameZ
max_wildcardsZpatspartsZleftmostZ	remainderZ	wildcardsfragZpatrrr_dnsname_matchbs*


rFcCs|stdg}|dd}|D]*\}}|dkr t||r@dS||q |s|ddD]6}|D],\}}|dkrdt||rdS||qdq\t|dkrtd	|d
tt|fn*t|dkrtd||dfntd
dS)a=Verify that *cert* (in decoded format as returned by
        SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
        rules are followed, but IP addresses are not accepted for *hostname*.

        CertificateError is raised on failure. On success, the function
        returns nothing.
        ztempty or no certificate, match_hostname needs a SSL socket or SSL context with either CERT_OPTIONAL or CERT_REQUIREDZsubjectAltNamerZDNSNZsubjectZ
commonNamerz&hostname %r doesn't match either of %s, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found)	
ValueErrorgetrFr=lenr3rAmapr;)ZcertrCZdnsnamesZsankeyvaluesubrrrr2s2


r2)SimpleNamespacec@seZdZdZddZdS)	ContainerzR
        A generic container for when multiple values need to be returned
        cKs|j|dSN__dict__update)selfkwargsrrr__init__szContainer.__init__N)r4r5r6__doc__rWrrrrrPsrP)whichcs"dd}tjr&||r"SdS|dkr>tjdtj}|sFdS|tj}tj	dkrtj
|krt|dtj
tjddtj}tfd	d
|Drg}q‡fdd|D}ng}t
}|D]P}tj|}||kr|||D](}	tj||	}
||
|r|
SqqdS)
aKGiven a command, mode, and a PATH string, return the path which
        conforms to the given mode on the PATH, or None if there is no such
        file.

        `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result
        of os.environ.get("PATH"), or can be overridden with a custom search
        path.

        cSs&tj|o$t||o$tj|SrQ)ospathexistsaccessisdir)fnmoderrr
_access_checkszwhich.._access_checkNPATHZwin32rZPATHEXTc3s |]}|VqdSrQ)r<endswith.0extcmdrr	szwhich..csg|]}|qSrrrerhrr
szwhich..)rZr[dirnameenvironrIdefpathr9pathsepsysplatformcurdirinsertanysetnormcaseaddrA)rir`r[raZpathextfilesseendirZnormdirZthefilenamerrhrrYs8





rY)ZipFile	__enter__)
ZipExtFilec@s$eZdZddZddZddZdS)r~cCs|j|jdSrQrR)rUbaserrrrWszZipExtFile.__init__cCs|SrQrrUrrrr}szZipExtFile.__enter__cGs|dSrQcloserUexc_inforrr__exit__szZipExtFile.__exit__N)r4r5r6rWr}rrrrrr~sr~c@s$eZdZddZddZddZdS)r|cCs|SrQrrrrrr}"szZipFile.__enter__cGs|dSrQrrrrrr%szZipFile.__exit__cOstj|f||}t|SrQ)BaseZipFileopenr~)rUargsrVrrrrr)szZipFile.openN)r4r5r6r}rrrrrrr|!sr|)python_implementationcCs0dtjkrdStjdkrdStjdr,dSdS)z6Return a string identifying the Python implementation.ZPyPyjavaZJythonZ
IronPythonZCPython)rpversionrZr{r>rrrrr0s

r)	sysconfig)CallablecCs
t|tSrQ)rr)objrrrcallableDsrrmbcsstrictsurrogateescapecCs:t|tr|St|tr$|ttStdt|jdSNzexpect bytes or str, not %s)	rbytes	text_typer_fsencoding	_fserrors	TypeErrortyper4filenamerrrfsencodeXs

rcCs:t|tr|St|tr$|ttStdt|jdSr)	rrrdecoderrrrr4rrrrfsdecodeas

r)detect_encoding)BOM_UTF8lookupzcoding[:=]\s*([-\w.]+)cCsH|dddd}|dks*|dr.dS|dks@|drDd	S|S)
z(Imitates get_normal_name in tokenizer.c.N_-rzutf-8-)zlatin-1
iso-8859-1ziso-latin-1)zlatin-1-ziso-8859-1-ziso-latin-1-r)r<r@r>)orig_encencrrr_get_normal_namersrcszjjWntk
r$dYnXdd}d}fdd}fdd}|}|trpd|d	d}d
}|s||gfS||}|r||gfS|}|s||gfS||}|r|||gfS|||gfS)a?
        The detect_encoding() function is used to detect the encoding that should
        be used to decode a Python source file.  It requires one argument, readline,
        in the same way as the tokenize() generator.

        It will call readline a maximum of twice, and return the encoding used
        (as a string) and a list of any lines (left as bytes) it has read in.

        It detects the encoding from the presence of a utf-8 bom or an encoding
        cookie as specified in pep-0263.  If both a bom and a cookie are present,
        but disagree, a SyntaxError will be raised.  If the encoding cookie is an
        invalid charset, raise a SyntaxError.  Note that if a utf-8 bom is found,
        'utf-8-sig' is returned.

        If no encoding is specified, then the default of 'utf-8' will be returned.
        NFrcs$zWStk
rYdSXdS)N)
StopIterationr)readlinerrread_or_stopsz%detect_encoding..read_or_stopcsz|d}Wn4tk
rBd}dk	r6d|}t|YnXt|}|sVdSt|d}zt|}Wn:tk
rdkrd|}nd|}t|YnXr|j	dkr؈dkrd}n
d}t||d	7}|S)
Nrz'invalid or missing encoding declarationz{} for {!r}rzunknown encoding: zunknown encoding for {!r}: {}zencoding problem: utf-8z encoding problem for {!r}: utf-8z-sig)
rUnicodeDecodeErrorformatSyntaxError	cookie_refindallrrLookupErrorr{)lineline_stringmsgZmatchesencodingcodec)	bom_foundrrrfind_cookies8



z$detect_encoding..find_cookieTrz	utf-8-sig)__self__r{AttributeErrorr>r)rrdefaultrrfirstsecondr)rrrrr}s4
&


r)r?r))r)unescape)ChainMap)MutableMapping)recursive_repr...csfdd}|S)zm
            Decorator to make a repr function return fillvalue for a recursive
            call
            csLtfdd}td|_td|_td|_tdi|_|S)Nc	sBt|tf}|krS|z|}W5|X|SrQ)id	get_identrwdiscard)rUrLresult)	fillvaluerepr_running
user_functionrrwrappers
z=_recursive_repr..decorating_function..wrapperr5rXr4__annotations__)rugetattrr5rXr4r)rrr)rrrdecorating_functionsz,_recursive_repr..decorating_functionr)rrrrr_recursive_reprsrc@seZdZdZddZddZddZd'd	d
ZddZd
dZ	ddZ
ddZeddZ
eddZddZeZddZeddZddZdd Zd!d"Zd#d$Zd%d&ZdS)(ra A ChainMap groups multiple dicts (or other mappings) together
        to create a single, updateable view.

        The underlying mappings are stored in a list.  That list is public and can
        accessed or updated using the *maps* attribute.  There is no other state.

        Lookups search the underlying mappings successively until a key is found.
        In contrast, writes, updates, and deletions only operate on the first
        mapping.

        cGst|pig|_dS)zInitialize a ChainMap by setting *maps* to the given mappings.
            If no mappings are provided, a single empty dictionary is used.

            N)listmaps)rUrrrrrWszChainMap.__init__cCst|dSrQ)KeyErrorrUrLrrr__missing__szChainMap.__missing__c	Cs:|jD](}z||WStk
r,YqXq||SrQ)rrr)rUrLmappingrrr__getitem__s
zChainMap.__getitem__NcCs||kr||S|SrQrrUrLrrrrrI%szChainMap.getcCsttj|jSrQ)rJruunionrrrrr__len__(szChainMap.__len__cCsttj|jSrQ)iterrurrrrrr__iter__+szChainMap.__iter__cstfdd|jDS)Nc3s|]}|kVqdSrQr)rfmrLrrrj/sz(ChainMap.__contains__..rtrrrrr__contains__.szChainMap.__contains__cCs
t|jSrQrrrrr__bool__1szChainMap.__bool__cCsd|dtt|jS)Nz{0.__class__.__name__}({1})rG)rrArKr;rrrrr__repr__4szChainMap.__repr__cGs|tj|f|S)z?Create a ChainMap with a single dict created from the iterable.)dictfromkeys)clsiterablerrrrr9szChainMap.fromkeyscCs$|j|jdf|jddS)zHNew ChainMap or subclass with a new copy of maps[0] and refs to maps[1:]rrN)	__class__rcopyrrrrr>sz
ChainMap.copycCs|jif|jS)z;New ChainMap with a new dict followed by all previous maps.rrrrrr	new_childDszChainMap.new_childcCs|j|jddS)zNew ChainMap from maps[1:].rNrrrrrparentsHszChainMap.parentscCs||jd|<dS)Nr)r)rUrLrMrrr__setitem__MszChainMap.__setitem__cCs8z|jd|=Wn"tk
r2td|YnXdS)Nr(Key not found in the first mapping: {!r})rrrrrrr__delitem__PszChainMap.__delitem__cCs2z|jdWStk
r,tdYnXdS)zPRemove and return an item pair from maps[0]. Raise KeyError is maps[0] is empty.rz#No keys found in the first mapping.N)rpopitemrrrrrrVszChainMap.popitemcGs@z|jdj|f|WStk
r:td|YnXdS)zWRemove *key* from maps[0] and return its value. Raise KeyError if *key* not in maps[0].rrN)rpoprr)rUrLrrrrr]szChainMap.popcCs|jddS)z'Clear maps[0], leaving maps[1:] intact.rN)rclearrrrrrdszChainMap.clear)N)r4r5r6rXrWrrrIrrrrrrclassmethodrr__copy__rpropertyrrrrrrrrrrrs.



r)cache_from_sourcecCs"|dkrd}|rd}nd}||S)NFcor)r[debug_overridesuffixrrrrnsr)OrderedDict)r)KeysView
ValuesView	ItemsViewc@seZdZdZddZejfddZejfddZdd	Zd
dZ	dd
Z
d6ddZddZddZ
ddZddZddZddZddZeZeZefdd Zd7d"d#Zd8d$d%Zd&d'Zd(d)Zed9d*d+Zd,d-Zd.d/Zd0d1Zd2d3Z d4d5Z!d!S):rz)Dictionary that remembers insertion ordercOsnt|dkrtdt|z
|jWn6tk
r\g|_}||dg|dd<i|_YnX|j||dS)zInitialize an ordered dictionary.  Signature is the same as for
            regular dictionaries, but keyword arguments are not recommended
            because their insertion order is arbitrary.

            rz$expected at most 1 arguments, got %dN)rJr_OrderedDict__rootr_OrderedDict__map_OrderedDict__update)rUrkwdsrootrrrrWs

zOrderedDict.__init__cCsF||kr6|j}|d}|||g|d<|d<|j|<||||dS)z!od.__setitem__(i, y) <==> od[i]=yrrN)rr)rUrLrMZdict_setitemrlastrrrrs
 zOrderedDict.__setitem__cCs0||||j|\}}}||d<||d<dS)z od.__delitem__(y) <==> del od[y]rrN)rr)rUrLZdict_delitem	link_prev	link_nextrrrrs
zOrderedDict.__delitem__ccs.|j}|d}||k	r*|dV|d}qdS)zod.__iter__() <==> iter(od)rr)NrrUrZcurrrrrrs

zOrderedDict.__iter__ccs.|j}|d}||k	r*|dV|d}qdS)z#od.__reversed__() <==> reversed(od)rr)Nrrrrr__reversed__s

zOrderedDict.__reversed__cCsdz@|jD]}|dd=q|j}||dg|dd<|jWntk
rTYnXt|dS)z.od.clear() -> None.  Remove all items from od.N)r
itervaluesrrrr)rUZnoderrrrrszOrderedDict.clearTcCs||std|j}|r8|d}|d}||d<||d<n |d}|d}||d<||d<|d}|j|=t||}||fS)zod.popitem() -> (k, v), return and remove a (key, value) pair.
            Pairs are returned in LIFO order if last is true or FIFO order if false.

            zdictionary is emptyrrr))rrrrr)rUrrlinkrrrLrMrrrrs 
zOrderedDict.popitemcCst|S)zod.keys() -> list of keys in od)rrrrrkeysszOrderedDict.keyscsfddDS)z#od.values() -> list of values in odcsg|]}|qSrrrfrLrrrrksz&OrderedDict.values..rrrrrvaluesszOrderedDict.valuescsfddDS)z.od.items() -> list of (key, value) pairs in odcsg|]}||fqSrrr
rrrrksz%OrderedDict.items..rrrrritemsszOrderedDict.itemscCst|S)z0od.iterkeys() -> an iterator over the keys in od)rrrrriterkeysszOrderedDict.iterkeysccs|D]}||VqdS)z2od.itervalues -> an iterator over the values in odNrrUkrrrr
szOrderedDict.itervaluesccs|D]}|||fVqdS)z=od.iteritems -> an iterator over the (key, value) items in odNrrrrr	iteritemsszOrderedDict.iteritemscOst|dkr tdt|fn|s,td|d}d}t|dkrL|d}t|trn|D]}||||<qZn None.  Update od from dict/iterable E and F.

            If E is a dict instance, does:           for k in E: od[k] = E[k]
            If E has a .keys() method, does:         for k in E.keys(): od[k] = E[k]
            Or if E is an iterable of items, does:   for k, v in E: od[k] = v
            In either case, this is followed by:     for k, v in F.items(): od[k] = v

            r)z8update() takes at most 2 positional arguments (%d given)z,update() takes at least 1 argument (0 given)rrrrN)rJrrrhasattrrr)rrrUotherrLrMrrrrTs(	


zOrderedDict.updatecCs0||kr||}||=|S||jkr,t||S)zod.pop(k[,d]) -> v, remove specified key and return the corresponding value.
            If key is not found, d is returned if given, otherwise KeyError is raised.

            )_OrderedDict__markerr)rUrLrrrrrr*s
zOrderedDict.popNcCs||kr||S|||<|S)zDod.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in odrrrrr
setdefault7szOrderedDict.setdefaultcCsf|si}t|tf}||kr"dSd||<z.|sBd|jjfWSd|jj|fWS||=XdS)zod.__repr__() <==> repr(od)rrz%s()z%s(%r)N)r
_get_identrr4r)rUZ
_repr_runningZcall_keyrrrr>szOrderedDict.__repr__csXfddD}t}ttD]}||dq(|rLj|f|fSj|ffS)z%Return state information for picklingcsg|]}||gqSrrrfrrrrrkNsz*OrderedDict.__reduce__..N)varsrrrr)rUrZ	inst_dictrrrr
__reduce__LszOrderedDict.__reduce__cCs
||S)z!od.copy() -> a shallow copy of od)rrrrrrVszOrderedDict.copycCs|}|D]}|||<q
|S)zOD.fromkeys(S[, v]) -> New ordered dictionary with keys from S
            and values equal to v (which defaults to None).

            r)rrrMdrLrrrrZs
zOrderedDict.fromkeyscCs6t|tr*t|t|ko(||kSt||S)zod.__eq__(y) <==> od==y.  Comparison to another OD is order-sensitive
            while comparison to a regular mapping is order-insensitive.

            )rrrJrr__eq__rUrrrrres
 zOrderedDict.__eq__cCs
||kSrQrrrrr__ne__nszOrderedDict.__ne__cCst|S)z@od.viewkeys() -> a set-like object providing a view on od's keys)rrrrrviewkeyssszOrderedDict.viewkeyscCst|S)z an object providing a view on od's values)rrrrr
viewvalueswszOrderedDict.viewvaluescCst|S)zBod.viewitems() -> a set-like object providing a view on od's items)rrrrr	viewitems{szOrderedDict.viewitems)T)N)N)N)"r4r5r6rXrWrrrrr	rrrrrrr
rrTrobjectrrrrrrrrrrr r!r"rrrrrs:
	




	r)BaseConfiguratorvalid_identz^[a-z_][a-z0-9_]*$cCst|}|std|dS)Nz!Not a valid Python identifier: %rT)
IDENTIFIERr-rH)rrrrrr%s
r%c@s"eZdZdZddZdddZdS)ConvertingDictz A converting dictionary wrapper.cCsJt||}|j|}||k	rF|||<t|tttfkrF||_||_	|SrQ)
rrconfiguratorconvertrr'ConvertingListConvertingTupleparentrLrUrLrMrrrrrs
zConvertingDict.__getitem__NcCsLt|||}|j|}||k	rH|||<t|tttfkrH||_||_	|SrQ)
rrIr(r)rr'r*r+r,rLrUrLrrMrrrrrIs
zConvertingDict.get)N)r4r5r6rXrrIrrrrr'sr'cCsDt|||}|j|}||k	r@t|tttfkr@||_||_	|SrQ)
rrr(r)rr'r*r+r,rLr.rrrrs
rc@s"eZdZdZddZdddZdS)	r*zA converting list wrapper.cCsJt||}|j|}||k	rF|||<t|tttfkrF||_||_	|SrQ)
rrr(r)rr'r*r+r,rLr-rrrrs
zConvertingList.__getitem__cCs<t||}|j|}||k	r8t|tttfkr8||_|SrQ)	rrr(r)rr'r*r+r,)rUidxrMrrrrrs
zConvertingList.popN)r/)r4r5r6rXrrrrrrr*sr*c@seZdZdZddZdS)r+zA converting tuple wrapper.cCsBt||}|j|}||k	r>t|tttfkr>||_||_	|SrQ)
tuplerr(r)rr'r*r+r,rLr-rrrrs
zConvertingTuple.__getitem__N)r4r5r6rXrrrrrr+sr+c@seZdZdZedZedZedZedZ	edZ
ddd	Zee
Zd
dZdd
ZddZddZddZddZddZdS)r$zQ
        The configurator base class which defines some useful defaults.
        z%^(?P[a-z]+)://(?P.*)$z^\s*(\w+)\s*z^\.\s*(\w+)\s*z^\[\s*(\w+)\s*\]\s*z^\d+$ext_convertcfg_convert)rgZcfgcCst||_||j_dSrQ)r'configr()rUr4rrrrWs
zBaseConfigurator.__init__c		Cs|d}|d}z^||}|D]H}|d|7}zt||}Wq$tk
rj||t||}Yq$Xq$|WStk
rtdd\}}td||f}|||_	|_
|YnXdS)zl
            Resolve strings to objects using standard import and attribute
            syntax.
            r7rrNzCannot resolve %r: %s)r9rimporterrrImportErrorrprrH	__cause__
__traceback__)	rUrr{ZusedfoundrEetbvrrrresolves"



zBaseConfigurator.resolvecCs
||S)z*Default converter for the ext:// protocol.)r=rUrMrrrr2szBaseConfigurator.ext_convertcCs|}|j|}|dkr&td|n||d}|j|d}|r|j|}|rn||d}nd|j|}|r|d}|j|s||}n2zt	|}||}Wnt
k
r||}YnX|r||d}qHtd||fqH|S)z*Default converter for the cfg:// protocol.NzUnable to convert %rrzUnable to convert %r at %r)WORD_PATTERNr-rHendr4groupsDOT_PATTERN
INDEX_PATTERN
DIGIT_PATTERNintr)rUrMrestrrr0nrrrr3s4
zBaseConfigurator.cfg_convertcCst|ts$t|tr$t|}||_nt|tsHt|trHt|}||_nzt|tslt|trlt|}||_nVt|tr|j	
|}|r|}|d}|j
|d}|r|d}t||}||}|S)z
            Convert values to an appropriate type. dicts, lists and tuples are
            replaced by their converting alternatives. Strings are checked to
            see if they have a conversion format and are converted if they do.
            prefixNr)rr'rr(r*rr+r1string_typesCONVERT_PATTERNr-	groupdictvalue_convertersrIr)rUrMrrrHZ	converterrrrrr)2s,


zBaseConfigurator.convertcsnd}t|s||}dd}tfddD}|f|}|rj|D]\}}t|||qT|S)z1Configure an object with a user-supplied factory.z()r7Ncs g|]}t|r||fqSr)r%rr4rrrkUsz5BaseConfigurator.configure_custom..)rrr=rrsetattr)rUr4rZpropsrVrr{rMrrMrconfigure_customNs


z!BaseConfigurator.configure_customcCst|trt|}|S)z0Utility function which converts lists to tuples.)rrr1r>rrras_tuple\s
zBaseConfigurator.as_tupleN)r4r5r6rXr+r,rJr?rBrCrDrLstaticmethod
__import__r5rWr=r2r3r)rOrPrrrrr$s"




"r$)r)r)N)N)Z
__future__rrZr+rpZsslr6version_inforZ
basestringrIrrtypesrZ	file_typeZ__builtin__builtinsZConfigParserZconfigparserZ	_backportrrr	r
rrZurllibr
rrrrrrrZurllib2rrrrr r!r"r#r$r%ZhttplibZ	xmlrpclibZQueueZqueuer&ZhtmlentitydefsZ	raw_input	itertoolsr'filterr(r1r*r/iostrr0Zurllib.parseZurllib.requestZurllib.errorZhttp.clientZclientZrequestZ
xmlrpc.clientZhtml.parserZ
html.entitiesZentitiesinputr2r3rHrFrOrPr#rYF_OKX_OKZzipfiler|rrr~ZBaseZipExtFilerqrrr	NameErrorcollectionsrrrrgetfilesystemencodingrrtokenizercodecsrrr,rrZhtmlr?Zcgirrrreprlibrrimportlib.utilrZimprthreadrrZdummy_threadZ_abcollrrrrZlogging.configr$r%Ir&r'rrr*r1r+rrrrs,
$,,0
2+A


		
[
b
w

site-packages/pip/_vendor/distlib/__pycache__/util.cpython-38.opt-1.pyc000064400000135707151464452710021744 0ustar00U

.e@sddlZddlmZddlZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZzddlZWnek
rdZYnXddlZddlZddlZddlZddlZzddlZWnek
rddlZYnXddlZddlmZddlmZmZmZmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0e
1e2Z3e
4dZ5e
4dZ6e
4d	Z7e
4d
Z8e
4dZ9e
4dZ:e
4d
Z;e
4dZddZ?ddZ@ddZAdddZBddZCddZDdd ZEejFd!d"ZGejFd#d$ZHejFdd&d'ZIGd(d)d)eJZKd*d+ZLGd,d-d-eJZMd.d/ZNGd0d1d1eJZOe
4d2e
jPZQd3d4ZRdd5d6ZSd7d8ZTd9d:ZUd;d<ZVd=d>ZWd?d@ZXe
4dAe
jYZZe
4dBZ[ddCdDZ\e
4dEZ]dFdGZ^dHdIZ_dJdKZ`dLZadMdNZbdOdPZcGdQdRdReJZdGdSdTdTeJZeGdUdVdVeJZfdWZgddYdZZhd[d\Zid]ZjGd^d_d_eJZke
4d`Zle
4daZme
4dbZndcddZdedfZoerddglmpZqmrZrmsZsGdhdidie$jtZtGdjdkdkeqZpGdldmdmepe'ZuejvddnZwewdokrGdpdqdqe$jxZxerGdrdsdse$jyZyGdtdudue%jzZzerBGdvdwdwe%j{Z{Gdxdydye%j|Z|dzd{Z}Gd|d}d}eJZ~Gd~dde~ZGddde~ZGddde(ZGdddeJZddZdS)N)deque)iglob)DistlibException)string_types	text_typeshutil	raw_inputStringIOcache_from_sourceurlopenurljoinhttplib	xmlrpclib	splittypeHTTPHandlerBaseConfiguratorvalid_ident	ContainerconfigparserURLErrorZipFilefsdecodeunquoteurlparsez^([\w\.-]+)\s*z^([\w\.*+-]+)\s*z^(<=?|>=?|={2,3}|[~!]=)\s*z*^((<=?)|(>=?)|={2,3}|[~!]=|in|not\s+in)\s*z^or\b\s*z	^and\b\s*z(\S+)\s*z(([\s\w\.{}()*+#:;,/?!~`@$%^&=|<>\[\]-]+)cs6ddfddfddfdd|S)	ae
    Parse a marker string and return a dictionary containing a marker expression.

    The dictionary will contain keys "op", "lhs" and "rhs" for non-terminals in
    the expression grammar, or strings. A string contained in quotes is to be
    interpreted as a literal string, and a string not contained in quotes is a
    variable (such as os_name).
    cSs.t|}|r,|d}||d}n|s:tdn|d}|dkrVtd|d|d}|dd}|g}|r|d|krqqt|d|kr|||dd}qtt|}|std|||d||d}qtd|}td|||d|}|dd	}||fS)	Nrzunexpected end of inputz'"zinvalid expression: %srzerror in string literal: %szunterminated string: %s)

IDENTIFIERmatchgroupsendSyntaxErrorreplaceappendSTRING_CHUNKjoinlstrip)	remainingmresultqZoqpartssr,.marker_varcs|rR|ddkrR|dd\}}|ddkr@td||dd}nZ|\}}|rt|}|srq|d}||d}|\}}|||d}q^|}||fS)Nr(r)unterminated parenthesis: %soplhsrhs)r%r 	MARKER_OPrrr)r&r(r4r'r3r5)markerr.r,r-marker_expres 
z!parse_marker..marker_exprcsR|\}}|rJt|}|s qJ||d}|\}}d||d}q||fS)Nandr2)ANDrrr&r4r'r5)r8r,r-
marker_andxs
z parse_marker..marker_andcsR|\}}|rJt|}|s qJ||d}|\}}d||d}q||fS)Norr2)ORrrr;)r<r,r-r7s
zparse_marker..markerr,)Z
marker_stringr,)r7r<r8r.r-parse_marker8s
	$r?cCs0|}|r|drdSt|}|s4td||d}||d}d}}}}|r:|ddkr:|dd}|dkrtd||d|}	||dd}g}|	r0t|	}|std	|	|	|d|	|d}	|	sq0|	dd
krtd|	|	dd}	q|s:d}|r|ddkr|dd}t
|}|sztd
||d}t|}
|
jr|
j
std|||d}ndd}|ddkr||\}}n|dd}|dkrtd||d|}	||dd}t|	r@||	\}}nXt|	}|s\td|	|d}
|	|d}	|	rtd|	d|
fg}|r|ddkrtd||dd}t|\}}|r|ddkrtd||s|}nd|ddd|Df}t||||||dS)z
    Parse a requirement passed in as a string. Return a Container
    whose attributes contain the various parts of the requirement.
    #Nzname expected: %sr[]rzunterminated extra: %szmalformed extra: %s,zcomma expected in extras: %s@zinvalid URI: %szInvalid URL: %scSst|}d}|rg}|d}||d}t|}|sLtd||d}|||f||d}|r|ddkrq|dd}t|}|std|q|sd}||fS)z|
                Return a list of operator, version tuples if any are
                specified, else None.
                Nrzinvalid version: %srCrinvalid constraint: %s)
COMPARE_OPrrrVERSION_IDENTIFIERr r"r%)Z
ver_remainingr'versionsr3vr,r,r-get_versionss*


z'parse_requirement..get_versionsr/r0r1rEz~=;zinvalid requirement: %szunexpected trailing data: %s%s %s, cSsg|]}d|qS)rLr,).0Zconr,r,r-
sz%parse_requirement..)nameextrasZconstraintsr7urlZrequirement)strip
startswithrrr rrfindr%r"	NON_SPACErschemenetlocrFrGr?r$r)reqr&r'ZdistnamerQZ	mark_exprrHuriir+trJ_rIZrsr,r,r-parse_requirements





r^cCsdd}i}|D]\}}}tj||}t|D]p}tj||}	t|	D]T}
|||
}|dkrn||dqJ|||
}|tjjdd}
|
d|||<qJq0q|S)z%Find destinations for resources filescSs6|tjjd}|tjjd}|t|ddSN/)r!ospathseplenr%)rootrbr,r,r-get_rel_path
sz)get_resources_dests..get_rel_pathNr`)rarbr$rpopr!rcrstrip)Zresources_rootZrulesrfZdestinationsbasesuffixdestprefixZabs_baseZabs_globZabs_pathZ
resource_fileZrel_pathZrel_destr,r,r-get_resources_dests
s

rmcCs(ttdrd}ntjttdtjk}|S)NZreal_prefixTbase_prefix)hasattrsysrlgetattrr(r,r,r-in_venv$s
rscCs$tjtj}t|ts t|}|SN)rarbnormcaserp
executable
isinstancerrrrr,r,r-get_executable.s

rxcCsN|}t|}|}|s|r|}|r|d}||kr6qJ|rd|||f}q|S)Nrz	%c: %s
%s)r	lower)promptZ
allowed_charsZerror_promptdefaultpr+cr,r,r-proceed>sr~cCs8t|tr|}i}|D]}||kr||||<q|Srt)rwrsplit)dkeysr(keyr,r,r-extract_by_keyNs
rcCs`tjddkrtd|}|}t|}z`t|}|ddd}|D]6\}}|D]$\}}d||f}t	|}	|	||<q`qP|WSt
k
r|ddYnXdd	}
t
}z|
||Wn<tjk
r|t|}t|}|
||YnXi}|D]D}i||<}||D]&\}
}d|
|f}t	|}	|	||
<q0q|S)
Nrutf-8
extensionszpython.exportsexportsz%s = %scSs$t|dr||n
||dS)N	read_file)rorZreadfp)cpstreamr,r,r-read_streamks
z!read_exports..read_stream)rpversion_infocodecs	getreaderreadr
jsonloaditemsget_export_entry	ExceptionseekrConfigParserZMissingSectionHeaderErrorclosetextwrapdedentZsections)rdataZjdatar(groupentrieskrIr+entryrrrrPvaluer,r,r-read_exportsWs@

rcCstjddkrtd|}t}|D]l\}}|||D]P}|j	dkr\|j
}nd|j
|j	f}|jrd|d|jf}|
||j|qFq,||dS)Nrrrz%s:%sz%s [%s]rM)rprr	getwriterrrrZadd_sectionvaluesrjrlflagsr$setrPwrite)rrrrrIrr+r,r,r-
write_exportss

rc	cs$t}z
|VW5t|XdSrt)tempfilemkdtemprrmtree)Ztdr,r,r-tempdirs
rc	cs.t}zt|dVW5t|XdSrt)ragetcwdchdir)rcwdr,r,r-rs


rc	cs.t}zt|dVW5t|XdSrt)socketZgetdefaulttimeoutZsetdefaulttimeout)ZsecondsZctor,r,r-socket_timeouts


rc@seZdZddZdddZdS)cached_propertycCs
||_dSrt)func)selfrr,r,r-__init__szcached_property.__init__NcCs,|dkr|S||}t||jj||Srt)robject__setattr____name__)robjclsrr,r,r-__get__s

zcached_property.__get__)N)r
__module____qualname__rrr,r,r,r-rsrcCs~tjdkr|S|s|S|ddkr.td||ddkrFtd||d}tj|krh|tjqP|srtjStjj|S)aReturn 'pathname' as a name that will work on the native filesystem.

    The path is split on '/' and put back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    r`rzpath '%s' cannot be absolutezpath '%s' cannot end with '/')rarc
ValueErrorrcurdirremoverbr$)pathnamepathsr,r,r-convert_paths



rc@seZdZd$ddZddZddZdd	Zd%ddZd&ddZddZ	ddZ
ddZddZddZ
d'ddZddZddZd d!Zd"d#Zd
S)(FileOperatorFcCs||_t|_|dSrt)dry_runrensured_init_record)rrr,r,r-rszFileOperator.__init__cCsd|_t|_t|_dSNF)recordr
files_writtendirs_createdrr,r,r-rszFileOperator._init_recordcCs|jr|j|dSrt)rradd)rrbr,r,r-record_as_writtenszFileOperator.record_as_writtencCsHtj|s tdtj|tj|s0dSt|jt|jkS)aTell if the target is newer than the source.

        Returns true if 'source' exists and is more recently modified than
        'target', or if 'source' exists and 'target' doesn't.

        Returns false if both exist and 'target' is the same age or younger
        than 'source'. Raise PackagingFileError if 'source' does not exist.

        Note that this test is not very accurate: files created in the same
        second will have the same "age".
        zfile '%r' does not existT)rarbexistsrabspathstatst_mtime)rsourcetargetr,r,r-newers
zFileOperator.newerTcCs|tj|td|||jsd}|rdtj|rDd|}n tj|rdtj	|sdd|}|rtt
|dt|||
|dS)z8Copy a file respecting dry-run and force flags.
        zCopying %s to %sNz%s is a symlinkz%s is a non-regular filez which would be overwritten)
ensure_dirrarbdirnameloggerinforislinkrisfilerrZcopyfiler)rZinfileoutfilecheckmsgr,r,r-	copy_files
zFileOperator.copy_fileNcCst|tj|td|||jsf|dkr:t|d}ntj|d|d}zt
||W5|	X||dS)NzCopying stream %s to %swbwencoding)
rrarbrrrropenrrrZcopyfileobjr)rZinstreamrrZ	outstreamr,r,r-copy_streams
zFileOperator.copy_streamc	Cs\|tj||jsNtj|r.t|t|d}||W5QRX|	|dS)Nr)
rrarbrrrrrrr)rrbrfr,r,r-write_binary_file!s
zFileOperator.write_binary_filecCs||||dSrt)rencode)rrbrrr,r,r-write_text_file*szFileOperator.write_text_filecCsntjdkstjdkrjtjdkrj|D]F}|jr:td|q"t|j|B|@}td||t||q"dS)Nposixjavazchanging mode of %szchanging mode of %s to %o)	rarP_namerrrrst_modechmod)rbitsmaskfilesrmoder,r,r-set_mode-szFileOperator.set_modecCs|dd|S)Nimi)r)r+rr,r,r-9zFileOperator.cCs|tj|}||jkrxtj|sx|j|tj|\}}||t	d||j
sft||jrx|j
|dS)NzCreating %s)rarbrrrrrrrrrmkdirrr)rrbrrr,r,r-r;s

zFileOperator.ensure_dirc	Cst||}td|||js||s0|||rJ|s:d}n|t|d}i}|rhttdrhtjj	|d<tj
|||df||||S)NzByte-compiling %s to %sPycInvalidationModeinvalidation_modeT)rrrrrrdro
py_compilerCHECKED_HASHcompiler)	rrboptimizeforcerlZhashed_invalidationZdpathZdiagpathZcompile_kwargsr,r,r-byte_compileGs
zFileOperator.byte_compilecCstj|rtj|r^tj|s^td||js@t	||j
r||jkr|j|nPtj|rpd}nd}td|||jst||j
r||j
kr|j
|dS)NzRemoving directory tree at %slinkfilezRemoving %s %s)rarbrisdirrrdebugrrrrrrr)rrbr+r,r,r-ensure_removedXs"



zFileOperator.ensure_removedcCsDd}|s@tj|r$t|tj}q@tj|}||kr:q@|}q|Sr)rarbraccessW_OKr)rrbr(parentr,r,r-is_writablemszFileOperator.is_writablecCs|j|jf}||S)zV
        Commit recorded changes, turn off recording, return
        changes.
        )rrr)rr(r,r,r-commityszFileOperator.commitcCs|jsxt|jD]}tj|rt|qt|jdd}|D]8}t	|}|rltj
||d}t|t|q>|dS)NT)reverser)
rlistrrarbrrsortedrlistdirr$rmdirr)rrdirsrflistZsdr,r,r-rollbacks

zFileOperator.rollback)F)T)N)FFNF)rrrrrrrrrrrrZset_executable_moderrrrrrr,r,r,r-rs 


	

rcCs^|tjkrtj|}nt|}|dkr,|}n.|d}t||d}|D]}t||}qJ|S)N.r)rpmodules
__import__rrqrg)Zmodule_nameZdotted_pathmodr(r*r|r,r,r-resolves

rc@s6eZdZddZeddZddZddZej	Z	d	S)
ExportEntrycCs||_||_||_||_dSrtrPrlrjr)rrPrlrjrr,r,r-rszExportEntry.__init__cCst|j|jSrt)rrlrjrr,r,r-rszExportEntry.valuecCsd|j|j|j|jfS)Nzrrr,r,r-__repr__s
zExportEntry.__repr__cCsDt|tsd}n0|j|jko>|j|jko>|j|jko>|j|jk}|Sr)rwrrPrlrjr)rotherr(r,r,r-__eq__s



zExportEntry.__eq__N)
rrrrrrrrr__hash__r,r,r,r-rs

rz(?P(\w|[-.+])+)
                      \s*=\s*(?P(\w+)([:\.]\w+)*)
                      \s*(\[\s*(?P\w+(=\w+)?(,\s*\w+(=\w+)?)*)\s*\])?
                      c
Cst|}|s0d}d|ks"d|krtd|n|}|d}|d}|d}|dkrf|d}}n"|dkrztd||d\}}|d	}	|	dkrd|ksd|krtd|g}	nd
d|	dD}	t||||	}|S)
NrArBzInvalid specification '%s'rPcallable:rrrcSsg|]}|qSr,)rS)rNrr,r,r-rOsz$get_export_entry..rC)ENTRY_REsearchr	groupdictcountrr)
Z
specificationr'r(rrPrbZcolonsrlrjrr,r,r-rs8

rcCs|dkrd}tjdkr.dtjkr.tjd}ntjd}tj|rft|tj}|st	
d|n|dd\}}d|kr.|}n|dd\}}|rJt|}|rVt|}|||fS)NrDrr)rsplitrr)rXZusernameZpasswordrlr,r,r-parse_credentials$sr1cCstd}t||S)N)raumaskrrr,r,r-get_process_umask3s

r4cCs0d}d}t|D]\}}t|tsd}q,q|S)NTF)	enumeraterwr)seqr(r[r+r,r,r-is_string_sequence8s
r7z3([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-([a-z0-9_.+-]+)z
-py(\d\.?\d?)cCsd}d}t|dd}t|}|r@|d}|d|}|rt|t|dkrtt	|d|}|r|
}|d|||dd|f}|dkrt|}|r|d|d|f}|S)zw
    Extract name, version, python version from a filename (no extension)

    Return name, version, pyver or None
    N -rz\br)rr!PYTHON_VERSIONrrstartrdrerescaperPROJECT_NAME_AND_VERSION)filenameZproject_namer(Zpyverr'nr,r,r-split_filenameGs"


rAz-(?P[\w .-]+)\s*\(\s*(?P[^\s)]+)\)$cCs:t|}|std||}|d|dfS)z
    A utility method used to get name and version from a string.

    From e.g. a Provides-Dist value.

    :param p: A value in a form 'foo (1.0)'
    :return: The name and version as a tuple.
    z$Ill-formed name/version string: '%s'rPZver)NAME_VERSION_RErrr rSry)r|r'rr,r,r-parse_name_and_versioncs
	
rCcCst}t|pg}t|pg}d|kr8|d||O}|D]x}|dkrT||q<|dr|dd}||krtd|||kr||q<||krtd|||q<|S)N*r9rzundeclared extra: %s)rrrrTrr()Z	requestedZ	availabler(rZunwantedr,r,r-
get_extrasrs&

rFc
Csi}zNt|}|}|d}|ds8td|ntd|}t	|}Wn0t
k
r}ztd||W5d}~XYnX|S)NzContent-Typezapplication/jsonz(Unexpected response for JSON request: %srz&Failed to get external data for %s: %s)rrgetrTrrrrrrr	exception)rRr(ZrespZheadersZctreaderer,r,r-_get_external_datas

 rKz'https://www.red-dove.com/pypi/projects/cCs*d|d|f}tt|}t|}|S)Nz%s/%s/project.jsonrupperr
_external_data_base_urlrK)rPrRr(r,r,r-get_project_datas
rOcCs(d|d||f}tt|}t|S)Nz%s/%s/package-%s.jsonrrL)rPversionrRr,r,r-get_package_datas
rQc@s(eZdZdZddZddZddZdS)	Cachez
    A class implementing a cache for resources that need to live in the file system
    e.g. shared libraries. This class was moved from resources to here because it
    could be used by other modules, e.g. the wheel module.
    cCsPtj|st|t|jd@dkr6td|tjtj	||_
dS)zu
        Initialise an instance.

        :param base: The base directory where the cache should be located.
        ?rzDirectory '%s' is not privateN)rarbrr)rrrr(rnormpathri)rrir,r,r-rs

zCache.__init__cCst|S)zN
        Converts a resource prefix to a directory name in the cache.
        )r-)rrlr,r,r-
prefix_to_dirszCache.prefix_to_dirc	Csg}t|jD]r}tj|j|}z>tj|s>tj|rJt|ntj|r`t	
|Wqtk
r||YqXq|S)z"
        Clear the cache.
        )
rarrirbr$rrrrrrrr")rZnot_removedfnr,r,r-clearszCache.clearN)rrr__doc__rrUrWr,r,r,r-rRsrRc@s:eZdZdZddZdddZddZd	d
ZddZd
S)
EventMixinz1
    A very simple publish/subscribe system.
    cCs
i|_dSrt)_subscribersrr,r,r-rszEventMixin.__init__TcCsD|j}||krt|g||<n"||}|r6||n
||dS)a`
        Add a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be added (and called when the
                           event is published).
        :param append: Whether to append or prepend the subscriber to an
                       existing subscriber list for the event.
        N)rZrr"
appendleft)revent
subscriberr"subsZsqr,r,r-rs
zEventMixin.addcCs,|j}||krtd||||dS)z
        Remove a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be removed.
        zNo subscribers: %rN)rZrr)rr\r]r^r,r,r-rszEventMixin.removecCst|j|dS)z
        Return an iterator for the subscribers for an event.
        :param event: The event to return subscribers for.
        r,)iterrZrG)rr\r,r,r-get_subscribersszEventMixin.get_subscribersc	Oslg}||D]F}z||f||}Wn"tk
rHtdd}YnX||qtd|||||S)a^
        Publish a event and return a list of values returned by its
        subscribers.

        :param event: The event to publish.
        :param args: The positional arguments to pass to the event's
                     subscribers.
        :param kwargs: The keyword arguments to pass to the event's
                       subscribers.
        z"Exception during event publicationNz/publish %s: args = %s, kwargs = %s, result = %s)r`rrrHr"r)rr\argskwargsr(r]rr,r,r-publishs

zEventMixin.publishN)T)	rrrrXrrrr`rcr,r,r,r-rYs
rYc@s^eZdZddZddZdddZdd	Zd
dZdd
ZddZ	e
ddZe
ddZdS)	SequencercCsi|_i|_t|_dSrt)_preds_succsr_nodesrr,r,r-r"szSequencer.__init__cCs|j|dSrt)rgrrnoder,r,r-add_node'szSequencer.add_nodeFcCs||jkr|j||rt|j|dD]}|||q,t|j|dD]}|||qPt|jD]\}}|sp|j|=qpt|jD]\}}|s|j|=qdS)Nr,)rgrrrerGrfr
r)rriZedgesr|r+rrIr,r,r-remove_node*s

zSequencer.remove_nodecCs0|j|t||j|t|dSrt)re
setdefaultrrrf)rpredsuccr,r,r-r:sz
Sequencer.addcCs|z|j|}|j|}Wn tk
r8td|YnXz||||Wn$tk
rvtd||fYnXdS)Nz%r not a successor of anythingz%r not a successor of %r)rerfKeyErrorrr)rrmrnpredsZsuccsr,r,r-r?s

zSequencer.removecCs||jkp||jkp||jkSrt)rerfrg)rstepr,r,r-is_stepLszSequencer.is_stepcCs||std|g}g}t}|||r|d}||krb||kr||||q.|||||j|d}|	|q.t
|S)NzUnknown: %rrr,)rrrrr"rgrrrerGextendreversed)rfinalr(Ztodoseenrqrpr,r,r-	get_stepsPs"





zSequencer.get_stepscsRdggiig|jfddD]}|kr8|q8S)Nrcsd|<d|<dd7<|z|}Wntk
rVg}YnX|D]J}|kr|t|||<q\|kr\t|||<q\||krg}}||||krqqt|}|dS)Nrr)r"rminrgtuple)riZ
successorsZ	successorZconnected_componentZ	componentZgraphindexZ
index_counterZlowlinksr(stack
strongconnectr,r-r}ts,


z3Sequencer.strong_connections..strongconnect)rfrhr,rzr-strong_connectionsis"
zSequencer.strong_connectionscCsfdg}|jD]*}|j|}|D]}|d||fqq|jD]}|d|q>|dd|S)Nzdigraph G {z  %s -> %s;z  %s;}
)rer"rgr$)rr(rnrprmrir,r,r-dots



z
Sequencer.dotN)F)
rrrrrjrkrrrrrwpropertyr~rr,r,r,r-rd!s


2rd).tar.gz.tar.bz2.tar.zip.tgz.tbz.whlTcsZfdd}tjtd}|dkr|dr>d}nH|drRd}d}n4|drfd	}d
}n |drzd}d
}ntd|z|dkrt|d
}|r|}|D]}||qn*t	
||}|r|}|D]}||q|dkr*tj
ddkr*|D]"}	t|	jts|	jd|	_qdd}
|
|_|W5|rT|XdS)NcsRt|ts|d}tjtj|}|rB|tjkrNt	d|dS)Nrzpath outside destination: %r)
rwrdecoderarbrr$rTrcr)rbr|dest_dirZplenr,r-
check_paths


zunarchive..check_path)rrzip)rrZtgzzr:gz)rrZtbzzr:bz2rZtarrEzUnknown format for %rrrrc
SsDzt||WStjk
r>}ztt|W5d}~XYnXdS)z:Run tarfile.tar_fillter, but raise the expected ValueErrorN)tarfileZ
tar_filterZFilterErrorrstr)memberrbexcr,r,r-extraction_filtersz$unarchive..extraction_filter)rarbrrdr.rrrZnamelistrrZgetnamesrprZ
getmembersrwrPrrrZ
extractall)Zarchive_filenamerformatrrarchivernamesrPZtarinforr,rr-	unarchivesL





rc	Cs~t}t|}t|dZ}t|D]H\}}}|D]8}tj||}||d}	tj|	|}
|||
q4q&W5QRX|S)z*zip a directory tree into a BytesIO objectrN)	ioBytesIOrdrrawalkrbr$r)Z	directoryr(ZdlenZzfrerrrPZfullZrelrkr,r,r-zip_dirsr)rKMGTPc@sreZdZdZdddZddZdd	Zd
dZdd
Ze	ddZ
e	ddZddZe	ddZ
e	ddZdS)ProgressZUNKNOWNrdcCs(||_|_||_d|_d|_d|_dS)NrF)rxcurmaxstartedelapseddone)rZminvalZmaxvalr,r,r-r	s
zProgress.__init__cCs0||_t}|jdkr ||_n||j|_dSrt)rtimerr)rZcurvalZnowr,r,r-updates

zProgress.updatecCs||j|dSrt)rr)rZincrr,r,r-	incrementszProgress.incrementcCs||j|Srt)rrxrr,r,r-r;szProgress.startcCs |jdk	r||jd|_dS)NT)rrrrr,r,r-stop#s
z
Progress.stopcCs|jdkr|jS|jSrt)runknownrr,r,r-maximum(szProgress.maximumcCsD|jrd}n4|jdkrd}n$d|j|j|j|j}d|}|S)Nz100 %z ?? %gY@z%3d %%)rrrrx)rr(rIr,r,r-
percentage,s
zProgress.percentagecCs:|dkr|jdks|j|jkr$d}ntdt|}|S)Nrz??:??:??z%H:%M:%S)rrrxrstrftimegmtime)rZdurationr(r,r,r-format_duration7szProgress.format_durationcCs|jrd}|j}n^d}|jdkr&d}nJ|jdks<|j|jkrBd}n.t|j|j}||j|j}|d|j}d|||fS)NZDonezETA rrrz%s: %s)rrrrrxfloatr)rrlr\r,r,r-ETA@s
zProgress.ETAcCsL|jdkrd}n|j|j|j}tD]}|dkr6q@|d}q&d||fS)Nrgig@@z%d %sB/s)rrrxUNITS)rr(Zunitr,r,r-speedSs

zProgress.speedN)rr)rrrrrrrr;rrrrrrrr,r,r,r-rs




	
rz\{([^}]*)\}z[^/\\,{]\*\*|\*\*[^/\\,}]z^[^{]*\}|\{[^}]*$cCs<t|rd}t||t|r4d}t||t|S)zAExtended globbing function that supports ** and {opt1,opt2,opt3}.z7invalid glob %r: recursive glob "**" must be used alonez2invalid glob %r: mismatching set marker '{' or '}')_CHECK_RECURSIVE_GLOBrr_CHECK_MISMATCH_SET_iglob)	path_globrr,r,r-rhs

rccst|d}t|dkrT|\}}}|dD]$}td|||fD]
}|VqDq,nd|krrt|D]
}|Vqdn~|dd\}}|dkrd}|dkrd}n|d}|d}t|D]4\}}}	tj	
|}ttj	||D]
}
|
VqqdS)	NrrCrz**rrDr`\)	RICH_GLOBrrdrr$	std_iglobr%rarrbrT)rZrich_path_globrlrrjitemrbZradicaldirrrVr,r,r-rss(



r)HTTPSHandlermatch_hostnameCertificateErrorc@seZdZdZdZddZdS)HTTPSConnectionNTcCsRt|j|jf|j}t|ddr0||_|tt	dsp|j
rHt	j}nt	j}t	j
||j|j|t	j|j
d|_nxt	t	j}|jt	jO_|jr||j|ji}|j
rt	j|_|j|j
dtt	ddr|j|d<|j
|f||_|j
rN|jrNz$t|j|jtd|jWn0tk
rL|jtj|jYnXdS)	NZ_tunnel_hostF
SSLContext)	cert_reqsZssl_versionca_certs)ZcafileZHAS_SNIZserver_hostnamezHost verified: %s) rZcreate_connectionhostporttimeoutrqsockZ_tunnelrosslrZ
CERT_REQUIREDZ	CERT_NONEZwrap_socketZkey_fileZ	cert_fileZPROTOCOL_SSLv23rZoptionsZOP_NO_SSLv2Zload_cert_chainZverify_modeZload_verify_locationscheck_domainrZgetpeercertrrrZshutdownZ	SHUT_RDWRr)rrrcontextrbr,r,r-connects@



zHTTPSConnection.connect)rrrrrrr,r,r,r-rsrc@s&eZdZd	ddZddZddZdS)
rTcCst|||_||_dSrt)BaseHTTPSHandlerrrr)rrrr,r,r-rs
zHTTPSHandler.__init__cOs$t||}|jr |j|_|j|_|S)a
            This is called to create a connection instance. Normally you'd
            pass a connection class to do_open, but it doesn't actually check for
            a class, and just expects a callable. As long as we behave just as a
            constructor would have, we should be OK. If it ever changes so that
            we *must* pass a class, we'll create an UnsafeHTTPSConnection class
            which just sets check_domain to False in the class definition, and
            choose which one to pass to do_open.
            )rrr)rrarbr(r,r,r-_conn_makers


zHTTPSHandler._conn_makerc
CsXz||j|WStk
rR}z$dt|jkr@td|jnW5d}~XYnXdS)Nzcertificate verify failedz*Unable to verify server certificate for %s)Zdo_openrrrreasonrr)rrYrJr,r,r-
https_openszHTTPSHandler.https_openN)T)rrrrrrr,r,r,r-rs
rc@seZdZddZdS)HTTPSOnlyHandlercCstd|dS)NzAUnexpected HTTP request on what should be a secure connection: %s)r)rrYr,r,r-	http_openszHTTPSOnlyHandler.http_openN)rrrrr,r,r,r-rsrrc@seZdZdddZdS)HTTPrNcKs&|dkrd}||j||f|dSNr_setupZ_connection_classrrrrbr,r,r-rsz
HTTP.__init__)rNrrrrr,r,r,r-rsrc@seZdZdddZdS)HTTPSrNcKs&|dkrd}||j||f|dSrrrr,r,r-rszHTTPS.__init__)rNrr,r,r,r-rsrc@seZdZdddZddZdS)	TransportrcCs||_tj||dSrt)rrrrrruse_datetimer,r,r-rszTransport.__init__cCs`||\}}}tdkr(t||jd}n4|jr<||jdkrR||_|t|f|_|jd}|S)Nr)rrr)
get_host_info	_ver_inforr_connection_extra_headersrZHTTPConnection)rrhehZx509r(r,r,r-make_connection
s
zTransport.make_connectionN)rrrrrrr,r,r,r-rs
rc@seZdZdddZddZdS)
SafeTransportrcCs||_tj||dSrt)rrrrrr,r,r-rszSafeTransport.__init__cCsx||\}}}|si}|j|d<tdkr:t|df|}n:|jrN||jdkrj||_|tj|df|f|_|jd}|S)Nrrrr)rrrrrrrr)rrrrrbr(r,r,r-rs

zSafeTransport.make_connectionN)rrr,r,r,r-rs
rc@seZdZddZdS)ServerProxyc	Kst|dd|_}|dk	r^t|\}}|dd}|dkr@t}nt}|||d|d<}||_tjj	||f|dS)NrrrZhttps)r	transport)
rgrrrGrrrrrr)	rrZrbrrWr]rZtclsr\r,r,r-r,szServerProxy.__init__Nrr,r,r,r-r+srcKs6tjddkr|d7}nd|d<d|d<t||f|S)Nrrbrnewlinerr)rprr)rVrrbr,r,r-	_csv_open@s

rc@s4eZdZedededdZddZddZd	S)
CSVBaserC"r)Z	delimiterZ	quotecharZlineterminatorcCs|Srtr,rr,r,r-	__enter__RszCSVBase.__enter__cGs|jdSrt)rr)rr$r,r,r-__exit__UszCSVBase.__exit__N)rrrrdefaultsrrr,r,r,r-rKsrc@s(eZdZddZddZddZeZdS)	CSVReadercKs\d|kr4|d}tjddkr,td|}||_nt|dd|_tj|jf|j|_dS)NrrrrrbrE)	rprrrrrcsvrIr)rrbrr,r,r-rZszCSVReader.__init__cCs|Srtr,rr,r,r-__iter__eszCSVReader.__iter__cCsFt|j}tjddkrBt|D] \}}t|ts |d||<q |SNrrr)nextrIrprr5rwrr)rr(r[rr,r,r-rhs

zCSVReader.nextN)rrrrrr__next__r,r,r,r-rYsrc@seZdZddZddZdS)	CSVWritercKs$t|d|_tj|jf|j|_dS)Nr)rrrwriterr)rrVrbr,r,r-rsszCSVWriter.__init__cCsNtjddkr>g}|D]"}t|tr.|d}||q|}|j|dSr)rprrwrrr"rwriterow)rrowrErr,r,r-rws

zCSVWriter.writerowN)rrrrrr,r,r,r-rrsrcsHeZdZeejZded<dfdd	ZddZdd	Zd
dZ	Z
S)
Configuratorinc_convertZincNcs"tt|||pt|_dSrt)superrrrarri)rconfigri	__class__r,r-rszConfigurator.__init__c
sfddd}t|s*|}dd}dd}|r\tfdd|D}fd	dD}t|}|||}|r|D]\}}	t|||	q|S)
Ncsvt|ttfr*t|fdd|D}nHt|trhd|krH|}qri}|D]}||||<qPn
|}|S)Ncsg|]}|qSr,r,)rNr[convertr,r-rOszBConfigurator.configure_custom..convert..())rwr
rytypedictconfigure_customr)or(r)rrr,r-rs

z.Configurator.configure_custom..convertrrz[]r,csg|]}|qSr,r,)rNr
rr,r-rOsz1Configurator.configure_custom..cs$g|]}t|r||fqSr,)r)rNr)rrr,r-rOs)rgrrryrrsetattr)
rrr}Zpropsrarrbr(r@rIr,)rrrr-r	s


zConfigurator.configure_customcCs4|j|}t|tr0d|kr0|||j|<}|S)Nr)rrwrr	)rrr(r,r,r-__getitem__s
zConfigurator.__getitem__c	CsFtj|stj|j|}tj|ddd}t|}W5QRX|S)z*Default converter for the inc:// protocol.rErr)	rarbisabsr$rirrrr)rrrr(r,r,r-rs
zConfigurator.inc_convert)N)rrrrrZvalue_convertersrr	rr
__classcell__r,r,rr-rs
rc@s*eZdZdZd
ddZddZdd	ZdS)SubprocessMixinzC
    Mixin for running subprocesses and capturing their output
    FNcCs||_||_dSrt)verboseprogress)rrrr,r,r-rszSubprocessMixin.__init__cCsj|j}|j}|}|sq^|dk	r.|||q|s@tjdntj|dtjq|dS)z
        Read lines from a subprocess' output stream and either pass to a progress
        callable (if specified) or write progress information to sys.stderr.
        Nrr)	rrreadlinerpstderrrrflushr)rrrrrr+r,r,r-rIszSubprocessMixin.readercKstj|ftjtjd|}tj|j|jdfd}|tj|j|jdfd}||	|
|
|jdk	r|ddn|jrt
jd|S)N)stdoutrr)rrarzdone.mainzdone.
)
subprocessPopenPIPE	threadingZThreadrIrr;rwaitr$rrrpr)rcmdrbr|t1t2r,r,r-run_commands"
zSubprocessMixin.run_command)FN)rrrrXrrIrr,r,r,r-rs
rcCstdd|S)z,Normalize a python package name a la PEP 503z[-_.]+r9)r<subry)rPr,r,r-normalize_namesr!)NN)r)N)N)NT)rcollectionsr
contextlibrZglobrrrrZloggingrarr<rrImportErrorrrprrrrZdummy_threadingrrrcompatrrrr	r
rrr
rrrrrrrrrrrrrZ	getLoggerrrrrrGrFr6r>r:rVr#r?r^rmrsrxr~rrrcontextmanagerrrrrrrrrrVERBOSErrr+r-r/r1r4r7Ir>r:rArBrCrFrKrNrOrQrRrYrdZARCHIVE_EXTENSIONSrrrrrrrrrrrrrrrrrrrrrrrrrrrr!r,r,r,r-s
\








Yy

	/

		8
)



,H
C]


*)	
7.site-packages/pip/_vendor/distlib/__pycache__/manifest.cpython-38.pyc000064400000023725151464452710021632 0ustar00U

.e9@sdZddlZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZdgZe
eZedejZed	ejejBZejdd
ZGdddeZdS)zu
Class representing the list of files in a distribution.

Equivalent to distutils.filelist, but fixes some problems.
N)DistlibException)fsdecodeconvert_pathManifestz\\w*
z#.*?(?=
)|
(?=$)c@szeZdZdZdddZddZddZd	d
Zddd
ZddZ	ddZ
ddZdddZd ddZ
d!ddZddZdS)"rz~A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.
    NcCs>tjtj|pt|_|jtj|_d|_t	|_
dS)zd
        Initialise an instance.

        :param base: The base directory to explore under.
        N)ospathabspathnormpathgetcwdbasesepprefixallfilessetfiles)selfrr@/usr/lib/python3.8/site-packages/pip/_vendor/distlib/manifest.py__init__*szManifest.__init__cCsddlm}m}m}g|_}|j}|g}|j}|j}|r|}t	|}	|	D]R}
tj
||
}t|}|j}
||
r|t
|qN||
rN||
sN||qNq6dS)zmFind all files under the base and set ``allfiles`` to the absolute
        pathnames of files found.
        r)S_ISREGS_ISDIRS_ISLNKN)statrrrrrpopappendr	listdirr
joinst_moder)rrrrrrootstackrpushnamesnamefullnamermoderrrfindall9s"


zManifest.findallcCs4||jstj|j|}|jtj|dS)zz
        Add a file to the manifest.

        :param item: The pathname to add. This can be relative to the base.
        N)	
startswithrr	r
rrraddr)ritemrrrr*TszManifest.addcCs|D]}||qdS)z
        Add a list of files to the manifest.

        :param items: The pathnames to add. These can be relative to the base.
        N)r*)ritemsr+rrradd_many^szManifest.add_manyFcsbfddtj}|rFt}|D]}|tj|q&||O}ddtdd|DDS)z8
        Return sorted files in directory order
        csJ||td||jkrFtj|\}}|dks.add_dircSsg|]}tjj|qSr)r	r
r).0Z
path_tuplerrr
zsz#Manifest.sorted..css|]}tj|VqdS)N)r	r
r2)r:r
rrr	{sz"Manifest.sorted..)rrr	r
dirnamesorted)rZwantdirsresultr4frr8rr>gs
zManifest.sortedcCst|_g|_dS)zClear all collected files.N)rrr)rrrrclear}szManifest.clearcCsj||\}}}}|dkrB|D]}|j|ddstd|qn$|dkrf|D]}|j|dd}qNn|dkr|D]}|j|ddsrtd|qrn|d	kr|D]}|j|dd}qn|d
kr|D] }|j||dstd||qn|d
kr
|D]}|j||d}qn\|dkr2|jd|dsftd|n4|dkrZ|jd|dsftd|ntd|dS)av
        Process a directive which either adds some files from ``allfiles`` to
        ``files``, or removes some files from ``files``.

        :param directive: The directive to process. This should be in a format
                     compatible with distutils ``MANIFEST.in`` files:

                     http://docs.python.org/distutils/sourcedist.html#commands
        includeT)anchorzno files found matching %rexcludeglobal-includeFz3no files found matching %r anywhere in distributionglobal-excluderecursive-include)rz-no files found matching %r under directory %rrecursive-excludegraftNz no directories found matching %rprunez4no previously-included directories found matching %rzinvalid action %r)_parse_directive_include_patternr0Zwarning_exclude_patternr)r	directiveactionpatternsthedirZ
dirpatternpatternfoundrrrprocess_directivesP


zManifest.process_directivecCs|}t|dkr,|ddkr,|dd|d}d}}}|dkrxt|dkr`td|d	d
|ddD}n~|dkrt|dkrtd
|t|d}dd
|ddD}n:|dkrt|dkrtd|t|d}ntd|||||fS)z
        Validate a directive.
        :param directive: The directive to validate.
        :return: A tuple of action, patterns, thedir, dir_patterns
        rr)rBrDrErFrGrHrIrJrBN)rBrDrErFrz$%r expects   ...cSsg|]}t|qSrrr:Zwordrrrr;sz-Manifest._parse_directive..)rGrHz*%r expects    ...cSsg|]}t|qSrrrUrrrr;s)rIrJz!%r expects a single zunknown action %r)r2leninsertrr)rrNZwordsrOrPrQZdir_patternrrrrKs4zManifest._parse_directiveTcCsPd}|||||}|jdkr&||jD]}||r,|j|d}q,|S)aSelect strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.

        Patterns are not quite the same as implemented by the 'fnmatch'
        module: '*' and '?'  match non-special characters, where "special"
        is platform-dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found.
        FNT)_translate_patternrr(searchrr*)rrRrCris_regexrS
pattern_rer%rrrrLs


zManifest._include_patterncCsBd}|||||}t|jD]}||r|j|d}q|S)atRemove strings (presumably filenames) from 'files' that match
        'pattern'.

        Other parameters are the same as for 'include_pattern()', above.
        The list 'self.files' is modified in place. Return True if files are
        found.

        This API is public to allow e.g. exclusion of SCM subdirs, e.g. when
        packaging source distributions
        FT)rYlistrrZremove)rrRrCrr[rSr\r@rrrrM)s
zManifest._exclude_patternc
Cs|rt|trt|S|Stdkr:|dd\}}}|rj||}tdkrn||rd||snt	nd}t
tj
|jd}	|dk	rftdkr|d}
||dt|
}n>||}||r||st	|t|t|t|}tj}tjdkrd}tdkr4d|	|
|d|f}n0|t|t|t|}d	||	||||f}n8|rtdkrd|	|}nd
||	|t|df}t|S)aTranslate a shell-like wildcard pattern to a compiled regular
        expression.

        Return the compiled regex.  If 'is_regex' true,
        then 'pattern' is directly compiled to a regex (if it's a string)
        or just returned as-is (assumes it's a regex object).
        )rVrr7r.N\z\\^z.*z%s%s%s%s.*%s%sz%s%s%s)
isinstancestrrecompile_PYTHON_VERSION_glob_to_re	partitionr)endswithr3escaper	r
rrrWr)
rrRrCrr[startr7endr\rZ
empty_patternZ	prefix_rerrrrrY=sH	









zManifest._translate_patterncCs8t|}tj}tjdkrd}d|}td||}|S)zTranslate a shell-like glob pattern to a regular expression.

        Return a string containing the regex.  Differs from
        'fnmatch.translate()' in that '*' does not match "special characters"
        (which are platform-specific).
        r_z\\\\z\1[^%s]z((?rArTrKrLrMrYrfrrrrr%s&

	
O/
)

7)rrrlZloggingr	rcsysr.rcompatrutilr__all__Z	getLoggerror0rdMZ_COLLAPSE_PATTERNSZ_COMMENTED_LINEversion_inforeobjectrrrrrs
site-packages/pip/_vendor/distlib/__pycache__/wheel.cpython-38.pyc000064400000061656151464452710021135 0ustar00U

.e@sddlmZddlZddlZddlZddlZddlmZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlZddlZddlZddlmZmZddlmZmZmZmZmZddlmZddlm Z m!Z!m"Z"dd	l#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,dd
l-m.Z.m/Z/e0e1Z2da3e4edr8dZ5n*ej67d
rLdZ5nej6dkr^dZ5ndZ5e8dZ9e9sdej:ddZ9de9Z;e5e9Zdd>ddZ?e8dZ@e@re@7dre@>ddZ@nddZAeAZ@[AeBdejCejDBZEeBdejCejDBZFeBdZGeBd ZHd!ZId"ZJe
jKd#krBd$d%ZLnd&d%ZLGd'd(d(eMZNeNZOGd)d*d*eMZPd+d,ZQeQZR[Qd/d-d.ZSdS)0)unicode_literalsN)message_from_file)__version__DistlibException)	sysconfigZipFilefsdecode	text_typefilter)InstalledDistribution)MetadataMETADATA_FILENAMEWHEEL_METADATA_FILENAME)	FileOperatorconvert_path	CSVReader	CSVWriterCachecached_propertyget_cache_baseread_exportstempdir)NormalizedVersionUnsupportedVersionErrorZpypy_version_infoZppjavaZjyZcliZipcpZpy_version_nodotz%s%spy-_.ZSOABIzcpython-cCsRdtg}tdr|dtdr0|dtddkrH|dd	|S)
NrZPy_DEBUGdZ
WITH_PYMALLOCmZPy_UNICODE_SIZEu)
VER_SUFFIXrget_config_varappendjoin)partsr,=/usr/lib/python3.8/site-packages/pip/_vendor/distlib/wheel.py_derive_abi;s




r.zz
(?P[^-]+)
-(?P\d+[^-]*)
(-(?P\d+[^-]*))?
-(?P\w+\d+(\.\w+\d+)*)
-(?P\w+)
-(?P\w+(\.\w+)*)
\.whl$
z7
(?P[^-]+)
-(?P\d+[^-]*)
(-(?P\d+[^-]*))?$
s
\s*#![^\r\n]*s^(\s*#!("[^"]+"|\S+))\s+(.*)$s#!pythons	#!pythonw/cCs|SNr,or,r,r-]r3cCs|tjdS)Nr/)replaceossepr1r,r,r-r3_r4c@s6eZdZddZddZddZddd	Zd
dZdS)
MountercCsi|_i|_dSr0)
impure_wheelslibsselfr,r,r-__init__cszMounter.__init__cCs||j|<|j|dSr0)r9r:update)r<pathname
extensionsr,r,r-addgs
zMounter.addcCs0|j|}|D]\}}||jkr|j|=qdSr0)r9popr:)r<r?r@kvr,r,r-removeks
zMounter.removeNcCs||jkr|}nd}|Sr0)r:)r<fullnamepathresultr,r,r-find_moduleqs
zMounter.find_modulecCsj|tjkrtj|}nP||jkr,td|t||j|}||_|dd}t|dkrf|d|_	|S)Nzunable to find extension for %sr!rr)
sysmodulesr:ImportErrorimpZload_dynamic
__loader__rsplitlen__package__)r<rFrHr+r,r,r-load_modulexs


zMounter.load_module)N)__name__
__module____qualname__r=rArErIrRr,r,r,r-r8bs

r8c@seZdZdZdZdZd4ddZedd	Zed
dZ	edd
Z
eddZddZ
eddZddZd5ddZddZddZddZd6ddZd d!Zd"d#Zd$d%Zd&d'Zd(d)Zd*d+Zd7d,d-Zd.d/Zd0d1Zd8d2d3ZdS)9Wheelz@
    Class to build and install from Wheel files (PEP 427).
    rrZsha256NFcCs8||_||_d|_tg|_dg|_dg|_t|_	|dkrRd|_
d|_|j|_
nt|}|r|d}|d|_
|dd	d
|_|d|_|j|_
ntj|\}}t|}|std||rtj||_	||_
|d}|d|_
|d|_|d|_|d
d|_|dd|_|dd|_dS)zB
        Initialise an instance using a (valid) filename.
        r&noneanyNZdummyz0.1ZnmZvnr rZbnzInvalid name or filename: %rrr!ZbiZar)signZ
should_verifybuildverPYVERpyverabiarchr6getcwddirnamenameversionfilenameZ	_filenameNAME_VERSION_REmatch	groupdictr5rGsplitFILENAME_RErabspath)r<rdrZverifyr#inforar,r,r-r=sD











zWheel.__init__cCs^|jrd|j}nd}d|j}d|j}d|j}|jdd}d|j|||||fS)zJ
        Build and return a filename from the various components.
        rr&r!r z%s-%s%s-%s-%s-%s.whl)r[r*r]r^r_rcr5rb)r<r[r]r^r_rcr,r,r-rds
zWheel.filenamecCstj|j|j}tj|Sr0)r6rGr*rardisfile)r<rGr,r,r-existsszWheel.existsccs4|jD](}|jD]}|jD]}|||fVqqqdSr0)r]r^r_)r<r]r^r_r,r,r-tagss


z
Wheel.tagscCstj|j|j}d|j|jf}d|}td}t	|d}|
|}|ddd}tdd	|D}|d
krt
tdg}	nt
tg}	d}
|	D]f}zLt||}||,}
||
}t|d}
|
rW5QRWqW5QRXWqtk
rYqXq|
std
d|	W5QRX|
S)N%s-%s%s.dist-infoutf-8r
Wheel-Versionr!rcSsg|]}t|qSr,int.0ir,r,r-
sz"Wheel.metadata..rWZMETADATA)Zfileobjz8Invalid wheel, because metadata is missing: looked in %sz, )r6rGr*rardrbrccodecs	getreaderrget_wheel_metadatarhtuplerr	posixpathopenr
KeyError
ValueError)r<r?name_verinfo_dirwrapperzfwheel_metadatawvfile_versionZfnsrHfnmetadata_filenamebfwfr,r,r-metadatas6


zWheel.metadatac	CsXd|j|jf}d|}t|d}||}td|}t|}W5QRXt|S)NrprqWHEELrr)	rbrcrr*rr{r|rdict)r<rrrrrrmessager,r,r-r}szWheel.get_wheel_metadatac	Cs6tj|j|j}t|d}||}W5QRX|S)Nrs)r6rGr*rardrr})r<r?rrHr,r,r-rlsz
Wheel.infocCst|}|r||}|d|||d}}d|krBt}nt}t|}|rfd|d}nd}||}||}nT|d}|d}	|dks||	krd}
n|||dd	krd	}
nd}
t|
|}|S)
Nspythonw r4

rrs
)	
SHEBANG_RErfendlowerSHEBANG_PYTHONWSHEBANG_PYTHONSHEBANG_DETAIL_REgroupsfind)r<datar#rZshebangZdata_after_shebangZshebang_pythonargsZcrZlfZtermr,r,r-process_shebangs,




zWheel.process_shebangcCsh|dkr|j}ztt|}Wn tk
r<td|YnX||}t|d	d}||fS)NzUnsupported hash algorithm: %r=ascii)
	hash_kindgetattrhashlibAttributeErrorrdigestbase64Zurlsafe_b64encoderstripdecode)r<rrhasherrHr,r,r-get_hash"szWheel.get_hashc	Cs^t|}ttj||}||ddf|t|}|D]}||q@W5QRXdS)Nr&)	listto_posixr6rGrelpathr)sortrZwriterow)r<recordsZrecord_pathbasepwriterrowr,r,r-write_record-s
zWheel.write_recordc
Csg}|\}}tt|j}|D]P\}}	t|	d}
|
}W5QRXd||}tj|	}
|	|||
fqtj
|d}	|||	|ttj
|d}|	||	fdS)Nrbz%s=%sRECORD)
rrrrreadrr6rGgetsizer)r*rr)r<rllibdir
archive_pathsrdistinforraprfrrsizer,r,r-
write_records6szWheel.write_recordsc	CsFt|dtj.}|D]"\}}td|||||qW5QRXdS)NwzWrote %s to %s in wheel)rzipfileZZIP_DEFLATEDloggerdebugwrite)r<r?rrrrr,r,r-	build_zipFszWheel.build_zipc!s~|dkri}ttfdddd}|dkrFd}tg}tg}tg}nd}tg}d	g}d
g}|d||_|d||_|d
||_	|}	d|j
|jf}
d|
}d|
}g}
dD]}|krq|}tj
|rt|D]\}}}|D]}ttj
||}tj
||}ttj
|||}|
||f|dkr|dst|d}|}W5QRX||}t|d}||W5QRXqqq|	}d}t|D]\}}}||krt|D]8\}}t|}|drtj
||}||=qq|std|D]H}t|dr(qtj
||}ttj
||}|
||fqqt|}|D]B}|dkrlttj
||}ttj
||}|
||fqld|p|jdtd|g}|jD] \}}}|d|||fqtj
|d}t|d}|d |W5QRXttj
|d}|
||f| ||f|	|
tj
|j!|j"} |#| |
| S)!z
        Build a wheel from files in specified paths, and use any specified tags
        when determining the name of the wheel.
        Ncs|kSr0r,r1pathsr,r-r3Tr4zWheel.build..)purelibplatlibrrZfalsetruerXrYr]r^r_rp%s.datarq)rZheadersscriptsr.exerwbz
.dist-infoz(.dist-info directory expected, not found)z.pycz.pyo)rZ	INSTALLERZSHAREDrzWheel-Version: %d.%dzGenerator: distlib %szRoot-Is-Purelib: %sz
Tag: %s-%s-%srr
)$rrIMPVERABIARCHr\getr]r^r_rbrcr6rGisdirwalkr	r*rrr)endswithrrrr	enumerateAssertionErrorlistdir
wheel_versionrrorrardr)!r<rrorZlibkeyZis_pureZ
default_pyverZdefault_abiZdefault_archrrdata_dirrrkeyrGrootdirsfilesrrrprrrrrydnrr]r^r_r?r,rr-buildLs




zWheel.buildcCs
|dS)zl
        Determine whether an archive entry should be skipped when verifying
        or installing.
        )r/z/RECORD.jws)r)r<arcnamer,r,r-
skip_entryszWheel.skip_entrycCKsf|j}|d}|dd}|dd}tj|j|j}d|j|jf}	d|	}
d|	}t	|t
}t	|d}
t	|d	}td
}t
|d}||
}||}t|}W5QRX|dd
d}tdd|D}||jkr|r||j||ddkr|d}n|d}i}||8}t|d"}|D]}|d}|||<q8W5QRXW5QRXt	|
d}t	|d}t	|
dd}t|d}d|_tj} g}!t}"|"|_d|_zz\|D]}#|#j}$t|$t r|$}%n
|$!d
}%|"|%rq||%}|dr4t#|#j$|dkr4t%d|%|dr|ddd\}&}'||$}|&}(W5QRX|'|(|&\})}*|*|'krt%d|$|r|%(||frt)*d |%q|%(|o|%+d!}+|%(|r
|%d"d\})},}-tj||,t,|-}.n$|%|
|fkrqtj|t,|%}.|+s
||$}|-||.W5QRX|!.|.|s|drt|.d#4}|&}(|'|(|&\})}/|/|*krt%d$|.W5QRX| r~|.+d%r~z|j/|.|d&}0|!.|0Wn$t0k
rt)j1d'dd(YnXnttj2t,|$}1tj|"|1}2||$}|-||2W5QRXtj|.\}3}1|3|_|3|1}4|4|4|!5|4q|rt)*d)d}5nnd}6|j6d}|d*krzt	|d+}7z||7}t7|}8W5QRXi}6d,D]l}9d-|9}:|:|8kri|6d.|9<};|8|:8D]6}|6d6i}?|>s|?r|dd}@tj>|@s*t?d7|@|_|>@D]*\}:}.zRoot-Is-Purelibrrrstreamrr&r)dry_runTNrsize mismatch for %s=digest mismatch for %szlib_only: skipping %srr/rzdigest mismatch on write for %sz.py)Zhashed_invalidationzByte-compilation failed)exc_infozlib_only: returning Nonez1.0zentry_points.txt)Zconsoleguiz
%s_scriptszwrap_%sz%s:%sz %szAUnable to read legacy script metadata, so cannot generate scriptsr@zpython.commandsz8Unable to read JSON metadata, so cannot generate scriptsZwrap_consoleZwrap_guizValid script path not specifiedz%s = %srlibprefixzinstallation failed.)Grrr6rGr*rardrbrcrrr{r|rrrrhr~rrrrecordrJdont_write_bytecodetempfileZmkdtempZ
source_dirZ
target_dirshutilZrmtreeinfolist
isinstancer
rrstr	file_sizerrr
startswithrrrrZcopy_streamr)Zbyte_compile	ExceptionZwarningbasenameZmakeZset_executable_modeextendrlrvaluesrsuffixflagsjsonloadrritemsrrZwrite_shared_locationsZwrite_installed_filesZ	exceptionZrollback)Cr<rZmakerkwargsrrrZbc_hashed_invalidationr?rrr
metadata_namewheel_metadata_namerecord_namerrbwfrrrrrrrreaderrrZdata_pfxZinfo_pfxZ
script_pfxZfileopZbcZoutfilesworkdirzinfor	u_arcnamekindvaluerr rZ	is_scriptwhererZoutfileZ	newdigestZpycrZworknamer	filenamesZdistZcommandsZepZepdatarrCr"rDsZconsole_scriptsZgui_scriptsZ
script_dirZscriptZoptionsr,r,r-installsT


 















z
Wheel.installcCs4tdkr0tjttdtjdd}t|atS)Nzdylib-cache)	cacher6rGr*rrrJrcr)r<rr,r,r-_get_dylib_cacheszWheel._get_dylib_cachec
Cshtj|j|j}d|j|jf}d|}t|d}t	d}g}t
|d}z||}||}	t
|	}
|}||}tj|j|}
tj|
st|
|
D]\}}tj|
t|}tj|sd}n6t|j}tj|}||}tj|j}||k}|r&|||
|||fqW5QRXWntk
rXYnXW5QRX|S)NrprqZ
EXTENSIONSrrrsT)r6rGr*rardrbrcrr{r|rrrrrZ
prefix_to_dirrrmakedirsrrrnstatst_mtimedatetimeZ
fromtimestampZgetinfoZ	date_timeextractr)r)r<r?rrrrrHrrrr@rrZ
cache_baserbrdestrZ	file_timerlZ
wheel_timer,r,r-_get_extensionss>




zWheel._get_extensionscCst|S)zM
        Determine if a wheel is compatible with the running system.
        )
is_compatibler;r,r,r-rszWheel.is_compatiblecCsdS)zP
        Determine if a wheel is asserted as mountable by its metadata.
        Tr,r;r,r,r-is_mountableszWheel.is_mountablecCstjtj|j|j}|s2d|}t||sJd|}t||t	jkrbt
d|nN|rtt	j|nt	j
d||}|rtt	jkrt	jtt||dS)Nz)Wheel %s not compatible with this Python.z$Wheel %s is marked as not mountable.z%s already in pathr)r6rGrjr*rardrrrrJrrr)insertr_hook	meta_pathrA)r<r)r?msgr@r,r,r-mounts"

zWheel.mountcCsrtjtj|j|j}|tjkr2td|n.rrr/..invalid entry in wheel: %rrrrr)r6rGr*rardrbrcrrr{r|rrrrhr~rrrr
rrrrrrr)r<r?rrrrrrrrr	rrrrrrr
rrrrr
rrrr rr,r,r-rks\





zWheel.verifyc
Ksdd}dd}tj|j|j}d|j|jf}d|}t|d}	t}
t	|d|}i}|
D]h}
|
j}t|tr|}n
|
d	}||	krqhd
|krtd|||
|
tj|
t|}|||<qhW5QRX|||\}}||f|}|r|||\}}|r$||kr$||||dkrNtjd
d|
d\}}t|n*tj|shtd|tj||j}t|}tj|
|}||f}|||
|||||dkrt||W5QRX|S)a
        Update the contents of a wheel in a generic way. The modifier should
        be a callable which expects a dictionary argument: its keys are
        archive-entry paths, and its values are absolute filesystem paths
        where the contents the corresponding archive entries can be found. The
        modifier is free to change the contents of the files pointed to, add
        new entries and remove entries, before returning. This method will
        extract the entire contents of the wheel to a temporary location, call
        the modifier, and then use the passed (and possibly updated)
        dictionary to write a new wheel. If ``dest_dir`` is specified, the new
        wheel is written there -- otherwise, the original wheel is overwritten.

        The modifier should return True if it updated the wheel, else False.
        This method returns the same value the modifier returns.
        cSsHd}}d|tf}||kr$d|}||kr@||}t|dj}||fS)Nz%s/%sz%s/PKG-INFOrG)rr
rc)path_maprrcrGrr,r,r-get_versionKsz!Wheel.update..get_versioncSsd}z|t|}|d}|dkr*d|}nTdd||dddD}|dd7<d	|d|dd
d|Df}Wn tk
rtd|YnX|rt|d
}||_|	t
}|j||dtd||dS)Nrrz%s+1cSsg|]}t|qSr,ru)rxrr,r,r-rz]sz8Wheel.update..update_version..rr!rz%s+%scss|]}t|VqdSr0)rrwr,r,r-	`sz7Wheel.update..update_version..z0Cannot update non-compliant (PEP-440) version %rr()rGlegacyzVersion updated from %r to %r)rrrhr*rrrr
rcrrr)rcrGupdatedrDryr+Zmdr,r,r,r-update_versionUs.

 


z$Wheel.update..update_versionrprqrrsrrr&r'Nz.whlz
wheel-update-)rrdirzNot a directory: %r)r6rGr*rardrbrcrrrrrr
rrrrrZmkstempcloserrrrrrZcopyfile)r<ZmodifierZdest_dirrr*r.r?rrrrrr)rrr
rGZoriginal_versionr ZmodifiedZcurrent_versionfdnewpathrrrlr,r,r-r>:s\







zWheel.update)NFF)N)NN)F)N)rSrTrU__doc__rrr=propertyrdrnrorrr}rlrrrrrrrrrrrrr$r%rkr>r,r,r,r-rVs@
)



	

	
hg	"
8rVc	Csxtg}td}ttjddddD]}|d|t|gq$g}tD]*\}}}|	drN||
dddqN|tdkr|
dt|dg}tg}tjd	krtd
t}|r|\}	}}}
t|}|
g}|
dkr|d|
d
kr|d|
dkr |d|
dkr4|d|
dkrH|d|dkr|D]*}d|	|||f}
|
tkrV||
qV|d8}qH|D]0}|D]$}
|dt|df||
fqqt|D]L\}}|dt|fddf|dkr|dt|dfddfqt|D]L\}}|dd|fddf|dkr"|dd|dfddfq"t|S)zG
    Return (pyver, abi, arch) tuples compatible with this Python.
    rrrr&z.abir!rrXdarwinz(\w+)_(\d+)_(\d+)_(\w+)$)i386ppcZfat)r6r7x86_64Zfat3)ppc64r8Zfat64)r6r8intel)r6r8r:r7r9Z	universalz%s_%s_%s_%srYr)r'rangerJversion_infor)r*rrMZget_suffixesrrhrrr rplatformrerfrrv
IMP_PREFIXrset)ZversionsmajorminorZabisrr rHZarchesr#rbr_Zmatchesrfrr^ryrcr,r,r-compatible_tagss`












&
"
"rCcCs\t|tst|}d}|dkr"t}|D]0\}}}||jkr&||jkr&||jkr&d}qXq&|S)NFT)rrVCOMPATIBLE_TAGSr]r^r_)ZwheelrorHZverr^r_r,r,r-rs
r)N)TZ
__future__rrr{rZdistutils.utilZ	distutilsZemailrrrMrZloggingr6rr>rrJrrr&rrcompatrrr	r
rZdatabaserrr
rrutilrrrrrrrrrrcrrZ	getLoggerrSrrhasattrr?r=rr(r'r<r\rZget_platformr5rrr.compile
IGNORECASEVERBOSErirerrrrr7robjectr8r!rVrCrDrr,r,r,r-s,


	





#>site-packages/pip/_vendor/distlib/__pycache__/markers.cpython-38.pyc000064400000010555151464452710021465 0ustar00U

.e#@sdZddlZddlZddlZddlZddlmZmZmZddl	m
Z
mZdgZddZ
Gd	d
d
eZddZeZ[eZdd
dZdS)zG
Parser for the environment markers micro-language defined in PEP 508.
N)python_implementationurlparsestring_types)in_venvparse_marker	interpretcCst|tr|sdS|ddkS)NFr'")
isinstancer)or?/usr/lib/python3.8/site-packages/pip/_vendor/distlib/markers.py_is_literalsrc
@sfeZdZdZddddddddddddd	dd
dddddd
ddddZddZdS)	Evaluatorz;
    This class is used to evaluate marker expessions.
    cCs||kSNrxyrrr
$zEvaluator.cCs||kSrrrrrr
r%rcCs||kp||kSrrrrrr
r&rcCs||kSrrrrrr
r'rcCs||kSrrrrrr
r(rcCs||kp||kSrrrrrr
r)rcCs||kSrrrrrr
r*rcCs||kp||kSrrrrrr
r+rcCs|o|Srrrrrr
r,rcCs|p|Srrrrrr
r-rcCs||kSrrrrrr
r.rcCs||kSrrrrrr
r/r)z==z===z~=z!=z>=andorinznot inc	Cst|trB|ddkr$|dd}q||kr8td|||}nt|tsPt|d}||jkrntd||d}|d	}t|drt|d	rtd
|||f|||}|||}|j|||}|S)z
        Evaluate a marker expression returned by the :func:`parse_requirement`
        function in the specified context.
        rr	rzunknown variable: %sopzop not implemented: %slhsrhszinvalid comparison: %s %s %s)	r
rSyntaxErrordictAssertionError
operationsNotImplementedErrorrevaluate)	selfexprcontextresultrZelhsZerhsrrrrr
r$2s$


zEvaluator.evaluateN)__name__
__module____qualname____doc__r"r$rrrr
rsrc
Csdd}ttdr(|tjj}tjj}nd}d}||tjttt	t
tttt
t
ddtjd}|S)NcSs<d|j|j|jf}|j}|dkr8||dt|j7}|S)Nz%s.%s.%sfinalr)majorminormicroreleaselevelstrserial)infoversionZkindrrr
format_full_versionNs
z,default_context..format_full_versionimplementation0)implementation_nameimplementation_versionZos_nameZplatform_machineZplatform_python_implementationZplatform_releaseZplatform_systemZplatform_versionZplatform_in_venvZpython_full_versionpython_versionZsys_platform)hasattrsysr7r5nameosplatformmachinerreleasesystemr2rr=)r6r<r;r(rrr
default_contextMs(

rFc
Cszt|\}}Wn2tk
rB}ztd||fW5d}~XYnX|rd|ddkrdtd||ftt}|rz||t||S)z
    Interpret a marker and return a result depending on environment.

    :param marker: The marker to interpret.
    :type marker: str
    :param execution_context: The context used for name lookup.
    :type execution_context: mapping
    z)Unable to interpret marker syntax: %s: %sNr#z*unexpected trailing data in marker: %s: %s)r	Exceptionrr DEFAULT_CONTEXTupdate	evaluatorr$)ZmarkerZexecution_contextr&rester'rrr
rqs	"
)N)r,rAr?rBrecompatrrrutilrr__all__robjectrrFrIrKrrrrr
s/site-packages/pip/_vendor/distlib/util.py000064400000165770151464452710014522 0ustar00#
# Copyright (C) 2012-2017 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
import codecs
from collections import deque
import contextlib
import csv
from glob import iglob as std_iglob
import io
import json
import logging
import os
import py_compile
import re
import socket
try:
    import ssl
except ImportError:  # pragma: no cover
    ssl = None
import subprocess
import sys
import tarfile
import tempfile
import textwrap

try:
    import threading
except ImportError:  # pragma: no cover
    import dummy_threading as threading
import time

from . import DistlibException
from .compat import (string_types, text_type, shutil, raw_input, StringIO,
                     cache_from_source, urlopen, urljoin, httplib, xmlrpclib,
                     splittype, HTTPHandler, BaseConfigurator, valid_ident,
                     Container, configparser, URLError, ZipFile, fsdecode,
                     unquote, urlparse)

logger = logging.getLogger(__name__)

#
# Requirement parsing code as per PEP 508
#

IDENTIFIER = re.compile(r'^([\w\.-]+)\s*')
VERSION_IDENTIFIER = re.compile(r'^([\w\.*+-]+)\s*')
COMPARE_OP = re.compile(r'^(<=?|>=?|={2,3}|[~!]=)\s*')
MARKER_OP = re.compile(r'^((<=?)|(>=?)|={2,3}|[~!]=|in|not\s+in)\s*')
OR = re.compile(r'^or\b\s*')
AND = re.compile(r'^and\b\s*')
NON_SPACE = re.compile(r'(\S+)\s*')
STRING_CHUNK = re.compile(r'([\s\w\.{}()*+#:;,/?!~`@$%^&=|<>\[\]-]+)')


def parse_marker(marker_string):
    """
    Parse a marker string and return a dictionary containing a marker expression.

    The dictionary will contain keys "op", "lhs" and "rhs" for non-terminals in
    the expression grammar, or strings. A string contained in quotes is to be
    interpreted as a literal string, and a string not contained in quotes is a
    variable (such as os_name).
    """
    def marker_var(remaining):
        # either identifier, or literal string
        m = IDENTIFIER.match(remaining)
        if m:
            result = m.groups()[0]
            remaining = remaining[m.end():]
        elif not remaining:
            raise SyntaxError('unexpected end of input')
        else:
            q = remaining[0]
            if q not in '\'"':
                raise SyntaxError('invalid expression: %s' % remaining)
            oq = '\'"'.replace(q, '')
            remaining = remaining[1:]
            parts = [q]
            while remaining:
                # either a string chunk, or oq, or q to terminate
                if remaining[0] == q:
                    break
                elif remaining[0] == oq:
                    parts.append(oq)
                    remaining = remaining[1:]
                else:
                    m = STRING_CHUNK.match(remaining)
                    if not m:
                        raise SyntaxError('error in string literal: %s' % remaining)
                    parts.append(m.groups()[0])
                    remaining = remaining[m.end():]
            else:
                s = ''.join(parts)
                raise SyntaxError('unterminated string: %s' % s)
            parts.append(q)
            result = ''.join(parts)
            remaining = remaining[1:].lstrip() # skip past closing quote
        return result, remaining

    def marker_expr(remaining):
        if remaining and remaining[0] == '(':
            result, remaining = marker(remaining[1:].lstrip())
            if remaining[0] != ')':
                raise SyntaxError('unterminated parenthesis: %s' % remaining)
            remaining = remaining[1:].lstrip()
        else:
            lhs, remaining = marker_var(remaining)
            while remaining:
                m = MARKER_OP.match(remaining)
                if not m:
                    break
                op = m.groups()[0]
                remaining = remaining[m.end():]
                rhs, remaining = marker_var(remaining)
                lhs = {'op': op, 'lhs': lhs, 'rhs': rhs}
            result = lhs
        return result, remaining

    def marker_and(remaining):
        lhs, remaining = marker_expr(remaining)
        while remaining:
            m = AND.match(remaining)
            if not m:
                break
            remaining = remaining[m.end():]
            rhs, remaining = marker_expr(remaining)
            lhs = {'op': 'and', 'lhs': lhs, 'rhs': rhs}
        return lhs, remaining

    def marker(remaining):
        lhs, remaining = marker_and(remaining)
        while remaining:
            m = OR.match(remaining)
            if not m:
                break
            remaining = remaining[m.end():]
            rhs, remaining = marker_and(remaining)
            lhs = {'op': 'or', 'lhs': lhs, 'rhs': rhs}
        return lhs, remaining

    return marker(marker_string)


def parse_requirement(req):
    """
    Parse a requirement passed in as a string. Return a Container
    whose attributes contain the various parts of the requirement.
    """
    remaining = req.strip()
    if not remaining or remaining.startswith('#'):
        return None
    m = IDENTIFIER.match(remaining)
    if not m:
        raise SyntaxError('name expected: %s' % remaining)
    distname = m.groups()[0]
    remaining = remaining[m.end():]
    extras = mark_expr = versions = uri = None
    if remaining and remaining[0] == '[':
        i = remaining.find(']', 1)
        if i < 0:
            raise SyntaxError('unterminated extra: %s' % remaining)
        s = remaining[1:i]
        remaining = remaining[i + 1:].lstrip()
        extras = []
        while s:
            m = IDENTIFIER.match(s)
            if not m:
                raise SyntaxError('malformed extra: %s' % s)
            extras.append(m.groups()[0])
            s = s[m.end():]
            if not s:
                break
            if s[0] != ',':
                raise SyntaxError('comma expected in extras: %s' % s)
            s = s[1:].lstrip()
        if not extras:
            extras = None
    if remaining:
        if remaining[0] == '@':
            # it's a URI
            remaining = remaining[1:].lstrip()
            m = NON_SPACE.match(remaining)
            if not m:
                raise SyntaxError('invalid URI: %s' % remaining)
            uri = m.groups()[0]
            t = urlparse(uri)
            # there are issues with Python and URL parsing, so this test
            # is a bit crude. See bpo-20271, bpo-23505. Python doesn't
            # always parse invalid URLs correctly - it should raise
            # exceptions for malformed URLs
            if not (t.scheme and t.netloc):
                raise SyntaxError('Invalid URL: %s' % uri)
            remaining = remaining[m.end():].lstrip()
        else:

            def get_versions(ver_remaining):
                """
                Return a list of operator, version tuples if any are
                specified, else None.
                """
                m = COMPARE_OP.match(ver_remaining)
                versions = None
                if m:
                    versions = []
                    while True:
                        op = m.groups()[0]
                        ver_remaining = ver_remaining[m.end():]
                        m = VERSION_IDENTIFIER.match(ver_remaining)
                        if not m:
                            raise SyntaxError('invalid version: %s' % ver_remaining)
                        v = m.groups()[0]
                        versions.append((op, v))
                        ver_remaining = ver_remaining[m.end():]
                        if not ver_remaining or ver_remaining[0] != ',':
                            break
                        ver_remaining = ver_remaining[1:].lstrip()
                        m = COMPARE_OP.match(ver_remaining)
                        if not m:
                            raise SyntaxError('invalid constraint: %s' % ver_remaining)
                    if not versions:
                        versions = None
                return versions, ver_remaining

            if remaining[0] != '(':
                versions, remaining = get_versions(remaining)
            else:
                i = remaining.find(')', 1)
                if i < 0:
                    raise SyntaxError('unterminated parenthesis: %s' % remaining)
                s = remaining[1:i]
                remaining = remaining[i + 1:].lstrip()
                # As a special diversion from PEP 508, allow a version number
                # a.b.c in parentheses as a synonym for ~= a.b.c (because this
                # is allowed in earlier PEPs)
                if COMPARE_OP.match(s):
                    versions, _ = get_versions(s)
                else:
                    m = VERSION_IDENTIFIER.match(s)
                    if not m:
                        raise SyntaxError('invalid constraint: %s' % s)
                    v = m.groups()[0]
                    s = s[m.end():].lstrip()
                    if s:
                        raise SyntaxError('invalid constraint: %s' % s)
                    versions = [('~=', v)]

    if remaining:
        if remaining[0] != ';':
            raise SyntaxError('invalid requirement: %s' % remaining)
        remaining = remaining[1:].lstrip()

        mark_expr, remaining = parse_marker(remaining)

    if remaining and remaining[0] != '#':
        raise SyntaxError('unexpected trailing data: %s' % remaining)

    if not versions:
        rs = distname
    else:
        rs = '%s %s' % (distname, ', '.join(['%s %s' % con for con in versions]))
    return Container(name=distname, extras=extras, constraints=versions,
                     marker=mark_expr, url=uri, requirement=rs)


def get_resources_dests(resources_root, rules):
    """Find destinations for resources files"""

    def get_rel_path(root, path):
        # normalizes and returns a lstripped-/-separated path
        root = root.replace(os.path.sep, '/')
        path = path.replace(os.path.sep, '/')
        assert path.startswith(root)
        return path[len(root):].lstrip('/')

    destinations = {}
    for base, suffix, dest in rules:
        prefix = os.path.join(resources_root, base)
        for abs_base in iglob(prefix):
            abs_glob = os.path.join(abs_base, suffix)
            for abs_path in iglob(abs_glob):
                resource_file = get_rel_path(resources_root, abs_path)
                if dest is None:  # remove the entry if it was here
                    destinations.pop(resource_file, None)
                else:
                    rel_path = get_rel_path(abs_base, abs_path)
                    rel_dest = dest.replace(os.path.sep, '/').rstrip('/')
                    destinations[resource_file] = rel_dest + '/' + rel_path
    return destinations


def in_venv():
    if hasattr(sys, 'real_prefix'):
        # virtualenv venvs
        result = True
    else:
        # PEP 405 venvs
        result = sys.prefix != getattr(sys, 'base_prefix', sys.prefix)
    return result


def get_executable():
# The __PYVENV_LAUNCHER__ dance is apparently no longer needed, as
# changes to the stub launcher mean that sys.executable always points
# to the stub on OS X
#    if sys.platform == 'darwin' and ('__PYVENV_LAUNCHER__'
#                                     in os.environ):
#        result =  os.environ['__PYVENV_LAUNCHER__']
#    else:
#        result = sys.executable
#    return result
    result = os.path.normcase(sys.executable)
    if not isinstance(result, text_type):
        result = fsdecode(result)
    return result


def proceed(prompt, allowed_chars, error_prompt=None, default=None):
    p = prompt
    while True:
        s = raw_input(p)
        p = prompt
        if not s and default:
            s = default
        if s:
            c = s[0].lower()
            if c in allowed_chars:
                break
            if error_prompt:
                p = '%c: %s\n%s' % (c, error_prompt, prompt)
    return c


def extract_by_key(d, keys):
    if isinstance(keys, string_types):
        keys = keys.split()
    result = {}
    for key in keys:
        if key in d:
            result[key] = d[key]
    return result

def read_exports(stream):
    if sys.version_info[0] >= 3:
        # needs to be a text stream
        stream = codecs.getreader('utf-8')(stream)
    # Try to load as JSON, falling back on legacy format
    data = stream.read()
    stream = StringIO(data)
    try:
        jdata = json.load(stream)
        result = jdata['extensions']['python.exports']['exports']
        for group, entries in result.items():
            for k, v in entries.items():
                s = '%s = %s' % (k, v)
                entry = get_export_entry(s)
                assert entry is not None
                entries[k] = entry
        return result
    except Exception:
        stream.seek(0, 0)

    def read_stream(cp, stream):
        if hasattr(cp, 'read_file'):
            cp.read_file(stream)
        else:
            cp.readfp(stream)

    cp = configparser.ConfigParser()
    try:
        read_stream(cp, stream)
    except configparser.MissingSectionHeaderError:
        stream.close()
        data = textwrap.dedent(data)
        stream = StringIO(data)
        read_stream(cp, stream)

    result = {}
    for key in cp.sections():
        result[key] = entries = {}
        for name, value in cp.items(key):
            s = '%s = %s' % (name, value)
            entry = get_export_entry(s)
            assert entry is not None
            #entry.dist = self
            entries[name] = entry
    return result


def write_exports(exports, stream):
    if sys.version_info[0] >= 3:
        # needs to be a text stream
        stream = codecs.getwriter('utf-8')(stream)
    cp = configparser.ConfigParser()
    for k, v in exports.items():
        # TODO check k, v for valid values
        cp.add_section(k)
        for entry in v.values():
            if entry.suffix is None:
                s = entry.prefix
            else:
                s = '%s:%s' % (entry.prefix, entry.suffix)
            if entry.flags:
                s = '%s [%s]' % (s, ', '.join(entry.flags))
            cp.set(k, entry.name, s)
    cp.write(stream)


@contextlib.contextmanager
def tempdir():
    td = tempfile.mkdtemp()
    try:
        yield td
    finally:
        shutil.rmtree(td)

@contextlib.contextmanager
def chdir(d):
    cwd = os.getcwd()
    try:
        os.chdir(d)
        yield
    finally:
        os.chdir(cwd)


@contextlib.contextmanager
def socket_timeout(seconds=15):
    cto = socket.getdefaulttimeout()
    try:
        socket.setdefaulttimeout(seconds)
        yield
    finally:
        socket.setdefaulttimeout(cto)


class cached_property(object):
    def __init__(self, func):
        self.func = func
        #for attr in ('__name__', '__module__', '__doc__'):
        #    setattr(self, attr, getattr(func, attr, None))

    def __get__(self, obj, cls=None):
        if obj is None:
            return self
        value = self.func(obj)
        object.__setattr__(obj, self.func.__name__, value)
        #obj.__dict__[self.func.__name__] = value = self.func(obj)
        return value

def convert_path(pathname):
    """Return 'pathname' as a name that will work on the native filesystem.

    The path is split on '/' and put back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    """
    if os.sep == '/':
        return pathname
    if not pathname:
        return pathname
    if pathname[0] == '/':
        raise ValueError("path '%s' cannot be absolute" % pathname)
    if pathname[-1] == '/':
        raise ValueError("path '%s' cannot end with '/'" % pathname)

    paths = pathname.split('/')
    while os.curdir in paths:
        paths.remove(os.curdir)
    if not paths:
        return os.curdir
    return os.path.join(*paths)


class FileOperator(object):
    def __init__(self, dry_run=False):
        self.dry_run = dry_run
        self.ensured = set()
        self._init_record()

    def _init_record(self):
        self.record = False
        self.files_written = set()
        self.dirs_created = set()

    def record_as_written(self, path):
        if self.record:
            self.files_written.add(path)

    def newer(self, source, target):
        """Tell if the target is newer than the source.

        Returns true if 'source' exists and is more recently modified than
        'target', or if 'source' exists and 'target' doesn't.

        Returns false if both exist and 'target' is the same age or younger
        than 'source'. Raise PackagingFileError if 'source' does not exist.

        Note that this test is not very accurate: files created in the same
        second will have the same "age".
        """
        if not os.path.exists(source):
            raise DistlibException("file '%r' does not exist" %
                                   os.path.abspath(source))
        if not os.path.exists(target):
            return True

        return os.stat(source).st_mtime > os.stat(target).st_mtime

    def copy_file(self, infile, outfile, check=True):
        """Copy a file respecting dry-run and force flags.
        """
        self.ensure_dir(os.path.dirname(outfile))
        logger.info('Copying %s to %s', infile, outfile)
        if not self.dry_run:
            msg = None
            if check:
                if os.path.islink(outfile):
                    msg = '%s is a symlink' % outfile
                elif os.path.exists(outfile) and not os.path.isfile(outfile):
                    msg = '%s is a non-regular file' % outfile
            if msg:
                raise ValueError(msg + ' which would be overwritten')
            shutil.copyfile(infile, outfile)
        self.record_as_written(outfile)

    def copy_stream(self, instream, outfile, encoding=None):
        assert not os.path.isdir(outfile)
        self.ensure_dir(os.path.dirname(outfile))
        logger.info('Copying stream %s to %s', instream, outfile)
        if not self.dry_run:
            if encoding is None:
                outstream = open(outfile, 'wb')
            else:
                outstream = codecs.open(outfile, 'w', encoding=encoding)
            try:
                shutil.copyfileobj(instream, outstream)
            finally:
                outstream.close()
        self.record_as_written(outfile)

    def write_binary_file(self, path, data):
        self.ensure_dir(os.path.dirname(path))
        if not self.dry_run:
            if os.path.exists(path):
                os.remove(path)
            with open(path, 'wb') as f:
                f.write(data)
        self.record_as_written(path)

    def write_text_file(self, path, data, encoding):
        self.write_binary_file(path, data.encode(encoding))

    def set_mode(self, bits, mask, files):
        if os.name == 'posix' or (os.name == 'java' and os._name == 'posix'):
            # Set the executable bits (owner, group, and world) on
            # all the files specified.
            for f in files:
                if self.dry_run:
                    logger.info("changing mode of %s", f)
                else:
                    mode = (os.stat(f).st_mode | bits) & mask
                    logger.info("changing mode of %s to %o", f, mode)
                    os.chmod(f, mode)

    set_executable_mode = lambda s, f: s.set_mode(0o555, 0o7777, f)

    def ensure_dir(self, path):
        path = os.path.abspath(path)
        if path not in self.ensured and not os.path.exists(path):
            self.ensured.add(path)
            d, f = os.path.split(path)
            self.ensure_dir(d)
            logger.info('Creating %s' % path)
            if not self.dry_run:
                os.mkdir(path)
            if self.record:
                self.dirs_created.add(path)

    def byte_compile(self, path, optimize=False, force=False, prefix=None, hashed_invalidation=False):
        dpath = cache_from_source(path, not optimize)
        logger.info('Byte-compiling %s to %s', path, dpath)
        if not self.dry_run:
            if force or self.newer(path, dpath):
                if not prefix:
                    diagpath = None
                else:
                    assert path.startswith(prefix)
                    diagpath = path[len(prefix):]
            compile_kwargs = {}
            if hashed_invalidation and hasattr(py_compile, 'PycInvalidationMode'):
                compile_kwargs['invalidation_mode'] = py_compile.PycInvalidationMode.CHECKED_HASH
            py_compile.compile(path, dpath, diagpath, True, **compile_kwargs)     # raise error
        self.record_as_written(dpath)
        return dpath

    def ensure_removed(self, path):
        if os.path.exists(path):
            if os.path.isdir(path) and not os.path.islink(path):
                logger.debug('Removing directory tree at %s', path)
                if not self.dry_run:
                    shutil.rmtree(path)
                if self.record:
                    if path in self.dirs_created:
                        self.dirs_created.remove(path)
            else:
                if os.path.islink(path):
                    s = 'link'
                else:
                    s = 'file'
                logger.debug('Removing %s %s', s, path)
                if not self.dry_run:
                    os.remove(path)
                if self.record:
                    if path in self.files_written:
                        self.files_written.remove(path)

    def is_writable(self, path):
        result = False
        while not result:
            if os.path.exists(path):
                result = os.access(path, os.W_OK)
                break
            parent = os.path.dirname(path)
            if parent == path:
                break
            path = parent
        return result

    def commit(self):
        """
        Commit recorded changes, turn off recording, return
        changes.
        """
        assert self.record
        result = self.files_written, self.dirs_created
        self._init_record()
        return result

    def rollback(self):
        if not self.dry_run:
            for f in list(self.files_written):
                if os.path.exists(f):
                    os.remove(f)
            # dirs should all be empty now, except perhaps for
            # __pycache__ subdirs
            # reverse so that subdirs appear before their parents
            dirs = sorted(self.dirs_created, reverse=True)
            for d in dirs:
                flist = os.listdir(d)
                if flist:
                    assert flist == ['__pycache__']
                    sd = os.path.join(d, flist[0])
                    os.rmdir(sd)
                os.rmdir(d)     # should fail if non-empty
        self._init_record()

def resolve(module_name, dotted_path):
    if module_name in sys.modules:
        mod = sys.modules[module_name]
    else:
        mod = __import__(module_name)
    if dotted_path is None:
        result = mod
    else:
        parts = dotted_path.split('.')
        result = getattr(mod, parts.pop(0))
        for p in parts:
            result = getattr(result, p)
    return result


class ExportEntry(object):
    def __init__(self, name, prefix, suffix, flags):
        self.name = name
        self.prefix = prefix
        self.suffix = suffix
        self.flags = flags

    @cached_property
    def value(self):
        return resolve(self.prefix, self.suffix)

    def __repr__(self):  # pragma: no cover
        return '' % (self.name, self.prefix,
                                                self.suffix, self.flags)

    def __eq__(self, other):
        if not isinstance(other, ExportEntry):
            result = False
        else:
            result = (self.name == other.name and
                      self.prefix == other.prefix and
                      self.suffix == other.suffix and
                      self.flags == other.flags)
        return result

    __hash__ = object.__hash__


ENTRY_RE = re.compile(r'''(?P(\w|[-.+])+)
                      \s*=\s*(?P(\w+)([:\.]\w+)*)
                      \s*(\[\s*(?P\w+(=\w+)?(,\s*\w+(=\w+)?)*)\s*\])?
                      ''', re.VERBOSE)

def get_export_entry(specification):
    m = ENTRY_RE.search(specification)
    if not m:
        result = None
        if '[' in specification or ']' in specification:
            raise DistlibException("Invalid specification "
                                   "'%s'" % specification)
    else:
        d = m.groupdict()
        name = d['name']
        path = d['callable']
        colons = path.count(':')
        if colons == 0:
            prefix, suffix = path, None
        else:
            if colons != 1:
                raise DistlibException("Invalid specification "
                                       "'%s'" % specification)
            prefix, suffix = path.split(':')
        flags = d['flags']
        if flags is None:
            if '[' in specification or ']' in specification:
                raise DistlibException("Invalid specification "
                                       "'%s'" % specification)
            flags = []
        else:
            flags = [f.strip() for f in flags.split(',')]
        result = ExportEntry(name, prefix, suffix, flags)
    return result


def get_cache_base(suffix=None):
    """
    Return the default base location for distlib caches. If the directory does
    not exist, it is created. Use the suffix provided for the base directory,
    and default to '.distlib' if it isn't provided.

    On Windows, if LOCALAPPDATA is defined in the environment, then it is
    assumed to be a directory, and will be the parent directory of the result.
    On POSIX, and on Windows if LOCALAPPDATA is not defined, the user's home
    directory - using os.expanduser('~') - will be the parent directory of
    the result.

    The result is just the directory '.distlib' in the parent directory as
    determined above, or with the name specified with ``suffix``.
    """
    if suffix is None:
        suffix = '.distlib'
    if os.name == 'nt' and 'LOCALAPPDATA' in os.environ:
        result = os.path.expandvars('$localappdata')
    else:
        # Assume posix, or old Windows
        result = os.path.expanduser('~')
    # we use 'isdir' instead of 'exists', because we want to
    # fail if there's a file with that name
    if os.path.isdir(result):
        usable = os.access(result, os.W_OK)
        if not usable:
            logger.warning('Directory exists but is not writable: %s', result)
    else:
        try:
            os.makedirs(result)
            usable = True
        except OSError:
            logger.warning('Unable to create %s', result, exc_info=True)
            usable = False
    if not usable:
        result = tempfile.mkdtemp()
        logger.warning('Default location unusable, using %s', result)
    return os.path.join(result, suffix)


def path_to_cache_dir(path):
    """
    Convert an absolute path to a directory name for use in a cache.

    The algorithm used is:

    #. On Windows, any ``':'`` in the drive is replaced with ``'---'``.
    #. Any occurrence of ``os.sep`` is replaced with ``'--'``.
    #. ``'.cache'`` is appended.
    """
    d, p = os.path.splitdrive(os.path.abspath(path))
    if d:
        d = d.replace(':', '---')
    p = p.replace(os.sep, '--')
    return d + p + '.cache'


def ensure_slash(s):
    if not s.endswith('/'):
        return s + '/'
    return s


def parse_credentials(netloc):
    username = password = None
    if '@' in netloc:
        prefix, netloc = netloc.rsplit('@', 1)
        if ':' not in prefix:
            username = prefix
        else:
            username, password = prefix.split(':', 1)
    if username:
        username = unquote(username)
    if password:
        password = unquote(password)
    return username, password, netloc


def get_process_umask():
    result = os.umask(0o22)
    os.umask(result)
    return result

def is_string_sequence(seq):
    result = True
    i = None
    for i, s in enumerate(seq):
        if not isinstance(s, string_types):
            result = False
            break
    assert i is not None
    return result

PROJECT_NAME_AND_VERSION = re.compile('([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-'
                                      '([a-z0-9_.+-]+)', re.I)
PYTHON_VERSION = re.compile(r'-py(\d\.?\d?)')


def split_filename(filename, project_name=None):
    """
    Extract name, version, python version from a filename (no extension)

    Return name, version, pyver or None
    """
    result = None
    pyver = None
    filename = unquote(filename).replace(' ', '-')
    m = PYTHON_VERSION.search(filename)
    if m:
        pyver = m.group(1)
        filename = filename[:m.start()]
    if project_name and len(filename) > len(project_name) + 1:
        m = re.match(re.escape(project_name) + r'\b', filename)
        if m:
            n = m.end()
            result = filename[:n], filename[n + 1:], pyver
    if result is None:
        m = PROJECT_NAME_AND_VERSION.match(filename)
        if m:
            result = m.group(1), m.group(3), pyver
    return result

# Allow spaces in name because of legacy dists like "Twisted Core"
NAME_VERSION_RE = re.compile(r'(?P[\w .-]+)\s*'
                             r'\(\s*(?P[^\s)]+)\)$')

def parse_name_and_version(p):
    """
    A utility method used to get name and version from a string.

    From e.g. a Provides-Dist value.

    :param p: A value in a form 'foo (1.0)'
    :return: The name and version as a tuple.
    """
    m = NAME_VERSION_RE.match(p)
    if not m:
        raise DistlibException('Ill-formed name/version string: \'%s\'' % p)
    d = m.groupdict()
    return d['name'].strip().lower(), d['ver']

def get_extras(requested, available):
    result = set()
    requested = set(requested or [])
    available = set(available or [])
    if '*' in requested:
        requested.remove('*')
        result |= available
    for r in requested:
        if r == '-':
            result.add(r)
        elif r.startswith('-'):
            unwanted = r[1:]
            if unwanted not in available:
                logger.warning('undeclared extra: %s' % unwanted)
            if unwanted in result:
                result.remove(unwanted)
        else:
            if r not in available:
                logger.warning('undeclared extra: %s' % r)
            result.add(r)
    return result
#
# Extended metadata functionality
#

def _get_external_data(url):
    result = {}
    try:
        # urlopen might fail if it runs into redirections,
        # because of Python issue #13696. Fixed in locators
        # using a custom redirect handler.
        resp = urlopen(url)
        headers = resp.info()
        ct = headers.get('Content-Type')
        if not ct.startswith('application/json'):
            logger.debug('Unexpected response for JSON request: %s', ct)
        else:
            reader = codecs.getreader('utf-8')(resp)
            #data = reader.read().decode('utf-8')
            #result = json.loads(data)
            result = json.load(reader)
    except Exception as e:
        logger.exception('Failed to get external data for %s: %s', url, e)
    return result

_external_data_base_url = 'https://www.red-dove.com/pypi/projects/'

def get_project_data(name):
    url = '%s/%s/project.json' % (name[0].upper(), name)
    url = urljoin(_external_data_base_url, url)
    result = _get_external_data(url)
    return result

def get_package_data(name, version):
    url = '%s/%s/package-%s.json' % (name[0].upper(), name, version)
    url = urljoin(_external_data_base_url, url)
    return _get_external_data(url)


class Cache(object):
    """
    A class implementing a cache for resources that need to live in the file system
    e.g. shared libraries. This class was moved from resources to here because it
    could be used by other modules, e.g. the wheel module.
    """

    def __init__(self, base):
        """
        Initialise an instance.

        :param base: The base directory where the cache should be located.
        """
        # we use 'isdir' instead of 'exists', because we want to
        # fail if there's a file with that name
        if not os.path.isdir(base):  # pragma: no cover
            os.makedirs(base)
        if (os.stat(base).st_mode & 0o77) != 0:
            logger.warning('Directory \'%s\' is not private', base)
        self.base = os.path.abspath(os.path.normpath(base))

    def prefix_to_dir(self, prefix):
        """
        Converts a resource prefix to a directory name in the cache.
        """
        return path_to_cache_dir(prefix)

    def clear(self):
        """
        Clear the cache.
        """
        not_removed = []
        for fn in os.listdir(self.base):
            fn = os.path.join(self.base, fn)
            try:
                if os.path.islink(fn) or os.path.isfile(fn):
                    os.remove(fn)
                elif os.path.isdir(fn):
                    shutil.rmtree(fn)
            except Exception:
                not_removed.append(fn)
        return not_removed


class EventMixin(object):
    """
    A very simple publish/subscribe system.
    """
    def __init__(self):
        self._subscribers = {}

    def add(self, event, subscriber, append=True):
        """
        Add a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be added (and called when the
                           event is published).
        :param append: Whether to append or prepend the subscriber to an
                       existing subscriber list for the event.
        """
        subs = self._subscribers
        if event not in subs:
            subs[event] = deque([subscriber])
        else:
            sq = subs[event]
            if append:
                sq.append(subscriber)
            else:
                sq.appendleft(subscriber)

    def remove(self, event, subscriber):
        """
        Remove a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be removed.
        """
        subs = self._subscribers
        if event not in subs:
            raise ValueError('No subscribers: %r' % event)
        subs[event].remove(subscriber)

    def get_subscribers(self, event):
        """
        Return an iterator for the subscribers for an event.
        :param event: The event to return subscribers for.
        """
        return iter(self._subscribers.get(event, ()))

    def publish(self, event, *args, **kwargs):
        """
        Publish a event and return a list of values returned by its
        subscribers.

        :param event: The event to publish.
        :param args: The positional arguments to pass to the event's
                     subscribers.
        :param kwargs: The keyword arguments to pass to the event's
                       subscribers.
        """
        result = []
        for subscriber in self.get_subscribers(event):
            try:
                value = subscriber(event, *args, **kwargs)
            except Exception:
                logger.exception('Exception during event publication')
                value = None
            result.append(value)
        logger.debug('publish %s: args = %s, kwargs = %s, result = %s',
                     event, args, kwargs, result)
        return result

#
# Simple sequencing
#
class Sequencer(object):
    def __init__(self):
        self._preds = {}
        self._succs = {}
        self._nodes = set()     # nodes with no preds/succs

    def add_node(self, node):
        self._nodes.add(node)

    def remove_node(self, node, edges=False):
        if node in self._nodes:
            self._nodes.remove(node)
        if edges:
            for p in set(self._preds.get(node, ())):
                self.remove(p, node)
            for s in set(self._succs.get(node, ())):
                self.remove(node, s)
            # Remove empties
            for k, v in list(self._preds.items()):
                if not v:
                    del self._preds[k]
            for k, v in list(self._succs.items()):
                if not v:
                    del self._succs[k]

    def add(self, pred, succ):
        assert pred != succ
        self._preds.setdefault(succ, set()).add(pred)
        self._succs.setdefault(pred, set()).add(succ)

    def remove(self, pred, succ):
        assert pred != succ
        try:
            preds = self._preds[succ]
            succs = self._succs[pred]
        except KeyError:  # pragma: no cover
            raise ValueError('%r not a successor of anything' % succ)
        try:
            preds.remove(pred)
            succs.remove(succ)
        except KeyError:  # pragma: no cover
            raise ValueError('%r not a successor of %r' % (succ, pred))

    def is_step(self, step):
        return (step in self._preds or step in self._succs or
                step in self._nodes)

    def get_steps(self, final):
        if not self.is_step(final):
            raise ValueError('Unknown: %r' % final)
        result = []
        todo = []
        seen = set()
        todo.append(final)
        while todo:
            step = todo.pop(0)
            if step in seen:
                # if a step was already seen,
                # move it to the end (so it will appear earlier
                # when reversed on return) ... but not for the
                # final step, as that would be confusing for
                # users
                if step != final:
                    result.remove(step)
                    result.append(step)
            else:
                seen.add(step)
                result.append(step)
                preds = self._preds.get(step, ())
                todo.extend(preds)
        return reversed(result)

    @property
    def strong_connections(self):
        #http://en.wikipedia.org/wiki/Tarjan%27s_strongly_connected_components_algorithm
        index_counter = [0]
        stack = []
        lowlinks = {}
        index = {}
        result = []

        graph = self._succs

        def strongconnect(node):
            # set the depth index for this node to the smallest unused index
            index[node] = index_counter[0]
            lowlinks[node] = index_counter[0]
            index_counter[0] += 1
            stack.append(node)

            # Consider successors
            try:
                successors = graph[node]
            except Exception:
                successors = []
            for successor in successors:
                if successor not in lowlinks:
                    # Successor has not yet been visited
                    strongconnect(successor)
                    lowlinks[node] = min(lowlinks[node],lowlinks[successor])
                elif successor in stack:
                    # the successor is in the stack and hence in the current
                    # strongly connected component (SCC)
                    lowlinks[node] = min(lowlinks[node],index[successor])

            # If `node` is a root node, pop the stack and generate an SCC
            if lowlinks[node] == index[node]:
                connected_component = []

                while True:
                    successor = stack.pop()
                    connected_component.append(successor)
                    if successor == node: break
                component = tuple(connected_component)
                # storing the result
                result.append(component)

        for node in graph:
            if node not in lowlinks:
                strongconnect(node)

        return result

    @property
    def dot(self):
        result = ['digraph G {']
        for succ in self._preds:
            preds = self._preds[succ]
            for pred in preds:
                result.append('  %s -> %s;' % (pred, succ))
        for node in self._nodes:
            result.append('  %s;' % node)
        result.append('}')
        return '\n'.join(result)

#
# Unarchiving functionality for zip, tar, tgz, tbz, whl
#

ARCHIVE_EXTENSIONS = ('.tar.gz', '.tar.bz2', '.tar', '.zip',
                      '.tgz', '.tbz', '.whl')

def unarchive(archive_filename, dest_dir, format=None, check=True):

    def check_path(path):
        if not isinstance(path, text_type):
            path = path.decode('utf-8')
        p = os.path.abspath(os.path.join(dest_dir, path))
        if not p.startswith(dest_dir) or p[plen] != os.sep:
            raise ValueError('path outside destination: %r' % p)

    dest_dir = os.path.abspath(dest_dir)
    plen = len(dest_dir)
    archive = None
    if format is None:
        if archive_filename.endswith(('.zip', '.whl')):
            format = 'zip'
        elif archive_filename.endswith(('.tar.gz', '.tgz')):
            format = 'tgz'
            mode = 'r:gz'
        elif archive_filename.endswith(('.tar.bz2', '.tbz')):
            format = 'tbz'
            mode = 'r:bz2'
        elif archive_filename.endswith('.tar'):
            format = 'tar'
            mode = 'r'
        else:  # pragma: no cover
            raise ValueError('Unknown format for %r' % archive_filename)
    try:
        if format == 'zip':
            archive = ZipFile(archive_filename, 'r')
            if check:
                names = archive.namelist()
                for name in names:
                    check_path(name)
        else:
            archive = tarfile.open(archive_filename, mode)
            if check:
                names = archive.getnames()
                for name in names:
                    check_path(name)
        if format != 'zip' and sys.version_info[0] < 3:
            # See Python issue 17153. If the dest path contains Unicode,
            # tarfile extraction fails on Python 2.x if a member path name
            # contains non-ASCII characters - it leads to an implicit
            # bytes -> unicode conversion using ASCII to decode.
            for tarinfo in archive.getmembers():
                if not isinstance(tarinfo.name, text_type):
                    tarinfo.name = tarinfo.name.decode('utf-8')

        # Limit extraction of dangerous items, if this Python
        # allows it easily. If not, just trust the input.
        # See: https://docs.python.org/3/library/tarfile.html#extraction-filters
        def extraction_filter(member, path):
            """Run tarfile.tar_fillter, but raise the expected ValueError"""
            # This is only called if the current Python has tarfile filters
            try:
                return tarfile.tar_filter(member, path)
            except tarfile.FilterError as exc:
                raise ValueError(str(exc))
        archive.extraction_filter = extraction_filter

        archive.extractall(dest_dir)

    finally:
        if archive:
            archive.close()


def zip_dir(directory):
    """zip a directory tree into a BytesIO object"""
    result = io.BytesIO()
    dlen = len(directory)
    with ZipFile(result, "w") as zf:
        for root, dirs, files in os.walk(directory):
            for name in files:
                full = os.path.join(root, name)
                rel = root[dlen:]
                dest = os.path.join(rel, name)
                zf.write(full, dest)
    return result

#
# Simple progress bar
#

UNITS = ('', 'K', 'M', 'G','T','P')


class Progress(object):
    unknown = 'UNKNOWN'

    def __init__(self, minval=0, maxval=100):
        assert maxval is None or maxval >= minval
        self.min = self.cur = minval
        self.max = maxval
        self.started = None
        self.elapsed = 0
        self.done = False

    def update(self, curval):
        assert self.min <= curval
        assert self.max is None or curval <= self.max
        self.cur = curval
        now = time.time()
        if self.started is None:
            self.started = now
        else:
            self.elapsed = now - self.started

    def increment(self, incr):
        assert incr >= 0
        self.update(self.cur + incr)

    def start(self):
        self.update(self.min)
        return self

    def stop(self):
        if self.max is not None:
            self.update(self.max)
        self.done = True

    @property
    def maximum(self):
        return self.unknown if self.max is None else self.max

    @property
    def percentage(self):
        if self.done:
            result = '100 %'
        elif self.max is None:
            result = ' ?? %'
        else:
            v = 100.0 * (self.cur - self.min) / (self.max - self.min)
            result = '%3d %%' % v
        return result

    def format_duration(self, duration):
        if (duration <= 0) and self.max is None or self.cur == self.min:
            result = '??:??:??'
        #elif duration < 1:
        #    result = '--:--:--'
        else:
            result = time.strftime('%H:%M:%S', time.gmtime(duration))
        return result

    @property
    def ETA(self):
        if self.done:
            prefix = 'Done'
            t = self.elapsed
            #import pdb; pdb.set_trace()
        else:
            prefix = 'ETA '
            if self.max is None:
                t = -1
            elif self.elapsed == 0 or (self.cur == self.min):
                t = 0
            else:
                #import pdb; pdb.set_trace()
                t = float(self.max - self.min)
                t /= self.cur - self.min
                t = (t - 1) * self.elapsed
        return '%s: %s' % (prefix, self.format_duration(t))

    @property
    def speed(self):
        if self.elapsed == 0:
            result = 0.0
        else:
            result = (self.cur - self.min) / self.elapsed
        for unit in UNITS:
            if result < 1000:
                break
            result /= 1000.0
        return '%d %sB/s' % (result, unit)

#
# Glob functionality
#

RICH_GLOB = re.compile(r'\{([^}]*)\}')
_CHECK_RECURSIVE_GLOB = re.compile(r'[^/\\,{]\*\*|\*\*[^/\\,}]')
_CHECK_MISMATCH_SET = re.compile(r'^[^{]*\}|\{[^}]*$')


def iglob(path_glob):
    """Extended globbing function that supports ** and {opt1,opt2,opt3}."""
    if _CHECK_RECURSIVE_GLOB.search(path_glob):
        msg = """invalid glob %r: recursive glob "**" must be used alone"""
        raise ValueError(msg % path_glob)
    if _CHECK_MISMATCH_SET.search(path_glob):
        msg = """invalid glob %r: mismatching set marker '{' or '}'"""
        raise ValueError(msg % path_glob)
    return _iglob(path_glob)


def _iglob(path_glob):
    rich_path_glob = RICH_GLOB.split(path_glob, 1)
    if len(rich_path_glob) > 1:
        assert len(rich_path_glob) == 3, rich_path_glob
        prefix, set, suffix = rich_path_glob
        for item in set.split(','):
            for path in _iglob(''.join((prefix, item, suffix))):
                yield path
    else:
        if '**' not in path_glob:
            for item in std_iglob(path_glob):
                yield item
        else:
            prefix, radical = path_glob.split('**', 1)
            if prefix == '':
                prefix = '.'
            if radical == '':
                radical = '*'
            else:
                # we support both
                radical = radical.lstrip('/')
                radical = radical.lstrip('\\')
            for path, dir, files in os.walk(prefix):
                path = os.path.normpath(path)
                for fn in _iglob(os.path.join(path, radical)):
                    yield fn

if ssl:
    from .compat import (HTTPSHandler as BaseHTTPSHandler, match_hostname,
                         CertificateError)


#
# HTTPSConnection which verifies certificates/matches domains
#

    class HTTPSConnection(httplib.HTTPSConnection):
        ca_certs = None # set this to the path to the certs file (.pem)
        check_domain = True # only used if ca_certs is not None

        # noinspection PyPropertyAccess
        def connect(self):
            sock = socket.create_connection((self.host, self.port), self.timeout)
            if getattr(self, '_tunnel_host', False):
                self.sock = sock
                self._tunnel()

            if not hasattr(ssl, 'SSLContext'):
                # For 2.x
                if self.ca_certs:
                    cert_reqs = ssl.CERT_REQUIRED
                else:
                    cert_reqs = ssl.CERT_NONE
                self.sock = ssl.wrap_socket(sock, self.key_file, self.cert_file,
                                            cert_reqs=cert_reqs,
                                            ssl_version=ssl.PROTOCOL_SSLv23,
                                            ca_certs=self.ca_certs)
            else:  # pragma: no cover
                context = ssl.SSLContext(ssl.PROTOCOL_SSLv23)
                context.options |= ssl.OP_NO_SSLv2
                if self.cert_file:
                    context.load_cert_chain(self.cert_file, self.key_file)
                kwargs = {}
                if self.ca_certs:
                    context.verify_mode = ssl.CERT_REQUIRED
                    context.load_verify_locations(cafile=self.ca_certs)
                    if getattr(ssl, 'HAS_SNI', False):
                        kwargs['server_hostname'] = self.host
                self.sock = context.wrap_socket(sock, **kwargs)
            if self.ca_certs and self.check_domain:
                try:
                    match_hostname(self.sock.getpeercert(), self.host)
                    logger.debug('Host verified: %s', self.host)
                except CertificateError:  # pragma: no cover
                    self.sock.shutdown(socket.SHUT_RDWR)
                    self.sock.close()
                    raise

    class HTTPSHandler(BaseHTTPSHandler):
        def __init__(self, ca_certs, check_domain=True):
            BaseHTTPSHandler.__init__(self)
            self.ca_certs = ca_certs
            self.check_domain = check_domain

        def _conn_maker(self, *args, **kwargs):
            """
            This is called to create a connection instance. Normally you'd
            pass a connection class to do_open, but it doesn't actually check for
            a class, and just expects a callable. As long as we behave just as a
            constructor would have, we should be OK. If it ever changes so that
            we *must* pass a class, we'll create an UnsafeHTTPSConnection class
            which just sets check_domain to False in the class definition, and
            choose which one to pass to do_open.
            """
            result = HTTPSConnection(*args, **kwargs)
            if self.ca_certs:
                result.ca_certs = self.ca_certs
                result.check_domain = self.check_domain
            return result

        def https_open(self, req):
            try:
                return self.do_open(self._conn_maker, req)
            except URLError as e:
                if 'certificate verify failed' in str(e.reason):
                    raise CertificateError('Unable to verify server certificate '
                                           'for %s' % req.host)
                else:
                    raise

    #
    # To prevent against mixing HTTP traffic with HTTPS (examples: A Man-In-The-
    # Middle proxy using HTTP listens on port 443, or an index mistakenly serves
    # HTML containing a http://xyz link when it should be https://xyz),
    # you can use the following handler class, which does not allow HTTP traffic.
    #
    # It works by inheriting from HTTPHandler - so build_opener won't add a
    # handler for HTTP itself.
    #
    class HTTPSOnlyHandler(HTTPSHandler, HTTPHandler):
        def http_open(self, req):
            raise URLError('Unexpected HTTP request on what should be a secure '
                           'connection: %s' % req)

#
# XML-RPC with timeouts
#

_ver_info = sys.version_info[:2]

if _ver_info == (2, 6):
    class HTTP(httplib.HTTP):
        def __init__(self, host='', port=None, **kwargs):
            if port == 0:   # 0 means use port 0, not the default port
                port = None
            self._setup(self._connection_class(host, port, **kwargs))


    if ssl:
        class HTTPS(httplib.HTTPS):
            def __init__(self, host='', port=None, **kwargs):
                if port == 0:   # 0 means use port 0, not the default port
                    port = None
                self._setup(self._connection_class(host, port, **kwargs))


class Transport(xmlrpclib.Transport):
    def __init__(self, timeout, use_datetime=0):
        self.timeout = timeout
        xmlrpclib.Transport.__init__(self, use_datetime)

    def make_connection(self, host):
        h, eh, x509 = self.get_host_info(host)
        if _ver_info == (2, 6):
            result = HTTP(h, timeout=self.timeout)
        else:
            if not self._connection or host != self._connection[0]:
                self._extra_headers = eh
                self._connection = host, httplib.HTTPConnection(h)
            result = self._connection[1]
        return result

if ssl:
    class SafeTransport(xmlrpclib.SafeTransport):
        def __init__(self, timeout, use_datetime=0):
            self.timeout = timeout
            xmlrpclib.SafeTransport.__init__(self, use_datetime)

        def make_connection(self, host):
            h, eh, kwargs = self.get_host_info(host)
            if not kwargs:
                kwargs = {}
            kwargs['timeout'] = self.timeout
            if _ver_info == (2, 6):
                result = HTTPS(host, None, **kwargs)
            else:
                if not self._connection or host != self._connection[0]:
                    self._extra_headers = eh
                    self._connection = host, httplib.HTTPSConnection(h, None,
                                                                     **kwargs)
                result = self._connection[1]
            return result


class ServerProxy(xmlrpclib.ServerProxy):
    def __init__(self, uri, **kwargs):
        self.timeout = timeout = kwargs.pop('timeout', None)
        # The above classes only come into play if a timeout
        # is specified
        if timeout is not None:
            scheme, _ = splittype(uri)
            use_datetime = kwargs.get('use_datetime', 0)
            if scheme == 'https':
                tcls = SafeTransport
            else:
                tcls = Transport
            kwargs['transport'] = t = tcls(timeout, use_datetime=use_datetime)
            self.transport = t
        xmlrpclib.ServerProxy.__init__(self, uri, **kwargs)

#
# CSV functionality. This is provided because on 2.x, the csv module can't
# handle Unicode. However, we need to deal with Unicode in e.g. RECORD files.
#

def _csv_open(fn, mode, **kwargs):
    if sys.version_info[0] < 3:
        mode += 'b'
    else:
        kwargs['newline'] = ''
        # Python 3 determines encoding from locale. Force 'utf-8'
        # file encoding to match other forced utf-8 encoding
        kwargs['encoding'] = 'utf-8'
    return open(fn, mode, **kwargs)


class CSVBase(object):
    defaults = {
        'delimiter': str(','),      # The strs are used because we need native
        'quotechar': str('"'),      # str in the csv API (2.x won't take
        'lineterminator': str('\n') # Unicode)
    }

    def __enter__(self):
        return self

    def __exit__(self, *exc_info):
        self.stream.close()


class CSVReader(CSVBase):
    def __init__(self, **kwargs):
        if 'stream' in kwargs:
            stream = kwargs['stream']
            if sys.version_info[0] >= 3:
                # needs to be a text stream
                stream = codecs.getreader('utf-8')(stream)
            self.stream = stream
        else:
            self.stream = _csv_open(kwargs['path'], 'r')
        self.reader = csv.reader(self.stream, **self.defaults)

    def __iter__(self):
        return self

    def next(self):
        result = next(self.reader)
        if sys.version_info[0] < 3:
            for i, item in enumerate(result):
                if not isinstance(item, text_type):
                    result[i] = item.decode('utf-8')
        return result

    __next__ = next

class CSVWriter(CSVBase):
    def __init__(self, fn, **kwargs):
        self.stream = _csv_open(fn, 'w')
        self.writer = csv.writer(self.stream, **self.defaults)

    def writerow(self, row):
        if sys.version_info[0] < 3:
            r = []
            for item in row:
                if isinstance(item, text_type):
                    item = item.encode('utf-8')
                r.append(item)
            row = r
        self.writer.writerow(row)

#
#   Configurator functionality
#

class Configurator(BaseConfigurator):

    value_converters = dict(BaseConfigurator.value_converters)
    value_converters['inc'] = 'inc_convert'

    def __init__(self, config, base=None):
        super(Configurator, self).__init__(config)
        self.base = base or os.getcwd()

    def configure_custom(self, config):
        def convert(o):
            if isinstance(o, (list, tuple)):
                result = type(o)([convert(i) for i in o])
            elif isinstance(o, dict):
                if '()' in o:
                    result = self.configure_custom(o)
                else:
                    result = {}
                    for k in o:
                        result[k] = convert(o[k])
            else:
                result = self.convert(o)
            return result

        c = config.pop('()')
        if not callable(c):
            c = self.resolve(c)
        props = config.pop('.', None)
        # Check for valid identifiers
        args = config.pop('[]', ())
        if args:
            args = tuple([convert(o) for o in args])
        items = [(k, convert(config[k])) for k in config if valid_ident(k)]
        kwargs = dict(items)
        result = c(*args, **kwargs)
        if props:
            for n, v in props.items():
                setattr(result, n, convert(v))
        return result

    def __getitem__(self, key):
        result = self.config[key]
        if isinstance(result, dict) and '()' in result:
            self.config[key] = result = self.configure_custom(result)
        return result

    def inc_convert(self, value):
        """Default converter for the inc:// protocol."""
        if not os.path.isabs(value):
            value = os.path.join(self.base, value)
        with codecs.open(value, 'r', encoding='utf-8') as f:
            result = json.load(f)
        return result


class SubprocessMixin(object):
    """
    Mixin for running subprocesses and capturing their output
    """
    def __init__(self, verbose=False, progress=None):
        self.verbose = verbose
        self.progress = progress

    def reader(self, stream, context):
        """
        Read lines from a subprocess' output stream and either pass to a progress
        callable (if specified) or write progress information to sys.stderr.
        """
        progress = self.progress
        verbose = self.verbose
        while True:
            s = stream.readline()
            if not s:
                break
            if progress is not None:
                progress(s, context)
            else:
                if not verbose:
                    sys.stderr.write('.')
                else:
                    sys.stderr.write(s.decode('utf-8'))
                sys.stderr.flush()
        stream.close()

    def run_command(self, cmd, **kwargs):
        p = subprocess.Popen(cmd, stdout=subprocess.PIPE,
                             stderr=subprocess.PIPE, **kwargs)
        t1 = threading.Thread(target=self.reader, args=(p.stdout, 'stdout'))
        t1.start()
        t2 = threading.Thread(target=self.reader, args=(p.stderr, 'stderr'))
        t2.start()
        p.wait()
        t1.join()
        t2.join()
        if self.progress is not None:
            self.progress('done.', 'main')
        elif self.verbose:
            sys.stderr.write('done.\n')
        return p


def normalize_name(name):
    """Normalize a python package name a la PEP 503"""
    # https://www.python.org/dev/peps/pep-0503/#normalized-names
    return re.sub('[-_.]+', '-', name).lower()
site-packages/pip/_vendor/distlib/metadata.py000064400000116452151464452710015316 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Implementation of the Metadata for Python packages PEPs.

Supports all metadata formats (1.0, 1.1, 1.2, and 2.0 experimental).
"""
from __future__ import unicode_literals

import codecs
from email import message_from_file
import json
import logging
import re


from . import DistlibException, __version__
from .compat import StringIO, string_types, text_type
from .markers import interpret
from .util import extract_by_key, get_extras
from .version import get_scheme, PEP440_VERSION_RE

logger = logging.getLogger(__name__)


class MetadataMissingError(DistlibException):
    """A required metadata is missing"""


class MetadataConflictError(DistlibException):
    """Attempt to read or write metadata fields that are conflictual."""


class MetadataUnrecognizedVersionError(DistlibException):
    """Unknown metadata version number."""


class MetadataInvalidError(DistlibException):
    """A metadata value is invalid"""

# public API of this module
__all__ = ['Metadata', 'PKG_INFO_ENCODING', 'PKG_INFO_PREFERRED_VERSION']

# Encoding used for the PKG-INFO files
PKG_INFO_ENCODING = 'utf-8'

# preferred version. Hopefully will be changed
# to 1.2 once PEP 345 is supported everywhere
PKG_INFO_PREFERRED_VERSION = '1.1'

_LINE_PREFIX_1_2 = re.compile('\n       \\|')
_LINE_PREFIX_PRE_1_2 = re.compile('\n        ')
_241_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'License')

_314_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Supported-Platform', 'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'License', 'Classifier', 'Download-URL', 'Obsoletes',
               'Provides', 'Requires')

_314_MARKERS = ('Obsoletes', 'Provides', 'Requires', 'Classifier',
                'Download-URL')

_345_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Supported-Platform', 'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'Maintainer', 'Maintainer-email', 'License',
               'Classifier', 'Download-URL', 'Obsoletes-Dist',
               'Project-URL', 'Provides-Dist', 'Requires-Dist',
               'Requires-Python', 'Requires-External')

_345_MARKERS = ('Provides-Dist', 'Requires-Dist', 'Requires-Python',
                'Obsoletes-Dist', 'Requires-External', 'Maintainer',
                'Maintainer-email', 'Project-URL')

_426_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Supported-Platform', 'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'Maintainer', 'Maintainer-email', 'License',
               'Classifier', 'Download-URL', 'Obsoletes-Dist',
               'Project-URL', 'Provides-Dist', 'Requires-Dist',
               'Requires-Python', 'Requires-External', 'Private-Version',
               'Obsoleted-By', 'Setup-Requires-Dist', 'Extension',
               'Provides-Extra')

_426_MARKERS = ('Private-Version', 'Provides-Extra', 'Obsoleted-By',
                'Setup-Requires-Dist', 'Extension')

# See issue #106: Sometimes 'Requires' and 'Provides' occur wrongly in
# the metadata. Include them in the tuple literal below to allow them
# (for now).
_566_FIELDS = _426_FIELDS + ('Description-Content-Type',
                             'Requires', 'Provides')

_566_MARKERS = ('Description-Content-Type',)

_ALL_FIELDS = set()
_ALL_FIELDS.update(_241_FIELDS)
_ALL_FIELDS.update(_314_FIELDS)
_ALL_FIELDS.update(_345_FIELDS)
_ALL_FIELDS.update(_426_FIELDS)
_ALL_FIELDS.update(_566_FIELDS)

EXTRA_RE = re.compile(r'''extra\s*==\s*("([^"]+)"|'([^']+)')''')


def _version2fieldlist(version):
    if version == '1.0':
        return _241_FIELDS
    elif version == '1.1':
        return _314_FIELDS
    elif version == '1.2':
        return _345_FIELDS
    elif version in ('1.3', '2.1'):
        return _345_FIELDS + _566_FIELDS
    elif version == '2.0':
        return _426_FIELDS
    raise MetadataUnrecognizedVersionError(version)


def _best_version(fields):
    """Detect the best version depending on the fields used."""
    def _has_marker(keys, markers):
        for marker in markers:
            if marker in keys:
                return True
        return False

    keys = []
    for key, value in fields.items():
        if value in ([], 'UNKNOWN', None):
            continue
        keys.append(key)

    possible_versions = ['1.0', '1.1', '1.2', '1.3', '2.0', '2.1']

    # first let's try to see if a field is not part of one of the version
    for key in keys:
        if key not in _241_FIELDS and '1.0' in possible_versions:
            possible_versions.remove('1.0')
            logger.debug('Removed 1.0 due to %s', key)
        if key not in _314_FIELDS and '1.1' in possible_versions:
            possible_versions.remove('1.1')
            logger.debug('Removed 1.1 due to %s', key)
        if key not in _345_FIELDS and '1.2' in possible_versions:
            possible_versions.remove('1.2')
            logger.debug('Removed 1.2 due to %s', key)
        if key not in _566_FIELDS and '1.3' in possible_versions:
            possible_versions.remove('1.3')
            logger.debug('Removed 1.3 due to %s', key)
        if key not in _566_FIELDS and '2.1' in possible_versions:
            if key != 'Description':  # In 2.1, description allowed after headers
                possible_versions.remove('2.1')
                logger.debug('Removed 2.1 due to %s', key)
        if key not in _426_FIELDS and '2.0' in possible_versions:
            possible_versions.remove('2.0')
            logger.debug('Removed 2.0 due to %s', key)

    # possible_version contains qualified versions
    if len(possible_versions) == 1:
        return possible_versions[0]   # found !
    elif len(possible_versions) == 0:
        logger.debug('Out of options - unknown metadata set: %s', fields)
        raise MetadataConflictError('Unknown metadata set')

    # let's see if one unique marker is found
    is_1_1 = '1.1' in possible_versions and _has_marker(keys, _314_MARKERS)
    is_1_2 = '1.2' in possible_versions and _has_marker(keys, _345_MARKERS)
    is_2_1 = '2.1' in possible_versions and _has_marker(keys, _566_MARKERS)
    is_2_0 = '2.0' in possible_versions and _has_marker(keys, _426_MARKERS)
    if int(is_1_1) + int(is_1_2) + int(is_2_1) + int(is_2_0) > 1:
        raise MetadataConflictError('You used incompatible 1.1/1.2/2.0/2.1 fields')

    # we have the choice, 1.0, or 1.2, or 2.0
    #   - 1.0 has a broken Summary field but works with all tools
    #   - 1.1 is to avoid
    #   - 1.2 fixes Summary but has little adoption
    #   - 2.0 adds more features and is very new
    if not is_1_1 and not is_1_2 and not is_2_1 and not is_2_0:
        # we couldn't find any specific marker
        if PKG_INFO_PREFERRED_VERSION in possible_versions:
            return PKG_INFO_PREFERRED_VERSION
    if is_1_1:
        return '1.1'
    if is_1_2:
        return '1.2'
    if is_2_1:
        return '2.1'

    return '2.0'

_ATTR2FIELD = {
    'metadata_version': 'Metadata-Version',
    'name': 'Name',
    'version': 'Version',
    'platform': 'Platform',
    'supported_platform': 'Supported-Platform',
    'summary': 'Summary',
    'description': 'Description',
    'keywords': 'Keywords',
    'home_page': 'Home-page',
    'author': 'Author',
    'author_email': 'Author-email',
    'maintainer': 'Maintainer',
    'maintainer_email': 'Maintainer-email',
    'license': 'License',
    'classifier': 'Classifier',
    'download_url': 'Download-URL',
    'obsoletes_dist': 'Obsoletes-Dist',
    'provides_dist': 'Provides-Dist',
    'requires_dist': 'Requires-Dist',
    'setup_requires_dist': 'Setup-Requires-Dist',
    'requires_python': 'Requires-Python',
    'requires_external': 'Requires-External',
    'requires': 'Requires',
    'provides': 'Provides',
    'obsoletes': 'Obsoletes',
    'project_url': 'Project-URL',
    'private_version': 'Private-Version',
    'obsoleted_by': 'Obsoleted-By',
    'extension': 'Extension',
    'provides_extra': 'Provides-Extra',
}

_PREDICATE_FIELDS = ('Requires-Dist', 'Obsoletes-Dist', 'Provides-Dist')
_VERSIONS_FIELDS = ('Requires-Python',)
_VERSION_FIELDS = ('Version',)
_LISTFIELDS = ('Platform', 'Classifier', 'Obsoletes',
               'Requires', 'Provides', 'Obsoletes-Dist',
               'Provides-Dist', 'Requires-Dist', 'Requires-External',
               'Project-URL', 'Supported-Platform', 'Setup-Requires-Dist',
               'Provides-Extra', 'Extension')
_LISTTUPLEFIELDS = ('Project-URL',)

_ELEMENTSFIELD = ('Keywords',)

_UNICODEFIELDS = ('Author', 'Maintainer', 'Summary', 'Description')

_MISSING = object()

_FILESAFE = re.compile('[^A-Za-z0-9.]+')


def _get_name_and_version(name, version, for_filename=False):
    """Return the distribution name with version.

    If for_filename is true, return a filename-escaped form."""
    if for_filename:
        # For both name and version any runs of non-alphanumeric or '.'
        # characters are replaced with a single '-'.  Additionally any
        # spaces in the version string become '.'
        name = _FILESAFE.sub('-', name)
        version = _FILESAFE.sub('-', version.replace(' ', '.'))
    return '%s-%s' % (name, version)


class LegacyMetadata(object):
    """The legacy metadata of a release.

    Supports versions 1.0, 1.1 and 1.2 (auto-detected). You can
    instantiate the class with one of these arguments (or none):
    - *path*, the path to a metadata file
    - *fileobj* give a file-like object with metadata as content
    - *mapping* is a dict-like object
    - *scheme* is a version scheme name
    """
    # TODO document the mapping API and UNKNOWN default key

    def __init__(self, path=None, fileobj=None, mapping=None,
                 scheme='default'):
        if [path, fileobj, mapping].count(None) < 2:
            raise TypeError('path, fileobj and mapping are exclusive')
        self._fields = {}
        self.requires_files = []
        self._dependencies = None
        self.scheme = scheme
        if path is not None:
            self.read(path)
        elif fileobj is not None:
            self.read_file(fileobj)
        elif mapping is not None:
            self.update(mapping)
            self.set_metadata_version()

    def set_metadata_version(self):
        self._fields['Metadata-Version'] = _best_version(self._fields)

    def _write_field(self, fileobj, name, value):
        fileobj.write('%s: %s\n' % (name, value))

    def __getitem__(self, name):
        return self.get(name)

    def __setitem__(self, name, value):
        return self.set(name, value)

    def __delitem__(self, name):
        field_name = self._convert_name(name)
        try:
            del self._fields[field_name]
        except KeyError:
            raise KeyError(name)

    def __contains__(self, name):
        return (name in self._fields or
                self._convert_name(name) in self._fields)

    def _convert_name(self, name):
        if name in _ALL_FIELDS:
            return name
        name = name.replace('-', '_').lower()
        return _ATTR2FIELD.get(name, name)

    def _default_value(self, name):
        if name in _LISTFIELDS or name in _ELEMENTSFIELD:
            return []
        return 'UNKNOWN'

    def _remove_line_prefix(self, value):
        if self.metadata_version in ('1.0', '1.1'):
            return _LINE_PREFIX_PRE_1_2.sub('\n', value)
        else:
            return _LINE_PREFIX_1_2.sub('\n', value)

    def __getattr__(self, name):
        if name in _ATTR2FIELD:
            return self[name]
        raise AttributeError(name)

    #
    # Public API
    #

#    dependencies = property(_get_dependencies, _set_dependencies)

    def get_fullname(self, filesafe=False):
        """Return the distribution name with version.

        If filesafe is true, return a filename-escaped form."""
        return _get_name_and_version(self['Name'], self['Version'], filesafe)

    def is_field(self, name):
        """return True if name is a valid metadata key"""
        name = self._convert_name(name)
        return name in _ALL_FIELDS

    def is_multi_field(self, name):
        name = self._convert_name(name)
        return name in _LISTFIELDS

    def read(self, filepath):
        """Read the metadata values from a file path."""
        fp = codecs.open(filepath, 'r', encoding='utf-8')
        try:
            self.read_file(fp)
        finally:
            fp.close()

    def read_file(self, fileob):
        """Read the metadata values from a file object."""
        msg = message_from_file(fileob)
        self._fields['Metadata-Version'] = msg['metadata-version']

        # When reading, get all the fields we can
        for field in _ALL_FIELDS:
            if field not in msg:
                continue
            if field in _LISTFIELDS:
                # we can have multiple lines
                values = msg.get_all(field)
                if field in _LISTTUPLEFIELDS and values is not None:
                    values = [tuple(value.split(',')) for value in values]
                self.set(field, values)
            else:
                # single line
                value = msg[field]
                if value is not None and value != 'UNKNOWN':
                    self.set(field, value)
        # logger.debug('Attempting to set metadata for %s', self)
        # self.set_metadata_version()

    def write(self, filepath, skip_unknown=False):
        """Write the metadata fields to filepath."""
        fp = codecs.open(filepath, 'w', encoding='utf-8')
        try:
            self.write_file(fp, skip_unknown)
        finally:
            fp.close()

    def write_file(self, fileobject, skip_unknown=False):
        """Write the PKG-INFO format data to a file object."""
        self.set_metadata_version()

        for field in _version2fieldlist(self['Metadata-Version']):
            values = self.get(field)
            if skip_unknown and values in ('UNKNOWN', [], ['UNKNOWN']):
                continue
            if field in _ELEMENTSFIELD:
                self._write_field(fileobject, field, ','.join(values))
                continue
            if field not in _LISTFIELDS:
                if field == 'Description':
                    if self.metadata_version in ('1.0', '1.1'):
                        values = values.replace('\n', '\n        ')
                    else:
                        values = values.replace('\n', '\n       |')
                values = [values]

            if field in _LISTTUPLEFIELDS:
                values = [','.join(value) for value in values]

            for value in values:
                self._write_field(fileobject, field, value)

    def update(self, other=None, **kwargs):
        """Set metadata values from the given iterable `other` and kwargs.

        Behavior is like `dict.update`: If `other` has a ``keys`` method,
        they are looped over and ``self[key]`` is assigned ``other[key]``.
        Else, ``other`` is an iterable of ``(key, value)`` iterables.

        Keys that don't match a metadata field or that have an empty value are
        dropped.
        """
        def _set(key, value):
            if key in _ATTR2FIELD and value:
                self.set(self._convert_name(key), value)

        if not other:
            # other is None or empty container
            pass
        elif hasattr(other, 'keys'):
            for k in other.keys():
                _set(k, other[k])
        else:
            for k, v in other:
                _set(k, v)

        if kwargs:
            for k, v in kwargs.items():
                _set(k, v)

    def set(self, name, value):
        """Control then set a metadata field."""
        name = self._convert_name(name)

        if ((name in _ELEMENTSFIELD or name == 'Platform') and
            not isinstance(value, (list, tuple))):
            if isinstance(value, string_types):
                value = [v.strip() for v in value.split(',')]
            else:
                value = []
        elif (name in _LISTFIELDS and
              not isinstance(value, (list, tuple))):
            if isinstance(value, string_types):
                value = [value]
            else:
                value = []

        if logger.isEnabledFor(logging.WARNING):
            project_name = self['Name']

            scheme = get_scheme(self.scheme)
            if name in _PREDICATE_FIELDS and value is not None:
                for v in value:
                    # check that the values are valid
                    if not scheme.is_valid_matcher(v.split(';')[0]):
                        logger.warning(
                            "'%s': '%s' is not valid (field '%s')",
                            project_name, v, name)
            # FIXME this rejects UNKNOWN, is that right?
            elif name in _VERSIONS_FIELDS and value is not None:
                if not scheme.is_valid_constraint_list(value):
                    logger.warning("'%s': '%s' is not a valid version (field '%s')",
                                   project_name, value, name)
            elif name in _VERSION_FIELDS and value is not None:
                if not scheme.is_valid_version(value):
                    logger.warning("'%s': '%s' is not a valid version (field '%s')",
                                   project_name, value, name)

        if name in _UNICODEFIELDS:
            if name == 'Description':
                value = self._remove_line_prefix(value)

        self._fields[name] = value

    def get(self, name, default=_MISSING):
        """Get a metadata field."""
        name = self._convert_name(name)
        if name not in self._fields:
            if default is _MISSING:
                default = self._default_value(name)
            return default
        if name in _UNICODEFIELDS:
            value = self._fields[name]
            return value
        elif name in _LISTFIELDS:
            value = self._fields[name]
            if value is None:
                return []
            res = []
            for val in value:
                if name not in _LISTTUPLEFIELDS:
                    res.append(val)
                else:
                    # That's for Project-URL
                    res.append((val[0], val[1]))
            return res

        elif name in _ELEMENTSFIELD:
            value = self._fields[name]
            if isinstance(value, string_types):
                return value.split(',')
        return self._fields[name]

    def check(self, strict=False):
        """Check if the metadata is compliant. If strict is True then raise if
        no Name or Version are provided"""
        self.set_metadata_version()

        # XXX should check the versions (if the file was loaded)
        missing, warnings = [], []

        for attr in ('Name', 'Version'):  # required by PEP 345
            if attr not in self:
                missing.append(attr)

        if strict and missing != []:
            msg = 'missing required metadata: %s' % ', '.join(missing)
            raise MetadataMissingError(msg)

        for attr in ('Home-page', 'Author'):
            if attr not in self:
                missing.append(attr)

        # checking metadata 1.2 (XXX needs to check 1.1, 1.0)
        if self['Metadata-Version'] != '1.2':
            return missing, warnings

        scheme = get_scheme(self.scheme)

        def are_valid_constraints(value):
            for v in value:
                if not scheme.is_valid_matcher(v.split(';')[0]):
                    return False
            return True

        for fields, controller in ((_PREDICATE_FIELDS, are_valid_constraints),
                                   (_VERSIONS_FIELDS,
                                    scheme.is_valid_constraint_list),
                                   (_VERSION_FIELDS,
                                    scheme.is_valid_version)):
            for field in fields:
                value = self.get(field, None)
                if value is not None and not controller(value):
                    warnings.append("Wrong value for '%s': %s" % (field, value))

        return missing, warnings

    def todict(self, skip_missing=False):
        """Return fields as a dict.

        Field names will be converted to use the underscore-lowercase style
        instead of hyphen-mixed case (i.e. home_page instead of Home-page).
        """
        self.set_metadata_version()

        mapping_1_0 = (
            ('metadata_version', 'Metadata-Version'),
            ('name', 'Name'),
            ('version', 'Version'),
            ('summary', 'Summary'),
            ('home_page', 'Home-page'),
            ('author', 'Author'),
            ('author_email', 'Author-email'),
            ('license', 'License'),
            ('description', 'Description'),
            ('keywords', 'Keywords'),
            ('platform', 'Platform'),
            ('classifiers', 'Classifier'),
            ('download_url', 'Download-URL'),
        )

        data = {}
        for key, field_name in mapping_1_0:
            if not skip_missing or field_name in self._fields:
                data[key] = self[field_name]

        if self['Metadata-Version'] == '1.2':
            mapping_1_2 = (
                ('requires_dist', 'Requires-Dist'),
                ('requires_python', 'Requires-Python'),
                ('requires_external', 'Requires-External'),
                ('provides_dist', 'Provides-Dist'),
                ('obsoletes_dist', 'Obsoletes-Dist'),
                ('project_url', 'Project-URL'),
                ('maintainer', 'Maintainer'),
                ('maintainer_email', 'Maintainer-email'),
            )
            for key, field_name in mapping_1_2:
                if not skip_missing or field_name in self._fields:
                    if key != 'project_url':
                        data[key] = self[field_name]
                    else:
                        data[key] = [','.join(u) for u in self[field_name]]

        elif self['Metadata-Version'] == '1.1':
            mapping_1_1 = (
                ('provides', 'Provides'),
                ('requires', 'Requires'),
                ('obsoletes', 'Obsoletes'),
            )
            for key, field_name in mapping_1_1:
                if not skip_missing or field_name in self._fields:
                    data[key] = self[field_name]

        return data

    def add_requirements(self, requirements):
        if self['Metadata-Version'] == '1.1':
            # we can't have 1.1 metadata *and* Setuptools requires
            for field in ('Obsoletes', 'Requires', 'Provides'):
                if field in self:
                    del self[field]
        self['Requires-Dist'] += requirements

    # Mapping API
    # TODO could add iter* variants

    def keys(self):
        return list(_version2fieldlist(self['Metadata-Version']))

    def __iter__(self):
        for key in self.keys():
            yield key

    def values(self):
        return [self[key] for key in self.keys()]

    def items(self):
        return [(key, self[key]) for key in self.keys()]

    def __repr__(self):
        return '<%s %s %s>' % (self.__class__.__name__, self.name,
                               self.version)


METADATA_FILENAME = 'pydist.json'
WHEEL_METADATA_FILENAME = 'metadata.json'
LEGACY_METADATA_FILENAME = 'METADATA'


class Metadata(object):
    """
    The metadata of a release. This implementation uses 2.0 (JSON)
    metadata where possible. If not possible, it wraps a LegacyMetadata
    instance which handles the key-value metadata format.
    """

    METADATA_VERSION_MATCHER = re.compile(r'^\d+(\.\d+)*$')

    NAME_MATCHER = re.compile('^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$', re.I)

    VERSION_MATCHER = PEP440_VERSION_RE

    SUMMARY_MATCHER = re.compile('.{1,2047}')

    METADATA_VERSION = '2.0'

    GENERATOR = 'distlib (%s)' % __version__

    MANDATORY_KEYS = {
        'name': (),
        'version': (),
        'summary': ('legacy',),
    }

    INDEX_KEYS = ('name version license summary description author '
                  'author_email keywords platform home_page classifiers '
                  'download_url')

    DEPENDENCY_KEYS = ('extras run_requires test_requires build_requires '
                       'dev_requires provides meta_requires obsoleted_by '
                       'supports_environments')

    SYNTAX_VALIDATORS = {
        'metadata_version': (METADATA_VERSION_MATCHER, ()),
        'name': (NAME_MATCHER, ('legacy',)),
        'version': (VERSION_MATCHER, ('legacy',)),
        'summary': (SUMMARY_MATCHER, ('legacy',)),
    }

    __slots__ = ('_legacy', '_data', 'scheme')

    def __init__(self, path=None, fileobj=None, mapping=None,
                 scheme='default'):
        if [path, fileobj, mapping].count(None) < 2:
            raise TypeError('path, fileobj and mapping are exclusive')
        self._legacy = None
        self._data = None
        self.scheme = scheme
        #import pdb; pdb.set_trace()
        if mapping is not None:
            try:
                self._validate_mapping(mapping, scheme)
                self._data = mapping
            except MetadataUnrecognizedVersionError:
                self._legacy = LegacyMetadata(mapping=mapping, scheme=scheme)
                self.validate()
        else:
            data = None
            if path:
                with open(path, 'rb') as f:
                    data = f.read()
            elif fileobj:
                data = fileobj.read()
            if data is None:
                # Initialised with no args - to be added
                self._data = {
                    'metadata_version': self.METADATA_VERSION,
                    'generator': self.GENERATOR,
                }
            else:
                if not isinstance(data, text_type):
                    data = data.decode('utf-8')
                try:
                    self._data = json.loads(data)
                    self._validate_mapping(self._data, scheme)
                except ValueError:
                    # Note: MetadataUnrecognizedVersionError does not
                    # inherit from ValueError (it's a DistlibException,
                    # which should not inherit from ValueError).
                    # The ValueError comes from the json.load - if that
                    # succeeds and we get a validation error, we want
                    # that to propagate
                    self._legacy = LegacyMetadata(fileobj=StringIO(data),
                                                  scheme=scheme)
                    self.validate()

    common_keys = set(('name', 'version', 'license', 'keywords', 'summary'))

    none_list = (None, list)
    none_dict = (None, dict)

    mapped_keys = {
        'run_requires': ('Requires-Dist', list),
        'build_requires': ('Setup-Requires-Dist', list),
        'dev_requires': none_list,
        'test_requires': none_list,
        'meta_requires': none_list,
        'extras': ('Provides-Extra', list),
        'modules': none_list,
        'namespaces': none_list,
        'exports': none_dict,
        'commands': none_dict,
        'classifiers': ('Classifier', list),
        'source_url': ('Download-URL', None),
        'metadata_version': ('Metadata-Version', None),
    }

    del none_list, none_dict

    def __getattribute__(self, key):
        common = object.__getattribute__(self, 'common_keys')
        mapped = object.__getattribute__(self, 'mapped_keys')
        if key in mapped:
            lk, maker = mapped[key]
            if self._legacy:
                if lk is None:
                    result = None if maker is None else maker()
                else:
                    result = self._legacy.get(lk)
            else:
                value = None if maker is None else maker()
                if key not in ('commands', 'exports', 'modules', 'namespaces',
                               'classifiers'):
                    result = self._data.get(key, value)
                else:
                    # special cases for PEP 459
                    sentinel = object()
                    result = sentinel
                    d = self._data.get('extensions')
                    if d:
                        if key == 'commands':
                            result = d.get('python.commands', value)
                        elif key == 'classifiers':
                            d = d.get('python.details')
                            if d:
                                result = d.get(key, value)
                        else:
                            d = d.get('python.exports')
                            if not d:
                                d = self._data.get('python.exports')
                            if d:
                                result = d.get(key, value)
                    if result is sentinel:
                        result = value
        elif key not in common:
            result = object.__getattribute__(self, key)
        elif self._legacy:
            result = self._legacy.get(key)
        else:
            result = self._data.get(key)
        return result

    def _validate_value(self, key, value, scheme=None):
        if key in self.SYNTAX_VALIDATORS:
            pattern, exclusions = self.SYNTAX_VALIDATORS[key]
            if (scheme or self.scheme) not in exclusions:
                m = pattern.match(value)
                if not m:
                    raise MetadataInvalidError("'%s' is an invalid value for "
                                               "the '%s' property" % (value,
                                                                    key))

    def __setattr__(self, key, value):
        self._validate_value(key, value)
        common = object.__getattribute__(self, 'common_keys')
        mapped = object.__getattribute__(self, 'mapped_keys')
        if key in mapped:
            lk, _ = mapped[key]
            if self._legacy:
                if lk is None:
                    raise NotImplementedError
                self._legacy[lk] = value
            elif key not in ('commands', 'exports', 'modules', 'namespaces',
                             'classifiers'):
                self._data[key] = value
            else:
                # special cases for PEP 459
                d = self._data.setdefault('extensions', {})
                if key == 'commands':
                    d['python.commands'] = value
                elif key == 'classifiers':
                    d = d.setdefault('python.details', {})
                    d[key] = value
                else:
                    d = d.setdefault('python.exports', {})
                    d[key] = value
        elif key not in common:
            object.__setattr__(self, key, value)
        else:
            if key == 'keywords':
                if isinstance(value, string_types):
                    value = value.strip()
                    if value:
                        value = value.split()
                    else:
                        value = []
            if self._legacy:
                self._legacy[key] = value
            else:
                self._data[key] = value

    @property
    def name_and_version(self):
        return _get_name_and_version(self.name, self.version, True)

    @property
    def provides(self):
        if self._legacy:
            result = self._legacy['Provides-Dist']
        else:
            result = self._data.setdefault('provides', [])
        s = '%s (%s)' % (self.name, self.version)
        if s not in result:
            result.append(s)
        return result

    @provides.setter
    def provides(self, value):
        if self._legacy:
            self._legacy['Provides-Dist'] = value
        else:
            self._data['provides'] = value

    def get_requirements(self, reqts, extras=None, env=None):
        """
        Base method to get dependencies, given a set of extras
        to satisfy and an optional environment context.
        :param reqts: A list of sometimes-wanted dependencies,
                      perhaps dependent on extras and environment.
        :param extras: A list of optional components being requested.
        :param env: An optional environment for marker evaluation.
        """
        if self._legacy:
            result = reqts
        else:
            result = []
            extras = get_extras(extras or [], self.extras)
            for d in reqts:
                if 'extra' not in d and 'environment' not in d:
                    # unconditional
                    include = True
                else:
                    if 'extra' not in d:
                        # Not extra-dependent - only environment-dependent
                        include = True
                    else:
                        include = d.get('extra') in extras
                    if include:
                        # Not excluded because of extras, check environment
                        marker = d.get('environment')
                        if marker:
                            include = interpret(marker, env)
                if include:
                    result.extend(d['requires'])
            for key in ('build', 'dev', 'test'):
                e = ':%s:' % key
                if e in extras:
                    extras.remove(e)
                    # A recursive call, but it should terminate since 'test'
                    # has been removed from the extras
                    reqts = self._data.get('%s_requires' % key, [])
                    result.extend(self.get_requirements(reqts, extras=extras,
                                                        env=env))
        return result

    @property
    def dictionary(self):
        if self._legacy:
            return self._from_legacy()
        return self._data

    @property
    def dependencies(self):
        if self._legacy:
            raise NotImplementedError
        else:
            return extract_by_key(self._data, self.DEPENDENCY_KEYS)

    @dependencies.setter
    def dependencies(self, value):
        if self._legacy:
            raise NotImplementedError
        else:
            self._data.update(value)

    def _validate_mapping(self, mapping, scheme):
        if mapping.get('metadata_version') != self.METADATA_VERSION:
            raise MetadataUnrecognizedVersionError()
        missing = []
        for key, exclusions in self.MANDATORY_KEYS.items():
            if key not in mapping:
                if scheme not in exclusions:
                    missing.append(key)
        if missing:
            msg = 'Missing metadata items: %s' % ', '.join(missing)
            raise MetadataMissingError(msg)
        for k, v in mapping.items():
            self._validate_value(k, v, scheme)

    def validate(self):
        if self._legacy:
            missing, warnings = self._legacy.check(True)
            if missing or warnings:
                logger.warning('Metadata: missing: %s, warnings: %s',
                               missing, warnings)
        else:
            self._validate_mapping(self._data, self.scheme)

    def todict(self):
        if self._legacy:
            return self._legacy.todict(True)
        else:
            result = extract_by_key(self._data, self.INDEX_KEYS)
            return result

    def _from_legacy(self):
        assert self._legacy and not self._data
        result = {
            'metadata_version': self.METADATA_VERSION,
            'generator': self.GENERATOR,
        }
        lmd = self._legacy.todict(True)     # skip missing ones
        for k in ('name', 'version', 'license', 'summary', 'description',
                  'classifier'):
            if k in lmd:
                if k == 'classifier':
                    nk = 'classifiers'
                else:
                    nk = k
                result[nk] = lmd[k]
        kw = lmd.get('Keywords', [])
        if kw == ['']:
            kw = []
        result['keywords'] = kw
        keys = (('requires_dist', 'run_requires'),
                ('setup_requires_dist', 'build_requires'))
        for ok, nk in keys:
            if ok in lmd and lmd[ok]:
                result[nk] = [{'requires': lmd[ok]}]
        result['provides'] = self.provides
        author = {}
        maintainer = {}
        return result

    LEGACY_MAPPING = {
        'name': 'Name',
        'version': 'Version',
        'license': 'License',
        'summary': 'Summary',
        'description': 'Description',
        'classifiers': 'Classifier',
    }

    def _to_legacy(self):
        def process_entries(entries):
            reqts = set()
            for e in entries:
                extra = e.get('extra')
                env = e.get('environment')
                rlist = e['requires']
                for r in rlist:
                    if not env and not extra:
                        reqts.add(r)
                    else:
                        marker = ''
                        if extra:
                            marker = 'extra == "%s"' % extra
                        if env:
                            if marker:
                                marker = '(%s) and %s' % (env, marker)
                            else:
                                marker = env
                        reqts.add(';'.join((r, marker)))
            return reqts

        assert self._data and not self._legacy
        result = LegacyMetadata()
        nmd = self._data
        for nk, ok in self.LEGACY_MAPPING.items():
            if nk in nmd:
                result[ok] = nmd[nk]
        r1 = process_entries(self.run_requires + self.meta_requires)
        r2 = process_entries(self.build_requires + self.dev_requires)
        if self.extras:
            result['Provides-Extra'] = sorted(self.extras)
        result['Requires-Dist'] = sorted(r1)
        result['Setup-Requires-Dist'] = sorted(r2)
        # TODO: other fields such as contacts
        return result

    def write(self, path=None, fileobj=None, legacy=False, skip_unknown=True):
        if [path, fileobj].count(None) != 1:
            raise ValueError('Exactly one of path and fileobj is needed')
        self.validate()
        if legacy:
            if self._legacy:
                legacy_md = self._legacy
            else:
                legacy_md = self._to_legacy()
            if path:
                legacy_md.write(path, skip_unknown=skip_unknown)
            else:
                legacy_md.write_file(fileobj, skip_unknown=skip_unknown)
        else:
            if self._legacy:
                d = self._from_legacy()
            else:
                d = self._data
            if fileobj:
                json.dump(d, fileobj, ensure_ascii=True, indent=2,
                          sort_keys=True)
            else:
                with codecs.open(path, 'w', 'utf-8') as f:
                    json.dump(d, f, ensure_ascii=True, indent=2,
                              sort_keys=True)

    def add_requirements(self, requirements):
        if self._legacy:
            self._legacy.add_requirements(requirements)
        else:
            run_requires = self._data.setdefault('run_requires', [])
            always = None
            for entry in run_requires:
                if 'environment' not in entry and 'extra' not in entry:
                    always = entry
                    break
            if always is None:
                always = { 'requires': requirements }
                run_requires.insert(0, always)
            else:
                rset = set(always['requires']) | set(requirements)
                always['requires'] = sorted(rset)

    def __repr__(self):
        name = self.name or '(no name)'
        version = self.version or 'no version'
        return '<%s %s %s (%s)>' % (self.__class__.__name__,
                                    self.metadata_version, name, version)
site-packages/pip/_vendor/distlib/locators.py000064400000145137151464452710015366 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2015 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#

import gzip
from io import BytesIO
import json
import logging
import os
import posixpath
import re
try:
    import threading
except ImportError:  # pragma: no cover
    import dummy_threading as threading
import zlib

from . import DistlibException
from .compat import (urljoin, urlparse, urlunparse, url2pathname, pathname2url,
                     queue, quote, unescape, string_types, build_opener,
                     HTTPRedirectHandler as BaseRedirectHandler, text_type,
                     Request, HTTPError, URLError)
from .database import Distribution, DistributionPath, make_dist
from .metadata import Metadata, MetadataInvalidError
from .util import (cached_property, parse_credentials, ensure_slash,
                   split_filename, get_project_data, parse_requirement,
                   parse_name_and_version, ServerProxy, normalize_name)
from .version import get_scheme, UnsupportedVersionError
from .wheel import Wheel, is_compatible

logger = logging.getLogger(__name__)

HASHER_HASH = re.compile(r'^(\w+)=([a-f0-9]+)')
CHARSET = re.compile(r';\s*charset\s*=\s*(.*)\s*$', re.I)
HTML_CONTENT_TYPE = re.compile('text/html|application/x(ht)?ml')
DEFAULT_INDEX = 'https://pypi.org/pypi'

def get_all_distribution_names(url=None):
    """
    Return all distribution names known by an index.
    :param url: The URL of the index.
    :return: A list of all known distribution names.
    """
    if url is None:
        url = DEFAULT_INDEX
    client = ServerProxy(url, timeout=3.0)
    try:
        return client.list_packages()
    finally:
        client('close')()

class RedirectHandler(BaseRedirectHandler):
    """
    A class to work around a bug in some Python 3.2.x releases.
    """
    # There's a bug in the base version for some 3.2.x
    # (e.g. 3.2.2 on Ubuntu Oneiric). If a Location header
    # returns e.g. /abc, it bails because it says the scheme ''
    # is bogus, when actually it should use the request's
    # URL for the scheme. See Python issue #13696.
    def http_error_302(self, req, fp, code, msg, headers):
        # Some servers (incorrectly) return multiple Location headers
        # (so probably same goes for URI).  Use first header.
        newurl = None
        for key in ('location', 'uri'):
            if key in headers:
                newurl = headers[key]
                break
        if newurl is None:  # pragma: no cover
            return
        urlparts = urlparse(newurl)
        if urlparts.scheme == '':
            newurl = urljoin(req.get_full_url(), newurl)
            if hasattr(headers, 'replace_header'):
                headers.replace_header(key, newurl)
            else:
                headers[key] = newurl
        return BaseRedirectHandler.http_error_302(self, req, fp, code, msg,
                                                  headers)

    http_error_301 = http_error_303 = http_error_307 = http_error_302

class Locator(object):
    """
    A base class for locators - things that locate distributions.
    """
    source_extensions = ('.tar.gz', '.tar.bz2', '.tar', '.zip', '.tgz', '.tbz')
    binary_extensions = ('.egg', '.exe', '.whl')
    excluded_extensions = ('.pdf',)

    # A list of tags indicating which wheels you want to match. The default
    # value of None matches against the tags compatible with the running
    # Python. If you want to match other values, set wheel_tags on a locator
    # instance to a list of tuples (pyver, abi, arch) which you want to match.
    wheel_tags = None

    downloadable_extensions = source_extensions + ('.whl',)

    def __init__(self, scheme='default'):
        """
        Initialise an instance.
        :param scheme: Because locators look for most recent versions, they
                       need to know the version scheme to use. This specifies
                       the current PEP-recommended scheme - use ``'legacy'``
                       if you need to support existing distributions on PyPI.
        """
        self._cache = {}
        self.scheme = scheme
        # Because of bugs in some of the handlers on some of the platforms,
        # we use our own opener rather than just using urlopen.
        self.opener = build_opener(RedirectHandler())
        # If get_project() is called from locate(), the matcher instance
        # is set from the requirement passed to locate(). See issue #18 for
        # why this can be useful to know.
        self.matcher = None
        self.errors = queue.Queue()

    def get_errors(self):
        """
        Return any errors which have occurred.
        """
        result = []
        while not self.errors.empty():  # pragma: no cover
            try:
                e = self.errors.get(False)
                result.append(e)
            except self.errors.Empty:
                continue
            self.errors.task_done()
        return result

    def clear_errors(self):
        """
        Clear any errors which may have been logged.
        """
        # Just get the errors and throw them away
        self.get_errors()

    def clear_cache(self):
        self._cache.clear()

    def _get_scheme(self):
        return self._scheme

    def _set_scheme(self, value):
        self._scheme = value

    scheme = property(_get_scheme, _set_scheme)

    def _get_project(self, name):
        """
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This should be implemented in subclasses.

        If called from a locate() request, self.matcher will be set to a
        matcher for the requirement to satisfy, otherwise it will be None.
        """
        raise NotImplementedError('Please implement in the subclass')

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        raise NotImplementedError('Please implement in the subclass')

    def get_project(self, name):
        """
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This calls _get_project to do all the work, and just implements a caching layer on top.
        """
        if self._cache is None:  # pragma: no cover
            result = self._get_project(name)
        elif name in self._cache:
            result = self._cache[name]
        else:
            self.clear_errors()
            result = self._get_project(name)
            self._cache[name] = result
        return result

    def score_url(self, url):
        """
        Give an url a score which can be used to choose preferred URLs
        for a given project release.
        """
        t = urlparse(url)
        basename = posixpath.basename(t.path)
        compatible = True
        is_wheel = basename.endswith('.whl')
        is_downloadable = basename.endswith(self.downloadable_extensions)
        if is_wheel:
            compatible = is_compatible(Wheel(basename), self.wheel_tags)
        return (t.scheme == 'https', 'pypi.org' in t.netloc,
                is_downloadable, is_wheel, compatible, basename)

    def prefer_url(self, url1, url2):
        """
        Choose one of two URLs where both are candidates for distribution
        archives for the same version of a distribution (for example,
        .tar.gz vs. zip).

        The current implementation favours https:// URLs over http://, archives
        from PyPI over those from other locations, wheel compatibility (if a
        wheel) and then the archive name.
        """
        result = url2
        if url1:
            s1 = self.score_url(url1)
            s2 = self.score_url(url2)
            if s1 > s2:
                result = url1
            if result != url2:
                logger.debug('Not replacing %r with %r', url1, url2)
            else:
                logger.debug('Replacing %r with %r', url1, url2)
        return result

    def split_filename(self, filename, project_name):
        """
        Attempt to split a filename in project name, version and Python version.
        """
        return split_filename(filename, project_name)

    def convert_url_to_download_info(self, url, project_name):
        """
        See if a URL is a candidate for a download URL for a project (the URL
        has typically been scraped from an HTML page).

        If it is, a dictionary is returned with keys "name", "version",
        "filename" and "url"; otherwise, None is returned.
        """
        def same_project(name1, name2):
            return normalize_name(name1) == normalize_name(name2)

        result = None
        scheme, netloc, path, params, query, frag = urlparse(url)
        if frag.lower().startswith('egg='):  # pragma: no cover
            logger.debug('%s: version hint in fragment: %r',
                         project_name, frag)
        m = HASHER_HASH.match(frag)
        if m:
            algo, digest = m.groups()
        else:
            algo, digest = None, None
        origpath = path
        if path and path[-1] == '/':  # pragma: no cover
            path = path[:-1]
        if path.endswith('.whl'):
            try:
                wheel = Wheel(path)
                if not is_compatible(wheel, self.wheel_tags):
                    logger.debug('Wheel not compatible: %s', path)
                else:
                    if project_name is None:
                        include = True
                    else:
                        include = same_project(wheel.name, project_name)
                    if include:
                        result = {
                            'name': wheel.name,
                            'version': wheel.version,
                            'filename': wheel.filename,
                            'url': urlunparse((scheme, netloc, origpath,
                                               params, query, '')),
                            'python-version': ', '.join(
                                ['.'.join(list(v[2:])) for v in wheel.pyver]),
                        }
            except Exception as e:  # pragma: no cover
                logger.warning('invalid path for wheel: %s', path)
        elif not path.endswith(self.downloadable_extensions):  # pragma: no cover
            logger.debug('Not downloadable: %s', path)
        else:  # downloadable extension
            path = filename = posixpath.basename(path)
            for ext in self.downloadable_extensions:
                if path.endswith(ext):
                    path = path[:-len(ext)]
                    t = self.split_filename(path, project_name)
                    if not t:  # pragma: no cover
                        logger.debug('No match for project/version: %s', path)
                    else:
                        name, version, pyver = t
                        if not project_name or same_project(project_name, name):
                            result = {
                                'name': name,
                                'version': version,
                                'filename': filename,
                                'url': urlunparse((scheme, netloc, origpath,
                                                   params, query, '')),
                                #'packagetype': 'sdist',
                            }
                            if pyver:  # pragma: no cover
                                result['python-version'] = pyver
                    break
        if result and algo:
            result['%s_digest' % algo] = digest
        return result

    def _get_digest(self, info):
        """
        Get a digest from a dictionary by looking at keys of the form
        'algo_digest'.

        Returns a 2-tuple (algo, digest) if found, else None. Currently
        looks only for SHA256, then MD5.
        """
        result = None
        for algo in ('sha256', 'md5'):
            key = '%s_digest' % algo
            if key in info:
                result = (algo, info[key])
                break
        return result

    def _update_version_data(self, result, info):
        """
        Update a result dictionary (the final result from _get_project) with a
        dictionary for a specific version, which typically holds information
        gleaned from a filename or URL for an archive for the distribution.
        """
        name = info.pop('name')
        version = info.pop('version')
        if version in result:
            dist = result[version]
            md = dist.metadata
        else:
            dist = make_dist(name, version, scheme=self.scheme)
            md = dist.metadata
        dist.digest = digest = self._get_digest(info)
        url = info['url']
        result['digests'][url] = digest
        if md.source_url != info['url']:
            md.source_url = self.prefer_url(md.source_url, url)
            result['urls'].setdefault(version, set()).add(url)
        dist.locator = self
        result[version] = dist

    def locate(self, requirement, prereleases=False):
        """
        Find the most recent distribution which matches the given
        requirement.

        :param requirement: A requirement of the form 'foo (1.0)' or perhaps
                            'foo (>= 1.0, < 2.0, != 1.3)'
        :param prereleases: If ``True``, allow pre-release versions
                            to be located. Otherwise, pre-release versions
                            are not returned.
        :return: A :class:`Distribution` instance, or ``None`` if no such
                 distribution could be located.
        """
        result = None
        r = parse_requirement(requirement)
        if r is None:  # pragma: no cover
            raise DistlibException('Not a valid requirement: %r' % requirement)
        scheme = get_scheme(self.scheme)
        self.matcher = matcher = scheme.matcher(r.requirement)
        logger.debug('matcher: %s (%s)', matcher, type(matcher).__name__)
        versions = self.get_project(r.name)
        if len(versions) > 2:   # urls and digests keys are present
            # sometimes, versions are invalid
            slist = []
            vcls = matcher.version_class
            for k in versions:
                if k in ('urls', 'digests'):
                    continue
                try:
                    if not matcher.match(k):
                        logger.debug('%s did not match %r', matcher, k)
                    else:
                        if prereleases or not vcls(k).is_prerelease:
                            slist.append(k)
                        else:
                            logger.debug('skipping pre-release '
                                         'version %s of %s', k, matcher.name)
                except Exception:  # pragma: no cover
                    logger.warning('error matching %s with %r', matcher, k)
                    pass # slist.append(k)
            if len(slist) > 1:
                slist = sorted(slist, key=scheme.key)
            if slist:
                logger.debug('sorted list: %s', slist)
                version = slist[-1]
                result = versions[version]
        if result:
            if r.extras:
                result.extras = r.extras
            result.download_urls = versions.get('urls', {}).get(version, set())
            d = {}
            sd = versions.get('digests', {})
            for url in result.download_urls:
                if url in sd:  # pragma: no cover
                    d[url] = sd[url]
            result.digests = d
        self.matcher = None
        return result


class PyPIRPCLocator(Locator):
    """
    This locator uses XML-RPC to locate distributions. It therefore
    cannot be used with simple mirrors (that only mirror file content).
    """
    def __init__(self, url, **kwargs):
        """
        Initialise an instance.

        :param url: The URL to use for XML-RPC.
        :param kwargs: Passed to the superclass constructor.
        """
        super(PyPIRPCLocator, self).__init__(**kwargs)
        self.base_url = url
        self.client = ServerProxy(url, timeout=3.0)

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        return set(self.client.list_packages())

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        versions = self.client.package_releases(name, True)
        for v in versions:
            urls = self.client.release_urls(name, v)
            data = self.client.release_data(name, v)
            metadata = Metadata(scheme=self.scheme)
            metadata.name = data['name']
            metadata.version = data['version']
            metadata.license = data.get('license')
            metadata.keywords = data.get('keywords', [])
            metadata.summary = data.get('summary')
            dist = Distribution(metadata)
            if urls:
                info = urls[0]
                metadata.source_url = info['url']
                dist.digest = self._get_digest(info)
                dist.locator = self
                result[v] = dist
                for info in urls:
                    url = info['url']
                    digest = self._get_digest(info)
                    result['urls'].setdefault(v, set()).add(url)
                    result['digests'][url] = digest
        return result

class PyPIJSONLocator(Locator):
    """
    This locator uses PyPI's JSON interface. It's very limited in functionality
    and probably not worth using.
    """
    def __init__(self, url, **kwargs):
        super(PyPIJSONLocator, self).__init__(**kwargs)
        self.base_url = ensure_slash(url)

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        raise NotImplementedError('Not available from this locator')

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        url = urljoin(self.base_url, '%s/json' % quote(name))
        try:
            resp = self.opener.open(url)
            data = resp.read().decode() # for now
            d = json.loads(data)
            md = Metadata(scheme=self.scheme)
            data = d['info']
            md.name = data['name']
            md.version = data['version']
            md.license = data.get('license')
            md.keywords = data.get('keywords', [])
            md.summary = data.get('summary')
            dist = Distribution(md)
            dist.locator = self
            urls = d['urls']
            result[md.version] = dist
            for info in d['urls']:
                url = info['url']
                dist.download_urls.add(url)
                dist.digests[url] = self._get_digest(info)
                result['urls'].setdefault(md.version, set()).add(url)
                result['digests'][url] = self._get_digest(info)
            # Now get other releases
            for version, infos in d['releases'].items():
                if version == md.version:
                    continue    # already done
                omd = Metadata(scheme=self.scheme)
                omd.name = md.name
                omd.version = version
                odist = Distribution(omd)
                odist.locator = self
                result[version] = odist
                for info in infos:
                    url = info['url']
                    odist.download_urls.add(url)
                    odist.digests[url] = self._get_digest(info)
                    result['urls'].setdefault(version, set()).add(url)
                    result['digests'][url] = self._get_digest(info)
#            for info in urls:
#                md.source_url = info['url']
#                dist.digest = self._get_digest(info)
#                dist.locator = self
#                for info in urls:
#                    url = info['url']
#                    result['urls'].setdefault(md.version, set()).add(url)
#                    result['digests'][url] = self._get_digest(info)
        except Exception as e:
            self.errors.put(text_type(e))
            logger.exception('JSON fetch failed: %s', e)
        return result


class Page(object):
    """
    This class represents a scraped HTML page.
    """
    # The following slightly hairy-looking regex just looks for the contents of
    # an anchor link, which has an attribute "href" either immediately preceded
    # or immediately followed by a "rel" attribute. The attribute values can be
    # declared with double quotes, single quotes or no quotes - which leads to
    # the length of the expression.
    _href = re.compile("""
(rel\\s*=\\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\\s\n]*))\\s+)?
href\\s*=\\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\\s\n]*))
(\\s+rel\\s*=\\s*(?:"(?P[^"]*)"|'(?P[^']*)'|(?P[^>\\s\n]*)))?
""", re.I | re.S | re.X)
    _base = re.compile(r"""]+)""", re.I | re.S)

    def __init__(self, data, url):
        """
        Initialise an instance with the Unicode page contents and the URL they
        came from.
        """
        self.data = data
        self.base_url = self.url = url
        m = self._base.search(self.data)
        if m:
            self.base_url = m.group(1)

    _clean_re = re.compile(r'[^a-z0-9$&+,/:;=?@.#%_\\|-]', re.I)

    @cached_property
    def links(self):
        """
        Return the URLs of all the links on a page together with information
        about their "rel" attribute, for determining which ones to treat as
        downloads and which ones to queue for further scraping.
        """
        def clean(url):
            "Tidy up an URL."
            scheme, netloc, path, params, query, frag = urlparse(url)
            return urlunparse((scheme, netloc, quote(path),
                               params, query, frag))

        result = set()
        for match in self._href.finditer(self.data):
            d = match.groupdict('')
            rel = (d['rel1'] or d['rel2'] or d['rel3'] or
                   d['rel4'] or d['rel5'] or d['rel6'])
            url = d['url1'] or d['url2'] or d['url3']
            url = urljoin(self.base_url, url)
            url = unescape(url)
            url = self._clean_re.sub(lambda m: '%%%2x' % ord(m.group(0)), url)
            result.add((url, rel))
        # We sort the result, hoping to bring the most recent versions
        # to the front
        result = sorted(result, key=lambda t: t[0], reverse=True)
        return result


class SimpleScrapingLocator(Locator):
    """
    A locator which scrapes HTML pages to locate downloads for a distribution.
    This runs multiple threads to do the I/O; performance is at least as good
    as pip's PackageFinder, which works in an analogous fashion.
    """

    # These are used to deal with various Content-Encoding schemes.
    decoders = {
        'deflate': zlib.decompress,
        'gzip': lambda b: gzip.GzipFile(fileobj=BytesIO(d)).read(),
        'none': lambda b: b,
    }

    def __init__(self, url, timeout=None, num_workers=10, **kwargs):
        """
        Initialise an instance.
        :param url: The root URL to use for scraping.
        :param timeout: The timeout, in seconds, to be applied to requests.
                        This defaults to ``None`` (no timeout specified).
        :param num_workers: The number of worker threads you want to do I/O,
                            This defaults to 10.
        :param kwargs: Passed to the superclass.
        """
        super(SimpleScrapingLocator, self).__init__(**kwargs)
        self.base_url = ensure_slash(url)
        self.timeout = timeout
        self._page_cache = {}
        self._seen = set()
        self._to_fetch = queue.Queue()
        self._bad_hosts = set()
        self.skip_externals = False
        self.num_workers = num_workers
        self._lock = threading.RLock()
        # See issue #45: we need to be resilient when the locator is used
        # in a thread, e.g. with concurrent.futures. We can't use self._lock
        # as it is for coordinating our internal threads - the ones created
        # in _prepare_threads.
        self._gplock = threading.RLock()
        self.platform_check = False  # See issue #112

    def _prepare_threads(self):
        """
        Threads are created only when get_project is called, and terminate
        before it returns. They are there primarily to parallelise I/O (i.e.
        fetching web pages).
        """
        self._threads = []
        for i in range(self.num_workers):
            t = threading.Thread(target=self._fetch)
            t.setDaemon(True)
            t.start()
            self._threads.append(t)

    def _wait_threads(self):
        """
        Tell all the threads to terminate (by sending a sentinel value) and
        wait for them to do so.
        """
        # Note that you need two loops, since you can't say which
        # thread will get each sentinel
        for t in self._threads:
            self._to_fetch.put(None)    # sentinel
        for t in self._threads:
            t.join()
        self._threads = []

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        with self._gplock:
            self.result = result
            self.project_name = name
            url = urljoin(self.base_url, '%s/' % quote(name))
            self._seen.clear()
            self._page_cache.clear()
            self._prepare_threads()
            try:
                logger.debug('Queueing %s', url)
                self._to_fetch.put(url)
                self._to_fetch.join()
            finally:
                self._wait_threads()
            del self.result
        return result

    platform_dependent = re.compile(r'\b(linux_(i\d86|x86_64|arm\w+)|'
                                    r'win(32|_amd64)|macosx_?\d+)\b', re.I)

    def _is_platform_dependent(self, url):
        """
        Does an URL refer to a platform-specific download?
        """
        return self.platform_dependent.search(url)

    def _process_download(self, url):
        """
        See if an URL is a suitable download for a project.

        If it is, register information in the result dictionary (for
        _get_project) about the specific version it's for.

        Note that the return value isn't actually used other than as a boolean
        value.
        """
        if self.platform_check and self._is_platform_dependent(url):
            info = None
        else:
            info = self.convert_url_to_download_info(url, self.project_name)
        logger.debug('process_download: %s -> %s', url, info)
        if info:
            with self._lock:    # needed because self.result is shared
                self._update_version_data(self.result, info)
        return info

    def _should_queue(self, link, referrer, rel):
        """
        Determine whether a link URL from a referring page and with a
        particular "rel" attribute should be queued for scraping.
        """
        scheme, netloc, path, _, _, _ = urlparse(link)
        if path.endswith(self.source_extensions + self.binary_extensions +
                         self.excluded_extensions):
            result = False
        elif self.skip_externals and not link.startswith(self.base_url):
            result = False
        elif not referrer.startswith(self.base_url):
            result = False
        elif rel not in ('homepage', 'download'):
            result = False
        elif scheme not in ('http', 'https', 'ftp'):
            result = False
        elif self._is_platform_dependent(link):
            result = False
        else:
            host = netloc.split(':', 1)[0]
            if host.lower() == 'localhost':
                result = False
            else:
                result = True
        logger.debug('should_queue: %s (%s) from %s -> %s', link, rel,
                     referrer, result)
        return result

    def _fetch(self):
        """
        Get a URL to fetch from the work queue, get the HTML page, examine its
        links for download candidates and candidates for further scraping.

        This is a handy method to run in a thread.
        """
        while True:
            url = self._to_fetch.get()
            try:
                if url:
                    page = self.get_page(url)
                    if page is None:    # e.g. after an error
                        continue
                    for link, rel in page.links:
                        if link not in self._seen:
                            try:
                                self._seen.add(link)
                                if (not self._process_download(link) and
                                    self._should_queue(link, url, rel)):
                                    logger.debug('Queueing %s from %s', link, url)
                                    self._to_fetch.put(link)
                            except MetadataInvalidError:  # e.g. invalid versions
                                pass
            except Exception as e:  # pragma: no cover
                self.errors.put(text_type(e))
            finally:
                # always do this, to avoid hangs :-)
                self._to_fetch.task_done()
            if not url:
                #logger.debug('Sentinel seen, quitting.')
                break

    def get_page(self, url):
        """
        Get the HTML for an URL, possibly from an in-memory cache.

        XXX TODO Note: this cache is never actually cleared. It's assumed that
        the data won't get stale over the lifetime of a locator instance (not
        necessarily true for the default_locator).
        """
        # http://peak.telecommunity.com/DevCenter/EasyInstall#package-index-api
        scheme, netloc, path, _, _, _ = urlparse(url)
        if scheme == 'file' and os.path.isdir(url2pathname(path)):
            url = urljoin(ensure_slash(url), 'index.html')

        if url in self._page_cache:
            result = self._page_cache[url]
            logger.debug('Returning %s from cache: %s', url, result)
        else:
            host = netloc.split(':', 1)[0]
            result = None
            if host in self._bad_hosts:
                logger.debug('Skipping %s due to bad host %s', url, host)
            else:
                req = Request(url, headers={'Accept-encoding': 'identity'})
                try:
                    logger.debug('Fetching %s', url)
                    resp = self.opener.open(req, timeout=self.timeout)
                    logger.debug('Fetched %s', url)
                    headers = resp.info()
                    content_type = headers.get('Content-Type', '')
                    if HTML_CONTENT_TYPE.match(content_type):
                        final_url = resp.geturl()
                        data = resp.read()
                        encoding = headers.get('Content-Encoding')
                        if encoding:
                            decoder = self.decoders[encoding]   # fail if not found
                            data = decoder(data)
                        encoding = 'utf-8'
                        m = CHARSET.search(content_type)
                        if m:
                            encoding = m.group(1)
                        try:
                            data = data.decode(encoding)
                        except UnicodeError:  # pragma: no cover
                            data = data.decode('latin-1')    # fallback
                        result = Page(data, final_url)
                        self._page_cache[final_url] = result
                except HTTPError as e:
                    if e.code != 404:
                        logger.exception('Fetch failed: %s: %s', url, e)
                except URLError as e:  # pragma: no cover
                    logger.exception('Fetch failed: %s: %s', url, e)
                    with self._lock:
                        self._bad_hosts.add(host)
                except Exception as e:  # pragma: no cover
                    logger.exception('Fetch failed: %s: %s', url, e)
                finally:
                    self._page_cache[url] = result   # even if None (failure)
        return result

    _distname_re = re.compile(']*>([^<]+)<')

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        result = set()
        page = self.get_page(self.base_url)
        if not page:
            raise DistlibException('Unable to get %s' % self.base_url)
        for match in self._distname_re.finditer(page.data):
            result.add(match.group(1))
        return result

class DirectoryLocator(Locator):
    """
    This class locates distributions in a directory tree.
    """

    def __init__(self, path, **kwargs):
        """
        Initialise an instance.
        :param path: The root of the directory tree to search.
        :param kwargs: Passed to the superclass constructor,
                       except for:
                       * recursive - if True (the default), subdirectories are
                         recursed into. If False, only the top-level directory
                         is searched,
        """
        self.recursive = kwargs.pop('recursive', True)
        super(DirectoryLocator, self).__init__(**kwargs)
        path = os.path.abspath(path)
        if not os.path.isdir(path):  # pragma: no cover
            raise DistlibException('Not a directory: %r' % path)
        self.base_dir = path

    def should_include(self, filename, parent):
        """
        Should a filename be considered as a candidate for a distribution
        archive? As well as the filename, the directory which contains it
        is provided, though not used by the current implementation.
        """
        return filename.endswith(self.downloadable_extensions)

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        for root, dirs, files in os.walk(self.base_dir):
            for fn in files:
                if self.should_include(fn, root):
                    fn = os.path.join(root, fn)
                    url = urlunparse(('file', '',
                                      pathname2url(os.path.abspath(fn)),
                                      '', '', ''))
                    info = self.convert_url_to_download_info(url, name)
                    if info:
                        self._update_version_data(result, info)
            if not self.recursive:
                break
        return result

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        result = set()
        for root, dirs, files in os.walk(self.base_dir):
            for fn in files:
                if self.should_include(fn, root):
                    fn = os.path.join(root, fn)
                    url = urlunparse(('file', '',
                                      pathname2url(os.path.abspath(fn)),
                                      '', '', ''))
                    info = self.convert_url_to_download_info(url, None)
                    if info:
                        result.add(info['name'])
            if not self.recursive:
                break
        return result

class JSONLocator(Locator):
    """
    This locator uses special extended metadata (not available on PyPI) and is
    the basis of performant dependency resolution in distlib. Other locators
    require archive downloads before dependencies can be determined! As you
    might imagine, that can be slow.
    """
    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        raise NotImplementedError('Not available from this locator')

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        data = get_project_data(name)
        if data:
            for info in data.get('files', []):
                if info['ptype'] != 'sdist' or info['pyversion'] != 'source':
                    continue
                # We don't store summary in project metadata as it makes
                # the data bigger for no benefit during dependency
                # resolution
                dist = make_dist(data['name'], info['version'],
                                 summary=data.get('summary',
                                                  'Placeholder for summary'),
                                 scheme=self.scheme)
                md = dist.metadata
                md.source_url = info['url']
                # TODO SHA256 digest
                if 'digest' in info and info['digest']:
                    dist.digest = ('md5', info['digest'])
                md.dependencies = info.get('requirements', {})
                dist.exports = info.get('exports', {})
                result[dist.version] = dist
                result['urls'].setdefault(dist.version, set()).add(info['url'])
        return result

class DistPathLocator(Locator):
    """
    This locator finds installed distributions in a path. It can be useful for
    adding to an :class:`AggregatingLocator`.
    """
    def __init__(self, distpath, **kwargs):
        """
        Initialise an instance.

        :param distpath: A :class:`DistributionPath` instance to search.
        """
        super(DistPathLocator, self).__init__(**kwargs)
        assert isinstance(distpath, DistributionPath)
        self.distpath = distpath

    def _get_project(self, name):
        dist = self.distpath.get_distribution(name)
        if dist is None:
            result = {'urls': {}, 'digests': {}}
        else:
            result = {
                dist.version: dist,
                'urls': {dist.version: set([dist.source_url])},
                'digests': {dist.version: set([None])}
            }
        return result


class AggregatingLocator(Locator):
    """
    This class allows you to chain and/or merge a list of locators.
    """
    def __init__(self, *locators, **kwargs):
        """
        Initialise an instance.

        :param locators: The list of locators to search.
        :param kwargs: Passed to the superclass constructor,
                       except for:
                       * merge - if False (the default), the first successful
                         search from any of the locators is returned. If True,
                         the results from all locators are merged (this can be
                         slow).
        """
        self.merge = kwargs.pop('merge', False)
        self.locators = locators
        super(AggregatingLocator, self).__init__(**kwargs)

    def clear_cache(self):
        super(AggregatingLocator, self).clear_cache()
        for locator in self.locators:
            locator.clear_cache()

    def _set_scheme(self, value):
        self._scheme = value
        for locator in self.locators:
            locator.scheme = value

    scheme = property(Locator.scheme.fget, _set_scheme)

    def _get_project(self, name):
        result = {}
        for locator in self.locators:
            d = locator.get_project(name)
            if d:
                if self.merge:
                    files = result.get('urls', {})
                    digests = result.get('digests', {})
                    # next line could overwrite result['urls'], result['digests']
                    result.update(d)
                    df = result.get('urls')
                    if files and df:
                        for k, v in files.items():
                            if k in df:
                                df[k] |= v
                            else:
                                df[k] = v
                    dd = result.get('digests')
                    if digests and dd:
                        dd.update(digests)
                else:
                    # See issue #18. If any dists are found and we're looking
                    # for specific constraints, we only return something if
                    # a match is found. For example, if a DirectoryLocator
                    # returns just foo (1.0) while we're looking for
                    # foo (>= 2.0), we'll pretend there was nothing there so
                    # that subsequent locators can be queried. Otherwise we
                    # would just return foo (1.0) which would then lead to a
                    # failure to find foo (>= 2.0), because other locators
                    # weren't searched. Note that this only matters when
                    # merge=False.
                    if self.matcher is None:
                        found = True
                    else:
                        found = False
                        for k in d:
                            if self.matcher.match(k):
                                found = True
                                break
                    if found:
                        result = d
                        break
        return result

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        result = set()
        for locator in self.locators:
            try:
                result |= locator.get_distribution_names()
            except NotImplementedError:
                pass
        return result


# We use a legacy scheme simply because most of the dists on PyPI use legacy
# versions which don't conform to PEP 426 / PEP 440.
default_locator = AggregatingLocator(
                    JSONLocator(),
                    SimpleScrapingLocator('https://pypi.org/simple/',
                                          timeout=3.0),
                    scheme='legacy')

locate = default_locator.locate

NAME_VERSION_RE = re.compile(r'(?P[\w-]+)\s*'
                             r'\(\s*(==\s*)?(?P[^)]+)\)$')

class DependencyFinder(object):
    """
    Locate dependencies for distributions.
    """

    def __init__(self, locator=None):
        """
        Initialise an instance, using the specified locator
        to locate distributions.
        """
        self.locator = locator or default_locator
        self.scheme = get_scheme(self.locator.scheme)

    def add_distribution(self, dist):
        """
        Add a distribution to the finder. This will update internal information
        about who provides what.
        :param dist: The distribution to add.
        """
        logger.debug('adding distribution %s', dist)
        name = dist.key
        self.dists_by_name[name] = dist
        self.dists[(name, dist.version)] = dist
        for p in dist.provides:
            name, version = parse_name_and_version(p)
            logger.debug('Add to provided: %s, %s, %s', name, version, dist)
            self.provided.setdefault(name, set()).add((version, dist))

    def remove_distribution(self, dist):
        """
        Remove a distribution from the finder. This will update internal
        information about who provides what.
        :param dist: The distribution to remove.
        """
        logger.debug('removing distribution %s', dist)
        name = dist.key
        del self.dists_by_name[name]
        del self.dists[(name, dist.version)]
        for p in dist.provides:
            name, version = parse_name_and_version(p)
            logger.debug('Remove from provided: %s, %s, %s', name, version, dist)
            s = self.provided[name]
            s.remove((version, dist))
            if not s:
                del self.provided[name]

    def get_matcher(self, reqt):
        """
        Get a version matcher for a requirement.
        :param reqt: The requirement
        :type reqt: str
        :return: A version matcher (an instance of
                 :class:`distlib.version.Matcher`).
        """
        try:
            matcher = self.scheme.matcher(reqt)
        except UnsupportedVersionError:  # pragma: no cover
            # XXX compat-mode if cannot read the version
            name = reqt.split()[0]
            matcher = self.scheme.matcher(name)
        return matcher

    def find_providers(self, reqt):
        """
        Find the distributions which can fulfill a requirement.

        :param reqt: The requirement.
         :type reqt: str
        :return: A set of distribution which can fulfill the requirement.
        """
        matcher = self.get_matcher(reqt)
        name = matcher.key   # case-insensitive
        result = set()
        provided = self.provided
        if name in provided:
            for version, provider in provided[name]:
                try:
                    match = matcher.match(version)
                except UnsupportedVersionError:
                    match = False

                if match:
                    result.add(provider)
                    break
        return result

    def try_to_replace(self, provider, other, problems):
        """
        Attempt to replace one provider with another. This is typically used
        when resolving dependencies from multiple sources, e.g. A requires
        (B >= 1.0) while C requires (B >= 1.1).

        For successful replacement, ``provider`` must meet all the requirements
        which ``other`` fulfills.

        :param provider: The provider we are trying to replace with.
        :param other: The provider we're trying to replace.
        :param problems: If False is returned, this will contain what
                         problems prevented replacement. This is currently
                         a tuple of the literal string 'cantreplace',
                         ``provider``, ``other``  and the set of requirements
                         that ``provider`` couldn't fulfill.
        :return: True if we can replace ``other`` with ``provider``, else
                 False.
        """
        rlist = self.reqts[other]
        unmatched = set()
        for s in rlist:
            matcher = self.get_matcher(s)
            if not matcher.match(provider.version):
                unmatched.add(s)
        if unmatched:
            # can't replace other with provider
            problems.add(('cantreplace', provider, other,
                          frozenset(unmatched)))
            result = False
        else:
            # can replace other with provider
            self.remove_distribution(other)
            del self.reqts[other]
            for s in rlist:
                self.reqts.setdefault(provider, set()).add(s)
            self.add_distribution(provider)
            result = True
        return result

    def find(self, requirement, meta_extras=None, prereleases=False):
        """
        Find a distribution and all distributions it depends on.

        :param requirement: The requirement specifying the distribution to
                            find, or a Distribution instance.
        :param meta_extras: A list of meta extras such as :test:, :build: and
                            so on.
        :param prereleases: If ``True``, allow pre-release versions to be
                            returned - otherwise, don't return prereleases
                            unless they're all that's available.

        Return a set of :class:`Distribution` instances and a set of
        problems.

        The distributions returned should be such that they have the
        :attr:`required` attribute set to ``True`` if they were
        from the ``requirement`` passed to ``find()``, and they have the
        :attr:`build_time_dependency` attribute set to ``True`` unless they
        are post-installation dependencies of the ``requirement``.

        The problems should be a tuple consisting of the string
        ``'unsatisfied'`` and the requirement which couldn't be satisfied
        by any distribution known to the locator.
        """

        self.provided = {}
        self.dists = {}
        self.dists_by_name = {}
        self.reqts = {}

        meta_extras = set(meta_extras or [])
        if ':*:' in meta_extras:
            meta_extras.remove(':*:')
            # :meta: and :run: are implicitly included
            meta_extras |= set([':test:', ':build:', ':dev:'])

        if isinstance(requirement, Distribution):
            dist = odist = requirement
            logger.debug('passed %s as requirement', odist)
        else:
            dist = odist = self.locator.locate(requirement,
                                               prereleases=prereleases)
            if dist is None:
                raise DistlibException('Unable to locate %r' % requirement)
            logger.debug('located %s', odist)
        dist.requested = True
        problems = set()
        todo = set([dist])
        install_dists = set([odist])
        while todo:
            dist = todo.pop()
            name = dist.key     # case-insensitive
            if name not in self.dists_by_name:
                self.add_distribution(dist)
            else:
                #import pdb; pdb.set_trace()
                other = self.dists_by_name[name]
                if other != dist:
                    self.try_to_replace(dist, other, problems)

            ireqts = dist.run_requires | dist.meta_requires
            sreqts = dist.build_requires
            ereqts = set()
            if meta_extras and dist in install_dists:
                for key in ('test', 'build', 'dev'):
                    e = ':%s:' % key
                    if e in meta_extras:
                        ereqts |= getattr(dist, '%s_requires' % key)
            all_reqts = ireqts | sreqts | ereqts
            for r in all_reqts:
                providers = self.find_providers(r)
                if not providers:
                    logger.debug('No providers found for %r', r)
                    provider = self.locator.locate(r, prereleases=prereleases)
                    # If no provider is found and we didn't consider
                    # prereleases, consider them now.
                    if provider is None and not prereleases:
                        provider = self.locator.locate(r, prereleases=True)
                    if provider is None:
                        logger.debug('Cannot satisfy %r', r)
                        problems.add(('unsatisfied', r))
                    else:
                        n, v = provider.key, provider.version
                        if (n, v) not in self.dists:
                            todo.add(provider)
                        providers.add(provider)
                        if r in ireqts and dist in install_dists:
                            install_dists.add(provider)
                            logger.debug('Adding %s to install_dists',
                                         provider.name_and_version)
                for p in providers:
                    name = p.key
                    if name not in self.dists_by_name:
                        self.reqts.setdefault(p, set()).add(r)
                    else:
                        other = self.dists_by_name[name]
                        if other != p:
                            # see if other can be replaced by p
                            self.try_to_replace(p, other, problems)

        dists = set(self.dists.values())
        for dist in dists:
            dist.build_time_dependency = dist not in install_dists
            if dist.build_time_dependency:
                logger.debug('%s is a build-time dependency only.',
                             dist.name_and_version)
        logger.debug('find done for %s', odist)
        return dists, problems
site-packages/pip/_vendor/distlib/wheel.py000064400000116765151464452710014651 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from __future__ import unicode_literals

import base64
import codecs
import datetime
import distutils.util
from email import message_from_file
import hashlib
import imp
import json
import logging
import os
import posixpath
import re
import shutil
import sys
import tempfile
import zipfile

from . import __version__, DistlibException
from .compat import sysconfig, ZipFile, fsdecode, text_type, filter
from .database import InstalledDistribution
from .metadata import Metadata, METADATA_FILENAME, WHEEL_METADATA_FILENAME
from .util import (FileOperator, convert_path, CSVReader, CSVWriter, Cache,
                   cached_property, get_cache_base, read_exports, tempdir)
from .version import NormalizedVersion, UnsupportedVersionError

logger = logging.getLogger(__name__)

cache = None    # created when needed

if hasattr(sys, 'pypy_version_info'):  # pragma: no cover
    IMP_PREFIX = 'pp'
elif sys.platform.startswith('java'):  # pragma: no cover
    IMP_PREFIX = 'jy'
elif sys.platform == 'cli':  # pragma: no cover
    IMP_PREFIX = 'ip'
else:
    IMP_PREFIX = 'cp'

VER_SUFFIX = sysconfig.get_config_var('py_version_nodot')
if not VER_SUFFIX:   # pragma: no cover
    VER_SUFFIX = '%s%s' % sys.version_info[:2]
PYVER = 'py' + VER_SUFFIX
IMPVER = IMP_PREFIX + VER_SUFFIX

ARCH = distutils.util.get_platform().replace('-', '_').replace('.', '_')

ABI = sysconfig.get_config_var('SOABI')
if ABI and ABI.startswith('cpython-'):
    ABI = ABI.replace('cpython-', 'cp')
else:
    def _derive_abi():
        parts = ['cp', VER_SUFFIX]
        if sysconfig.get_config_var('Py_DEBUG'):
            parts.append('d')
        if sysconfig.get_config_var('WITH_PYMALLOC'):
            parts.append('m')
        if sysconfig.get_config_var('Py_UNICODE_SIZE') == 4:
            parts.append('u')
        return ''.join(parts)
    ABI = _derive_abi()
    del _derive_abi

FILENAME_RE = re.compile(r'''
(?P[^-]+)
-(?P\d+[^-]*)
(-(?P\d+[^-]*))?
-(?P\w+\d+(\.\w+\d+)*)
-(?P\w+)
-(?P\w+(\.\w+)*)
\.whl$
''', re.IGNORECASE | re.VERBOSE)

NAME_VERSION_RE = re.compile(r'''
(?P[^-]+)
-(?P\d+[^-]*)
(-(?P\d+[^-]*))?$
''', re.IGNORECASE | re.VERBOSE)

SHEBANG_RE = re.compile(br'\s*#![^\r\n]*')
SHEBANG_DETAIL_RE = re.compile(br'^(\s*#!("[^"]+"|\S+))\s+(.*)$')
SHEBANG_PYTHON = b'#!python'
SHEBANG_PYTHONW = b'#!pythonw'

if os.sep == '/':
    to_posix = lambda o: o
else:
    to_posix = lambda o: o.replace(os.sep, '/')


class Mounter(object):
    def __init__(self):
        self.impure_wheels = {}
        self.libs = {}

    def add(self, pathname, extensions):
        self.impure_wheels[pathname] = extensions
        self.libs.update(extensions)

    def remove(self, pathname):
        extensions = self.impure_wheels.pop(pathname)
        for k, v in extensions:
            if k in self.libs:
                del self.libs[k]

    def find_module(self, fullname, path=None):
        if fullname in self.libs:
            result = self
        else:
            result = None
        return result

    def load_module(self, fullname):
        if fullname in sys.modules:
            result = sys.modules[fullname]
        else:
            if fullname not in self.libs:
                raise ImportError('unable to find extension for %s' % fullname)
            result = imp.load_dynamic(fullname, self.libs[fullname])
            result.__loader__ = self
            parts = fullname.rsplit('.', 1)
            if len(parts) > 1:
                result.__package__ = parts[0]
        return result

_hook = Mounter()


class Wheel(object):
    """
    Class to build and install from Wheel files (PEP 427).
    """

    wheel_version = (1, 1)
    hash_kind = 'sha256'

    def __init__(self, filename=None, sign=False, verify=False):
        """
        Initialise an instance using a (valid) filename.
        """
        self.sign = sign
        self.should_verify = verify
        self.buildver = ''
        self.pyver = [PYVER]
        self.abi = ['none']
        self.arch = ['any']
        self.dirname = os.getcwd()
        if filename is None:
            self.name = 'dummy'
            self.version = '0.1'
            self._filename = self.filename
        else:
            m = NAME_VERSION_RE.match(filename)
            if m:
                info = m.groupdict('')
                self.name = info['nm']
                # Reinstate the local version separator
                self.version = info['vn'].replace('_', '-')
                self.buildver = info['bn']
                self._filename = self.filename
            else:
                dirname, filename = os.path.split(filename)
                m = FILENAME_RE.match(filename)
                if not m:
                    raise DistlibException('Invalid name or '
                                           'filename: %r' % filename)
                if dirname:
                    self.dirname = os.path.abspath(dirname)
                self._filename = filename
                info = m.groupdict('')
                self.name = info['nm']
                self.version = info['vn']
                self.buildver = info['bn']
                self.pyver = info['py'].split('.')
                self.abi = info['bi'].split('.')
                self.arch = info['ar'].split('.')

    @property
    def filename(self):
        """
        Build and return a filename from the various components.
        """
        if self.buildver:
            buildver = '-' + self.buildver
        else:
            buildver = ''
        pyver = '.'.join(self.pyver)
        abi = '.'.join(self.abi)
        arch = '.'.join(self.arch)
        # replace - with _ as a local version separator
        version = self.version.replace('-', '_')
        return '%s-%s%s-%s-%s-%s.whl' % (self.name, version, buildver,
                                         pyver, abi, arch)

    @property
    def exists(self):
        path = os.path.join(self.dirname, self.filename)
        return os.path.isfile(path)

    @property
    def tags(self):
        for pyver in self.pyver:
            for abi in self.abi:
                for arch in self.arch:
                    yield pyver, abi, arch

    @cached_property
    def metadata(self):
        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        wrapper = codecs.getreader('utf-8')
        with ZipFile(pathname, 'r') as zf:
            wheel_metadata = self.get_wheel_metadata(zf)
            wv = wheel_metadata['Wheel-Version'].split('.', 1)
            file_version = tuple([int(i) for i in wv])
            if file_version < (1, 1):
                fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME, 'METADATA']
            else:
                fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME]
            result = None
            for fn in fns:
                try:
                    metadata_filename = posixpath.join(info_dir, fn)
                    with zf.open(metadata_filename) as bf:
                        wf = wrapper(bf)
                        result = Metadata(fileobj=wf)
                        if result:
                            break
                except KeyError:
                    pass
            if not result:
                raise ValueError('Invalid wheel, because metadata is '
                                 'missing: looked in %s' % ', '.join(fns))
        return result

    def get_wheel_metadata(self, zf):
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        metadata_filename = posixpath.join(info_dir, 'WHEEL')
        with zf.open(metadata_filename) as bf:
            wf = codecs.getreader('utf-8')(bf)
            message = message_from_file(wf)
        return dict(message)

    @cached_property
    def info(self):
        pathname = os.path.join(self.dirname, self.filename)
        with ZipFile(pathname, 'r') as zf:
            result = self.get_wheel_metadata(zf)
        return result

    def process_shebang(self, data):
        m = SHEBANG_RE.match(data)
        if m:
            end = m.end()
            shebang, data_after_shebang = data[:end], data[end:]
            # Preserve any arguments after the interpreter
            if b'pythonw' in shebang.lower():
                shebang_python = SHEBANG_PYTHONW
            else:
                shebang_python = SHEBANG_PYTHON
            m = SHEBANG_DETAIL_RE.match(shebang)
            if m:
                args = b' ' + m.groups()[-1]
            else:
                args = b''
            shebang = shebang_python + args
            data = shebang + data_after_shebang
        else:
            cr = data.find(b'\r')
            lf = data.find(b'\n')
            if cr < 0 or cr > lf:
                term = b'\n'
            else:
                if data[cr:cr + 2] == b'\r\n':
                    term = b'\r\n'
                else:
                    term = b'\r'
            data = SHEBANG_PYTHON + term + data
        return data

    def get_hash(self, data, hash_kind=None):
        if hash_kind is None:
            hash_kind = self.hash_kind
        try:
            hasher = getattr(hashlib, hash_kind)
        except AttributeError:
            raise DistlibException('Unsupported hash algorithm: %r' % hash_kind)
        result = hasher(data).digest()
        result = base64.urlsafe_b64encode(result).rstrip(b'=').decode('ascii')
        return hash_kind, result

    def write_record(self, records, record_path, base):
        records = list(records) # make a copy for sorting
        p = to_posix(os.path.relpath(record_path, base))
        records.append((p, '', ''))
        records.sort()
        with CSVWriter(record_path) as writer:
            for row in records:
                writer.writerow(row)

    def write_records(self, info, libdir, archive_paths):
        records = []
        distinfo, info_dir = info
        hasher = getattr(hashlib, self.hash_kind)
        for ap, p in archive_paths:
            with open(p, 'rb') as f:
                data = f.read()
            digest = '%s=%s' % self.get_hash(data)
            size = os.path.getsize(p)
            records.append((ap, digest, size))

        p = os.path.join(distinfo, 'RECORD')
        self.write_record(records, p, libdir)
        ap = to_posix(os.path.join(info_dir, 'RECORD'))
        archive_paths.append((ap, p))

    def build_zip(self, pathname, archive_paths):
        with ZipFile(pathname, 'w', zipfile.ZIP_DEFLATED) as zf:
            for ap, p in archive_paths:
                logger.debug('Wrote %s to %s in wheel', p, ap)
                zf.write(p, ap)

    def build(self, paths, tags=None, wheel_version=None):
        """
        Build a wheel from files in specified paths, and use any specified tags
        when determining the name of the wheel.
        """
        if tags is None:
            tags = {}

        libkey = list(filter(lambda o: o in paths, ('purelib', 'platlib')))[0]
        if libkey == 'platlib':
            is_pure = 'false'
            default_pyver = [IMPVER]
            default_abi = [ABI]
            default_arch = [ARCH]
        else:
            is_pure = 'true'
            default_pyver = [PYVER]
            default_abi = ['none']
            default_arch = ['any']

        self.pyver = tags.get('pyver', default_pyver)
        self.abi = tags.get('abi', default_abi)
        self.arch = tags.get('arch', default_arch)

        libdir = paths[libkey]

        name_ver = '%s-%s' % (self.name, self.version)
        data_dir = '%s.data' % name_ver
        info_dir = '%s.dist-info' % name_ver

        archive_paths = []

        # First, stuff which is not in site-packages
        for key in ('data', 'headers', 'scripts'):
            if key not in paths:
                continue
            path = paths[key]
            if os.path.isdir(path):
                for root, dirs, files in os.walk(path):
                    for fn in files:
                        p = fsdecode(os.path.join(root, fn))
                        rp = os.path.relpath(p, path)
                        ap = to_posix(os.path.join(data_dir, key, rp))
                        archive_paths.append((ap, p))
                        if key == 'scripts' and not p.endswith('.exe'):
                            with open(p, 'rb') as f:
                                data = f.read()
                            data = self.process_shebang(data)
                            with open(p, 'wb') as f:
                                f.write(data)

        # Now, stuff which is in site-packages, other than the
        # distinfo stuff.
        path = libdir
        distinfo = None
        for root, dirs, files in os.walk(path):
            if root == path:
                # At the top level only, save distinfo for later
                # and skip it for now
                for i, dn in enumerate(dirs):
                    dn = fsdecode(dn)
                    if dn.endswith('.dist-info'):
                        distinfo = os.path.join(root, dn)
                        del dirs[i]
                        break
                assert distinfo, '.dist-info directory expected, not found'

            for fn in files:
                # comment out next suite to leave .pyc files in
                if fsdecode(fn).endswith(('.pyc', '.pyo')):
                    continue
                p = os.path.join(root, fn)
                rp = to_posix(os.path.relpath(p, path))
                archive_paths.append((rp, p))

        # Now distinfo. Assumed to be flat, i.e. os.listdir is enough.
        files = os.listdir(distinfo)
        for fn in files:
            if fn not in ('RECORD', 'INSTALLER', 'SHARED', 'WHEEL'):
                p = fsdecode(os.path.join(distinfo, fn))
                ap = to_posix(os.path.join(info_dir, fn))
                archive_paths.append((ap, p))

        wheel_metadata = [
            'Wheel-Version: %d.%d' % (wheel_version or self.wheel_version),
            'Generator: distlib %s' % __version__,
            'Root-Is-Purelib: %s' % is_pure,
        ]
        for pyver, abi, arch in self.tags:
            wheel_metadata.append('Tag: %s-%s-%s' % (pyver, abi, arch))
        p = os.path.join(distinfo, 'WHEEL')
        with open(p, 'w') as f:
            f.write('\n'.join(wheel_metadata))
        ap = to_posix(os.path.join(info_dir, 'WHEEL'))
        archive_paths.append((ap, p))

        # Now, at last, RECORD.
        # Paths in here are archive paths - nothing else makes sense.
        self.write_records((distinfo, info_dir), libdir, archive_paths)
        # Now, ready to build the zip file
        pathname = os.path.join(self.dirname, self.filename)
        self.build_zip(pathname, archive_paths)
        return pathname

    def skip_entry(self, arcname):
        """
        Determine whether an archive entry should be skipped when verifying
        or installing.
        """
        # The signature file won't be in RECORD,
        # and we  don't currently don't do anything with it
        # We also skip directories, as they won't be in RECORD
        # either. See:
        #
        # https://github.com/pypa/wheel/issues/294
        # https://github.com/pypa/wheel/issues/287
        # https://github.com/pypa/wheel/pull/289
        #
        return arcname.endswith(('/', '/RECORD.jws'))

    def install(self, paths, maker, **kwargs):
        """
        Install a wheel to the specified paths. If kwarg ``warner`` is
        specified, it should be a callable, which will be called with two
        tuples indicating the wheel version of this software and the wheel
        version in the file, if there is a discrepancy in the versions.
        This can be used to issue any warnings to raise any exceptions.
        If kwarg ``lib_only`` is True, only the purelib/platlib files are
        installed, and the headers, scripts, data and dist-info metadata are
        not written. If kwarg ``bytecode_hashed_invalidation`` is True, written
        bytecode will try to use file-hash based invalidation (PEP-552) on
        supported interpreter versions (CPython 2.7+).

        The return value is a :class:`InstalledDistribution` instance unless
        ``options.lib_only`` is True, in which case the return value is ``None``.
        """

        dry_run = maker.dry_run
        warner = kwargs.get('warner')
        lib_only = kwargs.get('lib_only', False)
        bc_hashed_invalidation = kwargs.get('bytecode_hashed_invalidation', False)

        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        data_dir = '%s.data' % name_ver
        info_dir = '%s.dist-info' % name_ver

        metadata_name = posixpath.join(info_dir, METADATA_FILENAME)
        wheel_metadata_name = posixpath.join(info_dir, 'WHEEL')
        record_name = posixpath.join(info_dir, 'RECORD')

        wrapper = codecs.getreader('utf-8')

        with ZipFile(pathname, 'r') as zf:
            with zf.open(wheel_metadata_name) as bwf:
                wf = wrapper(bwf)
                message = message_from_file(wf)
            wv = message['Wheel-Version'].split('.', 1)
            file_version = tuple([int(i) for i in wv])
            if (file_version != self.wheel_version) and warner:
                warner(self.wheel_version, file_version)

            if message['Root-Is-Purelib'] == 'true':
                libdir = paths['purelib']
            else:
                libdir = paths['platlib']

            records = {}
            with zf.open(record_name) as bf:
                with CSVReader(stream=bf) as reader:
                    for row in reader:
                        p = row[0]
                        records[p] = row

            data_pfx = posixpath.join(data_dir, '')
            info_pfx = posixpath.join(info_dir, '')
            script_pfx = posixpath.join(data_dir, 'scripts', '')

            # make a new instance rather than a copy of maker's,
            # as we mutate it
            fileop = FileOperator(dry_run=dry_run)
            fileop.record = True    # so we can rollback if needed

            bc = not sys.dont_write_bytecode    # Double negatives. Lovely!

            outfiles = []   # for RECORD writing

            # for script copying/shebang processing
            workdir = tempfile.mkdtemp()
            # set target dir later
            # we default add_launchers to False, as the
            # Python Launcher should be used instead
            maker.source_dir = workdir
            maker.target_dir = None
            try:
                for zinfo in zf.infolist():
                    arcname = zinfo.filename
                    if isinstance(arcname, text_type):
                        u_arcname = arcname
                    else:
                        u_arcname = arcname.decode('utf-8')
                    if self.skip_entry(u_arcname):
                        continue
                    row = records[u_arcname]
                    if row[2] and str(zinfo.file_size) != row[2]:
                        raise DistlibException('size mismatch for '
                                               '%s' % u_arcname)
                    if row[1]:
                        kind, value = row[1].split('=', 1)
                        with zf.open(arcname) as bf:
                            data = bf.read()
                        _, digest = self.get_hash(data, kind)
                        if digest != value:
                            raise DistlibException('digest mismatch for '
                                                   '%s' % arcname)

                    if lib_only and u_arcname.startswith((info_pfx, data_pfx)):
                        logger.debug('lib_only: skipping %s', u_arcname)
                        continue
                    is_script = (u_arcname.startswith(script_pfx)
                                 and not u_arcname.endswith('.exe'))

                    if u_arcname.startswith(data_pfx):
                        _, where, rp = u_arcname.split('/', 2)
                        outfile = os.path.join(paths[where], convert_path(rp))
                    else:
                        # meant for site-packages.
                        if u_arcname in (wheel_metadata_name, record_name):
                            continue
                        outfile = os.path.join(libdir, convert_path(u_arcname))
                    if not is_script:
                        with zf.open(arcname) as bf:
                            fileop.copy_stream(bf, outfile)
                        outfiles.append(outfile)
                        # Double check the digest of the written file
                        if not dry_run and row[1]:
                            with open(outfile, 'rb') as bf:
                                data = bf.read()
                                _, newdigest = self.get_hash(data, kind)
                                if newdigest != digest:
                                    raise DistlibException('digest mismatch '
                                                           'on write for '
                                                           '%s' % outfile)
                        if bc and outfile.endswith('.py'):
                            try:
                                pyc = fileop.byte_compile(outfile,
                                                          hashed_invalidation=bc_hashed_invalidation)
                                outfiles.append(pyc)
                            except Exception:
                                # Don't give up if byte-compilation fails,
                                # but log it and perhaps warn the user
                                logger.warning('Byte-compilation failed',
                                               exc_info=True)
                    else:
                        fn = os.path.basename(convert_path(arcname))
                        workname = os.path.join(workdir, fn)
                        with zf.open(arcname) as bf:
                            fileop.copy_stream(bf, workname)

                        dn, fn = os.path.split(outfile)
                        maker.target_dir = dn
                        filenames = maker.make(fn)
                        fileop.set_executable_mode(filenames)
                        outfiles.extend(filenames)

                if lib_only:
                    logger.debug('lib_only: returning None')
                    dist = None
                else:
                    # Generate scripts

                    # Try to get pydist.json so we can see if there are
                    # any commands to generate. If this fails (e.g. because
                    # of a legacy wheel), log a warning but don't give up.
                    commands = None
                    file_version = self.info['Wheel-Version']
                    if file_version == '1.0':
                        # Use legacy info
                        ep = posixpath.join(info_dir, 'entry_points.txt')
                        try:
                            with zf.open(ep) as bwf:
                                epdata = read_exports(bwf)
                            commands = {}
                            for key in ('console', 'gui'):
                                k = '%s_scripts' % key
                                if k in epdata:
                                    commands['wrap_%s' % key] = d = {}
                                    for v in epdata[k].values():
                                        s = '%s:%s' % (v.prefix, v.suffix)
                                        if v.flags:
                                            s += ' %s' % v.flags
                                        d[v.name] = s
                        except Exception:
                            logger.warning('Unable to read legacy script '
                                           'metadata, so cannot generate '
                                           'scripts')
                    else:
                        try:
                            with zf.open(metadata_name) as bwf:
                                wf = wrapper(bwf)
                                commands = json.load(wf).get('extensions')
                                if commands:
                                    commands = commands.get('python.commands')
                        except Exception:
                            logger.warning('Unable to read JSON metadata, so '
                                           'cannot generate scripts')
                    if commands:
                        console_scripts = commands.get('wrap_console', {})
                        gui_scripts = commands.get('wrap_gui', {})
                        if console_scripts or gui_scripts:
                            script_dir = paths.get('scripts', '')
                            if not os.path.isdir(script_dir):
                                raise ValueError('Valid script path not '
                                                 'specified')
                            maker.target_dir = script_dir
                            for k, v in console_scripts.items():
                                script = '%s = %s' % (k, v)
                                filenames = maker.make(script)
                                fileop.set_executable_mode(filenames)

                            if gui_scripts:
                                options = {'gui': True }
                                for k, v in gui_scripts.items():
                                    script = '%s = %s' % (k, v)
                                    filenames = maker.make(script, options)
                                    fileop.set_executable_mode(filenames)

                    p = os.path.join(libdir, info_dir)
                    dist = InstalledDistribution(p)

                    # Write SHARED
                    paths = dict(paths)     # don't change passed in dict
                    del paths['purelib']
                    del paths['platlib']
                    paths['lib'] = libdir
                    p = dist.write_shared_locations(paths, dry_run)
                    if p:
                        outfiles.append(p)

                    # Write RECORD
                    dist.write_installed_files(outfiles, paths['prefix'],
                                               dry_run)
                return dist
            except Exception:  # pragma: no cover
                logger.exception('installation failed.')
                fileop.rollback()
                raise
            finally:
                shutil.rmtree(workdir)

    def _get_dylib_cache(self):
        global cache
        if cache is None:
            # Use native string to avoid issues on 2.x: see Python #20140.
            base = os.path.join(get_cache_base(), str('dylib-cache'),
                                sys.version[:3])
            cache = Cache(base)
        return cache

    def _get_extensions(self):
        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        arcname = posixpath.join(info_dir, 'EXTENSIONS')
        wrapper = codecs.getreader('utf-8')
        result = []
        with ZipFile(pathname, 'r') as zf:
            try:
                with zf.open(arcname) as bf:
                    wf = wrapper(bf)
                    extensions = json.load(wf)
                    cache = self._get_dylib_cache()
                    prefix = cache.prefix_to_dir(pathname)
                    cache_base = os.path.join(cache.base, prefix)
                    if not os.path.isdir(cache_base):
                        os.makedirs(cache_base)
                    for name, relpath in extensions.items():
                        dest = os.path.join(cache_base, convert_path(relpath))
                        if not os.path.exists(dest):
                            extract = True
                        else:
                            file_time = os.stat(dest).st_mtime
                            file_time = datetime.datetime.fromtimestamp(file_time)
                            info = zf.getinfo(relpath)
                            wheel_time = datetime.datetime(*info.date_time)
                            extract = wheel_time > file_time
                        if extract:
                            zf.extract(relpath, cache_base)
                        result.append((name, dest))
            except KeyError:
                pass
        return result

    def is_compatible(self):
        """
        Determine if a wheel is compatible with the running system.
        """
        return is_compatible(self)

    def is_mountable(self):
        """
        Determine if a wheel is asserted as mountable by its metadata.
        """
        return True # for now - metadata details TBD

    def mount(self, append=False):
        pathname = os.path.abspath(os.path.join(self.dirname, self.filename))
        if not self.is_compatible():
            msg = 'Wheel %s not compatible with this Python.' % pathname
            raise DistlibException(msg)
        if not self.is_mountable():
            msg = 'Wheel %s is marked as not mountable.' % pathname
            raise DistlibException(msg)
        if pathname in sys.path:
            logger.debug('%s already in path', pathname)
        else:
            if append:
                sys.path.append(pathname)
            else:
                sys.path.insert(0, pathname)
            extensions = self._get_extensions()
            if extensions:
                if _hook not in sys.meta_path:
                    sys.meta_path.append(_hook)
                _hook.add(pathname, extensions)

    def unmount(self):
        pathname = os.path.abspath(os.path.join(self.dirname, self.filename))
        if pathname not in sys.path:
            logger.debug('%s not in path', pathname)
        else:
            sys.path.remove(pathname)
            if pathname in _hook.impure_wheels:
                _hook.remove(pathname)
            if not _hook.impure_wheels:
                if _hook in sys.meta_path:
                    sys.meta_path.remove(_hook)

    def verify(self):
        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        data_dir = '%s.data' % name_ver
        info_dir = '%s.dist-info' % name_ver

        metadata_name = posixpath.join(info_dir, METADATA_FILENAME)
        wheel_metadata_name = posixpath.join(info_dir, 'WHEEL')
        record_name = posixpath.join(info_dir, 'RECORD')

        wrapper = codecs.getreader('utf-8')

        with ZipFile(pathname, 'r') as zf:
            with zf.open(wheel_metadata_name) as bwf:
                wf = wrapper(bwf)
                message = message_from_file(wf)
            wv = message['Wheel-Version'].split('.', 1)
            file_version = tuple([int(i) for i in wv])
            # TODO version verification

            records = {}
            with zf.open(record_name) as bf:
                with CSVReader(stream=bf) as reader:
                    for row in reader:
                        p = row[0]
                        records[p] = row

            for zinfo in zf.infolist():
                arcname = zinfo.filename
                if isinstance(arcname, text_type):
                    u_arcname = arcname
                else:
                    u_arcname = arcname.decode('utf-8')
                # See issue #115: some wheels have .. in their entries, but
                # in the filename ... e.g. __main__..py ! So the check is
                # updated to look for .. in the directory portions
                p = u_arcname.split('/')
                if '..' in p:
                    raise DistlibException('invalid entry in '
                                           'wheel: %r' % u_arcname)

                if self.skip_entry(u_arcname):
                    continue
                row = records[u_arcname]
                if row[2] and str(zinfo.file_size) != row[2]:
                    raise DistlibException('size mismatch for '
                                           '%s' % u_arcname)
                if row[1]:
                    kind, value = row[1].split('=', 1)
                    with zf.open(arcname) as bf:
                        data = bf.read()
                    _, digest = self.get_hash(data, kind)
                    if digest != value:
                        raise DistlibException('digest mismatch for '
                                               '%s' % arcname)

    def update(self, modifier, dest_dir=None, **kwargs):
        """
        Update the contents of a wheel in a generic way. The modifier should
        be a callable which expects a dictionary argument: its keys are
        archive-entry paths, and its values are absolute filesystem paths
        where the contents the corresponding archive entries can be found. The
        modifier is free to change the contents of the files pointed to, add
        new entries and remove entries, before returning. This method will
        extract the entire contents of the wheel to a temporary location, call
        the modifier, and then use the passed (and possibly updated)
        dictionary to write a new wheel. If ``dest_dir`` is specified, the new
        wheel is written there -- otherwise, the original wheel is overwritten.

        The modifier should return True if it updated the wheel, else False.
        This method returns the same value the modifier returns.
        """

        def get_version(path_map, info_dir):
            version = path = None
            key = '%s/%s' % (info_dir, METADATA_FILENAME)
            if key not in path_map:
                key = '%s/PKG-INFO' % info_dir
            if key in path_map:
                path = path_map[key]
                version = Metadata(path=path).version
            return version, path

        def update_version(version, path):
            updated = None
            try:
                v = NormalizedVersion(version)
                i = version.find('-')
                if i < 0:
                    updated = '%s+1' % version
                else:
                    parts = [int(s) for s in version[i + 1:].split('.')]
                    parts[-1] += 1
                    updated = '%s+%s' % (version[:i],
                                         '.'.join(str(i) for i in parts))
            except UnsupportedVersionError:
                logger.debug('Cannot update non-compliant (PEP-440) '
                             'version %r', version)
            if updated:
                md = Metadata(path=path)
                md.version = updated
                legacy = not path.endswith(METADATA_FILENAME)
                md.write(path=path, legacy=legacy)
                logger.debug('Version updated from %r to %r', version,
                             updated)

        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        record_name = posixpath.join(info_dir, 'RECORD')
        with tempdir() as workdir:
            with ZipFile(pathname, 'r') as zf:
                path_map = {}
                for zinfo in zf.infolist():
                    arcname = zinfo.filename
                    if isinstance(arcname, text_type):
                        u_arcname = arcname
                    else:
                        u_arcname = arcname.decode('utf-8')
                    if u_arcname == record_name:
                        continue
                    if '..' in u_arcname:
                        raise DistlibException('invalid entry in '
                                               'wheel: %r' % u_arcname)
                    zf.extract(zinfo, workdir)
                    path = os.path.join(workdir, convert_path(u_arcname))
                    path_map[u_arcname] = path

            # Remember the version.
            original_version, _ = get_version(path_map, info_dir)
            # Files extracted. Call the modifier.
            modified = modifier(path_map, **kwargs)
            if modified:
                # Something changed - need to build a new wheel.
                current_version, path = get_version(path_map, info_dir)
                if current_version and (current_version == original_version):
                    # Add or update local version to signify changes.
                    update_version(current_version, path)
                # Decide where the new wheel goes.
                if dest_dir is None:
                    fd, newpath = tempfile.mkstemp(suffix='.whl',
                                                   prefix='wheel-update-',
                                                   dir=workdir)
                    os.close(fd)
                else:
                    if not os.path.isdir(dest_dir):
                        raise DistlibException('Not a directory: %r' % dest_dir)
                    newpath = os.path.join(dest_dir, self.filename)
                archive_paths = list(path_map.items())
                distinfo = os.path.join(workdir, info_dir)
                info = distinfo, info_dir
                self.write_records(info, workdir, archive_paths)
                self.build_zip(newpath, archive_paths)
                if dest_dir is None:
                    shutil.copyfile(newpath, pathname)
        return modified

def compatible_tags():
    """
    Return (pyver, abi, arch) tuples compatible with this Python.
    """
    versions = [VER_SUFFIX]
    major = VER_SUFFIX[0]
    for minor in range(sys.version_info[1] - 1, - 1, -1):
        versions.append(''.join([major, str(minor)]))

    abis = []
    for suffix, _, _ in imp.get_suffixes():
        if suffix.startswith('.abi'):
            abis.append(suffix.split('.', 2)[1])
    abis.sort()
    if ABI != 'none':
        abis.insert(0, ABI)
    abis.append('none')
    result = []

    arches = [ARCH]
    if sys.platform == 'darwin':
        m = re.match(r'(\w+)_(\d+)_(\d+)_(\w+)$', ARCH)
        if m:
            name, major, minor, arch = m.groups()
            minor = int(minor)
            matches = [arch]
            if arch in ('i386', 'ppc'):
                matches.append('fat')
            if arch in ('i386', 'ppc', 'x86_64'):
                matches.append('fat3')
            if arch in ('ppc64', 'x86_64'):
                matches.append('fat64')
            if arch in ('i386', 'x86_64'):
                matches.append('intel')
            if arch in ('i386', 'x86_64', 'intel', 'ppc', 'ppc64'):
                matches.append('universal')
            while minor >= 0:
                for match in matches:
                    s = '%s_%s_%s_%s' % (name, major, minor, match)
                    if s != ARCH:   # already there
                        arches.append(s)
                minor -= 1

    # Most specific - our Python version, ABI and arch
    for abi in abis:
        for arch in arches:
            result.append((''.join((IMP_PREFIX, versions[0])), abi, arch))

    # where no ABI / arch dependency, but IMP_PREFIX dependency
    for i, version in enumerate(versions):
        result.append((''.join((IMP_PREFIX, version)), 'none', 'any'))
        if i == 0:
            result.append((''.join((IMP_PREFIX, version[0])), 'none', 'any'))

    # no IMP_PREFIX, ABI or arch dependency
    for i, version in enumerate(versions):
        result.append((''.join(('py', version)), 'none', 'any'))
        if i == 0:
            result.append((''.join(('py', version[0])), 'none', 'any'))
    return set(result)


COMPATIBLE_TAGS = compatible_tags()

del compatible_tags


def is_compatible(wheel, tags=None):
    if not isinstance(wheel, Wheel):
        wheel = Wheel(wheel)    # assume it's a filename
    result = False
    if tags is None:
        tags = COMPATIBLE_TAGS
    for ver, abi, arch in tags:
        if ver in wheel.pyver and abi in wheel.abi and arch in wheel.arch:
            result = True
            break
    return result
site-packages/pip/_vendor/distlib/compat.py000064400000120674151464452710015022 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from __future__ import absolute_import

import os
import re
import sys

try:
    import ssl
except ImportError:  # pragma: no cover
    ssl = None

if sys.version_info[0] < 3:  # pragma: no cover
    from StringIO import StringIO
    string_types = basestring,
    text_type = unicode
    from types import FileType as file_type
    import __builtin__ as builtins
    import ConfigParser as configparser
    from ._backport import shutil
    from urlparse import urlparse, urlunparse, urljoin, urlsplit, urlunsplit
    from urllib import (urlretrieve, quote as _quote, unquote, url2pathname,
                        pathname2url, ContentTooShortError, splittype)

    def quote(s):
        if isinstance(s, unicode):
            s = s.encode('utf-8')
        return _quote(s)

    import urllib2
    from urllib2 import (Request, urlopen, URLError, HTTPError,
                         HTTPBasicAuthHandler, HTTPPasswordMgr,
                         HTTPHandler, HTTPRedirectHandler,
                         build_opener)
    if ssl:
        from urllib2 import HTTPSHandler
    import httplib
    import xmlrpclib
    import Queue as queue
    from HTMLParser import HTMLParser
    import htmlentitydefs
    raw_input = raw_input
    from itertools import ifilter as filter
    from itertools import ifilterfalse as filterfalse

    _userprog = None
    def splituser(host):
        """splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'."""
        global _userprog
        if _userprog is None:
            import re
            _userprog = re.compile('^(.*)@(.*)$')

        match = _userprog.match(host)
        if match: return match.group(1, 2)
        return None, host

else:  # pragma: no cover
    from io import StringIO
    string_types = str,
    text_type = str
    from io import TextIOWrapper as file_type
    import builtins
    import configparser
    import shutil
    from urllib.parse import (urlparse, urlunparse, urljoin, splituser, quote,
                              unquote, urlsplit, urlunsplit, splittype)
    from urllib.request import (urlopen, urlretrieve, Request, url2pathname,
                                pathname2url,
                                HTTPBasicAuthHandler, HTTPPasswordMgr,
                                HTTPHandler, HTTPRedirectHandler,
                                build_opener)
    if ssl:
        from urllib.request import HTTPSHandler
    from urllib.error import HTTPError, URLError, ContentTooShortError
    import http.client as httplib
    import urllib.request as urllib2
    import xmlrpc.client as xmlrpclib
    import queue
    from html.parser import HTMLParser
    import html.entities as htmlentitydefs
    raw_input = input
    from itertools import filterfalse
    filter = filter

try:
    from ssl import match_hostname, CertificateError
except ImportError: # pragma: no cover
    class CertificateError(ValueError):
        pass


    def _dnsname_match(dn, hostname, max_wildcards=1):
        """Matching according to RFC 6125, section 6.4.3

        http://tools.ietf.org/html/rfc6125#section-6.4.3
        """
        pats = []
        if not dn:
            return False

        parts = dn.split('.')
        leftmost, remainder = parts[0], parts[1:]

        wildcards = leftmost.count('*')
        if wildcards > max_wildcards:
            # Issue #17980: avoid denials of service by refusing more
            # than one wildcard per fragment.  A survey of established
            # policy among SSL implementations showed it to be a
            # reasonable choice.
            raise CertificateError(
                "too many wildcards in certificate DNS name: " + repr(dn))

        # speed up common case w/o wildcards
        if not wildcards:
            return dn.lower() == hostname.lower()

        # RFC 6125, section 6.4.3, subitem 1.
        # The client SHOULD NOT attempt to match a presented identifier in which
        # the wildcard character comprises a label other than the left-most label.
        if leftmost == '*':
            # When '*' is a fragment by itself, it matches a non-empty dotless
            # fragment.
            pats.append('[^.]+')
        elif leftmost.startswith('xn--') or hostname.startswith('xn--'):
            # RFC 6125, section 6.4.3, subitem 3.
            # The client SHOULD NOT attempt to match a presented identifier
            # where the wildcard character is embedded within an A-label or
            # U-label of an internationalized domain name.
            pats.append(re.escape(leftmost))
        else:
            # Otherwise, '*' matches any dotless string, e.g. www*
            pats.append(re.escape(leftmost).replace(r'\*', '[^.]*'))

        # add the remaining fragments, ignore any wildcards
        for frag in remainder:
            pats.append(re.escape(frag))

        pat = re.compile(r'\A' + r'\.'.join(pats) + r'\Z', re.IGNORECASE)
        return pat.match(hostname)


    def match_hostname(cert, hostname):
        """Verify that *cert* (in decoded format as returned by
        SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
        rules are followed, but IP addresses are not accepted for *hostname*.

        CertificateError is raised on failure. On success, the function
        returns nothing.
        """
        if not cert:
            raise ValueError("empty or no certificate, match_hostname needs a "
                             "SSL socket or SSL context with either "
                             "CERT_OPTIONAL or CERT_REQUIRED")
        dnsnames = []
        san = cert.get('subjectAltName', ())
        for key, value in san:
            if key == 'DNS':
                if _dnsname_match(value, hostname):
                    return
                dnsnames.append(value)
        if not dnsnames:
            # The subject is only checked when there is no dNSName entry
            # in subjectAltName
            for sub in cert.get('subject', ()):
                for key, value in sub:
                    # XXX according to RFC 2818, the most specific Common Name
                    # must be used.
                    if key == 'commonName':
                        if _dnsname_match(value, hostname):
                            return
                        dnsnames.append(value)
        if len(dnsnames) > 1:
            raise CertificateError("hostname %r "
                "doesn't match either of %s"
                % (hostname, ', '.join(map(repr, dnsnames))))
        elif len(dnsnames) == 1:
            raise CertificateError("hostname %r "
                "doesn't match %r"
                % (hostname, dnsnames[0]))
        else:
            raise CertificateError("no appropriate commonName or "
                "subjectAltName fields were found")


try:
    from types import SimpleNamespace as Container
except ImportError:  # pragma: no cover
    class Container(object):
        """
        A generic container for when multiple values need to be returned
        """
        def __init__(self, **kwargs):
            self.__dict__.update(kwargs)


try:
    from shutil import which
except ImportError:  # pragma: no cover
    # Implementation from Python 3.3
    def which(cmd, mode=os.F_OK | os.X_OK, path=None):
        """Given a command, mode, and a PATH string, return the path which
        conforms to the given mode on the PATH, or None if there is no such
        file.

        `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result
        of os.environ.get("PATH"), or can be overridden with a custom search
        path.

        """
        # Check that a given file can be accessed with the correct mode.
        # Additionally check that `file` is not a directory, as on Windows
        # directories pass the os.access check.
        def _access_check(fn, mode):
            return (os.path.exists(fn) and os.access(fn, mode)
                    and not os.path.isdir(fn))

        # If we're given a path with a directory part, look it up directly rather
        # than referring to PATH directories. This includes checking relative to the
        # current directory, e.g. ./script
        if os.path.dirname(cmd):
            if _access_check(cmd, mode):
                return cmd
            return None

        if path is None:
            path = os.environ.get("PATH", os.defpath)
        if not path:
            return None
        path = path.split(os.pathsep)

        if sys.platform == "win32":
            # The current directory takes precedence on Windows.
            if not os.curdir in path:
                path.insert(0, os.curdir)

            # PATHEXT is necessary to check on Windows.
            pathext = os.environ.get("PATHEXT", "").split(os.pathsep)
            # See if the given file matches any of the expected path extensions.
            # This will allow us to short circuit when given "python.exe".
            # If it does match, only test that one, otherwise we have to try
            # others.
            if any(cmd.lower().endswith(ext.lower()) for ext in pathext):
                files = [cmd]
            else:
                files = [cmd + ext for ext in pathext]
        else:
            # On other platforms you don't have things like PATHEXT to tell you
            # what file suffixes are executable, so just pass on cmd as-is.
            files = [cmd]

        seen = set()
        for dir in path:
            normdir = os.path.normcase(dir)
            if not normdir in seen:
                seen.add(normdir)
                for thefile in files:
                    name = os.path.join(dir, thefile)
                    if _access_check(name, mode):
                        return name
        return None


# ZipFile is a context manager in 2.7, but not in 2.6

from zipfile import ZipFile as BaseZipFile

if hasattr(BaseZipFile, '__enter__'):  # pragma: no cover
    ZipFile = BaseZipFile
else:  # pragma: no cover
    from zipfile import ZipExtFile as BaseZipExtFile

    class ZipExtFile(BaseZipExtFile):
        def __init__(self, base):
            self.__dict__.update(base.__dict__)

        def __enter__(self):
            return self

        def __exit__(self, *exc_info):
            self.close()
            # return None, so if an exception occurred, it will propagate

    class ZipFile(BaseZipFile):
        def __enter__(self):
            return self

        def __exit__(self, *exc_info):
            self.close()
            # return None, so if an exception occurred, it will propagate

        def open(self, *args, **kwargs):
            base = BaseZipFile.open(self, *args, **kwargs)
            return ZipExtFile(base)

try:
    from platform import python_implementation
except ImportError: # pragma: no cover
    def python_implementation():
        """Return a string identifying the Python implementation."""
        if 'PyPy' in sys.version:
            return 'PyPy'
        if os.name == 'java':
            return 'Jython'
        if sys.version.startswith('IronPython'):
            return 'IronPython'
        return 'CPython'

try:
    import sysconfig
except ImportError: # pragma: no cover
    from ._backport import sysconfig

try:
    callable = callable
except NameError:   # pragma: no cover
    from collections import Callable

    def callable(obj):
        return isinstance(obj, Callable)


try:
    fsencode = os.fsencode
    fsdecode = os.fsdecode
except AttributeError:  # pragma: no cover
    # Issue #99: on some systems (e.g. containerised),
    # sys.getfilesystemencoding() returns None, and we need a real value,
    # so fall back to utf-8. From the CPython 2.7 docs relating to Unix and
    # sys.getfilesystemencoding(): the return value is "the user’s preference
    # according to the result of nl_langinfo(CODESET), or None if the
    # nl_langinfo(CODESET) failed."
    _fsencoding = sys.getfilesystemencoding() or 'utf-8'
    if _fsencoding == 'mbcs':
        _fserrors = 'strict'
    else:
        _fserrors = 'surrogateescape'

    def fsencode(filename):
        if isinstance(filename, bytes):
            return filename
        elif isinstance(filename, text_type):
            return filename.encode(_fsencoding, _fserrors)
        else:
            raise TypeError("expect bytes or str, not %s" %
                            type(filename).__name__)

    def fsdecode(filename):
        if isinstance(filename, text_type):
            return filename
        elif isinstance(filename, bytes):
            return filename.decode(_fsencoding, _fserrors)
        else:
            raise TypeError("expect bytes or str, not %s" %
                            type(filename).__name__)

try:
    from tokenize import detect_encoding
except ImportError: # pragma: no cover
    from codecs import BOM_UTF8, lookup
    import re

    cookie_re = re.compile(r"coding[:=]\s*([-\w.]+)")

    def _get_normal_name(orig_enc):
        """Imitates get_normal_name in tokenizer.c."""
        # Only care about the first 12 characters.
        enc = orig_enc[:12].lower().replace("_", "-")
        if enc == "utf-8" or enc.startswith("utf-8-"):
            return "utf-8"
        if enc in ("latin-1", "iso-8859-1", "iso-latin-1") or \
           enc.startswith(("latin-1-", "iso-8859-1-", "iso-latin-1-")):
            return "iso-8859-1"
        return orig_enc

    def detect_encoding(readline):
        """
        The detect_encoding() function is used to detect the encoding that should
        be used to decode a Python source file.  It requires one argument, readline,
        in the same way as the tokenize() generator.

        It will call readline a maximum of twice, and return the encoding used
        (as a string) and a list of any lines (left as bytes) it has read in.

        It detects the encoding from the presence of a utf-8 bom or an encoding
        cookie as specified in pep-0263.  If both a bom and a cookie are present,
        but disagree, a SyntaxError will be raised.  If the encoding cookie is an
        invalid charset, raise a SyntaxError.  Note that if a utf-8 bom is found,
        'utf-8-sig' is returned.

        If no encoding is specified, then the default of 'utf-8' will be returned.
        """
        try:
            filename = readline.__self__.name
        except AttributeError:
            filename = None
        bom_found = False
        encoding = None
        default = 'utf-8'
        def read_or_stop():
            try:
                return readline()
            except StopIteration:
                return b''

        def find_cookie(line):
            try:
                # Decode as UTF-8. Either the line is an encoding declaration,
                # in which case it should be pure ASCII, or it must be UTF-8
                # per default encoding.
                line_string = line.decode('utf-8')
            except UnicodeDecodeError:
                msg = "invalid or missing encoding declaration"
                if filename is not None:
                    msg = '{} for {!r}'.format(msg, filename)
                raise SyntaxError(msg)

            matches = cookie_re.findall(line_string)
            if not matches:
                return None
            encoding = _get_normal_name(matches[0])
            try:
                codec = lookup(encoding)
            except LookupError:
                # This behaviour mimics the Python interpreter
                if filename is None:
                    msg = "unknown encoding: " + encoding
                else:
                    msg = "unknown encoding for {!r}: {}".format(filename,
                            encoding)
                raise SyntaxError(msg)

            if bom_found:
                if codec.name != 'utf-8':
                    # This behaviour mimics the Python interpreter
                    if filename is None:
                        msg = 'encoding problem: utf-8'
                    else:
                        msg = 'encoding problem for {!r}: utf-8'.format(filename)
                    raise SyntaxError(msg)
                encoding += '-sig'
            return encoding

        first = read_or_stop()
        if first.startswith(BOM_UTF8):
            bom_found = True
            first = first[3:]
            default = 'utf-8-sig'
        if not first:
            return default, []

        encoding = find_cookie(first)
        if encoding:
            return encoding, [first]

        second = read_or_stop()
        if not second:
            return default, [first]

        encoding = find_cookie(second)
        if encoding:
            return encoding, [first, second]

        return default, [first, second]

# For converting & <-> & etc.
try:
    from html import escape
except ImportError:
    from cgi import escape
if sys.version_info[:2] < (3, 4):
    unescape = HTMLParser().unescape
else:
    from html import unescape

try:
    from collections import ChainMap
except ImportError: # pragma: no cover
    from collections import MutableMapping

    try:
        from reprlib import recursive_repr as _recursive_repr
    except ImportError:
        def _recursive_repr(fillvalue='...'):
            '''
            Decorator to make a repr function return fillvalue for a recursive
            call
            '''

            def decorating_function(user_function):
                repr_running = set()

                def wrapper(self):
                    key = id(self), get_ident()
                    if key in repr_running:
                        return fillvalue
                    repr_running.add(key)
                    try:
                        result = user_function(self)
                    finally:
                        repr_running.discard(key)
                    return result

                # Can't use functools.wraps() here because of bootstrap issues
                wrapper.__module__ = getattr(user_function, '__module__')
                wrapper.__doc__ = getattr(user_function, '__doc__')
                wrapper.__name__ = getattr(user_function, '__name__')
                wrapper.__annotations__ = getattr(user_function, '__annotations__', {})
                return wrapper

            return decorating_function

    class ChainMap(MutableMapping):
        ''' A ChainMap groups multiple dicts (or other mappings) together
        to create a single, updateable view.

        The underlying mappings are stored in a list.  That list is public and can
        accessed or updated using the *maps* attribute.  There is no other state.

        Lookups search the underlying mappings successively until a key is found.
        In contrast, writes, updates, and deletions only operate on the first
        mapping.

        '''

        def __init__(self, *maps):
            '''Initialize a ChainMap by setting *maps* to the given mappings.
            If no mappings are provided, a single empty dictionary is used.

            '''
            self.maps = list(maps) or [{}]          # always at least one map

        def __missing__(self, key):
            raise KeyError(key)

        def __getitem__(self, key):
            for mapping in self.maps:
                try:
                    return mapping[key]             # can't use 'key in mapping' with defaultdict
                except KeyError:
                    pass
            return self.__missing__(key)            # support subclasses that define __missing__

        def get(self, key, default=None):
            return self[key] if key in self else default

        def __len__(self):
            return len(set().union(*self.maps))     # reuses stored hash values if possible

        def __iter__(self):
            return iter(set().union(*self.maps))

        def __contains__(self, key):
            return any(key in m for m in self.maps)

        def __bool__(self):
            return any(self.maps)

        @_recursive_repr()
        def __repr__(self):
            return '{0.__class__.__name__}({1})'.format(
                self, ', '.join(map(repr, self.maps)))

        @classmethod
        def fromkeys(cls, iterable, *args):
            'Create a ChainMap with a single dict created from the iterable.'
            return cls(dict.fromkeys(iterable, *args))

        def copy(self):
            'New ChainMap or subclass with a new copy of maps[0] and refs to maps[1:]'
            return self.__class__(self.maps[0].copy(), *self.maps[1:])

        __copy__ = copy

        def new_child(self):                        # like Django's Context.push()
            'New ChainMap with a new dict followed by all previous maps.'
            return self.__class__({}, *self.maps)

        @property
        def parents(self):                          # like Django's Context.pop()
            'New ChainMap from maps[1:].'
            return self.__class__(*self.maps[1:])

        def __setitem__(self, key, value):
            self.maps[0][key] = value

        def __delitem__(self, key):
            try:
                del self.maps[0][key]
            except KeyError:
                raise KeyError('Key not found in the first mapping: {!r}'.format(key))

        def popitem(self):
            'Remove and return an item pair from maps[0]. Raise KeyError is maps[0] is empty.'
            try:
                return self.maps[0].popitem()
            except KeyError:
                raise KeyError('No keys found in the first mapping.')

        def pop(self, key, *args):
            'Remove *key* from maps[0] and return its value. Raise KeyError if *key* not in maps[0].'
            try:
                return self.maps[0].pop(key, *args)
            except KeyError:
                raise KeyError('Key not found in the first mapping: {!r}'.format(key))

        def clear(self):
            'Clear maps[0], leaving maps[1:] intact.'
            self.maps[0].clear()

try:
    from importlib.util import cache_from_source  # Python >= 3.4
except ImportError:  # pragma: no cover
    try:
        from imp import cache_from_source
    except ImportError:  # pragma: no cover
        def cache_from_source(path, debug_override=None):
            assert path.endswith('.py')
            if debug_override is None:
                debug_override = __debug__
            if debug_override:
                suffix = 'c'
            else:
                suffix = 'o'
            return path + suffix

try:
    from collections import OrderedDict
except ImportError: # pragma: no cover
## {{{ http://code.activestate.com/recipes/576693/ (r9)
# Backport of OrderedDict() class that runs on Python 2.4, 2.5, 2.6, 2.7 and pypy.
# Passes Python2.7's test suite and incorporates all the latest updates.
    try:
        from thread import get_ident as _get_ident
    except ImportError:
        from dummy_thread import get_ident as _get_ident

    try:
        from _abcoll import KeysView, ValuesView, ItemsView
    except ImportError:
        pass


    class OrderedDict(dict):
        'Dictionary that remembers insertion order'
        # An inherited dict maps keys to values.
        # The inherited dict provides __getitem__, __len__, __contains__, and get.
        # The remaining methods are order-aware.
        # Big-O running times for all methods are the same as for regular dictionaries.

        # The internal self.__map dictionary maps keys to links in a doubly linked list.
        # The circular doubly linked list starts and ends with a sentinel element.
        # The sentinel element never gets deleted (this simplifies the algorithm).
        # Each link is stored as a list of length three:  [PREV, NEXT, KEY].

        def __init__(self, *args, **kwds):
            '''Initialize an ordered dictionary.  Signature is the same as for
            regular dictionaries, but keyword arguments are not recommended
            because their insertion order is arbitrary.

            '''
            if len(args) > 1:
                raise TypeError('expected at most 1 arguments, got %d' % len(args))
            try:
                self.__root
            except AttributeError:
                self.__root = root = []                     # sentinel node
                root[:] = [root, root, None]
                self.__map = {}
            self.__update(*args, **kwds)

        def __setitem__(self, key, value, dict_setitem=dict.__setitem__):
            'od.__setitem__(i, y) <==> od[i]=y'
            # Setting a new item creates a new link which goes at the end of the linked
            # list, and the inherited dictionary is updated with the new key/value pair.
            if key not in self:
                root = self.__root
                last = root[0]
                last[1] = root[0] = self.__map[key] = [last, root, key]
            dict_setitem(self, key, value)

        def __delitem__(self, key, dict_delitem=dict.__delitem__):
            'od.__delitem__(y) <==> del od[y]'
            # Deleting an existing item uses self.__map to find the link which is
            # then removed by updating the links in the predecessor and successor nodes.
            dict_delitem(self, key)
            link_prev, link_next, key = self.__map.pop(key)
            link_prev[1] = link_next
            link_next[0] = link_prev

        def __iter__(self):
            'od.__iter__() <==> iter(od)'
            root = self.__root
            curr = root[1]
            while curr is not root:
                yield curr[2]
                curr = curr[1]

        def __reversed__(self):
            'od.__reversed__() <==> reversed(od)'
            root = self.__root
            curr = root[0]
            while curr is not root:
                yield curr[2]
                curr = curr[0]

        def clear(self):
            'od.clear() -> None.  Remove all items from od.'
            try:
                for node in self.__map.itervalues():
                    del node[:]
                root = self.__root
                root[:] = [root, root, None]
                self.__map.clear()
            except AttributeError:
                pass
            dict.clear(self)

        def popitem(self, last=True):
            '''od.popitem() -> (k, v), return and remove a (key, value) pair.
            Pairs are returned in LIFO order if last is true or FIFO order if false.

            '''
            if not self:
                raise KeyError('dictionary is empty')
            root = self.__root
            if last:
                link = root[0]
                link_prev = link[0]
                link_prev[1] = root
                root[0] = link_prev
            else:
                link = root[1]
                link_next = link[1]
                root[1] = link_next
                link_next[0] = root
            key = link[2]
            del self.__map[key]
            value = dict.pop(self, key)
            return key, value

        # -- the following methods do not depend on the internal structure --

        def keys(self):
            'od.keys() -> list of keys in od'
            return list(self)

        def values(self):
            'od.values() -> list of values in od'
            return [self[key] for key in self]

        def items(self):
            'od.items() -> list of (key, value) pairs in od'
            return [(key, self[key]) for key in self]

        def iterkeys(self):
            'od.iterkeys() -> an iterator over the keys in od'
            return iter(self)

        def itervalues(self):
            'od.itervalues -> an iterator over the values in od'
            for k in self:
                yield self[k]

        def iteritems(self):
            'od.iteritems -> an iterator over the (key, value) items in od'
            for k in self:
                yield (k, self[k])

        def update(*args, **kwds):
            '''od.update(E, **F) -> None.  Update od from dict/iterable E and F.

            If E is a dict instance, does:           for k in E: od[k] = E[k]
            If E has a .keys() method, does:         for k in E.keys(): od[k] = E[k]
            Or if E is an iterable of items, does:   for k, v in E: od[k] = v
            In either case, this is followed by:     for k, v in F.items(): od[k] = v

            '''
            if len(args) > 2:
                raise TypeError('update() takes at most 2 positional '
                                'arguments (%d given)' % (len(args),))
            elif not args:
                raise TypeError('update() takes at least 1 argument (0 given)')
            self = args[0]
            # Make progressively weaker assumptions about "other"
            other = ()
            if len(args) == 2:
                other = args[1]
            if isinstance(other, dict):
                for key in other:
                    self[key] = other[key]
            elif hasattr(other, 'keys'):
                for key in other.keys():
                    self[key] = other[key]
            else:
                for key, value in other:
                    self[key] = value
            for key, value in kwds.items():
                self[key] = value

        __update = update  # let subclasses override update without breaking __init__

        __marker = object()

        def pop(self, key, default=__marker):
            '''od.pop(k[,d]) -> v, remove specified key and return the corresponding value.
            If key is not found, d is returned if given, otherwise KeyError is raised.

            '''
            if key in self:
                result = self[key]
                del self[key]
                return result
            if default is self.__marker:
                raise KeyError(key)
            return default

        def setdefault(self, key, default=None):
            'od.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in od'
            if key in self:
                return self[key]
            self[key] = default
            return default

        def __repr__(self, _repr_running=None):
            'od.__repr__() <==> repr(od)'
            if not _repr_running: _repr_running = {}
            call_key = id(self), _get_ident()
            if call_key in _repr_running:
                return '...'
            _repr_running[call_key] = 1
            try:
                if not self:
                    return '%s()' % (self.__class__.__name__,)
                return '%s(%r)' % (self.__class__.__name__, self.items())
            finally:
                del _repr_running[call_key]

        def __reduce__(self):
            'Return state information for pickling'
            items = [[k, self[k]] for k in self]
            inst_dict = vars(self).copy()
            for k in vars(OrderedDict()):
                inst_dict.pop(k, None)
            if inst_dict:
                return (self.__class__, (items,), inst_dict)
            return self.__class__, (items,)

        def copy(self):
            'od.copy() -> a shallow copy of od'
            return self.__class__(self)

        @classmethod
        def fromkeys(cls, iterable, value=None):
            '''OD.fromkeys(S[, v]) -> New ordered dictionary with keys from S
            and values equal to v (which defaults to None).

            '''
            d = cls()
            for key in iterable:
                d[key] = value
            return d

        def __eq__(self, other):
            '''od.__eq__(y) <==> od==y.  Comparison to another OD is order-sensitive
            while comparison to a regular mapping is order-insensitive.

            '''
            if isinstance(other, OrderedDict):
                return len(self)==len(other) and self.items() == other.items()
            return dict.__eq__(self, other)

        def __ne__(self, other):
            return not self == other

        # -- the following methods are only used in Python 2.7 --

        def viewkeys(self):
            "od.viewkeys() -> a set-like object providing a view on od's keys"
            return KeysView(self)

        def viewvalues(self):
            "od.viewvalues() -> an object providing a view on od's values"
            return ValuesView(self)

        def viewitems(self):
            "od.viewitems() -> a set-like object providing a view on od's items"
            return ItemsView(self)

try:
    from logging.config import BaseConfigurator, valid_ident
except ImportError: # pragma: no cover
    IDENTIFIER = re.compile('^[a-z_][a-z0-9_]*$', re.I)


    def valid_ident(s):
        m = IDENTIFIER.match(s)
        if not m:
            raise ValueError('Not a valid Python identifier: %r' % s)
        return True


    # The ConvertingXXX classes are wrappers around standard Python containers,
    # and they serve to convert any suitable values in the container. The
    # conversion converts base dicts, lists and tuples to their wrapped
    # equivalents, whereas strings which match a conversion format are converted
    # appropriately.
    #
    # Each wrapper should have a configurator attribute holding the actual
    # configurator to use for conversion.

    class ConvertingDict(dict):
        """A converting dictionary wrapper."""

        def __getitem__(self, key):
            value = dict.__getitem__(self, key)
            result = self.configurator.convert(value)
            #If the converted value is different, save for next time
            if value is not result:
                self[key] = result
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

        def get(self, key, default=None):
            value = dict.get(self, key, default)
            result = self.configurator.convert(value)
            #If the converted value is different, save for next time
            if value is not result:
                self[key] = result
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

    def pop(self, key, default=None):
        value = dict.pop(self, key, default)
        result = self.configurator.convert(value)
        if value is not result:
            if type(result) in (ConvertingDict, ConvertingList,
                                ConvertingTuple):
                result.parent = self
                result.key = key
        return result

    class ConvertingList(list):
        """A converting list wrapper."""
        def __getitem__(self, key):
            value = list.__getitem__(self, key)
            result = self.configurator.convert(value)
            #If the converted value is different, save for next time
            if value is not result:
                self[key] = result
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

        def pop(self, idx=-1):
            value = list.pop(self, idx)
            result = self.configurator.convert(value)
            if value is not result:
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
            return result

    class ConvertingTuple(tuple):
        """A converting tuple wrapper."""
        def __getitem__(self, key):
            value = tuple.__getitem__(self, key)
            result = self.configurator.convert(value)
            if value is not result:
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

    class BaseConfigurator(object):
        """
        The configurator base class which defines some useful defaults.
        """

        CONVERT_PATTERN = re.compile(r'^(?P[a-z]+)://(?P.*)$')

        WORD_PATTERN = re.compile(r'^\s*(\w+)\s*')
        DOT_PATTERN = re.compile(r'^\.\s*(\w+)\s*')
        INDEX_PATTERN = re.compile(r'^\[\s*(\w+)\s*\]\s*')
        DIGIT_PATTERN = re.compile(r'^\d+$')

        value_converters = {
            'ext' : 'ext_convert',
            'cfg' : 'cfg_convert',
        }

        # We might want to use a different one, e.g. importlib
        importer = staticmethod(__import__)

        def __init__(self, config):
            self.config = ConvertingDict(config)
            self.config.configurator = self

        def resolve(self, s):
            """
            Resolve strings to objects using standard import and attribute
            syntax.
            """
            name = s.split('.')
            used = name.pop(0)
            try:
                found = self.importer(used)
                for frag in name:
                    used += '.' + frag
                    try:
                        found = getattr(found, frag)
                    except AttributeError:
                        self.importer(used)
                        found = getattr(found, frag)
                return found
            except ImportError:
                e, tb = sys.exc_info()[1:]
                v = ValueError('Cannot resolve %r: %s' % (s, e))
                v.__cause__, v.__traceback__ = e, tb
                raise v

        def ext_convert(self, value):
            """Default converter for the ext:// protocol."""
            return self.resolve(value)

        def cfg_convert(self, value):
            """Default converter for the cfg:// protocol."""
            rest = value
            m = self.WORD_PATTERN.match(rest)
            if m is None:
                raise ValueError("Unable to convert %r" % value)
            else:
                rest = rest[m.end():]
                d = self.config[m.groups()[0]]
                #print d, rest
                while rest:
                    m = self.DOT_PATTERN.match(rest)
                    if m:
                        d = d[m.groups()[0]]
                    else:
                        m = self.INDEX_PATTERN.match(rest)
                        if m:
                            idx = m.groups()[0]
                            if not self.DIGIT_PATTERN.match(idx):
                                d = d[idx]
                            else:
                                try:
                                    n = int(idx) # try as number first (most likely)
                                    d = d[n]
                                except TypeError:
                                    d = d[idx]
                    if m:
                        rest = rest[m.end():]
                    else:
                        raise ValueError('Unable to convert '
                                         '%r at %r' % (value, rest))
            #rest should be empty
            return d

        def convert(self, value):
            """
            Convert values to an appropriate type. dicts, lists and tuples are
            replaced by their converting alternatives. Strings are checked to
            see if they have a conversion format and are converted if they do.
            """
            if not isinstance(value, ConvertingDict) and isinstance(value, dict):
                value = ConvertingDict(value)
                value.configurator = self
            elif not isinstance(value, ConvertingList) and isinstance(value, list):
                value = ConvertingList(value)
                value.configurator = self
            elif not isinstance(value, ConvertingTuple) and\
                     isinstance(value, tuple):
                value = ConvertingTuple(value)
                value.configurator = self
            elif isinstance(value, string_types):
                m = self.CONVERT_PATTERN.match(value)
                if m:
                    d = m.groupdict()
                    prefix = d['prefix']
                    converter = self.value_converters.get(prefix, None)
                    if converter:
                        suffix = d['suffix']
                        converter = getattr(self, converter)
                        value = converter(suffix)
            return value

        def configure_custom(self, config):
            """Configure an object with a user-supplied factory."""
            c = config.pop('()')
            if not callable(c):
                c = self.resolve(c)
            props = config.pop('.', None)
            # Check for valid identifiers
            kwargs = dict([(k, config[k]) for k in config if valid_ident(k)])
            result = c(**kwargs)
            if props:
                for name, value in props.items():
                    setattr(result, name, value)
            return result

        def as_tuple(self, value):
            """Utility function which converts lists to tuples."""
            if isinstance(value, list):
                value = tuple(value)
            return value
site-packages/pip/_vendor/distlib/markers.py000064400000010443151464452710015173 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""
Parser for the environment markers micro-language defined in PEP 508.
"""

# Note: In PEP 345, the micro-language was Python compatible, so the ast
# module could be used to parse it. However, PEP 508 introduced operators such
# as ~= and === which aren't in Python, necessitating a different approach.

import os
import sys
import platform
import re

from .compat import python_implementation, urlparse, string_types
from .util import in_venv, parse_marker

__all__ = ['interpret']

def _is_literal(o):
    if not isinstance(o, string_types) or not o:
        return False
    return o[0] in '\'"'

class Evaluator(object):
    """
    This class is used to evaluate marker expessions.
    """

    operations = {
        '==': lambda x, y: x == y,
        '===': lambda x, y: x == y,
        '~=': lambda x, y: x == y or x > y,
        '!=': lambda x, y: x != y,
        '<':  lambda x, y: x < y,
        '<=':  lambda x, y: x == y or x < y,
        '>':  lambda x, y: x > y,
        '>=':  lambda x, y: x == y or x > y,
        'and': lambda x, y: x and y,
        'or': lambda x, y: x or y,
        'in': lambda x, y: x in y,
        'not in': lambda x, y: x not in y,
    }

    def evaluate(self, expr, context):
        """
        Evaluate a marker expression returned by the :func:`parse_requirement`
        function in the specified context.
        """
        if isinstance(expr, string_types):
            if expr[0] in '\'"':
                result = expr[1:-1]
            else:
                if expr not in context:
                    raise SyntaxError('unknown variable: %s' % expr)
                result = context[expr]
        else:
            assert isinstance(expr, dict)
            op = expr['op']
            if op not in self.operations:
                raise NotImplementedError('op not implemented: %s' % op)
            elhs = expr['lhs']
            erhs = expr['rhs']
            if _is_literal(expr['lhs']) and _is_literal(expr['rhs']):
                raise SyntaxError('invalid comparison: %s %s %s' % (elhs, op, erhs))

            lhs = self.evaluate(elhs, context)
            rhs = self.evaluate(erhs, context)
            result = self.operations[op](lhs, rhs)
        return result

def default_context():
    def format_full_version(info):
        version = '%s.%s.%s' % (info.major, info.minor, info.micro)
        kind = info.releaselevel
        if kind != 'final':
            version += kind[0] + str(info.serial)
        return version

    if hasattr(sys, 'implementation'):
        implementation_version = format_full_version(sys.implementation.version)
        implementation_name = sys.implementation.name
    else:
        implementation_version = '0'
        implementation_name = ''

    result = {
        'implementation_name': implementation_name,
        'implementation_version': implementation_version,
        'os_name': os.name,
        'platform_machine': platform.machine(),
        'platform_python_implementation': platform.python_implementation(),
        'platform_release': platform.release(),
        'platform_system': platform.system(),
        'platform_version': platform.version(),
        'platform_in_venv': str(in_venv()),
        'python_full_version': platform.python_version(),
        'python_version': platform.python_version()[:3],
        'sys_platform': sys.platform,
    }
    return result

DEFAULT_CONTEXT = default_context()
del default_context

evaluator = Evaluator()

def interpret(marker, execution_context=None):
    """
    Interpret a marker and return a result depending on environment.

    :param marker: The marker to interpret.
    :type marker: str
    :param execution_context: The context used for name lookup.
    :type execution_context: mapping
    """
    try:
        expr, rest = parse_marker(marker)
    except Exception as e:
        raise SyntaxError('Unable to interpret marker syntax: %s: %s' % (marker, e))
    if rest and rest[0] != '#':
        raise SyntaxError('unexpected trailing data in marker: %s: %s' % (marker, rest))
    context = dict(DEFAULT_CONTEXT)
    if execution_context:
        context.update(execution_context)
    return evaluator.evaluate(expr, context)
site-packages/pip/_vendor/distlib/__init__.py000064400000001113151464452710015260 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2019 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
import logging

__version__ = '0.2.9.post0'

class DistlibException(Exception):
    pass

try:
    from logging import NullHandler
except ImportError: # pragma: no cover
    class NullHandler(logging.Handler):
        def handle(self, record): pass
        def emit(self, record): pass
        def createLock(self): self.lock = None

logger = logging.getLogger(__name__)
logger.addHandler(NullHandler())
site-packages/pip/_vendor/distlib/index.py000064400000051112151464452710014634 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
import hashlib
import logging
import os
import shutil
import subprocess
import tempfile
try:
    from threading import Thread
except ImportError:
    from dummy_threading import Thread

from . import DistlibException
from .compat import (HTTPBasicAuthHandler, Request, HTTPPasswordMgr,
                     urlparse, build_opener, string_types)
from .util import cached_property, zip_dir, ServerProxy

logger = logging.getLogger(__name__)

DEFAULT_INDEX = 'https://pypi.org/pypi'
DEFAULT_REALM = 'pypi'

class PackageIndex(object):
    """
    This class represents a package index compatible with PyPI, the Python
    Package Index.
    """

    boundary = b'----------ThIs_Is_tHe_distlib_index_bouNdaRY_$'

    def __init__(self, url=None):
        """
        Initialise an instance.

        :param url: The URL of the index. If not specified, the URL for PyPI is
                    used.
        """
        self.url = url or DEFAULT_INDEX
        self.read_configuration()
        scheme, netloc, path, params, query, frag = urlparse(self.url)
        if params or query or frag or scheme not in ('http', 'https'):
            raise DistlibException('invalid repository: %s' % self.url)
        self.password_handler = None
        self.ssl_verifier = None
        self.gpg = None
        self.gpg_home = None
        with open(os.devnull, 'w') as sink:
            # Use gpg by default rather than gpg2, as gpg2 insists on
            # prompting for passwords
            for s in ('gpg', 'gpg2'):
                try:
                    rc = subprocess.check_call([s, '--version'], stdout=sink,
                                               stderr=sink)
                    if rc == 0:
                        self.gpg = s
                        break
                except OSError:
                    pass

    def _get_pypirc_command(self):
        """
        Get the distutils command for interacting with PyPI configurations.
        :return: the command.
        """
        from distutils.core import Distribution
        from distutils.config import PyPIRCCommand
        d = Distribution()
        return PyPIRCCommand(d)

    def read_configuration(self):
        """
        Read the PyPI access configuration as supported by distutils, getting
        PyPI to do the actual work. This populates ``username``, ``password``,
        ``realm`` and ``url`` attributes from the configuration.
        """
        # get distutils to do the work
        c = self._get_pypirc_command()
        c.repository = self.url
        cfg = c._read_pypirc()
        self.username = cfg.get('username')
        self.password = cfg.get('password')
        self.realm = cfg.get('realm', 'pypi')
        self.url = cfg.get('repository', self.url)

    def save_configuration(self):
        """
        Save the PyPI access configuration. You must have set ``username`` and
        ``password`` attributes before calling this method.

        Again, distutils is used to do the actual work.
        """
        self.check_credentials()
        # get distutils to do the work
        c = self._get_pypirc_command()
        c._store_pypirc(self.username, self.password)

    def check_credentials(self):
        """
        Check that ``username`` and ``password`` have been set, and raise an
        exception if not.
        """
        if self.username is None or self.password is None:
            raise DistlibException('username and password must be set')
        pm = HTTPPasswordMgr()
        _, netloc, _, _, _, _ = urlparse(self.url)
        pm.add_password(self.realm, netloc, self.username, self.password)
        self.password_handler = HTTPBasicAuthHandler(pm)

    def register(self, metadata):
        """
        Register a distribution on PyPI, using the provided metadata.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the distribution to be
                         registered.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        """
        self.check_credentials()
        metadata.validate()
        d = metadata.todict()
        d[':action'] = 'verify'
        request = self.encode_request(d.items(), [])
        response = self.send_request(request)
        d[':action'] = 'submit'
        request = self.encode_request(d.items(), [])
        return self.send_request(request)

    def _reader(self, name, stream, outbuf):
        """
        Thread runner for reading lines of from a subprocess into a buffer.

        :param name: The logical name of the stream (used for logging only).
        :param stream: The stream to read from. This will typically a pipe
                       connected to the output stream of a subprocess.
        :param outbuf: The list to append the read lines to.
        """
        while True:
            s = stream.readline()
            if not s:
                break
            s = s.decode('utf-8').rstrip()
            outbuf.append(s)
            logger.debug('%s: %s' % (name, s))
        stream.close()

    def get_sign_command(self, filename, signer, sign_password,
                         keystore=None):
        """
        Return a suitable command for signing a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The signing command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        """
        cmd = [self.gpg, '--status-fd', '2', '--no-tty']
        if keystore is None:
            keystore = self.gpg_home
        if keystore:
            cmd.extend(['--homedir', keystore])
        if sign_password is not None:
            cmd.extend(['--batch', '--passphrase-fd', '0'])
        td = tempfile.mkdtemp()
        sf = os.path.join(td, os.path.basename(filename) + '.asc')
        cmd.extend(['--detach-sign', '--armor', '--local-user',
                    signer, '--output', sf, filename])
        logger.debug('invoking: %s', ' '.join(cmd))
        return cmd, sf

    def run_command(self, cmd, input_data=None):
        """
        Run a command in a child process , passing it any input data specified.

        :param cmd: The command to run.
        :param input_data: If specified, this must be a byte string containing
                           data to be sent to the child process.
        :return: A tuple consisting of the subprocess' exit code, a list of
                 lines read from the subprocess' ``stdout``, and a list of
                 lines read from the subprocess' ``stderr``.
        """
        kwargs = {
            'stdout': subprocess.PIPE,
            'stderr': subprocess.PIPE,
        }
        if input_data is not None:
            kwargs['stdin'] = subprocess.PIPE
        stdout = []
        stderr = []
        p = subprocess.Popen(cmd, **kwargs)
        # We don't use communicate() here because we may need to
        # get clever with interacting with the command
        t1 = Thread(target=self._reader, args=('stdout', p.stdout, stdout))
        t1.start()
        t2 = Thread(target=self._reader, args=('stderr', p.stderr, stderr))
        t2.start()
        if input_data is not None:
            p.stdin.write(input_data)
            p.stdin.close()

        p.wait()
        t1.join()
        t2.join()
        return p.returncode, stdout, stderr

    def sign_file(self, filename, signer, sign_password, keystore=None):
        """
        Sign a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The absolute pathname of the file where the signature is
                 stored.
        """
        cmd, sig_file = self.get_sign_command(filename, signer, sign_password,
                                              keystore)
        rc, stdout, stderr = self.run_command(cmd,
                                              sign_password.encode('utf-8'))
        if rc != 0:
            raise DistlibException('sign command failed with error '
                                   'code %s' % rc)
        return sig_file

    def upload_file(self, metadata, filename, signer=None, sign_password=None,
                    filetype='sdist', pyversion='source', keystore=None):
        """
        Upload a release file to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the file to be uploaded.
        :param filename: The pathname of the file to be uploaded.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param filetype: The type of the file being uploaded. This is the
                        distutils command which produced that file, e.g.
                        ``sdist`` or ``bdist_wheel``.
        :param pyversion: The version of Python which the release relates
                          to. For code compatible with any Python, this would
                          be ``source``, otherwise it would be e.g. ``3.2``.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        """
        self.check_credentials()
        if not os.path.exists(filename):
            raise DistlibException('not found: %s' % filename)
        metadata.validate()
        d = metadata.todict()
        sig_file = None
        if signer:
            if not self.gpg:
                logger.warning('no signing program available - not signed')
            else:
                sig_file = self.sign_file(filename, signer, sign_password,
                                          keystore)
        with open(filename, 'rb') as f:
            file_data = f.read()
        md5_digest = hashlib.md5(file_data).hexdigest()
        sha256_digest = hashlib.sha256(file_data).hexdigest()
        d.update({
            ':action': 'file_upload',
            'protocol_version': '1',
            'filetype': filetype,
            'pyversion': pyversion,
            'md5_digest': md5_digest,
            'sha256_digest': sha256_digest,
        })
        files = [('content', os.path.basename(filename), file_data)]
        if sig_file:
            with open(sig_file, 'rb') as f:
                sig_data = f.read()
            files.append(('gpg_signature', os.path.basename(sig_file),
                         sig_data))
            shutil.rmtree(os.path.dirname(sig_file))
        request = self.encode_request(d.items(), files)
        return self.send_request(request)

    def upload_documentation(self, metadata, doc_dir):
        """
        Upload documentation to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the documentation to be
                         uploaded.
        :param doc_dir: The pathname of the directory which contains the
                        documentation. This should be the directory that
                        contains the ``index.html`` for the documentation.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        """
        self.check_credentials()
        if not os.path.isdir(doc_dir):
            raise DistlibException('not a directory: %r' % doc_dir)
        fn = os.path.join(doc_dir, 'index.html')
        if not os.path.exists(fn):
            raise DistlibException('not found: %r' % fn)
        metadata.validate()
        name, version = metadata.name, metadata.version
        zip_data = zip_dir(doc_dir).getvalue()
        fields = [(':action', 'doc_upload'),
                  ('name', name), ('version', version)]
        files = [('content', name, zip_data)]
        request = self.encode_request(fields, files)
        return self.send_request(request)

    def get_verify_command(self, signature_filename, data_filename,
                           keystore=None):
        """
        Return a suitable command for verifying a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The verifying command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        """
        cmd = [self.gpg, '--status-fd', '2', '--no-tty']
        if keystore is None:
            keystore = self.gpg_home
        if keystore:
            cmd.extend(['--homedir', keystore])
        cmd.extend(['--verify', signature_filename, data_filename])
        logger.debug('invoking: %s', ' '.join(cmd))
        return cmd

    def verify_signature(self, signature_filename, data_filename,
                         keystore=None):
        """
        Verify a signature for a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: True if the signature was verified, else False.
        """
        if not self.gpg:
            raise DistlibException('verification unavailable because gpg '
                                   'unavailable')
        cmd = self.get_verify_command(signature_filename, data_filename,
                                      keystore)
        rc, stdout, stderr = self.run_command(cmd)
        if rc not in (0, 1):
            raise DistlibException('verify command failed with error '
                             'code %s' % rc)
        return rc == 0

    def download_file(self, url, destfile, digest=None, reporthook=None):
        """
        This is a convenience method for downloading a file from an URL.
        Normally, this will be a file from the index, though currently
        no check is made for this (i.e. a file can be downloaded from
        anywhere).

        The method is just like the :func:`urlretrieve` function in the
        standard library, except that it allows digest computation to be
        done during download and checking that the downloaded data
        matched any expected value.

        :param url: The URL of the file to be downloaded (assumed to be
                    available via an HTTP GET request).
        :param destfile: The pathname where the downloaded file is to be
                         saved.
        :param digest: If specified, this must be a (hasher, value)
                       tuple, where hasher is the algorithm used (e.g.
                       ``'md5'``) and ``value`` is the expected value.
        :param reporthook: The same as for :func:`urlretrieve` in the
                           standard library.
        """
        if digest is None:
            digester = None
            logger.debug('No digest specified')
        else:
            if isinstance(digest, (list, tuple)):
                hasher, digest = digest
            else:
                hasher = 'md5'
            digester = getattr(hashlib, hasher)()
            logger.debug('Digest specified: %s' % digest)
        # The following code is equivalent to urlretrieve.
        # We need to do it this way so that we can compute the
        # digest of the file as we go.
        with open(destfile, 'wb') as dfp:
            # addinfourl is not a context manager on 2.x
            # so we have to use try/finally
            sfp = self.send_request(Request(url))
            try:
                headers = sfp.info()
                blocksize = 8192
                size = -1
                read = 0
                blocknum = 0
                if "content-length" in headers:
                    size = int(headers["Content-Length"])
                if reporthook:
                    reporthook(blocknum, blocksize, size)
                while True:
                    block = sfp.read(blocksize)
                    if not block:
                        break
                    read += len(block)
                    dfp.write(block)
                    if digester:
                        digester.update(block)
                    blocknum += 1
                    if reporthook:
                        reporthook(blocknum, blocksize, size)
            finally:
                sfp.close()

        # check that we got the whole file, if we can
        if size >= 0 and read < size:
            raise DistlibException(
                'retrieval incomplete: got only %d out of %d bytes'
                % (read, size))
        # if we have a digest, it must match.
        if digester:
            actual = digester.hexdigest()
            if digest != actual:
                raise DistlibException('%s digest mismatch for %s: expected '
                                       '%s, got %s' % (hasher, destfile,
                                                       digest, actual))
            logger.debug('Digest verified: %s', digest)

    def send_request(self, req):
        """
        Send a standard library :class:`Request` to PyPI and return its
        response.

        :param req: The request to send.
        :return: The HTTP response from PyPI (a standard library HTTPResponse).
        """
        handlers = []
        if self.password_handler:
            handlers.append(self.password_handler)
        if self.ssl_verifier:
            handlers.append(self.ssl_verifier)
        opener = build_opener(*handlers)
        return opener.open(req)

    def encode_request(self, fields, files):
        """
        Encode fields and files for posting to an HTTP server.

        :param fields: The fields to send as a list of (fieldname, value)
                       tuples.
        :param files: The files to send as a list of (fieldname, filename,
                      file_bytes) tuple.
        """
        # Adapted from packaging, which in turn was adapted from
        # http://code.activestate.com/recipes/146306

        parts = []
        boundary = self.boundary
        for k, values in fields:
            if not isinstance(values, (list, tuple)):
                values = [values]

            for v in values:
                parts.extend((
                    b'--' + boundary,
                    ('Content-Disposition: form-data; name="%s"' %
                     k).encode('utf-8'),
                    b'',
                    v.encode('utf-8')))
        for key, filename, value in files:
            parts.extend((
                b'--' + boundary,
                ('Content-Disposition: form-data; name="%s"; filename="%s"' %
                 (key, filename)).encode('utf-8'),
                b'',
                value))

        parts.extend((b'--' + boundary + b'--', b''))

        body = b'\r\n'.join(parts)
        ct = b'multipart/form-data; boundary=' + boundary
        headers = {
            'Content-type': ct,
            'Content-length': str(len(body))
        }
        return Request(self.url, body, headers)

    def search(self, terms, operator=None):
        if isinstance(terms, string_types):
            terms = {'name': terms}
        rpc_proxy = ServerProxy(self.url, timeout=3.0)
        try:
            return rpc_proxy.search(terms, operator or 'and')
        finally:
            rpc_proxy('close')()
site-packages/pip/_vendor/__pycache__/appdirs.cpython-38.opt-1.pyc000064400000047725151464452710021001 0ustar00U

.e_@s`dZdZdeeeZddlZddlZejddkZ	e	r>eZ
ejdrddlZe
ddZedrrdZqed	rd
ZqdZnejZd4d
dZd5ddZd6ddZd7ddZd8ddZd9ddZd:ddZGdddeZddZd d!Zd"d#Zd$d%Zedkrfzdd&lmZeZWnBe k
rdzddl!Z"eZWne k
r^eZYnXYnXe#d'kr\d(Z$d)Z%d*Z&e'd+ee'd,ee$e%d-d.Z(e&D]Z)e'd/e)e*e(e)fqe'd0ee$e%Z(e&D]Z)e'd/e)e*e(e)fqe'd1ee$Z(e&D]Z)e'd/e)e*e(e)fqe'd2ee$dd3Z(e&D]Z)e'd/e)e*e(e)fq>dS);zyUtilities for determining application-specific dirs.

See  for details and usage.
).NrjavaZWindowswin32ZMacdarwinZlinux2FcCstdkr^|dkr|}|rdpd}tjt|}|r|dk	rNtj|||}qtj||}nNtdkrtjd}|rtj||}n&tdtjd	}|rtj||}|r|rtj||}|S)
aJReturn full path to the user-specific data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            
            for a discussion of issues.

    Typical user data directories are:
        Mac OS X:               ~/Library/Application Support/
        Unix:                   ~/.local/share/    # or in $XDG_DATA_HOME, if defined
        Win XP (not roaming):   C:\Documents and Settings\\Application Data\\
        Win XP (roaming):       C:\Documents and Settings\\Local Settings\Application Data\\
        Win 7  (not roaming):   C:\Users\\AppData\Local\\
        Win 7  (roaming):       C:\Users\\AppData\Roaming\\

    For Unix, we follow the XDG spec and support $XDG_DATA_HOME.
    That means, by default "~/.local/share/".
    rN
CSIDL_APPDATACSIDL_LOCAL_APPDATAFrz~/Library/Application Support/Z
XDG_DATA_HOMEz~/.local/sharesystemospathnormpath_get_win_folderjoin
expandusergetenv)appname	appauthorversionroamingconstrr7/usr/lib/python3.8/site-packages/pip/_vendor/appdirs.py
user_data_dir-s& rcstdkrR|dkr}tjtd}r|dk	rBtj||}qtj|}ntdkrztjd}rtj|}nttdtjdd	g}d
d|	tjD}r|rtj|fdd|D}|rtj|}n|d
}|Sr|rtj||}|S)aiReturn full path to the user-shared data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "multipath" is an optional parameter only applicable to *nix
            which indicates that the entire list of data dirs should be
            returned. By default, the first item from XDG_DATA_DIRS is
            returned, or '/usr/local/share/',
            if XDG_DATA_DIRS is not set

    Typical site data directories are:
        Mac OS X:   /Library/Application Support/
        Unix:       /usr/local/share/ or /usr/share/
        Win XP:     C:\Documents and Settings\All Users\Application Data\\
        Vista:      (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.)
        Win 7:      C:\ProgramData\\   # Hidden, but writeable on Win 7.

    For Unix, this is using the $XDG_DATA_DIRS[0] default.

    WARNING: Do not use this on Windows. See the Vista-Fail note above for why.
    rNCSIDL_COMMON_APPDATAFrz/Library/Application SupportZ
XDG_DATA_DIRSz/usr/local/sharez
/usr/sharecSs g|]}tj|tjqSrr
rrrstripsep.0xrrr
sz!site_data_dir..csg|]}tj|gqSrr
rrr rrrr#sr)
rr
rrrrrrpathsepsplitrrr	multipathrZpathlistrr%r
site_data_dirds6r*cCsXtdkrt||d|}n&tdtjd}|r>tj||}|rT|rTtj||}|S)aReturn full path to the user-specific config dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            
            for a discussion of issues.

    Typical user config directories are:
        Mac OS X:               same as user_data_dir
        Unix:                   ~/.config/     # or in $XDG_CONFIG_HOME, if defined
        Win *:                  same as user_data_dir

    For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME.
    That means, by default "~/.config/".
    rrNZXDG_CONFIG_HOMEz	~/.configrrr
rrrrrrrrrrrruser_config_dirsr.cstdkr*t|}r|rtj||}ndtdd}dd|tjD}rt|rbtj|fdd|D}|rtj|}n|d}|S)aReturn full path to the user-shared data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "multipath" is an optional parameter only applicable to *nix
            which indicates that the entire list of config dirs should be
            returned. By default, the first item from XDG_CONFIG_DIRS is
            returned, or '/etc/xdg/', if XDG_CONFIG_DIRS is not set

    Typical site config directories are:
        Mac OS X:   same as site_data_dir
        Unix:       /etc/xdg/ or $XDG_CONFIG_DIRS[i]/ for each value in
                    $XDG_CONFIG_DIRS
        Win *:      same as site_data_dir
        Vista:      (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.)

    For Unix, this is using the $XDG_CONFIG_DIRS[0] default, if multipath=False

    WARNING: Do not use this on Windows. See the Vista-Fail note above for why.
    r+ZXDG_CONFIG_DIRSz/etc/xdgcSs g|]}tj|tjqSrrr rrrr#sz#site_config_dir..csg|]}tj|gqSrr$r r%rrr#sr)rr*r
rrrr'r&r(rr%rsite_config_dirs
r/TcCstdkrd|dkr|}tjtd}|r|dk	rBtj|||}ntj||}|rtj|d}nNtdkrtjd}|rtj||}n&tdtjd	}|rtj||}|r|rtj||}|S)
aReturn full path to the user-specific cache dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "opinion" (boolean) can be False to disable the appending of
            "Cache" to the base app data dir for Windows. See
            discussion below.

    Typical user cache directories are:
        Mac OS X:   ~/Library/Caches/
        Unix:       ~/.cache/ (XDG default)
        Win XP:     C:\Documents and Settings\\Local Settings\Application Data\\\Cache
        Vista:      C:\Users\\AppData\Local\\\Cache

    On Windows the only suggestion in the MSDN docs is that local settings go in
    the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming
    app data dir (the default returned by `user_data_dir` above). Apps typically
    put cache data somewhere *under* the given dir here. Some examples:
        ...\Mozilla\Firefox\Profiles\\Cache
        ...\Acme\SuperApp\Cache\1.0
    OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value.
    This can be disabled with the `opinion=False` option.
    rNr
FZCacherz~/Library/CachesZXDG_CACHE_HOMEz~/.cacherrrrZopinionrrrruser_cache_dirs(!r1cCsXtdkrt||d|}n&tdtjd}|r>tj||}|rT|rTtj||}|S)aReturn full path to the user-specific state dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            
            for a discussion of issues.

    Typical user state directories are:
        Mac OS X:  same as user_data_dir
        Unix:      ~/.local/state/   # or in $XDG_STATE_HOME, if defined
        Win *:     same as user_data_dir

    For Unix, we follow this Debian proposal 
    to extend the XDG spec and support $XDG_STATE_HOME.

    That means, by default "~/.local/state/".
    r+NZXDG_STATE_HOMEz~/.local/stater,r-rrruser_state_dir:sr2cCstdkr tjtjd|}nNtdkrLt|||}d}|rntj|d}n"t|||}d}|rntj|d}|r|rtj||}|S)aReturn full path to the user-specific log dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be ".".
            Only applied when appname is present.
        "opinion" (boolean) can be False to disable the appending of
            "Logs" to the base app data dir for Windows, and "log" to the
            base cache dir for Unix. See discussion below.

    Typical user log directories are:
        Mac OS X:   ~/Library/Logs/
        Unix:       ~/.cache//log  # or under $XDG_CACHE_HOME if defined
        Win XP:     C:\Documents and Settings\\Local Settings\Application Data\\\Logs
        Vista:      C:\Users\\AppData\Local\\\Logs

    On Windows the only suggestion in the MSDN docs is that local settings
    go in the `CSIDL_LOCAL_APPDATA` directory. (Note: I'm interested in
    examples of what some windows apps use for a logs dir.)

    OPINION: This function appends "Logs" to the `CSIDL_LOCAL_APPDATA`
    value for Windows and appends "log" to the user cache dir for Unix.
    This can be disabled with the `opinion=False` option.
    rz~/Library/LogsrFZLogslog)rr
rrrrr1r0rrruser_log_dirds" 
r4c@sneZdZdZdddZeddZedd	Zed
dZedd
Z	eddZ
eddZeddZdS)AppDirsz1Convenience wrapper for getting application dirs.NFcCs"||_||_||_||_||_dS)N)rrrrr))selfrrrrr)rrr__init__s
zAppDirs.__init__cCst|j|j|j|jdSN)rr)rrrrrr6rrrrs
zAppDirs.user_data_dircCst|j|j|j|jdSN)rr))r*rrrr)r9rrrr*s
zAppDirs.site_data_dircCst|j|j|j|jdSr8)r.rrrrr9rrrr.s
zAppDirs.user_config_dircCst|j|j|j|jdSr:)r/rrrr)r9rrrr/s
zAppDirs.site_config_dircCst|j|j|jdSNr)r1rrrr9rrrr1s
zAppDirs.user_cache_dircCst|j|j|jdSr;)r2rrrr9rrrr2s
zAppDirs.user_state_dircCst|j|j|jdSr;)r4rrrr9rrrr4s
zAppDirs.user_log_dir)NNNFF)
__name__
__module____qualname____doc__r7propertyrr*r.r/r1r2r4rrrrr5s&






r5cCsHtrddl}nddl}dddd|}||jd}|||\}}|S)zThis is a fallback technique at best. I'm not sure if using the
    registry for this guarantees us the correct answer for all CSIDL_*
    names.
    rNZAppDatazCommon AppDataz
Local AppDatar	rr
z@Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders)PY3winreg_winregOpenKeyHKEY_CURRENT_USERZQueryValueEx)
csidl_namerEZshell_folder_namekeydirtyperrr_get_win_folder_from_registrys
rLcCsddlm}m}|dt||dd}z^t|}d}|D]}t|dkr8d}qRq8|rzddl}||}Wnt	k
rYnXWnt
k
rYnX|S)Nr)shellconshellFT)Zwin32com.shellrMrNSHGetFolderPathgetattrunicodeordwin32apiGetShortPathNameImportErrorUnicodeError)rHrMrNrJ
has_high_charcrTrrr_get_win_folder_with_pywin32s$
rZcCsddl}dddd|}|d}|jjd|dd|d}|D]}t|dkr@d	}qZq@|r|d}|jj|j|dr|}|jS)
Nr#rBiFrOT)	ctypesZcreate_unicode_bufferwindllZshell32ZSHGetFolderPathWrSZkernel32ZGetShortPathNameWvalue)rHr^Zcsidl_constbufrXrYZbuf2rrr_get_win_folder_with_ctypess&

rbcCsddl}ddlm}ddlm}|jjd}|d|}|jj	}|
dt|j|d|jj
||j|d}d}|D]}	t|	dkr|d	}qq||r|d|}|jj	}
|
|||r|j|d}|S)
Nr)jna)rrYFrOT)arrayZcom.sunrcZcom.sun.jna.platformrZWinDefZMAX_PATHZzerosZShell32ZINSTANCErPrQZShlObjZSHGFP_TYPE_CURRENTZNativeZtoStringZtostringrrSZKernel32rU)rHrfrcrZbuf_sizerarNrJrXrYZkernelrrr_get_win_folder_with_jnas&rg)r___main__ZMyAppZ	MyCompany)rr.r1r2r4r*r/z-- app dirs %s --z%-- app dirs (with optional 'version')z1.0r<z%s: %sz)
-- app dirs (without optional 'version')z+
-- app dirs (without optional 'appauthor')z(
-- app dirs (with disabled 'appauthor'))r)NNNF)NNNF)NNNF)NNNF)NNNT)NNNF)NNNT)+r@Z__version_info__rmapstr__version__sysr
version_inforCrRplatform
startswithZjava_verZos_namerrr*r.r/r1r2r4objectr5rLrZrbrgr^r_rrVZcom.sun.jnaZcomr=rrZpropsprintdirsZproprQrrrrsv



7
B
(
3
9
*
30


site-packages/pip/_vendor/__pycache__/distro.cpython-38.opt-1.pyc000064400000107170151464452710020632 0ustar00U

.e@s~dZddlZddlZddlZddlZddlZddlZddlZddlZej	
ddZdZddiZ
dddd	Zd
diZedZedZd
ddedfZd@ddZddZdAddZdBddZdCddZdDddZdEdd ZdFd!d"Zd#d$Zd%d&ZdGd'd(Zd)d*Zd+d,Z d-d.Z!d/d0Z"d1d2Z#d3d4Z$d5d6Z%d7d8Z&Gd9d:d:e'Z(Gd;d<dZ+e,d?krze+dS)Ha
The ``distro`` package (``distro`` stands for Linux Distribution) provides
information about the Linux distribution it runs on, such as a reliable
machine-readable distro ID, or version information.

It is the recommended replacement for Python's original
:py:func:`platform.linux_distribution` function, but it provides much more
functionality. An alternative implementation became necessary because Python
3.5 deprecated this function, and Python 3.8 will remove it altogether.
Its predecessor function :py:func:`platform.dist` was already
deprecated since Python 2.6 and will also be removed in Python 3.8.
Still, there are many cases in which access to OS distribution information
is needed. See `Python issue 1322 `_ for
more information.
NZUNIXCONFDIRz/etcz
os-releaseZolZoracleZrhel)ZenterpriseenterpriseZredhatenterpriseworkstationZredhatenterpriseserverZredhatzA(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)z(\w+)[-_](release|version)$Zdebian_versionzlsb-releasezoem-releasezsystem-releaseTcCs
t|S)a
    Return information about the current OS distribution as a tuple
    ``(id_name, version, codename)`` with items as follows:

    * ``id_name``:  If *full_distribution_name* is false, the result of
      :func:`distro.id`. Otherwise, the result of :func:`distro.name`.

    * ``version``:  The result of :func:`distro.version`.

    * ``codename``:  The result of :func:`distro.codename`.

    The interface of this function is compatible with the original
    :py:func:`platform.linux_distribution` function, supporting a subset of
    its parameters.

    The data it returns may not exactly be the same, because it uses more data
    sources than the original function, and that may lead to different data if
    the OS distribution is not consistent across multiple data sources it
    provides (there are indeed such distributions ...).

    Another reason for differences is the fact that the :func:`distro.id`
    method normalizes the distro ID string to a reliable machine-readable value
    for a number of popular OS distributions.
    )_distrolinux_distribution)full_distribution_namer6/usr/lib/python3.8/site-packages/pip/_vendor/distro.pyrasrcCstS)a$
    Return the distro ID of the current distribution, as a
    machine-readable string.

    For a number of OS distributions, the returned distro ID value is
    *reliable*, in the sense that it is documented and that it does not change
    across releases of the distribution.

    This package maintains the following reliable distro ID values:

    ==============  =========================================
    Distro ID       Distribution
    ==============  =========================================
    "ubuntu"        Ubuntu
    "debian"        Debian
    "rhel"          RedHat Enterprise Linux
    "centos"        CentOS
    "fedora"        Fedora
    "sles"          SUSE Linux Enterprise Server
    "opensuse"      openSUSE
    "amazon"        Amazon Linux
    "arch"          Arch Linux
    "cloudlinux"    CloudLinux OS
    "exherbo"       Exherbo Linux
    "gentoo"        GenToo Linux
    "ibm_powerkvm"  IBM PowerKVM
    "kvmibm"        KVM for IBM z Systems
    "linuxmint"     Linux Mint
    "mageia"        Mageia
    "mandriva"      Mandriva Linux
    "parallels"     Parallels
    "pidora"        Pidora
    "raspbian"      Raspbian
    "oracle"        Oracle Linux (and Oracle Enterprise Linux)
    "scientific"    Scientific Linux
    "slackware"     Slackware
    "xenserver"     XenServer
    "openbsd"       OpenBSD
    "netbsd"        NetBSD
    "freebsd"       FreeBSD
    ==============  =========================================

    If you have a need to get distros for reliable IDs added into this set,
    or if you find that the :func:`distro.id` function returns a different
    distro ID for one of the listed distros, please create an issue in the
    `distro issue tracker`_.

    **Lookup hierarchy and transformations:**

    First, the ID is obtained from the following sources, in the specified
    order. The first available and non-empty value is used:

    * the value of the "ID" attribute of the os-release file,

    * the value of the "Distributor ID" attribute returned by the lsb_release
      command,

    * the first part of the file name of the distro release file,

    The so determined ID value then passes the following transformations,
    before it is returned by this method:

    * it is translated to lower case,

    * blanks (which should not be there anyway) are translated to underscores,

    * a normalization of the ID is performed, based upon
      `normalization tables`_. The purpose of this normalization is to ensure
      that the ID is as reliable as possible, even across incompatible changes
      in the OS distributions. A common reason for an incompatible change is
      the addition of an os-release file, or the addition of the lsb_release
      command, with ID values that differ from what was previously determined
      from the distro release file name.
    )ridrrrrr}sKrFcCs
t|S)ak
    Return the name of the current OS distribution, as a human-readable
    string.

    If *pretty* is false, the name is returned without version or codename.
    (e.g. "CentOS Linux")

    If *pretty* is true, the version and codename are appended.
    (e.g. "CentOS Linux 7.1.1503 (Core)")

    **Lookup hierarchy:**

    The name is obtained from the following sources, in the specified order.
    The first available and non-empty value is used:

    * If *pretty* is false:

      - the value of the "NAME" attribute of the os-release file,

      - the value of the "Distributor ID" attribute returned by the lsb_release
        command,

      - the value of the "" field of the distro release file.

    * If *pretty* is true:

      - the value of the "PRETTY_NAME" attribute of the os-release file,

      - the value of the "Description" attribute returned by the lsb_release
        command,

      - the value of the "" field of the distro release file, appended
        with the value of the pretty version ("" and ""
        fields) of the distro release file, if available.
    )rnameprettyrrrrs$rcCst||S)av
    Return the version of the current OS distribution, as a human-readable
    string.

    If *pretty* is false, the version is returned without codename (e.g.
    "7.0").

    If *pretty* is true, the codename in parenthesis is appended, if the
    codename is non-empty (e.g. "7.0 (Maipo)").

    Some distributions provide version numbers with different precisions in
    the different sources of distribution information. Examining the different
    sources in a fixed priority order does not always yield the most precise
    version (e.g. for Debian 8.2, or CentOS 7.1).

    The *best* parameter can be used to control the approach for the returned
    version:

    If *best* is false, the first non-empty version number in priority order of
    the examined sources is returned.

    If *best* is true, the most precise version number out of all examined
    sources is returned.

    **Lookup hierarchy:**

    In all cases, the version number is obtained from the following sources.
    If *best* is false, this order represents the priority order:

    * the value of the "VERSION_ID" attribute of the os-release file,
    * the value of the "Release" attribute returned by the lsb_release
      command,
    * the version number parsed from the "" field of the first line
      of the distro release file,
    * the version number parsed from the "PRETTY_NAME" attribute of the
      os-release file, if it follows the format of the distro release files.
    * the version number parsed from the "Description" attribute returned by
      the lsb_release command, if it follows the format of the distro release
      files.
    )rversionr
bestrrrrs)rcCs
t|S)a
    Return the version of the current OS distribution as a tuple
    ``(major, minor, build_number)`` with items as follows:

    * ``major``:  The result of :func:`distro.major_version`.

    * ``minor``:  The result of :func:`distro.minor_version`.

    * ``build_number``:  The result of :func:`distro.build_number`.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r
version_partsr
rrrrsrcCs
t|S)a5
    Return the major version of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The major version is the first
    part of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r
major_versionrrrrr/s
rcCs
t|S)a6
    Return the minor version of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The minor version is the second
    part of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r
minor_versionrrrrr<s
rcCs
t|S)a3
    Return the build number of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The build number is the third part
    of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )rbuild_numberrrrrrIs
rcCstS)a
    Return a space-separated list of distro IDs of distributions that are
    closely related to the current OS distribution in regards to packaging
    and programming interfaces, for example distributions the current
    distribution is a derivative from.

    **Lookup hierarchy:**

    This information item is only provided by the os-release file.
    For details, see the description of the "ID_LIKE" attribute in the
    `os-release man page
    `_.
    )rlikerrrrrVsrcCstS)a
    Return the codename for the release of the current OS distribution,
    as a string.

    If the distribution does not have a codename, an empty string is returned.

    Note that the returned codename is not always really a codename. For
    example, openSUSE returns "x86_64". This function does not handle such
    cases in any special way and just returns the string it finds, if any.

    **Lookup hierarchy:**

    * the codename within the "VERSION" attribute of the os-release file, if
      provided,

    * the value of the "Codename" attribute returned by the lsb_release
      command,

    * the value of the "" field of the distro release file.
    )rcodenamerrrrrgsrcCst||S)a
    Return certain machine-readable information items about the current OS
    distribution in a dictionary, as shown in the following example:

    .. sourcecode:: python

        {
            'id': 'rhel',
            'version': '7.0',
            'version_parts': {
                'major': '7',
                'minor': '0',
                'build_number': ''
            },
            'like': 'fedora',
            'codename': 'Maipo'
        }

    The dictionary structure and keys are always the same, regardless of which
    information items are available in the underlying data sources. The values
    for the various keys are as follows:

    * ``id``:  The result of :func:`distro.id`.

    * ``version``:  The result of :func:`distro.version`.

    * ``version_parts -> major``:  The result of :func:`distro.major_version`.

    * ``version_parts -> minor``:  The result of :func:`distro.minor_version`.

    * ``version_parts -> build_number``:  The result of
      :func:`distro.build_number`.

    * ``like``:  The result of :func:`distro.like`.

    * ``codename``:  The result of :func:`distro.codename`.

    For a description of the *pretty* and *best* parameters, see the
    :func:`distro.version` method.
    )rinforrrrrs)rcCstS)z
    Return a dictionary containing key-value pairs for the information items
    from the os-release file data source of the current OS distribution.

    See `os-release file`_ for details about these information items.
    )ros_release_inforrrrrsrcCstS)z
    Return a dictionary containing key-value pairs for the information items
    from the lsb_release command data source of the current OS distribution.

    See `lsb_release command output`_ for details about these information
    items.
    )rlsb_release_inforrrrrsrcCstS)z
    Return a dictionary containing key-value pairs for the information items
    from the distro release file data source of the current OS distribution.

    See `distro release file`_ for details about these information items.
    )rdistro_release_inforrrrrsrcCstS)z
    Return a dictionary containing key-value pairs for the information items
    from the distro release file data source of the current OS distribution.
    )r
uname_inforrrrrsrcCs
t|S)a
    Return a single named information item from the os-release file data source
    of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `os-release file`_ for details about these information items.
    )ros_release_attr	attributerrrrsrcCs
t|S)a
    Return a single named information item from the lsb_release command output
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `lsb_release command output`_ for details about these information
    items.
    )rlsb_release_attrrrrrrsrcCs
t|S)a
    Return a single named information item from the distro release file
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `distro release file`_ for details about these information items.
    )rdistro_release_attrrrrrrsrcCs
t|S)aZ
    Return a single named information item from the distro release file
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
                The empty string, if the item does not exist.
    )r
uname_attrrrrrrsrc@s eZdZdZddZddZdS)cached_propertyzA version of @property which caches the value.  On access, it calls the
    underlying function and sets the value in `__dict__` so future accesses
    will not re-call the property.
    cCs|j|_||_dSN)__name___fname_f)selffrrr__init__"szcached_property.__init__cCs||}|j|j<|Sr!)r$__dict__r#)r%objownerretrrr__get__&szcached_property.__get__N)r"
__module____qualname____doc__r'r,rrrrr sr c@s2eZdZdZdBddZddZdCdd	Zd
dZdDd
dZdEddZ	dFddZ
dGddZdHddZdIddZ
ddZddZdJddZdd Zd!d"Zd#d$Zd%d&Zd'd(Zd)d*Zd+d,Zd-d.Zed/d0Zed1d2Zed3d4Zed5d6Zed7d8Zed9d:Z ed;d<Z!d=d>Z"ed?d@Z#dAS)KLinuxDistributiona
    Provides information about a OS distribution.

    This package creates a private module-global instance of this class with
    default initialization arguments, that is used by the
    `consolidated accessor functions`_ and `single source accessor functions`_.
    By using default initialization arguments, that module-global instance
    returns data about the current OS distribution (i.e. the distro this
    package runs on).

    Normally, it is not necessary to create additional instances of this class.
    However, in situations where control is needed over the exact data sources
    that are used, instances of this class can be created with a specific
    distro release file, or a specific os-release file, or without invoking the
    lsb_release command.
    TcCs.|ptjtt|_|pd|_||_||_dS)ao
        The initialization method of this class gathers information from the
        available data sources, and stores that in private instance attributes.
        Subsequent access to the information items uses these private instance
        attributes, so that the data sources are read only once.

        Parameters:

        * ``include_lsb`` (bool): Controls whether the
          `lsb_release command output`_ is included as a data source.

          If the lsb_release command is not available in the program execution
          path, the data source for the lsb_release command will be empty.

        * ``os_release_file`` (string): The path name of the
          `os-release file`_ that is to be used as a data source.

          An empty string (the default) will cause the default path name to
          be used (see `os-release file`_ for details).

          If the specified or defaulted os-release file does not exist, the
          data source for the os-release file will be empty.

        * ``distro_release_file`` (string): The path name of the
          `distro release file`_ that is to be used as a data source.

          An empty string (the default) will cause a default search algorithm
          to be used (see `distro release file`_ for details).

          If the specified distro release file does not exist, or if no default
          distro release file can be found, the data source for the distro
          release file will be empty.

        * ``include_name`` (bool): Controls whether uname command output is
          included as a data source. If the uname command is not available in
          the program execution path the data source for the uname command will
          be empty.

        Public instance attributes:

        * ``os_release_file`` (string): The path name of the
          `os-release file`_ that is actually used as a data source. The
          empty string if no distro release file is used as a data source.

        * ``distro_release_file`` (string): The path name of the
          `distro release file`_ that is actually used as a data source. The
          empty string if no distro release file is used as a data source.

        * ``include_lsb`` (bool): The result of the ``include_lsb`` parameter.
          This controls whether the lsb information will be loaded.

        * ``include_uname`` (bool): The result of the ``include_uname``
          parameter. This controls whether the uname information will
          be loaded.

        Raises:

        * :py:exc:`IOError`: Some I/O issue with an os-release file or distro
          release file.

        * :py:exc:`subprocess.CalledProcessError`: The lsb_release command had
          some issue (other than not being available in the program execution
          path).

        * :py:exc:`UnicodeError`: A data source has unexpected characters or
          uses an unexpected encoding.
        r1N)	ospathjoin_UNIXCONFDIR_OS_RELEASE_BASENAMEos_release_filedistro_release_fileinclude_lsb
include_uname)r%r9r7r8r:rrrr'>sH
zLinuxDistribution.__init__cCsdj|dS)z Return repr of all info
        aeLinuxDistribution(os_release_file={self.os_release_file!r}, distro_release_file={self.distro_release_file!r}, include_lsb={self.include_lsb!r}, include_uname={self.include_uname!r}, _os_release_info={self._os_release_info!r}, _lsb_release_info={self._lsb_release_info!r}, _distro_release_info={self._distro_release_info!r}, _uname_info={self._uname_info!r})r%)formatr;rrr__repr__s	zLinuxDistribution.__repr__cCs"|r|n|||fS)z
        Return information about the OS distribution that is compatible
        with Python's :func:`platform.linux_distribution`, supporting a subset
        of its parameters.

        For details, see :func:`distro.linux_distribution`.
        )rrrr)r%rrrrrs	z$LinuxDistribution.linux_distributioncCsldd}|d}|r ||tS|d}|r8||tS|d}|rP||tS|d}|rh||tSdS)zoReturn the distro ID of the OS distribution, as a string.

        For details, see :func:`distro.id`.
        cSs|dd}|||S)N _)lowerreplaceget)	distro_idtablerrr	normalizesz'LinuxDistribution.id..normalizerdistributor_idr1)rNORMALIZED_OS_IDrNORMALIZED_LSB_IDrNORMALIZED_DISTRO_IDr)r%rErCrrrrs







zLinuxDistribution.idFcCs||dp&|dp&|dp&|d}|rt|dp>|d}|st|dpV|d}|jdd}|rt|d|}|pzdS)	zu
        Return the name of the OS distribution, as a string.

        For details, see :func:`distro.name`.
        rrFpretty_namedescriptionTr	r>r1)rrrrr)r%r
rrrrrrs&


zLinuxDistribution.namec	Cs|d|d|d||ddd||ddd|dg}d}|r|D]$}|d|dks||dkr\|}q\n|D]}|dkr|}qq|r|r|rd||}|S)z{
        Return the version of the OS distribution, as a string.

        For details, see :func:`distro.version`.
        
version_idreleaserJr1rK.z	{0} ({1}))	rrr_parse_distro_release_contentrBrcountrr<)r%r
r
Zversionsrvrrrrs:
zLinuxDistribution.versioncCsL|j|d}|rHtd}||}|rH|\}}}||p>d|pDdfSdS)z
        Return the version of the OS distribution, as a tuple of version
        numbers.

        For details, see :func:`distro.version_parts`.
        rz(\d+)\.?(\d+)?\.?(\d+)?r1)r1r1r1)rrecompilematchgroups)r%r
Zversion_strZ
version_regexmatchesmajorminorrrrrrs

zLinuxDistribution.version_partscCs||dS)z
        Return the major version number of the current distribution.

        For details, see :func:`distro.major_version`.
        rrr%r
rrrrszLinuxDistribution.major_versioncCs||dS)z
        Return the minor version number of the current distribution.

        For details, see :func:`distro.minor_version`.
        rYrZrrrrszLinuxDistribution.minor_versioncCs||dS)z}
        Return the build number of the current distribution.

        For details, see :func:`distro.build_number`.
        rYrZrrrrszLinuxDistribution.build_numbercCs|dpdS)z
        Return the IDs of distributions that are like the OS distribution.

        For details, see :func:`distro.like`.
        Zid_liker1)rr;rrrr#szLinuxDistribution.likecCs>z|jdWStk
r8|dp2|dp2dYSXdS)zp
        Return the codename of the OS distribution.

        For details, see :func:`distro.codename`.
        rr1N)_os_release_infoKeyErrorrrr;rrrr+s
zLinuxDistribution.codenamec	CsBt||||t||||||d||dS)z
        Return certain machine-readable information about the OS
        distribution.

        For details, see :func:`distro.info`.
        )rWrXr)rrrrr)dictrrrrrrr)r%r
r
rrrr:s
zLinuxDistribution.infocCs|jS)z
        Return a dictionary containing key-value pairs for the information
        items from the os-release file data source of the OS distribution.

        For details, see :func:`distro.os_release_info`.
        )r]r;rrrrMsz!LinuxDistribution.os_release_infocCs|jS)z
        Return a dictionary containing key-value pairs for the information
        items from the lsb_release command data source of the OS
        distribution.

        For details, see :func:`distro.lsb_release_info`.
        )_lsb_release_infor;rrrrVsz"LinuxDistribution.lsb_release_infocCs|jS)z
        Return a dictionary containing key-value pairs for the information
        items from the distro release file data source of the OS
        distribution.

        For details, see :func:`distro.distro_release_info`.
        )_distro_release_infor;rrrr`sz%LinuxDistribution.distro_release_infocCs|jS)z
        Return a dictionary containing key-value pairs for the information
        items from the uname command data source of the OS distribution.

        For details, see :func:`distro.uname_info`.
        )_uname_infor;rrrrjszLinuxDistribution.uname_infocCs|j|dS)z
        Return a single named information item from the os-release file data
        source of the OS distribution.

        For details, see :func:`distro.os_release_attr`.
        r1)r]rBr%rrrrrssz!LinuxDistribution.os_release_attrcCs|j|dS)z
        Return a single named information item from the lsb_release command
        output data source of the OS distribution.

        For details, see :func:`distro.lsb_release_attr`.
        r1)r`rBrcrrrr|sz"LinuxDistribution.lsb_release_attrcCs|j|dS)z
        Return a single named information item from the distro release file
        data source of the OS distribution.

        For details, see :func:`distro.distro_release_attr`.
        r1)rarBrcrrrrsz%LinuxDistribution.distro_release_attrcCs|j|dS)z
        Return a single named information item from the uname command
        output data source of the OS distribution.

        For details, see :func:`distro.uname_release_attr`.
        r1)rbrBrcrrrrszLinuxDistribution.uname_attrc
Cs:tj|jr6t|j}||W5QRSQRXiS)z
        Get the information items from the specified os-release file.

        Returns:
            A dictionary containing all information items.
        )r2r3isfiler7open_parse_os_release_content)r%Zrelease_filerrrr]sz"LinuxDistribution._os_release_infocCsi}tj|dd}d|_tjddkr@t|jtr@|jd|_t|}|D]>}d|krL|	dd\}}t|tr||d}|||
<qLqLd	|kr|d	|d
<nbd|kr|d|d
<nLd|krtd
|d}|r|
}|d}|d}|}||d
<|S)aD
        Parse the lines of an os-release file.

        Parameters:

        * lines: Iterable through the lines in the os-release file.
                 Each line must be a unicode string or a UTF-8 encoded byte
                 string.

        Returns:
            A dictionary containing all information items.
        T)posixrr\z
iso-8859-1=r[utf-8Zversion_codenamerZubuntu_codenamerz(\(\D+\))|,(\s+)?\D+z(),)shlexZwhitespace_splitsysversion_info
isinstanceZ	wordcharsbytesdecodelistsplitr@rRsearchgroupstrip)linespropsZlexertokenstokenkrQrrrrrfs4	




z+LinuxDistribution._parse_os_release_contentc
Csz|js
iSttjdB}zd}tj||d}Wn$tk
rRiYW5QRSXW5QRX|t	
}||S)z
        Get the information items from the lsb_release command output.

        Returns:
            A dictionary containing all information items.
        w)Zlsb_releasez-astderr)r9rer2devnull
subprocesscheck_outputOSErrorrprlgetfilesystemencoding
splitlines_parse_lsb_release_contentr%r~cmdstdoutZcontentrrrr`s z#LinuxDistribution._lsb_release_infocCsXi}|D]J}|ddd}t|dkr,q|\}}||dd|iq|S)aM
        Parse the output of the lsb_release command.

        Parameters:

        * lines: Iterable through the lines of the lsb_release output.
                 Each line must be a unicode string or a UTF-8 encoded byte
                 string.

        Returns:
            A dictionary containing all information items.
        
:r[r\r>r?)rurrlenupdaterAr@)rvrwlineZkvrzrQrrrrs z,LinuxDistribution._parse_lsb_release_contentc
CspttjdB}zd}tj||d}Wn$tk
rHiYW5QRSXW5QRX|t	}|
|S)Nr{)unamez-rsr|)rer2r~rrrrprlrr_parse_uname_contentrrrrrbs zLinuxDistribution._uname_infocCsTi}td|d}|rP|\}}|dkr4iS||d<||d<||d<|S)Nz^([^\s]+)\s+([\d\.]+)rZLinuxrrrM)rRrsrurUr@)rvrwrTrrrrrrsz&LinuxDistribution._parse_uname_contentcCs0|jrb||j}tj|j}t|}d|krLd|dkrLd|d<n|r^|d|d<|Szt	t
}|Wn6tk
rddddd	d
ddd
ddddddg}YnX|D]r}|t
krqt|}|rtjt
|}||}d|kr||_|d|d<d|dkrd|d<|SqiSdS)z
        Get the information items from the specified distro release file.

        Returns:
            A dictionary containing all information items.
        rZ
cloudlinuxrr[zSuSE-releasezarch-releasezbase-releasezcentos-releasezfedora-releasezgentoo-releasezmageia-releasezmandrake-releasezmandriva-releasezmandrivalinux-releasezmanjaro-releasezoracle-releasezredhat-releasez
sl-releasezslackware-versionN)r8_parse_distro_release_filer2r3basename _DISTRO_RELEASE_BASENAME_PATTERNrTr@rtlistdirr5sortr _DISTRO_RELEASE_IGNORE_BASENAMESr4)r%distro_inforrTZ	basenamesfilepathrrrra.s^






z&LinuxDistribution._distro_release_infoc
CsRz0t|}||W5QRWSQRXWnttfk
rLiYSXdS)z
        Parse a distro release file.

        Parameters:

        * filepath: Path name of the distro release file.

        Returns:
            A dictionary containing all information items.
        N)rerOreadlinerIOError)r%rfprrrrrs

&z,LinuxDistribution._parse_distro_release_filecCst|tr|d}t|ddd}i}|r|dddd|d<|drn|dddd|d<|dr|dddd|d	<n|r||d<|S)
a
        Parse a line from a distro release file.

        Parameters:
        * line: Line from the distro release file. Must be a unicode string
                or a UTF-8 encoded byte string.

        Returns:
            A dictionary containing all information items.
        riNrr\rLr[r)rnrorp(_DISTRO_RELEASE_CONTENT_REVERSED_PATTERNrTrurt)rrVrrrrrOs



z/LinuxDistribution._parse_distro_release_contentN)Tr1r1T)T)F)FF)F)F)F)F)FF)$r"r-r.r/r'r=rrrrrrrrrrrrrrrrrrrr r]staticmethodrfr`rrbrrarrOrrrrr0,sV
N


"




	

					

B





Cr0cCstt}|tj|ttjt	j
dd}|jddddd|}|j
rn|t
jtdd	d
n:|dtd	dtd	d}|d
|t}|d|dS)NzOS distro info tool)rKz--jsonz-jz!Output in machine readable format
store_true)helpactionT)indentZ	sort_keyszName: %sr	zVersion: %szCodename: %s)loggingZ	getLoggerr"ZsetLevelDEBUGZ
addHandlerZ
StreamHandlerrlrargparseArgumentParseradd_argument
parse_argsjsonrdumpsrrr)ZloggerparserargsZdistribution_versionZdistribution_codenamerrrmains$

r__main__)T)F)FF)F)F)F)F)FF)-r/r2rRrlrrkrrrenvironrBr5r6rGrHrIrSrrrrrrrrrrrrrrrrrrrrrrobjectr r0rrr"rrrrsz
	
N
'
,





,

}
site-packages/pip/_vendor/__pycache__/contextlib2.cpython-38.opt-1.pyc000064400000036202151464452710021560 0ustar00U

.eB	@sdZddlZddlZddlZddlmZddlmZddddd	d
ddd
g	Zedg7Zej	dddkrnej
Znede
fddiZddZddZGdddeZGdd	d	e
ZGdddeZddZGddde
ZGddde
ZGd ddeZGd!ddeZGd"d
d
e
Zej	dd#kZer@d$d%Zd&d'Znd(d%Zd)d'Zzdd*lmZWnek
rzeZ Yn
Xd+d,Z Gd-d
d
e
Z!Gd.dde!Z"Gd/ddeZ#dS)0zAcontextlib2 - backports and enhancements to the contextlib moduleN)dequewrapscontextmanagerclosingnullcontextAbstractContextManagerContextDecorator	ExitStackredirect_stdoutredirect_stderrsuppressContextStack)ABC	__slots__cCs0||krdS|||jD]}t||q|SN)append	__bases___classic_mro)CresultBrr;/usr/lib/python3.8/site-packages/pip/_vendor/contextlib2.pyrs

rcGsvz
|j}Wn"tk
r,tt|g}YnX|D]>}|D],}||jkr:|j|dkrbtSq2q:tSq2dS)NT)__mro__AttributeErrortupler__dict__NotImplemented)rmethodsmromethodrrrr_check_methods$s


r%c@s2eZdZdZddZejddZeddZ	dS)	rz,An abstract base class for context managers.cCs|S)z0Return `self` upon entering the runtime context.rselfrrr	__enter__8sz AbstractContextManager.__enter__cCsdS)z9Raise any exception triggered within the runtime context.Nr)r'exc_type	exc_value	tracebackrrr__exit__<szAbstractContextManager.__exit__cCs|tkrt|ddStS)z.innerr)r'r?r@rr>r__call__iszContextDecorator.__call__N)r/r0r1r2r:r9rArrrrr	Isc@s0eZdZdZddZddZddZdd	Zd
S)_GeneratorContextManagerz%Helper for @contextmanager decorator.cCsJ||||_||||_|_|_t|dd}|dkr@t|j}||_dS)Nr2)genr?r<r=getattrtyper2)r'r?r<r=docrrr__init__ts
z!_GeneratorContextManager.__init__cCs||j|j|jSr)	__class__r?r<r=r&rrrr9sz%_GeneratorContextManager._recreate_cmcCs.zt|jWStk
r(tdYnXdS)Nzgenerator didn't yield)nextrC
StopIterationRuntimeErrorr&rrrr(sz"_GeneratorContextManager.__enter__c
Cs|dkr8zt|jWntk
r,YdSXtdn|dkrF|}z|j|||tdWntk
r}z||k	WYSd}~XYnjtk
r}z0||krWY"dStr|j|krWY
dSW5d}~XYntd|k	rYnXdS)Nzgenerator didn't stopz#generator didn't stop after throw()F)	rIrCrJrKthrow_HAVE_EXCEPTION_CHAINING	__cause__sysexc_info)r'rEvaluer+excrrrr,s,


z!_GeneratorContextManager.__exit__N)r/r0r1r2rGr9r(r,rrrrrBqs
rBcstfdd}|S)a@contextmanager decorator.

    Typical usage:

        @contextmanager
        def some_generator():
            
            try:
                yield 
            finally:
                

    This makes this:

        with some_generator() as :
            

    equivalent to this:

        
        try:
             = 
            
        finally:
            

    cst||Sr)rBr;r?rrhelperszcontextmanager..helperr)r?rUrrTrrsc@s(eZdZdZddZddZddZdS)	ra2Context to automatically close something at the end of a block.

    Code like this:

        with closing(.open()) as f:
            

    is equivalent to this:

        f = .open()
        try:
            
        finally:
            f.close()

    cCs
||_dSrthing)r'rWrrrrGszclosing.__init__cCs|jSrrVr&rrrr(szclosing.__enter__cGs|jdSr)rWclose)r'rQrrrr,szclosing.__exit__Nr/r0r1r2rGr(r,rrrrrsc@s(eZdZdZddZddZddZdS)_RedirectStreamNcCs||_g|_dSr)_new_target_old_targets)r'
new_targetrrrrGsz_RedirectStream.__init__cCs*|jtt|jtt|j|j|jSr)r\rrDrP_streamsetattrr[r&rrrr(sz_RedirectStream.__enter__cCstt|j|jdSr)r_rPr^r\popr'exctypeexcinstexctbrrrr,sz_RedirectStream.__exit__)r/r0r1r^rGr(r,rrrrrZsrZc@seZdZdZdZdS)raAContext manager for temporarily redirecting stdout to another file.

        # How to send help() to stderr
        with redirect_stdout(sys.stderr):
            help(dir)

        # How to write help() to a file
        with open('help.txt', 'w') as f:
            with redirect_stdout(f):
                help(pow)
    stdoutNr/r0r1r2r^rrrrrsc@seZdZdZdZdS)rzCContext manager for temporarily redirecting stderr to another file.stderrNrfrrrrrsc@s(eZdZdZddZddZddZdS)	r
a?Context manager to suppress specified exceptions

    After the exception is suppressed, execution proceeds with the next
    statement following the with statement.

         with suppress(FileNotFoundError):
             os.remove(somefile)
         # Execution still resumes here if the file was already removed
    cGs
||_dSr)_exceptions)r'
exceptionsrrrrG)szsuppress.__init__cCsdSrrr&rrrr(,szsuppress.__enter__cCs|dk	ot||jSr)
issubclassrhrarrrr,/s
zsuppress.__exit__NrYrrrrr
s
rcsfdd}|S)Ncs4|j}||krdS|dks*|kr$q*|}q||_dSr)__context__)new_excold_excexc_context	frame_excrr_fix_exception_context@sz3_make_context_fixer.._fix_exception_contextr)rprqrror_make_context_fixer?s
rrcCs<z|dj}|dWn tk
r6||d_YnXdS)NrL)rk
BaseException)exc_details	fixed_ctxrrr_reraise_with_existing_contextOs

rvcCsddS)NcSsdSrr)rlrmrrr[z%_make_context_fixer..rrorrrrrZscCs|\}}}tddS)Nz!raise exc_type, exc_value, exc_tb)exec)rtr)r*Zexc_tbrrrrv_s
)InstanceTypecCst|}|tkr|jS|Sr)rErzrH)objZobj_typerrr	_get_typeksr|c@sXeZdZdZddZddZddZdd	Zd
dZdd
Z	ddZ
ddZddZdS)r
aContext manager for dynamic management of a stack of exit callbacks

    For example:

        with ExitStack() as stack:
            files = [stack.enter_context(open(fname)) for fname in filenames]
            # All opened files will automatically be closed at the end of
            # the with statement, even if attempts to open files later
            # in the list raise an exception

    cCst|_dSr)r_exit_callbacksr&rrrrGszExitStack.__init__cCst|}|j|_t|_|S)z?Preserve the context stack by transferring it to a new instance)rEr}r)r'	new_stackrrrpop_alls
zExitStack.pop_allcs"fdd}|_||dS)z:Helper to correctly register callbacks to __exit__ methodscsf|Srr)rtcmcm_exitrr
_exit_wrappersz.ExitStack._push_cm_exit.._exit_wrapperN)__self__push)r'rrrrrr
_push_cm_exitszExitStack._push_cm_exitcCsDt|}z
|j}Wn tk
r2|j|YnX||||S)aRegisters a callback with the standard __exit__ method signature

        Can suppress exceptions the same way __exit__ methods can.

        Also accepts any object with an __exit__ method (registering a call
        to the method instead of the object itself)
        )r|r,rr}rr)r'exit_cb_typeexit_methodrrrrs

zExitStack.pushcs$fdd}|_||S)z\Registers an arbitrary callback and arguments.

        Cannot suppress exceptions.
        csdSrr)r)rStbr<callbackr=rrrsz)ExitStack.callback.._exit_wrapper)__wrapped__r)r'rr<r=rrrrrs
zExitStack.callbackcCs(t|}|j}||}||||S)zEnters the supplied context manager

        If successful, also pushes its __exit__ method as a callback and
        returns the result of the __enter__ method.
        )r|r,r(r)r'r_cm_type_exitrrrr
enter_contexts

zExitStack.enter_contextcCs|ddddS)z$Immediately unwind the context stackN)r,r&rrrrXszExitStack.closecCs|Srrr&rrrr(szExitStack.__enter__c	Gs|ddk	}td}t|}d}d}|jr|j}z||rNd}d}d}Wq(t}||d|dd}|}Yq(Xq(|rt||o|S)NrrLFT)NNN)rPrQrrr}r`rv)	r'rtreceived_excrprqsuppressed_exc
pending_raisecbnew_exc_detailsrrrr,s(
zExitStack.__exit__N)
r/r0r1r2rGrrrrrrXr(r,rrrrr
ss

cs8eZdZdZfddZddZddZdd	ZZS)
rz+Backwards compatibility alias for ExitStackcstdttt|dS)Nz*ContextStack has been renamed to ExitStack)r6r7r8superrrGr&rHrrrGszContextStack.__init__cCs
||Sr)r)r'rrrr
register_exitszContextStack.register_exitcOs|j|f||Sr)r)r'rr<r=rrrregisterszContextStack.registercCs|Sr)rr&rrrpreserveszContextStack.preserve)	r/r0r1r2rGrrr
__classcell__rrrrrs
c@s*eZdZdZd	ddZddZddZdS)
raMContext manager that does no additional processing.
    Used as a stand-in for a normal context manager, when a particular
    block of code is only sometimes used with a normal context manager:
    cm = optional_cm if condition else nullcontext()
    with cm:
        # Perform operation, using optional_cm if condition is True
    NcCs
||_dSrenter_result)r'rrrrrGsznullcontext.__init__cCs|jSrrr&rrrr(sznullcontext.__enter__cGsdSrr)r'excinforrrr,sznullcontext.__exit__)NrYrrrrrs
)$r2r3rPr6collectionsr	functoolsr__all__version_inforZ_abc_ABCABCMetaobjectrr%rr	rBrrrZrrr
rNrrrvtypesrzImportErrorrEr|r
rrrrrrsT

(H"

qsite-packages/pip/_vendor/__pycache__/pyparsing.cpython-38.pyc000064400000705501151464452710020405 0ustar00U

.e,s@sdZdZdZdZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZddlmZddlZdd	lmZzdd
lmZWn ek
rddlmZYnXzddlmZWn ek
rddlmZYnXz dd
lmZddlmZm Z Wn2ek
rNdd
l
mZddl
mZm Z YnXzddl
m!Z"WnBek
rzddl#m!Z"Wnek
rdZ"YnXYnXzddlm$Z$Wn$ek
rGdddZ$YnXe$Z%de%_de%_&e$Z'de'_de'_(de'_)de'_*de'_+de'_,dddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dddddddddddgsZ-e.e	j/ddZ0e0ddkZ1e1rVe	j2Z3e4Z5e6Z7e4Z8e4Z9e:e;ee.e?e@eAeBeCgZDn`e	jEZ3eFZGddZ9gZDddlHZHdID]8ZJzeDKeLeHeJWneMk
rYq|YnXq|eNddeGdDZOddZPejQejRZSdZTeTdZUeSeTZVe6dZWdXddejYDZZdYddZ[Gdd4d4e\Z]Gdd6d6e]Z^Gdd8d8e]Z_Gdd:d:e_Z`Gdd=d=e\ZaGdddebZcGdd9d9ebZdeeedddQZfddbZgdd_ZhddZiddZjddZkddiZldZddZmGdd;d;ebZnGdddenZoGddCdCenZpGdd#d#epZqGdd.d.epZrGdd+d+epZsGdddesZtesZuesen_vGdd(d(epZwGdddesZxGdddewZyGdddepZzGddFdFepZ{Gddde{Z|GddJdJe|Z}Gdd>d>epZ~Gdd<dddeZejjjejjjejjjej_e1
rXeed?ejeed@ejeedAejeedBejeedCejeedDejeejdEejjeejdFejjeejdGejjeedHejeedIejeedJejedKkrexdLZexdMZe{eSeVdNZeedOddPeZeeedQZdReBZeedOddPeZeeedSZedTedQeedSZedUejdVejdVejdWddlZejeejejdXdS(aa

pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and
executing simple grammars, vs. the traditional lex/yacc approach, or the
use of regular expressions.  With pyparsing, you don't need to learn
a new syntax for defining grammars or matching expressions - the parsing
module provides a library of classes that you use to construct the
grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form
``", !"``), built up using :class:`Word`,
:class:`Literal`, and :class:`And` elements
(the :class:`'+'` operators create :class:`And` expressions,
and the strings are auto-converted to :class:`Literal` expressions)::

    from pip._vendor.pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the
self-explanatory class names, and the use of '+', '|' and '^' operators.

The :class:`ParseResults` object returned from
:class:`ParserElement.parseString` can be
accessed as a nested list, a dictionary, or an object with named
attributes.

The pyparsing module handles some of the problems that are typically
vexing when writing text parsers:

  - extra or missing whitespace (the above program will also handle
    "Hello,World!", "Hello  ,  World  !", etc.)
  - quoted strings
  - embedded comments


Getting Started -
-----------------
Visit the classes :class:`ParserElement` and :class:`ParseResults` to
see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:

 - construct literal match expressions from :class:`Literal` and
   :class:`CaselessLiteral` classes
 - construct character word-group expressions using the :class:`Word`
   class
 - see how to create repetitive expressions using :class:`ZeroOrMore`
   and :class:`OneOrMore` classes
 - use :class:`'+'`, :class:`'|'`, :class:`'^'`,
   and :class:`'&'` operators to combine simple expressions into
   more complex ones
 - associate names with your parsed results using
   :class:`ParserElement.setResultsName`
 - access the parsed data, which is returned as a :class:`ParseResults`
   object
 - find some helpful expression short-cuts like :class:`delimitedList`
   and :class:`oneOf`
 - find more useful common expressions in the :class:`pyparsing_common`
   namespace class
z2.4.2z29 Jul 2019 02:58 UTCz*Paul McGuire N)ref)datetime)
itemgetter)wraps)filterfalse)ifilterfalse)RLock)Iterable)MutableMappingMapping)OrderedDict)SimpleNamespacec@seZdZdS)r
N)__name__
__module____qualname__rr9/usr/lib/python3.8/site-packages/pip/_vendor/pyparsing.pyr
sr
aA
    A cross-version compatibility configuration for pyparsing features that will be
    released in a future version. By setting values in this configuration to True,
    those features can be enabled in prior versions for compatibility development
    and testing.

     - collect_all_And_tokens - flag to enable fix for Issue #63 that fixes erroneous grouping
       of results names when an And expression is nested within an Or or MatchFirst; set to
       True to enable bugfix released in pyparsing 2.3.0, or False to preserve
       pre-2.3.0 handling of named results
Ta
Diagnostic configuration (all default to False)
     - warn_multiple_tokens_in_named_alternation - flag to enable warnings when a results
       name is defined on a MatchFirst or Or expression with one or more And subexpressions
       (only warns if __compat__.collect_all_And_tokens is False)
     - warn_ungrouped_named_tokens_in_collection - flag to enable warnings when a results
       name is defined on a containing expression with ungrouped subexpressions that also
       have results names
     - warn_name_set_on_empty_Forward - flag to enable warnings whan a Forward is defined
       with a results name, but has no contents defined
     - warn_on_multiple_string_args_to_oneof - flag to enable warnings whan oneOf is
       incorrectly called with multiple str arguments
     - enable_debug_on_named_expressions - flag to auto-enable debug on all subsequent
       calls to ParserElement.setName()
F__version____versionTime__
__author__
__compat____diag__AndCaselessKeywordCaselessLiteral
CharsNotInCombineDictEachEmpty
FollowedByForward
GoToColumnGroupKeywordLineEnd	LineStartLiteral
PrecededBy
MatchFirstNoMatchNotAny	OneOrMoreOnlyOnceOptionalOrParseBaseExceptionParseElementEnhanceParseExceptionParseExpressionParseFatalExceptionParseResultsParseSyntaxException
ParserElementQuotedStringRecursiveGrammarExceptionRegexSkipTo	StringEndStringStartSuppressTokenTokenConverterWhiteWordWordEnd	WordStart
ZeroOrMoreChar	alphanumsalphas
alphas8bitanyCloseTag
anyOpenTag
cStyleCommentcolcommaSeparatedListcommonHTMLEntitycountedArraycppStyleCommentdblQuotedStringdblSlashComment
delimitedListdictOfdowncaseTokensemptyhexnumshtmlCommentjavaStyleCommentlinelineEnd	lineStartlinenomakeHTMLTagsmakeXMLTagsmatchOnlyAtColmatchPreviousExprmatchPreviousLiteral
nestedExprnullDebugActionnumsoneOfopAssocoperatorPrecedence
printablespunc8bitpythonStyleCommentquotedStringremoveQuotesreplaceHTMLEntityreplaceWith
restOfLinesglQuotedStringsrange	stringEndstringStarttraceParseAction
unicodeStringupcaseTokens
withAttribute
indentedBlockoriginalTextForungroup
infixNotationlocatedExpr	withClass
CloseMatchtokenMappyparsing_commonpyparsing_unicodeunicode_setconditionAsParseActioncCsft|tr|Sz
t|WStk
r`t|td}td}|dd|	|YSXdS)aDrop-in replacement for str(obj) that tries to be Unicode
        friendly. It first tries str(obj). If that fails with
        a UnicodeEncodeError, then it tries unicode(obj). It then
        < returns the unicode object | encodes it with the default
        encoding | ... >.
        xmlcharrefreplacez&#\d+;cSs$dtt|dddddS)Nz\ur)hexinttrrrz_ustr..N)

isinstanceunicodestrUnicodeEncodeErrorencodesysgetdefaultencodingr:setParseActiontransformString)objretZ
xmlcharrefrrr_ustrs

rz6sum len sorted reversed list tuple set any all min maxccs|]
}|VqdSNr).0yrrr	srcCs:d}dddD}t||D]\}}|||}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)&;Nr)rsrrrrsz_xml_escape..zamp gt lt quot apos)splitzipreplace)dataZfrom_symbolsZ
to_symbolsZfrom_Zto_rrr_xml_escapes
r
0123456789ZABCDEFabcdef\ccs|]}|tjkr|VqdSr)stringZ
whitespacercrrrrs
cs@|dk	r|nd|rtntttfdd}|S)Nzfailed user-defined conditioncs t|||s||dSr)boolrlrexc_typefnmsgrrpasz"conditionAsParseAction..pa)r4r2_trim_arityr)rmessagefatalrrrrrsc@sPeZdZdZdddZeddZdd	Zd
dZdd
Z	dddZ
ddZdS)r0z7base exception class for all parsing runtime exceptionsrNcCs>||_|dkr||_d|_n||_||_||_|||f|_dSNr)locrpstr
parserElementargs)selfrrrelemrrr__init__(szParseBaseException.__init__cCs||j|j|j|jS)z
        internal factory method to simplify creating one type of ParseException
        from another - avoids having __init__ signature conflicts among subclasses
        )rrrr)clsperrr_from_exception3sz"ParseBaseException._from_exceptioncCsN|dkrt|j|jS|dkr,t|j|jS|dkrBt|j|jSt|dS)zsupported attributes by name are:
           - lineno - returns the line number of the exception text
           - col - returns the column number of the exception text
           - line - returns the line containing the exception text
        r^)rMcolumnr[N)r^rrrMr[AttributeError)rZanamerrr__getattr__;szParseBaseException.__getattr__cCs^|jr@|jt|jkrd}qDd|j|j|jddd}nd}d|j||j|j|jfS)Nz, found end of textz
, found %rrz\\\rz%%s%s  (at char %d), (line:%d, col:%d))rrlenrrr^r)rZfoundstrrrr__str__Js$zParseBaseException.__str__cCst|Srrrrrr__repr__TszParseBaseException.__repr__>!z{0}
)inspectrgetrecursionlimitrr0appendr[rMformatrrZgetinnerframes
__traceback__set	enumeratef_localsgetr7f_codeco_nameaddrr)excdepthrrZcallersseeniZffZfrmZf_selfZ	self_typecoderrrexplainzsL



zParseException.explainN)r)rrrrstaticmethodrrrrrr2csc@seZdZdZdS)r4znuser-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediatelyNrrrrrrrrr4sc@seZdZdZdS)r6zjust like :class:`ParseFatalException`, but thrown internally
    when an :class:`ErrorStop` ('-' operator) indicates
    that parsing is to stop immediately because an unbacktrackable
    syntax error has been found.
    Nrrrrrr6sc@s eZdZdZddZddZdS)r9ziexception thrown by :class:`ParserElement.validate` if the
    grammar could be improperly recursive
    cCs
||_dSrZparseElementTracerparseElementListrrrrsz"RecursiveGrammarException.__init__cCs
d|jS)NzRecursiveGrammarException: %srrrrrrsz!RecursiveGrammarException.__str__N)rrrrrrrrrrr9sc@s,eZdZddZddZddZddZd	S)
_ParseResultsWithOffsetcCs||f|_dSrtup)rZp1Zp2rrrrsz _ParseResultsWithOffset.__init__cCs
|j|Srrrrrrr__getitem__sz#_ParseResultsWithOffset.__getitem__cCst|jdSNr)reprrrrrrrsz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSrrrrrr	setOffsetsz!_ParseResultsWithOffset.setOffsetN)rrrrrrrrrrrrsrc@seZdZdZd]ddZddddefddZdd	Zefd
dZdd
Z	ddZ
ddZddZeZ
ddZddZddZddZddZereZeZeZn$eZeZeZddZd d!Zd"d#Zd$d%Zd&d'Zd^d(d)Zd*d+Zd,d-Zd.d/Zd0d1Z d2d3Z!d4d5Z"d6d7Z#d8d9Z$d:d;Z%d`` - see :class:`ParserElement.setResultsName`)

    Example::

        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/'
                        + integer.setResultsName("month") + '/'
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)

    prints::

        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    NTcCs"t||r|St|}d|_|SNT)robject__new___ParseResults__doinit)rtoklistnameasListmodalZretobjrrrrs


zParseResults.__new__c
Csd|jrvd|_d|_d|_i|_||_||_|dkr6g}||trP|dd|_n||trft||_n|g|_t	|_
|dk	r`|r`|sd|j|<||trt|}||_||t
dttfr|ddgfks`||tr|g}|r*||trtt|jd||<ntt|dd||<|||_n6z|d||<Wn$tttfk
r^|||<YnXdS)NFrr)r_ParseResults__name_ParseResults__parent_ParseResults__accumNames_ParseResults__asList_ParseResults__modallist_ParseResults__toklist_generatorTypedict_ParseResults__tokdictrrr
basestringr5rKeyError	TypeError
IndexError)rrrrrrrrrr"sB



$
zParseResults.__init__cCsPt|ttfr|j|S||jkr4|j|ddStdd|j|DSdS)NrrcSsg|]}|dqSrrrvrrr
Psz,ParseResults.__getitem__..)rrslicerrrr5rrrrrIs


zParseResults.__getitem__cCs||tr0|j|t|g|j|<|d}nD||ttfrN||j|<|}n&|j|tt|dg|j|<|}||trt||_	dSr)
rrrrrrrr5wkrefr)rkrrsubrrr__setitem__Rs


"
zParseResults.__setitem__c
Cst|ttfrt|j}|j|=t|trH|dkr:||7}t||d}tt||}||j	
D]>\}}|D]0}t|D]"\}\}}	t||	|	|k||<qqxqln|j	|=dSNrr)
rrrrrrrangeindicesreverseritemsrr)
rrZmylenZremovedroccurrencesjrvaluepositionrrr__delitem___s

zParseResults.__delitem__cCs
||jkSr)r)rrrrr__contains__tszParseResults.__contains__cCs
t|jSr)rrrrrr__len__wszParseResults.__len__cCs
|jSrrrrrr__bool__zszParseResults.__bool__cCs
t|jSriterrrrrr__iter__~szParseResults.__iter__cCst|jdddSNrr+rrrr__reversed__szParseResults.__reversed__cCs$t|jdr|jSt|jSdS)Niterkeys)hasattrrr0r,rrrr	_iterkeyss
zParseResults._iterkeyscsfddDS)Nc3s|]}|VqdSrrrrrrrrsz+ParseResults._itervalues..r2rrrr_itervaluesszParseResults._itervaluescsfddDS)Nc3s|]}||fVqdSrrr3rrrrsz*ParseResults._iteritems..r4rrrr
_iteritemsszParseResults._iteritemscCst|S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)rr0rrrrkeysszParseResults.keyscCst|S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r
itervaluesrrrrvaluesszParseResults.valuescCst|S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r	iteritemsrrrrr!szParseResults.itemscCs
t|jS)zSince keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names.)rrrrrrhaskeysszParseResults.haskeyscOs|s
dg}|D]*\}}|dkr0|d|f}qtd|qt|dtsdt|dksd|d|kr~|d}||}||=|S|d}|SdS)a
        Removes and returns item at specified index (default= ``last``).
        Supports both ``list`` and ``dict`` semantics for ``pop()``. If
        passed no argument or an integer argument, it will use ``list``
        semantics and pop tokens from the list of parsed tokens. If passed
        a non-integer argument (most likely a string), it will use ``dict``
        semantics and pop the corresponding value from any defined results
        names. A second default return value argument is supported, just as in
        ``dict.pop()``.

        Example::

            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())

        prints::

            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        rdefaultrz-pop() got an unexpected keyword argument '%s'rN)r!rrrr)rrkwargsrrindexrZdefaultvaluerrrpops"%

zParseResults.popcCs||kr||S|SdS)a[
        Returns named result matching the given key, or if there is no
        such name, then returns the given ``defaultValue`` or ``None`` if no
        ``defaultValue`` is specified.

        Similar to ``dict.get()``.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        Nr)rkeydefaultValuerrrrszParseResults.getcCsR|j|||jD]4\}}t|D]"\}\}}t||||k||<q(qdS)a
        Inserts new element at location index in the list of parsed tokens.

        Similar to ``list.insert()``.

        Example::

            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        N)rinsertrr!rr)rr>ZinsStrrr"rr$r%rrrrBszParseResults.insertcCs|j|dS)a
        Add single element to end of ParseResults list of elements.

        Example::

            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        N)rr)ritemrrrrs
zParseResults.appendcCs&t|tr||n|j|dS)a	
        Add sequence of elements to end of ParseResults list of elements.

        Example::

            patt = OneOrMore(Word(alphas))

            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        N)rr5__iadd__rextend)rZitemseqrrrrE's
zParseResults.extendcCs|jdd=|jdS)z7
        Clear all elements and results names.
        N)rrclearrrrrrF:szParseResults.clearcCs&z
||WStk
r YdSXdSr)rrrrrrrAs
zParseResults.__getattr__cCs|}||7}|Srcopy)rotherrrrr__add__GszParseResults.__add__cs|jrjt|jfdd|j}fdd|D}|D],\}}|||<t|dtr.c	s4g|],\}}|D]}|t|d|dfqqSrr)rrrvlistr)	addoffsetrrrQsz)ParseResults.__iadd__..r)
rrrr!rr5rrrupdate)rrJZ
otheritemsZotherdictitemsrrr)rQrMrrDLs


zParseResults.__iadd__cCs&t|tr|dkr|S||SdSr)rrrIrrJrrr__radd__\szParseResults.__radd__cCsdt|jt|jfS)Nz(%s, %s))rrrrrrrrdszParseResults.__repr__cCsdddd|jDdS)N[, css(|] }t|trt|nt|VqdSr)rr5rrrrrrrrhsz'ParseResults.__str__..])rrrrrrrgszParseResults.__str__rcCsLg}|jD]<}|r |r ||t|tr8||7}q
|t|q
|Sr)rrrr5
_asStringListr)rsepoutrCrrrrYjs


zParseResults._asStringListcCsdd|jDS)ax
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::

            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # ->  ['sldkj', 'lsdkj', 'sldkj']

            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # ->  ['sldkj', 'lsdkj', 'sldkj']
        cSs"g|]}t|tr|n|qSr)rr5r)rresrrrrsz'ParseResults.asList..r)rrrrruszParseResults.asListcs6tr|j}n|j}fddtfdd|DS)a
        Returns the named parse results as a nested dictionary.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # ->  (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})

            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # ->  {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        cs6t|tr.|r|Sfdd|DSn|SdS)Ncsg|]}|qSrrrtoItemrrrsz7ParseResults.asDict..toItem..)rr5r;asDictrr]rrr^s

z#ParseResults.asDict..toItemc3s|]\}}||fVqdSrrrrrr]rrrsz&ParseResults.asDict..)PY_3r!r:r)rZitem_fnrr]rr_s
	zParseResults.asDictcCs<t|j}t|j|_|j|_|j|j|j|_|S)zG
        Returns a new copy of a :class:`ParseResults` object.
        )	r5rrrr!rrrRrrrrrrrIs
zParseResults.copyFcCsLd}g}tdd|jD}|d}|s8d}d}d}d}	|dk	rJ|}	n|jrV|j}	|	sf|rbdSd}	|||d|	d	g7}t|jD]\}
}t|tr|
|kr||||
|o|dk||g7}n||d|o|dk||g7}qd}|
|kr||
}|s|rqnd}t	t
|}
|||d|d	|
d
|d	g	7}q|||d
|	d	g7}d|S)z
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        rcss(|] \}}|D]}|d|fVqqdSrNrrOrrrrsz%ParseResults.asXML..  rNZITEM<>.z
%s%s- %s: rer)rjfullinclude_list_depthcss|]}t|tVqdSr)rr5)rvvrrrrKsz
%s%s[%d]:
%s%s%s)
rrrr;sortedr!rr5dumpranyrr)rrjrprqrrr[NLr!rrrrsrrrru!sP







zParseResults.dumpcOstj|f||dS)a#
        Pretty-printer for parsed results as a list, using the
        `pprint `_ module.
        Accepts additional positional or keyword args as defined for
        `pprint.pprint `_ .

        Example::

            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)

        prints::

            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        N)pprintrrrr=rrrrxbszParseResults.pprintcCs.|j|j|jdk	r|p d|j|jffSr)rrrIrrrrrrr__getstate__~szParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j||dk	rDt||_nd|_dSr)rrrrrRrr)rstaternZinAccumNamesrrr__setstate__s
zParseResults.__setstate__cCs|j|j|j|jfSr)rrr	r
rrrr__getnewargs__szParseResults.__getnewargs__cCstt|t|Sr)rrrr7rrrrrszParseResults.__dir__cCsrdd}|g}|D]>\}}t|tr>||j||d7}q|||g|||d7}q|dk	rn||g|d}|S)z
        Helper classmethod to construct a ParseResults from a dict, preserving the
        name-value relations as results names. If an optional 'name' argument is
        given, a nested ParseResults will be returned
        cSsHzt|Wntk
r"YdSXtr8t|ttfSt|tSdSNF)r,	Exceptionrbrrbytesrr`rrris_iterablesz+ParseResults.from_dict..is_iterabler)rrN)r!rr	from_dict)rrJrrrrrrrrrs
zParseResults.from_dict)NNTT)N)r)NFrT)rTTr)N)6rrrrrrrrrr&r'r(r*__nonzero__r-r/r2r5r6rbr7r9r!r0r8r:r;r?rrBrrErFrrKrDrTrrrYrr_rIrirlrorurxrzr|r}rrrrrrrr5sl*
	'	
7

$
=(
A
cCsF|}d|krt|kr4nn||ddkr4dS||dd|S)aReturns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See
   :class:`ParserElement.parseString` for more
   information on parsing strings containing ```` s, and suggested
   methods to maintain a consistent view of the parsed string, the parse
   location, and line and column positions within the parsed string.
   rrr)rrfind)rstrgrrrrrMscCs|dd|dS)aReturns current line number within a string, counting newlines as line separators.
    The first line is number 1.

    Note - the default parsing behavior is to expand tabs in the input string
    before starting the parsing process.  See :class:`ParserElement.parseString`
    for more information on parsing strings containing ```` s, and
    suggested methods to maintain a consistent view of the parsed string, the
    parse location, and line and column positions within the parsed string.
    rrr)count)rrrrrr^s
cCsF|dd|}|d|}|dkr2||d|S||ddSdS)zfReturns the line of text containing loc within a string, counting newlines as line separators.
       rrrN)rfind)rrZlastCRZnextCRrrrr[s
cCs8tdt|dt|dt||t||fdS)NzMatch z at loc z(%d,%d))printrr^rM)instringrexprrrr_defaultStartDebugActionsrcCs$tdt|dt|dS)NzMatched z -> )rrrr)rstartlocZendlocrtoksrrr_defaultSuccessDebugActionsrcCstdt|dS)NzException raised:)rr)rrrrrrr_defaultExceptionDebugActionsrcGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nr)rrrrresrcstkrfddSdgdgtdddkrFddd}dd	d
ntj}tjd}|ddd
}|d|d|ffdd}d}ztdtdj}Wntk
rt}YnX||_|S)Ncs|Srrr)funcrrrrz_trim_arity..rFr)rcSs8tdkrdnd}tj||dd|}|ddgS)N)rrrrlimitr)system_version	traceback
extract_stack)rrM
frame_summaryrrrrsz"_trim_arity..extract_stackcSs$tj||d}|d}|ddgS)Nrrr)r
extract_tb)tbrZframesrrrrrsz_trim_arity..extract_tbrrrcsz"|dd}dd<|WStk
rdr>nNz.td}|dddddksjW5z~Wntk
rYnXXdkrdd7<YqYqXqdS)NrTrrrr)r	NameErrorrexc_info)rrrrZ
foundArityrrmaxargsZpa_call_line_synthrrrs&z_trim_arity..wrapperzr	__class__)r)r)	singleArgBuiltinsrrrrgetattrrrr)rrrZ	LINE_DIFFZ	this_liner	func_namerrrrs,

rcseZdZdZdZdZeddZeddZddd	Z	d
dZ
dd
ZdddZdddZ
dddZddZddZddZddZddZdd Zdd!d"Zd#d$Zdd%d&Zd'd(Zd)d*ZGd+d,d,eZed-k	rGd.d/d/eZnGd0d/d/eZiZeZ d1d1gZ!dd2d3Z"eZ#ed4d5Z$dZ%edd7d8Z&dd9d:Z'e(dfd;d<Z)d=d>Z*e(fd?d@Z+e(dfdAdBZ,dCdDZ-dEdFZ.dGdHZ/dIdJZ0dKdLZ1dMdNZ2dOdPZ3dQdRZ4dSdTZ5dUdVZ6dWdXZ7dYdZZ8d[d\Z9d]d^Z:d_d`Z;ddadbZdgdhZ?didjZ@dkdlZAdmdnZBddodpZCdqdrZDdsdtZEdudvZFdwdxZGddydzZHdd{d|ZIfd}d~ZJddZKddZLddZMddZNdddZOdddZPZQS)r7z)Abstract base level parser element class.z 
	
FcCs
|t_dS)a
        Overrides the default whitespace chars

        Example::

            # default whitespace chars are space,  and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']

            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        N)r7DEFAULT_WHITE_CHARScharsrrrsetDefaultWhitespaceCharsDsz'ParserElement.setDefaultWhitespaceCharscCs
|t_dS)ah
        Set class to be used for inclusion of string literals into a parser.

        Example::

            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        N)r7_literalStringClassrrrrinlineLiteralsUsingTsz!ParserElement.inlineLiteralsUsingcCst|_d|_d|_d|_||_d|_ttj	|_
d|_d|_d|_
t|_d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr)NNN)rparseAction
failActionstrReprresultsName
saveAsListskipWhitespacerr7r
whiteCharscopyDefaultWhiteCharsmayReturnEmptykeepTabsignoreExprsdebugstreamlined
mayIndexErrorerrmsgmodalResultsdebugActionsrecallPreparse
callDuringTry)rsavelistrrrrjs(zParserElement.__init__cCs<t|}|jdd|_|jdd|_|jr8tj|_|S)a/
        Make a copy of this :class:`ParserElement`.  Useful for defining
        different parse actions for the same parsing pattern, using copies of
        the original parse element.

        Example::

            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0] * 1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M")

            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))

        prints::

            [5120, 100, 655360, 268435456]

        Equivalent form of ``expr.copy()`` is just ``expr()``::

            integerM = integer().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M")
        N)rIrrrr7rr)rZcpyrrrrIs
zParserElement.copycCs$||_d|j|_tjr ||S)a_
        Define name for this expression, makes debugging and exception messages clearer.

        Example::

            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        	Expected )rrr!enable_debug_on_named_expressionssetDebugrGrrrsetNames
	zParserElement.setNamecCs|||S)aO
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original :class:`ParserElement` object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        ``expr("name")`` in place of ``expr.setResultsName("name")``
        - see :class:`__call__`.

        Example::

            date_str = (integer.setResultsName("year") + '/'
                        + integer.setResultsName("month") + '/'
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        )_setResultsNamerrlistAllMatchesrrrsetResultsNameszParserElement.setResultsNamecCs4|}|dr"|dd}d}||_||_|S)N*rT)rIendswithrr)rrrZnewselfrrrrs
zParserElement._setResultsNameTcs@|r&|jdfdd	}|_||_nt|jdr<|jj|_|S)zMethod to invoke the Python pdb debugger when this element is
           about to be parsed. Set ``breakFlag`` to True to enable, False to
           disable.
        Tcsddl}|||||Sr)pdbZ	set_trace)rr	doActionscallPreParserZ_parseMethodrrbreakersz'ParserElement.setBreak..breaker_originalParseMethod)TT)_parserr1)rZ	breakFlagrrrrsetBreaks
zParserElement.setBreakcOsVt|dgkrg|_n`` s, and suggested
        methods to maintain a consistent view of the parsed string, the parse
        location, and line and column positions within the parsed string.

        Example::

            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        Ncss|]}t|VqdSr)callable)rrrrrrsz/ParserElement.setParseAction..zparse actions must be callablerF)rrallrmaprrrrfnsr=rrrrs(zParserElement.setParseActioncOs4|jtttt|7_|jp,|dd|_|S)z
        Add one or more parse actions to expression's list of parse actions. See :class:`setParseAction`.

        See examples in :class:`copy`.
        rF)rrrrrrrrrraddParseActionszParserElement.addParseActionc
OsF|D](}|jt||d|dddq|jp>|dd|_|S)aAdd a boolean predicate function to expression's list of parse actions. See
        :class:`setParseAction` for function call signatures. Unlike ``setParseAction``,
        functions passed to ``addCondition`` need to return boolean success/fail of the condition.

        Optional keyword arguments:
        - message = define a custom message to be used in the raised exception
        - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException

        Example::

            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        rrF)rrr)rrrrr)rrr=rrrraddConditions

zParserElement.addConditioncCs
||_|S)aDefine action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           ``fn(s, loc, expr, err)`` where:
           - s = string being parsed
           - loc = location where expression match was attempted and failed
           - expr = the parse expression that failed
           - err = the exception thrown
           The function returns no value.  It may throw :class:`ParseFatalException`
           if it is desired to stop parsing immediately.)rrrrrr
setFailAction4s
zParserElement.setFailActionc	CsNd}|rJd}|jD]4}z|||\}}d}qWqtk
rDYqXqq|SNTF)rrr2)rrrZ
exprsFoundeZdummyrrr_skipIgnorablesAs


zParserElement._skipIgnorablescCsH|jr|||}|jrD|j}t|}||krD|||krD|d7}q&|SNr)rrrrr)rrrZwtinstrlenrrrpreParseNs
zParserElement.preParsecCs|gfSrrrrrrrrrrZszParserElement.parseImplcCs|Srrrrr	tokenlistrrr	postParse]szParserElement.postParsecCsd\}}}|j}|s|jr"|j|r8|j||||z|rR|jrR|||}	n|}	|	}
|jsl|	t|krz|||	|\}}Wqtk
rt	|t||j
|YqXn|||	|\}}Wn\tk
r}z<|j|r|j|||
|||jr|||
||W5d}~XYnXn|r>|jr>|||}	n|}	|	}
|js\|	t|krz|||	|\}}Wn*tk
rt	|t||j
|YnXn|||	|\}}||||}t
||j|j|jd}
|jr`|s|jr`|rz|jD]}z|||
|
}Wn6tk
rD}zt	d}||_|W5d}~XYnX|dk	r||
k	rt
||j|jovt|t
tf|jd}
qWnFtk
r}z&|j|r|j|||
||W5d}~XYnXn|jD]}z|||
|
}Wn6tk
r }zt	d}||_|W5d}~XYnX|dk	r||
k	rt
||j|joRt|t
tf|jd}
q|r|j|r|j|||
|||
||
fS)N)rrr)rrz exception raised in parse action)rrrrrrrrrr2rrrr5rrrrr	__cause__rr)rrrrrZTRYZMATCHZFAILZ	debuggingprelocZtokensStarttokenserrZ	retTokensrZparse_action_excrrrrras






zParserElement._parseNoCachecCs@z|j||dddWStk
r:t|||j|YnXdS)NFrr)rr4r2rrrrrrrtryParseszParserElement.tryParsec	Cs4z|||Wnttfk
r*YdSXdSdS)NFT)rr2rrrrrcanParseNexts
zParserElement.canParseNextc@seZdZddZdS)zParserElement._UnboundedCachecs~it|_fdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS)	Ncs|Srrrr@cachenot_in_cacherrrsz3ParserElement._UnboundedCache.__init__..getcs||<dSrrrr@r$rrrrsz3ParserElement._UnboundedCache.__init__..setcsdSrrFrrrrrFsz5ParserElement._UnboundedCache.__init__..clearcstSrrrrrr	cache_lensz9ParserElement._UnboundedCache.__init__..cache_len)rrtypes
MethodTyperrrFr()rrrrFrrrrrsz&ParserElement._UnboundedCache.__init__Nrrrrrrrr_UnboundedCachesrNc@seZdZddZdS)ParserElement._FifoCachecst|_tfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS)	Ncs|Srrrrrrrs.ParserElement._FifoCache.__init__..getcs>||<tkr:zdWqtk
r6YqXqdSr~)rpopitemrr)rsizerrrs.ParserElement._FifoCache.__init__..setcsdSrrrrrrrFs0ParserElement._FifoCache.__init__..clearcstSrrrrrrrs4ParserElement._FifoCache.__init__..cache_len)	rr_OrderedDictrrrrrFr(rrrrrFrr)rrrrrs!ParserElement._FifoCache.__init__Nrrrrr
_FifoCachesrc@seZdZddZdS)rcst|_itgfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_	dS)	Ncs|Srrrrrrrsrcs4||<tkr&dq|dSr)rr?popleftrr)rkey_fiforrrrsrcsdSrrr)rr	rrrF
srcstSrrrrrrrsr)
rrcollectionsdequerrrrrFr(rr)rr	rrrrsrNrrrrrrsrcCsd\}}|||||f}tjtj}||}	|	|jkrtj|d7<z|||||}	Wn8tk
r}
z|||
j	|
j
W5d}
~
XYn.X|||	d|	df|	W5QRSn@tj|d7<t|	t
r|	|	d|	dfW5QRSW5QRXdS)NrNrr)r7packrat_cache_lock
packrat_cacherrpackrat_cache_statsrr0rrrrIrr)rrrrrZHITZMISSlookuprr$rrrr_parseCaches$


zParserElement._parseCachecCs(tjdgttjtjdd<dSr)r7r
rFrrrrrr
resetCache6s
zParserElement.resetCachecCs8tjs4dt_|dkr tt_nt|t_tjt_dS)aEnables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.

           Parameters:

           - cache_size_limit - (default= ``128``) - if an integer value is provided
             will limit the size of the packrat cache; if None is passed, then
             the cache size will be unbounded; if 0 is passed, the cache will
             be effectively disabled.

           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method :class:`ParserElement.enablePackrat`.
           For best results, call ``enablePackrat()`` immediately after
           importing pyparsing.

           Example::

               from pip._vendor import pyparsing
               pyparsing.ParserElement.enablePackrat()
        TN)r7_packratEnabledrr
rrr)Zcache_size_limitrrr
enablePackrat<szParserElement.enablePackratc
Cst|js||jD]}|q|js8|}z<||d\}}|rr|||}t	t
}|||Wn0tk
r}ztjrn|W5d}~XYnX|SdS)a
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        Returns the parsed data as a :class:`ParseResults` object, which may be
        accessed as a list, or as a dict or object with attributes if the given parser
        includes results names.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set ``parseAll`` to True (equivalent to ending
        the grammar with ``StringEnd()``).

        Note: ``parseString`` implicitly calls ``expandtabs()`` on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the ``loc`` argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:

        - calling ``parseWithTabs`` on your grammar before calling ``parseString``
          (see :class:`parseWithTabs`)
        - define your parse action using the full ``(s, loc, toks)`` signature, and
          reference the input string using the parse action's ``s`` argument
        - explictly expand the tabs in your input string before calling
          ``parseString``

        Example::

            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        rN)
r7rr
streamlinerr
expandtabsrrrr<r0verbose_stacktrace)rrparseAllrrrZserrrrparseString_s$!

zParserElement.parseStringc
cs6|js||jD]}|q|js4t|}t|}d}|j}|j}t	
d}	z||kr|	|krz |||}
|||
dd\}}Wntk
r|
d}YqZX||kr|	d7}	||
|fV|r|||}
|
|kr|}q|d7}q|}qZ|
d}qZWn4tk
r0}zt	j
rn|W5d}~XYnXdS)aq
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        ``maxMatches`` argument, to clip scanning after 'n' matches are found.  If
        ``overlap`` is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See :class:`parseString` for more information on parsing
        strings with embedded tabs.

        Example::

            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens, start, end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])

        prints::

            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        rFrrN)rrrrrrrrrr7rr2r0r)rr
maxMatchesZoverlaprrrZ
preparseFnZparseFnmatchesrZnextLocrZnextlocrrrr
scanStringsB




zParserElement.scanStringc
Csg}d}d|_z||D]Z\}}}|||||rpt|trR||7}nt|trf||7}n
|||}q|||ddd|D}dtt	t
|WStk
r}ztj
rƂn|W5d}~XYnXdS)a[
        Extension to :class:`scanString`, to modify matching text with modified tokens that may
        be returned from a parse action.  To use ``transformString``, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking ``transformString()`` on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  ``transformString()`` returns the resulting transformed string.

        Example::

            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())

            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))

        prints::

            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        rTNcSsg|]}|r|qSrrrorrrrsz1ParserElement.transformString..r)rrrrr5rrrrr_flattenr0r7r)rrr[ZlastErrrrrrrrs(



zParserElement.transformStringc
CsRztdd|||DWStk
rL}ztjr8n|W5d}~XYnXdS)a
        Another extension to :class:`scanString`, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        ``maxMatches`` argument, to clip searching after 'n' matches are found.

        Example::

            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())

            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))

        prints::

            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        cSsg|]\}}}|qSrr)rrrrrrrr#sz.ParserElement.searchString..N)r5rr0r7r)rrrrrrrsearchString
szParserElement.searchStringc	csTd}d}|j||dD]*\}}}|||V|r<|dV|}q||dVdS)aR
        Generator method to split a string using the given expression as a separator.
        May be called with optional ``maxsplit`` argument, to limit the number of splits;
        and the optional ``includeSeparators`` argument (default= ``False``), if the separating
        matching text should be included in the split results.

        Example::

            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))

        prints::

            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        r)rN)r)	rrmaxsplitZincludeSeparatorsZsplitsZlastrrrrrrr+s
zParserElement.splitcCsV|tkrt|St|tr$||}t|tsJtjdt|t	dddSt
||gS)a[
        Implementation of + operator - returns :class:`And`. Adding strings to a ParserElement
        converts them to :class:`Literal`s by default.

        Example::

            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))

        prints::

            Hello, World! -> ['Hello', ',', 'World', '!']

        ``...`` may be used as a parse expression as a short form of :class:`SkipTo`.

            Literal('start') + ... + Literal('end')

        is equivalent to:

            Literal('start') + SkipTo('end')("_skipped*") + Literal('end')

        Note that the skipped text is returned with '_skipped' as a results name,
        and to support having multiple skips in the same parser, the value returned is
        a list of all skipped text.
        4Cannot combine element of type %s with ParserElementr
stacklevelN)Ellipsis_PendingSkiprrrr7warningswarnr
SyntaxWarningrrSrrrrKDs


zParserElement.__add__cCsZ|tkrt|d|St|tr,||}t|tsRtjdt|t	dddS||S)z`
        Implementation of + operator when left operand is not a :class:`ParserElement`
        	_skipped*r#rr$N)
r&r;rrrr7r(r)rr*rSrrrrTjs


zParserElement.__radd__cCsJt|tr||}t|ts:tjdt|tdddS|t	|S)zT
        Implementation of - operator, returns :class:`And` with error stop
        r#rr$N)
rrrr7r(r)rr*r
_ErrorStoprSrrr__sub__ys


zParserElement.__sub__cCsBt|tr||}t|ts:tjdt|tdddS||S)z`
        Implementation of - operator when left operand is not a :class:`ParserElement`
        r#rr$Nrrrr7r(r)rr*rSrrr__rsub__s


zParserElement.__rsub__cs|tkrd}n8t|trF|ddtfkrFd|ddddd}t|tr\|d}}nt|trJtdd	|D}|d
dd}|ddkrd|df}t|dtr|ddkr|ddkrtS|ddkrtS|dtSnNt|dtr,t|dtr,|\}}||8}ntdt|dt|dntdt||dkrjtd
|dkr|td||krdkrnntd|rfdd|r|dkr҈|}nt	g||}n|}n|dkr}nt	g|}|S)a
        Implementation of * operator, allows use of ``expr * 3`` in place of
        ``expr + expr + expr``.  Expressions may also me multiplied by a 2-integer
        tuple, similar to ``{min, max}`` multipliers in regular expressions.  Tuples
        may also include ``None`` as in:
         - ``expr*(n, None)`` or ``expr*(n, )`` is equivalent
              to ``expr*n + ZeroOrMore(expr)``
              (read as "at least n instances of ``expr``")
         - ``expr*(None, n)`` is equivalent to ``expr*(0, n)``
              (read as "0 to n instances of ``expr``")
         - ``expr*(None, None)`` is equivalent to ``ZeroOrMore(expr)``
         - ``expr*(1, None)`` is equivalent to ``OneOrMore(expr)``

        Note that ``expr*(None, n)`` does not raise an exception if
        more than n exprs exist in the input stream; that is,
        ``expr*(None, n)`` does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        ``expr*(None, n) + ~expr``
        )rNNrrrrrcss|]}|tk	r|ndVqdSr)r&rrrrrsz(ParserElement.__mul__..)NNz8cannot multiply 'ParserElement' and ('%s', '%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez,cannot multiply ParserElement by 0 or (0, 0)cs(|dkrt|dStSdSr)r.nmakeOptionalListrrrr3sz/ParserElement.__mul__..makeOptionalList)
r&rtuplerrEr,rr
ValueErrorr)rrJZminElementsZoptElementsrrr2r__mul__sN
 





zParserElement.__mul__cCs
||Sr)r6rSrrr__rmul__szParserElement.__rmul__cCsZ|tkrt|ddSt|tr(||}t|tsNtjdt|t	dddSt
||gS)zL
        Implementation of | operator - returns :class:`MatchFirst`
        T)	must_skipr#rr$N)r&r'rrrr7r(r)rr*r)rSrrr__or__s


zParserElement.__or__cCsBt|tr||}t|ts:tjdt|tdddS||BS)z`
        Implementation of | operator when left operand is not a :class:`ParserElement`
        r#rr$Nr.rSrrr__ror__s


zParserElement.__ror__cCsFt|tr||}t|ts:tjdt|tdddSt||gS)zD
        Implementation of ^ operator - returns :class:`Or`
        r#rr$N)	rrrr7r(r)rr*r/rSrrr__xor__s


zParserElement.__xor__cCsBt|tr||}t|ts:tjdt|tdddS||AS)z`
        Implementation of ^ operator when left operand is not a :class:`ParserElement`
        r#rr$Nr.rSrrr__rxor__	s


zParserElement.__rxor__cCsFt|tr||}t|ts:tjdt|tdddSt||gS)zF
        Implementation of & operator - returns :class:`Each`
        r#rr$N)	rrrr7r(r)rr*rrSrrr__and__	s


zParserElement.__and__cCsBt|tr||}t|ts:tjdt|tdddS||@S)z`
        Implementation of & operator when left operand is not a :class:`ParserElement`
        r#rr$Nr.rSrrr__rand__	s


zParserElement.__rand__cCst|S)zH
        Implementation of ~ operator - returns :class:`NotAny`
        )r+rrrr
__invert__)	szParserElement.__invert__cCstd|jjdS)Nz%r object is not iterable)rrrrrrrr-/	szParserElement.__iter__c	Cszt|tr|f}t|Wntk
r8||f}YnXt|dkrztd|ddt|dkrrdt|nd|t|dd}|S)a
        use ``[]`` indexing notation as a short form for expression repetition:
         - ``expr[n]`` is equivalent to ``expr*n``
         - ``expr[m, n]`` is equivalent to ``expr*(m, n)``
         - ``expr[n, ...]`` or ``expr[n,]`` is equivalent
              to ``expr*n + ZeroOrMore(expr)``
              (read as "at least n instances of ``expr``")
         - ``expr[..., n]`` is equivalent to ``expr*(0, n)``
              (read as "0 to n instances of ``expr``")
         - ``expr[...]`` and ``expr[0, ...]`` are equivalent to ``ZeroOrMore(expr)``
         - ``expr[1, ...]`` is equivalent to ``OneOrMore(expr)``
         ``None`` may be used in place of ``...``.

        Note that ``expr[..., n]`` and ``expr[m, n]``do not raise an exception
        if more than ``n`` ``expr``s exist in the input stream.  If this behavior is
        desired, then write ``expr[..., n] + ~expr``.
       rz.only 1 or 2 index arguments supported ({0}{1})Nrz	... [{0}]r)	rrr,rrr(r)rr4)rr@rrrrr4	s

zParserElement.__getitem__cCs|dk	r||S|SdS)a
        Shortcut for :class:`setResultsName`, with ``listAllMatches=False``.

        If ``name`` is given with a trailing ``'*'`` character, then ``listAllMatches`` will be
        passed as ``True``.

        If ``name` is omitted, same as calling :class:`copy`.

        Example::

            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums + "-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums + "-")("socsecno")
        N)rrIrGrrr__call__X	s
zParserElement.__call__cCst|S)z
        Suppresses the output of this :class:`ParserElement`; useful to keep punctuation from
        cluttering up returned output.
        )r>rrrrsuppressl	szParserElement.suppresscCs
d|_|S)a
        Disables the skipping of whitespace before matching the characters in the
        :class:`ParserElement`'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        FrrrrrleaveWhitespaces	szParserElement.leaveWhitespacecCsd|_||_d|_|S)z8
        Overrides the default whitespace chars
        TF)rrr)rrrrrsetWhitespaceChars|	sz ParserElement.setWhitespaceCharscCs
d|_|S)z
        Overrides default behavior to expand ````s to spaces before parsing the input string.
        Must be called before ``parseString`` when the input grammar contains elements that
        match ```` characters.
        T)rrrrr
parseWithTabs	szParserElement.parseWithTabscCsLt|trt|}t|tr4||jkrH|j|n|jt||S)a
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.

        Example::

            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']

            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        )rrr>rrrIrSrrrignore	s


zParserElement.ignorecCs"|pt|pt|ptf|_d|_|S)zT
        Enable display of debugging messages while doing pattern matching.
        T)rrrrr)rZstartActionZ
successActionZexceptionActionrrrsetDebugActions	szParserElement.setDebugActionscCs|r|tttnd|_|S)a
        Enable display of debugging messages while doing pattern matching.
        Set ``flag`` to True to enable, False to disable.

        Example::

            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer

            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")

        prints::

            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using :class:`setDebugActions`. Prior to attempting
        to match the ``wd`` expression, the debugging message ``"Match  at loc (,)"``
        is shown. Then if the parse succeeds, a ``"Matched"`` message is shown, or an ``"Exception raised"``
        message is shown. Also note the use of :class:`setName` to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the :class:`Word` expression without calling ``setName`` is ``"W:(ABCD...)"``.
        F)rGrrrr)rflagrrrr	s%zParserElement.setDebugcCs|jSrrrrrrr	szParserElement.__str__cCst|Srrrrrrr	szParserElement.__repr__cCsd|_d|_|Sr)rrrrrrr	szParserElement.streamlinecCsdSrrrrrrcheckRecursion	szParserElement.checkRecursioncCs|gdS)zj
        Check defined expressions for valid structure, check for infinite recursive definitions.
        N)rI)r
validateTracerrrvalidate	szParserElement.validatecCsz|}Wn2tk
r>t|d}|}W5QRXYnXz|||WStk
r~}ztjrjn|W5d}~XYnXdS)z
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        rN)readropenrr0r7r)rZfile_or_filenamerZ
file_contentsfrrrr	parseFile	szParserElement.parseFilecsht|trBtr(||kp$tt||qd||kp>t|t|kSn"t|trV||Stt||kSdSr)rr7rbsuper__eq__varsrrrSrrrrR
s


zParserElement.__eq__cCs
||kSrrrSrrr__ne__
szParserElement.__ne__cCst|Sr)idrrrr__hash__
szParserElement.__hash__cCs||kSrrrSrrr__req__
szParserElement.__req__cCs
||kSrrrSrrr__rne__
szParserElement.__rne__cCs4z|jt||dWdStk
r.YdSXdS)a
        Method for quick testing of a parser against a test string. Good for simple
        inline microtests of sub expressions while building up larger parser.

        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests

        Example::

            expr = Word(nums)
            assert expr.matches("100")
        rTFN)rrr0)rZ
testStringrrrrr
s
zParserElement.matches#c	Cst|tr"tttj|}t|tr4t|}|dkrBt	j
}|j}	g}
g}d}tdt
dt}
d}|D]\}|dk	r||ds|r|s||qt|sqtd||g}g}z"|
||}|j||d}Wntk
r}zt|trdnd	}d|krB|t|j||d
t|j|dd|n|d
|jd||d
t||ot|}|}W5d}~XYntk
r}z$|dt||o|}|}W5d}~XYnX|o|}|dk	rzR|||}|dk	r(t|tr||n|t|n||WnRtk
r}z2||j|d|d|jt |j|W5d}~XYnXn||j|d|r|r|d	|	d||
||fqt||
fS)as

        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.

        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests
         - comment - (default= ``'#'``) - expression for indicating embedded comments in the test
              string; pass None to disable comment filtering
         - fullDump - (default= ``True``) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default= ``True``) prints test output to stdout
         - failureTests - (default= ``False``) indicates if these tests are expected to fail parsing
         - postParse - (default= ``None``) optional callback for successful parse results; called as
              `fn(test_string, parse_results)` and returns a string to be added to the test output
         - file - (default=``None``) optional file-like object to which test output will be written;
              if None, will default to ``sys.stdout``

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if ``failureTests`` is True), and the results contain a list of lines of each
        test's output

        Example::

            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")

        prints::

            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success

            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\n of strings that spans \n 3 lines")

        (Note that this is a raw string literal, you must include the leading 'r'.)
        NT\nruFrZz(FATAL)rrrrzFAIL: zFAIL-EXCEPTION: )rpz{0} failed: {1}: {2})!rrrrrrrstrip
splitlinesr'rstdoutwriterrprFrmrrrrlstriprr0r4r[rrMrr5rurrr)rZtestsrZcommentZfullDumpZprintResultsZfailureTestsrfileprint_Z
allResultsZcommentssuccessrwBOMrr[resultrrrZpp_valuerrrrrunTests.
sn`




$




0
zParserElement.runTests)F)F)F)T)T)TT)TT)r)F)N)T)N)F)T)Tr[TTFNN)RrrrrrrrrrrrIrrrrrrrrrrrrrrrrrrrr
rrrrrrrrr_MAX_INTrrr!rrKrTr-r/r6r7r9r:r;r<r=r>r?r-rr@rArCrDrErFrGrrrrrIrKrPrRrUrWrXrYrrg
__classcell__rrrTrr7?s




	
1




W

"
8H.&J$
			

+


cs6eZdZd
fdd	ZddZddZdd	ZZS)r'Fcs>tt|t|tdd|_|j|_||_||_	dS)Nr...)
rQr'rrrrrranchorr8)rrr8rTrrr
s
z_PendingSkip.__init__cs\t|dd}jrNdd}fdd}j||||B|Sj||S)Nrjr+cSs,|jr|jdgkr(|d=|dddS)Nrr_skipped)rlrr?rrrrr8
sz'_PendingSkip.__add__..must_skipcs<|jdddgkr8|d}dtjd|d<dS)Nrrrlz	missing .show_skip)r;rr8rkr)rrJZskipperr8rmrrrrK
s
z_PendingSkip.__add__cCs|jSr)rrrrrr
sz_PendingSkip.__repr__cGstddS)NzBuse of `...` expression without following SkipTo target expression)r)rrrrrr
sz_PendingSkip.parseImpl)F)rrrrrKrrrirrrTrr'
sr'cs eZdZdZfddZZS)r?zYAbstract :class:`ParserElement` subclass, for defining atomic
    matching patterns.
    cstt|jdddSNFr)rQr?rrrTrrr
szToken.__init__rrrrrrirrrTrr?
scs eZdZdZfddZZS)rz'An empty token, will always match.
    cs$tt|d|_d|_d|_dS)NrTF)rQrrrrrrrTrrr
szEmpty.__init__rprrrTrr
scs*eZdZdZfddZdddZZS)r*z#A token that will never match.
    cs*tt|d|_d|_d|_d|_dS)Nr*TFzUnmatchable token)rQr*rrrrrrrTrrrs
zNoMatch.__init__TcCst|||j|dSr)r2rrrrrrszNoMatch.parseImpl)TrrrrrrrirrrTrr*scs*eZdZdZfddZdddZZS)r'aToken to exactly match a specified string.

    Example::

        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"

    For case-insensitive matching, use :class:`CaselessLiteral`.

    For keyword matching (force word break before and after the matched string),
    use :class:`Keyword` or :class:`CaselessKeyword`.
    cstt|||_t||_z|d|_Wn*tk
rVtj	dt
ddt|_YnXdt
|j|_d|j|_d|_d|_|jdkrt|tkrt|_dS)	Nrz2null string passed to Literal; use Empty() insteadrr$"%s"rFr)rQr'rmatchrmatchLenfirstMatchCharrr(r)r*rrrrrrrr_SingleCharLiteralrmatchStringrTrrr s"
zLiteral.__init__TcCs@|||jkr,||j|r,||j|jfSt|||j|dSr)ru
startswithrsrtr2rrrrrr4szLiteral.parseImpl)TrqrrrTrr's
c@seZdZdddZdS)rvTcCs0|||jkr|d|jfSt|||j|dSr)rursr2rrrrrr:sz_SingleCharLiteral.parseImplN)Trrrrrrrrrv9srvcsLeZdZdZedZdfdd	Zddd	Zfd
dZe	dd
Z
ZS)r$aToken to exactly match a specified string as a keyword, that is,
    it must be immediately followed by a non-keyword character.  Compare
    with :class:`Literal`:

     - ``Literal("if")`` will match the leading ``'if'`` in
       ``'ifAndOnlyIf'``.
     - ``Keyword("if")`` will not; it will only match the leading
       ``'if'`` in ``'if x=1'``, or ``'if(y==2)'``

    Accepts two optional constructor arguments in addition to the
    keyword string:

     - ``identChars`` is a string of characters that would be valid
       identifier characters, defaulting to all alphanumerics + "_" and
       "$"
     - ``caseless`` allows case-insensitive matching, default is ``False``.

    Example::

        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use :class:`CaselessKeyword`.
    _$NFcstt||dkrtj}||_t||_z|d|_Wn$tk
r^t	j
dtddYnXd|j|_d|j|_
d|_d|_||_|r||_|}t||_dS)Nrz2null string passed to Keyword; use Empty() insteadrr$rrrF)rQr$rDEFAULT_KEYWORD_CHARSrsrrtrurr(r)r*rrrrcaselessupper
caselessmatchr
identChars)rrxrr}rTrrr]s*

zKeyword.__init__TcCs|jr|||||j|jkr|t||jksL|||j|jkr|dksj||d|jkr||j|jfSnv|||jkr|jdks||j|r|t||jks|||j|jkr|dks||d|jkr||j|jfSt	|||j
|dSr)r}rtr~rrrrsruryr2rrrrrrrs.zKeyword.parseImplcstt|}tj|_|Sr)rQr$rIr|r)rrrTrrrIszKeyword.copycCs
|t_dS)z,Overrides the default Keyword chars
        N)r$r|rrrrsetDefaultKeywordCharsszKeyword.setDefaultKeywordChars)NF)T)rrrrrGr|rrrIrrrirrrTrr$Bs
cs*eZdZdZfddZdddZZS)rafToken to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::

        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']

    (Contrast with example for :class:`CaselessKeyword`.)
    cs6tt||||_d|j|_d|j|_dS)Nz'%s'r)rQrrr~returnStringrrrwrTrrrszCaselessLiteral.__init__TcCs@||||j|jkr,||j|jfSt|||j|dSr)rtr~rsrr2rrrrrrszCaselessLiteral.parseImpl)TrqrrrTrrs
cs"eZdZdZdfdd	ZZS)rz
    Caseless version of :class:`Keyword`.

    Example::

        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']

    (Contrast with example for :class:`CaselessLiteral`.)
    Ncstt|j||dddS)NTr})rQrr)rrxrrTrrrszCaselessKeyword.__init__)NrprrrTrrs	cs,eZdZdZdfdd	Zd	ddZZS)
raA variation on :class:`Literal` which matches "close" matches,
    that is, strings with at most 'n' mismatching characters.
    :class:`CloseMatch` takes parameters:

     - ``match_string`` - string to be matched
     - ``maxMismatches`` - (``default=1``) maximum number of
       mismatches allowed to count as a match

    The results from a successful parse will contain the matched text
    from the input string and the following named results:

     - ``mismatches`` - a list of the positions within the
       match_string where mismatches were found
     - ``original`` - the original match_string used to compare
       against the input string

    If ``mismatches`` is an empty list, then the match was an exact
    match.

    Example::

        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    rcsBtt|||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F)	rQrrrmatch_string
maxMismatchesrrr)rrrrTrrrszCloseMatch.__init__TcCs|}t|}|t|j}||kr|j}d}g}	|j}
tt||||D]2\}}|\}}
||
krL|	|t|	|
krLqqL|d}t|||g}||d<|	|d<||fSt|||j|dS)Nrroriginal
mismatches)	rrrrrrr5r2r)rrrrstartrmaxlocrZmatch_stringlocrrZs_msrcmatresultsrrrrs(
zCloseMatch.parseImpl)r)TrqrrrTrrs 	cs8eZdZdZd
fdd	Zdd	d
ZfddZZS)rBaX	Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters, an
    optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for ``min`` is
    1 (a minimum value < 1 is not valid); the default values for
    ``max`` and ``exact`` are 0, meaning no maximum or exact
    length restriction. An optional ``excludeChars`` parameter can
    list characters that might be found in the input ``bodyChars``
    string; useful to define a word of all printables except for one or
    two characters, for instance.

    :class:`srange` is useful for defining custom character set strings
    for defining ``Word`` expressions, using range notation from
    regular expression character sets.

    A common mistake is to use :class:`Word` to match a specific literal
    string, as in ``Word("Address")``. Remember that :class:`Word`
    uses the string argument to define *sets* of matchable characters.
    This expression would match "Add", "AAA", "dAred", or any other word
    made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an
    exact literal string, use :class:`Literal` or :class:`Keyword`.

    pyparsing includes helper strings for building Words:

     - :class:`alphas`
     - :class:`nums`
     - :class:`alphanums`
     - :class:`hexnums`
     - :class:`alphas8bit` (alphabetic characters in ASCII range 128-255
       - accented, tilded, umlauted, etc.)
     - :class:`punc8bit` (non-alphabetic characters in ASCII range
       128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - :class:`printables` (any non-whitespace character)

    Example::

        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))

        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums + '-')

        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")

        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    NrrFcstt|rNtdfdd|D}|rNdfdd|D}||_t||_|rt||_t||_n||_t||_|dk|_	|dkrt
d||_|dkr||_nt
|_|dkr||_||_t||_d|j|_d	|_||_d
|j|jkr|dkr|dkr|dkr|j|jkr@dt|j|_nHt|jdkrndt|jt|jf|_nd
t|jt|jf|_|jrd|jd|_zt|j|_Wntk
rd|_YnX|jj|_t|_dS)Nrc3s|]}|kr|VqdSrrrexcludeCharsrrr2sz Word.__init__..c3s|]}|kr|VqdSrrrrrrr4srrzZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedrFrz[%s]+z%s[%s]*z	[%s][%s]*z\b)rQrBrrr
initCharsOrig	initChars
bodyCharsOrig	bodyCharsmaxSpecifiedr5minLenmaxLenrhrrrr	asKeyword_escapeRegexRangeCharsreStringrrescapecompilerrsre_match
_WordRegexr)rrrminmaxexactrrrTrrr.sZ



0


z
Word.__init__Tc	Cs|||jkrt|||j||}|d7}t|}|j}||j}t||}||krj|||krj|d7}qLd}|||jkrd}nV|jr||kr|||krd}n6|j	r|dkr||d|ks||kr|||krd}|rt|||j|||||fS)NrFTr)
rr2rrrrrrrr)	rrrrrrZ	bodycharsrZthrowExceptionrrrrgs2



zWord.parseImplcsvztt|WStk
r$YnX|jdkrpdd}|j|jkr`d||j||jf|_nd||j|_|jS)NcSs$t|dkr|dddS|SdS)Nrjrrrrr
charsAsStrsz Word.__str__..charsAsStrz
W:(%s, %s)zW:(%s))rQrBrrrrr)rrrTrrrs
zWord.__str__)NrrrFN)TrrrrrrrrirrrTrrBs49
c@seZdZdddZdS)rTcCs4|||}|s t|||j||}||fSr)rr2rendgroup)rrrrrfrrrrs
z_WordRegex.parseImplN)Trzrrrrrsrcs"eZdZdZdfdd	ZZS)rFzA short-cut class for defining ``Word(characters, exact=1)``,
    when defining a match of any single character in a string of
    characters.
    FNcsZtt|j|d||ddtd|j|_|r>d|j|_t|j|_|jj	|_
dS)Nr)rrr[%s]rz\b%s\b)rQrFrrrrrrrrsr)rcharsetrrrTrrrsz
Char.__init__)FNrprrrTrrFscsbeZdZdZeedZdfdd	Zddd	Z	dd
dZ
ddd
ZfddZddZ
ZS)r:aToken for matching strings that match a given regular
    expression. Defined with string specifying the regular expression in
    a form recognized by the stdlib Python  `re module `_.
    If the given regex contains named groups (defined using ``(?P...)``),
    these will be preserved as named parse results.

    Example::

        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P\d{4})-(?P\d\d?)-(?P\d\d?)')
        # ref: https://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
    z[A-Z]rFcstt|t|tr|s,tjdtdd||_||_	zt
|j|j	|_
|j|_Wqt
jk
rtjd|tddYqXn2t|tjr||_
t||_|_||_	ntd|j
j|_t||_d|j|_d|_d|_||_||_|jr|j|_|jr|j|_d	S)
aThe parameters ``pattern`` and ``flags`` are passed
        to the ``re.compile()`` function as-is. See the Python
        `re module `_ module for an
        explanation of the acceptable patterns and flags.
        z0null string passed to Regex; use Empty() insteadrr$$invalid pattern (%s) passed to RegexzCRegex may only be constructed with a string or a compiled RE objectrFTN)rQr:rrrr(r)r*patternflagsrrr
sre_constantserrorcompiledREtyperr5rsrrrrrrasGroupListasMatchparseImplAsGroupListrparseImplAsMatch)rrrrrrTrrrsD




zRegex.__init__Tc	Csb|||}|s t|||j||}t|}|}|rZ|D]\}}|||<qH||fSr)rr2rrr5r	groupdictr!)	rrrrrfrdrrrrrrs
zRegex.parseImplcCs8|||}|s t|||j||}|}||fSr)rr2rrgroupsrrrrrfrrrrrszRegex.parseImplAsGroupListcCs4|||}|s t|||j||}|}||fSr)rr2rrrrrrr
szRegex.parseImplAsMatchcsFztt|WStk
r$YnX|jdkr@dt|j|_|jS)NzRe:(%s))rQr:rrrrrrrTrrr
s
z
Regex.__str__csljrtjdtddtjr@tr@tjdtddtjrTfdd}nfdd}|S)a
        Return Regex with an attached parse action to transform the parsed
        result as if called using `re.sub(expr, repl, string) `_.

        Example::

            make_html = Regex(r"(\w+):(.*?):").sub(r"<\1>\2")
            print(make_html.transformString("h1:main title:"))
            # prints "

main title

" z-cannot use sub() with Regex(asGroupList=True)rr$z9cannot use sub() with a callable with Regex(asMatch=True)cs|dSr)expandr)replrrr, szRegex.sub..pacsj|dSr)rrrrrrrr/ s)rr(r)r* SyntaxErrorrrr)rrrrrrr s z Regex.sub)rFF)T)T)T)rrrrrrrrrrrrrrrirrrTrr: s - cs8eZdZdZd fdd Zd ddZfd d ZZS) r8a& Token for matching strings that are delimited by quoting characters. Defined with the following parameters: - quoteChar - string of one or more characters defining the quote delimiting string - escChar - character to escape quotes, typically backslash (default= ``None``) - escQuote - special quote sequence to escape an embedded quote string (such as SQL's ``""`` to escape an embedded ``"``) (default= ``None``) - multiline - boolean indicating whether quotes can span multiple lines (default= ``False``) - unquoteResults - boolean indicating whether the matched text should be unquoted (default= ``True``) - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default= ``None`` => same as quoteChar) - convertWhitespaceEscapes - convert escaped whitespace (``'\t'``, ``'\n'``, etc.) to actual whitespace (default= ``True``) Example:: qs = QuotedString('"') print(qs.searchString('lsjdf "This is the quote" sldjf')) complex_qs = QuotedString('{{', endQuoteChar='}}') print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf')) sql_qs = QuotedString('"', escQuote='""') print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf')) prints:: [['This is the quote']] [['This is the "quote"']] [['This is the quote with "embedded" quotes']] NFTc sXtt|}|s0tjdtddt|dkr>|}n"|}|s`tjdtddt|_t |_ |d_ |_ t |_ |_|_|_|_|rtjtjB_dtjtj d|dk rt|pdf_n.r)z|(?:%s)z|(?:%s.)z(.)z)*%srrFT)'rQr8rrr(r)r*r quoteCharr quoteCharLenfirstQuoteCharrendQuoteCharLenescCharescQuoteunquoteResultsconvertWhitespaceEscapesr MULTILINEDOTALLrrrrrrescCharReplacePatternrrrsrrrrrrrr)rrrrZ multilinerrrrTrrrZ sv           zQuotedString.__init__c Cs|||jkr|||pd}|s2t|||j||}|}|jr||j|j }t |t rd|kr|j rddddd}| D]\}}| ||}q|jrt|jd|}|jr| |j|j}||fS)Nr r  )\tr\z\fz\rz\g<1>)rrr2rrrrrrrrrr!rrrrrrr) rrrrrfrZws_mapZwslitZwscharrrrr s* zQuotedString.parseImplcsHztt|WStk r$YnX|jdkrBd|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)rQr8rrrrrrrTrrr s zQuotedString.__str__)NNFTNT)TrrrrTrr83 s&A #cs8eZdZdZd fdd Zd ddZfd d ZZS) raToken for matching words composed of characters *not* in a given set (will include whitespace in matched characters if not listed in the provided exclusion set - see example). Defined with string containing all disallowed characters, and an optional minimum, maximum, and/or exact length. The default value for ``min`` is 1 (a minimum value < 1 is not valid); the default values for ``max`` and ``exact`` are 0, meaning no maximum or exact length restriction. Example:: # define a comma-separated-value as anything that is not a ',' csv_value = CharsNotIn(',') print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213")) prints:: ['dkls', 'lsdkjf', 's12 34', '@!#', '213'] rrcstt|d|_||_|dkr*td||_|dkr@||_nt|_|dkrZ||_||_t ||_ d|j |_ |jdk|_ d|_ dS)NFrzfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr)rQrrrnotCharsr5rrrhrrrrr)rrrrrrTrrr s    zCharsNotIn.__init__TcCs|||jkrt|||j||}|d7}|j}t||jt|}||krb|||krb|d7}qD|||jkrt|||j|||||fSr)rr2rrrrr)rrrrrZnotcharsmaxlenrrrr s zCharsNotIn.parseImplcsfztt|WStk r$YnX|jdkr`t|jdkrTd|jdd|_n d|j|_|jS)Nrz !W:(%s...)z!W:(%s))rQrrrrrrrrTrrrs  zCharsNotIn.__str__)rrr)TrrrrTrr s cs`eZdZdZdddddddd d d d d ddddddddddddZd"fdd Zd#d d!ZZS)$rAaSpecial matching class for matching whitespace. Normally, whitespace is ignored by pyparsing grammars. This class is included when some whitespace structures are significant. Define with a string containing the whitespace characters to be matched; default is ``" \t\r\n"``. Also takes optional ``min``, ``max``, and ``exact`` arguments, as defined for the :class:`Word` class. zzzzzzzzz z z z zzzzzz z zzzz)rrrrrzuA0zu80zu80Euu€0uu€1uu€2uu€3uu€4uu€5uu€6uu€7uu€8uu€9uu€Auu€Buu‚Fuu…FuuÀ0 rrcstt|_dfddjDdddjD_d_dj_ |_ |dkrt|_ nt _ |dkr|_ |_ dS)Nrc3s|]}|jkr|VqdSr) matchWhiterrrrr:s z!White.__init__..css|]}tj|VqdSr)rA whiteStrsrrrrr<sTrr) rQrArrrDrrrrrrrrh)rZwsrrrrTrrr7s  zWhite.__init__TcCs|||jkrt|||j||}|d7}||j}t|t|}||krb|||jkrb|d7}qB|||jkrt|||j|||||fSr)rr2rrrrr)rrrrrrrrrrKs  zWhite.parseImpl)rrrr)T)rrrrrrrrirrrTrrAs6 cseZdZfddZZS)_PositionTokencs(tt||jj|_d|_d|_dSr)rQrrrrrrrrrTrrr\s z_PositionToken.__init__rrrrrirrrTrr[srcs2eZdZdZfddZddZd ddZZS) r"zaToken to advance to a specific column of input text; useful for tabular report scraping. cstt|||_dSr)rQr"rrM)rcolnorTrrrfszGoToColumn.__init__cCs\t|||jkrXt|}|jr*|||}||krX||rXt|||jkrX|d7}q*|Sr)rMrrrisspace)rrrrrrrrjs $ zGoToColumn.preParseTcCsDt||}||jkr"t||d|||j|}|||}||fS)NzText not in expected columnrMr2)rrrrZthiscolZnewlocrrrrrss    zGoToColumn.parseImpl)T)rrrrrrrrirrrTrr"bs  cs*eZdZdZfddZdddZZS)r&aMatches if current position is at the beginning of a line within the parse string Example:: test = '''\ AAA this line AAA and this line AAA but not this one B AAA and definitely not this one ''' for t in (LineStart() + 'AAA' + restOfLine).searchString(test): print(t) prints:: ['AAA', ' this line'] ['AAA', ' and this line'] cstt|d|_dS)NzExpected start of line)rQr&rrrrTrrrszLineStart.__init__TcCs*t||dkr|gfSt|||j|dSr)rMr2rrrrrrszLineStart.parseImpl)TrqrrrTrr&|s cs*eZdZdZfddZdddZZS)r%zTMatches if current position is at the end of a line within the parse string cs,tt||tjddd|_dS)NrrzExpected end of line)rQr%rrDr7rrrrrTrrrszLineEnd.__init__TcCsb|t|kr6||dkr$|ddfSt|||j|n(|t|krN|dgfSt|||j|dS)Nrrrr2rrrrrrs     zLineEnd.parseImpl)TrqrrrTrr%s cs*eZdZdZfddZdddZZS)r=zLMatches if current position is at the beginning of the parse string cstt|d|_dS)NzExpected start of text)rQr=rrrrTrrrszStringStart.__init__TcCs0|dkr(|||dkr(t|||j||gfSr)rr2rrrrrrszStringStart.parseImpl)TrqrrrTrr=s cs*eZdZdZfddZdddZZS)r<zBMatches if current position is at the end of the parse string cstt|d|_dS)NzExpected end of text)rQr<rrrrTrrrszStringEnd.__init__TcCs^|t|krt|||j|n<|t|kr6|dgfS|t|krJ|gfSt|||j|dSrrrrrrrs    zStringEnd.parseImpl)TrqrrrTrr<s cs.eZdZdZeffdd ZdddZZS)rDayMatches if the current position is at the beginning of a Word, and is not preceded by any character in a given set of ``wordChars`` (default= ``printables``). To emulate the ```` behavior of regular expressions, use ``WordStart(alphanums)``. ``WordStart`` will also match at the beginning of the string being parsed, or at the beginning of a line. cs"tt|t||_d|_dS)NzNot at the start of a word)rQrDrr wordCharsrrrrTrrrs zWordStart.__init__TcCs@|dkr8||d|jks(|||jkr8t|||j||gfSr)rr2rrrrrrs  zWordStart.parseImpl)TrrrrrjrrrirrrTrrDscs.eZdZdZeffdd ZdddZZS)rCa_Matches if the current position is at the end of a Word, and is not followed by any character in a given set of ``wordChars`` (default= ``printables``). To emulate the ```` behavior of regular expressions, use ``WordEnd(alphanums)``. ``WordEnd`` will also match at the end of the string being parsed, or at the end of a line. cs(tt|t||_d|_d|_dS)NFzNot at the end of a word)rQrCrrrrrrrTrrrs zWordEnd.__init__TcCsPt|}|dkrH||krH|||jks8||d|jkrHt|||j||gfSr)rrr2r)rrrrrrrrrszWordEnd.parseImpl)TrrrrTrrCscszeZdZdZdfdd ZddZddZfd d Zfd d Zfd dZ dddZ fddZ dfdd Z Z S)r3z]Abstract subclass of ParserElement, for combining and post-processing parsed tokens. Fcstt|t|tr"t|}t|tr<|g_nt|t rP|g_nnt|t rt|}t dd|Drfdd|D}t|_n,zt|_Wnt k r|g_YnXd_ dS)Ncss|]}t|tVqdSr)rr)rrrrrr sz+ParseExpression.__init__..c3s&|]}t|tr|n|VqdSr)rrrrrrrrr sF)rQr3rrr r rrexprsr7r rvrrrrrrTrrrs"      zParseExpression.__init__cCs|j|d|_|Sr)rrrrSrrrrs zParseExpression.appendcCs0d|_dd|jD|_|jD] }|q|S)zExtends ``leaveWhitespace`` defined in base class, and also invokes ``leaveWhitespace`` on all contained expressions.FcSsg|] }|qSrrHrrrrrsz3ParseExpression.leaveWhitespace..)rrrC)rrrrrrCs   zParseExpression.leaveWhitespacecsrt|trB||jkrntt|||jD]}||jdq*n,tt|||jD]}||jdqX|Sr.)rr>rrQr3rFr)rrJrrTrrrF#s    zParseExpression.ignorecsNztt|WStk r$YnX|jdkrHd|jjt|jf|_|jSNz%s:(%s)) rQr3rrrrrrrrrTrrr/s zParseExpression.__str__cs*tt||jD] }|qt|jdkr|jd}t||jr|js|jdkr|j s|jdd|jdg|_d|_ |j |j O_ |j |j O_ |jd}t||jr|js|jdkr|j s|jdd|jdd|_d|_ |j |j O_ |j |j O_ dt ||_|S)Nrrrrr)rQr3rrrrrrrrrrrrr)rrrJrTrrr9s<     zParseExpression.streamlineNcCsB|dk r |ngdd|g}|jD]}||q$|gdSr)rrKrI)rrJtmprrrrrK[s  zParseExpression.validatecs$tt|}dd|jD|_|S)NcSsg|] }|qSrrHrrrrrcsz(ParseExpression.copy..)rQr3rIrrcrTrrrIaszParseExpression.copycsVtjrD|jD]6}t|tr |jr tjdd|t |j |jddq t t | ||S)N]{0}: setting results name {1!r} on {2} expression collides with {3!r} on contained expression)warn_ungrouped_named_tokens_in_collectionrr$)rrrrr7rr(r)rrrrQr3rrrrrrTrrrfs zParseExpression._setResultsName)F)N)F)rrrrrrrCrFrrrKrIrrirrrTrr3s "  cs`eZdZdZGdddeZdfdd ZfddZdd d Zd d Z d dZ ddZ Z S)ra Requires all given :class:`ParseExpression` s to be found in the given order. Expressions may be separated by whitespace. May be constructed using the ``'+'`` operator. May also be constructed using the ``'-'`` operator, which will suppress backtracking. Example:: integer = Word(nums) name_expr = OneOrMore(Word(alphas)) expr = And([integer("id"), name_expr("name"), integer("age")]) # more easily written as: expr = integer("id") + name_expr("name") + integer("age") cseZdZfddZZS)zAnd._ErrorStopcs&ttj|j||d|_|dS)N-)rQrr,rrrCryrTrrrszAnd._ErrorStop.__init__rrrrTrr,sr,Tcs|rt|krg}t|D]`\}}|tkrn|t|dkrdt||djd}|t|dqxtdq||q||dd<tt | ||t dd|jD|_ | |jdj|jdj|_d|_dS) Nrrr+z0cannot construct And with sequence ending in ...css|] }|jVqdSrrrrrrrszAnd.__init__..rT)r&rrrrrr;rrQrrrrrDrrr)rrrrrrZ skipto_argrTrrrs    z And.__init__cs|jrtdd|jddDrt|jddD]^\}}|dkrFq4t|tr4|jr4t|jdtr4|jd|j|d|jd<d|j|d<q4dd|jD|_tt|t dd|jD|_ |S)Ncss.|]&}t|to$|jo$t|jdtVqdSrN)rr3rr'rrrrrsz!And.streamline..rrcSsg|]}|dk r|qSrrrrrrrsz"And.streamline..css|] }|jVqdSrrrrrrrs) rrvrrr3r'rQrrrr)rrrrTrrrs$  zAnd.streamlinec Cs|jdj|||dd\}}d}|jddD]}t|tjrDd}q.|rz||||\}}Wqtk rtYqtk r}zd|_t|W5d}~XYqt k rt|t ||j |YqXn||||\}}|s| r.||7}q.||fS)NrFrrT) rrrrr,r6r0rrrrrr;) rrrr resultlistZ errorStoprZ exprtokensrrrrrs(   z And.parseImplcCst|tr||}||SrrrrrrSrrrrDs  z And.__iadd__cCs6|dd|g}|jD]}|||jsq2qdSr)rrIrrrsubRecCheckListrrrrrIs   zAnd.checkRecursioncCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nr{rcss|]}t|VqdSrrrrrrrszAnd.__str__..}r1rrrrrrrrrs    z And.__str__)T)T) rrrrrr,rrrrDrIrrirrrTrrts  cs^eZdZdZdfdd ZfddZddd Zd d Zd d ZddZ dfdd Z Z S)r/aRequires that at least one :class:`ParseExpression` is found. If two expressions match, the expression that matches the longest string will be used. May be constructed using the ``'^'`` operator. Example:: # construct Or using '^' operator number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) prints:: [['123'], ['3.1416'], ['789']] Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|] }|jVqdSrrrrrrrszOr.__init__..T)rQr/rrrvrrrTrrrsz Or.__init__cs.tt|tjr*tdd|jD|_|S)Ncss|] }|jVqdSrrrrrrrsz Or.streamline..)rQr/rrcollect_all_And_tokensrvrrrrTrrrsz Or.streamlineTc Csd}d}g}|jD]}z|||}Wnvtk rb} zd| _| j|krR| }| j}W5d} ~ XYqtk rt||krt|t||j|}t|}YqX|||fq|r|j t ddd|s|dd} | |||Sd} |D]\} } | | dkr | Sz| |||\}}Wn@tk r`} z d| _| j|krP| }| j}W5d} ~ XYqX|| krx||fS|| dkr||f} q| dkr| S|dk r|j|_ |nt||d|dS)NrrT)r@r rr no defined alternatives to match) rrr2rrrrrrsortrrr)rrrr maxExcLoc maxExceptionrrZloc2rZ best_exprZlongestZloc1Zexpr1rrrrrsT            z Or.parseImplcCst|tr||}||SrrrSrrr__ixor__;s  z Or.__ixor__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz ^ css|]}t|VqdSrrrrrrrEszOr.__str__..rrrrrrr@s    z Or.__str__cCs,|dd|g}|jD]}||qdSrrrIrrrrrIIs zOr.checkRecursioncsPtjs>tjr>tdd|jDr>tjdd|t |j ddt t | ||S)Ncss|]}t|tVqdSrrrrrrrrQsz%Or._setResultsName..{0}: setting results name {1!r} on {2} expression may only return a single token for an And alternative, in future will return the full list of tokens)warn_multiple_tokens_in_named_alternationrr$)rrrrrvrr(r)rrrrQr/rrrTrrrNszOr._setResultsName)F)T)F) rrrrrrrrrrIrrirrrTrr/s  = cs^eZdZdZdfdd ZfddZddd Zd d Zd d ZddZ dfdd Z Z S)r)aRequires that at least one :class:`ParseExpression` is found. If two expressions match, the first one listed is the one that will match. May be constructed using the ``'|'`` operator. Example:: # construct MatchFirst using '|' operator # watch the order of expressions to match number = Word(nums) | Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) # Fail! -> [['123'], ['3'], ['1416'], ['789']] # put more selective expression first number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums) print(number.searchString("123 3.1416 789")) # Better -> [['123'], ['3.1416'], ['789']] Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|] }|jVqdSrrrrrrrosz&MatchFirst.__init__..T)rQr)rrrvrrrTrrrlszMatchFirst.__init__cs.tt|tjr*tdd|jD|_|S)Ncss|] }|jVqdSrrrrrrrvsz(MatchFirst.streamline..)rQr)rrrrvrrrrTrrrsszMatchFirst.streamlineTc Csd}d}|jD]}z||||}|WStk r`}z|j|krP|}|j}W5d}~XYqtk rt||krt|t||j|}t|}YqXq|dk r|j|_|nt||d|dS)Nrr)rrr2rrrrr) rrrrrrrrrrrrrys$    zMatchFirst.parseImplcCst|tr||}||SrrrSrrr__ior__s  zMatchFirst.__ior__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrr | css|]}t|VqdSrrrrrrrsz%MatchFirst.__str__..rrrrrrrs    zMatchFirst.__str__cCs,|dd|g}|jD]}||qdSrrrrrrrIs zMatchFirst.checkRecursioncsPtjs>tjr>tdd|jDr>tjdd|t |j ddt t | ||S)Ncss|]}t|tVqdSrrrrrrrsz-MatchFirst._setResultsName..rrrr$)rrrrrvrr(r)rrrrQr)rrrTrrrszMatchFirst._setResultsName)F)T)F) rrrrrrrrrrIrrirrrTrr)[s   csHeZdZdZd fdd ZfddZdddZd d Zd d ZZ S)rasRequires all given :class:`ParseExpression` s to be found, but in any order. Expressions may be separated by whitespace. May be constructed using the ``'&'`` operator. Example:: color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN") shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON") integer = Word(nums) shape_attr = "shape:" + shape_type("shape") posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn") color_attr = "color:" + color("color") size_attr = "size:" + integer("size") # use Each (using operator '&') to accept attributes in any order # (shape and posn are required, color and size are optional) shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr) shape_spec.runTests(''' shape: SQUARE color: BLACK posn: 100, 120 shape: CIRCLE size: 50 color: BLUE posn: 50,80 color:GREEN size:20 shape:TRIANGLE posn:20,40 ''' ) prints:: shape: SQUARE color: BLACK posn: 100, 120 ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']] - color: BLACK - posn: ['100', ',', '120'] - x: 100 - y: 120 - shape: SQUARE shape: CIRCLE size: 50 color: BLUE posn: 50,80 ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']] - color: BLUE - posn: ['50', ',', '80'] - x: 50 - y: 80 - shape: CIRCLE - size: 50 color: GREEN size: 20 shape: TRIANGLE posn: 20,40 ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']] - color: GREEN - posn: ['20', ',', '40'] - x: 20 - y: 40 - shape: TRIANGLE - size: 20 Tcs>tt|||tdd|jD|_d|_d|_d|_dS)Ncss|] }|jVqdSrrrrrrrsz Each.__init__..T) rQrrrrrrinitExprGroupsrrrTrrrs z Each.__init__cs(tt|tdd|jD|_|S)Ncss|] }|jVqdSrrrrrrrsz"Each.streamline..)rQrrrrrrrTrrrszEach.streamlinec s|jrtdd|jD|_dd|jD}dd|jD}|||_dd|jD|_dd|jD|_dd|jD|_|j|j7_d |_|}|jdd}|jddg}d } | rj||j|j} g} | D]v} z| ||}Wn t k r| | YqX| |j t | | | |kr@| | q| kr܈ | qt| t| krd } q|rd d d|D} t ||d | |fdd|jD7}g}|D]"} | |||\}}| |qt|tg}||fS)Ncss&|]}t|trt|j|fVqdSr)rr.rVrrrrrrs z!Each.parseImpl..cSsg|]}t|tr|jqSrrr.rrrrrrs z"Each.parseImpl..cSs g|]}|jrt|ts|qSr)rrr.rrrrrs cSsg|]}t|tr|jqSr)rrErrrrrrs cSsg|]}t|tr|jqSr)rr,rrrrrrs cSs g|]}t|tttfs|qSr)rr.rEr,rrrrrsFTrVcss|]}t|VqdSrrrrrrrsz*Missing one or more required elements (%s)cs$g|]}t|tr|jkr|qSrrrZtmpOptrrrs )rrrZopt1mapZ optionalsZmultioptionalsZ multirequiredZrequiredrr2rrrVremoverrrsumr5)rrrrZopt1Zopt2ZtmpLocZtmpReqdZ matchOrderZ keepMatchingZtmpExprsZfailedrZmissingrrZ finalResultsrrrrsP    zEach.parseImplcCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz & css|]}t|VqdSrrrrrrr,szEach.__str__..rrrrrrr's    z Each.__str__cCs,|dd|g}|jD]}||qdSrrrrrrrI0s zEach.checkRecursion)T)T) rrrrrrrrrIrirrrTrrs 8  1 csjeZdZdZdfdd ZdddZdd Zfd d Zfd d ZddZ dddZ fddZ Z S)r1zfAbstract subclass of :class:`ParserElement`, for combining and post-processing parsed tokens. Fcstt||t|tr@t|jtr2||}n|t|}||_ d|_ |dk r|j |_ |j |_ | |j|j|_|j|_|j|_|j|jdSr)rQr1rrr issubclassrr?r'rrrrrDrrrrrrErrrrTrrr:s    zParseElementEnhance.__init__TcCs2|jdk r|jj|||ddStd||j|dS)NFrr)rrr2rrrrrrLs zParseElementEnhance.parseImplcCs*d|_|j|_|jdk r&|j|Sr~)rrrIrCrrrrrCRs    z#ParseElementEnhance.leaveWhitespacecsrt|trB||jkrntt|||jdk rn|j|jdn,tt|||jdk rn|j|jd|Sr.)rr>rrQr1rFrrSrTrrrFYs    zParseElementEnhance.ignorecs&tt||jdk r"|j|Sr)rQr1rrrrTrrres  zParseElementEnhance.streamlinecCsB||krt||g|dd|g}|jdk r>|j|dSr)r9rrI)rrrrrrrIks  z"ParseElementEnhance.checkRecursionNcCsB|dkr g}|dd|g}|jdk r4|j||gdSrrrKrIrrJrrrrrKrs   zParseElementEnhance.validatecsXztt|WStk r$YnX|jdkrR|jdk rRd|jjt|jf|_|jSr) rQr1rrrrrrrrrTrrrzszParseElementEnhance.__str__)F)T)N) rrrrrrrCrFrrIrKrrirrrTrr16s   cs*eZdZdZfddZdddZZS)r abLookahead matching of the given parse expression. ``FollowedBy`` does *not* advance the parsing position within the input string, it only verifies that the specified parse expression matches at the current position. ``FollowedBy`` always returns a null token list. If any results names are defined in the lookahead expression, those *will* be returned for access by name. Example:: # use FollowedBy to match a label only if it is followed by a ':' data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint() prints:: [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']] cstt||d|_dSr)rQr rrrrrTrrrszFollowedBy.__init__TcCs(|jj|||d\}}|dd=||fS)Nr)rr)rrrr_rrrrrs zFollowedBy.parseImpl)TrqrrrTrr s cs,eZdZdZd fdd Zd ddZZS) r(apLookbehind matching of the given parse expression. ``PrecededBy`` does not advance the parsing position within the input string, it only verifies that the specified parse expression matches prior to the current position. ``PrecededBy`` always returns a null token list, but if a results name is defined on the given expression, it is returned. Parameters: - expr - expression that must match prior to the current parse location - retreat - (default= ``None``) - (int) maximum number of characters to lookbehind prior to the current parse location If the lookbehind expression is a string, Literal, Keyword, or a Word or CharsNotIn with a specified exact or maximum length, then the retreat parameter is not required. Otherwise, retreat must be specified to give a maximum number of characters to look back from the current parse position for a lookbehind match. Example:: # VB-style variable names with type prefixes int_var = PrecededBy("#") + pyparsing_common.identifier str_var = PrecededBy("$") + pyparsing_common.identifier Ncstt||||_d|_d|_d|_t|t rJt |}d|_nVt|t t frf|j }d|_n:t|ttfr|jtkr|j}d|_nt|trd}d|_||_dt ||_d|_dS)NTFrznot preceded by )rQr(rrrCrrrrrrr'r$rtrBrrrhrretreatrr)rrrrTrrrs(  zPrecededBy.__init__rTc Cs|jr<||jkrt|||j||j}|j||\}}n|jt}|d|}t|||j} tdt||jdD]H} z|||| \}}Wn&t k r} z| } W5d} ~ XYqxXqqx| |dd=||fSr) rrr2rrrr<rrr0) rrrrrrrZ test_exprZinstring_sliceZ last_exprrMZpberrrrs"     zPrecededBy.parseImpl)N)rTrqrrrTrr(scs2eZdZdZfddZd ddZddZZS) r+aLookahead to disallow matching with the given parse expression. ``NotAny`` does *not* advance the parsing position within the input string, it only verifies that the specified parse expression does *not* match at the current position. Also, ``NotAny`` does *not* skip over leading whitespace. ``NotAny`` always returns a null token list. May be constructed using the '~' operator. Example:: AND, OR, NOT = map(CaselessKeyword, "AND OR NOT".split()) # take care not to mistake keywords for identifiers ident = ~(AND | OR | NOT) + Word(alphas) boolean_term = Optional(NOT) + ident # very crude boolean expression - to support parenthesis groups and # operation hierarchy, use infixNotation boolean_expr = boolean_term + ZeroOrMore((AND | OR) + boolean_term) # integers that are followed by "." are actually floats integer = Word(nums) + ~Char(".") cs0tt||d|_d|_dt|j|_dS)NFTzFound unwanted token, )rQr+rrrrrrrrTrrr szNotAny.__init__TcCs&|j||rt|||j||gfSr)rrr2rrrrrrszNotAny.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrz~{rr1rrrrrrrrrs   zNotAny.__str__)TrrrrTrr+s  cs>eZdZd fdd ZddZd ddZdfd d ZZS)_MultipleMatchNcs<tt||d|_|}t|tr.||}||dSr)rQrrrrrrstopOn)rrrenderrTrrr s   z_MultipleMatch.__init__cCs,t|tr||}|dk r"|nd|_|Sr)rrr not_ender)rrrrrr(s  z_MultipleMatch.stopOnTc Cs|jj}|j}|jdk }|r$|jj}|r2|||||||dd\}}zV|j } |r`|||| rp|||} n|} ||| |\}} | s| rR|| 7}qRWnttfk rYnX||fSNFr) rrrr rrr;r2r) rrrrZself_expr_parseZself_skip_ignorablesZ check_enderZ try_not_enderrZhasIgnoreExprsrZ tmptokensrrrr.s*      z_MultipleMatch.parseImplFcsftjrT|jgt|jdgD]6}t|tr|jrtjd d|t |j |jddqt t |||S)Nrrrrr$)rrrrrr7rr(r)rrrrQrrrrTrrrKsz_MultipleMatch._setResultsName)N)T)F)rrrrrrrrirrrTrrs rc@seZdZdZddZdS)r,ajRepetition of one or more of the given expression. Parameters: - expr - expression that must match one or more times - stopOn - (default= ``None``) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: BLACK" OneOrMore(attr_expr).parseString(text).pprint() # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']] # use stopOn attribute for OneOrMore to avoid reading label string as part of the data attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']] # could also be written as (attr_expr * (1,)).parseString(text).pprint() cCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrrz}...rrrrrrss   zOneOrMore.__str__N)rrrrrrrrrr,Yscs8eZdZdZd fdd Zd fdd Zdd ZZS) rEakOptional repetition of zero or more of the given expression. Parameters: - expr - expression that must match zero or more times - stopOn - (default= ``None``) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example: similar to :class:`OneOrMore` Ncstt|j||dd|_dS)N)rT)rQrErr)rrrrTrrrszZeroOrMore.__init__Tc s<ztt||||WSttfk r6|gfYSXdSr)rQrErr2rrrTrrrszZeroOrMore.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)NrrU]...rrrrrrs   zZeroOrMore.__str__)N)TrrrrTrrE|s c@s eZdZddZeZddZdS) _NullTokencCsdSr~rrrrrr*sz_NullToken.__bool__cCsdSrrrrrrrsz_NullToken.__str__N)rrrr*rrrrrrr sr cs<eZdZdZeZeffdd Zd ddZddZZ S) r.aGOptional matching of the given expression. Parameters: - expr - expression that must match zero or more times - default (optional) - value to be returned if the optional expression is not found. Example:: # US postal code can be a 5-digit zip, plus optional 4-digit qualifier zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4))) zip.runTests(''' # traditional ZIP code 12345 # ZIP+4 form 12101-0001 # invalid ZIP 98765- ''') prints:: # traditional ZIP code 12345 ['12345'] # ZIP+4 form 12101-0001 ['12101-0001'] # invalid ZIP 98765- ^ FAIL: Expected end of text (at char 5), (line:1, col:6) cs.tt|j|dd|jj|_||_d|_dS)NFroT)rQr.rrrrAr)rrr<rTrrrs zOptional.__init__Tc Cs|z|jj|||dd\}}WnVttfk rr|j|jk rj|jjr`t|jg}|j||jj<qn|jg}ng}YnX||fSr )rrr2rrA_Optional__optionalNotMatchedrr5)rrrrrrrrrs    zOptional.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)NrrUrXrrrrrrs   zOptional.__str__)T) rrrrr r rrrrirrrTrr.s $ cs,eZdZdZd fdd Zd ddZZS) r;a Token for skipping over all undefined text until the matched expression is found. Parameters: - expr - target expression marking the end of the data to be skipped - include - (default= ``False``) if True, the target expression is also parsed (the skipped text and target expression are returned as a 2-element list). - ignore - (default= ``None``) used to define grammars (typically quoted strings and comments) that might contain false matches to the target expression - failOn - (default= ``None``) define expressions that are not allowed to be included in the skipped test; if found before the target expression is found, the SkipTo is not a match Example:: report = ''' Outstanding Issues Report - 1 Jan 2000 # | Severity | Description | Days Open -----+----------+-------------------------------------------+----------- 101 | Critical | Intermittent system crash | 6 94 | Cosmetic | Spelling error on Login ('log|n') | 14 79 | Minor | System slow when running too many reports | 47 ''' integer = Word(nums) SEP = Suppress('|') # use SkipTo to simply match everything up until the next SEP # - ignore quoted strings, so that a '|' character inside a quoted string does not match # - parse action will call token.strip() for each matched token, i.e., the description body string_data = SkipTo(SEP, ignore=quotedString) string_data.setParseAction(tokenMap(str.strip)) ticket_expr = (integer("issue_num") + SEP + string_data("sev") + SEP + string_data("desc") + SEP + integer("days_open")) for tkt in ticket_expr.searchString(report): print tkt.dump() prints:: ['101', 'Critical', 'Intermittent system crash', '6'] - days_open: 6 - desc: Intermittent system crash - issue_num: 101 - sev: Critical ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14'] - days_open: 14 - desc: Spelling error on Login ('log|n') - issue_num: 94 - sev: Cosmetic ['79', 'Minor', 'System slow when running too many reports', '47'] - days_open: 47 - desc: System slow when running too many reports - issue_num: 79 - sev: Minor FNcs`tt||||_d|_d|_||_d|_t|t rF| ||_ n||_ dt |j |_dS)NTFzNo match found for )rQr;r ignoreExprrr includeMatchrrrrfailOnrrr)rrJZincluderFrrTrrr s zSkipTo.__init__Tc Cs&|}t|}|j}|jj}|jdk r,|jjnd}|jdk rB|jjnd} |} | |kr|dk rf||| rfq| dk rz| || } Wqntk rYqYqnXqnz||| dddWqtt fk r| d7} YqJXqqJt|||j || }|||} t | } |j r||||dd\}} | | 7} || fS)NF)rrrr) rrrrrrrr0r2rrr5r)rrrrrrrZ expr_parseZself_failOn_canParseNextZself_ignoreExpr_tryParseZtmplocZskiptextZ skipresultrrrrr-s:   zSkipTo.parseImpl)FNN)TrqrrrTrr;s9 csneZdZdZdfdd ZddZddZd d Zd d Zdd dZ ddZ fddZ dfdd Z Z S)r!a_Forward declaration of an expression to be defined later - used for recursive grammars, such as algebraic infix notation. When the expression is known, it is assigned to the ``Forward`` variable using the '<<' operator. Note: take care when assigning to ``Forward`` not to overlook precedence of operators. Specifically, '|' has a lower precedence than '<<', so that:: fwdExpr << a | b | c will actually be evaluated as:: (fwdExpr << a) | b | c thereby leaving b and c out as parseable alternatives. It is recommended that you explicitly group the values inserted into the ``Forward``:: fwdExpr << (a | b | c) Converting to use the '<<=' operator instead will avoid this problem. See :class:`ParseResults.pprint` for an example of a recursive parser created using ``Forward``. Ncstt|j|dddSrn)rQr!rrSrTrrrwszForward.__init__cCsjt|tr||}||_d|_|jj|_|jj|_||jj|jj |_ |jj |_ |j |jj |Sr) rrrrrrrrDrrrrrErSrrr __lshift__zs      zForward.__lshift__cCs||>SrrrSrrr __ilshift__szForward.__ilshift__cCs d|_|Sr~rBrrrrrCszForward.leaveWhitespacecCs$|js d|_|jdk r |j|Sr)rrrrrrrrs   zForward.streamlinecCsJ|dkr g}||kr<|dd|g}|jdk r<|j||gdSrrrrrrrKs  zForward.validatecCslt|dr|jS|jdk r |jSd|_d}z&|jdk rJt|jdd}nd}W5|jjd||_X|jS)Nrz: ...rjz: iNone)r1rrrrrr)rZ retStringrrrrs   zForward.__str__cs.|jdk rtt|St}||K}|SdSr)rrQr!rIrcrTrrrIs  z Forward.copyFcs@tjr.|jdkr.tjdd|t|jddtt | ||S)NzR{0}: setting results name {0!r} on {1} expression that has no contained expressionwarn_name_set_on_empty_Forwardrr$) rrrr(r)rrrrQr!rrrTrrrs zForward._setResultsName)N)N)F)rrrrrrrrCrrKrrIrrirrrTrr!\s   cs"eZdZdZdfdd ZZS)r@zW Abstract subclass of :class:`ParseExpression`, for converting parsed results. Fcstt||d|_dSr~)rQr@rrrrTrrrszTokenConverter.__init__)FrprrrTrr@scs6eZdZdZd fdd ZfddZdd ZZS) raConverter to concatenate all matching tokens to a single string. By default, the matching patterns must also be contiguous in the input string; this can be disabled by specifying ``'adjacent=False'`` in the constructor. Example:: real = Word(nums) + '.' + Word(nums) print(real.parseString('3.1416')) # -> ['3', '.', '1416'] # will also erroneously match the following print(real.parseString('3. 1416')) # -> ['3', '.', '1416'] real = Combine(Word(nums) + '.' + Word(nums)) print(real.parseString('3.1416')) # -> ['3.1416'] # no match when there are internal spaces print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...) rTcs8tt|||r|||_d|_||_d|_dSr)rQrrrCadjacentr joinStringr)rrrrrTrrrszCombine.__init__cs(|jrt||ntt|||Sr)rr7rFrQrrSrTrrrFszCombine.ignorecCsP|}|dd=|td||jg|jd7}|jrH|rH|gS|SdS)Nr)r)rIr5rrYrrrr;)rrrrZretToksrrrrs  "zCombine.postParse)rT)rrrrrrFrrirrrTrrs cs(eZdZdZfddZddZZS)r#aConverter to return the matched tokens as a list - useful for returning tokens of :class:`ZeroOrMore` and :class:`OneOrMore` expressions. Example:: ident = Word(alphas) num = Word(nums) term = ident | num func = ident + Optional(delimitedList(term)) print(func.parseString("fn a, b, 100")) # -> ['fn', 'a', 'b', '100'] func = ident + Group(Optional(delimitedList(term))) print(func.parseString("fn a, b, 100")) # -> ['fn', ['a', 'b', '100']] cstt||d|_dSr)rQr#rrrrTrrr szGroup.__init__cCs|gSrrrrrrrszGroup.postParserrrrrrrirrrTrr#s cs(eZdZdZfddZddZZS)ra?Converter to return a repetitive expression as a list, but also as a dictionary. Each element can also be referenced using the first token in the expression as its key. Useful for tabular report scraping when the first column can be used as a item key. Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) # print attributes as plain groups print(OneOrMore(attr_expr).parseString(text).dump()) # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names result = Dict(OneOrMore(Group(attr_expr))).parseString(text) print(result.dump()) # access named fields as dict entries, or output as dict print(result['shape']) print(result.asDict()) prints:: ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap'] [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'} See more examples at :class:`ParseResults` of accessing fields by results name. cstt||d|_dSr)rQrrrrrTrrr8sz Dict.__init__cCst|D]\}}t|dkrq|d}t|tr@t|d}t|dkr\td|||<qt|dkrt|dtst|d|||<q|}|d=t|dkst|tr| rt||||<qt|d|||<q|j r|gS|SdS)Nrrrr) rrrrrrrr5rIr;r)rrrrrtokZikeyZ dictvaluerrrr<s$   zDict.postParserrrrTrrs& c@s eZdZdZddZddZdS)r>a[Converter for ignoring the results of a parsed expression. Example:: source = "a, b, c,d" wd = Word(alphas) wd_list1 = wd + ZeroOrMore(',' + wd) print(wd_list1.parseString(source)) # often, delimiters that are useful during parsing are just in the # way afterward - use Suppress to keep them out of the parsed output wd_list2 = wd + ZeroOrMore(Suppress(',') + wd) print(wd_list2.parseString(source)) prints:: ['a', ',', 'b', ',', 'c', ',', 'd'] ['a', 'b', 'c', 'd'] (See also :class:`delimitedList`.) cCsgSrrrrrrrkszSuppress.postParsecCs|SrrrrrrrAnszSuppress.suppressN)rrrrrrArrrrr>Usc@s(eZdZdZddZddZddZdS) r-zDWrapper for parse actions, to ensure they are only called once. cCst||_d|_dSr~)rrcalled)rZ methodCallrrrrus zOnlyOnce.__init__cCs.|js||||}d|_|St||ddS)NTr)rrr2)rrrrrrrrr@xs zOnlyOnce.__call__cCs d|_dSr~)rrrrrreset~szOnlyOnce.resetN)rrrrrr@rrrrrr-rscs:tfdd}z j|_Wntk r4YnX|S)aqDecorator for debugging parse actions. When the parse action is called, this decorator will print ``">> entering method-name(line:, , )"``. When the parse action completes, the decorator will print ``"<<"`` followed by the returned value, or any exception that the parse action raised. Example:: wd = Word(alphas) @traceParseAction def remove_duplicate_chars(tokens): return ''.join(sorted(set(''.join(tokens)))) wds = OneOrMore(wd).setParseAction(remove_duplicate_chars) print(wds.parseString("slkdjs sld sldd sdlf sdljf")) prints:: >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {})) <>entering %s(line: '%s', %d, %r) z<.z)rrr)rOrrrrrvs  ,cCs`t|dt|dt|d}|rBt|t|||S|tt|||SdS)aHelper to define a delimited list of expressions - the delimiter defaults to ','. By default, the list elements and delimiters can have intervening whitespace, and comments, but this can be overridden by passing ``combine=True`` in the constructor. If ``combine`` is set to ``True``, the matching tokens are returned as a single token string, with the delimiters included; otherwise, the matching tokens are returned as a list of tokens, with the delimiters suppressed. Example:: delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc'] delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE'] z [rr N)rrrErr>)rZdelimcombineZdlNamerrrrTs$csjtfdd}|dkr0ttdd}n|}|d|j|dd|d td S) a>Helper to define a counted list of expressions. This helper defines a pattern of the form:: integer expr expr expr... where the leading integer tells how many expr expressions follow. The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed. If ``intExpr`` is specified, it should be a pyparsing expression that produces an integer value. Example:: countedArray(Word(alphas)).parseString('2 ab cd ef') # -> ['ab', 'cd'] # in this parser, the leading integer value is given in binary, # '10' indicating that 2 values are in the array binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2)) countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef') # -> ['ab', 'cd'] cs.|d}|r ttg|p&tt>gSr)r#rrW)rrrr1Z arrayExprrrrcountFieldParseActions"z+countedArray..countFieldParseActionNcSs t|dSr)rrrrrrrzcountedArray..ZarrayLenTrz(len) rj)r!rBrfrrIrrr)rZintExprr"rr!rrPs cCs6g}|D](}t|tr&|t|q||q|Sr)rr rEr r)Lrrrrrr s   r cs6tfdd}|j|dddt|S)a4Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousLiteral(first) matchExpr = first + ":" + second will match ``"1:1"``, but not ``"1:2"``. Because this matches a previous literal, will also match the leading ``"1:1"`` in ``"1:10"``. If this is not desired, use :class:`matchPreviousExpr`. Do *not* use with packrat parsing enabled. csP|rBt|dkr|d>qLt|}tdd|D>n t>dS)Nrrcss|]}t|VqdSr)r'rZttrrrr szDmatchPreviousLiteral..copyTokenToRepeater..)rr rrr)rrrZtflatZreprrcopyTokenToRepeaters   z1matchPreviousLiteral..copyTokenToRepeaterTr#(prev) )r!rrr)rr'rr&rrcs  csFt|}|Kfdd}|j|dddt|S)aTHelper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousExpr(first) matchExpr = first + ":" + second will match ``"1:1"``, but not ``"1:2"``. Because this matches by expressions, will *not* match the leading ``"1:1"`` in ``"1:10"``; the expressions are evaluated first, and then compared, so ``"1"`` is compared with ``"10"``. Do *not* use with packrat parsing enabled. cs*t|fdd}j|dddS)Ncs$t|}|kr tddddS)Nrr)r rr2)rrrZ theseTokensZ matchTokensrrmustMatchTheseTokens%s zLmatchPreviousExpr..copyTokenToRepeater..mustMatchTheseTokensTr#)r rr)rrrr*r&r)rr'#s  z.matchPreviousExpr..copyTokenToRepeaterTr#r()r!rIrrr)rZe2r'rr&rrbs cCs:dD]}||t|}q|dd}|dd}t|S)Nz\^-]rr\rr)r_bslashr)rrrrrr.s   rc st|trtjddd|r:dd}dd}|r4tntndd}dd}|rRtntg}t|trn|}n$t|t rt |}ntjd t dd|st S|s.d }|t |d kr.||}t||d d D]N\} } || |r||| d =qq||| r||| d =||| qq|d 7}q|s|s|rzlt |t d |krtdd dd|Dd|WStddd|Dd|WSWn&tk rtjdt ddYnXtfdd|Dd|S)aHelper to quickly define a set of alternative Literals, and makes sure to do longest-first testing when there is a conflict, regardless of the input order, but returns a :class:`MatchFirst` for best performance. Parameters: - strs - a string of space-delimited literals, or a collection of string literals - caseless - (default= ``False``) - treat all literals as caseless - useRegex - (default= ``True``) - as an optimization, will generate a Regex object; otherwise, will generate a :class:`MatchFirst` object (if ``caseless=True`` or ``asKeyword=True``, or if creating a :class:`Regex` raises an exception) - asKeyword - (default=``False``) - enforce Keyword-style matching on the generated expressions Example:: comp_oper = oneOf("< = > <= >= !=") var = Word(alphas) number = Word(nums) term = var | number comparison_expr = term + comp_oper + term print(comparison_expr.searchString("B = 12 AA=23 B<=AA AA>12")) prints:: [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']] z_More than one string argument passed to oneOf, pass choices as a list or space-delimited stringrr$cSs||kSr)r~rLbrrrr[rzoneOf..cSs||Sr)r~ryr,rrrr\rcSs||kSrrr,rrrr_rcSs ||Sr)ryr,rrrr`rz6Invalid argument to oneOf, expected string or iterablerrNrrcss|]}t|VqdSr)rrZsymrrrrszoneOf..r|css|]}t|VqdSr)rrr.rrrrsz7Exception creating Regex for oneOf, building MatchFirstc3s|]}|VqdSrrr.ZparseElementClassrrrs)rrr(r)rrr$r'rr r r*r*rrrBrr:rrr)) Zstrsr}ZuseRegexrZisequalZmasksZsymbolsrZcurr#rJrr0rrg6s\          ** cCsttt||S)aHelper to easily and clearly define a dictionary by specifying the respective patterns for the key and value. Takes care of defining the :class:`Dict`, :class:`ZeroOrMore`, and :class:`Group` tokens in the proper order. The key pattern can include delimiting markers or punctuation, as long as they are suppressed, thereby leaving the significant key text. The value pattern can include named results, so that the :class:`Dict` results can include named token fields. Example:: text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) print(OneOrMore(attr_expr).parseString(text).dump()) attr_label = label attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join) # similar to Dict, but simpler call format result = dictOf(attr_label, attr_value).parseString(text) print(result.dump()) print(result['shape']) print(result.shape) # object attribute access works too print(result.asDict()) prints:: [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE SQUARE {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'} )rr,r#)r@r$rrrrUs%cCs^tdd}|}d|_|d||d}|r@dd}ndd}|||j|_|S) aHelper to return the original, untokenized text for a given expression. Useful to restore the parsed fields of an HTML start tag into the raw tag text itself, or to revert separate tokens with intervening whitespace back to the original matching input text. By default, returns astring containing the original parsed text. If the optional ``asString`` argument is passed as ``False``, then the return value is a :class:`ParseResults` containing any results names that were originally matched, and a single token containing the original matched text from the input string. So if the expression passed to :class:`originalTextFor` contains expressions with defined results names, you must set ``asString`` to ``False`` if you want to preserve those results name values. Example:: src = "this is test bold text normal text " for tag in ("b", "i"): opener, closer = makeHTMLTags(tag) patt = originalTextFor(opener + SkipTo(closer) + closer) print(patt.searchString(src)[0]) prints:: [' bold text '] ['text'] cSs|Srr)rrrrrrrrz!originalTextFor..F_original_start _original_endcSs||j|jSr)r1r2rrrrrrcSs&||d|dg|dd<dS)Nr1r2)r?rrrr extractTextsz$originalTextFor..extractText)rrrIrr)rZasStringZ locMarkerZ endlocMarker matchExprr3rrrr{s  cCst|ddS)zkHelper to undo pyparsing's default grouping of And expressions, even if all but one are non-empty. cSs|dSrrrrrrrrzungroup..)r@r)rrrrr|scCs4tdd}t|d|d|dS)aHelper to decorate a returned token with its starting and ending locations in the input string. This helper adds the following results names: - locn_start = location where matched expression begins - locn_end = location where matched expression ends - value = the actual parsed results Be careful if the input text contains ```` characters, you may want to call :class:`ParserElement.parseWithTabs` Example:: wd = Word(alphas) for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"): print(match) prints:: [[0, 'ljsdf', 5]] [[8, 'lksdjjf', 15]] [[18, 'lkkjj', 23]] cSs|SrrrrrrrrzlocatedExpr..Z locn_startr$Zlocn_end)rrr#rIrC)rZlocatorrrrr~sz\[]-*.$+^?()~ rcCs |ddSrrrrrrrrrz\\0?[xX][0-9a-fA-F]+cCstt|dddS)Nrz\0xr)unichrrrarrrrr rz \\0[0-7]+cCstt|ddddS)Nrr)r6rrrrrr rz\]rrUrnegatebodyrXcsFddz"dfddt|jDWStk r@YdSXdS)aHelper to easily define string ranges for use in Word construction. Borrows syntax from regexp '[]' string range definitions:: srange("[0-9]") -> "0123456789" srange("[a-z]") -> "abcdefghijklmnopqrstuvwxyz" srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_" The input string must be enclosed in []'s, and the returned string is the expanded character set joined into a single string. The values enclosed in the []'s may be: - a single character - an escaped character with a leading backslash (such as ``\-`` or ``\]``) - an escaped hex character with a leading ``'\x'`` (``\x21``, which is a ``'!'`` character) (``\0x##`` is also supported for backwards compatibility) - an escaped octal character with a leading ``'\0'`` (``\041``, which is a ``'!'`` character) - a range of any of the above, separated by a dash (``'a-z'``, etc.) - any combination of the above (``'aeiouy'``, ``'a-zA-Z0-9_$'``, etc.) cSs<t|ts|Sdddtt|dt|ddDS)Nrcss|]}t|VqdSrr6rrrrr)sz+srange....rr)rr5rrord)prrrr)rzsrange..rc3s|]}|VqdSrr)rpartZ _expandedrrr+szsrange..N)r_reBracketExprrr9rrrr>rrss "csfdd}|S)zoHelper method for defining parse actions that require matching at a specific column in the input text. cs"t||krt||ddS)Nzmatched token not at column %dr)rZlocnrr0rr verifyCol3sz!matchOnlyAtCol..verifyColr)r1r@rr0rra/s cs fddS)aHelper method for common parse actions that simply return a literal value. Especially useful when used with :class:`transformString` (). Example:: num = Word(nums).setParseAction(lambda toks: int(toks[0])) na = oneOf("N/A NA").setParseAction(replaceWith(math.nan)) term = na | num OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234] csgSrrrZreplStrrrrErzreplaceWith..rrArrArrp8s cCs|dddS)aHelper parse action for removing quotation marks from parsed quoted strings. Example:: # by default, quotation marks are included in parsed results quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"] # use removeQuotes to strip quotation marks from parsed results quotedString.setParseAction(removeQuotes) quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"] rrrrrrrrrnGs csNfdd}ztdtdj}Wntk rBt}YnX||_|S)aLHelper to define a parse action by mapping a function to all elements of a ParseResults list. If any additional args are passed, they are forwarded to the given function as additional arguments after the token, as in ``hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))``, which will convert the parsed data to an integer using base 16. Example (compare the last to example in :class:`ParserElement.transformString`:: hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16)) hex_ints.runTests(''' 00 11 22 aa FF 0a 0d 1a ''') upperword = Word(alphas).setParseAction(tokenMap(str.upper)) OneOrMore(upperword).runTests(''' my kingdom for a horse ''') wd = Word(alphas).setParseAction(tokenMap(str.title)) OneOrMore(wd).setParseAction(' '.join).runTests(''' now is the winter of our discontent made glorious summer by this sun of york ''') prints:: 00 11 22 aa FF 0a 0d 1a [0, 17, 34, 170, 255, 10, 13, 26] my kingdom for a horse ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE'] now is the winter of our discontent made glorious summer by this sun of york ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York'] csfdd|DS)Ncsg|]}|fqSrr)rZtoknrrrrr{sz(tokenMap..pa..rrrBrrrzsztokenMap..parr)rrrr)rrrrrrBrrVs$ cCs t|Srrr~rrrrrrcCs t|Srrlowerrrrrrrrfrgcs~t|tr|t|| d}n|jtttd}|rt t }||dt t t |td|tddgdd d d |}nlt t ttd d B}||dt t t | tttd|tddgdd d d |}ttd|d dd}|d|fdd |ddddd}|_|_t||_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namerz_-:tag=/Fr<rWcSs |ddkSNrrHrrrrrrrz_makeTags..rgrcSs |ddkSrJrrrrrrrrh)rz<%s>c s*|dddd|S)Nrr:r)rrrtitlerrIrZresnamerrrrrrrKrz)rrr$rrBrHrGrRrIrrnrrEr#r>r.rmrjrVr_LrrrrrLrrFr;Ztag_body)tagStrZxmlZ suppress_LTZ suppress_GTZ tagAttrNameZ tagAttrValueZopenTagZcloseTagrrMr _makeTagssH , rPcCs t|dS)aKHelper to construct opening and closing tag expressions for HTML, given a tag name. Matches tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values. Example:: text = 'More info at the
pyparsing wiki page' # makeHTMLTags returns pyparsing expressions for the opening and # closing tags as a 2-tuple a, a_end = makeHTMLTags("A") link_expr = a + SkipTo(a_end)("link_text") + a_end for link in link_expr.searchString(text): # attributes in the tag (like "href" shown here) are # also accessible as named results print(link.link_text, '->', link.href) prints:: pyparsing -> https://github.com/pyparsing/pyparsing/wiki FrPrOrrrr_scCs t|dS)zHelper to construct opening and closing tag expressions for XML, given a tag name. Matches tags only in the given upper/lower case. Example: similar to :class:`makeHTMLTags` TrQrRrrrr`scs8|r|ddn|ddDfdd}|S)a7Helper to create a validating parse action to be used with start tags created with :class:`makeXMLTags` or :class:`makeHTMLTags`. Use ``withAttribute`` to qualify a starting tag with a required attribute value, to avoid false matches on common tags such as ```` or ``
``. Call ``withAttribute`` with a series of attribute names and values. Specify the list of filter attributes names and values as: - keyword arguments, as in ``(align="right")``, or - as an explicit dict with ``**`` operator, when an attribute name is also a Python reserved word, as in ``**{"class":"Customer", "align":"right"}`` - a list of name-value tuples, as in ``(("ns1:class", "Customer"), ("ns2:align", "right"))`` For attribute names with a namespace prefix, you must use the second form. Attribute names are matched insensitive to upper/lower case. If just testing for ``class`` (with or without a namespace), use :class:`withClass`. To verify that the attribute exists, but without specifying a value, pass ``withAttribute.ANY_VALUE`` as the value. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this has no type
''' div,div_end = makeHTMLTags("div") # only match div tag having a type attribute with value "grid" div_grid = div().setParseAction(withAttribute(type="grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) # construct a match with any div tag having a type attribute, regardless of the value div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 NcSsg|]\}}||fqSrrrarrrrsz!withAttribute..csZD]P\}}||kr$t||d||tjkr|||krt||d||||fqdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r2ry ANY_VALUE)rrrZattrNameZ attrValueZattrsrrrs  zwithAttribute..pa)r!)rZattrDictrrrTrrys 8 cCs|r d|nd}tf||iS)aSimplified version of :class:`withAttribute` when matching on a div class - made difficult because ``class`` is a reserved word in Python. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this <div> has no class
''' div,div_end = makeHTMLTags("div") div_grid = div().setParseAction(withClass("grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 z%s:classclass)ry)Z classname namespaceZ classattrrrrrs#(rcCsGdddt}t}||||B}t|D]n\}}|ddd\} } } } | dkr`d| nd| } | dkr| dkst| d krtd | \}}t| }| tjkrp| d kr||| t|t | }n| d kr*| dk r ||| |t|t | |}n|||t|t |}nD| dkrf||||||t|||||}ntd n| tj krT| d krt | t st | } || j |t| |}n| d kr| dk r||| |t|t | |}n|||t|t |}nD| dkrJ||||||t|||||}ntd ntd | rt | ttfr~|j| n || ||| |BK}|}q.||K}|S)al Helper method for constructing grammars of expressions made up of operators working in a precedence hierarchy. Operators may be unary or binary, left- or right-associative. Parse actions can also be attached to operator expressions. The generated parser will also recognize the use of parentheses to override operator precedences (see example below). Note: if you define a deep operator list, you may see performance issues when using infixNotation. See :class:`ParserElement.enablePackrat` for a mechanism to potentially improve your parser performance. Parameters: - baseExpr - expression representing the most basic element for the nested - opList - list of tuples, one for each operator precedence level in the expression grammar; each tuple is of the form ``(opExpr, numTerms, rightLeftAssoc, parseAction)``, where: - opExpr is the pyparsing expression for the operator; may also be a string, which will be converted to a Literal; if numTerms is 3, opExpr is a tuple of two expressions, for the two operators separating the 3 terms - numTerms is the number of terms for this operator (must be 1, 2, or 3) - rightLeftAssoc is the indicator whether the operator is right or left associative, using the pyparsing-defined constants ``opAssoc.RIGHT`` and ``opAssoc.LEFT``. - parseAction is the parse action to be associated with expressions matching this operator expression (the parse action tuple member may be omitted); if the parse action is passed a tuple or list of functions, this is equivalent to calling ``setParseAction(*fn)`` (:class:`ParserElement.setParseAction`) - lpar - expression for matching left-parentheses (default= ``Suppress('(')``) - rpar - expression for matching right-parentheses (default= ``Suppress(')')``) Example:: # simple example of four-function arithmetic with ints and # variable names integer = pyparsing_common.signed_integer varname = pyparsing_common.identifier arith_expr = infixNotation(integer | varname, [ ('-', 1, opAssoc.RIGHT), (oneOf('* /'), 2, opAssoc.LEFT), (oneOf('+ -'), 2, opAssoc.LEFT), ]) arith_expr.runTests(''' 5+3*6 (5+3)*6 -2--11 ''', fullDump=False) prints:: 5+3*6 [[5, '+', [3, '*', 6]]] (5+3)*6 [[[5, '+', 3], '*', 6]] -2--11 [[['-', 2], '-', ['-', 11]]] c@seZdZdddZdS)zinfixNotation.._FBTcSs|j|||gfSr)rrrrrrrsz$infixNotation.._FB.parseImplN)Trzrrrr_FBsrXrNrrz%s termz %s%s termrz@if numterms=3, opExpr must be a tuple or list of two expressionsrz6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)r r!rrr5rrhLEFTr#r,RIGHTrr.rr4r r)ZbaseExprZopListZlparZrparrXrZlastExprrZoperDefZopExprZarityZrightLeftAssocrZtermNameZopExpr1ZopExpr2ZthisExprr4rrrr}Ds`H    &       &    z4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*'z string enclosed in single quotesz*quotedString using single or double quotesuzunicode string literalcCs||krtd|dkr*t|tr"t|tr"t|dkrt|dkr|dk rtt|t||tjdd dd}n$t t||tj dd}nx|dk rtt|t |t |ttjdd dd}n4ttt |t |ttjdd d d}ntd t }|dk rd|tt|t||B|Bt|K}n$|tt|t||Bt|K}|d ||f|S) a Helper method for defining nested lists enclosed in opening and closing delimiters ("(" and ")" are the default). Parameters: - opener - opening character for a nested list (default= ``"("``); can also be a pyparsing expression - closer - closing character for a nested list (default= ``")"``); can also be a pyparsing expression - content - expression for items within the nested lists (default= ``None``) - ignoreExpr - expression for ignoring opening and closing delimiters (default= :class:`quotedString`) If an expression is not provided for the content argument, the nested expression will capture all whitespace-delimited content between delimiters as a list of separate values. Use the ``ignoreExpr`` argument to define expressions that may contain opening or closing characters that should not be treated as opening or closing characters for nesting, such as quotedString or a comment expression. Specify multiple expressions using an :class:`Or` or :class:`MatchFirst`. The default is :class:`quotedString`, but if no expressions are to be ignored, then pass ``None`` for this argument. Example:: data_type = oneOf("void int short long char float double") decl_data_type = Combine(data_type + Optional(Word('*'))) ident = Word(alphas+'_', alphanums+'_') number = pyparsing_common.number arg = Group(decl_data_type + ident) LPAR, RPAR = map(Suppress, "()") code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment)) c_function = (decl_data_type("type") + ident("name") + LPAR + Optional(delimitedList(arg), [])("args") + RPAR + code_body("body")) c_function.ignore(cStyleComment) source_code = ''' int is_odd(int x) { return (x%2); } int dec_to_hex(char hchar) { if (hchar >= '0' && hchar <= '9') { return (ord(hchar)-ord('0')); } else { return (10+ord(hchar)-ord('A')); } } ''' for func in c_function.searchString(source_code): print("%(name)s (%(type)s) args: %(args)s" % func) prints:: is_odd (int) args: [['int', 'x']] dec_to_hex (int) args: [['char', 'hchar']] z.opening and closing strings cannot be the sameNrr5cSs |dSrrrrrrrrznestedExpr..cSs |dSrr^rrrrr rcSs |dSrr^rrrrr'rcSs |dSrr^rrrrr,rzOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)r5rrrrr,rr7rrrWrIr'r!r#r>rEr)ZopenerZcloserZcontentrrrrrrds`A      *$c sddfddfdd}fdd}fdd }ttd }tt|d }t|d }t|d } |rtt||t|t|t|| } n(tt|t|t|t|| } | fdd| t t| dS)aHelper method for defining space-delimited indentation blocks, such as those used to define block statements in Python source code. Parameters: - blockStatementExpr - expression defining syntax of statement that is repeated within the indented block - indentStack - list created by caller to manage indentation stack (multiple statementWithIndentedBlock expressions within a single grammar should share a common indentStack) - indent - boolean indicating whether block must be indented beyond the current level; set to False for block of left-most statements (default= ``True``) A valid block must contain at least one ``blockStatement``. Example:: data = ''' def A(z): A1 B = 100 G = A2 A2 A3 B def BB(a,b,c): BB1 def BBA(): bba1 bba2 bba3 C D def spam(x,y): def eggs(z): pass ''' indentStack = [1] stmt = Forward() identifier = Word(alphas, alphanums) funcDecl = ("def" + identifier + Group("(" + Optional(delimitedList(identifier)) + ")") + ":") func_body = indentedBlock(stmt, indentStack) funcDef = Group(funcDecl + func_body) rvalue = Forward() funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")") rvalue << (funcCall | identifier | Word(nums)) assignment = Group(identifier + "=" + rvalue) stmt << (funcDef | assignment | identifier) module_body = OneOrMore(stmt) parseTree = module_body.parseString(data) parseTree.pprint() prints:: [['def', 'A', ['(', 'z', ')'], ':', [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]], 'B', ['def', 'BB', ['(', 'a', 'b', 'c', ')'], ':', [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]], 'C', 'D', ['def', 'spam', ['(', 'x', 'y', ')'], ':', [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] Ncsdd<dSrrr) backup_stack indentStackrr reset_stacksz"indentedBlock..reset_stackcsN|t|krdSt||}|dkrJ|dkr>t||dt||ddS)Nrzillegal nestingznot a peer entry)rrMr2rrrZcurColr`rrcheckPeerIndents     z&indentedBlock..checkPeerIndentcs2t||}|dkr"|n t||ddS)Nrznot a subentry)rMrr2rbrcrrcheckSubIndents   z%indentedBlock..checkSubIndentcsJ|t|krdSt||}r&|ks2t||d|dkrFdS)Nznot an unindentr)rrMr2r?rbrcrr checkUnindents     z$indentedBlock..checkUnindentz INDENTrZUNINDENTcsSrr)rLr-rr)rarrrrzindentedBlock..zindented block) r,r%rDrArrrr#r.rrFr+) ZblockStatementExprr`rjrdrerfrwrgZPEERZUNDENTZsmExprr)r_r`rarrz7s2Q    z#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z &(?Pr/z);zcommon HTML entitycCs t|jS)zRHelper parser action to replace common HTML entities with their special characters)_htmlEntityMaprZentityrrrrrosz/\*(?:[^*]|\*(?!/))*z*/zC style commentzz HTML commentz.*z rest of linez//(?:\\\n|[^\n])*z // commentzC++ style commentz#.*zPython style commentr commaItemrIc@seZdZdZeeZeeZe e  d eZ e e d eedZed d eZe ede e dZed d eeeed eB d Zeeed  d eZed d eZeeBeBZed d eZe eded dZed dZ ed dZ!e!de!d dZ"ee!de!ddee!de!d dZ#e#$dd d e  d!Z%e&e"e%Be#B d" d"Z'ed# d$Z(e)d=d&d'Z*e)d>d)d*Z+ed+ d,Z,ed- d.Z-ed/ d0Z.e/e0BZ1e)d1d2Z2e&e3e4d3e5e e6d3d4ee7d5 d6Z8e9ee:;e8Bd7d8 d9Zd`, :class:`reals`, :class:`scientific notation`) - common :class:`programming identifiers` - network addresses (:class:`MAC`, :class:`IPv4`, :class:`IPv6`) - ISO8601 :class:`dates` and :class:`datetime` - :class:`UUID` - :class:`comma-separated list` Parse actions: - :class:`convertToInteger` - :class:`convertToFloat` - :class:`convertToDate` - :class:`convertToDatetime` - :class:`stripHTMLTags` - :class:`upcaseTokens` - :class:`downcaseTokens` Example:: pyparsing_common.number.runTests(''' # any int or real number, returned as the appropriate type 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.fnumber.runTests(''' # any int or real number, returned as float 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.hex_integer.runTests(''' # hex numbers 100 FF ''') pyparsing_common.fraction.runTests(''' # fractions 1/2 -3/4 ''') pyparsing_common.mixed_integer.runTests(''' # mixed fractions 1 1/2 -3/4 1-3/4 ''') import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(''' # uuid 12345678-1234-5678-1234-567812345678 ''') prints:: # any int or real number, returned as the appropriate type 100 [100] -100 [-100] +100 [100] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # any int or real number, returned as float 100 [100.0] -100 [-100.0] +100 [100.0] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # hex numbers 100 [256] FF [255] # fractions 1/2 [0.5] -3/4 [-0.75] # mixed fractions 1 [1] 1/2 [0.5] -3/4 [-0.75] 1-3/4 [1.75] # uuid 12345678-1234-5678-1234-567812345678 [UUID('12345678-1234-5678-1234-567812345678')] integerz hex integerrz[+-]?\d+zsigned integerrHfractioncCs|d|dS)Nrrrrrrrrrzpyparsing_common.rz"fraction or mixed integer-fractionz[+-]?(:?\d+\.\d*|\.\d+)z real numberz@[+-]?(:?\d+(:?[eE][+-]?\d+)|(:?\d+\.\d*|\.\d+)(:?[eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?fnumberr identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}z IPv4 addressz[0-9a-fA-F]{1,4} hex_integerrKzfull IPv6 address)rrz::zshort IPv6 addresscCstdd|DdkS)Ncss|]}tj|rdVqdSrd)r _ipv6_partrr%rrrrs z,pyparsing_common...r7)rrrrrrrz::ffff:zmixed IPv6 addressz IPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}z MAC address%Y-%m-%dcsfdd}|S)a Helper to create a parse action for converting parsed date string to Python datetime.date Params - - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%d"``) Example:: date_expr = pyparsing_common.iso8601_date.copy() date_expr.setParseAction(pyparsing_common.convertToDate()) print(date_expr.parseString("1999-12-31")) prints:: [datetime.date(1999, 12, 31)] c sNzt|dWStk rH}zt||t|W5d}~XYnXdSr)rstrptimedater5r2rrrrZvefmtrrcvt_fnsz.pyparsing_common.convertToDate..cvt_fnrrwrxrrvr convertToDates zpyparsing_common.convertToDate%Y-%m-%dT%H:%M:%S.%fcsfdd}|S)aHelper to create a parse action for converting parsed datetime string to Python datetime.datetime Params - - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%dT%H:%M:%S.%f"``) Example:: dt_expr = pyparsing_common.iso8601_datetime.copy() dt_expr.setParseAction(pyparsing_common.convertToDatetime()) print(dt_expr.parseString("1999-12-31T23:59:59.999")) prints:: [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)] c sJzt|dWStk rD}zt||t|W5d}~XYnXdSr)rrsr5r2rrurvrrrxsz2pyparsing_common.convertToDatetime..cvt_fnrryrrvrconvertToDatetimes z"pyparsing_common.convertToDatetimez7(?P\d{4})(?:-(?P\d\d)(?:-(?P\d\d))?)?z ISO8601 datez(?P\d{4})-(?P\d\d)-(?P\d\d)[T ](?P\d\d):(?P\d\d)(:(?P\d\d(\.\d*)?)?)?(?PZ|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}UUIDcCstj|dS)aParse action to remove HTML tags from web page HTML source Example:: # strip HTML links from normal text text = 'More info at the
pyparsing wiki page' td, td_end = makeHTMLTags("TD") table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end print(table_text.parseString(text).body) Prints:: More info at the pyparsing wiki page r)r_html_stripperr)rrrrrr stripHTMLTagsszpyparsing_common.stripHTMLTagsrrrirjrrIzcomma separated listcCs t|SrrCrrrrrrcCs t|SrrDrrrrrrN)rr)r{)?rrrrrrZconvertToIntegerfloatZconvertToFloatrBrfrrrkrXror:Zsigned_integerrlrr.rAZ mixed_integerrrealZsci_realrnumberrmrHrGrnZ ipv4_addressrqZ_full_ipv6_addressZ_short_ipv6_addressrZ_mixed_ipv6_addressrZ ipv6_addressZ mac_addressrrzr|Z iso8601_dateZiso8601_datetimeuuidrKrJr~rr,r'r%rjrA _commasepitemrTrmrIZcomma_separated_listrxrVrrrrrsv""        c@seZdZddZddZdS)_lazyclasspropertycCs||_|j|_|j|_dSr)rrrrrrrr sz_lazyclassproperty.__init__csldkrt|tdr:tfddjddDr@i_|jj}|jkrb|j|<j|S)N_internc3s |]}jt|dgkVqdS)rN)rr)rZ superclassrrrrsz-_lazyclassproperty.__get__..r)rr1rv__mro__rrr)rrrattrnamerrr__get__s  z_lazyclassproperty.__get__N)rrrrrrrrrr src@sPeZdZdZgZeddZeddZeddZ edd Z ed d Z d S) ra A set of Unicode characters, for language-specific strings for ``alphas``, ``nums``, ``alphanums``, and ``printables``. A unicode_set is defined by a list of ranges in the Unicode character set, in a class attribute ``_ranges``, such as:: _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),] A unicode set can also be defined using multiple inheritance of other unicode sets:: class CJK(Chinese, Japanese, Korean): pass cCsZg}|jD]8}|tkrqD|jD] }|t|d|ddq q ddtt|DS)NrrrcSsg|] }t|qSrr:rrrrr4sz5unicode_set._get_chars_for_ranges..)rr_rangesrErrtr)rrZccZrrrrr_get_chars_for_ranges,s   z!unicode_set._get_chars_for_rangescCsdttj|S)z+all non-whitespace characters in this ranger)rrrrrrrrrrj6szunicode_set.printablescCsdttj|S)z'all alphabetic characters in this ranger)rfilterrisalpharrrrrrH;szunicode_set.alphascCsdttj|S)z*all numeric digit characters in this ranger)rrrisdigitrrrrrrf@szunicode_set.numscCs |j|jS)z)all alphanumeric characters in this range)rHrfrrrrrGEszunicode_set.alphanumsN) rrrrrrrrrjrHrfrGrrrrrs     c@seZdZdZdejfgZGdddeZGdddeZ GdddeZ Gd d d eZ Gd d d eZ Gd ddeZ GdddeZGdddeZGddde eeZGdddeZGdddeZGdddeZGdddeZdS)rzF A namespace class for defining common language unicode_sets. c@seZdZdZddgZdS)zpyparsing_unicode.Latin1z/Unicode set for Latin-1 Unicode Character Range)r~)NrrrrrrrrrLatin1Qsrc@seZdZdZdgZdS)zpyparsing_unicode.LatinAz/Unicode set for Latin-A Unicode Character Range)iNrrrrrLatinAUsrc@seZdZdZdgZdS)zpyparsing_unicode.LatinBz/Unicode set for Latin-B Unicode Character Range)iiONrrrrrLatinBYsrc@s6eZdZdZdddddddd d d d d dddddgZdS)zpyparsing_unicode.Greekz.Unicode set for Greek Unicode Character Ranges)ipi)ii)ii)i iE)iHiM)iPiW)iY)i[)i])i_i})ii)ii)ii)ii)ii)ii)iiNrrrrrGreek]s&rc@seZdZdZdgZdS)zpyparsing_unicode.Cyrillicz0Unicode set for Cyrillic Unicode Character Range)iiNrrrrrCyrillicesrc@seZdZdZddgZdS)zpyparsing_unicode.Chinesez/Unicode set for Chinese Unicode Character Range)Nii0i?0NrrrrrChineseisrc@sDeZdZdZgZGdddeZGdddeZGdddeZdS) zpyparsing_unicode.Japanesez`Unicode set for Japanese Unicode Character Range, combining Kanji, Hiragana, and Katakana rangesc@seZdZdZddgZdS)z pyparsing_unicode.Japanese.Kanjiz-Unicode set for Kanji Unicode Character Range)rirNrrrrrKanjiqsrc@seZdZdZdgZdS)z#pyparsing_unicode.Japanese.Hiraganaz0Unicode set for Hiragana Unicode Character Range)i@0i0NrrrrrHiraganausrc@seZdZdZdgZdS)z#pyparsing_unicode.Japanese.Katakanaz1Unicode set for Katakana Unicode Character Range)i0i0NrrrrrKatakanaysrN) rrrrrrrrrrrrrJapanesems rc@s eZdZdZddddddgZdS) zpyparsing_unicode.Koreanz.Unicode set for Korean Unicode Character Range)ii)ii)i01i1)i`i)iirNrrrrrKorean}src@seZdZdZdS)zpyparsing_unicode.CJKzTUnicode set for combined Chinese, Japanese, and Korean (CJK) Unicode Character RangeNrrrrrCJKsrc@seZdZdZddgZdS)zpyparsing_unicode.Thaiz,Unicode set for Thai Unicode Character Range)ii:)i?i[NrrrrrThaisrc@seZdZdZdddgZdS)zpyparsing_unicode.Arabicz.Unicode set for Arabic Unicode Character Range)ii)ii)iiNrrrrrArabicsrc@seZdZdZdgZdS)zpyparsing_unicode.Hebrewz.Unicode set for Hebrew Unicode Character Range)iiNrrrrrHebrewsrc@seZdZdZddgZdS)zpyparsing_unicode.Devanagariz2Unicode set for Devanagari Unicode Character Range)i i )iiNrrrrr DevanagarisrN)rrrrr maxunicoderrrrrrrrrrrrrrrrrrrrKs uالعربيةu中文uкириллицаuΕλληνικάuעִברִיתu 日本語u漢字u カタカナu ひらがなu 한국어u ไทยuदेवनागरी__main__Zselectfromr{r)r columnsrZtablesZcommandaK # '*' as column list and dotted table name select * from SYS.XYZZY # caseless match on "SELECT", and casts back to "select" SELECT * from XYZZY, ABC # list of column names, and mixed case SELECT keyword Select AA,BB,CC from Sys.dual # multiple tables Select A, B, C from Sys.dual, Table2 # invalid SELECT keyword - should fail Xelect A, B, C from Sys.dual # incomplete command - should fail Select # invalid column name - should fail Select ^^^ frox Sys.dual z] 100 -100 +100 3.14159 6.02e23 1e-12 z 100 FF z6 12345678-1234-5678-1234-567812345678 )NF)r)rF)N)FTF)T)r)T( rrrrrweakrefrrrIrr(rrr rxrrroperatorr itertools functoolsrr ImportErrorr_threadrZ threadingZcollections.abcr r r r rZ ordereddictr rrrrrrZ%warn_on_multiple_string_args_to_oneofr__all__r4 version_inforrbmaxsizerhrrchrr6rrrrrtreversedr rrvrrrrZmaxintZxrangerZ __builtin__rZfnamerrrrr rZascii_uppercaseZascii_lowercaserHrfrXrGr+rZ printablerjrrr0r2r4r6r9rrr5registerrMr^r[rrrrerr7r'r?rr*r'rvrNrr$rrrrBrrFr:r8rrArr"r&r%r=r<rDrCr3rr/r)rr1r r(r+rr,rEr r.r;r!r@rr#rr>r-rvrTrPr rcrbrrgrUr{r|r~rrWr]r\rurtrZ _escapedPuncZ_escapedHexCharZ_escapedOctCharZ _singleCharZ _charRangerr?rsrarprnrrxrVrPr_r`ryrSrrhrYrZr}rirRrrrmrwrdrzrIrkrKrJrrrhr7rOrorLrYrCrqrSrQrZrlrrrNrrrrrrrrrsetattrrrrrrrrrrZ selectTokenZ fromTokenZidentZ columnNameZcolumnNameListZ columnSpecZ tableNameZ tableNameListZ simpleSQLrgrrmrorr}rrrrsH            ?]  H    D! 'N E  KFyl{VO#K,:#Dvj-D0  $  W' *     4   0%   E&h~   (     ./J    &   site-packages/pip/_vendor/__pycache__/__init__.cpython-38.pyc000064400000005415151464452710020125 0ustar00U .e1@sdZddlmZddlZddlZddlZdZej ej e Z ddZ ereeje dejejdd<e de d e d e d e d e d e de de de de de de de de de de de de de de de de de de d e d!e d"e d#e d$e d%e d&e d'e d(e d)e d*e d+e d,e d-e d.e d/e d0e d1e d2e d3e d4e d5e d6e d7dS)8z pip._vendor is for vendoring dependencies of pip to prevent needing pip to depend on something external. Files inside of pip._vendor should be considered immutable and should only be updated to versions from upstream. )absolute_importNFcCsvdt|}zt|ttddWntk r8Yn:Xtj|tj|<|dd\}}t tj||tj|dS)Nz{0}.{1}r)level.) format__name__ __import__globalslocals ImportErrorsysmodulesrsplitsetattr)Z modulenameZ vendored_namebaseheadr8/usr/lib/python3.8/site-packages/pip/_vendor/__init__.pyvendoreds  rz*.whlZ cachecontrolZcoloramaZ contextlib2ZdistlibZdistroZhtml5libZsixz six.moveszsix.moves.urllibzsix.moves.urllib.parseZ packagingzpackaging.versionzpackaging.specifiersZpep517Z pkg_resourcesZprogressZpytomlZretryingZrequestszrequests.exceptionszrequests.packageszrequests.packages.urllib3z&requests.packages.urllib3._collectionsz$requests.packages.urllib3.connectionz(requests.packages.urllib3.connectionpoolz!requests.packages.urllib3.contribz*requests.packages.urllib3.contrib.ntlmpoolz+requests.packages.urllib3.contrib.pyopensslz$requests.packages.urllib3.exceptionsz requests.packages.urllib3.fieldsz"requests.packages.urllib3.filepostz"requests.packages.urllib3.packagesz/requests.packages.urllib3.packages.ordered_dictz&requests.packages.urllib3.packages.sixz5requests.packages.urllib3.packages.ssl_match_hostnamezErequests.packages.urllib3.packages.ssl_match_hostname._implementationz%requests.packages.urllib3.poolmanagerz!requests.packages.urllib3.requestz"requests.packages.urllib3.responsezrequests.packages.urllib3.utilz)requests.packages.urllib3.util.connectionz&requests.packages.urllib3.util.requestz'requests.packages.urllib3.util.responsez$requests.packages.urllib3.util.retryz#requests.packages.urllib3.util.ssl_z&requests.packages.urllib3.util.timeoutz"requests.packages.urllib3.util.urlZurllib3)__doc__Z __future__rZglobZos.pathosr Z DEBUNDLEDpathabspathdirname__file__Z WHEEL_DIRrjoinrrrrsr $site-packages/pip/_vendor/__pycache__/six.cpython-38.opt-1.pyc000064400000063434151464452710020135 0ustar00U .e~C@sdZddlmZddlZddlZddlZddlZddlZdZdZ ej ddkZ ej ddkZ ej dddkZ e refZefZefZeZeZejZn~efZeefZeejfZeZeZejd red ZnHGd d d eZ ze!e Wne"k red ZYn Xed Z[ ddZ#ddZ$GdddeZ%Gddde%Z&Gdddej'Z(Gddde%Z)GdddeZ*e*e+Z,Gddde(Z-e)dddd e)d!d"d#d$d!e)d%d"d"d&d%e)d'd(d#d)d'e)d*d(d+e)d,d"d#d-d,e)d.d/d/d0d.e)d1d/d/d.d1e)d2d3d4e)d5d(d#d6d5e)d7d(e rd8nd9d:e)d;d(d<e)d=d>d?d@e)d d de)dAdAdBe)dCdCdBe)dDdDdBe)d6d(d#d6d5e)dEd"d#dFdEe)dGd"d"dHdGe&d#d(e&dIdJe&dKdLe&dMdNdOe&dPdQdPe&dRdSdTe&dUdVdWe&dXdYdZe&d[d\d]e&d^d_d`e&dadbdce&dddedfe&dgdhdie&djdkdle&dmdndoe&dpdpdqe&drdrdqe&dsdsdqe&dtdtdue&dvdwe&dxdye&dzd{e&d|d}d|e&d~de&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&de+dde&de+dde&de+de+de&ddde&ddde&dddg@Z.ejdkrfe.e&ddg7Z.e.D]2Z/e0e-e/j1e/e2e/e&rje,3e/de/j1qj[/e.e-_.e-e+dZ4e,3e4dGddde(Z5e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d@dde)ddde)ddde)ddde)ddddd΃e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)dddgZ6e6D]Z/e0e5e/j1e/q[/e6e5_.e,3e5e+dddڡGdd܄de(Z7e)ddde)ddde)dddgZ8e8D]Z/e0e7e/j1e/q>[/e8e7_.e,3e7e+dddGddde(Z9e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d ddg#Z:e:D]Z/e0e9e/j1e/q[/e:e9_.e,3e9e+d d d Gd dde(Z;e)ddde)ddde)ddde)dddgZe>D]Z/e0e=e/j1e/q[/e>e=_.e,3e=e+dddGdddej'Z?e,3e?e+ddd d!Z@d"d#ZAe rd$ZBd%ZCd&ZDd'ZEd(ZFd)ZGn$d*ZBd+ZCd,ZDd-ZEd.ZFd/ZGzeHZIWn"eJk rd0d1ZIYnXeIZHzeKZKWn"eJk rd2d3ZKYnXe rFd4d5ZLejMZNd6d7ZOeZPn>d8d5ZLd9d:ZNd;d7ZOGd<d=d=eZPeKZKe#eLd>eQeBZReQeCZSeQeDZTeQeEZUeQeFZVeQeGZWe r(d?d@ZXdAdBZYdCdDZZdEdFZ[e\dGZ]e\dHZ^e\dIZ_nTdJd@ZXdKdBZYdLdDZZdMdFZ[e\dNZ]e\dOZ^e\dPZ_e#eXdQe#eYdRe#eZdSe#e[dTe rDdUdVZ`dWdXZaebZcddldZdededYjfZg[dehdZiejjZkelZmddlnZnenjoZoenjpZpdZZqej d[d[k r6d\Zrd]Zsn d^Zrd_Zsnjd`dVZ`dadXZaecZcebZgdbdcZidddeZketejuevZmddloZoeojoZoZpdfZqd\Zrd]Zse#e`dge#eadhdidZZwdjd^Zxdkd_Zye reze4j{dlZ|ddmdnZ}nddodpZ|e|dqej dddrk rLe|dsn.ej dddrk rne|dtn dudvZ~eze4j{dwdZedk rdxdyZej dddzk reZd{dyZe#e}d|ej dddk rejejfd}d~ZnejZddZddZdddZdddZdddZddZgZe+Zeddk rzge_ejreejD]4\ZZeej+dkrej1e+kreje=qq[[eje,dS(z6Utilities for writing code that runs on Python 2 and 3)absolute_importNz'Benjamin Peterson z1.12.0)rjavaic@seZdZddZdS)XcCsdS)Nlselfrr3/usr/lib/python3.8/site-packages/pip/_vendor/six.py__len__>sz X.__len__N)__name__ __module__ __qualname__r rrrr r<srlcCs ||_dS)z Add documentation to a function.N)__doc__)funcdocrrr _add_docKsrcCst|tj|S)z7Import module, returning the module after the last dot.) __import__sysmodulesnamerrr _import_modulePsrc@seZdZddZddZdS) _LazyDescrcCs ||_dSNrr rrrr __init__Xsz_LazyDescr.__init__cCsB|}t||j|zt|j|jWntk r<YnX|Sr)_resolvesetattrrdelattr __class__AttributeError)r objtpresultrrr __get__[sz_LazyDescr.__get__N)r rrrr&rrrr rVsrcs.eZdZdfdd ZddZddZZS) MovedModuleNcs2tt||tr(|dkr |}||_n||_dSr)superr'rPY3mod)r roldnewr!rr ris zMovedModule.__init__cCs t|jSr)rr*r rrr rrszMovedModule._resolvecCs"|}t||}t||||Sr)rgetattrr)r attr_modulevaluerrr __getattr__us  zMovedModule.__getattr__)N)r rrrrr2 __classcell__rrr-r r'gs r'cs(eZdZfddZddZgZZS) _LazyModulecstt|||jj|_dSr)r(r4rr!rrr-rr r~sz_LazyModule.__init__cCs ddg}|dd|jD7}|S)Nrr cSsg|] }|jqSrr).0r/rrr sz'_LazyModule.__dir__..)_moved_attributes)r Zattrsrrr __dir__sz_LazyModule.__dir__)r rrrr8r7r3rrr-r r4|s r4cs&eZdZdfdd ZddZZS)MovedAttributeNcsdtt||trH|dkr |}||_|dkr@|dkr<|}n|}||_n||_|dkrZ|}||_dSr)r(r9rr)r*r/)r rZold_modZnew_modZold_attrZnew_attrr-rr rszMovedAttribute.__init__cCst|j}t||jSr)rr*r.r/)r modulerrr rs zMovedAttribute._resolve)NN)r rrrrr3rrr-r r9sr9c@sVeZdZdZddZddZddZdd d Zd d Zd dZ ddZ ddZ e Z dS)_SixMetaPathImporterz A meta path importer to import six.moves and its submodules. This class implements a PEP302 finder and loader. It should be compatible with Python 2.5 and all existing versions of Python3 cCs||_i|_dSr)r known_modules)r Zsix_module_namerrr rsz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdSN.r<r)r r*Z fullnamesfullnamerrr _add_modulesz _SixMetaPathImporter._add_modulecCs|j|jd|Sr=r?r r@rrr _get_modulesz _SixMetaPathImporter._get_moduleNcCs||jkr|SdSr)r<)r r@pathrrr find_modules z _SixMetaPathImporter.find_modulecCs2z |j|WStk r,td|YnXdS)Nz!This loader does not know module )r<KeyError ImportErrorrBrrr Z __get_modules z!_SixMetaPathImporter.__get_modulecCsTz tj|WStk r YnX||}t|tr@|}n||_|tj|<|Sr)rrrF _SixMetaPathImporter__get_module isinstancer'r __loader__)r r@r*rrr load_modules     z _SixMetaPathImporter.load_modulecCst||dS)z Return true, if the named module is a package. We need this method to get correct spec objects with Python 3.4 (see PEP451) __path__)hasattrrHrBrrr is_packagesz_SixMetaPathImporter.is_packagecCs||dS)z;Return None Required, if is_package is implementedN)rHrBrrr get_codes z_SixMetaPathImporter.get_code)N) r rrrrrArCrErHrKrNrO get_sourcerrrr r;s  r;c@seZdZdZgZdS) _MovedItemszLazy loading of moved objectsN)r rrrrLrrrr rQsrQZ cStringIOioStringIOfilter itertoolsbuiltinsZifilter filterfalseZ ifilterfalseinputZ __builtin__Z raw_inputinternrmapimapgetcwdosZgetcwdugetcwdbZ getoutputZcommands subprocessrangeZxrangeZ reload_module importlibZimpreloadreduce functoolsZ shlex_quoteZpipesZshlexZquoteUserDict collectionsUserList UserStringzipZizip zip_longestZ izip_longestZ configparserZ ConfigParsercopyregZcopy_regZdbm_gnuZgdbmzdbm.gnuZ _dummy_threadZ dummy_threadZhttp_cookiejarZ cookielibzhttp.cookiejarZ http_cookiesZCookiez http.cookiesZ html_entitiesZhtmlentitydefsz html.entitiesZ html_parserZ HTMLParserz html.parserZ http_clientZhttplibz http.clientZemail_mime_basezemail.MIMEBasezemail.mime.baseZemail_mime_imagezemail.MIMEImagezemail.mime.imageZemail_mime_multipartzemail.MIMEMultipartzemail.mime.multipartZemail_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipartZemail_mime_textzemail.MIMETextzemail.mime.textZBaseHTTPServerz http.serverZ CGIHTTPServerZSimpleHTTPServerZcPicklepickleZqueueZQueuereprlibreprZ socketserverZ SocketServer_threadthreadZtkinterZTkinterZtkinter_dialogZDialogztkinter.dialogZtkinter_filedialogZ FileDialogztkinter.filedialogZtkinter_scrolledtextZ ScrolledTextztkinter.scrolledtextZtkinter_simpledialogZ SimpleDialogztkinter.simpledialogZ tkinter_tixZTixz tkinter.tixZ tkinter_ttkZttkz tkinter.ttkZtkinter_constantsZ Tkconstantsztkinter.constantsZ tkinter_dndZTkdndz tkinter.dndZtkinter_colorchooserZtkColorChooserztkinter.colorchooserZtkinter_commondialogZtkCommonDialogztkinter.commondialogZtkinter_tkfiledialogZ tkFileDialogZ tkinter_fontZtkFontz tkinter.fontZtkinter_messageboxZ tkMessageBoxztkinter.messageboxZtkinter_tksimpledialogZtkSimpleDialogZ urllib_parsez.moves.urllib_parsez urllib.parseZ urllib_errorz.moves.urllib_errorz urllib.errorZurllibz .moves.urllibZurllib_robotparser robotparserzurllib.robotparserZ xmlrpc_clientZ xmlrpclibz xmlrpc.clientZ xmlrpc_serverZSimpleXMLRPCServerz xmlrpc.serverZwin32winreg_winregzmoves.z.movesmovesc@seZdZdZdS)Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseNr rrrrrrr ruBsruZ ParseResultZurlparseZ SplitResultZparse_qsZ parse_qslZ urldefragZurljoinZurlsplitZ urlunparseZ urlunsplitZ quote_plusZunquoteZ unquote_plusZunquote_to_bytesZ urlencodeZ splitqueryZsplittagZ splituserZ splitvalueZ uses_fragmentZ uses_netlocZ uses_paramsZ uses_queryZ uses_relativemoves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNrvrrrr rxlsrxZURLErrorZurllib2Z HTTPErrorZContentTooShortErrorz.moves.urllib.errormoves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNrvrrrr rzsrzZurlopenzurllib.requestZinstall_openerZ build_openerZ pathname2urlZ url2pathnameZ getproxiesZRequestZOpenerDirectorZHTTPDefaultErrorHandlerZHTTPRedirectHandlerZHTTPCookieProcessorZ ProxyHandlerZ BaseHandlerZHTTPPasswordMgrZHTTPPasswordMgrWithDefaultRealmZAbstractBasicAuthHandlerZHTTPBasicAuthHandlerZProxyBasicAuthHandlerZAbstractDigestAuthHandlerZHTTPDigestAuthHandlerZProxyDigestAuthHandlerZ HTTPHandlerZ HTTPSHandlerZ FileHandlerZ FTPHandlerZCacheFTPHandlerZUnknownHandlerZHTTPErrorProcessorZ urlretrieveZ urlcleanupZ URLopenerZFancyURLopenerZ proxy_bypassZparse_http_listZparse_keqv_listz.moves.urllib.requestmoves.urllib_requestzmoves.urllib.requestc@seZdZdZdS) Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNrvrrrr r|sr|Zaddbasezurllib.responseZ addclosehookZaddinfoZ addinfourlz.moves.urllib.responsemoves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNrvrrrr r~sr~ZRobotFileParserz.moves.urllib.robotparsermoves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZedZedZedZ edZ edZ ddZ d S) Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacerwryr{r}rcCsdddddgS)Nparseerrorrequestresponserqrr rrr r8szModule_six_moves_urllib.__dir__N) r rrrrL _importerrCrrrrrqr8rrrr rs     rz moves.urllibcCstt|j|dS)zAdd an item to six.moves.N)rrQr)Zmoverrr add_movesrc CsXztt|WnDtk rRz tj|=Wn"tk rLtd|fYnXYnXdS)zRemove item from six.moves.zno such move, %rN)r rQr"rt__dict__rFrrrr remove_moves r__func____self__ __closure____code__ __defaults__ __globals__im_funcZim_selfZ func_closureZ func_codeZ func_defaultsZ func_globalscCs|Sr)next)itrrr advance_iteratorsrcCstddt|jDS)Ncss|]}d|jkVqdS)__call__N)r)r5klassrrr szcallable..)anytype__mro__)r#rrr callablesrcCs|SrrZunboundrrr get_unbound_functionsrcCs|Srrrclsrrr create_unbound_method#srcCs|jSr)rrrrr r(scCst|||jSr)types MethodTyper!)rr#rrr create_bound_method+srcCst|d|Sr)rrrrrr r.sc@seZdZddZdS)IteratorcCst||Sr)r__next__r rrr r3sz Iterator.nextN)r rrrrrrr r1srz3Get the function out of a possibly unbound functioncKst|jf|Sr)iterkeysdkwrrr iterkeysDsrcKst|jf|Sr)rvaluesrrrr itervaluesGsrcKst|jf|Sr)ritemsrrrr iteritemsJsrcKst|jf|Sr)rZlistsrrrr iterlistsMsrrrrcKs |jf|Sr)rrrrr rVscKs |jf|Sr)rrrrr rYscKs |jf|Sr)rrrrr r\scKs |jf|Sr)rrrrr r_sviewkeys viewvalues viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs |dS)Nzlatin-1)encodesrrr bqsrcCs|Srrrrrr utsrz>BassertCountEqualZassertRaisesRegexpZassertRegexpMatchesassertRaisesRegex assertRegexcCs|Srrrrrr rscCst|dddS)Nz\\z\\\\Zunicode_escape)unicodereplacerrrr rscCs t|dS)Nrord)Zbsrrr byte2intsrcCs t||Srr)Zbufirrr indexbytessrZassertItemsEqualz Byte literalz Text literalcOst|t||Sr)r._assertCountEqualr argskwargsrrr rscOst|t||Sr)r._assertRaisesRegexrrrr rscOst|t||Sr)r. _assertRegexrrrr rsexeccCs:z*|dkr|}|j|k r$|||W5d}d}XdSr) __traceback__with_traceback)r$r1tbrrr reraises  rcCsB|dkr*td}|j}|dkr&|j}~n |dkr6|}tddS)zExecute code in a namespace.Nrzexec _code_ in _globs_, _locs_)r _getframe f_globalsf_localsr)Z_code_Z_globs_Z_locs_framerrr exec_s rzedef reraise(tp, value, tb=None): try: raise tp, value, tb finally: tb = None )rrzdef raise_from(value, from_value): try: if from_value is None: raise value raise value from from_value finally: value = None zrdef raise_from(value, from_value): try: raise value from from_value finally: value = None cCs|dSrr)r1Z from_valuerrr raise_fromsrprintc s.|dtjdkrdSfdd}d}|dd}|dk r`t|trNd}nt|ts`td|d d}|dk rt|trd}nt|tstd |rtd |s|D]}t|trd}qq|rtd }td }nd }d }|dkr|}|dkr|}t|D] \} }| r||||q||dS)z4The new-style print function for Python 2.4 and 2.5.fileNcsdt|tst|}ttrVt|trVjdk rVtdd}|dkrHd}|j|}|dS)Nerrorsstrict) rI basestringstrrrencodingr.rwrite)datarfprr rs   zprint_..writeFsepTzsep must be None or a stringendzend must be None or a stringz$invalid keyword arguments to print()  )poprstdoutrIrr TypeError enumerate) rrrZ want_unicoderrargnewlineZspacerrrr print_sL          r)rrcOs<|dtj}|dd}t|||r8|dk r8|dS)NrflushF)getrrr_printr)rrrrrrr rs    zReraise an exception.csfdd}|S)Ncst|}|_|Sr)rdwraps __wrapped__)fassignedupdatedwrappedrr wrapper*szwraps..wrapperr)rrrrrrr r(srcs&Gfdddt}t|ddiS)z%Create a base class with a metaclass.cs,eZdZfddZefddZdS)z!with_metaclass..metaclasscs ||Srr)rr this_basesrbasesmetarr __new__:sz)with_metaclass..metaclass.__new__cs |Sr) __prepare__)rrrrrr r=sz-with_metaclass..metaclass.__prepare__N)r rrr classmethodrrrrr metaclass8srZtemporary_classr)rr)rrrrrr with_metaclass3srcsfdd}|S)z6Class decorator for creating a class with a metaclass.cs||j}|d}|dk r@t|tr,|g}|D]}||q0|dd|ddt|drl|j|d<|j|j |S)N __slots__r __weakref__r) rcopyrrIrrrMrr __bases__)rZ orig_varsslotsZ slots_varrrr rEs        zadd_metaclass..wrapperr)rrrrr add_metaclassCs rutf-8rcCs8t|tr|||St|tr$|Stdt|dS)zCoerce **s** to six.binary_type. For Python 2: - `unicode` -> encoded to `str` - `str` -> `str` For Python 3: - `str` -> encoded to `bytes` - `bytes` -> `bytes` not expecting type '%s'N)rI text_typer binary_typerrrrrrrr ensure_binaryUs   r cCsXt|ttfstdt|tr:t|tr:|||}ntrTt|trT|||}|S)zCoerce *s* to `str`. For Python 2: - `unicode` -> encoded to `str` - `str` -> `str` For Python 3: - `str` -> `str` - `bytes` -> decoded to `str` r) rIrrrrPY2rr)decoderrrr ensure_strhs  r cCs8t|tr|||St|tr$|Stdt|dS)zCoerce *s* to six.text_type. For Python 2: - `unicode` -> `unicode` - `str` -> `unicode` For Python 3: - `str` -> `str` - `bytes` -> decoded to `str` rN)rIrr rrrrrrr ensure_text|s   r cCs2tr.d|jkrtd|j|j|_dd|_|S)a A decorator that defines __unicode__ and __str__ methods under Python 2. Under Python 3 it does nothing. To support Python 2 and 3 with a single code base, define a __str__ method returning text and apply this decorator to the class. __str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|dS)Nr) __unicode__rr rrr z-python_2_unicode_compatible..)r r ValueErrorr rr)rrrr python_2_unicode_compatibles  r__spec__)N)NN)rr)rr)rr)rZ __future__rrdrUoperatorrr __author__ __version__ version_infor r)ZPY34rZ string_typesintZ integer_typesrZ class_typesrbytesrmaxsizeZMAXSIZErZlongZ ClassTyperplatform startswithobjectrlen OverflowErrorrrrr' ModuleTyper4r9r;r rrQr7r/rrrIrArtruZ_urllib_parse_moved_attributesrxZ_urllib_error_moved_attributesrzZ _urllib_request_moved_attributesr|Z!_urllib_response_moved_attributesr~Z$_urllib_robotparser_moved_attributesrrrZ _meth_funcZ _meth_selfZ _func_closureZ _func_codeZ_func_defaultsZ _func_globalsrr NameErrorrrrrrr attrgetterZget_method_functionZget_method_selfZget_function_closureZget_function_codeZget_function_defaultsZget_function_globalsrrrr methodcallerrrrrrchrZunichrstructStructpackZint2byte itemgetterrgetitemrrZ iterbytesrRrSBytesIOrrrpartialr[rrrrr.rVrrrrrWRAPPER_ASSIGNMENTSWRAPPER_UPDATESrrrr r r rrL __package__globalsrrsubmodule_search_locations meta_pathrrZimporterappendrrrr s     >                                        F                                                                  %                                                  5      site-packages/pip/_vendor/__pycache__/retrying.cpython-38.opt-1.pyc000064400000017506151464452710021174 0ustar00U .e&@slddlZddlmZddlZddlZddlZdZddZGdddeZ Gdd d eZ Gd d d e Z dS) N)sixi?csBtdkr,tdr,dd}|dSfdd}|SdS)z Decorator function that instantiates the Retrying object @param *dargs: positional arguments passed to Retrying object @param **dkw: keyword arguments passed to the Retrying object rcstfdd}|S)Ncstjf||SNRetryingcallargskwf8/usr/lib/python3.8/site-packages/pip/_vendor/retrying.py wrapped_f$sz-retry..wrap_simple..wrapped_frwrapsr rr r r wrap_simple"szretry..wrap_simplecstfdd}|S)Ncstjf||Srrr)dargsdkwr r rr/sz&retry..wrap..wrapped_frrrrr rwrap-szretry..wrapN)lencallable)rrrrr rrretrys  rc@sneZdZdddZddZddZd d Zd d Zd dZddZ ddZ ddZ ddZ ddZ ddZdS)rNFcs |dkr dn||_|dkrdn||_|dkr0dn||_|dkrBdn||_|dkrTdn||_|dkrfdn||_| dkrxdn| |_| dkrdn| |_| dkrtn| |_ |dkrdn||_ g|dk r̈ |j |dk r |j |dk r||_n&|dkr fdd|_n t|||_ddg|dk r6 |j|dk sJ|dk rV |j|dk sj| dk rv |j| dk s| dk r |j|dk r||_n&|dkr‡fd d|_n t|||_| dkr|j|_n| |_| dkr|j|_n| |_||_dS) NdrrcstfddDS)Nc3s|]}|VqdSrr .0r ZattemptsZdelayr r as6Retrying.__init__....)anyr ) stop_funcsr raz#Retrying.__init__..c_sdS)Nrr )r kwargsr r rr%hr&cstfddDS)Nc3s|]}|VqdSrr rr r rr!ysr")maxr ) wait_funcsr rr%yr&)_stop_max_attempt_number_stop_max_delay _wait_fixed_wait_random_min_wait_random_max_wait_incrementing_start_wait_incrementing_increment_wait_exponential_multiplierMAX_WAIT_wait_exponential_max_wait_jitter_maxappendstop_after_attemptstop_after_delaystopgetattr fixed_sleep random_sleepincrementing_sleepexponential_sleepwait always_reject_retry_on_exception never_reject_retry_on_result_wrap_exception)selfr8r>Zstop_max_attempt_numberZstop_max_delayZ wait_fixedZwait_random_minZwait_random_maxZwait_incrementing_startZwait_incrementing_incrementZwait_exponential_multiplierZwait_exponential_maxZretry_on_exceptionZretry_on_resultwrap_exceptionZ stop_funcZ wait_funcZwait_jitter_maxr )r$r)r__init__:sR                 zRetrying.__init__cCs ||jkS)z;Stop after the previous attempt >= stop_max_attempt_number.)r*rDprevious_attempt_numberdelay_since_first_attempt_msr r rr6szRetrying.stop_after_attemptcCs ||jkS)z=Stop after the time from the first attempt >= stop_max_delay.)r+rGr r rr7szRetrying.stop_after_delaycCsdS)z#Don't sleep at all before retrying.rr rGr r rno_sleepszRetrying.no_sleepcCs|jS)z0Sleep a fixed amount of time between each retry.)r,rGr r rr:szRetrying.fixed_sleepcCst|j|jS)zISleep a random amount of time between wait_random_min and wait_random_max)randomZrandintr-r.rGr r rr;szRetrying.random_sleepcCs$|j|j|d}|dkr d}|S)z Sleep an incremental amount of time after each attempt, starting at wait_incrementing_start and incrementing by wait_incrementing_increment rr)r/r0)rDrHrIresultr r rr<szRetrying.incrementing_sleepcCs2d|}|j|}||jkr"|j}|dkr.d}|S)Nr)r1r3)rDrHrIZexprLr r rr=s  zRetrying.exponential_sleepcCsdS)NFr rDrLr r rrAszRetrying.never_rejectcCsdS)NTr rNr r rr?szRetrying.always_rejectcCs4d}|jr |||jdO}n|||jO}|S)NFr) has_exceptionr@valuerB)rDattemptZrejectr r r should_rejects zRetrying.should_rejectc Ostttd}d}zt||||d}Wn t}t||d}YnX||sf||jStttd|}| ||r|js|j r|qt |n<| ||} |j rt|j } | td| } t| d|d7}qdS)NrrFTrg@@)introundtimeAttemptsysexc_inforRgetrCr8rO RetryErrorr>r4rKr(sleep) rDfnr r'Z start_timeattempt_numberrQtbrIr[Zjitterr r rrs(       z Retrying.call)NNNNNNNNNNNNNFNNN)__name__ __module__ __qualname__rFr6r7rJr:r;r<r=rAr?rRrr r r rr8s: S   rc@s*eZdZdZddZd ddZddZd S) rVz An Attempt encapsulates a call to a target function that may end as a normal return value from the function or an Exception depending on what occurred during the execution. cCs||_||_||_dSr)rPr]rO)rDrPr]rOr r rrFszAttempt.__init__FcCs@|jr6|rt|qs *!site-packages/pip/_vendor/__pycache__/ipaddress.cpython-38.opt-1.pyc000064400000176207151464452710021313 0ustar00U .e7@s`dZddlmZddlZddlZdZefZz eefZWne k rJYnXze Z Wne k rle Z YnXddkrddZ nddZ z ejZWnek rdd ZYnXd d Zeed rd dZnddZdNddZGdddeZdZdZGdddeZGdddeZddZdOddZd d!Zd"d#Zd$d%Z d&d'Z!d(d)Z"d*d+Z#d,d-Z$d.d/Z%d0d1Z&d2d3Z'Gd4d5d5eZ(Gd6d7d7e(Z)Gd8d9d9e(Z*Gd:d;d;eZ+Gdd?d?e,Z-Gd@dAdAe+e*Z.GdBdCdCeZ/e/e,_0GdDdEdEeZ1GdFdGdGe1e)Z2GdHdIdIe2Z3GdJdKdKe1e*Z4GdLdMdMeZ5e5e2_0dS)PzA fast, lightweight IPv4/IPv6 manipulation library in Python. This library is used to create/poke/manipulate IPv4 and IPv6 addresses and networks. )unicode_literalsNz1.0.22cCs|SNZbytrr9/usr/lib/python3.8/site-packages/pip/_vendor/ipaddress.py_compat_bytes_to_byte_valssrcCsdd|DS)NcSsg|]}td|dqS)!Br)structunpack.0brrr #sz._compat_bytes_to_byte_vals..rrrrrr"scCsd}|D]}|d>|}q|S)Nrr)Zbytvals endianessresZbvrrr_compat_int_from_byte_vals'srcCst|dkr.|dks|dkr"tdtd|S|dkrj|dksJ|ddkrTtd td |d ?|d @StdS) Nrlz(integer out of range for 'I' format codes!Iz)integer out of range for 'QQ' format codes!QQ@l)r errorpackNotImplementedError)ZintvalZlengthrrrr_compat_to_bytes0s   r bit_lengthcCs|Sr)r)irrr_compat_bit_lengthAsrcCs&tD]}||?dkr|SqdSNr) itertoolscount)rrrrrrDs  ccs |}||kr|V||7}qdSrr)startendsteprrrr _compat_rangeJsr&c@s@eZdZdZddZddZddZdd Zd d Zd d Z dS)_TotalOrderingMixinrcCstdSrrselfotherrrr__eq__Ysz_TotalOrderingMixin.__eq__cCs||}|tkrtS| Sr)r,NotImplemented)r*r+equalrrr__ne__\s z_TotalOrderingMixin.__ne__cCstdSrr(r)rrr__lt__bsz_TotalOrderingMixin.__lt__cCs$||}|tks|s ||S|Srr0r-r,r*r+lessrrr__le__es   z_TotalOrderingMixin.__le__cCs6||}|tkrtS||}|tkr,tS|p2| Srr1)r*r+r3r.rrr__gt__ks  z_TotalOrderingMixin.__gt__cCs||}|tkrtS| Sr)r0r-r2rrr__ge__ts z_TotalOrderingMixin.__ge__N) __name__ __module__ __qualname__ __slots__r,r/r0r4r5r6rrrrr'Rs r' rc@seZdZdZdS)AddressValueErrorz%A Value Error related to the address.Nr7r8r9__doc__rrrrr<sr<c@seZdZdZdS)NetmaskValueErrorz%A Value Error related to the netmask.Nr=rrrrr?sr?c Csnz t|WSttfk r"YnXz t|WSttfk rFYnXt|tr^td|td|dS)aTake an IP string/int and return an object of the correct type. Args: address: A string or integer, the IP address. Either IPv4 or IPv6 addresses may be supplied; integers less than 2**32 will be considered to be IPv4 by default. Returns: An IPv4Address or IPv6Address object. Raises: ValueError: if the *address* passed isn't either a v4 or a v6 address zx%r does not appear to be an IPv4 or IPv6 address. Did you pass in a bytes (str in Python 2) instead of a unicode object?z0%r does not appear to be an IPv4 or IPv6 addressN) IPv4Addressr<r? IPv6Address isinstancebytes ValueErroraddressrrr ip_addresss"   rGTc Csrz t||WSttfk r$YnXz t||WSttfk rJYnXt|trbtd|td|dS)aTake an IP string/int and return an object of the correct type. Args: address: A string or integer, the IP network. Either IPv4 or IPv6 networks may be supplied; integers less than 2**32 will be considered to be IPv4 by default. Returns: An IPv4Network or IPv6Network object. Raises: ValueError: if the string passed isn't either a v4 or a v6 address. Or if the network has host bits set. zx%r does not appear to be an IPv4 or IPv6 network. Did you pass in a bytes (str in Python 2) instead of a unicode object?z0%r does not appear to be an IPv4 or IPv6 networkN) IPv4Networkr<r? IPv6NetworkrBrCrD)rFstrictrrr ip_networks"   rKc CsXz t|WSttfk r"YnXz t|WSttfk rFYnXtd|dS)agTake an IP string/int and return an object of the correct type. Args: address: A string or integer, the IP address. Either IPv4 or IPv6 addresses may be supplied; integers less than 2**32 will be considered to be IPv4 by default. Returns: An IPv4Interface or IPv6Interface object. Raises: ValueError: if the string passed isn't either a v4 or a v6 address. Notes: The IPv?Interface classes describe an Address on a particular Network, so they're basically a combination of both the Address and Network classes. z2%r does not appear to be an IPv4 or IPv6 interfaceN) IPv4Interfacer<r? IPv6InterfacerDrErrr ip_interfaces  rNc Cs6zt|ddWStjtfk r0tdYnXdS)a`Represent an address as 4 packed bytes in network (big-endian) order. Args: address: An integer representation of an IPv4 IP address. Returns: The integer address packed as 4 bytes in network (big-endian) order. Raises: ValueError: If the integer is negative or too large to be an IPv4 IP address. rbigz&Address negative or too large for IPv4Nrr r OverflowErrorrDrErrrv4_int_to_packedsrRc Cs6zt|ddWStjtfk r0tdYnXdS)zRepresent an address as 16 packed bytes in network (big-endian) order. Args: address: An integer representation of an IPv6 IP address. Returns: The integer address packed as 16 bytes in network (big-endian) order. rrOz&Address negative or too large for IPv6NrPrErrrv6_int_to_packeds rScCs*t|d}t|dkr&td||S)zAHelper to split the netmask and raise AddressValueError if needed/rzOnly one '/' permitted in %r) _compat_strsplitlenr<)rFaddrrrr_split_optional_netmasks  rYccsNt|}t|}}|D]&}|j|jdkr:||fV|}|}q||fVdS)zFind a sequence of sorted deduplicated IPv#Address. Args: addresses: a list of IPv#Address objects. Yields: A tuple containing the first and last IP addresses in the sequence. r"N)iternext_ip) addressesitfirstlastiprrr_find_address_ranges   rbcCs$|dkr |St|t||d@S)zCount the number of zero bits on the right hand side. Args: number: an integer. bits: maximum number of bits to count. Returns: The number of zero bits on the right hand side of the number. rr")minr)Znumberbitsrrr_count_righthand_zero_bits2s reccst|trt|tstd|j|jkr8td||f||krHtd|jdkrXt}n|jdkrht}ntd|j}|j}|j}||krt t ||t ||dd}||||f}|V|d|>7}|d|j krqqdS) aSummarize a network range given the first and last IP addresses. Example: >>> list(summarize_address_range(IPv4Address('192.0.2.0'), ... IPv4Address('192.0.2.130'))) ... #doctest: +NORMALIZE_WHITESPACE [IPv4Network('192.0.2.0/25'), IPv4Network('192.0.2.128/31'), IPv4Network('192.0.2.130/32')] Args: first: the first IPv4Address or IPv6Address in the range. last: the last IPv4Address or IPv6Address in the range. Returns: An iterator of the summarized IPv(4|6) network objects. Raise: TypeError: If the first and last objects are not IP addresses. If the first and last objects are not the same version. ValueError: If the last object is not greater than the first. If the version of the first address is not 4 or 6. z1first and last must be IP addresses, not networks%%s and %s are not of the same versionz*last IP address must be greater than firstrzunknown IP versionr"N) rB _BaseAddress TypeErrorversionrDrHrI_max_prefixlenr\rcrer _ALL_ONES)r_r`raZip_bitsZ first_intZlast_intZnbitsnetrrrsummarize_address_rangeBs8      rnccst|}i}|rV|}|}||}|dkr<|||<q ||kr ||=||q d}t|D]$}|dk r|j|jkrqf|V|}qfdS)auLoops through the addresses, collapsing concurrent netblocks. Example: ip1 = IPv4Network('192.0.2.0/26') ip2 = IPv4Network('192.0.2.64/26') ip3 = IPv4Network('192.0.2.128/26') ip4 = IPv4Network('192.0.2.192/26') _collapse_addresses_internal([ip1, ip2, ip3, ip4]) -> [IPv4Network('192.0.2.0/24')] This shouldn't be called directly; it is called via collapse_addresses([]). Args: addresses: A list of IPv4Network's or IPv6Network's Returns: A list of IPv4Network's or IPv6Network's depending on what we were passed. N)listpopsupernetgetappendsortedvaluesbroadcast_address)r]Zto_mergesubnetsrmrqZexistingr`rrr_collapse_addresses_internalys$    rxc Cs0g}g}g}|D]}t|trR|rF|dj|jkrFtd||df||q|j|jkr|r|dj|jkrtd||dfz||jWqtk r||j YqXq|r|dj|jkrtd||df||qt t |}|r$t |D]\}}| t||qt||S)aCollapse a list of IP objects. Example: collapse_addresses([IPv4Network('192.0.2.0/25'), IPv4Network('192.0.2.128/25')]) -> [IPv4Network('192.0.2.0/24')] Args: addresses: An iterator of IPv4Network or IPv6Network objects. Returns: An iterator of the collapsed IPv(4|6)Network objects. Raises: TypeError: If passed a list of mixed version objects. rf)rBrh_versionrirs _prefixlenrkraAttributeErrornetwork_addressrtsetrbextendrnrx)r]ZaddrsZipsZnetsrar_r`rrrcollapse_addressess@     rcCs(t|tr|St|tr$|StS)a2Return a key suitable for sorting between networks and addresses. Address and Network objects are not sortable by default; they're fundamentally different so the expression IPv4Address('192.0.2.0') <= IPv4Network('192.0.2.0/24') doesn't make any sense. There are some times however, where you may wish to have ipaddress sort these for you anyway. If you need to do this, you can use this function as the key= argument to sorted(). Args: obj: either a Network or Address object. Returns: appropriate key. )rB _BaseNetwork_get_networks_keyrh_get_address_keyr-)objrrrget_mixed_type_keys   rc@seZdZdZdZeddZeddZeddZed d Z d d Z d dZ e ddZ e ddZe ddZe ddZe ddZddZdS)_IPAddressBasezThe mother class.rcCs|S)z:Return the longhand version of the IP address as a string.)_explode_shorthand_ip_stringr*rrrexplodedsz_IPAddressBase.explodedcCst|S)z;Return the shorthand version of the IP address as a string.rUrrrr compressedsz_IPAddressBase.compressedcCs|S)aIThe name of the reverse DNS pointer for the IP address, e.g.: >>> ipaddress.ip_address("127.0.0.1").reverse_pointer '1.0.0.127.in-addr.arpa' >>> ipaddress.ip_address("2001:db8::1").reverse_pointer '1.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.8.b.d.0.1.0.0.2.ip6.arpa' )_reverse_pointerrrrrreverse_pointer s z_IPAddressBase.reverse_pointercCsdt|f}t|dS)Nz%200s has no version specifiedtyperr*msgrrrrjsz_IPAddressBase.versioncCsF|dkrd}t|||jf||jkrBd}t|||j|jfdS)Nrz-%d (< 0) is not permitted as an IPv%d addressz2%d (>= 2**%d) is not permitted as an IPv%d address)r<rzrlrk)r*rFrrrr_check_int_addresss  z!_IPAddressBase._check_int_addresscCs.t|}||kr*d}t|||||jfdS)Nz~%r (len %d != %d) is not permitted as an IPv%d address. Did you pass in a bytes (str in Python 2) instead of a unicode object?)rWr<rz)r*rFZ expected_lenZ address_lenrrrr_check_packed_address"sz$_IPAddressBase._check_packed_addresscCs|j|j|?AS)zTurn the prefix length into a bitwise netmask Args: prefixlen: An integer, the prefix length. Returns: An integer. )rl)cls prefixlenrrr_ip_int_from_prefix,s z"_IPAddressBase._ip_int_from_prefixc Cs\t||j}|j|}||?}d|>d}||krX|jd}t||d}d}t|||S)aReturn prefix length from the bitwise netmask. Args: ip_int: An integer, the netmask in expanded bitwise format Returns: An integer, the prefix length. Raises: ValueError: If the input intermingles zeroes & ones r"rrOz&Netmask pattern %r mixes zeroes & ones)rerkrrD) rip_intZtrailing_zeroesrZ leading_onesZall_onesZbyteslenZdetailsrrrr_prefix_from_ip_int9s      z"_IPAddressBase._prefix_from_ip_intcCsd|}t|dS)Nz%r is not a valid netmask)r?)rZ netmask_strrrrr_report_invalid_netmaskRsz&_IPAddressBase._report_invalid_netmaskcCshtj|s||z t|}Wntk r@||YnXd|krX|jksdn|||S)a Return prefix length from a numeric string Args: prefixlen_str: The string to be converted Returns: An integer, the prefix length. Raises: NetmaskValueError: If the input is not a valid netmask r)_BaseV4_DECIMAL_DIGITS issupersetrintrDrk)rZ prefixlen_strrrrr_prefix_from_prefix_stringWs    z)_IPAddressBase._prefix_from_prefix_stringcCsz||}Wntk r,||YnXz ||WStk rNYnX||jN}z ||WStk r||YnXdS)aTurn a netmask/hostmask string into a prefix length Args: ip_str: The netmask/hostmask to be converted Returns: An integer, the prefix length. Raises: NetmaskValueError: If the input is not a valid netmask/hostmask N)_ip_int_from_stringr<rrrDrl)rip_strrrrr_prefix_from_ip_stringps   z%_IPAddressBase._prefix_from_ip_stringcCs|jt|ffSr) __class__rUrrrr __reduce__sz_IPAddressBase.__reduce__N)r7r8r9r>r:propertyrrrrjrr classmethodrrrrrrrrrrrs.           !rc@sdeZdZdZdZddZddZddZd d Zd d Z d dZ ddZ ddZ ddZ ddZdS)rhzA generic IP object. This IP class contains the version independent methods which are used by single IP addresses. rcCs|jSrr\rrrr__int__sz_BaseAddress.__int__cCs8z|j|jko|j|jkWStk r2tYSXdSr)r\rzr|r-r)rrrr,s   z_BaseAddress.__eq__cCs`t|tstSt|ts(td||f|j|jkrDtd||f|j|jkr\|j|jkSdSN"%s and %s are not of the same typerfF)rBrr-rhrirzr\r)rrrr0s     z_BaseAddress.__lt__cCs t|tstS|t||SrrB_compat_int_typesr-rrr)rrr__add__s z_BaseAddress.__add__cCs t|tstS|t||Srrr)rrr__sub__s z_BaseAddress.__sub__cCsd|jjt|fSNz%s(%r)rr7rUrrrr__repr__sz_BaseAddress.__repr__cCst||jSr)rU_string_from_ip_intr\rrrr__str__sz_BaseAddress.__str__cCsttt|jSr)hashhexrr\rrrr__hash__sz_BaseAddress.__hash__cCs |j|fSrrzrrrrrsz_BaseAddress._get_address_keycCs|j|jffSr)rr\rrrrrsz_BaseAddress.__reduce__N)r7r8r9r>r:rr,r0rrrrrrrrrrrrhsrhc@sdeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ ddZ ddZeddZeddZeddZeddZed d!Zed"d#Zed$d%Zed&d'Zd(d)Zd*d+Zd,d-ZdHd0d1ZdId2d3Zed4d5Zed6d7Zd8d9Z d:d;Z!edd?Z#ed@dAZ$edBdCZ%edDdEZ&edFdGZ'd/S)Jrz~A generic IP network object. This IP class contains the version independent methods which are used by networks. cCs i|_dSr)_cache)r*rFrrr__init__sz_BaseNetwork.__init__cCsd|jjt|fSrrrrrrrsz_BaseNetwork.__repr__cCsd|j|jfSN%s/%d)r}rrrrrrsz_BaseNetwork.__str__ccs8t|j}t|j}t|d|D]}||Vq"dS)zGenerate Iterator over usable hosts in a network. This is like __iter__ except it doesn't return the network or broadcast addresses. r"Nrr}rvr&_address_classr*network broadcastxrrrhostss  z_BaseNetwork.hostsccs8t|j}t|j}t||dD]}||Vq"dSNr"rrrrr__iter__s  z_BaseNetwork.__iter__cCslt|j}t|j}|dkr>|||kr0td|||S|d7}|||krZtd|||SdS)Nrzaddress out of ranger")rr}rv IndexErrorr)r*nrrrrr __getitem__s    z_BaseNetwork.__getitem__cCsxt|tstSt|ts(td||f|j|jkrDtd||f|j|jkr\|j|jkS|j|jkrt|j|jkSdSr)rBrr-rrirzr}netmaskr)rrrr0s"       z_BaseNetwork.__lt__cCsLz.|j|jko,|j|jko,t|jt|jkWStk rFtYSXdSr)rzr}rrr|r-r)rrrr,s  z_BaseNetwork.__eq__cCstt|jt|jASr)rrr}rrrrrrsz_BaseNetwork.__hash__cCsL|j|jkrdSt|trdSt|jt|jkoBt|jkSSdSNF)rzrBrrr}r\rvr)rrr __contains__s  z_BaseNetwork.__contains__cCs(|j|kp&|j|kp&|j|kp&|j|kS)z*Tell if self is partly contained in other.)r}rvr)rrroverlaps*s    z_BaseNetwork.overlapscCs<|jd}|dkr8|t|jt|jB}||jd<|S)Nrv)rrrrrr}hostmaskr*rrrrrv1s   z_BaseNetwork.broadcast_addresscCs8|jd}|dkr4|t|j|jA}||jd<|S)Nr)rrrrrrrlrrrrr:s   z_BaseNetwork.hostmaskcCsd|j|jfSr)r}r{rrrrwith_prefixlenBsz_BaseNetwork.with_prefixlencCsd|j|jfSN%s/%s)r}rrrrr with_netmaskFsz_BaseNetwork.with_netmaskcCsd|j|jfSr)r}rrrrr with_hostmaskJsz_BaseNetwork.with_hostmaskcCst|jt|jdS)z&Number of hosts in the current subnet.r")rrvr}rrrr num_addressesNsz_BaseNetwork.num_addressescCsdt|f}t|dS)Nz%%200s has no associated address classrrrrrrSsz_BaseNetwork._address_classcCs|jSr)r{rrrrr[sz_BaseNetwork.prefixlenccs|j|jkstd||ft|ts2td|||sLtd||f||krXdS|d|j|jf}| \}}||kr||kr||r|V| \}}qz||r|V| \}}qzt d|||fqz||kr|Vn"||kr|Vnt d|||fdS)aRemove an address from a larger block. For example: addr1 = ip_network('192.0.2.0/28') addr2 = ip_network('192.0.2.1/32') list(addr1.address_exclude(addr2)) = [IPv4Network('192.0.2.0/32'), IPv4Network('192.0.2.2/31'), IPv4Network('192.0.2.4/30'), IPv4Network('192.0.2.8/29')] or IPv6: addr1 = ip_network('2001:db8::1/32') addr2 = ip_network('2001:db8::1/128') list(addr1.address_exclude(addr2)) = [ip_network('2001:db8::1/128'), ip_network('2001:db8::2/127'), ip_network('2001:db8::4/126'), ip_network('2001:db8::8/125'), ... ip_network('2001:db8:8000::/33')] Args: other: An IPv4Network or IPv6Network object of the same type. Returns: An iterator of the IPv(4|6)Network objects which is self minus other. Raises: TypeError: If self and other are of differing address versions, or if other is not a network object. ValueError: If other is not completely contained by self. rfz%s is not a network objectz%s not contained in %sNrz3Error performing exclusion: s1: %s s2: %s other: %s) rzrirBr subnet_ofrDrr}rrwAssertionError)r*r+s1s2rrraddress_exclude_s@$        z_BaseNetwork.address_excludecCs`|j|jkrtd||f|j|jkr,dS|j|jkr other eg: IPv4Network('192.0.2.128/25') > IPv4Network('192.0.2.0/25') IPv6Network('2001:db8::2000/124') > IPv6Network('2001:db8::1000/124') Raises: TypeError if the IP versions are different. rryr"r)rzrir}rr)rrrcompare_networkss!     z_BaseNetwork.compare_networkscCs|j|j|jfS)zNetwork-only key function. Returns an object that identifies this address' network and netmask. This function is a suitable "key" argument for sorted() and list.sort(). )rzr}rrrrrrsz_BaseNetwork._get_networks_keyr"Nc cs|j|jkr|VdS|dk rJ||jkr0td|dkr@td||j}|dkrZtd|j|}||jkr~td||ft|j}t|jd}t|jd|?}t|||D]}|||f}|VqdS)aThe subnets which join to make the current subnet. In the case that self contains only one IP (self._prefixlen == 32 for IPv4 or self._prefixlen == 128 for IPv6), yield an iterator with just ourself. Args: prefixlen_diff: An integer, the amount the prefix length should be increased by. This should not be set if new_prefix is also set. new_prefix: The desired new prefix length. This must be a larger number (smaller prefix) than the existing prefix. This should not be set if prefixlen_diff is also set. Returns: An iterator of IPv(4|6) objects. Raises: ValueError: The prefixlen_diff is too small or too large. OR prefixlen_diff and new_prefix are both set or new_prefix is a smaller number than the current prefix (smaller number means a larger network) Nznew prefix must be longerr"(cannot set prefixlen_diff and new_prefixrzprefix length diff must be > 0z0prefix length diff %d is invalid for netblock %s) r{rkrDrr}rvrr&r) r*prefixlen_diff new_prefix new_prefixlenr#r$r%Znew_addrZcurrentrrrrws2      z_BaseNetwork.subnetscCs|jdkr|S|dk rB||jkr(td|dkr8td|j|}|j|}|dkrftd|j|f|t|jt|j|>@|fS)aThe supernet containing the current network. Args: prefixlen_diff: An integer, the amount the prefix length of the network should be decreased by. For example, given a /24 network and a prefixlen_diff of 3, a supernet with a /21 netmask is returned. Returns: An IPv4 network object. Raises: ValueError: If self.prefixlen - prefixlen_diff < 0. I.e., you have a negative prefix length. OR If prefixlen_diff and new_prefix are both set or new_prefix is a larger number than the current prefix (larger number means a smaller network) rNznew prefix must be shorterr"rz;current prefixlen is %d, cannot have a prefixlen_diff of %d)r{rDrrrr}r)r*rrrrrrrqs&    z_BaseNetwork.supernetcCs|jjo|jjSzTest if the address is reserved for multicast use. Returns: A boolean, True if the address is a multicast address. See RFC 2373 2.7 for details. )r} is_multicastrvrrrrrAs z_BaseNetwork.is_multicastcCs^z4|j|jkrtd|||j|jko2|j|jkWStk rXtd||fYnXdS)Nrfz3Unable to test subnet containment between %s and %s)rzrir}rvr|)ar rrr _is_subnet_ofMs   z_BaseNetwork._is_subnet_ofcCs |||S)z1Return True if this network is a subnet of other.rr)rrrrYsz_BaseNetwork.subnet_ofcCs |||S)z3Return True if this network is a supernet of other.rr)rrr supernet_of]sz_BaseNetwork.supernet_ofcCs|jjo|jjS)Test if the address is otherwise IETF reserved. Returns: A boolean, True if the address is within one of the reserved IPv6 Network ranges. )r} is_reservedrvrrrrras z_BaseNetwork.is_reservedcCs|jjo|jjSzTest if the address is reserved for link-local. Returns: A boolean, True if the address is reserved per RFC 4291. )r} is_link_localrvrrrrrmsz_BaseNetwork.is_link_localcCs|jjo|jjS)zTest if this address is allocated for private networks. Returns: A boolean, True if the address is reserved per iana-ipv4-special-registry or iana-ipv6-special-registry. )r} is_privatervrrrrrxs z_BaseNetwork.is_privatecCs|j S)zTest if this address is allocated for public networks. Returns: A boolean, True if the address is not reserved per iana-ipv4-special-registry or iana-ipv6-special-registry. rrrrr is_globals z_BaseNetwork.is_globalcCs|jjo|jjS)Test if the address is unspecified. Returns: A boolean, True if this is the unspecified address as defined in RFC 2373 2.5.2. )r}is_unspecifiedrvrrrrrs z_BaseNetwork.is_unspecifiedcCs|jjo|jjS)Test if the address is a loopback address. Returns: A boolean, True if the address is a loopback address as defined in RFC 2373 2.5.3. )r} is_loopbackrvrrrrrs z_BaseNetwork.is_loopback)r"N)r"N)(r7r8r9r>rrrrrrr0r,rrrrrvrrrrrrrrrrrwrqr staticmethodrrrrrrrrrrrrrrsf           K0 5 (       rc @seZdZdZdZdZdedZedZ eddd d d d d ddg Z eZ iZ ddZ eddZeddZeddZeddZddZddZeddZed d!Zd"S)#rzyBase IPv4 object. The following methods are used by IPv4 objects in both single IP addresses and networks. rrrr" 0123456789rrcCst|Srrrrrrrsz$_BaseV4._explode_shorthand_ip_stringcCsn||jkrdt|tr|}n.z||}Wntk rF||}YnXt||}||f|j|<|j|SaMake a (netmask, prefix_len) tuple from the given argument. Argument can be: - an integer (the prefix length) - a string representing the prefix length (e.g. "24") - a string representing the prefix netmask (e.g. "255.255.255.0") )_netmask_cacherBrrr?rr@rrargrrrrr _make_netmasks  z_BaseV4._make_netmaskc Csz|s td|d}t|dkr.td|ztt|j|dWStk rt}ztd||fW5d}~XYnXdS)aTurn the given IP string into an integer for comparison. Args: ip_str: A string, the IP ip_str. Returns: The IP ip_str as an integer. Raises: AddressValueError: if ip_str isn't a valid IPv4 Address. Address cannot be empty.rzExpected 4 octets in %rrO%s in %rN)r<rVrWrmap _parse_octetrD)rrZoctetsexcrrrrs    z_BaseV4._ip_int_from_stringcCs|s td|j|s(d}t||t|dkrDd}t||t|d}|dkrr|ddkrrd }t|||d krtd ||S) a Convert a decimal octet into an integer. Args: octet_str: A string, the number to parse. Returns: The octet as an integer. Raises: ValueError: if the octet isn't strictly a decimal from [0..255]. zEmpty octet not permittedz#Only decimal digits permitted in %rz$At most 3 characters permitted in %r r0z3Ambiguous (octal/decimal) value in %r not permittedrzOctet %d (> 255) not permitted)rDrrrWr)rZ octet_strrZ octet_intrrrrs       z_BaseV4._parse_octetcCsdddt|ddDS)zTurns a 32-bit integer into dotted decimal notation. Args: ip_int: An integer, the IP address. Returns: The IP address as a string in dotted decimal notation. rcss0|](}tt|tr"td|dn|VqdS)rrN)rUrBrCr r r rrr %s z._BaseV4._string_from_ip_int..rrO)joinr)rrrrrrs  z_BaseV4._string_from_ip_intcsj|d}zfddtt|D}Wntk r<YdSXt|t|krRdS|d|dkrfdSdS)zTest if the IP string is a hostmask (rather than a netmask). Args: ip_str: A string, the potential hostmask. Returns: A boolean, True if the IP string is a hostmask. rcsg|]}|jkr|qSr)_valid_mask_octetsr rrrrr6s z(_BaseV4._is_hostmask..FrryT)rVrrrDrW)r*rrdpartsrrr _is_hostmask*s z_BaseV4._is_hostmaskcCs&t|dddd}d|dS)zReturn the reverse DNS pointer name for the IPv4 address. This implements the method described in RFC1035 3.5. rNryz .in-addr.arpa)rUrVr)r*Zreverse_octetsrrrr?sz_BaseV4._reverse_pointercCs|jSrrkrrrr max_prefixlenHsz_BaseV4.max_prefixlencCs|jSrrrrrrrjLsz_BaseV4.versionN)r7r8r9r>r:rz IPV4LENGTHrl frozensetrrrkrrrrrrrr rrr rjrrrrrs.    $   rc@s|eZdZdZdZddZeddZeddZed d Z ed d Z ed dZ eddZ eddZ eddZdS)r@z/Represent and manipulate single IPv4 Addresses.r\ __weakref__cCsxt|tr||||_dSt|trL||dt|}t|d|_dSt|}d|krht d|| ||_dS)a Args: address: A string or integer representing the IP Additionally, an integer can be passed, so IPv4Address('192.0.2.1') == IPv4Address(3221225985). or, more generally IPv4Address(int(IPv4Address('192.0.2.1'))) == IPv4Address('192.0.2.1') Raises: AddressValueError: If ipaddress isn't a valid IPv4 address. NrrOrTUnexpected '/' in %r rBrrr\rCrrrrUr<rr*rFZbvsZaddr_strrrrrWs      zIPv4Address.__init__cCs t|jSz*The binary representation of this address.)rRr\rrrrpacked{szIPv4Address.packedcCs ||jjkS)zTest if the address is otherwise IETF reserved. Returns: A boolean, True if the address is within the reserved IPv4 Network range. ) _constants_reserved_networkrrrrrs zIPv4Address.is_reservedcstfddjjDS)zTest if this address is allocated for private networks. Returns: A boolean, True if the address is reserved per iana-ipv4-special-registry. c3s|]}|kVqdSrrr rmrrrrsz)IPv4Address.is_private..anyr_private_networksrrrrrs zIPv4Address.is_privatecCs||jjko|j Sr)r_public_networkrrrrrrs zIPv4Address.is_globalcCs ||jjkS)zTest if the address is reserved for multicast use. Returns: A boolean, True if the address is multicast. See RFC 3171 for details. r_multicast_networkrrrrrs zIPv4Address.is_multicastcCs ||jjkS)zTest if the address is unspecified. Returns: A boolean, True if this is the unspecified address as defined in RFC 5735 3. )r_unspecified_addressrrrrrs zIPv4Address.is_unspecifiedcCs ||jjkS)zTest if the address is a loopback address. Returns: A boolean, True if the address is a loopback per RFC 3330. )r_loopback_networkrrrrrszIPv4Address.is_loopbackcCs ||jjkS)zTest if the address is reserved for link-local. Returns: A boolean, True if the address is link-local per RFC 3927. r_linklocal_networkrrrrrszIPv4Address.is_link_localN)r7r8r9r>r:rrrrrrrrrrrrrrr@Qs&$       r@c@sjeZdZddZddZddZddZd d Zej Z e d d Z e d dZ e ddZ e ddZdS)rLcCst|ttfr2t||t|j|_|j|_ dSt|t rt||dt |dkrht |d|_ n|j|_ t|dd|_|jj |_ |jj|_dSt|}t||dt|dd|_|jj |_ |jj |_ |jj|_dSNrr"F)rJ)rBrCrr@rrHr\rrkr{tuplerWrrrrYr*rFrXrrrrs(        zIPv4Interface.__init__cCsd||j|jjfSrrr\rrrrrrrs zIPv4Interface.__str__cCsFt||}|r|tkr|Sz|j|jkWStk r@YdSXdSr)r@r,r-rr|r*r+Z address_equalrrrr,s  zIPv4Interface.__eq__cCsRt||}|tkrtSz|j|jkp4|j|jko4|WStk rLYdSXdSr)r@r0r-rr|r*r+Z address_lessrrrr0s  zIPv4Interface.__lt__cCs|j|jAt|jjASrr\r{rrr}rrrrrszIPv4Interface.__hash__cCs t|jSr)r@r\rrrrraszIPv4Interface.ipcCsd||j|jfSrrr\r{rrrrr s zIPv4Interface.with_prefixlencCsd||j|jfSrrr\rrrrrrs zIPv4Interface.with_netmaskcCsd||j|jfSrrr\rrrrrrs zIPv4Interface.with_hostmaskN)r7r8r9rrr,r0rrrrrarrrrrrrrLs     rLc@s*eZdZdZeZdddZeddZdS) rHaeThis class represents and manipulates 32-bit IPv4 network + addresses.. Attributes: [examples for IPv4Network('192.0.2.0/27')] .network_address: IPv4Address('192.0.2.0') .hostmask: IPv4Address('0.0.0.31') .broadcast_address: IPv4Address('192.0.2.32') .netmask: IPv4Address('255.255.255.224') .prefixlen: 27 TcCs|t||t|ttfrr@rrrrrrrrrHs   UrHc@seZdZedZedZedZedZededededededed ed ed ed ed edededgZedZ e dZ dS)_IPv4Constantsz169.254.0.0/16z 127.0.0.0/8z 224.0.0.0/4r0z 0.0.0.0/8z 10.0.0.0/8z 172.16.0.0/12z 192.0.0.0/29z192.0.0.170/31z 192.0.2.0/24z192.168.0.0/16z 198.18.0.0/15z198.51.100.0/24z203.0.113.0/24z 240.0.0.0/4z255.255.255.255/32z0.0.0.0N) r7r8r9rHr#r!rrrrr@r rrrrr1s*r1c@seZdZdZdZdZdedZdZe dZ eZ iZ e dd Ze d d Ze d d Ze ddZe dddZddZddZeddZeddZdS)_BaseV6zyBase IPv6 object. The following methods are used by IPv6 objects in both single IP addresses and networks. rrgrr"rZ0123456789ABCDEFabcdefcCsJ||jkr@t|tr|}n ||}t||}||f|j|<|j|Sr)rrBrrrArrrrrrs   z_BaseV6._make_netmaskc Cs|s td|d}d}t||kr:d||f}t|d|dkrzt|j}Wn2tk r}ztd||fW5d}~XYnX|d |d ?d @|d |d @|jd }t||krd |d |f}t|d}td t|d D]*} || s|dk rd|}t|| }q|dk r|} t||d } |dsj| d 8} | rjd}t|||ds| d 8} | rd}t|||j| | } | d kr0d}t||jd |fnjt||jkrd}t||j|f|dsd}t|||ds d}t||t|} d} d} znd} t | D] } | d K} | | || O} q>| d | K} t | dD] } | d K} | | || O} qx| WSt k r}ztd||fW5d}~XYnXdS)zTurn an IPv6 ip_str into an integer. Args: ip_str: A string, the IPv6 ip_str. Returns: An int, the IPv6 address Raises: AddressValueError: if ip_str isn't a valid IPv6 Address. r:rz At least %d parts expected in %rrryrN%xrr"z!At most %d colons permitted in %rz At most one '::' permitted in %rrz0Leading ':' only permitted as part of '::' in %rz1Trailing ':' only permitted as part of '::' in %rz/Expected at most %d other parts with '::' in %rz,Exactly %d parts expected without '::' in %r) r<rVrWr@rpr\rs _HEXTET_COUNTr&range _parse_hextetrD)rrr Z _min_partsrZipv4_intrZ _max_partsZ skip_indexrZparts_hiZparts_loZ parts_skippedrrrrrs    "               z_BaseV6._ip_int_from_stringcCs>|j|std|t|dkr4d}t||t|dS)a&Convert an IPv6 hextet string into an integer. Args: hextet_str: A string, the number to parse. Returns: The hextet as an integer. Raises: ValueError: if the input isn't strictly a hex number from [0..FFFF]. zOnly hex digits permitted in %rrz$At most 4 characters permitted in %rr) _HEX_DIGITSrrDrWr)rZ hextet_strrrrrr8>s     z_BaseV6._parse_hextetc Csd}d}d}d}t|D]>\}}|dkrN|d7}|dkr<|}||krV|}|}qd}d}q|dkr||}|t|kr~|dg7}dg|||<|dkrdg|}|S)aCompresses a list of hextets. Compresses a list of strings, replacing the longest continuous sequence of "0" in the list with "" and adding empty strings at the beginning or at the end of the string such that subsequently calling ":".join(hextets) will produce the compressed version of the IPv6 address. Args: hextets: A list of strings, the hextets to compress. Returns: A list of strings. ryrrr") enumeraterW) rhextetsZbest_doublecolon_startZbest_doublecolon_lenZdoublecolon_startZdoublecolon_lenindexZhextetZbest_doublecolon_endrrr_compress_hextetsXs0   z_BaseV6._compress_hextetsNcsZ|dkrt|j}||jkr$tdd|fddtdddD}||}d |S) a,Turns a 128-bit integer into hexadecimal notation. Args: ip_int: An integer, the IP address. Returns: A string, the hexadecimal representation of the address. Raises: ValueError: The address is bigger than 128 bits of all ones. NzIPv6 address is too large%032xcs&g|]}dt||ddqS)r4rr)rr Zhex_strrrrsz/_BaseV6._string_from_ip_int..rr;rr3)rr\rlrDr7r>r)rrr<rr@rrs   z_BaseV6._string_from_ip_intcst|trt|j}nt|tr,t|j}nt|}||}d|fddtdddD}t|ttfrdd ||j fSd |S) zExpand a shortened IPv6 address. Args: ip_str: A string, the IPv6 address. Returns: A string, the expanded IPv6 address. r?csg|]}||dqS)rrr r@rrrsz8_BaseV6._explode_shorthand_ip_string..rr;rrr3) rBrIrUr}rMrarr7rrr{)r*rrr rr@rrs     z$_BaseV6._explode_shorthand_ip_stringcCs&|jddddd}d|dS)zReturn the reverse DNS pointer name for the IPv6 address. This implements the method described in RFC3596 2.5. Nryr3r:rz .ip6.arpa)rreplacer)r*Z reverse_charsrrrrsz_BaseV6._reverse_pointercCs|jSrr rrrrr sz_BaseV6.max_prefixlencCs|jSrrrrrrrjsz_BaseV6.version)N)r7r8r9r>r:rz IPV6LENGTHrlr6rr9rkrrrrr8r>rrrrr rjrrrrr2s0   h  /   r2c@seZdZdZdZddZeddZeddZed d Z ed d Z ed dZ eddZ eddZ eddZeddZeddZeddZeddZdS)rAz/Represent and manipulate single IPv6 Addresses.rcCsxt|tr||||_dSt|trL||dt|}t|d|_dSt|}d|krht d|| ||_dS)aInstantiate a new IPv6 address object. Args: address: A string or integer representing the IP Additionally, an integer can be passed, so IPv6Address('2001:db8::') == IPv6Address(42540766411282592856903984951653826560) or, more generally IPv6Address(int(IPv6Address('2001:db8::'))) == IPv6Address('2001:db8::') Raises: AddressValueError: If address isn't a valid IPv6 address. NrrOrTrrrrrrrs      zIPv6Address.__init__cCs t|jSr)rSr\rrrrrszIPv6Address.packedcCs ||jjkSrrrrrrrs zIPv6Address.is_multicastcstfddjjDS)rc3s|]}|kVqdSrrr rrrrsz*IPv6Address.is_reserved..)rr_reserved_networksrrrrrs zIPv6Address.is_reservedcCs ||jjkSrr"rrrrrszIPv6Address.is_link_localcCs ||jjkSa`Test if the address is reserved for site-local. Note that the site-local address space has been deprecated by RFC 3879. Use is_private to test if this address is in the space of unique local addresses as defined by RFC 4193. Returns: A boolean, True if the address is reserved per RFC 3513 2.5.6. )r_sitelocal_networkrrrr is_site_locals zIPv6Address.is_site_localcstfddjjDS)zTest if this address is allocated for private networks. Returns: A boolean, True if the address is reserved per iana-ipv6-special-registry. c3s|]}|kVqdSrrrrrrr3sz)IPv6Address.is_private..rrrrrr*s zIPv6Address.is_privatecCs|j S)zTest if this address is allocated for public networks. Returns: A boolean, true if the address is not reserved per iana-ipv6-special-registry. rrrrrr5s zIPv6Address.is_globalcCs |jdkS)rrrrrrrr@s zIPv6Address.is_unspecifiedcCs |jdkS)rr"rrrrrrKs zIPv6Address.is_loopbackcCs |jd?dkrdSt|jd@S)zReturn the IPv4 mapped address. Returns: If the IPv6 address is a v4 mapped address, return the IPv4 mapped address. Return None otherwise. r;r5Nr\r@rrrr ipv4_mappedVs zIPv6Address.ipv4_mappedcCs4|jd?dkrdSt|jd?d@t|jd@fS)zTuple of embedded teredo IPs. Returns: Tuple of the (server, client) IPs or None if the address doesn't appear to be a teredo address (doesn't start with 2001::/32) `i NrrGrHrrrrteredocs zIPv6Address.teredocCs$|jd?dkrdSt|jd?d@S)zReturn the IPv4 6to4 embedded address. Returns: The IPv4 6to4-embedded address if present or None if the address doesn't appear to contain a 6to4 embedded address. pi NPrGrHrrrr sixtofourrs zIPv6Address.sixtofourN)r7r8r9r>r:rrrrrrrFrrrrrIrKrNrrrrrAs6%           rAc@seZdZddZddZddZddZd d Zej Z e d d Z e d dZ e ddZ e ddZe ddZe ddZdS)rMcCst|ttfr2t||t|j|_|j|_ dSt|t rt||dt |dkrht |d|_ n|j|_ t|dd|_|jj |_ |jj|_dSt|}t||dt|dd|_|jj |_ |jj |_ |jj|_dSr$)rBrCrrArrIr\rrkr{r%rWrrrrYr&rrrrs(        zIPv6Interface.__init__cCsd||j|jjfSrr'rrrrrs zIPv6Interface.__str__cCsFt||}|r|tkr|Sz|j|jkWStk r@YdSXdSr)rAr,r-rr|r(rrrr,s  zIPv6Interface.__eq__cCsRt||}|tkrtSz|j|jkp4|j|jko4|WStk rLYdSXdSr)rAr0r-rr|r)rrrr0s  zIPv6Interface.__lt__cCs|j|jAt|jjASrr*rrrrrszIPv6Interface.__hash__cCs t|jSr)rAr\rrrrraszIPv6Interface.ipcCsd||j|jfSrr+rrrrrs zIPv6Interface.with_prefixlencCsd||j|jfSrr,rrrrrs zIPv6Interface.with_netmaskcCsd||j|jfSrr-rrrrrs zIPv6Interface.with_hostmaskcCs|jdko|jjSr)r\rrrrrrrszIPv6Interface.is_unspecifiedcCs|jdko|jjSr)r\rrrrrrrszIPv6Interface.is_loopbackN)r7r8r9rrr,r0rrrrrarrrrrrrrrrMs$       rMc@s2eZdZdZeZd ddZddZeddZ d S) rIavThis class represents and manipulates 128-bit IPv6 networks. Attributes: [examples for IPv6('2001:db8::1000/124')] .network_address: IPv6Address('2001:db8::1000') .hostmask: IPv6Address('::f') .broadcast_address: IPv6Address('2001:db8::100f') .netmask: IPv6Address('ffff:ffff:ffff:ffff:ffff:ffff:ffff:fff0') .prefixlen: 124 TcCs|t||t|ttfrrArrrrrFrrrrrIs   O rIc@seZdZedZedZededededededed ed ed edg Zed ed edededededededededededededgZedZdS)_IPv6Constantsz fe80::/10zff00::/8z::1/128z::/128z ::ffff:0:0/96z100::/64z 2001::/23z 2001:2::/48z 2001:db8::/32z 2001:10::/28zfc00::/7z::/8z100::/8z200::/7z400::/6z800::/5z1000::/4z4000::/3z6000::/3z8000::/3zA000::/3zC000::/3zE000::/4zF000::/5zF800::/6zFE00::/9z fec0::/10N) r7r8r9rIr#rrrCrErrrrrOR s< rO)r")T)6r>Z __future__rr r __version__rrZlong NameErrorZunicoderUstrr from_bytesrr|rhasattrrr&objectr'rrBrDr<r?rGrKrNrRrSrYrbrernrxrrrrhrrr@rLrHr1rr2rArMrIrOrrrrs        )$ $#716=X*vSr 5W{!site-packages/pip/_vendor/__pycache__/ipaddress.cpython-38.pyc000064400000176416151464452710020356 0ustar00U .e7@sldZddlmZddlZddlZdZefZz eefZWne k rJYnXze Z Wn$e k rxe Z e e ksttYnXddkrddZnddZz ejZWnek rdd ZYnXd d Zeed rd dZnddZdNddZGdddeZdZdZGdddeZGdddeZddZdOddZd d!Z d"d#Z!d$d%Z"d&d'Z#d(d)Z$d*d+Z%d,d-Z&d.d/Z'd0d1Z(d2d3Z)Gd4d5d5eZ*Gd6d7d7e*Z+Gd8d9d9e*Z,Gd:d;d;eZ-Gdd?d?e.Z/Gd@dAdAe-e,Z0GdBdCdCeZ1e1e._2GdDdEdEeZ3GdFdGdGe3e+Z4GdHdIdIe4Z5GdJdKdKe3e,Z6GdLdMdMeZ7e7e4_2dS)PzA fast, lightweight IPv4/IPv6 manipulation library in Python. This library is used to create/poke/manipulate IPv4 and IPv6 addresses and networks. )unicode_literalsNz1.0.22cCs|SNZbytrr9/usr/lib/python3.8/site-packages/pip/_vendor/ipaddress.py_compat_bytes_to_byte_valssrcCsdd|DS)NcSsg|]}td|dqS)!Br)structunpack.0brrr #sz._compat_bytes_to_byte_vals..rrrrrr"scCs8|dks td}|D]}t|ts&t|d>|}q|S)Nbigr)AssertionError isinstance_compat_int_types)Zbytvals endianessresZbvrrr_compat_int_from_byte_vals's  rcCst|tst|dkst|dkrH|dks2|dkr7}|d|j krqqdS) aSummarize a network range given the first and last IP addresses. Example: >>> list(summarize_address_range(IPv4Address('192.0.2.0'), ... IPv4Address('192.0.2.130'))) ... #doctest: +NORMALIZE_WHITESPACE [IPv4Network('192.0.2.0/25'), IPv4Network('192.0.2.128/31'), IPv4Network('192.0.2.130/32')] Args: first: the first IPv4Address or IPv6Address in the range. last: the last IPv4Address or IPv6Address in the range. Returns: An iterator of the summarized IPv(4|6) network objects. Raise: TypeError: If the first and last objects are not IP addresses. If the first and last objects are not the same version. ValueError: If the last object is not greater than the first. If the version of the first address is not 4 or 6. z1first and last must be IP addresses, not networks%%s and %s are not of the same versionz*last IP address must be greater than firstrzunknown IP versionr&N) r _BaseAddress TypeErrorversionrGrKrL_max_prefixlenr^rergr" _ALL_ONES)rarbrcZip_bitsZ first_intZlast_intZnbitsnetrrrsummarize_address_rangeBs8      rpccst|}i}|rV|}|}||}|dkr<|||<q ||kr ||=||q d}t|D]$}|dk r|j|jkrqf|V|}qfdS)auLoops through the addresses, collapsing concurrent netblocks. Example: ip1 = IPv4Network('192.0.2.0/26') ip2 = IPv4Network('192.0.2.64/26') ip3 = IPv4Network('192.0.2.128/26') ip4 = IPv4Network('192.0.2.192/26') _collapse_addresses_internal([ip1, ip2, ip3, ip4]) -> [IPv4Network('192.0.2.0/24')] This shouldn't be called directly; it is called via collapse_addresses([]). Args: addresses: A list of IPv4Network's or IPv6Network's Returns: A list of IPv4Network's or IPv6Network's depending on what we were passed. N)listpopsupernetgetappendsortedvaluesbroadcast_address)r_Zto_mergesubnetsrorsZexistingrbrrr_collapse_addresses_internalys$    rzc Cs0g}g}g}|D]}t|trR|rF|dj|jkrFtd||df||q|j|jkr|r|dj|jkrtd||dfz||jWqtk r||j YqXq|r|dj|jkrtd||df||qt t |}|r$t |D]\}}| t||qt||S)aCollapse a list of IP objects. Example: collapse_addresses([IPv4Network('192.0.2.0/25'), IPv4Network('192.0.2.128/25')]) -> [IPv4Network('192.0.2.0/24')] Args: addresses: An iterator of IPv4Network or IPv6Network objects. Returns: An iterator of the collapsed IPv(4|6)Network objects. Raises: TypeError: If passed a list of mixed version objects. rh)rrj_versionrkru _prefixlenrmrcAttributeErrornetwork_addressrvsetrdextendrprz)r_ZaddrsZipsZnetsrcrarbrrrcollapse_addressess@     rcCs(t|tr|St|tr$|StS)a2Return a key suitable for sorting between networks and addresses. Address and Network objects are not sortable by default; they're fundamentally different so the expression IPv4Address('192.0.2.0') <= IPv4Network('192.0.2.0/24') doesn't make any sense. There are some times however, where you may wish to have ipaddress sort these for you anyway. If you need to do this, you can use this function as the key= argument to sorted(). Args: obj: either a Network or Address object. Returns: appropriate key. )r _BaseNetwork_get_networks_keyrj_get_address_keyr1)objrrrget_mixed_type_keys   rc@seZdZdZdZeddZeddZeddZed d Z d d Z d dZ e ddZ e ddZe ddZe ddZe ddZddZdS)_IPAddressBasezThe mother class.rcCs|S)z:Return the longhand version of the IP address as a string.)_explode_shorthand_ip_stringr.rrrexplodedsz_IPAddressBase.explodedcCst|S)z;Return the shorthand version of the IP address as a string.rWrrrr compressedsz_IPAddressBase.compressedcCs|S)aIThe name of the reverse DNS pointer for the IP address, e.g.: >>> ipaddress.ip_address("127.0.0.1").reverse_pointer '1.0.0.127.in-addr.arpa' >>> ipaddress.ip_address("2001:db8::1").reverse_pointer '1.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.8.b.d.0.1.0.0.2.ip6.arpa' )_reverse_pointerrrrrreverse_pointer s z_IPAddressBase.reverse_pointercCsdt|f}t|dS)Nz%200s has no version specifiedtyperr.msgrrrrlsz_IPAddressBase.versioncCsF|dkrd}t|||jf||jkrBd}t|||j|jfdS)Nrz-%d (< 0) is not permitted as an IPv%d addressz2%d (>= 2**%d) is not permitted as an IPv%d address)r@r|rnrm)r.rIrrrr_check_int_addresss  z!_IPAddressBase._check_int_addresscCs.t|}||kr*d}t|||||jfdS)Nz~%r (len %d != %d) is not permitted as an IPv%d address. Did you pass in a bytes (str in Python 2) instead of a unicode object?)rYr@r|)r.rIZ expected_lenZ address_lenrrrr_check_packed_address"sz$_IPAddressBase._check_packed_addresscCs|j|j|?AS)zTurn the prefix length into a bitwise netmask Args: prefixlen: An integer, the prefix length. Returns: An integer. )rn)cls prefixlenrrr_ip_int_from_prefix,s z"_IPAddressBase._ip_int_from_prefixc Cs\t||j}|j|}||?}d|>d}||krX|jd}t||d}d}t|||S)aReturn prefix length from the bitwise netmask. Args: ip_int: An integer, the netmask in expanded bitwise format Returns: An integer, the prefix length. Raises: ValueError: If the input intermingles zeroes & ones r&rrz&Netmask pattern %r mixes zeroes & ones)rgrmrrG) rip_intZtrailing_zeroesrZ leading_onesZall_onesZbyteslenZdetailsrrrr_prefix_from_ip_int9s      z"_IPAddressBase._prefix_from_ip_intcCsd|}t|dS)Nz%r is not a valid netmask)rC)rZ netmask_strrrrr_report_invalid_netmaskRsz&_IPAddressBase._report_invalid_netmaskcCshtj|s||z t|}Wntk r@||YnXd|krX|jksdn|||S)a Return prefix length from a numeric string Args: prefixlen_str: The string to be converted Returns: An integer, the prefix length. Raises: NetmaskValueError: If the input is not a valid netmask r)_BaseV4_DECIMAL_DIGITS issupersetrintrGrm)rZ prefixlen_strrrrr_prefix_from_prefix_stringWs    z)_IPAddressBase._prefix_from_prefix_stringcCsz||}Wntk r,||YnXz ||WStk rNYnX||jN}z ||WStk r||YnXdS)aTurn a netmask/hostmask string into a prefix length Args: ip_str: The netmask/hostmask to be converted Returns: An integer, the prefix length. Raises: NetmaskValueError: If the input is not a valid netmask/hostmask N)_ip_int_from_stringr@rrrGrn)rip_strrrrr_prefix_from_ip_stringps   z%_IPAddressBase._prefix_from_ip_stringcCs|jt|ffSr) __class__rWrrrr __reduce__sz_IPAddressBase.__reduce__N)r;r<r=rBr>propertyrrrrlrr classmethodrrrrrrrrrrrs.           !rc@sdeZdZdZdZddZddZddZd d Zd d Z d dZ ddZ ddZ ddZ ddZdS)rjzA generic IP object. This IP class contains the version independent methods which are used by single IP addresses. rcCs|jSrr^rrrr__int__sz_BaseAddress.__int__cCs8z|j|jko|j|jkWStk r2tYSXdSr)r^r|r~r1r-rrrr0s   z_BaseAddress.__eq__cCs`t|tstSt|ts(td||f|j|jkrDtd||f|j|jkr\|j|jkSdSN"%s and %s are not of the same typerhF)rrr1rjrkr|r^r-rrrr4s     z_BaseAddress.__lt__cCs t|tstS|t||Srrrr1rrr-rrr__add__s z_BaseAddress.__add__cCs t|tstS|t||Srrr-rrr__sub__s z_BaseAddress.__sub__cCsd|jjt|fSNz%s(%r)rr;rWrrrr__repr__sz_BaseAddress.__repr__cCst||jSr)rW_string_from_ip_intr^rrrr__str__sz_BaseAddress.__str__cCsttt|jSr)hashhexrr^rrrr__hash__sz_BaseAddress.__hash__cCs |j|fSrr|rrrrrsz_BaseAddress._get_address_keycCs|j|jffSr)rr^rrrrrsz_BaseAddress.__reduce__N)r;r<r=rBr>rr0r4rrrrrrrrrrrrjsrjc@sdeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ ddZ ddZeddZeddZeddZeddZed d!Zed"d#Zed$d%Zed&d'Zd(d)Zd*d+Zd,d-ZdHd0d1ZdId2d3Zed4d5Zed6d7Zd8d9Z d:d;Z!edd?Z#ed@dAZ$edBdCZ%edDdEZ&edFdGZ'd/S)Jrz~A generic IP network object. This IP class contains the version independent methods which are used by networks. cCs i|_dSr)_cache)r.rIrrr__init__sz_BaseNetwork.__init__cCsd|jjt|fSrrrrrrrsz_BaseNetwork.__repr__cCsd|j|jfSN%s/%d)rrrrrrrsz_BaseNetwork.__str__ccs8t|j}t|j}t|d|D]}||Vq"dS)zGenerate Iterator over usable hosts in a network. This is like __iter__ except it doesn't return the network or broadcast addresses. r&Nrrrxr*_address_classr.network broadcastxrrrhostss  z_BaseNetwork.hostsccs8t|j}t|j}t||dD]}||Vq"dSNr&rrrrr__iter__s  z_BaseNetwork.__iter__cCslt|j}t|j}|dkr>|||kr0td|||S|d7}|||krZtd|||SdS)Nrzaddress out of ranger&)rrrx IndexErrorr)r.nrrrrr __getitem__s    z_BaseNetwork.__getitem__cCsxt|tstSt|ts(td||f|j|jkrDtd||f|j|jkr\|j|jkS|j|jkrt|j|jkSdSr)rrr1rrkr|rnetmaskr-rrrr4s"       z_BaseNetwork.__lt__cCsLz.|j|jko,|j|jko,t|jt|jkWStk rFtYSXdSr)r|rrrr~r1r-rrrr0s  z_BaseNetwork.__eq__cCstt|jt|jASr)rrrrrrrrrsz_BaseNetwork.__hash__cCsL|j|jkrdSt|trdSt|jt|jkoBt|jkSSdSNF)r|rrrrr^rxr-rrr __contains__s  z_BaseNetwork.__contains__cCs(|j|kp&|j|kp&|j|kp&|j|kS)z*Tell if self is partly contained in other.)rrxr-rrroverlaps*s    z_BaseNetwork.overlapscCs<|jd}|dkr8|t|jt|jB}||jd<|S)Nrx)rrtrrrhostmaskr.rrrrrx1s   z_BaseNetwork.broadcast_addresscCs8|jd}|dkr4|t|j|jA}||jd<|S)Nr)rrtrrrrnrrrrr:s   z_BaseNetwork.hostmaskcCsd|j|jfSr)rr}rrrrwith_prefixlenBsz_BaseNetwork.with_prefixlencCsd|j|jfSN%s/%s)rrrrrr with_netmaskFsz_BaseNetwork.with_netmaskcCsd|j|jfSr)rrrrrr with_hostmaskJsz_BaseNetwork.with_hostmaskcCst|jt|jdS)z&Number of hosts in the current subnet.r&)rrxrrrrr num_addressesNsz_BaseNetwork.num_addressescCsdt|f}t|dS)Nz%%200s has no associated address classrrrrrrSsz_BaseNetwork._address_classcCs|jSr)r}rrrrr[sz_BaseNetwork.prefixlenccs|j|jkstd||ft|ts2td|||sLtd||f||krXdS|d|j|jf}| \}}||kr||kr||r|V| \}}qz||r|V| \}}qzt d|||fqz||kr|Vn"||kr|Vnt d|||fdS)aRemove an address from a larger block. For example: addr1 = ip_network('192.0.2.0/28') addr2 = ip_network('192.0.2.1/32') list(addr1.address_exclude(addr2)) = [IPv4Network('192.0.2.0/32'), IPv4Network('192.0.2.2/31'), IPv4Network('192.0.2.4/30'), IPv4Network('192.0.2.8/29')] or IPv6: addr1 = ip_network('2001:db8::1/32') addr2 = ip_network('2001:db8::1/128') list(addr1.address_exclude(addr2)) = [ip_network('2001:db8::1/128'), ip_network('2001:db8::2/127'), ip_network('2001:db8::4/126'), ip_network('2001:db8::8/125'), ... ip_network('2001:db8:8000::/33')] Args: other: An IPv4Network or IPv6Network object of the same type. Returns: An iterator of the IPv(4|6)Network objects which is self minus other. Raises: TypeError: If self and other are of differing address versions, or if other is not a network object. ValueError: If other is not completely contained by self. rhz%s is not a network objectz%s not contained in %sNrz3Error performing exclusion: s1: %s s2: %s other: %s) r|rkrr subnet_ofrGrrrryr)r.r/s1s2rrraddress_exclude_s@$        z_BaseNetwork.address_excludecCs`|j|jkrtd||f|j|jkr,dS|j|jkr other eg: IPv4Network('192.0.2.128/25') > IPv4Network('192.0.2.0/25') IPv6Network('2001:db8::2000/124') > IPv6Network('2001:db8::1000/124') Raises: TypeError if the IP versions are different. rr{r&r)r|rkrrr-rrrcompare_networkss!     z_BaseNetwork.compare_networkscCs|j|j|jfS)zNetwork-only key function. Returns an object that identifies this address' network and netmask. This function is a suitable "key" argument for sorted() and list.sort(). )r|rrrrrrrsz_BaseNetwork._get_networks_keyr&Nc cs|j|jkr|VdS|dk rJ||jkr0td|dkr@td||j}|dkrZtd|j|}||jkr~td||ft|j}t|jd}t|jd|?}t|||D]}|||f}|VqdS)aThe subnets which join to make the current subnet. In the case that self contains only one IP (self._prefixlen == 32 for IPv4 or self._prefixlen == 128 for IPv6), yield an iterator with just ourself. Args: prefixlen_diff: An integer, the amount the prefix length should be increased by. This should not be set if new_prefix is also set. new_prefix: The desired new prefix length. This must be a larger number (smaller prefix) than the existing prefix. This should not be set if prefixlen_diff is also set. Returns: An iterator of IPv(4|6) objects. Raises: ValueError: The prefixlen_diff is too small or too large. OR prefixlen_diff and new_prefix are both set or new_prefix is a smaller number than the current prefix (smaller number means a larger network) Nznew prefix must be longerr&(cannot set prefixlen_diff and new_prefixrzprefix length diff must be > 0z0prefix length diff %d is invalid for netblock %s) r}rmrGrrrxrr*r) r.prefixlen_diff new_prefix new_prefixlenr'r(r)Znew_addrZcurrentrrrrys2      z_BaseNetwork.subnetscCs|jdkr|S|dk rB||jkr(td|dkr8td|j|}|j|}|dkrftd|j|f|t|jt|j|>@|fS)aThe supernet containing the current network. Args: prefixlen_diff: An integer, the amount the prefix length of the network should be decreased by. For example, given a /24 network and a prefixlen_diff of 3, a supernet with a /21 netmask is returned. Returns: An IPv4 network object. Raises: ValueError: If self.prefixlen - prefixlen_diff < 0. I.e., you have a negative prefix length. OR If prefixlen_diff and new_prefix are both set or new_prefix is a larger number than the current prefix (larger number means a smaller network) rNznew prefix must be shorterr&rz;current prefixlen is %d, cannot have a prefixlen_diff of %d)r}rGrrrrr)r.rrrrrrrss&    z_BaseNetwork.supernetcCs|jjo|jjSzTest if the address is reserved for multicast use. Returns: A boolean, True if the address is a multicast address. See RFC 2373 2.7 for details. )r is_multicastrxrrrrrAs z_BaseNetwork.is_multicastcCs^z4|j|jkrtd|||j|jko2|j|jkWStk rXtd||fYnXdS)Nrhz3Unable to test subnet containment between %s and %s)r|rkrrxr~)ar rrr _is_subnet_ofMs   z_BaseNetwork._is_subnet_ofcCs |||S)z1Return True if this network is a subnet of other.rr-rrrrYsz_BaseNetwork.subnet_ofcCs |||S)z3Return True if this network is a supernet of other.rr-rrr supernet_of]sz_BaseNetwork.supernet_ofcCs|jjo|jjS)Test if the address is otherwise IETF reserved. Returns: A boolean, True if the address is within one of the reserved IPv6 Network ranges. )r is_reservedrxrrrrras z_BaseNetwork.is_reservedcCs|jjo|jjSzTest if the address is reserved for link-local. Returns: A boolean, True if the address is reserved per RFC 4291. )r is_link_localrxrrrrrmsz_BaseNetwork.is_link_localcCs|jjo|jjS)zTest if this address is allocated for private networks. Returns: A boolean, True if the address is reserved per iana-ipv4-special-registry or iana-ipv6-special-registry. )r is_privaterxrrrrrxs z_BaseNetwork.is_privatecCs|j S)zTest if this address is allocated for public networks. Returns: A boolean, True if the address is not reserved per iana-ipv4-special-registry or iana-ipv6-special-registry. rrrrr is_globals z_BaseNetwork.is_globalcCs|jjo|jjS)Test if the address is unspecified. Returns: A boolean, True if this is the unspecified address as defined in RFC 2373 2.5.2. )ris_unspecifiedrxrrrrrs z_BaseNetwork.is_unspecifiedcCs|jjo|jjS)Test if the address is a loopback address. Returns: A boolean, True if the address is a loopback address as defined in RFC 2373 2.5.3. )r is_loopbackrxrrrrrs z_BaseNetwork.is_loopback)r&N)r&N)(r;r<r=rBrrrrrrr4r0rrrrrxrrrrrrrrrrryrsr staticmethodrrrrrrrrrrrrrrsf           K0 5 (       rc @seZdZdZdZdZdedZedZ eddd d d d d ddg Z eZ iZ ddZ eddZeddZeddZeddZddZddZeddZed d!Zd"S)#rzyBase IPv4 object. The following methods are used by IPv4 objects in both single IP addresses and networks. rrrr& 0123456789rrcCst|Srrrrrrrsz$_BaseV4._explode_shorthand_ip_stringcCsn||jkrdt|tr|}n.z||}Wntk rF||}YnXt||}||f|j|<|j|SaMake a (netmask, prefix_len) tuple from the given argument. Argument can be: - an integer (the prefix length) - a string representing the prefix length (e.g. "24") - a string representing the prefix netmask (e.g. "255.255.255.0") )_netmask_cacherrrrCrrDrrargrrrrr _make_netmasks  z_BaseV4._make_netmaskc Csz|s td|d}t|dkr.td|ztt|j|dWStk rt}ztd||fW5d}~XYnXdS)aTurn the given IP string into an integer for comparison. Args: ip_str: A string, the IP ip_str. Returns: The IP ip_str as an integer. Raises: AddressValueError: if ip_str isn't a valid IPv4 Address. Address cannot be empty.rzExpected 4 octets in %rr%s in %rN)r@rXrYrmap _parse_octetrG)rrZoctetsexcrrrrs    z_BaseV4._ip_int_from_stringcCs|s td|j|s(d}t||t|dkrDd}t||t|d}|dkrr|ddkrrd }t|||d krtd ||S) a Convert a decimal octet into an integer. Args: octet_str: A string, the number to parse. Returns: The octet as an integer. Raises: ValueError: if the octet isn't strictly a decimal from [0..255]. zEmpty octet not permittedz#Only decimal digits permitted in %rz$At most 3 characters permitted in %r r0z3Ambiguous (octal/decimal) value in %r not permittedrzOctet %d (> 255) not permitted)rGrrrYr)rZ octet_strrZ octet_intrrrrs       z_BaseV4._parse_octetcCsdddt|ddDS)zTurns a 32-bit integer into dotted decimal notation. Args: ip_int: An integer, the IP address. Returns: The IP address as a string in dotted decimal notation. rcss0|](}tt|tr"td|dn|VqdS)rrN)rWrrFr r r rrr %s z._BaseV4._string_from_ip_int..rr)joinr)rrrrrrs  z_BaseV4._string_from_ip_intcsj|d}zfddtt|D}Wntk r<YdSXt|t|krRdS|d|dkrfdSdS)zTest if the IP string is a hostmask (rather than a netmask). Args: ip_str: A string, the potential hostmask. Returns: A boolean, True if the IP string is a hostmask. rcsg|]}|jkr|qSr)_valid_mask_octetsr rrrrr6s z(_BaseV4._is_hostmask..Frr{T)rXrrrGrY)r.rrfpartsrrr _is_hostmask*s z_BaseV4._is_hostmaskcCs&t|dddd}d|dS)zReturn the reverse DNS pointer name for the IPv4 address. This implements the method described in RFC1035 3.5. rNr{z .in-addr.arpa)rWrXr)r.Zreverse_octetsrrrr?sz_BaseV4._reverse_pointercCs|jSrrmrrrr max_prefixlenHsz_BaseV4.max_prefixlencCs|jSrrrrrrrlLsz_BaseV4.versionN)r;r<r=rBr>r| IPV4LENGTHrn frozensetrrrmrrrrrrrr rrr rlrrrrrs.    $   rc@s|eZdZdZdZddZeddZeddZed d Z ed d Z ed dZ eddZ eddZ eddZdS)rDz/Represent and manipulate single IPv4 Addresses.r^ __weakref__cCsxt|tr||||_dSt|trL||dt|}t|d|_dSt|}d|krht d|| ||_dS)a Args: address: A string or integer representing the IP Additionally, an integer can be passed, so IPv4Address('192.0.2.1') == IPv4Address(3221225985). or, more generally IPv4Address(int(IPv4Address('192.0.2.1'))) == IPv4Address('192.0.2.1') Raises: AddressValueError: If ipaddress isn't a valid IPv4 address. NrrrVUnexpected '/' in %r rrrr^rFrrrrWr@rr.rIZbvsZaddr_strrrrrWs      zIPv4Address.__init__cCs t|jSz*The binary representation of this address.)rTr^rrrrpacked{szIPv4Address.packedcCs ||jjkS)zTest if the address is otherwise IETF reserved. Returns: A boolean, True if the address is within the reserved IPv4 Network range. ) _constants_reserved_networkrrrrrs zIPv4Address.is_reservedcstfddjjDS)zTest if this address is allocated for private networks. Returns: A boolean, True if the address is reserved per iana-ipv4-special-registry. c3s|]}|kVqdSrrr rorrrrsz)IPv4Address.is_private..anyr_private_networksrrrrrs zIPv4Address.is_privatecCs||jjko|j Sr)r_public_networkrrrrrrs zIPv4Address.is_globalcCs ||jjkS)zTest if the address is reserved for multicast use. Returns: A boolean, True if the address is multicast. See RFC 3171 for details. r_multicast_networkrrrrrs zIPv4Address.is_multicastcCs ||jjkS)zTest if the address is unspecified. Returns: A boolean, True if this is the unspecified address as defined in RFC 5735 3. )r_unspecified_addressrrrrrs zIPv4Address.is_unspecifiedcCs ||jjkS)zTest if the address is a loopback address. Returns: A boolean, True if the address is a loopback per RFC 3330. )r_loopback_networkrrrrrszIPv4Address.is_loopbackcCs ||jjkS)zTest if the address is reserved for link-local. Returns: A boolean, True if the address is link-local per RFC 3927. r_linklocal_networkrrrrrszIPv4Address.is_link_localN)r;r<r=rBr>rrrrrrrrrrrrrrrDQs&$       rDc@sjeZdZddZddZddZddZd d Zej Z e d d Z e d dZ e ddZ e ddZdS)rOcCst|ttfr2t||t|j|_|j|_ dSt|t rt||dt |dkrht |d|_ n|j|_ t|dd|_|jj |_ |jj|_dSt|}t||dt|dd|_|jj |_ |jj |_ |jj|_dSNrr&F)rM)rrFrrDrrKr^rrmr}tuplerYrrrr[r.rIrZrrrrs(        zIPv4Interface.__init__cCsd||j|jjfSrrr^rrrrrrrs zIPv4Interface.__str__cCsFt||}|r|tkr|Sz|j|jkWStk r@YdSXdSr)rDr0r1rr~r.r/Z address_equalrrrr0s  zIPv4Interface.__eq__cCsRt||}|tkrtSz|j|jkp4|j|jko4|WStk rLYdSXdSr)rDr4r1rr~r.r/Z address_lessrrrr4s  zIPv4Interface.__lt__cCs|j|jAt|jjASrr^r}rrrrrrrrszIPv4Interface.__hash__cCs t|jSr)rDr^rrrrrcszIPv4Interface.ipcCsd||j|jfSrrr^r}rrrrr s zIPv4Interface.with_prefixlencCsd||j|jfSrrr^rrrrrrs zIPv4Interface.with_netmaskcCsd||j|jfSrrr^rrrrrrs zIPv4Interface.with_hostmaskN)r;r<r=rrr0r4rrrrrcrrrrrrrrOs     rOc@s*eZdZdZeZdddZeddZdS) rKaeThis class represents and manipulates 32-bit IPv4 network + addresses.. Attributes: [examples for IPv4Network('192.0.2.0/27')] .network_address: IPv4Address('192.0.2.0') .hostmask: IPv4Address('0.0.0.31') .broadcast_address: IPv4Address('192.0.2.32') .netmask: IPv4Address('255.255.255.224') .prefixlen: 27 TcCs|t||t|ttfr| d | K} t | dD] } | d K} | | || O} qx| WSt k r}ztd||fW5d}~XYnXdS)zTurn an IPv6 ip_str into an integer. Args: ip_str: A string, the IPv6 ip_str. Returns: An int, the IPv6 address Raises: AddressValueError: if ip_str isn't a valid IPv6 Address. r:rz At least %d parts expected in %rrr{rN%xrr&z!At most %d colons permitted in %rz At most one '::' permitted in %rrz0Leading ':' only permitted as part of '::' in %rz1Trailing ':' only permitted as part of '::' in %rz/Expected at most %d other parts with '::' in %rz,Exactly %d parts expected without '::' in %r) r@rXrYrDrrr^ru _HEXTET_COUNTr*range _parse_hextetrG)rrr Z _min_partsrZipv4_intrZ _max_partsZ skip_indexr!Zparts_hiZparts_loZ parts_skippedrrrrrs    "               z_BaseV6._ip_int_from_stringcCs>|j|std|t|dkr4d}t||t|dS)a&Convert an IPv6 hextet string into an integer. Args: hextet_str: A string, the number to parse. Returns: The hextet as an integer. Raises: ValueError: if the input isn't strictly a hex number from [0..FFFF]. zOnly hex digits permitted in %rrz$At most 4 characters permitted in %rr) _HEX_DIGITSrrGrYr)rZ hextet_strrrrrr8>s     z_BaseV6._parse_hextetc Csd}d}d}d}t|D]>\}}|dkrN|d7}|dkr<|}||krV|}|}qd}d}q|dkr||}|t|kr~|dg7}dg|||<|dkrdg|}|S)aCompresses a list of hextets. Compresses a list of strings, replacing the longest continuous sequence of "0" in the list with "" and adding empty strings at the beginning or at the end of the string such that subsequently calling ":".join(hextets) will produce the compressed version of the IPv6 address. Args: hextets: A list of strings, the hextets to compress. Returns: A list of strings. r{rrr&) enumeraterY) rhextetsZbest_doublecolon_startZbest_doublecolon_lenZdoublecolon_startZdoublecolon_lenindexZhextetZbest_doublecolon_endrrr_compress_hextetsXs0   z_BaseV6._compress_hextetsNcsZ|dkrt|j}||jkr$tdd|fddtdddD}||}d |S) a,Turns a 128-bit integer into hexadecimal notation. Args: ip_int: An integer, the IP address. Returns: A string, the hexadecimal representation of the address. Raises: ValueError: The address is bigger than 128 bits of all ones. NzIPv6 address is too large%032xcs&g|]}dt||ddqS)r4rr)rr Zhex_strrrrsz/_BaseV6._string_from_ip_int..rr?rr3)rr^rnrGr7r>r)rrr<rr@rrs   z_BaseV6._string_from_ip_intcst|trt|j}nt|tr,t|j}nt|}||}d|fddtdddD}t|ttfrdd ||j fSd |S) zExpand a shortened IPv6 address. Args: ip_str: A string, the IPv6 address. Returns: A string, the expanded IPv6 address. r?csg|]}||dqS)rrr r@rrrsz8_BaseV6._explode_shorthand_ip_string..rr?rrr3) rrLrWrrPrcrr7rrr})r.rrr rr@rrs     z$_BaseV6._explode_shorthand_ip_stringcCs&|jddddd}d|dS)zReturn the reverse DNS pointer name for the IPv6 address. This implements the method described in RFC3596 2.5. Nr{r3r:rz .ip6.arpa)rreplacer)r.Z reverse_charsrrrrsz_BaseV6._reverse_pointercCs|jSrr rrrrr sz_BaseV6.max_prefixlencCs|jSrrrrrrrlsz_BaseV6.version)N)r;r<r=rBr>r| IPV6LENGTHrnr6rr9rmrrrrr8r>rrrrr rlrrrrr2s0   h  /   r2c@seZdZdZdZddZeddZeddZed d Z ed d Z ed dZ eddZ eddZ eddZeddZeddZeddZeddZdS)rEz/Represent and manipulate single IPv6 Addresses.rcCsxt|tr||||_dSt|trL||dt|}t|d|_dSt|}d|krht d|| ||_dS)aInstantiate a new IPv6 address object. Args: address: A string or integer representing the IP Additionally, an integer can be passed, so IPv6Address('2001:db8::') == IPv6Address(42540766411282592856903984951653826560) or, more generally IPv6Address(int(IPv6Address('2001:db8::'))) == IPv6Address('2001:db8::') Raises: AddressValueError: If address isn't a valid IPv6 address. NrrrVrrrrrrrs      zIPv6Address.__init__cCs t|jSr)rUr^rrrrrszIPv6Address.packedcCs ||jjkSrrrrrrrs zIPv6Address.is_multicastcstfddjjDS)rc3s|]}|kVqdSrrr rrrrsz*IPv6Address.is_reserved..)rr_reserved_networksrrrrrs zIPv6Address.is_reservedcCs ||jjkSrr"rrrrrszIPv6Address.is_link_localcCs ||jjkSa`Test if the address is reserved for site-local. Note that the site-local address space has been deprecated by RFC 3879. Use is_private to test if this address is in the space of unique local addresses as defined by RFC 4193. Returns: A boolean, True if the address is reserved per RFC 3513 2.5.6. )r_sitelocal_networkrrrr is_site_locals zIPv6Address.is_site_localcstfddjjDS)zTest if this address is allocated for private networks. Returns: A boolean, True if the address is reserved per iana-ipv6-special-registry. c3s|]}|kVqdSrrrrrrr3sz)IPv6Address.is_private..rrrrrr*s zIPv6Address.is_privatecCs|j S)zTest if this address is allocated for public networks. Returns: A boolean, true if the address is not reserved per iana-ipv6-special-registry. rrrrrr5s zIPv6Address.is_globalcCs |jdkS)rrrrrrrr@s zIPv6Address.is_unspecifiedcCs |jdkS)rr&rrrrrrKs zIPv6Address.is_loopbackcCs |jd?dkrdSt|jd@S)zReturn the IPv4 mapped address. Returns: If the IPv6 address is a v4 mapped address, return the IPv4 mapped address. Return None otherwise. r?r5Nr^rDrrrr ipv4_mappedVs zIPv6Address.ipv4_mappedcCs4|jd?dkrdSt|jd?d@t|jd@fS)zTuple of embedded teredo IPs. Returns: Tuple of the (server, client) IPs or None if the address doesn't appear to be a teredo address (doesn't start with 2001::/32) `i NrrGrHrrrrteredocs zIPv6Address.teredocCs$|jd?dkrdSt|jd?d@S)zReturn the IPv4 6to4 embedded address. Returns: The IPv4 6to4-embedded address if present or None if the address doesn't appear to contain a 6to4 embedded address. pi NPrGrHrrrr sixtofourrs zIPv6Address.sixtofourN)r;r<r=rBr>rrrrrrrFrrrrrIrKrNrrrrrEs6%           rEc@seZdZddZddZddZddZd d Zej Z e d d Z e d dZ e ddZ e ddZe ddZe ddZdS)rPcCst|ttfr2t||t|j|_|j|_ dSt|t rt||dt |dkrht |d|_ n|j|_ t|dd|_|jj |_ |jj|_dSt|}t||dt|dd|_|jj |_ |jj |_ |jj|_dSr$)rrFrrErrLr^rrmr}r%rYrrrr[r&rrrrs(        zIPv6Interface.__init__cCsd||j|jjfSrr'rrrrrs zIPv6Interface.__str__cCsFt||}|r|tkr|Sz|j|jkWStk r@YdSXdSr)rEr0r1rr~r(rrrr0s  zIPv6Interface.__eq__cCsRt||}|tkrtSz|j|jkp4|j|jko4|WStk rLYdSXdSr)rEr4r1rr~r)rrrr4s  zIPv6Interface.__lt__cCs|j|jAt|jjASrr*rrrrrszIPv6Interface.__hash__cCs t|jSr)rEr^rrrrrcszIPv6Interface.ipcCsd||j|jfSrr+rrrrrs zIPv6Interface.with_prefixlencCsd||j|jfSrr,rrrrrs zIPv6Interface.with_netmaskcCsd||j|jfSrr-rrrrrs zIPv6Interface.with_hostmaskcCs|jdko|jjSr#)r^rrrrrrrszIPv6Interface.is_unspecifiedcCs|jdko|jjSr)r^rrrrrrrszIPv6Interface.is_loopbackN)r;r<r=rrr0r4rrrrrcrrrrrrrrrrPs$       rPc@s2eZdZdZeZd ddZddZeddZ d S) rLavThis class represents and manipulates 128-bit IPv6 networks. Attributes: [examples for IPv6('2001:db8::1000/124')] .network_address: IPv6Address('2001:db8::1000') .hostmask: IPv6Address('::f') .broadcast_address: IPv6Address('2001:db8::100f') .netmask: IPv6Address('ffff:ffff:ffff:ffff:ffff:ffff:ffff:fff0') .prefixlen: 124 TcCs|t||t|ttfrs        )$ $#716=X*vSr 5W{!site-packages/pip/_vendor/__pycache__/appdirs.cpython-38.pyc000064400000047725151464452710020042 0ustar00U .e_ @s`dZdZdeeeZddlZddlZejddkZ e r>eZ ej drddl Z e ddZe drrdZqe d rd Zqd Znej Zd4d dZd5ddZd6ddZd7ddZd8ddZd9ddZd:ddZGdddeZddZd d!Zd"d#Zd$d%Zedkrfzdd&lmZeZWnBe k rdzddl!Z"eZWne k r^eZYnXYnXe#d'kr\d(Z$d)Z%d*Z&e'd+ee'd,ee$e%d-d.Z(e&D]Z)e'd/e)e*e(e)fqe'd0ee$e%Z(e&D]Z)e'd/e)e*e(e)fqe'd1ee$Z(e&D]Z)e'd/e)e*e(e)fqe'd2ee$d d3Z(e&D]Z)e'd/e)e*e(e)fq>dS);zyUtilities for determining application-specific dirs. See for details and usage. ).NrjavaZWindowswin32ZMacdarwinZlinux2FcCstdkr^|dkr|}|rdpd}tjt|}|r|dk rNtj|||}qtj||}nNtdkrtjd}|rtj||}n&tdtjd }|rtj||}|r|rtj||}|S) aJReturn full path to the user-specific data dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "roaming" (boolean, default False) can be set True to use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user data directories are: Mac OS X: ~/Library/Application Support/ Unix: ~/.local/share/ # or in $XDG_DATA_HOME, if defined Win XP (not roaming): C:\Documents and Settings\\Application Data\\ Win XP (roaming): C:\Documents and Settings\\Local Settings\Application Data\\ Win 7 (not roaming): C:\Users\\AppData\Local\\ Win 7 (roaming): C:\Users\\AppData\Roaming\\ For Unix, we follow the XDG spec and support $XDG_DATA_HOME. That means, by default "~/.local/share/". rN CSIDL_APPDATACSIDL_LOCAL_APPDATAFrz~/Library/Application Support/Z XDG_DATA_HOMEz~/.local/sharesystemospathnormpath_get_win_folderjoin expandusergetenv)appname appauthorversionroamingconstrr7/usr/lib/python3.8/site-packages/pip/_vendor/appdirs.py user_data_dir-s&   rcs tdkrR|dkr}tjtd}r|dk rBtj||}qtj|}ntdkrztjd}rtj|}nttdtjdd g}d d | tjD}r|rtj|fd d |D}|rtj|}n|d }|Sr|rtj||}|S)aiReturn full path to the user-shared data dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "multipath" is an optional parameter only applicable to *nix which indicates that the entire list of data dirs should be returned. By default, the first item from XDG_DATA_DIRS is returned, or '/usr/local/share/', if XDG_DATA_DIRS is not set Typical site data directories are: Mac OS X: /Library/Application Support/ Unix: /usr/local/share/ or /usr/share/ Win XP: C:\Documents and Settings\All Users\Application Data\\ Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) Win 7: C:\ProgramData\\ # Hidden, but writeable on Win 7. For Unix, this is using the $XDG_DATA_DIRS[0] default. WARNING: Do not use this on Windows. See the Vista-Fail note above for why. rNCSIDL_COMMON_APPDATAFrz/Library/Application SupportZ XDG_DATA_DIRSz/usr/local/sharez /usr/sharecSs g|]}tj|tjqSrr rrrstripsep.0xrrr sz!site_data_dir..csg|]}tj|gqSrr rrr rrrr#sr) r r rrrrrrpathsepsplitrrr multipathrZpathlistrr%r site_data_dirds6  r*cCsXtdkrt||d|}n&tdtjd}|r>tj||}|rT|rTtj||}|S)aReturn full path to the user-specific config dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "roaming" (boolean, default False) can be set True to use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user config directories are: Mac OS X: same as user_data_dir Unix: ~/.config/ # or in $XDG_CONFIG_HOME, if defined Win *: same as user_data_dir For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME. That means, by default "~/.config/". rrNZXDG_CONFIG_HOMEz ~/.configr rr rrrrrrrrrrrruser_config_dirsr.cstdkr*t|}r|rtj||}ndtdd}dd|tjD}rt|rbtj|fdd|D}|rtj|}n|d}|S)aReturn full path to the user-shared data dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "multipath" is an optional parameter only applicable to *nix which indicates that the entire list of config dirs should be returned. By default, the first item from XDG_CONFIG_DIRS is returned, or '/etc/xdg/', if XDG_CONFIG_DIRS is not set Typical site config directories are: Mac OS X: same as site_data_dir Unix: /etc/xdg/ or $XDG_CONFIG_DIRS[i]/ for each value in $XDG_CONFIG_DIRS Win *: same as site_data_dir Vista: (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.) For Unix, this is using the $XDG_CONFIG_DIRS[0] default, if multipath=False WARNING: Do not use this on Windows. See the Vista-Fail note above for why. r+ZXDG_CONFIG_DIRSz/etc/xdgcSs g|]}tj|tjqSrrr rrrr#sz#site_config_dir..csg|]}tj|gqSrr$r r%rrr#sr)r r*r rrrr'r&r(rr%rsite_config_dirs  r/TcCstdkrd|dkr|}tjtd}|r|dk rBtj|||}ntj||}|rtj|d}nNtdkrtjd}|rtj||}n&tdtjd }|rtj||}|r|rtj||}|S) aReturn full path to the user-specific cache dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "opinion" (boolean) can be False to disable the appending of "Cache" to the base app data dir for Windows. See discussion below. Typical user cache directories are: Mac OS X: ~/Library/Caches/ Unix: ~/.cache/ (XDG default) Win XP: C:\Documents and Settings\\Local Settings\Application Data\\\Cache Vista: C:\Users\\AppData\Local\\\Cache On Windows the only suggestion in the MSDN docs is that local settings go in the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming app data dir (the default returned by `user_data_dir` above). Apps typically put cache data somewhere *under* the given dir here. Some examples: ...\Mozilla\Firefox\Profiles\\Cache ...\Acme\SuperApp\Cache\1.0 OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value. This can be disabled with the `opinion=False` option. rNr FZCacherz~/Library/CachesZXDG_CACHE_HOMEz~/.cacher rrrZopinionrrrruser_cache_dirs(! r1cCsXtdkrt||d|}n&tdtjd}|r>tj||}|rT|rTtj||}|S)aReturn full path to the user-specific state dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "roaming" (boolean, default False) can be set True to use the Windows roaming appdata directory. That means that for users on a Windows network setup for roaming profiles, this user data will be sync'd on login. See for a discussion of issues. Typical user state directories are: Mac OS X: same as user_data_dir Unix: ~/.local/state/ # or in $XDG_STATE_HOME, if defined Win *: same as user_data_dir For Unix, we follow this Debian proposal to extend the XDG spec and support $XDG_STATE_HOME. That means, by default "~/.local/state/". r+NZXDG_STATE_HOMEz~/.local/stater,r-rrruser_state_dir:sr2cCstdkr tjtjd|}nNtdkrLt|||}d}|rntj|d}n"t|||}d}|rntj|d}|r|rtj||}|S)aReturn full path to the user-specific log dir for this application. "appname" is the name of application. If None, just the system directory is returned. "appauthor" (only used on Windows) is the name of the appauthor or distributing body for this application. Typically it is the owning company name. This falls back to appname. You may pass False to disable it. "version" is an optional version path element to append to the path. You might want to use this if you want multiple versions of your app to be able to run independently. If used, this would typically be ".". Only applied when appname is present. "opinion" (boolean) can be False to disable the appending of "Logs" to the base app data dir for Windows, and "log" to the base cache dir for Unix. See discussion below. Typical user log directories are: Mac OS X: ~/Library/Logs/ Unix: ~/.cache//log # or under $XDG_CACHE_HOME if defined Win XP: C:\Documents and Settings\\Local Settings\Application Data\\\Logs Vista: C:\Users\\AppData\Local\\\Logs On Windows the only suggestion in the MSDN docs is that local settings go in the `CSIDL_LOCAL_APPDATA` directory. (Note: I'm interested in examples of what some windows apps use for a logs dir.) OPINION: This function appends "Logs" to the `CSIDL_LOCAL_APPDATA` value for Windows and appends "log" to the user cache dir for Unix. This can be disabled with the `opinion=False` option. rz~/Library/LogsrFZLogslog)r r rrrrr1r0rrr user_log_dirds"    r4c@sneZdZdZdddZeddZedd Zed d Zed d Z eddZ eddZ eddZ dS)AppDirsz1Convenience wrapper for getting application dirs.NFcCs"||_||_||_||_||_dS)N)rrrrr))selfrrrrr)rrr__init__s zAppDirs.__init__cCst|j|j|j|jdSN)rr)rrrrrr6rrrrs zAppDirs.user_data_dircCst|j|j|j|jdSN)rr))r*rrrr)r9rrrr*s zAppDirs.site_data_dircCst|j|j|j|jdSr8)r.rrrrr9rrrr.s zAppDirs.user_config_dircCst|j|j|j|jdSr:)r/rrrr)r9rrrr/s zAppDirs.site_config_dircCst|j|j|jdSNr)r1rrrr9rrrr1s zAppDirs.user_cache_dircCst|j|j|jdSr;)r2rrrr9rrrr2s zAppDirs.user_state_dircCst|j|j|jdSr;)r4rrrr9rrrr4s zAppDirs.user_log_dir)NNNFF) __name__ __module__ __qualname____doc__r7propertyrr*r.r/r1r2r4rrrrr5s&       r5cCsHtrddl}nddl}dddd|}||jd}|||\}}|S)zThis is a fallback technique at best. I'm not sure if using the registry for this guarantees us the correct answer for all CSIDL_* names. rNZAppDatazCommon AppDataz Local AppDatar rr z@Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders)PY3winreg_winregOpenKeyHKEY_CURRENT_USERZ QueryValueEx) csidl_namerEZshell_folder_namekeydirtyperrr_get_win_folder_from_registrys rLcCsddlm}m}|dt||dd}z^t|}d}|D]}t|dkr8d}qRq8|rzddl}||}Wnt k rYnXWnt k rYnX|S)Nr)shellconshellFT) Zwin32com.shellrMrNSHGetFolderPathgetattrunicodeordwin32apiGetShortPathName ImportError UnicodeError)rHrMrNrJ has_high_charcrTrrr_get_win_folder_with_pywin32s$  rZcCsddl}dddd|}|d}|jjd|dd|d}|D]}t|dkr@d }qZq@|r|d}|jj|j|dr|}|jS) Nr#rBiFrOT) ctypesZcreate_unicode_bufferwindllZshell32ZSHGetFolderPathWrSZkernel32ZGetShortPathNameWvalue)rHr^Z csidl_constbufrXrYZbuf2rrr_get_win_folder_with_ctypess&   rbc Csddl}ddlm}ddlm}|jjd}|d|}|jj }| dt |j |d|j j ||j|d}d}|D]} t| dkr|d }qq||r|d|}|jj } | |||r|j|d}|S) Nr)jna)rrYFrOT)arrayZcom.sunrcZcom.sun.jna.platformrZWinDefZMAX_PATHZzerosZShell32ZINSTANCErPrQZShlObjZSHGFP_TYPE_CURRENTZNativeZtoStringZtostringrrSZKernel32rU) rHrfrcrZbuf_sizerarNrJrXrYZkernelrrr_get_win_folder_with_jnas&      rg)r___main__ZMyAppZ MyCompany)rr.r1r2r4r*r/z-- app dirs %s --z%-- app dirs (with optional 'version')z1.0r<z%s: %sz) -- app dirs (without optional 'version')z+ -- app dirs (without optional 'appauthor')z( -- app dirs (with disabled 'appauthor'))r)NNNF)NNNF)NNNF)NNNF)NNNT)NNNF)NNNT)+r@Z__version_info__rmapstr __version__sysr version_inforCrRplatform startswithZjava_verZos_namer rr*r.r/r1r2r4objectr5rLrZrbrgr^r_rrVZ com.sun.jnaZcomr=rrZpropsprintdirsZproprQrrrrsv     7 B ( 3 9 * 30      site-packages/pip/_vendor/__pycache__/retrying.cpython-38.pyc000064400000017506151464452710020235 0ustar00U .e&@slddlZddlmZddlZddlZddlZdZddZGdddeZ Gdd d eZ Gd d d e Z dS) N)sixi?csBtdkr,tdr,dd}|dSfdd}|SdS)z Decorator function that instantiates the Retrying object @param *dargs: positional arguments passed to Retrying object @param **dkw: keyword arguments passed to the Retrying object rcstfdd}|S)Ncstjf||SNRetryingcallargskwf8/usr/lib/python3.8/site-packages/pip/_vendor/retrying.py wrapped_f$sz-retry..wrap_simple..wrapped_frwrapsr rr r r wrap_simple"szretry..wrap_simplecstfdd}|S)Ncstjf||Srrr)dargsdkwr r rr/sz&retry..wrap..wrapped_frrrrr rwrap-szretry..wrapN)lencallable)rrrrr rrretrys  rc@sneZdZdddZddZddZd d Zd d Zd dZddZ ddZ ddZ ddZ ddZ ddZdS)rNFcs |dkr dn||_|dkrdn||_|dkr0dn||_|dkrBdn||_|dkrTdn||_|dkrfdn||_| dkrxdn| |_| dkrdn| |_| dkrtn| |_ |dkrdn||_ g|dk r̈ |j |dk r |j |dk r||_n&|dkr fdd|_n t|||_ddg|dk r6 |j|dk sJ|dk rV |j|dk sj| dk rv |j| dk s| dk r |j|dk r||_n&|dkr‡fd d|_n t|||_| dkr|j|_n| |_| dkr|j|_n| |_||_dS) NdrrcstfddDS)Nc3s|]}|VqdSrr .0r ZattemptsZdelayr r as6Retrying.__init__....)anyr ) stop_funcsr raz#Retrying.__init__..c_sdS)Nrr )r kwargsr r rr%hr&cstfddDS)Nc3s|]}|VqdSrr rr r rr!ysr")maxr ) wait_funcsr rr%yr&)_stop_max_attempt_number_stop_max_delay _wait_fixed_wait_random_min_wait_random_max_wait_incrementing_start_wait_incrementing_increment_wait_exponential_multiplierMAX_WAIT_wait_exponential_max_wait_jitter_maxappendstop_after_attemptstop_after_delaystopgetattr fixed_sleep random_sleepincrementing_sleepexponential_sleepwait always_reject_retry_on_exception never_reject_retry_on_result_wrap_exception)selfr8r>Zstop_max_attempt_numberZstop_max_delayZ wait_fixedZwait_random_minZwait_random_maxZwait_incrementing_startZwait_incrementing_incrementZwait_exponential_multiplierZwait_exponential_maxZretry_on_exceptionZretry_on_resultwrap_exceptionZ stop_funcZ wait_funcZwait_jitter_maxr )r$r)r__init__:sR                 zRetrying.__init__cCs ||jkS)z;Stop after the previous attempt >= stop_max_attempt_number.)r*rDprevious_attempt_numberdelay_since_first_attempt_msr r rr6szRetrying.stop_after_attemptcCs ||jkS)z=Stop after the time from the first attempt >= stop_max_delay.)r+rGr r rr7szRetrying.stop_after_delaycCsdS)z#Don't sleep at all before retrying.rr rGr r rno_sleepszRetrying.no_sleepcCs|jS)z0Sleep a fixed amount of time between each retry.)r,rGr r rr:szRetrying.fixed_sleepcCst|j|jS)zISleep a random amount of time between wait_random_min and wait_random_max)randomZrandintr-r.rGr r rr;szRetrying.random_sleepcCs$|j|j|d}|dkr d}|S)z Sleep an incremental amount of time after each attempt, starting at wait_incrementing_start and incrementing by wait_incrementing_increment rr)r/r0)rDrHrIresultr r rr<szRetrying.incrementing_sleepcCs2d|}|j|}||jkr"|j}|dkr.d}|S)Nr)r1r3)rDrHrIZexprLr r rr=s  zRetrying.exponential_sleepcCsdS)NFr rDrLr r rrAszRetrying.never_rejectcCsdS)NTr rNr r rr?szRetrying.always_rejectcCs4d}|jr |||jdO}n|||jO}|S)NFr) has_exceptionr@valuerB)rDattemptZrejectr r r should_rejects zRetrying.should_rejectc Ostttd}d}zt||||d}Wn t}t||d}YnX||sf||jStttd|}| ||r|js|j r|qt |n<| ||} |j rt|j } | td| } t| d|d7}qdS)NrrFTrg@@)introundtimeAttemptsysexc_inforRgetrCr8rO RetryErrorr>r4rKr(sleep) rDfnr r'Z start_timeattempt_numberrQtbrIr[Zjitterr r rrs(       z Retrying.call)NNNNNNNNNNNNNFNNN)__name__ __module__ __qualname__rFr6r7rJr:r;r<r=rAr?rRrr r r rr8s: S   rc@s*eZdZdZddZd ddZddZd S) rVz An Attempt encapsulates a call to a target function that may end as a normal return value from the function or an Exception depending on what occurred during the execution. cCs||_||_||_dSr)rPr]rO)rDrPr]rOr r rrFszAttempt.__init__FcCs@|jr6|rt|qs *!site-packages/pip/_vendor/__pycache__/pyparsing.cpython-38.opt-1.pyc000064400000705501151464452710021344 0ustar00U .e,s@sdZdZdZdZddlZddlmZddlZddl Z ddl Z ddl Z ddl Z ddl Z ddlZddlZddlZddlmZddlmZddlZdd lmZzdd lmZWn ek rdd lmZYnXzdd lmZWn ek rdd lmZYnXz dd lmZddlmZm Z Wn2ek rNdd l mZddl mZm Z YnXzddl m!Z"WnBek rzddl#m!Z"Wnek rdZ"YnXYnXzddlm$Z$Wn$ek rGdddZ$YnXe$Z%de%_de%_&e$Z'de'_de'_(de'_)de'_*de'_+de'_,dddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dddddddddddgsZ-e.e j/ddZ0e0ddkZ1e1rVe j2Z3e4Z5e6Z7e4Z8e4Z9e:e;ee.e?e@eAeBeCg ZDn`e jEZ3eFZGddZ9gZDddlHZHdID]8ZJzeDKeLeHeJWneMk rYq|YnXq|eNddeGdDZOddZPejQejRZSdZTeTdZUeSeTZVe6dZWdXddejYDZZdYddZ[Gdd4d4e\Z]Gdd6d6e]Z^Gdd8d8e]Z_Gdd:d:e_Z`Gdd=d=e\ZaGdddebZcGdd9d9ebZdeeedddQZfddbZgdd_ZhddZiddZjddZkddiZldZddZmGdd;d;ebZnGdddenZoGddCdCenZpGdd#d#epZqGdd.d.epZrGdd+d+epZsGdddesZtesZuesen_vGdd(d(epZwGdddesZxGdddewZyGdddepZzGddFdFepZ{Gddde{Z|GddJdJe|Z}Gdd>d>epZ~Gdd<dddeZejjjejjjejjjej_e1 rXeed?ejeed@ejeedAejeedBejeedCejeedDejeejdEejjeejdFejjeejdGejjeedHejeedIejeedJejedKkrexdLZexdMZe{eSeVdNZeedOddPeZeeedQZdReBZeedOddPeZeeedSZedTedQeedSZedUejdVejdVejdWddlZejeejejdXdS(aa pyparsing module - Classes and methods to define and execute parsing grammars ============================================================================= The pyparsing module is an alternative approach to creating and executing simple grammars, vs. the traditional lex/yacc approach, or the use of regular expressions. With pyparsing, you don't need to learn a new syntax for defining grammars or matching expressions - the parsing module provides a library of classes that you use to construct the grammar directly in Python. Here is a program to parse "Hello, World!" (or any greeting of the form ``", !"``), built up using :class:`Word`, :class:`Literal`, and :class:`And` elements (the :class:`'+'` operators create :class:`And` expressions, and the strings are auto-converted to :class:`Literal` expressions):: from pip._vendor.pyparsing import Word, alphas # define grammar of a greeting greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) The program outputs the following:: Hello, World! -> ['Hello', ',', 'World', '!'] The Python representation of the grammar is quite readable, owing to the self-explanatory class names, and the use of '+', '|' and '^' operators. The :class:`ParseResults` object returned from :class:`ParserElement.parseString` can be accessed as a nested list, a dictionary, or an object with named attributes. The pyparsing module handles some of the problems that are typically vexing when writing text parsers: - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello , World !", etc.) - quoted strings - embedded comments Getting Started - ----------------- Visit the classes :class:`ParserElement` and :class:`ParseResults` to see the base classes that most other pyparsing classes inherit from. Use the docstrings for examples of how to: - construct literal match expressions from :class:`Literal` and :class:`CaselessLiteral` classes - construct character word-group expressions using the :class:`Word` class - see how to create repetitive expressions using :class:`ZeroOrMore` and :class:`OneOrMore` classes - use :class:`'+'`, :class:`'|'`, :class:`'^'`, and :class:`'&'` operators to combine simple expressions into more complex ones - associate names with your parsed results using :class:`ParserElement.setResultsName` - access the parsed data, which is returned as a :class:`ParseResults` object - find some helpful expression short-cuts like :class:`delimitedList` and :class:`oneOf` - find more useful common expressions in the :class:`pyparsing_common` namespace class z2.4.2z29 Jul 2019 02:58 UTCz*Paul McGuire N)ref)datetime) itemgetter)wraps) filterfalse) ifilterfalse)RLock)Iterable)MutableMappingMapping) OrderedDict)SimpleNamespacec@s eZdZdS)r N)__name__ __module__ __qualname__rr9/usr/lib/python3.8/site-packages/pip/_vendor/pyparsing.pyr sr aA A cross-version compatibility configuration for pyparsing features that will be released in a future version. By setting values in this configuration to True, those features can be enabled in prior versions for compatibility development and testing. - collect_all_And_tokens - flag to enable fix for Issue #63 that fixes erroneous grouping of results names when an And expression is nested within an Or or MatchFirst; set to True to enable bugfix released in pyparsing 2.3.0, or False to preserve pre-2.3.0 handling of named results Ta Diagnostic configuration (all default to False) - warn_multiple_tokens_in_named_alternation - flag to enable warnings when a results name is defined on a MatchFirst or Or expression with one or more And subexpressions (only warns if __compat__.collect_all_And_tokens is False) - warn_ungrouped_named_tokens_in_collection - flag to enable warnings when a results name is defined on a containing expression with ungrouped subexpressions that also have results names - warn_name_set_on_empty_Forward - flag to enable warnings whan a Forward is defined with a results name, but has no contents defined - warn_on_multiple_string_args_to_oneof - flag to enable warnings whan oneOf is incorrectly called with multiple str arguments - enable_debug_on_named_expressions - flag to auto-enable debug on all subsequent calls to ParserElement.setName() F __version____versionTime__ __author__ __compat____diag__AndCaselessKeywordCaselessLiteral CharsNotInCombineDictEachEmpty FollowedByForward GoToColumnGroupKeywordLineEnd LineStartLiteral PrecededBy MatchFirstNoMatchNotAny OneOrMoreOnlyOnceOptionalOrParseBaseExceptionParseElementEnhanceParseExceptionParseExpressionParseFatalException ParseResultsParseSyntaxException ParserElement QuotedStringRecursiveGrammarExceptionRegexSkipTo StringEnd StringStartSuppressTokenTokenConverterWhiteWordWordEnd WordStart ZeroOrMoreChar alphanumsalphas alphas8bit anyCloseTag anyOpenTag cStyleCommentcolcommaSeparatedListcommonHTMLEntity countedArraycppStyleCommentdblQuotedStringdblSlashComment delimitedListdictOfdowncaseTokensemptyhexnums htmlCommentjavaStyleCommentlinelineEnd lineStartlineno makeHTMLTags makeXMLTagsmatchOnlyAtColmatchPreviousExprmatchPreviousLiteral nestedExprnullDebugActionnumsoneOfopAssocoperatorPrecedence printablespunc8bitpythonStyleComment quotedString removeQuotesreplaceHTMLEntity replaceWith restOfLinesglQuotedStringsrange stringEnd stringStarttraceParseAction unicodeString upcaseTokens withAttribute indentedBlockoriginalTextForungroup infixNotation locatedExpr withClass CloseMatchtokenMappyparsing_commonpyparsing_unicode unicode_setconditionAsParseActioncCsft|tr|Sz t|WStk r`t|td}td}|dd| |YSXdS)aDrop-in replacement for str(obj) that tries to be Unicode friendly. It first tries str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It then < returns the unicode object | encodes it with the default encoding | ... >. xmlcharrefreplacez&#\d+;cSs$dtt|dddddS)Nz\ur)hexinttrrrz_ustr..N) isinstanceunicodestrUnicodeEncodeErrorencodesysgetdefaultencodingr:setParseActiontransformString)objretZ xmlcharrefrrr_ustrs  rz6sum len sorted reversed list tuple set any all min maxccs|] }|VqdSNr).0yrrr srcCs:d}dddD}t||D]\}}|||}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)&;Nr)rsrrrr sz_xml_escape..zamp gt lt quot apos)splitzipreplace)dataZ from_symbolsZ to_symbolsZfrom_Zto_rrr _xml_escapes r 0123456789Z ABCDEFabcdef\ccs|]}|tjkr|VqdSr)stringZ whitespacercrrrrs cs@|dk r |nd|rtntttfdd}|S)Nzfailed user-defined conditioncs t|||s||dSr)boolrlrexc_typefnmsgrrpasz"conditionAsParseAction..pa)r4r2 _trim_arityr)rmessagefatalrrrrrs  c@sPeZdZdZdddZeddZdd Zd d Zd d Z dddZ ddZ dS)r0z7base exception class for all parsing runtime exceptionsrNcCs>||_|dkr||_d|_n ||_||_||_|||f|_dSNr)locrpstr parserElementargs)selfrrrelemrrr__init__(szParseBaseException.__init__cCs||j|j|j|jS)z internal factory method to simplify creating one type of ParseException from another - avoids having __init__ signature conflicts among subclasses )rrrr)clsperrr_from_exception3sz"ParseBaseException._from_exceptioncCsN|dkrt|j|jS|dkr,t|j|jS|dkrBt|j|jSt|dS)zsupported attributes by name are: - lineno - returns the line number of the exception text - col - returns the column number of the exception text - line - returns the line containing the exception text r^)rMcolumnr[N)r^rrrMr[AttributeError)rZanamerrr __getattr__;szParseBaseException.__getattr__cCs^|jr@|jt|jkrd}qDd|j|j|jddd}nd}d|j||j|j|jfS)Nz, found end of textz , found %rrz\\\rz%%s%s (at char %d), (line:%d, col:%d))rrlenrrr^r)rZfoundstrrrr__str__Js$zParseBaseException.__str__cCst|Srrrrrr__repr__TszParseBaseException.__repr__>!z{0} )inspectrgetrecursionlimitrr0appendr[rMformatrrZgetinnerframes __traceback__set enumeratef_localsgetr7f_codeco_nameaddrr) excdepthrrZcallersseeniZffZfrmZf_selfZ self_typecoderrrexplainzsL          zParseException.explainN)r)rrrr staticmethodrrrrrr2csc@seZdZdZdS)r4znuser-throwable exception thrown when inconsistent parse content is found; stops all parsing immediatelyNrrrrrrrrr4sc@seZdZdZdS)r6zjust like :class:`ParseFatalException`, but thrown internally when an :class:`ErrorStop` ('-' operator) indicates that parsing is to stop immediately because an unbacktrackable syntax error has been found. Nrrrrrr6sc@s eZdZdZddZddZdS)r9ziexception thrown by :class:`ParserElement.validate` if the grammar could be improperly recursive cCs ||_dSrZparseElementTracerparseElementListrrrrsz"RecursiveGrammarException.__init__cCs d|jS)NzRecursiveGrammarException: %srrrrrrsz!RecursiveGrammarException.__str__N)rrrrrrrrrrr9sc@s,eZdZddZddZddZddZd S) _ParseResultsWithOffsetcCs||f|_dSrtup)rZp1Zp2rrrrsz _ParseResultsWithOffset.__init__cCs |j|Srrrrrrr __getitem__sz#_ParseResultsWithOffset.__getitem__cCst|jdSNr)reprrrrrrrsz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSrrrrrr setOffsetsz!_ParseResultsWithOffset.setOffsetN)rrrrrrrrrrrrsrc@seZdZdZd]ddZddddefddZdd Zefd d Zd d Z ddZ ddZ ddZ e Z ddZddZddZddZddZereZeZeZn$eZeZeZddZd d!Zd"d#Zd$d%Zd&d'Zd^d(d)Zd*d+Zd,d-Zd.d/Zd0d1Z d2d3Z!d4d5Z"d6d7Z#d8d9Z$d:d;Z%d`` - see :class:`ParserElement.setResultsName`) Example:: integer = Word(nums) date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: # date_str = integer("year") + '/' + integer("month") + '/' + integer("day") # parseString returns a ParseResults object result = date_str.parseString("1999/12/31") def test(s, fn=repr): print("%s -> %s" % (s, fn(eval(s)))) test("list(result)") test("result[0]") test("result['month']") test("result.day") test("'month' in result") test("'minutes' in result") test("result.dump()", str) prints:: list(result) -> ['1999', '/', '12', '/', '31'] result[0] -> '1999' result['month'] -> '12' result.day -> '31' 'month' in result -> True 'minutes' in result -> False result.dump() -> ['1999', '/', '12', '/', '31'] - day: 31 - month: 12 - year: 1999 NTcCs"t||r|St|}d|_|SNT)robject__new___ParseResults__doinit)rtoklistnameasListmodalZretobjrrrrs   zParseResults.__new__c Csd|jrvd|_d|_d|_i|_||_||_|dkr6g}||trP|dd|_n||trft||_n|g|_t |_ |dk r`|r`|sd|j|<||t rt |}||_||t dttfr|ddgfks`||tr|g}|r*||trtt|jd||<ntt|dd||<|||_n6z|d||<Wn$tttfk r^|||<YnXdS)NFrr)r_ParseResults__name_ParseResults__parent_ParseResults__accumNames_ParseResults__asList_ParseResults__modallist_ParseResults__toklist_generatorTypedict_ParseResults__tokdictrrr basestringr5rKeyError TypeError IndexError)rrrrrrrrrr"sB     $   zParseResults.__init__cCsPt|ttfr|j|S||jkr4|j|ddStdd|j|DSdS)NrrcSsg|] }|dqSrrrvrrr Psz,ParseResults.__getitem__..)rrslicer rrr5rrrrrIs   zParseResults.__getitem__cCs||tr0|j|t|g|j|<|d}nD||ttfrN||j|<|}n&|j|tt|dg|j|<|}||trt||_ dSr) rrrr rrr r5wkrefr)rkrrsubrrr __setitem__Rs   " zParseResults.__setitem__c Cst|ttfrt|j}|j|=t|trH|dkr:||7}t||d}tt||}||j D]>\}}|D]0}t |D]"\}\}} t || | |k||<qqxqln|j |=dSNrr) rrrrr r rangeindicesreverseritemsrr) rrZmylenZremovedr occurrencesjrvaluepositionrrr __delitem___s  zParseResults.__delitem__cCs ||jkSr)r)rrrrr __contains__tszParseResults.__contains__cCs t|jSr)rr rrrr__len__wszParseResults.__len__cCs |j Srr rrrr__bool__zszParseResults.__bool__cCs t|jSriterr rrrr__iter__~szParseResults.__iter__cCst|jdddSNrr+rrrr __reversed__szParseResults.__reversed__cCs$t|jdr|jSt|jSdS)Niterkeys)hasattrrr0r,rrrr _iterkeyss  zParseResults._iterkeyscsfddDS)Nc3s|]}|VqdSrrrrrrrrsz+ParseResults._itervalues..r2rrrr _itervaluesszParseResults._itervaluescsfddDS)Nc3s|]}||fVqdSrrr3rrrrsz*ParseResults._iteritems..r4rrrr _iteritemsszParseResults._iteritemscCs t|S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)r r0rrrrkeysszParseResults.keyscCs t|S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r  itervaluesrrrrvaluesszParseResults.valuescCs t|S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r  iteritemsrrrrr!szParseResults.itemscCs t|jS)zSince keys() returns an iterator, this method is helpful in bypassing code that looks for the existence of any defined results names.)rrrrrrhaskeysszParseResults.haskeyscOs|s dg}|D]*\}}|dkr0|d|f}qtd|qt|dtsdt|dksd|d|kr~|d}||}||=|S|d}|SdS)a Removes and returns item at specified index (default= ``last``). Supports both ``list`` and ``dict`` semantics for ``pop()``. If passed no argument or an integer argument, it will use ``list`` semantics and pop tokens from the list of parsed tokens. If passed a non-integer argument (most likely a string), it will use ``dict`` semantics and pop the corresponding value from any defined results names. A second default return value argument is supported, just as in ``dict.pop()``. Example:: def remove_first(tokens): tokens.pop(0) print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321'] label = Word(alphas) patt = label("LABEL") + OneOrMore(Word(nums)) print(patt.parseString("AAB 123 321").dump()) # Use pop() in a parse action to remove named result (note that corresponding value is not # removed from list form of results) def remove_LABEL(tokens): tokens.pop("LABEL") return tokens patt.addParseAction(remove_LABEL) print(patt.parseString("AAB 123 321").dump()) prints:: ['AAB', '123', '321'] - LABEL: AAB ['AAB', '123', '321'] rdefaultrz-pop() got an unexpected keyword argument '%s'rN)r!rrrr)rrkwargsrrindexrZ defaultvaluerrrpops"%  zParseResults.popcCs||kr||S|SdS)a[ Returns named result matching the given key, or if there is no such name, then returns the given ``defaultValue`` or ``None`` if no ``defaultValue`` is specified. Similar to ``dict.get()``. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString("1999/12/31") print(result.get("year")) # -> '1999' print(result.get("hour", "not specified")) # -> 'not specified' print(result.get("hour")) # -> None Nr)rkey defaultValuerrrrszParseResults.getcCsR|j|||jD]4\}}t|D]"\}\}}t||||k||<q(qdS)a Inserts new element at location index in the list of parsed tokens. Similar to ``list.insert()``. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to insert the parse location in the front of the parsed results def insert_locn(locn, tokens): tokens.insert(0, locn) print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321'] N)r insertrr!rr)rr>ZinsStrrr"rr$r%rrrrBszParseResults.insertcCs|j|dS)a Add single element to end of ParseResults list of elements. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to compute the sum of the parsed integers, and add it to the end def append_sum(tokens): tokens.append(sum(map(int, tokens))) print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444] N)r r)ritemrrrrs zParseResults.appendcCs&t|tr||n |j|dS)a  Add sequence of elements to end of ParseResults list of elements. Example:: patt = OneOrMore(Word(alphas)) # use a parse action to append the reverse of the matched strings, to make a palindrome def make_palindrome(tokens): tokens.extend(reversed([t[::-1] for t in tokens])) return ''.join(tokens) print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl' N)rr5__iadd__r extend)rZitemseqrrrrE's  zParseResults.extendcCs|jdd=|jdS)z7 Clear all elements and results names. N)r rclearrrrrrF:s zParseResults.clearcCs&z ||WStk r YdSXdSr)rrrrrrrAs zParseResults.__getattr__cCs|}||7}|Srcopy)rotherrrrr__add__GszParseResults.__add__cs|jrjt|jfdd|j}fdd|D}|D],\}}|||<t|dtr.c s4g|],\}}|D]}|t|d|dfqqSrr)rrrvlistr) addoffsetrrrQsz)ParseResults.__iadd__..r) rrr r!rr5rrrupdate)rrJZ otheritemsZotherdictitemsrrr)rQrMrrDLs     zParseResults.__iadd__cCs&t|tr|dkr|S||SdSr)rrrIrrJrrr__radd__\szParseResults.__radd__cCsdt|jt|jfS)Nz(%s, %s))rr rrrrrrdszParseResults.__repr__cCsdddd|jDdS)N[, css(|] }t|trt|nt|VqdSr)rr5rrrrrrrrhsz'ParseResults.__str__..])rr rrrrrgszParseResults.__str__rcCsLg}|jD]<}|r |r ||t|tr8||7}q |t|q |Sr)r rrr5 _asStringListr)rsepoutrCrrrrYjs   zParseResults._asStringListcCsdd|jDS)ax Returns the parse results as a nested list of matching tokens, all converted to strings. Example:: patt = OneOrMore(Word(alphas)) result = patt.parseString("sldkj lsdkj sldkj") # even though the result prints in string-like form, it is actually a pyparsing ParseResults print(type(result), result) # -> ['sldkj', 'lsdkj', 'sldkj'] # Use asList() to create an actual list result_list = result.asList() print(type(result_list), result_list) # -> ['sldkj', 'lsdkj', 'sldkj'] cSs"g|]}t|tr|n|qSr)rr5r)rresrrrrsz'ParseResults.asList..r)rrrrruszParseResults.asListcs6tr |j}n|j}fddtfdd|DS)a Returns the named parse results as a nested dictionary. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString('12/31/1999') print(type(result), repr(result)) # -> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]}) result_dict = result.asDict() print(type(result_dict), repr(result_dict)) # -> {'day': '1999', 'year': '12', 'month': '31'} # even though a ParseResults supports dict-like access, sometime you just need to have a dict import json print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"} cs6t|tr.|r|Sfdd|DSn|SdS)Ncsg|] }|qSrrrtoItemrrrsz7ParseResults.asDict..toItem..)rr5r;asDictrr]rrr^s  z#ParseResults.asDict..toItemc3s|]\}}||fVqdSrrrrrr]rrrsz&ParseResults.asDict..)PY_3r!r:r)rZitem_fnrr]rr_s  zParseResults.asDictcCs<t|j}t|j|_|j|_|j|j|j|_|S)zG Returns a new copy of a :class:`ParseResults` object. ) r5r rrr!rrrRrrrrrrrIs  zParseResults.copyFc CsLd}g}tdd|jD}|d}|s8d}d}d}d} |dk rJ|} n |jrV|j} | sf|rbdSd} |||d| d g7}t|jD]\} } t| tr| |kr|| || |o|dk||g7}n|| d|o|dk||g7}qd} | |kr|| } | s|rqnd} t t | } |||d| d | d | d g 7}q|||d | d g7}d |S) z (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names. rcss(|] \}}|D]}|d|fVqqdSrNrrOrrrrsz%ParseResults.asXML.. rNZITEM<>.z %s%s- %s: rer)rjfull include_list_depthcss|]}t|tVqdSr)rr5)rvvrrrrKsz %s%s[%d]: %s%s%s) rrrr;sortedr!rr5dumpranyrr) rrjrprqrrr[NLr!rrrrsrrrru!sP         zParseResults.dumpcOstj|f||dS)a# Pretty-printer for parsed results as a list, using the `pprint `_ module. Accepts additional positional or keyword args as defined for `pprint.pprint `_ . Example:: ident = Word(alphas, alphanums) num = Word(nums) func = Forward() term = ident | num | Group('(' + func + ')') func <<= ident + Group(Optional(delimitedList(term))) result = func.parseString("fna a,b,(fnb c,d,200),100") result.pprint(width=40) prints:: ['fna', ['a', 'b', ['(', 'fnb', ['c', 'd', '200'], ')'], '100']] N)pprintrrrr=rrrrxbszParseResults.pprintcCs.|j|j|jdk r|p d|j|jffSr)r rrIrrrrrrr __getstate__~szParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j||dk rDt||_nd|_dSr)r rrrrRrr)rstaternZ inAccumNamesrrr __setstate__s   zParseResults.__setstate__cCs|j|j|j|jfSr)r rr r rrrr__getnewargs__szParseResults.__getnewargs__cCstt|t|Sr)rrr r7rrrrrszParseResults.__dir__cCsrdd}|g}|D]>\}}t|tr>||j||d7}q|||g|||d7}q|dk rn||g|d}|S)z Helper classmethod to construct a ParseResults from a dict, preserving the name-value relations as results names. If an optional 'name' argument is given, a nested ParseResults will be returned cSsHz t|Wntk r"YdSXtr8t|ttf St|t SdSNF)r, Exceptionrbrrbytesrr`rrr is_iterables z+ParseResults.from_dict..is_iterabler)rrN)r!rr from_dict)rrJrrrrrrrrrs  zParseResults.from_dict)NNTT)N)r)NFrT)rTTr)N)6rrrrrrrrrr&r'r(r* __nonzero__r-r/r2r5r6rbr7r9r!r0r8r:r;r?rrBrrErFrrKrDrTrrrYrr_rIrirlrorurxrzr|r}rrrrrrrr5sl* ' 7  $ =( A cCsF|}d|krt|kr4nn||ddkr4dS||dd|S)aReturns current column within a string, counting newlines as line separators. The first column is number 1. Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See :class:`ParserElement.parseString` for more information on parsing strings containing ```` s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. rrr)rrfind)rstrgrrrrrMs cCs|dd|dS)aReturns current line number within a string, counting newlines as line separators. The first line is number 1. Note - the default parsing behavior is to expand tabs in the input string before starting the parsing process. See :class:`ParserElement.parseString` for more information on parsing strings containing ```` s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. rrr)count)rrrrrr^s cCsF|dd|}|d|}|dkr2||d|S||ddSdS)zfReturns the line of text containing loc within a string, counting newlines as line separators. rrrN)rfind)rrZlastCRZnextCRrrrr[s  cCs8tdt|dt|dt||t||fdS)NzMatch z at loc z(%d,%d))printrr^rM)instringrexprrrr_defaultStartDebugActionsrcCs$tdt|dt|dS)NzMatched z -> )rrrr)rstartlocZendlocrtoksrrr_defaultSuccessDebugActionsrcCstdt|dS)NzException raised:)rr)rrrrrrr_defaultExceptionDebugActionsrcGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nr)rrrrresrcstkrfddSdgdgtdddkrFddd}dd d n tj}tjd }|dd d }|d|d|ffdd}d}ztdtdj}Wntk rt}YnX||_|S)Ncs|Srrr)funcrrrrz_trim_arity..rFr)rcSs8tdkr dnd}tj| |dd|}|ddgS)N)rrrrlimitr)system_version traceback extract_stack)rrM frame_summaryrrrrsz"_trim_arity..extract_stackcSs$tj||d}|d}|ddgS)Nrrr)r extract_tb)tbrZframesrrrrrsz_trim_arity..extract_tbrrrcsz"|dd}dd<|WStk rdr>nNz.td}|dddddksjW5z~Wntk rYnXXdkrdd7<YqYqXqdS)NrTrrrr)r NameErrorrexc_info)rrrrZ foundArityrrmaxargsZpa_call_line_synthrrrs&  z_trim_arity..wrapperzr __class__)r)r) singleArgBuiltinsrrrrgetattrrrr)rrrZ LINE_DIFFZ this_liner func_namerrrrs,    rcseZdZdZdZdZeddZeddZddd Z d d Z d d Z dddZ dddZ dddZddZddZddZddZddZdd Zdd!d"Zd#d$Zdd%d&Zd'd(Zd)d*ZGd+d,d,eZed-k rGd.d/d/eZnGd0d/d/eZiZeZ d1d1gZ!dd2d3Z"eZ#ed4d5Z$dZ%edd7d8Z&dd9d:Z'e(dfd;d<Z)d=d>Z*e(fd?d@Z+e(dfdAdBZ,dCdDZ-dEdFZ.dGdHZ/dIdJZ0dKdLZ1dMdNZ2dOdPZ3dQdRZ4dSdTZ5dUdVZ6dWdXZ7dYdZZ8d[d\Z9d]d^Z:d_d`Z;ddadbZdgdhZ?didjZ@dkdlZAdmdnZBddodpZCdqdrZDdsdtZEdudvZFdwdxZGddydzZHdd{d|ZIfd}d~ZJddZKddZLddZMddZNdddZOdddZPZQS)r7z)Abstract base level parser element class.z FcCs |t_dS)a Overrides the default whitespace chars Example:: # default whitespace chars are space, and newline OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def', 'ghi', 'jkl'] # change to just treat newline as significant ParserElement.setDefaultWhitespaceChars(" \t") OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def'] N)r7DEFAULT_WHITE_CHARScharsrrrsetDefaultWhitespaceCharsDsz'ParserElement.setDefaultWhitespaceCharscCs |t_dS)ah Set class to be used for inclusion of string literals into a parser. Example:: # default literal class used is Literal integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # change to Suppress ParserElement.inlineLiteralsUsing(Suppress) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '12', '31'] N)r7_literalStringClassrrrrinlineLiteralsUsingTsz!ParserElement.inlineLiteralsUsingcCst|_d|_d|_d|_||_d|_ttj |_ d|_ d|_ d|_ t|_d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr)NNN)r  parseAction failActionstrRepr resultsName saveAsListskipWhitespacerr7r whiteCharscopyDefaultWhiteCharsmayReturnEmptykeepTabs ignoreExprsdebug streamlined mayIndexErrorerrmsg modalResults debugActionsre callPreparse callDuringTry)rsavelistrrrrjs( zParserElement.__init__cCs<t|}|jdd|_|jdd|_|jr8tj|_|S)a/ Make a copy of this :class:`ParserElement`. Useful for defining different parse actions for the same parsing pattern, using copies of the original parse element. Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) integerK = integer.copy().addParseAction(lambda toks: toks[0] * 1024) + Suppress("K") integerM = integer.copy().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M") print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M")) prints:: [5120, 100, 655360, 268435456] Equivalent form of ``expr.copy()`` is just ``expr()``:: integerM = integer().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M") N)rIrrrr7rr)rZcpyrrrrIs  zParserElement.copycCs$||_d|j|_tjr ||S)a_ Define name for this expression, makes debugging and exception messages clearer. Example:: Word(nums).parseString("ABC") # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1) Word(nums).setName("integer").parseString("ABC") # -> Exception: Expected integer (at char 0), (line:1, col:1) Expected )rrr!enable_debug_on_named_expressionssetDebugrGrrrsetNames  zParserElement.setNamecCs |||S)aO Define name for referencing matching tokens as a nested attribute of the returned parse results. NOTE: this returns a *copy* of the original :class:`ParserElement` object; this is so that the client can define a basic element, such as an integer, and reference it in multiple places with different names. You can also set results names using the abbreviated syntax, ``expr("name")`` in place of ``expr.setResultsName("name")`` - see :class:`__call__`. Example:: date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: date_str = integer("year") + '/' + integer("month") + '/' + integer("day") )_setResultsNamerrlistAllMatchesrrrsetResultsNameszParserElement.setResultsNamecCs4|}|dr"|dd}d}||_| |_|S)N*rT)rIendswithrr)rrrZnewselfrrrrs  zParserElement._setResultsNameTcs@|r&|jdfdd }|_||_nt|jdr<|jj|_|S)zMethod to invoke the Python pdb debugger when this element is about to be parsed. Set ``breakFlag`` to True to enable, False to disable. Tcsddl}|||||Sr)pdbZ set_trace)rr doActions callPreParserZ _parseMethodrrbreakersz'ParserElement.setBreak..breaker_originalParseMethod)TT)_parserr1)rZ breakFlagrrrrsetBreaks  zParserElement.setBreakcOsVt|dgkrg|_n`` s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. Example:: integer = Word(nums) date_str = integer + '/' + integer + '/' + integer date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # use parse action to convert to ints at parse time integer = Word(nums).setParseAction(lambda toks: int(toks[0])) date_str = integer + '/' + integer + '/' + integer # note that integer fields are now ints, not strings date_str.parseString("1999/12/31") # -> [1999, '/', 12, '/', 31] Ncss|]}t|VqdSr)callable)rrrrrr sz/ParserElement.setParseAction..zparse actions must be callablerF)r rallrmaprrrrfnsr=rrrrs(zParserElement.setParseActioncOs4|jtttt|7_|jp,|dd|_|S)z Add one or more parse actions to expression's list of parse actions. See :class:`setParseAction`. See examples in :class:`copy`. rF)rr rrrrrrrraddParseActionszParserElement.addParseActionc OsF|D](}|jt||d|dddq|jp>|dd|_|S)aAdd a boolean predicate function to expression's list of parse actions. See :class:`setParseAction` for function call signatures. Unlike ``setParseAction``, functions passed to ``addCondition`` need to return boolean success/fail of the condition. Optional keyword arguments: - message = define a custom message to be used in the raised exception - fatal = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) year_int = integer.copy() year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later") date_str = year_int + '/' + integer + '/' + integer result = date_str.parseString("1999/12/31") # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1) rrF)rrr)rrrrr)rrr=rrrr addConditions  zParserElement.addConditioncCs ||_|S)aDefine action to perform if parsing fails at this expression. Fail acton fn is a callable function that takes the arguments ``fn(s, loc, expr, err)`` where: - s = string being parsed - loc = location where expression match was attempted and failed - expr = the parse expression that failed - err = the exception thrown The function returns no value. It may throw :class:`ParseFatalException` if it is desired to stop parsing immediately.)rrrrrr setFailAction4s zParserElement.setFailActionc CsNd}|rJd}|jD]4}z|||\}}d}qWqtk rDYqXqq|SNTF)rrr2)rrrZ exprsFoundeZdummyrrr_skipIgnorablesAs   zParserElement._skipIgnorablescCsH|jr|||}|jrD|j}t|}||krD|||krD|d7}q&|SNr)rrrrr)rrrZwtinstrlenrrrpreParseNs  zParserElement.preParsecCs|gfSrrrrrrrrrrZszParserElement.parseImplcCs|Srrrrr tokenlistrrr postParse]szParserElement.postParsec Csd\}}}|j}|s|jr"|j|r8|j||||z|rR|jrR|||} n|} | } |jsl| t|krz||| |\}} Wqtk rt |t||j |YqXn||| |\}} Wn\t k r} z<|j|r|j||| || |jr ||| || W5d} ~ XYnXn|r>|jr>|||} n|} | } |js\| t|krz||| |\}} Wn*tk rt |t||j |YnXn||| |\}} | ||| } t | |j|j|jd} |jr`|s|jr`|rz|jD]}z||| | } Wn6tk rD}zt d}||_|W5d}~XYnX| dk r| | k rt | |j|jovt| t tf|jd} qWnFt k r} z&|j|r|j||| || W5d} ~ XYnXn|jD]}z||| | } Wn6tk r }zt d}||_|W5d}~XYnX| dk r| | k rt | |j|joRt| t tf|jd} q|r|j|r|j||| ||| || fS)N)rrr)rrz exception raised in parse action)rrrrrrrrrr2rrrr5rrrrr __cause__rr )rrrrrZTRYZMATCHZFAILZ debuggingprelocZ tokensStarttokenserrZ retTokensrZparse_action_excrrrrras          zParserElement._parseNoCachecCs@z|j||dddWStk r:t|||j|YnXdS)NFrr)rr4r2rrrrrrrtryParseszParserElement.tryParsec Cs4z|||Wnttfk r*YdSXdSdS)NFT)rr2rrrrr canParseNexts zParserElement.canParseNextc@seZdZddZdS)zParserElement._UnboundedCachecs~it|_fdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS) Ncs |Srrrr@cache not_in_cacherrrsz3ParserElement._UnboundedCache.__init__..getcs ||<dSrrrr@r$rrrrsz3ParserElement._UnboundedCache.__init__..setcs dSrrFrrrrrFsz5ParserElement._UnboundedCache.__init__..clearcstSrrrrrr cache_lensz9ParserElement._UnboundedCache.__init__..cache_len)rrtypes MethodTyperrrFr()rrrrFrrrrrs    z&ParserElement._UnboundedCache.__init__Nrrrrrrrr_UnboundedCachesrNc@seZdZddZdS)ParserElement._FifoCachecst|_tfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_dS) Ncs |Srrrrrrrs.ParserElement._FifoCache.__init__..getcs>||<tkr:zdWqtk r6YqXqdSr~)rpopitemrr)rsizerrrs  .ParserElement._FifoCache.__init__..setcs dSrrrrrrrFs0ParserElement._FifoCache.__init__..clearcstSrrrrrrrs4ParserElement._FifoCache.__init__..cache_len) rr _OrderedDictrrrrrFr(rrrrrFrr)rrrrrs   !ParserElement._FifoCache.__init__Nrrrrr _FifoCachesrc@seZdZddZdS)rcst|_itgfdd}fdd}fdd}fdd}t|||_t|||_t|||_t|||_ dS) Ncs |Srrrrrrrsrcs4||<tkr&dq|dSr)rr?popleftrr)rkey_fiforrrrs rcsdSrrr)rr rrrF srcstSrrrrrrrsr) rr collectionsdequerrrrrFr(rr)rr rrrrs   rNrrrrrrsrc Csd\}}|||||f}tjtj}||} | |jkrtj|d7<z|||||} Wn8tk r} z||| j | j W5d} ~ XYn.X||| d| d f| W5QRSn@tj|d7<t | t r| | d| d fW5QRSW5QRXdS)NrNrr)r7packrat_cache_lock packrat_cacherrpackrat_cache_statsrr0rrrrIrr) rrrrrZHITZMISSlookuprr$rrrr _parseCaches$   zParserElement._parseCachecCs(tjdgttjtjdd<dSr)r7r rFrrrrrr resetCache6s zParserElement.resetCachecCs8tjs4dt_|dkr tt_n t|t_tjt_dS)aEnables "packrat" parsing, which adds memoizing to the parsing logic. Repeated parse attempts at the same string location (which happens often in many complex grammars) can immediately return a cached value, instead of re-executing parsing/validating code. Memoizing is done of both valid results and parsing exceptions. Parameters: - cache_size_limit - (default= ``128``) - if an integer value is provided will limit the size of the packrat cache; if None is passed, then the cache size will be unbounded; if 0 is passed, the cache will be effectively disabled. This speedup may break existing programs that use parse actions that have side-effects. For this reason, packrat parsing is disabled when you first import pyparsing. To activate the packrat feature, your program must call the class method :class:`ParserElement.enablePackrat`. For best results, call ``enablePackrat()`` immediately after importing pyparsing. Example:: from pip._vendor import pyparsing pyparsing.ParserElement.enablePackrat() TN)r7_packratEnabledrr rrr)Zcache_size_limitrrr enablePackrat<s   zParserElement.enablePackratc Cst|js||jD] }|q|js8|}z<||d\}}|rr|||}t t }|||Wn0t k r}ztj rn|W5d}~XYnX|SdS)a Execute the parse expression with the given string. This is the main interface to the client code, once the complete expression has been built. Returns the parsed data as a :class:`ParseResults` object, which may be accessed as a list, or as a dict or object with attributes if the given parser includes results names. If you want the grammar to require that the entire input string be successfully parsed, then set ``parseAll`` to True (equivalent to ending the grammar with ``StringEnd()``). Note: ``parseString`` implicitly calls ``expandtabs()`` on the input string, in order to report proper column numbers in parse actions. If the input string contains tabs and the grammar uses parse actions that use the ``loc`` argument to index into the string being parsed, you can ensure you have a consistent view of the input string by: - calling ``parseWithTabs`` on your grammar before calling ``parseString`` (see :class:`parseWithTabs`) - define your parse action using the full ``(s, loc, toks)`` signature, and reference the input string using the parse action's ``s`` argument - explictly expand the tabs in your input string before calling ``parseString`` Example:: Word('a').parseString('aaaaabaaa') # -> ['aaaaa'] Word('a').parseString('aaaaabaaa', parseAll=True) # -> Exception: Expected end of text rN) r7rr streamlinerr expandtabsrrrr<r0verbose_stacktrace)rrparseAllrrrZserrrr parseString_s$!    zParserElement.parseStringc cs6|js||jD] }|q|js4t|}t|}d}|j}|j}t d} z||kr| |krz |||} ||| dd\} } Wnt k r| d}YqZX| |kr| d7} | | | fV|r|||} | |kr| }q|d7}q| }qZ| d}qZWn4t k r0}zt j rn|W5d}~XYnXdS)aq Scan the input string for expression matches. Each match will return the matching tokens, start location, and end location. May be called with optional ``maxMatches`` argument, to clip scanning after 'n' matches are found. If ``overlap`` is specified, then overlapping matches will be reported. Note that the start and end locations are reported relative to the string being parsed. See :class:`parseString` for more information on parsing strings with embedded tabs. Example:: source = "sldjf123lsdjjkf345sldkjf879lkjsfd987" print(source) for tokens, start, end in Word(alphas).scanString(source): print(' '*start + '^'*(end-start)) print(' '*start + tokens[0]) prints:: sldjf123lsdjjkf345sldkjf879lkjsfd987 ^^^^^ sldjf ^^^^^^^ lsdjjkf ^^^^^^ sldkjf ^^^^^^ lkjsfd rFrrN)rrrrrrrrrr7rr2r0r)rr maxMatchesZoverlaprrrZ preparseFnZparseFnmatchesrZnextLocrZnextlocrrrr scanStringsB       zParserElement.scanStringc Csg}d}d|_z||D]Z\}}}|||||rpt|trR||7}nt|trf||7}n |||}q|||ddd|D}dtt t |WSt k r}zt j rƂn|W5d}~XYnXdS)a[ Extension to :class:`scanString`, to modify matching text with modified tokens that may be returned from a parse action. To use ``transformString``, define a grammar and attach a parse action to it that modifies the returned token list. Invoking ``transformString()`` on a target string will then scan for matches, and replace the matched text patterns according to the logic in the parse action. ``transformString()`` returns the resulting transformed string. Example:: wd = Word(alphas) wd.setParseAction(lambda toks: toks[0].title()) print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york.")) prints:: Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York. rTNcSsg|] }|r|qSrrrorrrrsz1ParserElement.transformString..r)rrrrr5rr rrr_flattenr0r7r)rrr[ZlastErrrrrrrrs(    zParserElement.transformStringc CsRztdd|||DWStk rL}ztjr8n|W5d}~XYnXdS)a Another extension to :class:`scanString`, simplifying the access to the tokens found to match the given parse expression. May be called with optional ``maxMatches`` argument, to clip searching after 'n' matches are found. Example:: # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters cap_word = Word(alphas.upper(), alphas.lower()) print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")) # the sum() builtin can be used to merge results into a single ParseResults object print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))) prints:: [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']] ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity'] cSsg|]\}}}|qSrr)rrrrrrrr#sz.ParserElement.searchString..N)r5rr0r7r)rrrrrrr searchString s zParserElement.searchStringc csTd}d}|j||dD]*\}}}|||V|r<|dV|}q||dVdS)aR Generator method to split a string using the given expression as a separator. May be called with optional ``maxsplit`` argument, to limit the number of splits; and the optional ``includeSeparators`` argument (default= ``False``), if the separating matching text should be included in the split results. Example:: punc = oneOf(list(".,;:/-!?")) print(list(punc.split("This, this?, this sentence, is badly punctuated!"))) prints:: ['This', ' this', '', ' this sentence', ' is badly punctuated', ''] r)rN)r) rrmaxsplitZincludeSeparatorsZsplitsZlastrrrrrrr+s zParserElement.splitcCsV|tkrt|St|tr$||}t|tsJtjdt|t dddSt ||gS)a[ Implementation of + operator - returns :class:`And`. Adding strings to a ParserElement converts them to :class:`Literal`s by default. Example:: greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) prints:: Hello, World! -> ['Hello', ',', 'World', '!'] ``...`` may be used as a parse expression as a short form of :class:`SkipTo`. Literal('start') + ... + Literal('end') is equivalent to: Literal('start') + SkipTo('end')("_skipped*") + Literal('end') Note that the skipped text is returned with '_skipped' as a results name, and to support having multiple skips in the same parser, the value returned is a list of all skipped text. 4Cannot combine element of type %s with ParserElementr stacklevelN) Ellipsis _PendingSkiprrrr7warningswarnr SyntaxWarningrrSrrrrKDs   zParserElement.__add__cCsZ|tkrt|d|St|tr,||}t|tsRtjdt|t dddS||S)z` Implementation of + operator when left operand is not a :class:`ParserElement` _skipped*r#rr$N) r&r;rrrr7r(r)rr*rSrrrrTjs   zParserElement.__radd__cCsJt|tr||}t|ts:tjdt|tdddS|t |S)zT Implementation of - operator, returns :class:`And` with error stop r#rr$N) rrrr7r(r)rr*r _ErrorStoprSrrr__sub__ys   zParserElement.__sub__cCsBt|tr||}t|ts:tjdt|tdddS||S)z` Implementation of - operator when left operand is not a :class:`ParserElement` r#rr$Nrrrr7r(r)rr*rSrrr__rsub__s   zParserElement.__rsub__cs|tkrd}n8t|trF|ddtfkrFd|ddddd}t|tr\|d}}nt|trJtdd |D}|d dd}|ddkrd|df}t|dtr|ddkr|ddkrtS|ddkrtS|dtSnNt|dtr,t|dtr,|\}}||8}ntd t|dt|dntd t||dkrjtd |dkr|td||krdkrnntd|rfdd|r|dkr҈|}nt g||}n|}n|dkr}nt g|}|S)a Implementation of * operator, allows use of ``expr * 3`` in place of ``expr + expr + expr``. Expressions may also me multiplied by a 2-integer tuple, similar to ``{min, max}`` multipliers in regular expressions. Tuples may also include ``None`` as in: - ``expr*(n, None)`` or ``expr*(n, )`` is equivalent to ``expr*n + ZeroOrMore(expr)`` (read as "at least n instances of ``expr``") - ``expr*(None, n)`` is equivalent to ``expr*(0, n)`` (read as "0 to n instances of ``expr``") - ``expr*(None, None)`` is equivalent to ``ZeroOrMore(expr)`` - ``expr*(1, None)`` is equivalent to ``OneOrMore(expr)`` Note that ``expr*(None, n)`` does not raise an exception if more than n exprs exist in the input stream; that is, ``expr*(None, n)`` does not enforce a maximum number of expr occurrences. If this behavior is desired, then write ``expr*(None, n) + ~expr`` )rNNrrrrrcss|]}|tk r|ndVqdSr)r&rrrrrsz(ParserElement.__mul__..)NNz8cannot multiply 'ParserElement' and ('%s', '%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez,cannot multiply ParserElement by 0 or (0, 0)cs(|dkrt|dStSdSr)r.nmakeOptionalListrrrr3sz/ParserElement.__mul__..makeOptionalList) r&rtuplerrEr,rr ValueErrorr)rrJZ minElementsZ optElementsrrr2r__mul__sN              zParserElement.__mul__cCs ||Sr)r6rSrrr__rmul__szParserElement.__rmul__cCsZ|tkrt|ddSt|tr(||}t|tsNtjdt|t dddSt ||gS)zL Implementation of | operator - returns :class:`MatchFirst` T) must_skipr#rr$N) r&r'rrrr7r(r)rr*r)rSrrr__or__s    zParserElement.__or__cCsBt|tr||}t|ts:tjdt|tdddS||BS)z` Implementation of | operator when left operand is not a :class:`ParserElement` r#rr$Nr.rSrrr__ror__s   zParserElement.__ror__cCsFt|tr||}t|ts:tjdt|tdddSt||gS)zD Implementation of ^ operator - returns :class:`Or` r#rr$N) rrrr7r(r)rr*r/rSrrr__xor__s   zParserElement.__xor__cCsBt|tr||}t|ts:tjdt|tdddS||AS)z` Implementation of ^ operator when left operand is not a :class:`ParserElement` r#rr$Nr.rSrrr__rxor__ s   zParserElement.__rxor__cCsFt|tr||}t|ts:tjdt|tdddSt||gS)zF Implementation of & operator - returns :class:`Each` r#rr$N) rrrr7r(r)rr*rrSrrr__and__ s   zParserElement.__and__cCsBt|tr||}t|ts:tjdt|tdddS||@S)z` Implementation of & operator when left operand is not a :class:`ParserElement` r#rr$Nr.rSrrr__rand__ s   zParserElement.__rand__cCst|S)zH Implementation of ~ operator - returns :class:`NotAny` )r+rrrr __invert__) szParserElement.__invert__cCstd|jjdS)Nz%r object is not iterable)rrrrrrrr-/ szParserElement.__iter__c Cszt|tr|f}t|Wntk r8||f}YnXt|dkrztd|ddt|dkrrdt|nd|t|dd}|S)a use ``[]`` indexing notation as a short form for expression repetition: - ``expr[n]`` is equivalent to ``expr*n`` - ``expr[m, n]`` is equivalent to ``expr*(m, n)`` - ``expr[n, ...]`` or ``expr[n,]`` is equivalent to ``expr*n + ZeroOrMore(expr)`` (read as "at least n instances of ``expr``") - ``expr[..., n]`` is equivalent to ``expr*(0, n)`` (read as "0 to n instances of ``expr``") - ``expr[...]`` and ``expr[0, ...]`` are equivalent to ``ZeroOrMore(expr)`` - ``expr[1, ...]`` is equivalent to ``OneOrMore(expr)`` ``None`` may be used in place of ``...``. Note that ``expr[..., n]`` and ``expr[m, n]``do not raise an exception if more than ``n`` ``expr``s exist in the input stream. If this behavior is desired, then write ``expr[..., n] + ~expr``. rz.only 1 or 2 index arguments supported ({0}{1})Nrz ... [{0}]r) rrr,rrr(r)rr4)rr@rrrrr4 s    zParserElement.__getitem__cCs|dk r||S|SdS)a Shortcut for :class:`setResultsName`, with ``listAllMatches=False``. If ``name`` is given with a trailing ``'*'`` character, then ``listAllMatches`` will be passed as ``True``. If ``name` is omitted, same as calling :class:`copy`. Example:: # these are equivalent userdata = Word(alphas).setResultsName("name") + Word(nums + "-").setResultsName("socsecno") userdata = Word(alphas)("name") + Word(nums + "-")("socsecno") N)rrIrGrrr__call__X s zParserElement.__call__cCst|S)z Suppresses the output of this :class:`ParserElement`; useful to keep punctuation from cluttering up returned output. )r>rrrrsuppressl szParserElement.suppresscCs d|_|S)a Disables the skipping of whitespace before matching the characters in the :class:`ParserElement`'s defined pattern. This is normally only used internally by the pyparsing module, but may be needed in some whitespace-sensitive grammars. FrrrrrleaveWhitespaces szParserElement.leaveWhitespacecCsd|_||_d|_|S)z8 Overrides the default whitespace chars TF)rrr)rrrrrsetWhitespaceChars| sz ParserElement.setWhitespaceCharscCs d|_|S)z Overrides default behavior to expand ````s to spaces before parsing the input string. Must be called before ``parseString`` when the input grammar contains elements that match ```` characters. T)rrrrr parseWithTabs szParserElement.parseWithTabscCsLt|trt|}t|tr4||jkrH|j|n|jt||S)a Define expression to be ignored (e.g., comments) while doing pattern matching; may be called repeatedly, to define multiple comment or other ignorable patterns. Example:: patt = OneOrMore(Word(alphas)) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj'] patt.ignore(cStyleComment) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd'] )rrr>rrrIrSrrrignore s   zParserElement.ignorecCs"|pt|p t|ptf|_d|_|S)zT Enable display of debugging messages while doing pattern matching. T)rrrrr)rZ startActionZ successActionZexceptionActionrrrsetDebugActions s zParserElement.setDebugActionscCs|r|tttnd|_|S)a Enable display of debugging messages while doing pattern matching. Set ``flag`` to True to enable, False to disable. Example:: wd = Word(alphas).setName("alphaword") integer = Word(nums).setName("numword") term = wd | integer # turn on debugging for wd wd.setDebug() OneOrMore(term).parseString("abc 123 xyz 890") prints:: Match alphaword at loc 0(1,1) Matched alphaword -> ['abc'] Match alphaword at loc 3(1,4) Exception raised:Expected alphaword (at char 4), (line:1, col:5) Match alphaword at loc 7(1,8) Matched alphaword -> ['xyz'] Match alphaword at loc 11(1,12) Exception raised:Expected alphaword (at char 12), (line:1, col:13) Match alphaword at loc 15(1,16) Exception raised:Expected alphaword (at char 15), (line:1, col:16) The output shown is that produced by the default debug actions - custom debug actions can be specified using :class:`setDebugActions`. Prior to attempting to match the ``wd`` expression, the debugging message ``"Match at loc (,)"`` is shown. Then if the parse succeeds, a ``"Matched"`` message is shown, or an ``"Exception raised"`` message is shown. Also note the use of :class:`setName` to assign a human-readable name to the expression, which makes debugging and exception messages easier to understand - for instance, the default name created for the :class:`Word` expression without calling ``setName`` is ``"W:(ABCD...)"``. F)rGrrrr)rflagrrrr s%zParserElement.setDebugcCs|jSrrrrrrr szParserElement.__str__cCst|Srrrrrrr szParserElement.__repr__cCsd|_d|_|Sr)rrrrrrr szParserElement.streamlinecCsdSrrrrrrcheckRecursion szParserElement.checkRecursioncCs|gdS)zj Check defined expressions for valid structure, check for infinite recursive definitions. N)rI)r validateTracerrrvalidate szParserElement.validatec Csz |}Wn2tk r>t|d}|}W5QRXYnXz|||WStk r~}ztjrjn|W5d}~XYnXdS)z Execute the parse expression on the given file or filename. If a filename is specified (instead of a file object), the entire file is opened, read, and closed before parsing. rN)readropenrr0r7r)rZfile_or_filenamerZ file_contentsfrrrr parseFile s  zParserElement.parseFilecsht|trBtr(||kp$tt||qd||kp>t|t|kSn"t|trV||Stt||kSdSr)rr7rbsuper__eq__varsrrrSrrrrR s   zParserElement.__eq__cCs ||k SrrrSrrr__ne__ szParserElement.__ne__cCst|Sr)idrrrr__hash__ szParserElement.__hash__cCs||kSrrrSrrr__req__ szParserElement.__req__cCs ||k SrrrSrrr__rne__ szParserElement.__rne__cCs4z|jt||dWdStk r.YdSXdS)a Method for quick testing of a parser against a test string. Good for simple inline microtests of sub expressions while building up larger parser. Parameters: - testString - to test against this expression for a match - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests Example:: expr = Word(nums) assert expr.matches("100") rTFN)rrr0)rZ testStringrrrrr s zParserElement.matches#c  Cst|tr"tttj|}t|tr4t|}|dkrBt j }|j } g} g} d} td t dt} d}|D]\}|dk r||ds| r|s| |qt|sqtd| |g}g} z"| ||}|j||d}Wntk r}zt|trdnd }d|krB|t|j||d t|j|d d |n|d |jd ||d t|| ot|} |}W5d}~XYntk r}z$|dt|| o|} |}W5d}~XYnX| o| } |dk rzR|||}|dk r(t|tr||n|t|n||WnRtk r}z2||j|d|d|jt |j|W5d}~XYnXn||j|d|r|r|d | d|| ||fqt| | fS)as Execute the parse expression on a series of test strings, showing each test, the parsed results or where the parse failed. Quick and easy way to run a parse expression against a list of sample strings. Parameters: - tests - a list of separate test strings, or a multiline string of test strings - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests - comment - (default= ``'#'``) - expression for indicating embedded comments in the test string; pass None to disable comment filtering - fullDump - (default= ``True``) - dump results as list followed by results names in nested outline; if False, only dump nested list - printResults - (default= ``True``) prints test output to stdout - failureTests - (default= ``False``) indicates if these tests are expected to fail parsing - postParse - (default= ``None``) optional callback for successful parse results; called as `fn(test_string, parse_results)` and returns a string to be added to the test output - file - (default=``None``) optional file-like object to which test output will be written; if None, will default to ``sys.stdout`` Returns: a (success, results) tuple, where success indicates that all tests succeeded (or failed if ``failureTests`` is True), and the results contain a list of lines of each test's output Example:: number_expr = pyparsing_common.number.copy() result = number_expr.runTests(''' # unsigned integer 100 # negative integer -100 # float with scientific notation 6.02e23 # integer with scientific notation 1e-12 ''') print("Success" if result[0] else "Failed!") result = number_expr.runTests(''' # stray character 100Z # missing leading digit before '.' -.100 # too many '.' 3.14.159 ''', failureTests=True) print("Success" if result[0] else "Failed!") prints:: # unsigned integer 100 [100] # negative integer -100 [-100] # float with scientific notation 6.02e23 [6.02e+23] # integer with scientific notation 1e-12 [1e-12] Success # stray character 100Z ^ FAIL: Expected end of text (at char 3), (line:1, col:4) # missing leading digit before '.' -.100 ^ FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1) # too many '.' 3.14.159 ^ FAIL: Expected end of text (at char 4), (line:1, col:5) Success Each test string must be on a single line. If you want to test a string that spans multiple lines, create a test like this:: expr.runTest(r"this is a test\n of strings that spans \n 3 lines") (Note that this is a raw string literal, you must include the leading 'r'.) NT\nruFrZz(FATAL)rrrrzFAIL: zFAIL-EXCEPTION: )rpz{0} failed: {1}: {2})!rrr rrrrstrip splitlinesr'rstdoutwriterrprFrmrrrrlstriprr0r4r[rrMrr5rurrr)rZtestsrZcommentZfullDumpZ printResultsZ failureTestsrfileprint_Z allResultsZcommentssuccessrwBOMrr[resultrrrZpp_valuerrrrrunTests. sn`     $       0 zParserElement.runTests)F)F)F)T)T)TT)TT)r)F)N)T)N)F)T)Tr[TTFNN)RrrrrrrrrrrrIrrrrrrrrrrrrrrrrrrrr rr rrrrrrr_MAX_INTrrr!rrKrTr-r/r6r7r9r:r;r<r=r>r?r-rr@rArCrDrErFrGrrrrrIrKrPrRrUrWrXrYrrg __classcell__rrrTrr7?s     1     W   " 8H. &  J     $     +    cs6eZdZd fdd ZddZddZdd ZZS) r'Fcs>tt|t|tdd|_|j|_||_||_ dS)Nr...) rQr'rrrrrranchorr8)rrr8rTrrr s z_PendingSkip.__init__cs\t|dd}jrNdd}fdd}j||||B|Sj||S)Nrjr+cSs,|jr|jdgkr(|d=|dddS)Nrr_skipped)rlrr?rrrrr8 sz'_PendingSkip.__add__..must_skipcs<|jdddgkr8|d}dtjd|d<dS)Nrrrlz missing .show_skip)r;rr8rkr)rrJZskipperr8rmrrrrK s  z_PendingSkip.__add__cCs|jSr)rrrrrr sz_PendingSkip.__repr__cGs tddS)NzBuse of `...` expression without following SkipTo target expression)r)rrrrrr sz_PendingSkip.parseImpl)F)rrrrrKrrrirrrTrr' sr'cs eZdZdZfddZZS)r?zYAbstract :class:`ParserElement` subclass, for defining atomic matching patterns. cstt|jdddSNFr)rQr?rrrTrrr szToken.__init__rrrrrrirrrTrr? scs eZdZdZfddZZS)rz'An empty token, will always match. cs$tt|d|_d|_d|_dS)NrTF)rQrrrrrrrTrrr szEmpty.__init__rprrrTrr scs*eZdZdZfddZdddZZS)r*z#A token that will never match. cs*tt|d|_d|_d|_d|_dS)Nr*TFzUnmatchable token)rQr*rrrrrrrTrrr s zNoMatch.__init__TcCst|||j|dSr)r2rrrrrr szNoMatch.parseImpl)TrrrrrrrirrrTrr* s cs*eZdZdZfddZdddZZS)r'aToken to exactly match a specified string. Example:: Literal('blah').parseString('blah') # -> ['blah'] Literal('blah').parseString('blahfooblah') # -> ['blah'] Literal('blah').parseString('bla') # -> Exception: Expected "blah" For case-insensitive matching, use :class:`CaselessLiteral`. For keyword matching (force word break before and after the matched string), use :class:`Keyword` or :class:`CaselessKeyword`. cstt|||_t||_z|d|_Wn*tk rVtj dt ddt |_ YnXdt |j|_d|j|_d|_d|_|jdkrt|tkrt|_ dS) Nrz2null string passed to Literal; use Empty() insteadrr$"%s"rFr)rQr'rmatchrmatchLenfirstMatchCharrr(r)r*rrrrrrrr_SingleCharLiteralr matchStringrTrrr s"   zLiteral.__init__TcCs@|||jkr,||j|r,||j|jfSt|||j|dSr)ru startswithrsrtr2rrrrrr4 szLiteral.parseImpl)TrqrrrTrr' s c@seZdZdddZdS)rvTcCs0|||jkr|d|jfSt|||j|dSr)rursr2rrrrrr: sz_SingleCharLiteral.parseImplN)Trrrrrrrrrv9 srvcsLeZdZdZedZdfdd Zddd Zfd d Ze d d Z Z S)r$aToken to exactly match a specified string as a keyword, that is, it must be immediately followed by a non-keyword character. Compare with :class:`Literal`: - ``Literal("if")`` will match the leading ``'if'`` in ``'ifAndOnlyIf'``. - ``Keyword("if")`` will not; it will only match the leading ``'if'`` in ``'if x=1'``, or ``'if(y==2)'`` Accepts two optional constructor arguments in addition to the keyword string: - ``identChars`` is a string of characters that would be valid identifier characters, defaulting to all alphanumerics + "_" and "$" - ``caseless`` allows case-insensitive matching, default is ``False``. Example:: Keyword("start").parseString("start") # -> ['start'] Keyword("start").parseString("starting") # -> Exception For case-insensitive matching, use :class:`CaselessKeyword`. _$NFcstt||dkrtj}||_t||_z|d|_Wn$tk r^t j dt ddYnXd|j|_ d|j |_ d|_d|_||_|r||_|}t||_dS)Nrz2null string passed to Keyword; use Empty() insteadrr$rrrF)rQr$rDEFAULT_KEYWORD_CHARSrsrrtrurr(r)r*rrrrcaselessupper caselessmatchr identChars)rrxrr}rTrrr] s*     zKeyword.__init__TcCs|jr|||||j|jkr|t||jksL|||j|jkr|dksj||d|jkr||j|jfSnv|||jkr|jdks||j|r|t||jks|||j|jkr|dks||d|jkr||j|jfSt |||j |dSr) r}rtr~rrrrsruryr2rrrrrrr s.zKeyword.parseImplcstt|}tj|_|Sr)rQr$rIr|r)rrrTrrrI sz Keyword.copycCs |t_dS)z,Overrides the default Keyword chars N)r$r|rrrrsetDefaultKeywordChars szKeyword.setDefaultKeywordChars)NF)T) rrrrrGr|rrrIrrrirrrTrr$B s  cs*eZdZdZfddZdddZZS)rafToken to match a specified string, ignoring case of letters. Note: the matched results will always be in the case of the given match string, NOT the case of the input text. Example:: OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD'] (Contrast with example for :class:`CaselessKeyword`.) cs6tt||||_d|j|_d|j|_dS)Nz'%s'r)rQrrr~ returnStringrrrwrTrrr s zCaselessLiteral.__init__TcCs@||||j|jkr,||j|jfSt|||j|dSr)rtr~rsrr2rrrrrr szCaselessLiteral.parseImpl)TrqrrrTrr s cs"eZdZdZdfdd ZZS)rz Caseless version of :class:`Keyword`. Example:: OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD'] (Contrast with example for :class:`CaselessLiteral`.) Ncstt|j||dddS)NTr})rQrr)rrxrrTrrr szCaselessKeyword.__init__)NrprrrTrr s cs,eZdZdZdfdd Zd ddZZS) raA variation on :class:`Literal` which matches "close" matches, that is, strings with at most 'n' mismatching characters. :class:`CloseMatch` takes parameters: - ``match_string`` - string to be matched - ``maxMismatches`` - (``default=1``) maximum number of mismatches allowed to count as a match The results from a successful parse will contain the matched text from the input string and the following named results: - ``mismatches`` - a list of the positions within the match_string where mismatches were found - ``original`` - the original match_string used to compare against the input string If ``mismatches`` is an empty list, then the match was an exact match. Example:: patt = CloseMatch("ATCATCGAATGGA") patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']}) patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1) # exact match patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']}) # close match allowing up to 2 mismatches patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2) patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']}) rcsBtt|||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F) rQrrr match_string maxMismatchesrrr)rrrrTrrr szCloseMatch.__init__TcCs|}t|}|t|j}||kr|j}d}g} |j} tt||||D]2\}} | \} } | | krL| |t| | krLqqL|d}t|||g}||d<| |d<||fSt|||j|dS)Nrroriginal mismatches) rrrrrrr5r2r)rrrrstartrmaxlocrZmatch_stringlocrrZs_msrcmatresultsrrrr s(  zCloseMatch.parseImpl)r)TrqrrrTrr s  cs8eZdZdZd fdd Zdd d Zfd d ZZS)rBaX Token for matching words composed of allowed character sets. Defined with string containing all allowed initial characters, an optional string containing allowed body characters (if omitted, defaults to the initial character set), and an optional minimum, maximum, and/or exact length. The default value for ``min`` is 1 (a minimum value < 1 is not valid); the default values for ``max`` and ``exact`` are 0, meaning no maximum or exact length restriction. An optional ``excludeChars`` parameter can list characters that might be found in the input ``bodyChars`` string; useful to define a word of all printables except for one or two characters, for instance. :class:`srange` is useful for defining custom character set strings for defining ``Word`` expressions, using range notation from regular expression character sets. A common mistake is to use :class:`Word` to match a specific literal string, as in ``Word("Address")``. Remember that :class:`Word` uses the string argument to define *sets* of matchable characters. This expression would match "Add", "AAA", "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an exact literal string, use :class:`Literal` or :class:`Keyword`. pyparsing includes helper strings for building Words: - :class:`alphas` - :class:`nums` - :class:`alphanums` - :class:`hexnums` - :class:`alphas8bit` (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.) - :class:`punc8bit` (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.) - :class:`printables` (any non-whitespace character) Example:: # a word composed of digits integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9")) # a word with a leading capital, and zero or more lowercase capital_word = Word(alphas.upper(), alphas.lower()) # hostnames are alphanumeric, with leading alpha, and '-' hostname = Word(alphas, alphanums + '-') # roman numeral (not a strict parser, accepts invalid mix of characters) roman = Word("IVXLCDM") # any string of non-whitespace characters, except for ',' csv_value = Word(printables, excludeChars=",") NrrFcstt|rNtdfdd|D}|rNdfdd|D}||_t||_|rt||_t||_n||_t||_|dk|_ |dkrt d||_ |dkr||_ nt |_ |dkr||_ ||_ t||_d|j|_d |_||_d |j|jkr|dkr|dkr|dkr|j|jkr@d t|j|_nHt|jdkrnd t|jt|jf|_nd t|jt|jf|_|jrd|jd|_zt|j|_Wntk rd|_YnX|jj|_t|_dS)Nrc3s|]}|kr|VqdSrrr excludeCharsrrr2 sz Word.__init__..c3s|]}|kr|VqdSrrrrrrr4 srrzZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedrFrz[%s]+z%s[%s]*z [%s][%s]*z\b)rQrBrrr initCharsOrig initChars bodyCharsOrig bodyChars maxSpecifiedr5minLenmaxLenrhrrrr asKeyword_escapeRegexRangeCharsreStringrrescapecompilerrsre_match _WordRegexr)rrrminmaxexactrrrTrrr. sZ      0     z Word.__init__Tc Cs|||jkrt|||j||}|d7}t|}|j}||j}t||}||krj|||krj|d7}qLd}|||jkrd}nV|jr||kr|||krd}n6|j r|dkr||d|ks||kr|||krd}|rt|||j|||||fS)NrFTr) rr2rrrrrrrr) rrrrrrZ bodycharsrZthrowExceptionrrrrg s2    zWord.parseImplcsvztt|WStk r$YnX|jdkrpdd}|j|jkr`d||j||jf|_nd||j|_|jS)NcSs$t|dkr|dddS|SdS)Nrjrrrrr charsAsStr s z Word.__str__..charsAsStrz W:(%s, %s)zW:(%s))rQrBrrrrr)rrrTrrr s  z Word.__str__)NrrrFN)TrrrrrrrrirrrTrrB s49 c@seZdZdddZdS)rTcCs4|||}|s t|||j||}||fSr)rr2rendgroup)rrrrrfrrrr s  z_WordRegex.parseImplN)Trzrrrrr srcs"eZdZdZdfdd ZZS)rFzA short-cut class for defining ``Word(characters, exact=1)``, when defining a match of any single character in a string of characters. FNcsZtt|j|d||ddtd|j|_|r>d|j|_t|j|_|jj |_ dS)Nr)rrr[%s]rz\b%s\b) rQrFrrrrrrrrsr)rcharsetrrrTrrr s  z Char.__init__)FNrprrrTrrF scsbeZdZdZeedZdfdd Zddd Z dd d Z dd d Z fddZ ddZ ZS)r:aToken for matching strings that match a given regular expression. Defined with string specifying the regular expression in a form recognized by the stdlib Python `re module `_. If the given regex contains named groups (defined using ``(?P...)``), these will be preserved as named parse results. Example:: realnum = Regex(r"[+-]?\d+\.\d*") date = Regex(r'(?P\d{4})-(?P\d\d?)-(?P\d\d?)') # ref: https://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression roman = Regex(r"M{0,4}(CM|CD|D?{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})") z[A-Z]rFcstt|t|tr|s,tjdtdd||_||_ zt |j|j |_ |j|_ Wqt jk rtjd|tddYqXn2t|tjr||_ t||_|_ ||_ ntd|j j|_t||_d|j|_d|_d|_||_||_|jr|j|_|jr|j|_d S) aThe parameters ``pattern`` and ``flags`` are passed to the ``re.compile()`` function as-is. See the Python `re module `_ module for an explanation of the acceptable patterns and flags. z0null string passed to Regex; use Empty() insteadrr$$invalid pattern (%s) passed to RegexzCRegex may only be constructed with a string or a compiled RE objectrFTN)rQr:rrrr(r)r*patternflagsrrr sre_constantserrorcompiledREtyperr5rsrrrrrr asGroupListasMatchparseImplAsGroupListrparseImplAsMatch)rrrrrrTrrr sD        zRegex.__init__Tc Csb|||}|s t|||j||}t|}|}|rZ|D]\}}|||<qH||fSr)rr2rrr5r groupdictr!) rrrrrfrdrrrrrr s   zRegex.parseImplcCs8|||}|s t|||j||}|}||fSr)rr2rrgroupsrrrrrfrrrrr s  zRegex.parseImplAsGroupListcCs4|||}|s t|||j||}|}||fSr)rr2rrrrrrr s  zRegex.parseImplAsMatchcsFztt|WStk r$YnX|jdkr@dt|j|_|jS)NzRe:(%s))rQr:rrrrrrrTrrr s z Regex.__str__csljrtjdtddtjr@tr@tjdtddtjrTfdd}nfdd}|S)a Return Regex with an attached parse action to transform the parsed result as if called using `re.sub(expr, repl, string) `_. Example:: make_html = Regex(r"(\w+):(.*?):").sub(r"<\1>\2") print(make_html.transformString("h1:main title:")) # prints "

main title

" z-cannot use sub() with Regex(asGroupList=True)rr$z9cannot use sub() with a callable with Regex(asMatch=True)cs|dSr)expandr)replrrr, szRegex.sub..pacsj|dSr)rrrrrrrr/ s)rr(r)r* SyntaxErrorrrr)rrrrrrr s z Regex.sub)rFF)T)T)T)rrrrrrrrrrrrrrrirrrTrr: s - cs8eZdZdZd fdd Zd ddZfd d ZZS) r8a& Token for matching strings that are delimited by quoting characters. Defined with the following parameters: - quoteChar - string of one or more characters defining the quote delimiting string - escChar - character to escape quotes, typically backslash (default= ``None``) - escQuote - special quote sequence to escape an embedded quote string (such as SQL's ``""`` to escape an embedded ``"``) (default= ``None``) - multiline - boolean indicating whether quotes can span multiple lines (default= ``False``) - unquoteResults - boolean indicating whether the matched text should be unquoted (default= ``True``) - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default= ``None`` => same as quoteChar) - convertWhitespaceEscapes - convert escaped whitespace (``'\t'``, ``'\n'``, etc.) to actual whitespace (default= ``True``) Example:: qs = QuotedString('"') print(qs.searchString('lsjdf "This is the quote" sldjf')) complex_qs = QuotedString('{{', endQuoteChar='}}') print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf')) sql_qs = QuotedString('"', escQuote='""') print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf')) prints:: [['This is the quote']] [['This is the "quote"']] [['This is the quote with "embedded" quotes']] NFTc sXtt|}|s0tjdtddt|dkr>|}n"|}|s`tjdtddt|_t |_ |d_ |_ t |_ |_|_|_|_|rtjtjB_dtjtj d|dk rt|pdf_n.r)z|(?:%s)z|(?:%s.)z(.)z)*%srrFT)'rQr8rrr(r)r*r quoteCharr quoteCharLenfirstQuoteCharrendQuoteCharLenescCharescQuoteunquoteResultsconvertWhitespaceEscapesr MULTILINEDOTALLrrrrrrescCharReplacePatternrrrsrrrrrrrr)rrrrZ multilinerrrrTrrrZ sv           zQuotedString.__init__c Cs|||jkr|||pd}|s2t|||j||}|}|jr||j|j }t |t rd|kr|j rddddd}| D]\}}| ||}q|jrt|jd|}|jr| |j|j}||fS)Nr r  )\tr\z\fz\rz\g<1>)rrr2rrrrrrrrrr!rrrrrrr) rrrrrfrZws_mapZwslitZwscharrrrr s* zQuotedString.parseImplcsHztt|WStk r$YnX|jdkrBd|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)rQr8rrrrrrrTrrr s zQuotedString.__str__)NNFTNT)TrrrrTrr83 s&A #cs8eZdZdZd fdd Zd ddZfd d ZZS) raToken for matching words composed of characters *not* in a given set (will include whitespace in matched characters if not listed in the provided exclusion set - see example). Defined with string containing all disallowed characters, and an optional minimum, maximum, and/or exact length. The default value for ``min`` is 1 (a minimum value < 1 is not valid); the default values for ``max`` and ``exact`` are 0, meaning no maximum or exact length restriction. Example:: # define a comma-separated-value as anything that is not a ',' csv_value = CharsNotIn(',') print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213")) prints:: ['dkls', 'lsdkjf', 's12 34', '@!#', '213'] rrcstt|d|_||_|dkr*td||_|dkr@||_nt|_|dkrZ||_||_t ||_ d|j |_ |jdk|_ d|_ dS)NFrzfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr)rQrrrnotCharsr5rrrhrrrrr)rrrrrrTrrr s    zCharsNotIn.__init__TcCs|||jkrt|||j||}|d7}|j}t||jt|}||krb|||krb|d7}qD|||jkrt|||j|||||fSr)rr2rrrrr)rrrrrZnotcharsmaxlenrrrr s zCharsNotIn.parseImplcsfztt|WStk r$YnX|jdkr`t|jdkrTd|jdd|_n d|j|_|jS)Nrz !W:(%s...)z!W:(%s))rQrrrrrrrrTrrrs  zCharsNotIn.__str__)rrr)TrrrrTrr s cs`eZdZdZdddddddd d d d d ddddddddddddZd"fdd Zd#d d!ZZS)$rAaSpecial matching class for matching whitespace. Normally, whitespace is ignored by pyparsing grammars. This class is included when some whitespace structures are significant. Define with a string containing the whitespace characters to be matched; default is ``" \t\r\n"``. Also takes optional ``min``, ``max``, and ``exact`` arguments, as defined for the :class:`Word` class. zzzzzzzzz z z z zzzzzz z zzzz)rrrrrzuA0zu80zu80Euu€0uu€1uu€2uu€3uu€4uu€5uu€6uu€7uu€8uu€9uu€Auu€Buu‚Fuu…FuuÀ0 rrcstt|_dfddjDdddjD_d_dj_ |_ |dkrt|_ nt _ |dkr|_ |_ dS)Nrc3s|]}|jkr|VqdSr) matchWhiterrrrr:s z!White.__init__..css|]}tj|VqdSr)rA whiteStrsrrrrr<sTrr) rQrArrrDrrrrrrrrh)rZwsrrrrTrrr7s  zWhite.__init__TcCs|||jkrt|||j||}|d7}||j}t|t|}||krb|||jkrb|d7}qB|||jkrt|||j|||||fSr)rr2rrrrr)rrrrrrrrrrKs  zWhite.parseImpl)rrrr)T)rrrrrrrrirrrTrrAs6 cseZdZfddZZS)_PositionTokencs(tt||jj|_d|_d|_dSr)rQrrrrrrrrrTrrr\s z_PositionToken.__init__rrrrrirrrTrr[srcs2eZdZdZfddZddZd ddZZS) r"zaToken to advance to a specific column of input text; useful for tabular report scraping. cstt|||_dSr)rQr"rrM)rcolnorTrrrfszGoToColumn.__init__cCs\t|||jkrXt|}|jr*|||}||krX||rXt|||jkrX|d7}q*|Sr)rMrrrisspace)rrrrrrrrjs $ zGoToColumn.preParseTcCsDt||}||jkr"t||d|||j|}|||}||fS)NzText not in expected columnrMr2)rrrrZthiscolZnewlocrrrrrss    zGoToColumn.parseImpl)T)rrrrrrrrirrrTrr"bs  cs*eZdZdZfddZdddZZS)r&aMatches if current position is at the beginning of a line within the parse string Example:: test = '''\ AAA this line AAA and this line AAA but not this one B AAA and definitely not this one ''' for t in (LineStart() + 'AAA' + restOfLine).searchString(test): print(t) prints:: ['AAA', ' this line'] ['AAA', ' and this line'] cstt|d|_dS)NzExpected start of line)rQr&rrrrTrrrszLineStart.__init__TcCs*t||dkr|gfSt|||j|dSr)rMr2rrrrrrszLineStart.parseImpl)TrqrrrTrr&|s cs*eZdZdZfddZdddZZS)r%zTMatches if current position is at the end of a line within the parse string cs,tt||tjddd|_dS)NrrzExpected end of line)rQr%rrDr7rrrrrTrrrszLineEnd.__init__TcCsb|t|kr6||dkr$|ddfSt|||j|n(|t|krN|dgfSt|||j|dS)Nrrrr2rrrrrrs     zLineEnd.parseImpl)TrqrrrTrr%s cs*eZdZdZfddZdddZZS)r=zLMatches if current position is at the beginning of the parse string cstt|d|_dS)NzExpected start of text)rQr=rrrrTrrrszStringStart.__init__TcCs0|dkr(|||dkr(t|||j||gfSr)rr2rrrrrrszStringStart.parseImpl)TrqrrrTrr=s cs*eZdZdZfddZdddZZS)r<zBMatches if current position is at the end of the parse string cstt|d|_dS)NzExpected end of text)rQr<rrrrTrrrszStringEnd.__init__TcCs^|t|krt|||j|n<|t|kr6|dgfS|t|krJ|gfSt|||j|dSrrrrrrrs    zStringEnd.parseImpl)TrqrrrTrr<s cs.eZdZdZeffdd ZdddZZS)rDayMatches if the current position is at the beginning of a Word, and is not preceded by any character in a given set of ``wordChars`` (default= ``printables``). To emulate the ```` behavior of regular expressions, use ``WordStart(alphanums)``. ``WordStart`` will also match at the beginning of the string being parsed, or at the beginning of a line. cs"tt|t||_d|_dS)NzNot at the start of a word)rQrDrr wordCharsrrrrTrrrs zWordStart.__init__TcCs@|dkr8||d|jks(|||jkr8t|||j||gfSr)rr2rrrrrrs  zWordStart.parseImpl)TrrrrrjrrrirrrTrrDscs.eZdZdZeffdd ZdddZZS)rCa_Matches if the current position is at the end of a Word, and is not followed by any character in a given set of ``wordChars`` (default= ``printables``). To emulate the ```` behavior of regular expressions, use ``WordEnd(alphanums)``. ``WordEnd`` will also match at the end of the string being parsed, or at the end of a line. cs(tt|t||_d|_d|_dS)NFzNot at the end of a word)rQrCrrrrrrrTrrrs zWordEnd.__init__TcCsPt|}|dkrH||krH|||jks8||d|jkrHt|||j||gfSr)rrr2r)rrrrrrrrrszWordEnd.parseImpl)TrrrrTrrCscszeZdZdZdfdd ZddZddZfd d Zfd d Zfd dZ dddZ fddZ dfdd Z Z S)r3z]Abstract subclass of ParserElement, for combining and post-processing parsed tokens. Fcstt|t|tr"t|}t|tr<|g_nt|t rP|g_nnt|t rt|}t dd|Drfdd|D}t|_n,zt|_Wnt k r|g_YnXd_ dS)Ncss|]}t|tVqdSr)rr)rrrrrr sz+ParseExpression.__init__..c3s&|]}t|tr|n|VqdSr)rrrrrrrrr sF)rQr3rrr r rrexprsr7r rvrrrrrrTrrrs"      zParseExpression.__init__cCs|j|d|_|Sr)rrrrSrrrrs zParseExpression.appendcCs0d|_dd|jD|_|jD] }|q|S)zExtends ``leaveWhitespace`` defined in base class, and also invokes ``leaveWhitespace`` on all contained expressions.FcSsg|] }|qSrrHrrrrrsz3ParseExpression.leaveWhitespace..)rrrC)rrrrrrCs   zParseExpression.leaveWhitespacecsrt|trB||jkrntt|||jD]}||jdq*n,tt|||jD]}||jdqX|Sr.)rr>rrQr3rFr)rrJrrTrrrF#s    zParseExpression.ignorecsNztt|WStk r$YnX|jdkrHd|jjt|jf|_|jSNz%s:(%s)) rQr3rrrrrrrrrTrrr/s zParseExpression.__str__cs*tt||jD] }|qt|jdkr|jd}t||jr|js|jdkr|j s|jdd|jdg|_d|_ |j |j O_ |j |j O_ |jd}t||jr|js|jdkr|j s|jdd|jdd|_d|_ |j |j O_ |j |j O_ dt ||_|S)Nrrrrr)rQr3rrrrrrrrrrrrr)rrrJrTrrr9s<     zParseExpression.streamlineNcCsB|dk r |ngdd|g}|jD]}||q$|gdSr)rrKrI)rrJtmprrrrrK[s  zParseExpression.validatecs$tt|}dd|jD|_|S)NcSsg|] }|qSrrHrrrrrcsz(ParseExpression.copy..)rQr3rIrrcrTrrrIaszParseExpression.copycsVtjrD|jD]6}t|tr |jr tjdd|t |j |jddq t t | ||S)N]{0}: setting results name {1!r} on {2} expression collides with {3!r} on contained expression)warn_ungrouped_named_tokens_in_collectionrr$)rrrrr7rr(r)rrrrQr3rrrrrrTrrrfs zParseExpression._setResultsName)F)N)F)rrrrrrrCrFrrrKrIrrirrrTrr3s "  cs`eZdZdZGdddeZdfdd ZfddZdd d Zd d Z d dZ ddZ Z S)ra Requires all given :class:`ParseExpression` s to be found in the given order. Expressions may be separated by whitespace. May be constructed using the ``'+'`` operator. May also be constructed using the ``'-'`` operator, which will suppress backtracking. Example:: integer = Word(nums) name_expr = OneOrMore(Word(alphas)) expr = And([integer("id"), name_expr("name"), integer("age")]) # more easily written as: expr = integer("id") + name_expr("name") + integer("age") cseZdZfddZZS)zAnd._ErrorStopcs&ttj|j||d|_|dS)N-)rQrr,rrrCryrTrrrszAnd._ErrorStop.__init__rrrrTrr,sr,Tcs|rt|krg}t|D]`\}}|tkrn|t|dkrdt||djd}|t|dqxtdq||q||dd<tt | ||t dd|jD|_ | |jdj|jdj|_d|_dS) Nrrr+z0cannot construct And with sequence ending in ...css|] }|jVqdSrrrrrrrszAnd.__init__..rT)r&rrrrrr;rrQrrrrrDrrr)rrrrrrZ skipto_argrTrrrs    z And.__init__cs|jrtdd|jddDrt|jddD]^\}}|dkrFq4t|tr4|jr4t|jdtr4|jd|j|d|jd<d|j|d<q4dd|jD|_tt|t dd|jD|_ |S)Ncss.|]&}t|to$|jo$t|jdtVqdSrN)rr3rr'rrrrrsz!And.streamline..rrcSsg|]}|dk r|qSrrrrrrrsz"And.streamline..css|] }|jVqdSrrrrrrrs) rrvrrr3r'rQrrrr)rrrrTrrrs$  zAnd.streamlinec Cs|jdj|||dd\}}d}|jddD]}t|tjrDd}q.|rz||||\}}Wqtk rtYqtk r}zd|_t|W5d}~XYqt k rt|t ||j |YqXn||||\}}|s| r.||7}q.||fS)NrFrrT) rrrrr,r6r0rrrrrr;) rrrr resultlistZ errorStoprZ exprtokensrrrrrs(   z And.parseImplcCst|tr||}||SrrrrrrSrrrrDs  z And.__iadd__cCs6|dd|g}|jD]}|||jsq2qdSr)rrIrrrsubRecCheckListrrrrrIs   zAnd.checkRecursioncCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nr{rcss|]}t|VqdSrrrrrrrszAnd.__str__..}r1rrrrrrrrrs    z And.__str__)T)T) rrrrrr,rrrrDrIrrirrrTrrts  cs^eZdZdZdfdd ZfddZddd Zd d Zd d ZddZ dfdd Z Z S)r/aRequires that at least one :class:`ParseExpression` is found. If two expressions match, the expression that matches the longest string will be used. May be constructed using the ``'^'`` operator. Example:: # construct Or using '^' operator number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) prints:: [['123'], ['3.1416'], ['789']] Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|] }|jVqdSrrrrrrrszOr.__init__..T)rQr/rrrvrrrTrrrsz Or.__init__cs.tt|tjr*tdd|jD|_|S)Ncss|] }|jVqdSrrrrrrrsz Or.streamline..)rQr/rrcollect_all_And_tokensrvrrrrTrrrsz Or.streamlineTc Csd}d}g}|jD]}z|||}Wnvtk rb} zd| _| j|krR| }| j}W5d} ~ XYqtk rt||krt|t||j|}t|}YqX|||fq|r|j t ddd|s|dd} | |||Sd} |D]\} } | | dkr | Sz| |||\}}Wn@tk r`} z d| _| j|krP| }| j}W5d} ~ XYqX|| krx||fS|| dkr||f} q| dkr| S|dk r|j|_ |nt||d|dS)NrrT)r@r rr no defined alternatives to match) rrr2rrrrrrsortrrr)rrrr maxExcLoc maxExceptionrrZloc2rZ best_exprZlongestZloc1Zexpr1rrrrrsT            z Or.parseImplcCst|tr||}||SrrrSrrr__ixor__;s  z Or.__ixor__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz ^ css|]}t|VqdSrrrrrrrEszOr.__str__..rrrrrrr@s    z Or.__str__cCs,|dd|g}|jD]}||qdSrrrIrrrrrIIs zOr.checkRecursioncsPtjs>tjr>tdd|jDr>tjdd|t |j ddt t | ||S)Ncss|]}t|tVqdSrrrrrrrrQsz%Or._setResultsName..{0}: setting results name {1!r} on {2} expression may only return a single token for an And alternative, in future will return the full list of tokens)warn_multiple_tokens_in_named_alternationrr$)rrrrrvrr(r)rrrrQr/rrrTrrrNszOr._setResultsName)F)T)F) rrrrrrrrrrIrrirrrTrr/s  = cs^eZdZdZdfdd ZfddZddd Zd d Zd d ZddZ dfdd Z Z S)r)aRequires that at least one :class:`ParseExpression` is found. If two expressions match, the first one listed is the one that will match. May be constructed using the ``'|'`` operator. Example:: # construct MatchFirst using '|' operator # watch the order of expressions to match number = Word(nums) | Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) # Fail! -> [['123'], ['3'], ['1416'], ['789']] # put more selective expression first number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums) print(number.searchString("123 3.1416 789")) # Better -> [['123'], ['3.1416'], ['789']] Fcs:tt||||jr0tdd|jD|_nd|_dS)Ncss|] }|jVqdSrrrrrrrosz&MatchFirst.__init__..T)rQr)rrrvrrrTrrrlszMatchFirst.__init__cs.tt|tjr*tdd|jD|_|S)Ncss|] }|jVqdSrrrrrrrvsz(MatchFirst.streamline..)rQr)rrrrvrrrrTrrrsszMatchFirst.streamlineTc Csd}d}|jD]}z||||}|WStk r`}z|j|krP|}|j}W5d}~XYqtk rt||krt|t||j|}t|}YqXq|dk r|j|_|nt||d|dS)Nrr)rrr2rrrrr) rrrrrrrrrrrrrys$    zMatchFirst.parseImplcCst|tr||}||SrrrSrrr__ior__s  zMatchFirst.__ior__cCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrr | css|]}t|VqdSrrrrrrrsz%MatchFirst.__str__..rrrrrrrs    zMatchFirst.__str__cCs,|dd|g}|jD]}||qdSrrrrrrrIs zMatchFirst.checkRecursioncsPtjs>tjr>tdd|jDr>tjdd|t |j ddt t | ||S)Ncss|]}t|tVqdSrrrrrrrsz-MatchFirst._setResultsName..rrrr$)rrrrrvrr(r)rrrrQr)rrrTrrrszMatchFirst._setResultsName)F)T)F) rrrrrrrrrrIrrirrrTrr)[s   csHeZdZdZd fdd ZfddZdddZd d Zd d ZZ S)rasRequires all given :class:`ParseExpression` s to be found, but in any order. Expressions may be separated by whitespace. May be constructed using the ``'&'`` operator. Example:: color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN") shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON") integer = Word(nums) shape_attr = "shape:" + shape_type("shape") posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn") color_attr = "color:" + color("color") size_attr = "size:" + integer("size") # use Each (using operator '&') to accept attributes in any order # (shape and posn are required, color and size are optional) shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr) shape_spec.runTests(''' shape: SQUARE color: BLACK posn: 100, 120 shape: CIRCLE size: 50 color: BLUE posn: 50,80 color:GREEN size:20 shape:TRIANGLE posn:20,40 ''' ) prints:: shape: SQUARE color: BLACK posn: 100, 120 ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']] - color: BLACK - posn: ['100', ',', '120'] - x: 100 - y: 120 - shape: SQUARE shape: CIRCLE size: 50 color: BLUE posn: 50,80 ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']] - color: BLUE - posn: ['50', ',', '80'] - x: 50 - y: 80 - shape: CIRCLE - size: 50 color: GREEN size: 20 shape: TRIANGLE posn: 20,40 ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']] - color: GREEN - posn: ['20', ',', '40'] - x: 20 - y: 40 - shape: TRIANGLE - size: 20 Tcs>tt|||tdd|jD|_d|_d|_d|_dS)Ncss|] }|jVqdSrrrrrrrsz Each.__init__..T) rQrrrrrrinitExprGroupsrrrTrrrs z Each.__init__cs(tt|tdd|jD|_|S)Ncss|] }|jVqdSrrrrrrrsz"Each.streamline..)rQrrrrrrrTrrrszEach.streamlinec s|jrtdd|jD|_dd|jD}dd|jD}|||_dd|jD|_dd|jD|_dd|jD|_|j|j7_d |_|}|jdd}|jddg}d } | rj||j|j} g} | D]v} z| ||}Wn t k r| | YqX| |j t | | | |kr@| | q| kr܈ | qt| t| krd } q|rd d d|D} t ||d | |fdd|jD7}g}|D]"} | |||\}}| |qt|tg}||fS)Ncss&|]}t|trt|j|fVqdSr)rr.rVrrrrrrs z!Each.parseImpl..cSsg|]}t|tr|jqSrrr.rrrrrrs z"Each.parseImpl..cSs g|]}|jrt|ts|qSr)rrr.rrrrrs cSsg|]}t|tr|jqSr)rrErrrrrrs cSsg|]}t|tr|jqSr)rr,rrrrrrs cSs g|]}t|tttfs|qSr)rr.rEr,rrrrrsFTrVcss|]}t|VqdSrrrrrrrsz*Missing one or more required elements (%s)cs$g|]}t|tr|jkr|qSrrrZtmpOptrrrs )rrrZopt1mapZ optionalsZmultioptionalsZ multirequiredZrequiredrr2rrrVremoverrrsumr5)rrrrZopt1Zopt2ZtmpLocZtmpReqdZ matchOrderZ keepMatchingZtmpExprsZfailedrZmissingrrZ finalResultsrrrrsP    zEach.parseImplcCs@t|dr|jS|jdkr:dddd|jDd|_|jS)Nrrz & css|]}t|VqdSrrrrrrr,szEach.__str__..rrrrrrr's    z Each.__str__cCs,|dd|g}|jD]}||qdSrrrrrrrI0s zEach.checkRecursion)T)T) rrrrrrrrrIrirrrTrrs 8  1 csjeZdZdZdfdd ZdddZdd Zfd d Zfd d ZddZ dddZ fddZ Z S)r1zfAbstract subclass of :class:`ParserElement`, for combining and post-processing parsed tokens. Fcstt||t|tr@t|jtr2||}n|t|}||_ d|_ |dk r|j |_ |j |_ | |j|j|_|j|_|j|_|j|jdSr)rQr1rrr issubclassrr?r'rrrrrDrrrrrrErrrrTrrr:s    zParseElementEnhance.__init__TcCs2|jdk r|jj|||ddStd||j|dS)NFrr)rrr2rrrrrrLs zParseElementEnhance.parseImplcCs*d|_|j|_|jdk r&|j|Sr~)rrrIrCrrrrrCRs    z#ParseElementEnhance.leaveWhitespacecsrt|trB||jkrntt|||jdk rn|j|jdn,tt|||jdk rn|j|jd|Sr.)rr>rrQr1rFrrSrTrrrFYs    zParseElementEnhance.ignorecs&tt||jdk r"|j|Sr)rQr1rrrrTrrres  zParseElementEnhance.streamlinecCsB||krt||g|dd|g}|jdk r>|j|dSr)r9rrI)rrrrrrrIks  z"ParseElementEnhance.checkRecursionNcCsB|dkr g}|dd|g}|jdk r4|j||gdSrrrKrIrrJrrrrrKrs   zParseElementEnhance.validatecsXztt|WStk r$YnX|jdkrR|jdk rRd|jjt|jf|_|jSr) rQr1rrrrrrrrrTrrrzszParseElementEnhance.__str__)F)T)N) rrrrrrrCrFrrIrKrrirrrTrr16s   cs*eZdZdZfddZdddZZS)r abLookahead matching of the given parse expression. ``FollowedBy`` does *not* advance the parsing position within the input string, it only verifies that the specified parse expression matches at the current position. ``FollowedBy`` always returns a null token list. If any results names are defined in the lookahead expression, those *will* be returned for access by name. Example:: # use FollowedBy to match a label only if it is followed by a ':' data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint() prints:: [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']] cstt||d|_dSr)rQr rrrrrTrrrszFollowedBy.__init__TcCs(|jj|||d\}}|dd=||fS)Nr)rr)rrrr_rrrrrs zFollowedBy.parseImpl)TrqrrrTrr s cs,eZdZdZd fdd Zd ddZZS) r(apLookbehind matching of the given parse expression. ``PrecededBy`` does not advance the parsing position within the input string, it only verifies that the specified parse expression matches prior to the current position. ``PrecededBy`` always returns a null token list, but if a results name is defined on the given expression, it is returned. Parameters: - expr - expression that must match prior to the current parse location - retreat - (default= ``None``) - (int) maximum number of characters to lookbehind prior to the current parse location If the lookbehind expression is a string, Literal, Keyword, or a Word or CharsNotIn with a specified exact or maximum length, then the retreat parameter is not required. Otherwise, retreat must be specified to give a maximum number of characters to look back from the current parse position for a lookbehind match. Example:: # VB-style variable names with type prefixes int_var = PrecededBy("#") + pyparsing_common.identifier str_var = PrecededBy("$") + pyparsing_common.identifier Ncstt||||_d|_d|_d|_t|t rJt |}d|_nVt|t t frf|j }d|_n:t|ttfr|jtkr|j}d|_nt|trd}d|_||_dt ||_d|_dS)NTFrznot preceded by )rQr(rrrCrrrrrrr'r$rtrBrrrhrretreatrr)rrrrTrrrs(  zPrecededBy.__init__rTc Cs|jr<||jkrt|||j||j}|j||\}}n|jt}|d|}t|||j} tdt||jdD]H} z|||| \}}Wn&t k r} z| } W5d} ~ XYqxXqqx| |dd=||fSr) rrr2rrrr<rrr0) rrrrrrrZ test_exprZinstring_sliceZ last_exprrMZpberrrrs"     zPrecededBy.parseImpl)N)rTrqrrrTrr(scs2eZdZdZfddZd ddZddZZS) r+aLookahead to disallow matching with the given parse expression. ``NotAny`` does *not* advance the parsing position within the input string, it only verifies that the specified parse expression does *not* match at the current position. Also, ``NotAny`` does *not* skip over leading whitespace. ``NotAny`` always returns a null token list. May be constructed using the '~' operator. Example:: AND, OR, NOT = map(CaselessKeyword, "AND OR NOT".split()) # take care not to mistake keywords for identifiers ident = ~(AND | OR | NOT) + Word(alphas) boolean_term = Optional(NOT) + ident # very crude boolean expression - to support parenthesis groups and # operation hierarchy, use infixNotation boolean_expr = boolean_term + ZeroOrMore((AND | OR) + boolean_term) # integers that are followed by "." are actually floats integer = Word(nums) + ~Char(".") cs0tt||d|_d|_dt|j|_dS)NFTzFound unwanted token, )rQr+rrrrrrrrTrrr szNotAny.__init__TcCs&|j||rt|||j||gfSr)rrr2rrrrrrszNotAny.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrz~{rr1rrrrrrrrrs   zNotAny.__str__)TrrrrTrr+s  cs>eZdZd fdd ZddZd ddZdfd d ZZS)_MultipleMatchNcs<tt||d|_|}t|tr.||}||dSr)rQrrrrrrstopOn)rrrenderrTrrr s   z_MultipleMatch.__init__cCs,t|tr||}|dk r"|nd|_|Sr)rrr not_ender)rrrrrr(s  z_MultipleMatch.stopOnTc Cs|jj}|j}|jdk }|r$|jj}|r2|||||||dd\}}zV|j } |r`|||| rp|||} n|} ||| |\}} | s| rR|| 7}qRWnttfk rYnX||fSNFr) rrrr rrr;r2r) rrrrZself_expr_parseZself_skip_ignorablesZ check_enderZ try_not_enderrZhasIgnoreExprsrZ tmptokensrrrr.s*      z_MultipleMatch.parseImplFcsftjrT|jgt|jdgD]6}t|tr|jrtjd d|t |j |jddqt t |||S)Nrrrrr$)rrrrrr7rr(r)rrrrQrrrrTrrrKsz_MultipleMatch._setResultsName)N)T)F)rrrrrrrrirrrTrrs rc@seZdZdZddZdS)r,ajRepetition of one or more of the given expression. Parameters: - expr - expression that must match one or more times - stopOn - (default= ``None``) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: BLACK" OneOrMore(attr_expr).parseString(text).pprint() # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']] # use stopOn attribute for OneOrMore to avoid reading label string as part of the data attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']] # could also be written as (attr_expr * (1,)).parseString(text).pprint() cCs4t|dr|jS|jdkr.dt|jd|_|jS)Nrrz}...rrrrrrss   zOneOrMore.__str__N)rrrrrrrrrr,Yscs8eZdZdZd fdd Zd fdd Zdd ZZS) rEakOptional repetition of zero or more of the given expression. Parameters: - expr - expression that must match zero or more times - stopOn - (default= ``None``) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example: similar to :class:`OneOrMore` Ncstt|j||dd|_dS)N)rT)rQrErr)rrrrTrrrszZeroOrMore.__init__Tc s<ztt||||WSttfk r6|gfYSXdSr)rQrErr2rrrTrrrszZeroOrMore.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)NrrU]...rrrrrrs   zZeroOrMore.__str__)N)TrrrrTrrE|s c@s eZdZddZeZddZdS) _NullTokencCsdSr~rrrrrr*sz_NullToken.__bool__cCsdSrrrrrrrsz_NullToken.__str__N)rrrr*rrrrrrr sr cs<eZdZdZeZeffdd Zd ddZddZZ S) r.aGOptional matching of the given expression. Parameters: - expr - expression that must match zero or more times - default (optional) - value to be returned if the optional expression is not found. Example:: # US postal code can be a 5-digit zip, plus optional 4-digit qualifier zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4))) zip.runTests(''' # traditional ZIP code 12345 # ZIP+4 form 12101-0001 # invalid ZIP 98765- ''') prints:: # traditional ZIP code 12345 ['12345'] # ZIP+4 form 12101-0001 ['12101-0001'] # invalid ZIP 98765- ^ FAIL: Expected end of text (at char 5), (line:1, col:6) cs.tt|j|dd|jj|_||_d|_dS)NFroT)rQr.rrrrAr)rrr<rTrrrs zOptional.__init__Tc Cs|z|jj|||dd\}}WnVttfk rr|j|jk rj|jjr`t|jg}|j||jj<qn|jg}ng}YnX||fSr )rrr2rrA_Optional__optionalNotMatchedrr5)rrrrrrrrrs    zOptional.parseImplcCs4t|dr|jS|jdkr.dt|jd|_|jS)NrrUrXrrrrrrs   zOptional.__str__)T) rrrrr r rrrrirrrTrr.s $ cs,eZdZdZd fdd Zd ddZZS) r;a Token for skipping over all undefined text until the matched expression is found. Parameters: - expr - target expression marking the end of the data to be skipped - include - (default= ``False``) if True, the target expression is also parsed (the skipped text and target expression are returned as a 2-element list). - ignore - (default= ``None``) used to define grammars (typically quoted strings and comments) that might contain false matches to the target expression - failOn - (default= ``None``) define expressions that are not allowed to be included in the skipped test; if found before the target expression is found, the SkipTo is not a match Example:: report = ''' Outstanding Issues Report - 1 Jan 2000 # | Severity | Description | Days Open -----+----------+-------------------------------------------+----------- 101 | Critical | Intermittent system crash | 6 94 | Cosmetic | Spelling error on Login ('log|n') | 14 79 | Minor | System slow when running too many reports | 47 ''' integer = Word(nums) SEP = Suppress('|') # use SkipTo to simply match everything up until the next SEP # - ignore quoted strings, so that a '|' character inside a quoted string does not match # - parse action will call token.strip() for each matched token, i.e., the description body string_data = SkipTo(SEP, ignore=quotedString) string_data.setParseAction(tokenMap(str.strip)) ticket_expr = (integer("issue_num") + SEP + string_data("sev") + SEP + string_data("desc") + SEP + integer("days_open")) for tkt in ticket_expr.searchString(report): print tkt.dump() prints:: ['101', 'Critical', 'Intermittent system crash', '6'] - days_open: 6 - desc: Intermittent system crash - issue_num: 101 - sev: Critical ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14'] - days_open: 14 - desc: Spelling error on Login ('log|n') - issue_num: 94 - sev: Cosmetic ['79', 'Minor', 'System slow when running too many reports', '47'] - days_open: 47 - desc: System slow when running too many reports - issue_num: 79 - sev: Minor FNcs`tt||||_d|_d|_||_d|_t|t rF| ||_ n||_ dt |j |_dS)NTFzNo match found for )rQr;r ignoreExprrr includeMatchrrrrfailOnrrr)rrJZincluderFrrTrrr s zSkipTo.__init__Tc Cs&|}t|}|j}|jj}|jdk r,|jjnd}|jdk rB|jjnd} |} | |kr|dk rf||| rfq| dk rz| || } Wqntk rYqYqnXqnz||| dddWqtt fk r| d7} YqJXqqJt|||j || }|||} t | } |j r||||dd\}} | | 7} || fS)NF)rrrr) rrrrrrrr0r2rrr5r)rrrrrrrZ expr_parseZself_failOn_canParseNextZself_ignoreExpr_tryParseZtmplocZskiptextZ skipresultrrrrr-s:   zSkipTo.parseImpl)FNN)TrqrrrTrr;s9 csneZdZdZdfdd ZddZddZd d Zd d Zdd dZ ddZ fddZ dfdd Z Z S)r!a_Forward declaration of an expression to be defined later - used for recursive grammars, such as algebraic infix notation. When the expression is known, it is assigned to the ``Forward`` variable using the '<<' operator. Note: take care when assigning to ``Forward`` not to overlook precedence of operators. Specifically, '|' has a lower precedence than '<<', so that:: fwdExpr << a | b | c will actually be evaluated as:: (fwdExpr << a) | b | c thereby leaving b and c out as parseable alternatives. It is recommended that you explicitly group the values inserted into the ``Forward``:: fwdExpr << (a | b | c) Converting to use the '<<=' operator instead will avoid this problem. See :class:`ParseResults.pprint` for an example of a recursive parser created using ``Forward``. Ncstt|j|dddSrn)rQr!rrSrTrrrwszForward.__init__cCsjt|tr||}||_d|_|jj|_|jj|_||jj|jj |_ |jj |_ |j |jj |Sr) rrrrrrrrDrrrrrErSrrr __lshift__zs      zForward.__lshift__cCs||>SrrrSrrr __ilshift__szForward.__ilshift__cCs d|_|Sr~rBrrrrrCszForward.leaveWhitespacecCs$|js d|_|jdk r |j|Sr)rrrrrrrrs   zForward.streamlinecCsJ|dkr g}||kr<|dd|g}|jdk r<|j||gdSrrrrrrrKs  zForward.validatecCslt|dr|jS|jdk r |jSd|_d}z&|jdk rJt|jdd}nd}W5|jjd||_X|jS)Nrz: ...rjz: iNone)r1rrrrrr)rZ retStringrrrrs   zForward.__str__cs.|jdk rtt|St}||K}|SdSr)rrQr!rIrcrTrrrIs  z Forward.copyFcs@tjr.|jdkr.tjdd|t|jddtt | ||S)NzR{0}: setting results name {0!r} on {1} expression that has no contained expressionwarn_name_set_on_empty_Forwardrr$) rrrr(r)rrrrQr!rrrTrrrs zForward._setResultsName)N)N)F)rrrrrrrrCrrKrrIrrirrrTrr!\s   cs"eZdZdZdfdd ZZS)r@zW Abstract subclass of :class:`ParseExpression`, for converting parsed results. Fcstt||d|_dSr~)rQr@rrrrTrrrszTokenConverter.__init__)FrprrrTrr@scs6eZdZdZd fdd ZfddZdd ZZS) raConverter to concatenate all matching tokens to a single string. By default, the matching patterns must also be contiguous in the input string; this can be disabled by specifying ``'adjacent=False'`` in the constructor. Example:: real = Word(nums) + '.' + Word(nums) print(real.parseString('3.1416')) # -> ['3', '.', '1416'] # will also erroneously match the following print(real.parseString('3. 1416')) # -> ['3', '.', '1416'] real = Combine(Word(nums) + '.' + Word(nums)) print(real.parseString('3.1416')) # -> ['3.1416'] # no match when there are internal spaces print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...) rTcs8tt|||r|||_d|_||_d|_dSr)rQrrrCadjacentr joinStringr)rrrrrTrrrszCombine.__init__cs(|jrt||ntt|||Sr)rr7rFrQrrSrTrrrFszCombine.ignorecCsP|}|dd=|td||jg|jd7}|jrH|rH|gS|SdS)Nr)r)rIr5rrYrrrr;)rrrrZretToksrrrrs  "zCombine.postParse)rT)rrrrrrFrrirrrTrrs cs(eZdZdZfddZddZZS)r#aConverter to return the matched tokens as a list - useful for returning tokens of :class:`ZeroOrMore` and :class:`OneOrMore` expressions. Example:: ident = Word(alphas) num = Word(nums) term = ident | num func = ident + Optional(delimitedList(term)) print(func.parseString("fn a, b, 100")) # -> ['fn', 'a', 'b', '100'] func = ident + Group(Optional(delimitedList(term))) print(func.parseString("fn a, b, 100")) # -> ['fn', ['a', 'b', '100']] cstt||d|_dSr)rQr#rrrrTrrr szGroup.__init__cCs|gSrrrrrrrszGroup.postParserrrrrrrirrrTrr#s cs(eZdZdZfddZddZZS)ra?Converter to return a repetitive expression as a list, but also as a dictionary. Each element can also be referenced using the first token in the expression as its key. Useful for tabular report scraping when the first column can be used as a item key. Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) # print attributes as plain groups print(OneOrMore(attr_expr).parseString(text).dump()) # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names result = Dict(OneOrMore(Group(attr_expr))).parseString(text) print(result.dump()) # access named fields as dict entries, or output as dict print(result['shape']) print(result.asDict()) prints:: ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap'] [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'} See more examples at :class:`ParseResults` of accessing fields by results name. cstt||d|_dSr)rQrrrrrTrrr8sz Dict.__init__cCst|D]\}}t|dkrq|d}t|tr@t|d}t|dkr\td|||<qt|dkrt|dtst|d|||<q|}|d=t|dkst|tr| rt||||<qt|d|||<q|j r|gS|SdS)Nrrrr) rrrrrrrr5rIr;r)rrrrrtokZikeyZ dictvaluerrrr<s$   zDict.postParserrrrTrrs& c@s eZdZdZddZddZdS)r>a[Converter for ignoring the results of a parsed expression. Example:: source = "a, b, c,d" wd = Word(alphas) wd_list1 = wd + ZeroOrMore(',' + wd) print(wd_list1.parseString(source)) # often, delimiters that are useful during parsing are just in the # way afterward - use Suppress to keep them out of the parsed output wd_list2 = wd + ZeroOrMore(Suppress(',') + wd) print(wd_list2.parseString(source)) prints:: ['a', ',', 'b', ',', 'c', ',', 'd'] ['a', 'b', 'c', 'd'] (See also :class:`delimitedList`.) cCsgSrrrrrrrkszSuppress.postParsecCs|SrrrrrrrAnszSuppress.suppressN)rrrrrrArrrrr>Usc@s(eZdZdZddZddZddZdS) r-zDWrapper for parse actions, to ensure they are only called once. cCst||_d|_dSr~)rrcalled)rZ methodCallrrrrus zOnlyOnce.__init__cCs.|js||||}d|_|St||ddS)NTr)rrr2)rrrrrrrrr@xs zOnlyOnce.__call__cCs d|_dSr~)rrrrrreset~szOnlyOnce.resetN)rrrrrr@rrrrrr-rscs:tfdd}z j|_Wntk r4YnX|S)aqDecorator for debugging parse actions. When the parse action is called, this decorator will print ``">> entering method-name(line:, , )"``. When the parse action completes, the decorator will print ``"<<"`` followed by the returned value, or any exception that the parse action raised. Example:: wd = Word(alphas) @traceParseAction def remove_duplicate_chars(tokens): return ''.join(sorted(set(''.join(tokens)))) wds = OneOrMore(wd).setParseAction(remove_duplicate_chars) print(wds.parseString("slkdjs sld sldd sdlf sdljf")) prints:: >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {})) <>entering %s(line: '%s', %d, %r) z<.z)rrr)rOrrrrrvs  ,cCs`t|dt|dt|d}|rBt|t|||S|tt|||SdS)aHelper to define a delimited list of expressions - the delimiter defaults to ','. By default, the list elements and delimiters can have intervening whitespace, and comments, but this can be overridden by passing ``combine=True`` in the constructor. If ``combine`` is set to ``True``, the matching tokens are returned as a single token string, with the delimiters included; otherwise, the matching tokens are returned as a list of tokens, with the delimiters suppressed. Example:: delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc'] delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE'] z [rr N)rrrErr>)rZdelimcombineZdlNamerrrrTs$csjtfdd}|dkr0ttdd}n|}|d|j|dd|d td S) a>Helper to define a counted list of expressions. This helper defines a pattern of the form:: integer expr expr expr... where the leading integer tells how many expr expressions follow. The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed. If ``intExpr`` is specified, it should be a pyparsing expression that produces an integer value. Example:: countedArray(Word(alphas)).parseString('2 ab cd ef') # -> ['ab', 'cd'] # in this parser, the leading integer value is given in binary, # '10' indicating that 2 values are in the array binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2)) countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef') # -> ['ab', 'cd'] cs.|d}|r ttg|p&tt>gSr)r#rrW)rrrr1Z arrayExprrrrcountFieldParseActions"z+countedArray..countFieldParseActionNcSs t|dSr)rrrrrrrzcountedArray..ZarrayLenTrz(len) rj)r!rBrfrrIrrr)rZintExprr"rr!rrPs cCs6g}|D](}t|tr&|t|q||q|Sr)rr rEr r)Lrrrrrr s   r cs6tfdd}|j|dddt|S)a4Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousLiteral(first) matchExpr = first + ":" + second will match ``"1:1"``, but not ``"1:2"``. Because this matches a previous literal, will also match the leading ``"1:1"`` in ``"1:10"``. If this is not desired, use :class:`matchPreviousExpr`. Do *not* use with packrat parsing enabled. csP|rBt|dkr|d>qLt|}tdd|D>n t>dS)Nrrcss|]}t|VqdSr)r'rZttrrrr szDmatchPreviousLiteral..copyTokenToRepeater..)rr rrr)rrrZtflatZreprrcopyTokenToRepeaters   z1matchPreviousLiteral..copyTokenToRepeaterTr#(prev) )r!rrr)rr'rr&rrcs  csFt|}|Kfdd}|j|dddt|S)aTHelper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousExpr(first) matchExpr = first + ":" + second will match ``"1:1"``, but not ``"1:2"``. Because this matches by expressions, will *not* match the leading ``"1:1"`` in ``"1:10"``; the expressions are evaluated first, and then compared, so ``"1"`` is compared with ``"10"``. Do *not* use with packrat parsing enabled. cs*t|fdd}j|dddS)Ncs$t|}|kr tddddS)Nrr)r rr2)rrrZ theseTokensZ matchTokensrrmustMatchTheseTokens%s zLmatchPreviousExpr..copyTokenToRepeater..mustMatchTheseTokensTr#)r rr)rrrr*r&r)rr'#s  z.matchPreviousExpr..copyTokenToRepeaterTr#r()r!rIrrr)rZe2r'rr&rrbs cCs:dD]}||t|}q|dd}|dd}t|S)Nz\^-]rr\rr)r_bslashr)rrrrrr.s   rc st|trtjddd|r:dd}dd}|r4tntndd}dd}|rRtntg}t|trn|}n$t|t rt |}ntjd t dd|st S|s.d }|t |d kr.||}t||d d D]N\} } || |r||| d =qq||| r||| d =||| qq|d 7}q|s|s|rzlt |t d |krtdd dd|Dd|WStddd|Dd|WSWn&tk rtjdt ddYnXtfdd|Dd|S)aHelper to quickly define a set of alternative Literals, and makes sure to do longest-first testing when there is a conflict, regardless of the input order, but returns a :class:`MatchFirst` for best performance. Parameters: - strs - a string of space-delimited literals, or a collection of string literals - caseless - (default= ``False``) - treat all literals as caseless - useRegex - (default= ``True``) - as an optimization, will generate a Regex object; otherwise, will generate a :class:`MatchFirst` object (if ``caseless=True`` or ``asKeyword=True``, or if creating a :class:`Regex` raises an exception) - asKeyword - (default=``False``) - enforce Keyword-style matching on the generated expressions Example:: comp_oper = oneOf("< = > <= >= !=") var = Word(alphas) number = Word(nums) term = var | number comparison_expr = term + comp_oper + term print(comparison_expr.searchString("B = 12 AA=23 B<=AA AA>12")) prints:: [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']] z_More than one string argument passed to oneOf, pass choices as a list or space-delimited stringrr$cSs||kSr)r~rLbrrrr[rzoneOf..cSs||Sr)r~ryr,rrrr\rcSs||kSrrr,rrrr_rcSs ||Sr)ryr,rrrr`rz6Invalid argument to oneOf, expected string or iterablerrNrrcss|]}t|VqdSr)rrZsymrrrrszoneOf..r|css|]}t|VqdSr)rrr.rrrrsz7Exception creating Regex for oneOf, building MatchFirstc3s|]}|VqdSrrr.ZparseElementClassrrrs)rrr(r)rrr$r'rr r r*r*rrrBrr:rrr)) Zstrsr}ZuseRegexrZisequalZmasksZsymbolsrZcurr#rJrr0rrg6s\          ** cCsttt||S)aHelper to easily and clearly define a dictionary by specifying the respective patterns for the key and value. Takes care of defining the :class:`Dict`, :class:`ZeroOrMore`, and :class:`Group` tokens in the proper order. The key pattern can include delimiting markers or punctuation, as long as they are suppressed, thereby leaving the significant key text. The value pattern can include named results, so that the :class:`Dict` results can include named token fields. Example:: text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) print(OneOrMore(attr_expr).parseString(text).dump()) attr_label = label attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join) # similar to Dict, but simpler call format result = dictOf(attr_label, attr_value).parseString(text) print(result.dump()) print(result['shape']) print(result.shape) # object attribute access works too print(result.asDict()) prints:: [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE SQUARE {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'} )rr,r#)r@r$rrrrUs%cCs^tdd}|}d|_|d||d}|r@dd}ndd}|||j|_|S) aHelper to return the original, untokenized text for a given expression. Useful to restore the parsed fields of an HTML start tag into the raw tag text itself, or to revert separate tokens with intervening whitespace back to the original matching input text. By default, returns astring containing the original parsed text. If the optional ``asString`` argument is passed as ``False``, then the return value is a :class:`ParseResults` containing any results names that were originally matched, and a single token containing the original matched text from the input string. So if the expression passed to :class:`originalTextFor` contains expressions with defined results names, you must set ``asString`` to ``False`` if you want to preserve those results name values. Example:: src = "this is test bold text normal text " for tag in ("b", "i"): opener, closer = makeHTMLTags(tag) patt = originalTextFor(opener + SkipTo(closer) + closer) print(patt.searchString(src)[0]) prints:: [' bold text '] ['text'] cSs|Srr)rrrrrrrrz!originalTextFor..F_original_start _original_endcSs||j|jSr)r1r2rrrrrrcSs&||d|dg|dd<dS)Nr1r2)r?rrrr extractTextsz$originalTextFor..extractText)rrrIrr)rZasStringZ locMarkerZ endlocMarker matchExprr3rrrr{s  cCst|ddS)zkHelper to undo pyparsing's default grouping of And expressions, even if all but one are non-empty. cSs|dSrrrrrrrrzungroup..)r@r)rrrrr|scCs4tdd}t|d|d|dS)aHelper to decorate a returned token with its starting and ending locations in the input string. This helper adds the following results names: - locn_start = location where matched expression begins - locn_end = location where matched expression ends - value = the actual parsed results Be careful if the input text contains ```` characters, you may want to call :class:`ParserElement.parseWithTabs` Example:: wd = Word(alphas) for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"): print(match) prints:: [[0, 'ljsdf', 5]] [[8, 'lksdjjf', 15]] [[18, 'lkkjj', 23]] cSs|SrrrrrrrrzlocatedExpr..Z locn_startr$Zlocn_end)rrr#rIrC)rZlocatorrrrr~sz\[]-*.$+^?()~ rcCs |ddSrrrrrrrrrz\\0?[xX][0-9a-fA-F]+cCstt|dddS)Nrz\0xr)unichrrrarrrrr rz \\0[0-7]+cCstt|ddddS)Nrr)r6rrrrrr rz\]rrUrnegatebodyrXcsFddz"dfddt|jDWStk r@YdSXdS)aHelper to easily define string ranges for use in Word construction. Borrows syntax from regexp '[]' string range definitions:: srange("[0-9]") -> "0123456789" srange("[a-z]") -> "abcdefghijklmnopqrstuvwxyz" srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_" The input string must be enclosed in []'s, and the returned string is the expanded character set joined into a single string. The values enclosed in the []'s may be: - a single character - an escaped character with a leading backslash (such as ``\-`` or ``\]``) - an escaped hex character with a leading ``'\x'`` (``\x21``, which is a ``'!'`` character) (``\0x##`` is also supported for backwards compatibility) - an escaped octal character with a leading ``'\0'`` (``\041``, which is a ``'!'`` character) - a range of any of the above, separated by a dash (``'a-z'``, etc.) - any combination of the above (``'aeiouy'``, ``'a-zA-Z0-9_$'``, etc.) cSs<t|ts|Sdddtt|dt|ddDS)Nrcss|]}t|VqdSrr6rrrrr)sz+srange....rr)rr5rrord)prrrr)rzsrange..rc3s|]}|VqdSrr)rpartZ _expandedrrr+szsrange..N)r_reBracketExprrr9rrrr>rrss "csfdd}|S)zoHelper method for defining parse actions that require matching at a specific column in the input text. cs"t||krt||ddS)Nzmatched token not at column %dr)rZlocnrr0rr verifyCol3sz!matchOnlyAtCol..verifyColr)r1r@rr0rra/s cs fddS)aHelper method for common parse actions that simply return a literal value. Especially useful when used with :class:`transformString` (). Example:: num = Word(nums).setParseAction(lambda toks: int(toks[0])) na = oneOf("N/A NA").setParseAction(replaceWith(math.nan)) term = na | num OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234] csgSrrrZreplStrrrrErzreplaceWith..rrArrArrp8s cCs|dddS)aHelper parse action for removing quotation marks from parsed quoted strings. Example:: # by default, quotation marks are included in parsed results quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"] # use removeQuotes to strip quotation marks from parsed results quotedString.setParseAction(removeQuotes) quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"] rrrrrrrrrnGs csNfdd}ztdtdj}Wntk rBt}YnX||_|S)aLHelper to define a parse action by mapping a function to all elements of a ParseResults list. If any additional args are passed, they are forwarded to the given function as additional arguments after the token, as in ``hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))``, which will convert the parsed data to an integer using base 16. Example (compare the last to example in :class:`ParserElement.transformString`:: hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16)) hex_ints.runTests(''' 00 11 22 aa FF 0a 0d 1a ''') upperword = Word(alphas).setParseAction(tokenMap(str.upper)) OneOrMore(upperword).runTests(''' my kingdom for a horse ''') wd = Word(alphas).setParseAction(tokenMap(str.title)) OneOrMore(wd).setParseAction(' '.join).runTests(''' now is the winter of our discontent made glorious summer by this sun of york ''') prints:: 00 11 22 aa FF 0a 0d 1a [0, 17, 34, 170, 255, 10, 13, 26] my kingdom for a horse ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE'] now is the winter of our discontent made glorious summer by this sun of york ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York'] csfdd|DS)Ncsg|]}|fqSrr)rZtoknrrrrr{sz(tokenMap..pa..rrrBrrrzsztokenMap..parr)rrrr)rrrrrrBrrVs$ cCs t|Srrr~rrrrrrcCs t|Srrlowerrrrrrrrfrgcs~t|tr|t|| d}n|jtttd}|rt t }||dt t t |td|tddgdd d d |}nlt t ttd d B}||dt t t | tttd|tddgdd d d |}ttd|d dd}|d|fdd |ddddd}|_|_t||_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namerz_-:tag=/Fr<rWcSs |ddkSNrrHrrrrrrrz_makeTags..rgrcSs |ddkSrJrrrrrrrrh)rz<%s>c s*|dddd|S)Nrr:r)rrrtitlerrIrZresnamerrrrrrrKrz)rrr$rrBrHrGrRrIrrnrrEr#r>r.rmrjrVr_LrrrrrLrrFr;Ztag_body)tagStrZxmlZ suppress_LTZ suppress_GTZ tagAttrNameZ tagAttrValueZopenTagZcloseTagrrMr _makeTagssH , rPcCs t|dS)aKHelper to construct opening and closing tag expressions for HTML, given a tag name. Matches tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values. Example:: text = 'More info at the
pyparsing wiki page' # makeHTMLTags returns pyparsing expressions for the opening and # closing tags as a 2-tuple a, a_end = makeHTMLTags("A") link_expr = a + SkipTo(a_end)("link_text") + a_end for link in link_expr.searchString(text): # attributes in the tag (like "href" shown here) are # also accessible as named results print(link.link_text, '->', link.href) prints:: pyparsing -> https://github.com/pyparsing/pyparsing/wiki FrPrOrrrr_scCs t|dS)zHelper to construct opening and closing tag expressions for XML, given a tag name. Matches tags only in the given upper/lower case. Example: similar to :class:`makeHTMLTags` TrQrRrrrr`scs8|r|ddn|ddDfdd}|S)a7Helper to create a validating parse action to be used with start tags created with :class:`makeXMLTags` or :class:`makeHTMLTags`. Use ``withAttribute`` to qualify a starting tag with a required attribute value, to avoid false matches on common tags such as ```` or ``
``. Call ``withAttribute`` with a series of attribute names and values. Specify the list of filter attributes names and values as: - keyword arguments, as in ``(align="right")``, or - as an explicit dict with ``**`` operator, when an attribute name is also a Python reserved word, as in ``**{"class":"Customer", "align":"right"}`` - a list of name-value tuples, as in ``(("ns1:class", "Customer"), ("ns2:align", "right"))`` For attribute names with a namespace prefix, you must use the second form. Attribute names are matched insensitive to upper/lower case. If just testing for ``class`` (with or without a namespace), use :class:`withClass`. To verify that the attribute exists, but without specifying a value, pass ``withAttribute.ANY_VALUE`` as the value. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this has no type
''' div,div_end = makeHTMLTags("div") # only match div tag having a type attribute with value "grid" div_grid = div().setParseAction(withAttribute(type="grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) # construct a match with any div tag having a type attribute, regardless of the value div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 NcSsg|]\}}||fqSrrrarrrrsz!withAttribute..csZD]P\}}||kr$t||d||tjkr|||krt||d||||fqdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r2ry ANY_VALUE)rrrZattrNameZ attrValueZattrsrrrs  zwithAttribute..pa)r!)rZattrDictrrrTrrys 8 cCs|r d|nd}tf||iS)aSimplified version of :class:`withAttribute` when matching on a div class - made difficult because ``class`` is a reserved word in Python. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this <div> has no class
''' div,div_end = makeHTMLTags("div") div_grid = div().setParseAction(withClass("grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 z%s:classclass)ry)Z classname namespaceZ classattrrrrrs#(rcCsGdddt}t}||||B}t|D]n\}}|ddd\} } } } | dkr`d| nd| } | dkr| dkst| d krtd | \}}t| }| tjkrp| d kr||| t|t | }n| d kr*| dk r ||| |t|t | |}n|||t|t |}nD| dkrf||||||t|||||}ntd n| tj krT| d krt | t st | } || j |t| |}n| d kr| dk r||| |t|t | |}n|||t|t |}nD| dkrJ||||||t|||||}ntd ntd | rt | ttfr~|j| n || ||| |BK}|}q.||K}|S)al Helper method for constructing grammars of expressions made up of operators working in a precedence hierarchy. Operators may be unary or binary, left- or right-associative. Parse actions can also be attached to operator expressions. The generated parser will also recognize the use of parentheses to override operator precedences (see example below). Note: if you define a deep operator list, you may see performance issues when using infixNotation. See :class:`ParserElement.enablePackrat` for a mechanism to potentially improve your parser performance. Parameters: - baseExpr - expression representing the most basic element for the nested - opList - list of tuples, one for each operator precedence level in the expression grammar; each tuple is of the form ``(opExpr, numTerms, rightLeftAssoc, parseAction)``, where: - opExpr is the pyparsing expression for the operator; may also be a string, which will be converted to a Literal; if numTerms is 3, opExpr is a tuple of two expressions, for the two operators separating the 3 terms - numTerms is the number of terms for this operator (must be 1, 2, or 3) - rightLeftAssoc is the indicator whether the operator is right or left associative, using the pyparsing-defined constants ``opAssoc.RIGHT`` and ``opAssoc.LEFT``. - parseAction is the parse action to be associated with expressions matching this operator expression (the parse action tuple member may be omitted); if the parse action is passed a tuple or list of functions, this is equivalent to calling ``setParseAction(*fn)`` (:class:`ParserElement.setParseAction`) - lpar - expression for matching left-parentheses (default= ``Suppress('(')``) - rpar - expression for matching right-parentheses (default= ``Suppress(')')``) Example:: # simple example of four-function arithmetic with ints and # variable names integer = pyparsing_common.signed_integer varname = pyparsing_common.identifier arith_expr = infixNotation(integer | varname, [ ('-', 1, opAssoc.RIGHT), (oneOf('* /'), 2, opAssoc.LEFT), (oneOf('+ -'), 2, opAssoc.LEFT), ]) arith_expr.runTests(''' 5+3*6 (5+3)*6 -2--11 ''', fullDump=False) prints:: 5+3*6 [[5, '+', [3, '*', 6]]] (5+3)*6 [[[5, '+', 3], '*', 6]] -2--11 [[['-', 2], '-', ['-', 11]]] c@seZdZdddZdS)zinfixNotation.._FBTcSs|j|||gfSr)rrrrrrrsz$infixNotation.._FB.parseImplN)Trzrrrr_FBsrXrNrrz%s termz %s%s termrz@if numterms=3, opExpr must be a tuple or list of two expressionsrz6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)r r!rrr5rrhLEFTr#r,RIGHTrr.rr4r r)ZbaseExprZopListZlparZrparrXrZlastExprrZoperDefZopExprZarityZrightLeftAssocrZtermNameZopExpr1ZopExpr2ZthisExprr4rrrr}Ds`H    &       &    z4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*'z string enclosed in single quotesz*quotedString using single or double quotesuzunicode string literalcCs||krtd|dkr*t|tr"t|tr"t|dkrt|dkr|dk rtt|t||tjdd dd}n$t t||tj dd}nx|dk rtt|t |t |ttjdd dd}n4ttt |t |ttjdd d d}ntd t }|dk rd|tt|t||B|Bt|K}n$|tt|t||Bt|K}|d ||f|S) a Helper method for defining nested lists enclosed in opening and closing delimiters ("(" and ")" are the default). Parameters: - opener - opening character for a nested list (default= ``"("``); can also be a pyparsing expression - closer - closing character for a nested list (default= ``")"``); can also be a pyparsing expression - content - expression for items within the nested lists (default= ``None``) - ignoreExpr - expression for ignoring opening and closing delimiters (default= :class:`quotedString`) If an expression is not provided for the content argument, the nested expression will capture all whitespace-delimited content between delimiters as a list of separate values. Use the ``ignoreExpr`` argument to define expressions that may contain opening or closing characters that should not be treated as opening or closing characters for nesting, such as quotedString or a comment expression. Specify multiple expressions using an :class:`Or` or :class:`MatchFirst`. The default is :class:`quotedString`, but if no expressions are to be ignored, then pass ``None`` for this argument. Example:: data_type = oneOf("void int short long char float double") decl_data_type = Combine(data_type + Optional(Word('*'))) ident = Word(alphas+'_', alphanums+'_') number = pyparsing_common.number arg = Group(decl_data_type + ident) LPAR, RPAR = map(Suppress, "()") code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment)) c_function = (decl_data_type("type") + ident("name") + LPAR + Optional(delimitedList(arg), [])("args") + RPAR + code_body("body")) c_function.ignore(cStyleComment) source_code = ''' int is_odd(int x) { return (x%2); } int dec_to_hex(char hchar) { if (hchar >= '0' && hchar <= '9') { return (ord(hchar)-ord('0')); } else { return (10+ord(hchar)-ord('A')); } } ''' for func in c_function.searchString(source_code): print("%(name)s (%(type)s) args: %(args)s" % func) prints:: is_odd (int) args: [['int', 'x']] dec_to_hex (int) args: [['char', 'hchar']] z.opening and closing strings cannot be the sameNrr5cSs |dSrrrrrrrrznestedExpr..cSs |dSrr^rrrrr rcSs |dSrr^rrrrr'rcSs |dSrr^rrrrr,rzOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)r5rrrrr,rr7rrrWrIr'r!r#r>rEr)ZopenerZcloserZcontentrrrrrrds`A      *$c sddfddfdd}fdd}fdd }ttd }tt|d }t|d }t|d } |rtt||t|t|t|| } n(tt|t|t|t|| } | fdd| t t| dS)aHelper method for defining space-delimited indentation blocks, such as those used to define block statements in Python source code. Parameters: - blockStatementExpr - expression defining syntax of statement that is repeated within the indented block - indentStack - list created by caller to manage indentation stack (multiple statementWithIndentedBlock expressions within a single grammar should share a common indentStack) - indent - boolean indicating whether block must be indented beyond the current level; set to False for block of left-most statements (default= ``True``) A valid block must contain at least one ``blockStatement``. Example:: data = ''' def A(z): A1 B = 100 G = A2 A2 A3 B def BB(a,b,c): BB1 def BBA(): bba1 bba2 bba3 C D def spam(x,y): def eggs(z): pass ''' indentStack = [1] stmt = Forward() identifier = Word(alphas, alphanums) funcDecl = ("def" + identifier + Group("(" + Optional(delimitedList(identifier)) + ")") + ":") func_body = indentedBlock(stmt, indentStack) funcDef = Group(funcDecl + func_body) rvalue = Forward() funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")") rvalue << (funcCall | identifier | Word(nums)) assignment = Group(identifier + "=" + rvalue) stmt << (funcDef | assignment | identifier) module_body = OneOrMore(stmt) parseTree = module_body.parseString(data) parseTree.pprint() prints:: [['def', 'A', ['(', 'z', ')'], ':', [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]], 'B', ['def', 'BB', ['(', 'a', 'b', 'c', ')'], ':', [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]], 'C', 'D', ['def', 'spam', ['(', 'x', 'y', ')'], ':', [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] Ncsdd<dSrrr) backup_stack indentStackrr reset_stacksz"indentedBlock..reset_stackcsN|t|krdSt||}|dkrJ|dkr>t||dt||ddS)Nrzillegal nestingznot a peer entry)rrMr2rrrZcurColr`rrcheckPeerIndents     z&indentedBlock..checkPeerIndentcs2t||}|dkr"|n t||ddS)Nrznot a subentry)rMrr2rbrcrrcheckSubIndents   z%indentedBlock..checkSubIndentcsJ|t|krdSt||}r&|ks2t||d|dkrFdS)Nznot an unindentr)rrMr2r?rbrcrr checkUnindents     z$indentedBlock..checkUnindentz INDENTrZUNINDENTcsSrr)rLr-rr)rarrrrzindentedBlock..zindented block) r,r%rDrArrrr#r.rrFr+) ZblockStatementExprr`rjrdrerfrwrgZPEERZUNDENTZsmExprr)r_r`rarrz7s2Q    z#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z &(?Pr/z);zcommon HTML entitycCs t|jS)zRHelper parser action to replace common HTML entities with their special characters)_htmlEntityMaprZentityrrrrrosz/\*(?:[^*]|\*(?!/))*z*/zC style commentzz HTML commentz.*z rest of linez//(?:\\\n|[^\n])*z // commentzC++ style commentz#.*zPython style commentr commaItemrIc@seZdZdZeeZeeZe e  d eZ e e d eedZed d eZe ede e dZed d eeeed eB d Zeeed  d eZed d eZeeBeBZed d eZe eded dZed dZ ed dZ!e!de!d dZ"ee!de!ddee!de!d dZ#e#$dd d e  d!Z%e&e"e%Be#B d" d"Z'ed# d$Z(e)d=d&d'Z*e)d>d)d*Z+ed+ d,Z,ed- d.Z-ed/ d0Z.e/e0BZ1e)d1d2Z2e&e3e4d3e5e e6d3d4ee7d5 d6Z8e9ee:;e8Bd7d8 d9Zd`, :class:`reals`, :class:`scientific notation`) - common :class:`programming identifiers` - network addresses (:class:`MAC`, :class:`IPv4`, :class:`IPv6`) - ISO8601 :class:`dates` and :class:`datetime` - :class:`UUID` - :class:`comma-separated list` Parse actions: - :class:`convertToInteger` - :class:`convertToFloat` - :class:`convertToDate` - :class:`convertToDatetime` - :class:`stripHTMLTags` - :class:`upcaseTokens` - :class:`downcaseTokens` Example:: pyparsing_common.number.runTests(''' # any int or real number, returned as the appropriate type 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.fnumber.runTests(''' # any int or real number, returned as float 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.hex_integer.runTests(''' # hex numbers 100 FF ''') pyparsing_common.fraction.runTests(''' # fractions 1/2 -3/4 ''') pyparsing_common.mixed_integer.runTests(''' # mixed fractions 1 1/2 -3/4 1-3/4 ''') import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(''' # uuid 12345678-1234-5678-1234-567812345678 ''') prints:: # any int or real number, returned as the appropriate type 100 [100] -100 [-100] +100 [100] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # any int or real number, returned as float 100 [100.0] -100 [-100.0] +100 [100.0] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # hex numbers 100 [256] FF [255] # fractions 1/2 [0.5] -3/4 [-0.75] # mixed fractions 1 [1] 1/2 [0.5] -3/4 [-0.75] 1-3/4 [1.75] # uuid 12345678-1234-5678-1234-567812345678 [UUID('12345678-1234-5678-1234-567812345678')] integerz hex integerrz[+-]?\d+zsigned integerrHfractioncCs|d|dS)Nrrrrrrrrrzpyparsing_common.rz"fraction or mixed integer-fractionz[+-]?(:?\d+\.\d*|\.\d+)z real numberz@[+-]?(:?\d+(:?[eE][+-]?\d+)|(:?\d+\.\d*|\.\d+)(:?[eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?fnumberr identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}z IPv4 addressz[0-9a-fA-F]{1,4} hex_integerrKzfull IPv6 address)rrz::zshort IPv6 addresscCstdd|DdkS)Ncss|]}tj|rdVqdSrd)r _ipv6_partrr%rrrrs z,pyparsing_common...r7)rrrrrrrz::ffff:zmixed IPv6 addressz IPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}z MAC address%Y-%m-%dcsfdd}|S)a Helper to create a parse action for converting parsed date string to Python datetime.date Params - - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%d"``) Example:: date_expr = pyparsing_common.iso8601_date.copy() date_expr.setParseAction(pyparsing_common.convertToDate()) print(date_expr.parseString("1999-12-31")) prints:: [datetime.date(1999, 12, 31)] c sNzt|dWStk rH}zt||t|W5d}~XYnXdSr)rstrptimedater5r2rrrrZvefmtrrcvt_fnsz.pyparsing_common.convertToDate..cvt_fnrrwrxrrvr convertToDates zpyparsing_common.convertToDate%Y-%m-%dT%H:%M:%S.%fcsfdd}|S)aHelper to create a parse action for converting parsed datetime string to Python datetime.datetime Params - - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%dT%H:%M:%S.%f"``) Example:: dt_expr = pyparsing_common.iso8601_datetime.copy() dt_expr.setParseAction(pyparsing_common.convertToDatetime()) print(dt_expr.parseString("1999-12-31T23:59:59.999")) prints:: [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)] c sJzt|dWStk rD}zt||t|W5d}~XYnXdSr)rrsr5r2rrurvrrrxsz2pyparsing_common.convertToDatetime..cvt_fnrryrrvrconvertToDatetimes z"pyparsing_common.convertToDatetimez7(?P\d{4})(?:-(?P\d\d)(?:-(?P\d\d))?)?z ISO8601 datez(?P\d{4})-(?P\d\d)-(?P\d\d)[T ](?P\d\d):(?P\d\d)(:(?P\d\d(\.\d*)?)?)?(?PZ|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}UUIDcCstj|dS)aParse action to remove HTML tags from web page HTML source Example:: # strip HTML links from normal text text = 'More info at the
pyparsing wiki page' td, td_end = makeHTMLTags("TD") table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end print(table_text.parseString(text).body) Prints:: More info at the pyparsing wiki page r)r_html_stripperr)rrrrrr stripHTMLTagsszpyparsing_common.stripHTMLTagsrrrirjrrIzcomma separated listcCs t|SrrCrrrrrrcCs t|SrrDrrrrrrN)rr)r{)?rrrrrrZconvertToIntegerfloatZconvertToFloatrBrfrrrkrXror:Zsigned_integerrlrr.rAZ mixed_integerrrealZsci_realrnumberrmrHrGrnZ ipv4_addressrqZ_full_ipv6_addressZ_short_ipv6_addressrZ_mixed_ipv6_addressrZ ipv6_addressZ mac_addressrrzr|Z iso8601_dateZiso8601_datetimeuuidrKrJr~rr,r'r%rjrA _commasepitemrTrmrIZcomma_separated_listrxrVrrrrrsv""        c@seZdZddZddZdS)_lazyclasspropertycCs||_|j|_|j|_dSr)rrrrrrrr sz_lazyclassproperty.__init__csldkrt|tdr:tfddjddDr@i_|jj}|jkrb|j|<j|S)N_internc3s |]}jt|dgkVqdS)rN)rr)rZ superclassrrrrsz-_lazyclassproperty.__get__..r)rr1rv__mro__rrr)rrrattrnamerrr__get__s  z_lazyclassproperty.__get__N)rrrrrrrrrr src@sPeZdZdZgZeddZeddZeddZ edd Z ed d Z d S) ra A set of Unicode characters, for language-specific strings for ``alphas``, ``nums``, ``alphanums``, and ``printables``. A unicode_set is defined by a list of ranges in the Unicode character set, in a class attribute ``_ranges``, such as:: _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),] A unicode set can also be defined using multiple inheritance of other unicode sets:: class CJK(Chinese, Japanese, Korean): pass cCsZg}|jD]8}|tkrqD|jD] }|t|d|ddq q ddtt|DS)NrrrcSsg|] }t|qSrr:rrrrr4sz5unicode_set._get_chars_for_ranges..)rr_rangesrErrtr)rrZccZrrrrr_get_chars_for_ranges,s   z!unicode_set._get_chars_for_rangescCsdttj|S)z+all non-whitespace characters in this ranger)rrrrrrrrrrj6szunicode_set.printablescCsdttj|S)z'all alphabetic characters in this ranger)rfilterrisalpharrrrrrH;szunicode_set.alphascCsdttj|S)z*all numeric digit characters in this ranger)rrrisdigitrrrrrrf@szunicode_set.numscCs |j|jS)z)all alphanumeric characters in this range)rHrfrrrrrGEszunicode_set.alphanumsN) rrrrrrrrrjrHrfrGrrrrrs     c@seZdZdZdejfgZGdddeZGdddeZ GdddeZ Gd d d eZ Gd d d eZ Gd ddeZ GdddeZGdddeZGddde eeZGdddeZGdddeZGdddeZGdddeZdS)rzF A namespace class for defining common language unicode_sets. c@seZdZdZddgZdS)zpyparsing_unicode.Latin1z/Unicode set for Latin-1 Unicode Character Range)r~)NrrrrrrrrrLatin1Qsrc@seZdZdZdgZdS)zpyparsing_unicode.LatinAz/Unicode set for Latin-A Unicode Character Range)iNrrrrrLatinAUsrc@seZdZdZdgZdS)zpyparsing_unicode.LatinBz/Unicode set for Latin-B Unicode Character Range)iiONrrrrrLatinBYsrc@s6eZdZdZdddddddd d d d d dddddgZdS)zpyparsing_unicode.Greekz.Unicode set for Greek Unicode Character Ranges)ipi)ii)ii)i iE)iHiM)iPiW)iY)i[)i])i_i})ii)ii)ii)ii)ii)ii)iiNrrrrrGreek]s&rc@seZdZdZdgZdS)zpyparsing_unicode.Cyrillicz0Unicode set for Cyrillic Unicode Character Range)iiNrrrrrCyrillicesrc@seZdZdZddgZdS)zpyparsing_unicode.Chinesez/Unicode set for Chinese Unicode Character Range)Nii0i?0NrrrrrChineseisrc@sDeZdZdZgZGdddeZGdddeZGdddeZdS) zpyparsing_unicode.Japanesez`Unicode set for Japanese Unicode Character Range, combining Kanji, Hiragana, and Katakana rangesc@seZdZdZddgZdS)z pyparsing_unicode.Japanese.Kanjiz-Unicode set for Kanji Unicode Character Range)rirNrrrrrKanjiqsrc@seZdZdZdgZdS)z#pyparsing_unicode.Japanese.Hiraganaz0Unicode set for Hiragana Unicode Character Range)i@0i0NrrrrrHiraganausrc@seZdZdZdgZdS)z#pyparsing_unicode.Japanese.Katakanaz1Unicode set for Katakana Unicode Character Range)i0i0NrrrrrKatakanaysrN) rrrrrrrrrrrrrJapanesems rc@s eZdZdZddddddgZdS) zpyparsing_unicode.Koreanz.Unicode set for Korean Unicode Character Range)ii)ii)i01i1)i`i)iirNrrrrrKorean}src@seZdZdZdS)zpyparsing_unicode.CJKzTUnicode set for combined Chinese, Japanese, and Korean (CJK) Unicode Character RangeNrrrrrCJKsrc@seZdZdZddgZdS)zpyparsing_unicode.Thaiz,Unicode set for Thai Unicode Character Range)ii:)i?i[NrrrrrThaisrc@seZdZdZdddgZdS)zpyparsing_unicode.Arabicz.Unicode set for Arabic Unicode Character Range)ii)ii)iiNrrrrrArabicsrc@seZdZdZdgZdS)zpyparsing_unicode.Hebrewz.Unicode set for Hebrew Unicode Character Range)iiNrrrrrHebrewsrc@seZdZdZddgZdS)zpyparsing_unicode.Devanagariz2Unicode set for Devanagari Unicode Character Range)i i )iiNrrrrr DevanagarisrN)rrrrr maxunicoderrrrrrrrrrrrrrrrrrrrKs uالعربيةu中文uкириллицаuΕλληνικάuעִברִיתu 日本語u漢字u カタカナu ひらがなu 한국어u ไทยuदेवनागरी__main__Zselectfromr{r)r columnsrZtablesZcommandaK # '*' as column list and dotted table name select * from SYS.XYZZY # caseless match on "SELECT", and casts back to "select" SELECT * from XYZZY, ABC # list of column names, and mixed case SELECT keyword Select AA,BB,CC from Sys.dual # multiple tables Select A, B, C from Sys.dual, Table2 # invalid SELECT keyword - should fail Xelect A, B, C from Sys.dual # incomplete command - should fail Select # invalid column name - should fail Select ^^^ frox Sys.dual z] 100 -100 +100 3.14159 6.02e23 1e-12 z 100 FF z6 12345678-1234-5678-1234-567812345678 )NF)r)rF)N)FTF)T)r)T( rrrrrweakrefrrrIrr(rrr rxrrroperatorr itertools functoolsrr ImportErrorr_threadrZ threadingZcollections.abcr r r r rZ ordereddictr rrrrrrZ%warn_on_multiple_string_args_to_oneofr__all__r4 version_inforrbmaxsizerhrrchrr6rrrrrtreversedr rrvrrrrZmaxintZxrangerZ __builtin__rZfnamerrrrr rZascii_uppercaseZascii_lowercaserHrfrXrGr+rZ printablerjrrr0r2r4r6r9rrr5registerrMr^r[rrrrerr7r'r?rr*r'rvrNrr$rrrrBrrFr:r8rrArr"r&r%r=r<rDrCr3rr/r)rr1r r(r+rr,rEr r.r;r!r@rr#rr>r-rvrTrPr rcrbrrgrUr{r|r~rrWr]r\rurtrZ _escapedPuncZ_escapedHexCharZ_escapedOctCharZ _singleCharZ _charRangerr?rsrarprnrrxrVrPr_r`ryrSrrhrYrZr}rirRrrrmrwrdrzrIrkrKrJrrrhr7rOrorLrYrCrqrSrQrZrlrrrNrrrrrrrrrsetattrrrrrrrrrrZ selectTokenZ fromTokenZidentZ columnNameZcolumnNameListZ columnSpecZ tableNameZ tableNameListZ simpleSQLrgrrmrorr}rrrrsH            ?]  H    D! 'N E  KFyl{VO#K,:#Dvj-D0  $  W' *     4   0%   E&h~   (     ./J    &   site-packages/pip/_vendor/__pycache__/distro.cpython-38.pyc000064400000107275151464452710017701 0ustar00U .e@s~dZddlZddlZddlZddlZddlZddlZddlZddlZej ddZ dZ ddiZ dddd Zd diZed Zed Zd dde dfZd@ddZddZdAddZdBddZdCddZdDddZdEdd ZdFd!d"Zd#d$Zd%d&ZdGd'd(Zd)d*Zd+d,Z d-d.Z!d/d0Z"d1d2Z#d3d4Z$d5d6Z%d7d8Z&Gd9d:d:e'Z(Gd;d<dZ+e,d?krze+dS)Ha  The ``distro`` package (``distro`` stands for Linux Distribution) provides information about the Linux distribution it runs on, such as a reliable machine-readable distro ID, or version information. It is the recommended replacement for Python's original :py:func:`platform.linux_distribution` function, but it provides much more functionality. An alternative implementation became necessary because Python 3.5 deprecated this function, and Python 3.8 will remove it altogether. Its predecessor function :py:func:`platform.dist` was already deprecated since Python 2.6 and will also be removed in Python 3.8. Still, there are many cases in which access to OS distribution information is needed. See `Python issue 1322 `_ for more information. NZ UNIXCONFDIRz/etcz os-releaseZolZoracleZrhel)ZenterpriseenterpriseZredhatenterpriseworkstationZredhatenterpriseserverZredhatzA(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)z(\w+)[-_](release|version)$Zdebian_versionz lsb-releasez oem-releasezsystem-releaseTcCs t|S)a Return information about the current OS distribution as a tuple ``(id_name, version, codename)`` with items as follows: * ``id_name``: If *full_distribution_name* is false, the result of :func:`distro.id`. Otherwise, the result of :func:`distro.name`. * ``version``: The result of :func:`distro.version`. * ``codename``: The result of :func:`distro.codename`. The interface of this function is compatible with the original :py:func:`platform.linux_distribution` function, supporting a subset of its parameters. The data it returns may not exactly be the same, because it uses more data sources than the original function, and that may lead to different data if the OS distribution is not consistent across multiple data sources it provides (there are indeed such distributions ...). Another reason for differences is the fact that the :func:`distro.id` method normalizes the distro ID string to a reliable machine-readable value for a number of popular OS distributions. )_distrolinux_distribution)full_distribution_namer6/usr/lib/python3.8/site-packages/pip/_vendor/distro.pyrasrcCstS)a$ Return the distro ID of the current distribution, as a machine-readable string. For a number of OS distributions, the returned distro ID value is *reliable*, in the sense that it is documented and that it does not change across releases of the distribution. This package maintains the following reliable distro ID values: ============== ========================================= Distro ID Distribution ============== ========================================= "ubuntu" Ubuntu "debian" Debian "rhel" RedHat Enterprise Linux "centos" CentOS "fedora" Fedora "sles" SUSE Linux Enterprise Server "opensuse" openSUSE "amazon" Amazon Linux "arch" Arch Linux "cloudlinux" CloudLinux OS "exherbo" Exherbo Linux "gentoo" GenToo Linux "ibm_powerkvm" IBM PowerKVM "kvmibm" KVM for IBM z Systems "linuxmint" Linux Mint "mageia" Mageia "mandriva" Mandriva Linux "parallels" Parallels "pidora" Pidora "raspbian" Raspbian "oracle" Oracle Linux (and Oracle Enterprise Linux) "scientific" Scientific Linux "slackware" Slackware "xenserver" XenServer "openbsd" OpenBSD "netbsd" NetBSD "freebsd" FreeBSD ============== ========================================= If you have a need to get distros for reliable IDs added into this set, or if you find that the :func:`distro.id` function returns a different distro ID for one of the listed distros, please create an issue in the `distro issue tracker`_. **Lookup hierarchy and transformations:** First, the ID is obtained from the following sources, in the specified order. The first available and non-empty value is used: * the value of the "ID" attribute of the os-release file, * the value of the "Distributor ID" attribute returned by the lsb_release command, * the first part of the file name of the distro release file, The so determined ID value then passes the following transformations, before it is returned by this method: * it is translated to lower case, * blanks (which should not be there anyway) are translated to underscores, * a normalization of the ID is performed, based upon `normalization tables`_. The purpose of this normalization is to ensure that the ID is as reliable as possible, even across incompatible changes in the OS distributions. A common reason for an incompatible change is the addition of an os-release file, or the addition of the lsb_release command, with ID values that differ from what was previously determined from the distro release file name. )ridrrrrr}sKrFcCs t|S)ak Return the name of the current OS distribution, as a human-readable string. If *pretty* is false, the name is returned without version or codename. (e.g. "CentOS Linux") If *pretty* is true, the version and codename are appended. (e.g. "CentOS Linux 7.1.1503 (Core)") **Lookup hierarchy:** The name is obtained from the following sources, in the specified order. The first available and non-empty value is used: * If *pretty* is false: - the value of the "NAME" attribute of the os-release file, - the value of the "Distributor ID" attribute returned by the lsb_release command, - the value of the "" field of the distro release file. * If *pretty* is true: - the value of the "PRETTY_NAME" attribute of the os-release file, - the value of the "Description" attribute returned by the lsb_release command, - the value of the "" field of the distro release file, appended with the value of the pretty version ("" and "" fields) of the distro release file, if available. )rnameprettyrrrrs$rcCs t||S)av Return the version of the current OS distribution, as a human-readable string. If *pretty* is false, the version is returned without codename (e.g. "7.0"). If *pretty* is true, the codename in parenthesis is appended, if the codename is non-empty (e.g. "7.0 (Maipo)"). Some distributions provide version numbers with different precisions in the different sources of distribution information. Examining the different sources in a fixed priority order does not always yield the most precise version (e.g. for Debian 8.2, or CentOS 7.1). The *best* parameter can be used to control the approach for the returned version: If *best* is false, the first non-empty version number in priority order of the examined sources is returned. If *best* is true, the most precise version number out of all examined sources is returned. **Lookup hierarchy:** In all cases, the version number is obtained from the following sources. If *best* is false, this order represents the priority order: * the value of the "VERSION_ID" attribute of the os-release file, * the value of the "Release" attribute returned by the lsb_release command, * the version number parsed from the "" field of the first line of the distro release file, * the version number parsed from the "PRETTY_NAME" attribute of the os-release file, if it follows the format of the distro release files. * the version number parsed from the "Description" attribute returned by the lsb_release command, if it follows the format of the distro release files. )rversionr bestrrrr s)r cCs t|S)a Return the version of the current OS distribution as a tuple ``(major, minor, build_number)`` with items as follows: * ``major``: The result of :func:`distro.major_version`. * ``minor``: The result of :func:`distro.minor_version`. * ``build_number``: The result of :func:`distro.build_number`. For a description of the *best* parameter, see the :func:`distro.version` method. )r version_partsr rrrrsrcCs t|S)a5 Return the major version of the current OS distribution, as a string, if provided. Otherwise, the empty string is returned. The major version is the first part of the dot-separated version string. For a description of the *best* parameter, see the :func:`distro.version` method. )r major_versionrrrrr/s rcCs t|S)a6 Return the minor version of the current OS distribution, as a string, if provided. Otherwise, the empty string is returned. The minor version is the second part of the dot-separated version string. For a description of the *best* parameter, see the :func:`distro.version` method. )r minor_versionrrrrr<s rcCs t|S)a3 Return the build number of the current OS distribution, as a string, if provided. Otherwise, the empty string is returned. The build number is the third part of the dot-separated version string. For a description of the *best* parameter, see the :func:`distro.version` method. )r build_numberrrrrrIs rcCstS)a Return a space-separated list of distro IDs of distributions that are closely related to the current OS distribution in regards to packaging and programming interfaces, for example distributions the current distribution is a derivative from. **Lookup hierarchy:** This information item is only provided by the os-release file. For details, see the description of the "ID_LIKE" attribute in the `os-release man page `_. )rlikerrrrrVsrcCstS)a Return the codename for the release of the current OS distribution, as a string. If the distribution does not have a codename, an empty string is returned. Note that the returned codename is not always really a codename. For example, openSUSE returns "x86_64". This function does not handle such cases in any special way and just returns the string it finds, if any. **Lookup hierarchy:** * the codename within the "VERSION" attribute of the os-release file, if provided, * the value of the "Codename" attribute returned by the lsb_release command, * the value of the "" field of the distro release file. )rcodenamerrrrrgsrcCs t||S)a Return certain machine-readable information items about the current OS distribution in a dictionary, as shown in the following example: .. sourcecode:: python { 'id': 'rhel', 'version': '7.0', 'version_parts': { 'major': '7', 'minor': '0', 'build_number': '' }, 'like': 'fedora', 'codename': 'Maipo' } The dictionary structure and keys are always the same, regardless of which information items are available in the underlying data sources. The values for the various keys are as follows: * ``id``: The result of :func:`distro.id`. * ``version``: The result of :func:`distro.version`. * ``version_parts -> major``: The result of :func:`distro.major_version`. * ``version_parts -> minor``: The result of :func:`distro.minor_version`. * ``version_parts -> build_number``: The result of :func:`distro.build_number`. * ``like``: The result of :func:`distro.like`. * ``codename``: The result of :func:`distro.codename`. For a description of the *pretty* and *best* parameters, see the :func:`distro.version` method. )rinfor rrrrs)rcCstS)z Return a dictionary containing key-value pairs for the information items from the os-release file data source of the current OS distribution. See `os-release file`_ for details about these information items. )ros_release_inforrrrrsrcCstS)z Return a dictionary containing key-value pairs for the information items from the lsb_release command data source of the current OS distribution. See `lsb_release command output`_ for details about these information items. )rlsb_release_inforrrrrsrcCstS)z Return a dictionary containing key-value pairs for the information items from the distro release file data source of the current OS distribution. See `distro release file`_ for details about these information items. )rdistro_release_inforrrrrsrcCstS)z Return a dictionary containing key-value pairs for the information items from the distro release file data source of the current OS distribution. )r uname_inforrrrrsrcCs t|S)a Return a single named information item from the os-release file data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. See `os-release file`_ for details about these information items. )ros_release_attr attributerrrrsrcCs t|S)a Return a single named information item from the lsb_release command output data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. See `lsb_release command output`_ for details about these information items. )rlsb_release_attrrrrrrsrcCs t|S)a Return a single named information item from the distro release file data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. See `distro release file`_ for details about these information items. )rdistro_release_attrrrrrrsrcCs t|S)aZ Return a single named information item from the distro release file data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. )r uname_attrrrrrr src@s eZdZdZddZddZdS)cached_propertyzA version of @property which caches the value. On access, it calls the underlying function and sets the value in `__dict__` so future accesses will not re-call the property. cCs|j|_||_dS)N)__name___fname_f)selffrrr__init__"szcached_property.__init__cCs2|dk std|j||}|j|j<|S)Nzcall {} on an instance)AssertionErrorformatr"r#__dict__)r$objownerretrrr__get__&szcached_property.__get__N)r! __module__ __qualname____doc__r&r-rrrrr sr c@s2eZdZdZdBddZddZdCdd Zd d ZdDd dZdEddZ dFddZ dGddZ dHddZ dIddZ ddZddZdJddZdd Zd!d"Zd#d$Zd%d&Zd'd(Zd)d*Zd+d,Zd-d.Zed/d0Zed1d2Zed3d4Zed5d6Zed7d8Zed9d:Z ed;d<Z!d=d>Z"ed?d@Z#dAS)KLinuxDistributiona  Provides information about a OS distribution. This package creates a private module-global instance of this class with default initialization arguments, that is used by the `consolidated accessor functions`_ and `single source accessor functions`_. By using default initialization arguments, that module-global instance returns data about the current OS distribution (i.e. the distro this package runs on). Normally, it is not necessary to create additional instances of this class. However, in situations where control is needed over the exact data sources that are used, instances of this class can be created with a specific distro release file, or a specific os-release file, or without invoking the lsb_release command. TcCs.|ptjtt|_|pd|_||_||_dS)ao The initialization method of this class gathers information from the available data sources, and stores that in private instance attributes. Subsequent access to the information items uses these private instance attributes, so that the data sources are read only once. Parameters: * ``include_lsb`` (bool): Controls whether the `lsb_release command output`_ is included as a data source. If the lsb_release command is not available in the program execution path, the data source for the lsb_release command will be empty. * ``os_release_file`` (string): The path name of the `os-release file`_ that is to be used as a data source. An empty string (the default) will cause the default path name to be used (see `os-release file`_ for details). If the specified or defaulted os-release file does not exist, the data source for the os-release file will be empty. * ``distro_release_file`` (string): The path name of the `distro release file`_ that is to be used as a data source. An empty string (the default) will cause a default search algorithm to be used (see `distro release file`_ for details). If the specified distro release file does not exist, or if no default distro release file can be found, the data source for the distro release file will be empty. * ``include_name`` (bool): Controls whether uname command output is included as a data source. If the uname command is not available in the program execution path the data source for the uname command will be empty. Public instance attributes: * ``os_release_file`` (string): The path name of the `os-release file`_ that is actually used as a data source. The empty string if no distro release file is used as a data source. * ``distro_release_file`` (string): The path name of the `distro release file`_ that is actually used as a data source. The empty string if no distro release file is used as a data source. * ``include_lsb`` (bool): The result of the ``include_lsb`` parameter. This controls whether the lsb information will be loaded. * ``include_uname`` (bool): The result of the ``include_uname`` parameter. This controls whether the uname information will be loaded. Raises: * :py:exc:`IOError`: Some I/O issue with an os-release file or distro release file. * :py:exc:`subprocess.CalledProcessError`: The lsb_release command had some issue (other than not being available in the program execution path). * :py:exc:`UnicodeError`: A data source has unexpected characters or uses an unexpected encoding. r2N) ospathjoin _UNIXCONFDIR_OS_RELEASE_BASENAMEos_release_filedistro_release_file include_lsb include_uname)r$r:r8r9r;rrrr&>s H  zLinuxDistribution.__init__cCs dj|dS)z Return repr of all info aeLinuxDistribution(os_release_file={self.os_release_file!r}, distro_release_file={self.distro_release_file!r}, include_lsb={self.include_lsb!r}, include_uname={self.include_uname!r}, _os_release_info={self._os_release_info!r}, _lsb_release_info={self._lsb_release_info!r}, _distro_release_info={self._distro_release_info!r}, _uname_info={self._uname_info!r})r$)r(r<rrr__repr__s zLinuxDistribution.__repr__cCs"|r |n|||fS)z Return information about the OS distribution that is compatible with Python's :func:`platform.linux_distribution`, supporting a subset of its parameters. For details, see :func:`distro.linux_distribution`. )rrr r)r$rrrrrs z$LinuxDistribution.linux_distributioncCsldd}|d}|r ||tS|d}|r8||tS|d}|rP||tS|d}|rh||tSdS)zoReturn the distro ID of the OS distribution, as a string. For details, see :func:`distro.id`. cSs|dd}|||S)N _)lowerreplaceget) distro_idtablerrr normalizesz'LinuxDistribution.id..normalizerdistributor_idr2)rNORMALIZED_OS_IDrNORMALIZED_LSB_IDrNORMALIZED_DISTRO_IDr)r$rErCrrrrs        zLinuxDistribution.idFcCs||dp&|dp&|dp&|d}|rt|dp>|d}|st|dpV|d}|jdd}|rt|d|}|pzdS) zu Return the name of the OS distribution, as a string. For details, see :func:`distro.name`. rrF pretty_name descriptionTr r>r2)rrrrr )r$r rr rrrrs&     zLinuxDistribution.namec Cs|d|d|d||ddd||ddd|dg}d}|r|D]$}|d|dks||dkr\|}q\n|D]}|dkr|}qq|r|r|rd||}|S)z{ Return the version of the OS distribution, as a string. For details, see :func:`distro.version`. version_idreleaserJr2rK.z {0} ({1})) rrr_parse_distro_release_contentrBrcountrr()r$r r Zversionsr vrrrr s: zLinuxDistribution.versioncCsL|j|d}|rHtd}||}|rH|\}}}||p>d|pDdfSdS)z Return the version of the OS distribution, as a tuple of version numbers. For details, see :func:`distro.version_parts`. rz(\d+)\.?(\d+)?\.?(\d+)?r2)r2r2r2)r recompilematchgroups)r$r Z version_strZ version_regexmatchesmajorminorrrrrrs   zLinuxDistribution.version_partscCs||dS)z Return the major version number of the current distribution. For details, see :func:`distro.major_version`. rrr$r rrrr szLinuxDistribution.major_versioncCs||dS)z Return the minor version number of the current distribution. For details, see :func:`distro.minor_version`. rYrZrrrrszLinuxDistribution.minor_versioncCs||dS)z} Return the build number of the current distribution. For details, see :func:`distro.build_number`. rYrZrrrrszLinuxDistribution.build_numbercCs|dp dS)z Return the IDs of distributions that are like the OS distribution. For details, see :func:`distro.like`. Zid_liker2)rr<rrrr#szLinuxDistribution.likecCs>z |jdWStk r8|dp2|dp2dYSXdS)zp Return the codename of the OS distribution. For details, see :func:`distro.codename`. rr2N)_os_release_infoKeyErrorrrr<rrrr+s  zLinuxDistribution.codenamec CsBt||||t||||||d||dS)z Return certain machine-readable information about the OS distribution. For details, see :func:`distro.info`. )rWrXr)rr rrr)dictrr rrrrr)r$r r rrrr:s zLinuxDistribution.infocCs|jS)z Return a dictionary containing key-value pairs for the information items from the os-release file data source of the OS distribution. For details, see :func:`distro.os_release_info`. )r]r<rrrrMsz!LinuxDistribution.os_release_infocCs|jS)z Return a dictionary containing key-value pairs for the information items from the lsb_release command data source of the OS distribution. For details, see :func:`distro.lsb_release_info`. )_lsb_release_infor<rrrrVsz"LinuxDistribution.lsb_release_infocCs|jS)z Return a dictionary containing key-value pairs for the information items from the distro release file data source of the OS distribution. For details, see :func:`distro.distro_release_info`. )_distro_release_infor<rrrr`sz%LinuxDistribution.distro_release_infocCs|jS)z Return a dictionary containing key-value pairs for the information items from the uname command data source of the OS distribution. For details, see :func:`distro.uname_info`. ) _uname_infor<rrrrjszLinuxDistribution.uname_infocCs|j|dS)z Return a single named information item from the os-release file data source of the OS distribution. For details, see :func:`distro.os_release_attr`. r2)r]rBr$rrrrrssz!LinuxDistribution.os_release_attrcCs|j|dS)z Return a single named information item from the lsb_release command output data source of the OS distribution. For details, see :func:`distro.lsb_release_attr`. r2)r`rBrcrrrr|sz"LinuxDistribution.lsb_release_attrcCs|j|dS)z Return a single named information item from the distro release file data source of the OS distribution. For details, see :func:`distro.distro_release_attr`. r2)rarBrcrrrrsz%LinuxDistribution.distro_release_attrcCs|j|dS)z Return a single named information item from the uname command output data source of the OS distribution. For details, see :func:`distro.uname_release_attr`. r2)rbrBrcrrrrszLinuxDistribution.uname_attrc Cs:tj|jr6t|j}||W5QRSQRXiS)z Get the information items from the specified os-release file. Returns: A dictionary containing all information items. )r3r4isfiler8open_parse_os_release_content)r$Z release_filerrrr]s z"LinuxDistribution._os_release_infocCsi}tj|dd}d|_tjddkr@t|jtr@|jd|_t|}|D]>}d|krL| dd\}}t|tr||d}||| <qLqLd |kr|d |d <nbd |kr|d |d <nLd |krt d |d }|r| }|d}|d}|}||d <|S)aD Parse the lines of an os-release file. Parameters: * lines: Iterable through the lines in the os-release file. Each line must be a unicode string or a UTF-8 encoded byte string. Returns: A dictionary containing all information items. T)posixrr\z iso-8859-1=r[utf-8Zversion_codenamerZubuntu_codenamer z(\(\D+\))|,(\s+)?\D+z(),)shlexZwhitespace_splitsys version_info isinstanceZ wordcharsbytesdecodelistsplitr@rRsearchgroupstrip)linespropsZlexertokenstokenkrQrrrrrfs4      z+LinuxDistribution._parse_os_release_contentc Csz|js iSttjdB}zd}tj||d}Wn$tk rRiYW5QRSXW5QRX|t  }| |S)z Get the information items from the lsb_release command output. Returns: A dictionary containing all information items. w)Z lsb_releasez-astderr) r:rer3devnull subprocess check_outputOSErrorrprlgetfilesystemencoding splitlines_parse_lsb_release_contentr$r~cmdstdoutZcontentrrrr`s z#LinuxDistribution._lsb_release_infocCsXi}|D]J}|ddd}t|dkr,q|\}}||dd|iq|S)aM Parse the output of the lsb_release command. Parameters: * lines: Iterable through the lines of the lsb_release output. Each line must be a unicode string or a UTF-8 encoded byte string. Returns: A dictionary containing all information items.  :r[r\r>r?)rurrlenupdaterAr@)rvrwlineZkvrzrQrrrrs  z,LinuxDistribution._parse_lsb_release_contentc CspttjdB}zd}tj||d}Wn$tk rHiYW5QRSXW5QRX|t }| |S)Nr{)unamez-rsr|) rer3r~rrrrprlrr_parse_uname_contentrrrrrbs zLinuxDistribution._uname_infocCsTi}td|d}|rP|\}}|dkr4iS||d<||d<||d<|S)Nz^([^\s]+)\s+([\d\.]+)rZLinuxrrrM)rRrsrurUr@)rvrwrTrr rrrrs  z&LinuxDistribution._parse_uname_contentcCs0|jrb||j}tj|j}t|}d|krLd|dkrLd|d<n|r^|d|d<|Szt t }| Wn6t k rddddd d d d d ddddddg}YnX|D]r}|t krqt|}|rtjt |}||}d|kr||_|d|d<d|dkrd|d<|SqiSdS)z Get the information items from the specified distro release file. Returns: A dictionary containing all information items. rZ cloudlinuxrr[z SuSE-releasez arch-releasez base-releasezcentos-releasezfedora-releasezgentoo-releasezmageia-releasezmandrake-releasezmandriva-releasezmandrivalinux-releasezmanjaro-releasezoracle-releasezredhat-releasez sl-releasezslackware-versionN)r9_parse_distro_release_filer3r4basename _DISTRO_RELEASE_BASENAME_PATTERNrTr@rtlistdirr6sortr _DISTRO_RELEASE_IGNORE_BASENAMESr5)r$ distro_inforrTZ basenamesfilepathrrrra.s^        z&LinuxDistribution._distro_release_infoc CsRz0t|}||W5QRWSQRXWnttfk rLiYSXdS)z Parse a distro release file. Parameters: * filepath: Path name of the distro release file. Returns: A dictionary containing all information items. N)rerOreadlinerIOError)r$rfprrrrrs  &z,LinuxDistribution._parse_distro_release_filecCst|tr|d}t|ddd}i}|r|dddd|d<|drn|dddd|d<|dr|dddd|d <n|r||d<|S) a Parse a line from a distro release file. Parameters: * line: Line from the distro release file. Must be a unicode string or a UTF-8 encoded byte string. Returns: A dictionary containing all information items. riNrr\rLr[r)rnrorp(_DISTRO_RELEASE_CONTENT_REVERSED_PATTERNrTrurt)rrVrrrrrOs     z/LinuxDistribution._parse_distro_release_contentN)Tr2r2T)T)F)FF)F)F)F)F)FF)$r!r.r/r0r&r=rrrr rrrrrrrrrrrrrrrr r] staticmethodrfr`rrbrrarrOrrrrr1,sV N   "               B     Cr1cCstt}|tj|ttjt j dd}|j ddddd| }|j rn|t jtdd d n:|d td d td d }|d |t}|d|dS)NzOS distro info tool)rKz--jsonz-jz!Output in machine readable format store_true)helpactionT)indentZ sort_keyszName: %sr z Version: %sz Codename: %s)loggingZ getLoggerr!ZsetLevelDEBUGZ addHandlerZ StreamHandlerrlrargparseArgumentParser add_argument parse_argsjsonrdumpsrr r)ZloggerparserargsZdistribution_versionZdistribution_codenamerrrmains$     r__main__)T)F)FF)F)F)F)F)FF)-r0r3rRrlrrkrrrenvironrBr6r7rGrHrIrSrrrrrrr rrrrrrrrrrrrrrrobjectr r1rrr!rrrrsz   N ' ,   ,   } site-packages/pip/_vendor/__pycache__/__init__.cpython-38.opt-1.pyc000064400000005415151464452710021064 0ustar00U .e1@sdZddlmZddlZddlZddlZdZej ej e Z ddZ ereeje dejejdd<e de d e d e d e d e d e de de de de de de de de de de de de de de de de de de d e d!e d"e d#e d$e d%e d&e d'e d(e d)e d*e d+e d,e d-e d.e d/e d0e d1e d2e d3e d4e d5e d6e d7dS)8z pip._vendor is for vendoring dependencies of pip to prevent needing pip to depend on something external. Files inside of pip._vendor should be considered immutable and should only be updated to versions from upstream. )absolute_importNFcCsvdt|}zt|ttddWntk r8Yn:Xtj|tj|<|dd\}}t tj||tj|dS)Nz{0}.{1}r)level.) format__name__ __import__globalslocals ImportErrorsysmodulesrsplitsetattr)Z modulenameZ vendored_namebaseheadr8/usr/lib/python3.8/site-packages/pip/_vendor/__init__.pyvendoreds  rz*.whlZ cachecontrolZcoloramaZ contextlib2ZdistlibZdistroZhtml5libZsixz six.moveszsix.moves.urllibzsix.moves.urllib.parseZ packagingzpackaging.versionzpackaging.specifiersZpep517Z pkg_resourcesZprogressZpytomlZretryingZrequestszrequests.exceptionszrequests.packageszrequests.packages.urllib3z&requests.packages.urllib3._collectionsz$requests.packages.urllib3.connectionz(requests.packages.urllib3.connectionpoolz!requests.packages.urllib3.contribz*requests.packages.urllib3.contrib.ntlmpoolz+requests.packages.urllib3.contrib.pyopensslz$requests.packages.urllib3.exceptionsz requests.packages.urllib3.fieldsz"requests.packages.urllib3.filepostz"requests.packages.urllib3.packagesz/requests.packages.urllib3.packages.ordered_dictz&requests.packages.urllib3.packages.sixz5requests.packages.urllib3.packages.ssl_match_hostnamezErequests.packages.urllib3.packages.ssl_match_hostname._implementationz%requests.packages.urllib3.poolmanagerz!requests.packages.urllib3.requestz"requests.packages.urllib3.responsezrequests.packages.urllib3.utilz)requests.packages.urllib3.util.connectionz&requests.packages.urllib3.util.requestz'requests.packages.urllib3.util.responsez$requests.packages.urllib3.util.retryz#requests.packages.urllib3.util.ssl_z&requests.packages.urllib3.util.timeoutz"requests.packages.urllib3.util.urlZurllib3)__doc__Z __future__rZglobZos.pathosr Z DEBUNDLEDpathabspathdirname__file__Z WHEEL_DIRrjoinrrrrsr $site-packages/pip/_vendor/__pycache__/six.cpython-38.pyc000064400000063434151464452710017176 0ustar00U .e~C@sdZddlmZddlZddlZddlZddlZddlZdZdZ ej ddkZ ej ddkZ ej dddkZ e refZefZefZeZeZejZn~efZeefZeejfZeZeZejd red ZnHGd d d eZ ze!e Wne"k red ZYn Xed Z[ ddZ#ddZ$GdddeZ%Gddde%Z&Gdddej'Z(Gddde%Z)GdddeZ*e*e+Z,Gddde(Z-e)dddd e)d!d"d#d$d!e)d%d"d"d&d%e)d'd(d#d)d'e)d*d(d+e)d,d"d#d-d,e)d.d/d/d0d.e)d1d/d/d.d1e)d2d3d4e)d5d(d#d6d5e)d7d(e rd8nd9d:e)d;d(d<e)d=d>d?d@e)d d de)dAdAdBe)dCdCdBe)dDdDdBe)d6d(d#d6d5e)dEd"d#dFdEe)dGd"d"dHdGe&d#d(e&dIdJe&dKdLe&dMdNdOe&dPdQdPe&dRdSdTe&dUdVdWe&dXdYdZe&d[d\d]e&d^d_d`e&dadbdce&dddedfe&dgdhdie&djdkdle&dmdndoe&dpdpdqe&drdrdqe&dsdsdqe&dtdtdue&dvdwe&dxdye&dzd{e&d|d}d|e&d~de&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&ddde&de+dde&de+dde&de+de+de&ddde&ddde&dddg@Z.ejdkrfe.e&ddg7Z.e.D]2Z/e0e-e/j1e/e2e/e&rje,3e/de/j1qj[/e.e-_.e-e+dZ4e,3e4dGddde(Z5e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d@dde)ddde)ddde)ddde)ddddd΃e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)dddgZ6e6D]Z/e0e5e/j1e/q[/e6e5_.e,3e5e+dddڡGdd܄de(Z7e)ddde)ddde)dddgZ8e8D]Z/e0e7e/j1e/q>[/e8e7_.e,3e7e+dddGddde(Z9e)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)ddde)d ddg#Z:e:D]Z/e0e9e/j1e/q[/e:e9_.e,3e9e+d d d Gd dde(Z;e)ddde)ddde)ddde)dddgZe>D]Z/e0e=e/j1e/q[/e>e=_.e,3e=e+dddGdddej'Z?e,3e?e+ddd d!Z@d"d#ZAe rd$ZBd%ZCd&ZDd'ZEd(ZFd)ZGn$d*ZBd+ZCd,ZDd-ZEd.ZFd/ZGzeHZIWn"eJk rd0d1ZIYnXeIZHzeKZKWn"eJk rd2d3ZKYnXe rFd4d5ZLejMZNd6d7ZOeZPn>d8d5ZLd9d:ZNd;d7ZOGd<d=d=eZPeKZKe#eLd>eQeBZReQeCZSeQeDZTeQeEZUeQeFZVeQeGZWe r(d?d@ZXdAdBZYdCdDZZdEdFZ[e\dGZ]e\dHZ^e\dIZ_nTdJd@ZXdKdBZYdLdDZZdMdFZ[e\dNZ]e\dOZ^e\dPZ_e#eXdQe#eYdRe#eZdSe#e[dTe rDdUdVZ`dWdXZaebZcddldZdededYjfZg[dehdZiejjZkelZmddlnZnenjoZoenjpZpdZZqej d[d[k r6d\Zrd]Zsn d^Zrd_Zsnjd`dVZ`dadXZaecZcebZgdbdcZidddeZketejuevZmddloZoeojoZoZpdfZqd\Zrd]Zse#e`dge#eadhdidZZwdjd^Zxdkd_Zye reze4j{dlZ|ddmdnZ}nddodpZ|e|dqej dddrk rLe|dsn.ej dddrk rne|dtn dudvZ~eze4j{dwdZedk rdxdyZej dddzk reZd{dyZe#e}d|ej dddk rejejfd}d~ZnejZddZddZdddZdddZdddZddZgZe+Zeddk rzge_ejreejD]4\ZZeej+dkrej1e+kreje=qq[[eje,dS(z6Utilities for writing code that runs on Python 2 and 3)absolute_importNz'Benjamin Peterson z1.12.0)rjavaic@seZdZddZdS)XcCsdS)Nlselfrr3/usr/lib/python3.8/site-packages/pip/_vendor/six.py__len__>sz X.__len__N)__name__ __module__ __qualname__r rrrr r<srlcCs ||_dS)z Add documentation to a function.N)__doc__)funcdocrrr _add_docKsrcCst|tj|S)z7Import module, returning the module after the last dot.) __import__sysmodulesnamerrr _import_modulePsrc@seZdZddZddZdS) _LazyDescrcCs ||_dSNrr rrrr __init__Xsz_LazyDescr.__init__cCsB|}t||j|zt|j|jWntk r<YnX|Sr)_resolvesetattrrdelattr __class__AttributeError)r objtpresultrrr __get__[sz_LazyDescr.__get__N)r rrrr&rrrr rVsrcs.eZdZdfdd ZddZddZZS) MovedModuleNcs2tt||tr(|dkr |}||_n||_dSr)superr'rPY3mod)r roldnewr!rr ris zMovedModule.__init__cCs t|jSr)rr*r rrr rrszMovedModule._resolvecCs"|}t||}t||||Sr)rgetattrr)r attr_modulevaluerrr __getattr__us  zMovedModule.__getattr__)N)r rrrrr2 __classcell__rrr-r r'gs r'cs(eZdZfddZddZgZZS) _LazyModulecstt|||jj|_dSr)r(r4rr!rrr-rr r~sz_LazyModule.__init__cCs ddg}|dd|jD7}|S)Nrr cSsg|] }|jqSrr).0r/rrr sz'_LazyModule.__dir__..)_moved_attributes)r Zattrsrrr __dir__sz_LazyModule.__dir__)r rrrr8r7r3rrr-r r4|s r4cs&eZdZdfdd ZddZZS)MovedAttributeNcsdtt||trH|dkr |}||_|dkr@|dkr<|}n|}||_n||_|dkrZ|}||_dSr)r(r9rr)r*r/)r rZold_modZnew_modZold_attrZnew_attrr-rr rszMovedAttribute.__init__cCst|j}t||jSr)rr*r.r/)r modulerrr rs zMovedAttribute._resolve)NN)r rrrrr3rrr-r r9sr9c@sVeZdZdZddZddZddZdd d Zd d Zd dZ ddZ ddZ e Z dS)_SixMetaPathImporterz A meta path importer to import six.moves and its submodules. This class implements a PEP302 finder and loader. It should be compatible with Python 2.5 and all existing versions of Python3 cCs||_i|_dSr)r known_modules)r Zsix_module_namerrr rsz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdSN.r<r)r r*Z fullnamesfullnamerrr _add_modulesz _SixMetaPathImporter._add_modulecCs|j|jd|Sr=r?r r@rrr _get_modulesz _SixMetaPathImporter._get_moduleNcCs||jkr|SdSr)r<)r r@pathrrr find_modules z _SixMetaPathImporter.find_modulecCs2z |j|WStk r,td|YnXdS)Nz!This loader does not know module )r<KeyError ImportErrorrBrrr Z __get_modules z!_SixMetaPathImporter.__get_modulecCsTz tj|WStk r YnX||}t|tr@|}n||_|tj|<|Sr)rrrF _SixMetaPathImporter__get_module isinstancer'r __loader__)r r@r*rrr load_modules     z _SixMetaPathImporter.load_modulecCst||dS)z Return true, if the named module is a package. We need this method to get correct spec objects with Python 3.4 (see PEP451) __path__)hasattrrHrBrrr is_packagesz_SixMetaPathImporter.is_packagecCs||dS)z;Return None Required, if is_package is implementedN)rHrBrrr get_codes z_SixMetaPathImporter.get_code)N) r rrrrrArCrErHrKrNrO get_sourcerrrr r;s  r;c@seZdZdZgZdS) _MovedItemszLazy loading of moved objectsN)r rrrrLrrrr rQsrQZ cStringIOioStringIOfilter itertoolsbuiltinsZifilter filterfalseZ ifilterfalseinputZ __builtin__Z raw_inputinternrmapimapgetcwdosZgetcwdugetcwdbZ getoutputZcommands subprocessrangeZxrangeZ reload_module importlibZimpreloadreduce functoolsZ shlex_quoteZpipesZshlexZquoteUserDict collectionsUserList UserStringzipZizip zip_longestZ izip_longestZ configparserZ ConfigParsercopyregZcopy_regZdbm_gnuZgdbmzdbm.gnuZ _dummy_threadZ dummy_threadZhttp_cookiejarZ cookielibzhttp.cookiejarZ http_cookiesZCookiez http.cookiesZ html_entitiesZhtmlentitydefsz html.entitiesZ html_parserZ HTMLParserz html.parserZ http_clientZhttplibz http.clientZemail_mime_basezemail.MIMEBasezemail.mime.baseZemail_mime_imagezemail.MIMEImagezemail.mime.imageZemail_mime_multipartzemail.MIMEMultipartzemail.mime.multipartZemail_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipartZemail_mime_textzemail.MIMETextzemail.mime.textZBaseHTTPServerz http.serverZ CGIHTTPServerZSimpleHTTPServerZcPicklepickleZqueueZQueuereprlibreprZ socketserverZ SocketServer_threadthreadZtkinterZTkinterZtkinter_dialogZDialogztkinter.dialogZtkinter_filedialogZ FileDialogztkinter.filedialogZtkinter_scrolledtextZ ScrolledTextztkinter.scrolledtextZtkinter_simpledialogZ SimpleDialogztkinter.simpledialogZ tkinter_tixZTixz tkinter.tixZ tkinter_ttkZttkz tkinter.ttkZtkinter_constantsZ Tkconstantsztkinter.constantsZ tkinter_dndZTkdndz tkinter.dndZtkinter_colorchooserZtkColorChooserztkinter.colorchooserZtkinter_commondialogZtkCommonDialogztkinter.commondialogZtkinter_tkfiledialogZ tkFileDialogZ tkinter_fontZtkFontz tkinter.fontZtkinter_messageboxZ tkMessageBoxztkinter.messageboxZtkinter_tksimpledialogZtkSimpleDialogZ urllib_parsez.moves.urllib_parsez urllib.parseZ urllib_errorz.moves.urllib_errorz urllib.errorZurllibz .moves.urllibZurllib_robotparser robotparserzurllib.robotparserZ xmlrpc_clientZ xmlrpclibz xmlrpc.clientZ xmlrpc_serverZSimpleXMLRPCServerz xmlrpc.serverZwin32winreg_winregzmoves.z.movesmovesc@seZdZdZdS)Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseNr rrrrrrr ruBsruZ ParseResultZurlparseZ SplitResultZparse_qsZ parse_qslZ urldefragZurljoinZurlsplitZ urlunparseZ urlunsplitZ quote_plusZunquoteZ unquote_plusZunquote_to_bytesZ urlencodeZ splitqueryZsplittagZ splituserZ splitvalueZ uses_fragmentZ uses_netlocZ uses_paramsZ uses_queryZ uses_relativemoves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNrvrrrr rxlsrxZURLErrorZurllib2Z HTTPErrorZContentTooShortErrorz.moves.urllib.errormoves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNrvrrrr rzsrzZurlopenzurllib.requestZinstall_openerZ build_openerZ pathname2urlZ url2pathnameZ getproxiesZRequestZOpenerDirectorZHTTPDefaultErrorHandlerZHTTPRedirectHandlerZHTTPCookieProcessorZ ProxyHandlerZ BaseHandlerZHTTPPasswordMgrZHTTPPasswordMgrWithDefaultRealmZAbstractBasicAuthHandlerZHTTPBasicAuthHandlerZProxyBasicAuthHandlerZAbstractDigestAuthHandlerZHTTPDigestAuthHandlerZProxyDigestAuthHandlerZ HTTPHandlerZ HTTPSHandlerZ FileHandlerZ FTPHandlerZCacheFTPHandlerZUnknownHandlerZHTTPErrorProcessorZ urlretrieveZ urlcleanupZ URLopenerZFancyURLopenerZ proxy_bypassZparse_http_listZparse_keqv_listz.moves.urllib.requestmoves.urllib_requestzmoves.urllib.requestc@seZdZdZdS) Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNrvrrrr r|sr|Zaddbasezurllib.responseZ addclosehookZaddinfoZ addinfourlz.moves.urllib.responsemoves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNrvrrrr r~sr~ZRobotFileParserz.moves.urllib.robotparsermoves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZedZedZedZ edZ edZ ddZ d S) Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacerwryr{r}rcCsdddddgS)Nparseerrorrequestresponserqrr rrr r8szModule_six_moves_urllib.__dir__N) r rrrrL _importerrCrrrrrqr8rrrr rs     rz moves.urllibcCstt|j|dS)zAdd an item to six.moves.N)rrQr)Zmoverrr add_movesrc CsXztt|WnDtk rRz tj|=Wn"tk rLtd|fYnXYnXdS)zRemove item from six.moves.zno such move, %rN)r rQr"rt__dict__rFrrrr remove_moves r__func____self__ __closure____code__ __defaults__ __globals__im_funcZim_selfZ func_closureZ func_codeZ func_defaultsZ func_globalscCs|Sr)next)itrrr advance_iteratorsrcCstddt|jDS)Ncss|]}d|jkVqdS)__call__N)r)r5klassrrr szcallable..)anytype__mro__)r#rrr callablesrcCs|SrrZunboundrrr get_unbound_functionsrcCs|Srrrclsrrr create_unbound_method#srcCs|jSr)rrrrr r(scCst|||jSr)types MethodTyper!)rr#rrr create_bound_method+srcCst|d|Sr)rrrrrr r.sc@seZdZddZdS)IteratorcCst||Sr)r__next__r rrr r3sz Iterator.nextN)r rrrrrrr r1srz3Get the function out of a possibly unbound functioncKst|jf|Sr)iterkeysdkwrrr iterkeysDsrcKst|jf|Sr)rvaluesrrrr itervaluesGsrcKst|jf|Sr)ritemsrrrr iteritemsJsrcKst|jf|Sr)rZlistsrrrr iterlistsMsrrrrcKs |jf|Sr)rrrrr rVscKs |jf|Sr)rrrrr rYscKs |jf|Sr)rrrrr r\scKs |jf|Sr)rrrrr r_sviewkeys viewvalues viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs |dS)Nzlatin-1)encodesrrr bqsrcCs|Srrrrrr utsrz>BassertCountEqualZassertRaisesRegexpZassertRegexpMatchesassertRaisesRegex assertRegexcCs|Srrrrrr rscCst|dddS)Nz\\z\\\\Zunicode_escape)unicodereplacerrrr rscCs t|dS)Nrord)Zbsrrr byte2intsrcCs t||Srr)Zbufirrr indexbytessrZassertItemsEqualz Byte literalz Text literalcOst|t||Sr)r._assertCountEqualr argskwargsrrr rscOst|t||Sr)r._assertRaisesRegexrrrr rscOst|t||Sr)r. _assertRegexrrrr rsexeccCs:z*|dkr|}|j|k r$|||W5d}d}XdSr) __traceback__with_traceback)r$r1tbrrr reraises  rcCsB|dkr*td}|j}|dkr&|j}~n |dkr6|}tddS)zExecute code in a namespace.Nrzexec _code_ in _globs_, _locs_)r _getframe f_globalsf_localsr)Z_code_Z_globs_Z_locs_framerrr exec_s rzedef reraise(tp, value, tb=None): try: raise tp, value, tb finally: tb = None )rrzdef raise_from(value, from_value): try: if from_value is None: raise value raise value from from_value finally: value = None zrdef raise_from(value, from_value): try: raise value from from_value finally: value = None cCs|dSrr)r1Z from_valuerrr raise_fromsrprintc s.|dtjdkrdSfdd}d}|dd}|dk r`t|trNd}nt|ts`td|d d}|dk rt|trd}nt|tstd |rtd |s|D]}t|trd}qq|rtd }td }nd }d }|dkr|}|dkr|}t|D] \} }| r||||q||dS)z4The new-style print function for Python 2.4 and 2.5.fileNcsdt|tst|}ttrVt|trVjdk rVtdd}|dkrHd}|j|}|dS)Nerrorsstrict) rI basestringstrrrencodingr.rwrite)datarfprr rs   zprint_..writeFsepTzsep must be None or a stringendzend must be None or a stringz$invalid keyword arguments to print()  )poprstdoutrIrr TypeError enumerate) rrrZ want_unicoderrargnewlineZspacerrrr print_sL          r)rrcOs<|dtj}|dd}t|||r8|dk r8|dS)NrflushF)getrrr_printr)rrrrrrr rs    zReraise an exception.csfdd}|S)Ncst|}|_|Sr)rdwraps __wrapped__)fassignedupdatedwrappedrr wrapper*szwraps..wrapperr)rrrrrrr r(srcs&Gfdddt}t|ddiS)z%Create a base class with a metaclass.cs,eZdZfddZefddZdS)z!with_metaclass..metaclasscs ||Srr)rr this_basesrbasesmetarr __new__:sz)with_metaclass..metaclass.__new__cs |Sr) __prepare__)rrrrrr r=sz-with_metaclass..metaclass.__prepare__N)r rrr classmethodrrrrr metaclass8srZtemporary_classr)rr)rrrrrr with_metaclass3srcsfdd}|S)z6Class decorator for creating a class with a metaclass.cs||j}|d}|dk r@t|tr,|g}|D]}||q0|dd|ddt|drl|j|d<|j|j |S)N __slots__r __weakref__r) rcopyrrIrrrMrr __bases__)rZ orig_varsslotsZ slots_varrrr rEs        zadd_metaclass..wrapperr)rrrrr add_metaclassCs rutf-8rcCs8t|tr|||St|tr$|Stdt|dS)zCoerce **s** to six.binary_type. For Python 2: - `unicode` -> encoded to `str` - `str` -> `str` For Python 3: - `str` -> encoded to `bytes` - `bytes` -> `bytes` not expecting type '%s'N)rI text_typer binary_typerrrrrrrr ensure_binaryUs   r cCsXt|ttfstdt|tr:t|tr:|||}ntrTt|trT|||}|S)zCoerce *s* to `str`. For Python 2: - `unicode` -> encoded to `str` - `str` -> `str` For Python 3: - `str` -> `str` - `bytes` -> decoded to `str` r) rIrrrrPY2rr)decoderrrr ensure_strhs  r cCs8t|tr|||St|tr$|Stdt|dS)zCoerce *s* to six.text_type. For Python 2: - `unicode` -> `unicode` - `str` -> `unicode` For Python 3: - `str` -> `str` - `bytes` -> decoded to `str` rN)rIrr rrrrrrr ensure_text|s   r cCs2tr.d|jkrtd|j|j|_dd|_|S)a A decorator that defines __unicode__ and __str__ methods under Python 2. Under Python 3 it does nothing. To support Python 2 and 3 with a single code base, define a __str__ method returning text and apply this decorator to the class. __str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|dS)Nr) __unicode__rr rrr z-python_2_unicode_compatible..)r r ValueErrorr rr)rrrr python_2_unicode_compatibles  r__spec__)N)NN)rr)rr)rr)rZ __future__rrdrUoperatorrr __author__ __version__ version_infor r)ZPY34rZ string_typesintZ integer_typesrZ class_typesrbytesrmaxsizeZMAXSIZErZlongZ ClassTyperplatform startswithobjectrlen OverflowErrorrrrr' ModuleTyper4r9r;r rrQr7r/rrrIrArtruZ_urllib_parse_moved_attributesrxZ_urllib_error_moved_attributesrzZ _urllib_request_moved_attributesr|Z!_urllib_response_moved_attributesr~Z$_urllib_robotparser_moved_attributesrrrZ _meth_funcZ _meth_selfZ _func_closureZ _func_codeZ_func_defaultsZ _func_globalsrr NameErrorrrrrrr attrgetterZget_method_functionZget_method_selfZget_function_closureZget_function_codeZget_function_defaultsZget_function_globalsrrrr methodcallerrrrrrchrZunichrstructStructpackZint2byte itemgetterrgetitemrrZ iterbytesrRrSBytesIOrrrpartialr[rrrrr.rVrrrrrWRAPPER_ASSIGNMENTSWRAPPER_UPDATESrrrr r r rrL __package__globalsrrsubmodule_search_locations meta_pathrrZimporterappendrrrr s     >                                        F                                                                  %                                                  5      site-packages/pip/_vendor/__pycache__/contextlib2.cpython-38.pyc000064400000036202151464452710020621 0ustar00U .eB @sdZddlZddlZddlZddlmZddlmZddddd d d d d g Zedg7Zej dddkrnej Z ne de fddiZ ddZddZGddde ZGdd d e ZGdddeZddZGddde ZGddde ZGd d d eZGd!d d eZGd"d d e Zej dd#kZer@d$d%Zd&d'Znd(d%Zd)d'Zzdd*lmZWnek rzeZ Yn Xd+d,Z Gd-d d e Z!Gd.dde!Z"Gd/ddeZ#dS)0zAcontextlib2 - backports and enhancements to the contextlib moduleN)dequewrapscontextmanagerclosing nullcontextAbstractContextManagerContextDecorator ExitStackredirect_stdoutredirect_stderrsuppress ContextStack)ABC __slots__cCs0||kr dS|||jD]}t||q|SN)append __bases__ _classic_mro)CresultBrr;/usr/lib/python3.8/site-packages/pip/_vendor/contextlib2.pyrs    rcGsvz |j}Wn"tk r,tt|g}YnX|D]>}|D],}||jkr:|j|dkrbtSq2q:tSq2dS)NT)__mro__AttributeErrortupler__dict__NotImplemented)rmethodsmromethodrrrr_check_methods$s    r%c@s2eZdZdZddZejddZeddZ dS) rz,An abstract base class for context managers.cCs|S)z0Return `self` upon entering the runtime context.rselfrrr __enter__8sz AbstractContextManager.__enter__cCsdS)z9Raise any exception triggered within the runtime context.Nr)r'exc_type exc_value tracebackrrr__exit__<szAbstractContextManager.__exit__cCs|tkrt|ddStS)z.innerr)r'r?r@rr>r__call__iszContextDecorator.__call__N)r/r0r1r2r:r9rArrrrr Is c@s0eZdZdZddZddZddZdd Zd S) _GeneratorContextManagerz%Helper for @contextmanager decorator.cCsJ||||_||||_|_|_t|dd}|dkr@t|j}||_dS)Nr2)genr?r<r=getattrtyper2)r'r?r<r=docrrr__init__ts    z!_GeneratorContextManager.__init__cCs||j|j|jSr) __class__r?r<r=r&rrrr9sz%_GeneratorContextManager._recreate_cmcCs.z t|jWStk r(tdYnXdS)Nzgenerator didn't yield)nextrC StopIteration RuntimeErrorr&rrrr(s z"_GeneratorContextManager.__enter__c Cs|dkr8zt|jWntk r,YdSXtdn|dkrF|}z|j|||tdWntk r}z||k WYSd}~XYnjtk r}z0||krWY"dStr|j|krWY dSW5d}~XYntd|k rYnXdS)Nzgenerator didn't stopz#generator didn't stop after throw()F) rIrCrJrKthrow_HAVE_EXCEPTION_CHAINING __cause__sysexc_info)r'rEvaluer+excrrrr,s,    z!_GeneratorContextManager.__exit__N)r/r0r1r2rGr9r(r,rrrrrBqs rBcstfdd}|S)a@contextmanager decorator. Typical usage: @contextmanager def some_generator(): try: yield finally: This makes this: with some_generator() as : equivalent to this: try: = finally: cs t||Sr)rBr;r?rrhelperszcontextmanager..helperr)r?rUrrTrrsc@s(eZdZdZddZddZddZdS) ra2Context to automatically close something at the end of a block. Code like this: with closing(.open()) as f: is equivalent to this: f = .open() try: finally: f.close() cCs ||_dSrthing)r'rWrrrrGszclosing.__init__cCs|jSrrVr&rrrr(szclosing.__enter__cGs|jdSr)rWclose)r'rQrrrr,szclosing.__exit__Nr/r0r1r2rGr(r,rrrrrsc@s(eZdZdZddZddZddZdS)_RedirectStreamNcCs||_g|_dSr) _new_target _old_targets)r' new_targetrrrrGsz_RedirectStream.__init__cCs*|jtt|jtt|j|j|jSr)r\rrDrP_streamsetattrr[r&rrrr(sz_RedirectStream.__enter__cCstt|j|jdSr)r_rPr^r\popr'exctypeexcinstexctbrrrr,sz_RedirectStream.__exit__)r/r0r1r^rGr(r,rrrrrZsrZc@seZdZdZdZdS)r aAContext manager for temporarily redirecting stdout to another file. # How to send help() to stderr with redirect_stdout(sys.stderr): help(dir) # How to write help() to a file with open('help.txt', 'w') as f: with redirect_stdout(f): help(pow) stdoutNr/r0r1r2r^rrrrr s c@seZdZdZdZdS)r zCContext manager for temporarily redirecting stderr to another file.stderrNrfrrrrr sc@s(eZdZdZddZddZddZdS) r a?Context manager to suppress specified exceptions After the exception is suppressed, execution proceeds with the next statement following the with statement. with suppress(FileNotFoundError): os.remove(somefile) # Execution still resumes here if the file was already removed cGs ||_dSr) _exceptions)r' exceptionsrrrrG)szsuppress.__init__cCsdSrrr&rrrr(,szsuppress.__enter__cCs|dk ot||jSr) issubclassrhrarrrr,/s zsuppress.__exit__NrYrrrrr s rcsfdd}|S)Ncs4|j}||krdS|dks*|kr$q*|}q||_dSr) __context__)new_excold_exc exc_context frame_excrr_fix_exception_context@sz3_make_context_fixer.._fix_exception_contextr)rprqrror_make_context_fixer?s rrcCs<z|dj}|dWn tk r6||d_YnXdS)NrL)rk BaseException) exc_details fixed_ctxrrr_reraise_with_existing_contextOs    rvcCsddS)NcSsdSrr)rlrmrrr[z%_make_context_fixer..rrorrrrrZscCs|\}}}tddS)Nz!raise exc_type, exc_value, exc_tb)exec)rtr)r*Zexc_tbrrrrv_s ) InstanceTypecCst|}|tkr|jS|Sr)rErzrH)objZobj_typerrr _get_typeksr|c@sXeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ ddZ dS)r aContext manager for dynamic management of a stack of exit callbacks For example: with ExitStack() as stack: files = [stack.enter_context(open(fname)) for fname in filenames] # All opened files will automatically be closed at the end of # the with statement, even if attempts to open files later # in the list raise an exception cCs t|_dSr)r_exit_callbacksr&rrrrGszExitStack.__init__cCst|}|j|_t|_|S)z?Preserve the context stack by transferring it to a new instance)rEr}r)r' new_stackrrrpop_alls zExitStack.pop_allcs"fdd}|_||dS)z:Helper to correctly register callbacks to __exit__ methodscsf|Srr)rtcmcm_exitrr _exit_wrappersz.ExitStack._push_cm_exit.._exit_wrapperN)__self__push)r'rrrrrr _push_cm_exitszExitStack._push_cm_exitcCsDt|}z |j}Wn tk r2|j|YnX||||S)aRegisters a callback with the standard __exit__ method signature Can suppress exceptions the same way __exit__ methods can. Also accepts any object with an __exit__ method (registering a call to the method instead of the object itself) )r|r,rr}rr)r'exit_cb_type exit_methodrrrrs   zExitStack.pushcs$fdd}|_||S)z\Registers an arbitrary callback and arguments. Cannot suppress exceptions. csdSrr)r)rStbr<callbackr=rrrsz)ExitStack.callback.._exit_wrapper) __wrapped__r)r'rr<r=rrrrrs zExitStack.callbackcCs(t|}|j}||}||||S)zEnters the supplied context manager If successful, also pushes its __exit__ method as a callback and returns the result of the __enter__ method. )r|r,r(r)r'r_cm_type_exitrrrr enter_contexts   zExitStack.enter_contextcCs|ddddS)z$Immediately unwind the context stackN)r,r&rrrrXszExitStack.closecCs|Srrr&rrrr(szExitStack.__enter__c Gs|ddk }td}t|}d}d}|jr|j}z||rNd}d}d}Wq(t}||d|dd}|}Yq(Xq(|rt||o|S)NrrLFT)NNN)rPrQrrr}r`rv) r'rt received_excrprqsuppressed_exc pending_raisecbnew_exc_detailsrrrr,s(    zExitStack.__exit__N) r/r0r1r2rGrrrrrrXr(r,rrrrr ss   cs8eZdZdZfddZddZddZdd ZZS) rz+Backwards compatibility alias for ExitStackcstdttt|dS)Nz*ContextStack has been renamed to ExitStack)r6r7r8superrrGr&rHrrrGszContextStack.__init__cCs ||Sr)r)r'rrrr register_exitszContextStack.register_exitcOs|j|f||Sr)r)r'rr<r=rrrregisterszContextStack.registercCs|Sr)rr&rrrpreserveszContextStack.preserve) r/r0r1r2rGrrr __classcell__rrrrrs  c@s*eZdZdZd ddZddZddZdS) raMContext manager that does no additional processing. Used as a stand-in for a normal context manager, when a particular block of code is only sometimes used with a normal context manager: cm = optional_cm if condition else nullcontext() with cm: # Perform operation, using optional_cm if condition is True NcCs ||_dSr enter_result)r'rrrrrGsznullcontext.__init__cCs|jSrrr&rrrr(sznullcontext.__enter__cGsdSrr)r'excinforrrr,sznullcontext.__exit__)NrYrrrrrs )$r2r3rPr6 collectionsr functoolsr__all__ version_inforZ_abc_ABCABCMetaobjectrr%rr rBrrrZr r r rNrrrvtypesrz ImportErrorrEr|r rrrrrrsT    (H"  qsite-packages/pip/_vendor/pkg_resources/py31compat.py000064400000001062151464452710016745 0ustar00import os import errno import sys from pip._vendor import six def _makedirs_31(path, exist_ok=False): try: os.makedirs(path) except OSError as exc: if not exist_ok or exc.errno != errno.EEXIST: raise # rely on compatibility behavior until mode considerations # and exists_ok considerations are disentangled. # See https://github.com/pypa/setuptools/pull/1083#issuecomment-315168663 needs_makedirs = ( six.PY2 or (3, 4) <= sys.version_info < (3, 4, 1) ) makedirs = _makedirs_31 if needs_makedirs else os.makedirs site-packages/pip/_vendor/pkg_resources/__pycache__/py31compat.cpython-38.pyc000064400000001144151464452710023234 0ustar00U .e2@s`ddlZddlZddlZddlmZdddZejpLdejkoHdknZerVenej Z dS) N)sixFc CsHzt|Wn4tk rB}z|r0|jtjkr2W5d}~XYnXdS)N)osmakedirsOSErrorerrnoZEEXIST)pathexist_okexcr H/usr/lib/python3.8/site-packages/pip/_vendor/pkg_resources/py31compat.py _makedirs_31s r ))r r)F) rrsysZ pip._vendorrr ZPY2 version_infoZneeds_makedirsrr r r r s  site-packages/pip/_vendor/pkg_resources/__pycache__/__init__.cpython-38.pyc000064400000303710151464452710022777 0ustar00U .eͦG@sdZddlmZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl Z ddl Z ddl Z ddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlmZz ddlZWnek rddlZYnXze Wne!k r*e"Z YnXddl#m$Z$ddl%m&Z&m'Z'm(Z(ddlm)Z)zddlm*Z*m+Z+m,Z,d Z-Wnek rd Z-YnXdd lm.Z/dd l0m1Z1m2Z2zddl3m4Z5e5j6Wnek rdZ5YnXd dl7m8Z8ddl#m9Z9ddl#m:Z:e;de;de;de;dekr@dkrLnne?de$j@r\dZAdZBdZCdZDdZEdZFdZGdZHdZIdZJdZKdZLdZMdZNdZOdZPdZQdZRdZSGdddeTZUddZViZWddZXddZYd d!ZZd"d#Z[d$d%Z\d&d'Z]d(d)Z^d*d+Z_Z`d,d-Zad.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMddNddOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrgGZbGdsdIdIecZdGdtdJdJedZeGdudvdveeZfGdwdKdKedZgGdxdLdLedZhiZiejjddyZkdyZldzZmd ZndZod{Zpd|dmZqd}d0Zrgfd~dZsddZtddZuevdZwevdZxeuZyddRZzdd/Z{e{Z|dd1Z}dd2Z~ddd3Zdd4ZGdd`d`ZGddadaeZGddDdDZGdddeZGddCdCZeZGddMdMe?ZGddEdEZddBZddOZddPZddUZddVZddWZdddXZGddgdgZeqeeGddhdheZGddidieZeGddedeeZeZGdddeZGdddeZGddjdjeZeqe jeGddbdbeZGddcdceZGdddddeZeXdidddkZddd?ZdddZee jedddZeeeddZdddZddZGdddZddZddZddZddZeejeee5dree5jeeXdideXdidddlZddĄZddƄZdd<ZdddnZddʄZeejeee jeee5dr,ee5jedd̄ZeeeddZZddτZifddфZddӄZddՄZddׄZddSZevd١jZevdejejBjZGddHdHZdd݄Zdd߄ZGddFdFZGdddeăZGdddeăZeeedZddZGdddeɃZddNZGddGdGe:jj̓ZddZddZddYZddZddTZddZe jdeUd dddZeeփfddZeddZGddpdpeكZdS(aZ Package resource API -------------------- A resource is a logical file contained within a package, or a logical subdirectory thereof. The package resource API expects resource names to have their path parts separated with ``/``, *not* whatever the local path separator is. Do not use os.path operations to manipulate resource names being passed into the API. The package resource API is designed to work with normal filesystem packages, .egg files, and unpacked .egg files. It can also work in a limited way with .zip files and with custom PEP 302 loaders that support the ``get_data()`` method. )absolute_importN) get_importer)six)urllibmapfilter)utime)mkdirrenameunlinkTF)open)isdirsplit) py31compat)appdirs) packagingzpip._vendor.packaging.versionz pip._vendor.packaging.specifiersz"pip._vendor.packaging.requirementszpip._vendor.packaging.markers)r)rzPython 3.4 or later is requiredc@seZdZdZdS) PEP440Warningza Used when there is an issue with a version or specifier not complying with PEP 440. N__name__ __module__ __qualname____doc__rrF/usr/lib/python3.8/site-packages/pip/_vendor/pkg_resources/__init__.pyrxsrcCs8ztj|WStjjk r2tj|YSXdSN)rversionVersionInvalidVersion LegacyVersion)vrrr parse_versionsr#cKs"t|tt||dSr)globalsupdate _state_varsdictfromkeys)Zvartypekwrrr_declare_states r*cCs8i}t}tD] \}}|d|||||<q|S)NZ_sget_)r$r&itemsstategkr"rrr __getstate__s r0cCs8t}|D]$\}}|dt|||||q|S)NZ_sset_)r$r+r&r,rrr __setstate__sr1cCs|Sr)copyvalrrr _sget_dictsr5cCs|||dSr)clearr%keyobr-rrr _sset_dictsr:cCs|Sr)r0r3rrr _sget_objectsr;cCs||dSr)r1r7rrr _sset_objectsr<cGsdSrrargsrrrr?cCsbt}t|}|dk r^tjdkr^z&ddtdd|df}Wntk r\YnX|S)aZReturn this platform's maximum compatible version. distutils.util.get_platform() normally reports the minimum version of Mac OS X that would be required to *use* extensions produced by distutils. But what we want when checking compatibility is to know the version of Mac OS X that we are *running*. To allow usage of packages that explicitly require a newer version of Mac OS X, we must also know the current version of the OS. If this condition occurs for any other platform with a version in its platform strings, this function should be extended accordingly. Ndarwinz macosx-%s-%s.r) get_build_platformmacosVersionStringmatchsysplatformjoin _macosx_versgroup ValueError)platmrrrget_supported_platforms  &rOrequire run_script get_providerget_distributionload_entry_point get_entry_mapget_entry_infoiter_entry_pointsresource_stringresource_streamresource_filenameresource_listdirresource_existsresource_isdirdeclare_namespace working_setadd_activation_listenerfind_distributionsset_extraction_pathcleanup_resourcesget_default_cache Environment WorkingSetResourceManager Distribution Requirement EntryPointResolutionErrorVersionConflictDistributionNotFound UnknownExtraExtractionErrorparse_requirements safe_name safe_version get_platformcompatible_platforms yield_linessplit_sections safe_extra to_filenameinvalid_markerevaluate_markerensure_directorynormalize_pathEGG_DIST BINARY_DIST SOURCE_DIST CHECKOUT_DIST DEVELOP_DISTIMetadataProviderIResourceProvider FileMetadata PathMetadata EggMetadata EmptyProviderempty_provider NullProvider EggProviderDefaultProvider ZipProviderregister_finderregister_namespace_handlerregister_loader_typefixup_namespace_packagesrPkgResourcesDeprecationWarningrun_mainAvailableDistributionsc@seZdZdZddZdS)rkz.Abstract base for dependency resolution errorscCs|jjt|jSr) __class__rreprr>selfrrr__repr__szResolutionError.__repr__N)rrrrrrrrrrksc@s<eZdZdZdZeddZeddZddZd d Z d S) rlz An already-installed version conflicts with the requested version. Should be initialized with the installed Distribution and the requested Requirement. z3{self.dist} is installed but {self.req} is requiredcCs |jdSNrr=rrrrdist szVersionConflict.distcCs |jdSNrr=rrrrreqszVersionConflict.reqcCs|jjftSr _templateformatlocalsrrrrreportszVersionConflict.reportcCs|s|S|j|f}t|S)zt If required_by is non-empty, return a version of self that is a ContextualVersionConflict. )r>ContextualVersionConflict)r required_byr>rrr with_contexts zVersionConflict.with_contextN) rrrrrpropertyrrrrrrrrrls  c@s&eZdZdZejdZeddZdS)rz A VersionConflict that accepts a third parameter, the set of the requirements that required the installed Distribution. z by {self.required_by}cCs |jdS)NrCr=rrrrr*sz%ContextualVersionConflict.required_byN)rrrrrlrrrrrrrr"s rc@sHeZdZdZdZeddZeddZeddZd d Z d d Z d S)rmz&A requested distribution was not foundzSThe '{self.req}' distribution was not found and is required by {self.requirers_str}cCs |jdSrr=rrrrr5szDistributionNotFound.reqcCs |jdSrr=rrrr requirers9szDistributionNotFound.requirerscCs|js dSd|jS)Nzthe applicationz, )rrIrrrr requirers_str=sz"DistributionNotFound.requirers_strcCs|jjftSrrrrrrrCszDistributionNotFound.reportcCs|Sr)rrrrr__str__FszDistributionNotFound.__str__N) rrrrrrrrrrrrrrrrm/s   c@seZdZdZdS)rnz>Distribution doesn't have an "extra feature" of the given nameNrrrrrrnJsrrCcCs |t|<dS)aRegister `provider_factory` to make providers for `loader_type` `loader_type` is the type or class of a PEP 302 ``module.__loader__``, and `provider_factory` is a function that, passed a *module* object, returns an ``IResourceProvider`` for that module. N)_provider_factories)Z loader_typeZprovider_factoryrrrrXscCstt|tr$t|p"tt|dSztj|}Wn&tk rXt |tj|}YnXt |dd}t t ||S)z?Return an IResourceProvider for the named module or requirementr __loader__N) isinstancerir_findrPstrrGmodulesKeyError __import__getattr _find_adapterr)Z moduleOrReqmoduleloaderrrrrRbs  cCsd|s\td}|dkrLd}tj|rLttdrLt|}d|krL|d}|| d|dS)Nrz0/System/Library/CoreServices/SystemVersion.plist readPlistZProductVersionrB) rHZmac_verospathexistshasattrplistlibrappendr)_cacherZplistZ plist_contentrrrrJos    rJcCsddd||S)NZppc)ZPowerPCZPower_Macintosh)get)machinerrr _macosx_archsrcCs~ddlm}|}tjdkrz|dszz>t}tddd}dt |dt |d t |fWSt k rxYnX|S) zReturn this platform's string for platform-specific distributions XXX Currently this is the same as ``distutils.util.get_platform()``, but it needs some hacks for Linux and Mac OS X. r)rsrAzmacosx-r _zmacosx-%d.%d-%sr) Z sysconfigrsrGrH startswithrJrunamereplaceintrrL)rsrMrrrrrrDs  rDzmacosx-(\d+)\.(\d+)-(.*)zdarwin-(\d+)\.(\d+)\.(\d+)-(.*)cCs|dks|dks||krdSt|}|rt|}|st|}|rt|d}d|d|df}|dkr||dks|dkr|d krdSd S|d|dks|d |d krd St|dt|dkrd SdSd S) zCan code for the `provided` platform run on the `required` platform? Returns true if either platform is ``None``, or the platforms are equal. XXX Needs compatibility checks for Linux and other unixy OSes. NTrz%s.%srCz10.3z10.4Fr)rErFdarwinVersionStringrrK)ZprovidedZrequiredZreqMacZprovMacZ provDarwinZdversionZ macosversionrrrrts2   cCs<tdj}|d}|||d<t|d||dS)z@Locate distribution `dist_spec` and run its `script_name` scriptrrrNrG _getframe f_globalsr6rPrQ)Z dist_spec script_namensnamerrrrQs  cCs@t|tjrt|}t|tr(t|}t|tsIs the named metadata a directory? (like ``os.path.isdir()``)Nrrrrrmetadata_isdirsz IMetadataProvider.metadata_isdircCsdS)z?List of metadata names in the directory (like ``os.listdir()``)Nrrrrrmetadata_listdirsz"IMetadataProvider.metadata_listdircCsdS)z=Execute the named script in the supplied namespace dictionaryNr)r namespacerrrrQ szIMetadataProvider.run_scriptN) rrrrrrrrrQrrrrrs c@s@eZdZdZddZddZddZdd Zd d Zd d Z dS)rz3An object that provides access to package resourcescCsdS)zdReturn a true filesystem path for `resource_name` `manager` must be an ``IResourceManager``Nrmanager resource_namerrrget_resource_filenamesz'IResourceProvider.get_resource_filenamecCsdS)ziReturn a readable file-like object for `resource_name` `manager` must be an ``IResourceManager``Nrrrrrget_resource_streamsz%IResourceProvider.get_resource_streamcCsdS)zmReturn a string containing the contents of `resource_name` `manager` must be an ``IResourceManager``Nrrrrrget_resource_stringsz%IResourceProvider.get_resource_stringcCsdS)z,Does the package contain the named resource?Nrrrrr has_resourceszIResourceProvider.has_resourcecCsdS)z>Is the named resource a directory? (like ``os.path.isdir()``)Nrrrrrr]"sz IResourceProvider.resource_isdircCsdS)z?List of resource names in the directory (like ``os.listdir()``)Nrrrrrr[%sz"IResourceProvider.resource_listdirN) rrrrrrrrr]r[rrrrr sc@seZdZdZd'ddZeddZeddZd d Zd d Z d dZ d(ddZ ddZ ddZ d)ddZd*ddZd+ddZddZd,dd Zd!d"Zd#d$Zd%d&ZdS)-rfzDA collection of active distributions on sys.path (or a similar list)NcCs>g|_i|_i|_g|_|dkr&tj}|D]}||q*dS)z?Create working set from list of path entries (default=sys.path)N)entries entry_keysby_key callbacksrGr add_entry)rrentryrrr__init__,szWorkingSet.__init__cCsb|}zddlm}Wntk r.|YSXz||Wntk r\||YSX|S)z1 Prepare the master working set. r) __requires__)__main__r ImportErrorrPrl_build_from_requirements)clswsrrrr _build_master9s zWorkingSet._build_mastercCsf|g}t|}||t}|D]}||q"tjD]}||jkr8||q8|jtjdd<|S)zQ Build a working set from a requirement spec. Rewrites sys.path. N)rpresolvereaddrGrrr)rZreq_specrreqsdistsrrrrrrMs    z#WorkingSet._build_from_requirementscCs<|j|g|j|t|dD]}|||dq$dS)aAdd a path item to ``.entries``, finding any distributions on it ``find_distributions(entry, True)`` is used to find distributions corresponding to the path entry, and they are added. `entry` is always appended to ``.entries``, even if it is already present. (This is because ``sys.path`` can contain the same value more than once, and the ``.entries`` of the ``sys.path`` WorkingSet should always equal ``sys.path``.) TFN)r setdefaultrrrar)rrrrrrrcs  zWorkingSet.add_entrycCs|j|j|kS)z9True if `dist` is the active distribution for its project)rrr8rrrrr __contains__rszWorkingSet.__contains__cCs,|j|j}|dk r(||kr(t|||S)aFind a distribution matching requirement `req` If there is an active distribution for the requested project, this returns it as long as it meets the version requirement specified by `req`. But, if there is an active distribution for the project and it does *not* meet the `req` requirement, ``VersionConflict`` is raised. If there is no active distribution for the requested project, ``None`` is returned. N)rrr8rl)rrrrrrrvs  zWorkingSet.findcsfdd|DS)aYield entry point objects from `group` matching `name` If `name` is None, yields all entry points in `group` from all distributions in the working set, otherwise only ones matching both `group` and `name` are yielded (in distribution order). c3s8|]0}|D]}dks*|jkr|VqqdSr)rUvaluesr).0rrrKrrr s  z/WorkingSet.iter_entry_points..rrrKrrrrrWs zWorkingSet.iter_entry_pointscCs>tdj}|d}|||d<||d||dS)z?Locate distribution for `requires` and run `script_name` scriptrrrNr)rrequiresrrrrrrrQs  zWorkingSet.run_scriptccsLi}|jD]<}||jkrq |j|D] }||kr$d||<|j|Vq$q dS)zYield distributions for non-duplicate projects in the working set The yield order is the order in which the items' path entries were added to the working set. rN)rrr)rseenitemr8rrr__iter__s  zWorkingSet.__iter__TFcCs|r|j|j||d|dkr$|j}|j|g}|j|jg}|sV|j|jkrVdS||j|j<|j|krx||j|j|kr||j||dS)aAdd `dist` to working set, associated with `entry` If `entry` is unspecified, it defaults to the ``.location`` of `dist`. On exit from this routine, `entry` is added to the end of the working set's ``.entries`` (if it wasn't already present). `dist` is only added to the working set if it's for a project that doesn't already have a distribution in the set, unless `replace=True`. If it's added, any callbacks registered with the ``subscribe()`` method will be called. rN) insert_onrlocationrrr8rr _added_new)rrrinsertrkeysZkeys2rrrrs      zWorkingSet.addcCsxt|ddd}i}i}g}t} tt} |rt|d} | |krHq.| | |sVq.|| j} | dkr|j | j} | dks| | kr|r|} |dkr| dkrt |j }nt g}t g} |j | | ||d} || j<| dkr| | d}t| ||| | | kr$| | }t| | || | jddd}|||D] }| || j| j| |<qHd|| <q.|S)aList all distributions needed to (recursively) meet `requirements` `requirements` must be a sequence of ``Requirement`` objects. `env`, if supplied, should be an ``Environment`` instance. If not supplied, it defaults to all distributions available within any entry or distribution in the working set. `installer`, if supplied, will be invoked with each requirement that cannot be met by an already-installed distribution; it should return a ``Distribution`` or ``None``. Unless `replace_conflicting=True`, raises a VersionConflict exception if any requirements are found on the path that have the correct name but the wrong version. Otherwise, if an `installer` is supplied it will be invoked to obtain the correct version of the requirement and activate it. `extras` is a list of the extras to be used with these requirements. This is important because extra requirements may look like `my_req; extra = "my_extra"`, which would otherwise be interpreted as a purely optional requirement. Instead, we want to be able to assert that these requirements are truly required. Nrr)replace_conflictingT)list _ReqExtras collections defaultdictsetpop markers_passrr8rrerrf best_matchrmrrlrrextrasextendr project_name)r requirementsenv installerr rZ processedZbestZ to_activateZ req_extrasrrrrrZ dependent_reqZnew_requirementsZnew_requirementrrrrsT            zWorkingSet.resolvec Cs t|}|i}i}|dkr4t|j}||7}n||}|g} tt| j||D]} || D]} | g} z| | ||} WnBt k r}z$||| <|rWYqfn WYqZW5d}~XYqfXtt| j| | t | qZqfqZt|}|||fS)asFind all activatable distributions in `plugin_env` Example usage:: distributions, errors = working_set.find_plugins( Environment(plugin_dirlist) ) # add plugins+libs to sys.path map(working_set.add, distributions) # display errors print('Could not load', errors) The `plugin_env` should be an ``Environment`` instance that contains only distributions that are in the project's "plugin directory" or directories. The `full_env`, if supplied, should be an ``Environment`` contains all currently-available distributions. If `full_env` is not supplied, one is created automatically from the ``WorkingSet`` this method is called on, which will typically mean that every directory on ``sys.path`` will be scanned for distributions. `installer` is a standard installer callback as used by the ``resolve()`` method. The `fallback` flag indicates whether we should attempt to resolve older versions of a plugin if the newest version cannot be resolved. This method returns a 2-tuple: (`distributions`, `error_info`), where `distributions` is a list of the distributions found in `plugin_env` that were loadable, along with any other distributions that are needed to resolve their dependencies. `error_info` is a dictionary mapping unloadable plugin distributions to an exception instance describing the error that occurred. Usually this will be a ``DistributionNotFound`` or ``VersionConflict`` instance. N) r sortrerrrras_requirementrrkr%r'r()rZ plugin_envZfull_envrZfallbackZplugin_projectsZ error_infoZ distributionsrZ shadow_setrrrZ resolveesr"rrr find_plugins's4$      zWorkingSet.find_pluginscGs&|t|}|D]}||q|S)aEnsure that distributions matching `requirements` are activated `requirements` must be a string or a (possibly-nested) sequence thereof, specifying the distributions and versions required. The return value is a sequence of the distributions that needed to be activated to fulfill the requirements; all relevant distributions are included, even if they were already activated in this working set. )rrpr)rrZneededrrrrrP{s  zWorkingSet.requirecCs8||jkrdS|j||s"dS|D] }||q&dS)zInvoke `callback` for all distributions If `existing=True` (default), call on all existing ones, as well. N)rr)rcallbackexistingrrrr subscribes  zWorkingSet.subscribecCs|jD] }||qdSr)r)rrrrrrrs zWorkingSet._added_newcCs,|jdd|j|j|jddfSr)rrr2rrrrrrr0s   zWorkingSet.__getstate__cCs@|\}}}}|dd|_||_||_|dd|_dSr)rr2rrr)rZe_k_b_crr rrrrrr1s    zWorkingSet.__setstate__)N)N)NTF)NNFN)NNT)T)rrrrr classmethodrrrrrrWrQrrrrrPrrr0r1rrrrrf)s4      ] T c@seZdZdZdddZdS)r z> Map each requirement to the extras that demanded it. Ncs2fdd|d|pdD}j p0t|S)z Evaluate markers for req against each extra that demanded it. Return False if the req has a marker and fails evaluation. Otherwise, return True. c3s|]}jd|iVqdS)extraNmarkerevaluate)rr!rrrrsz*_ReqExtras.markers_pass..rr)rr#any)rrrZ extra_evalsrr%rrs z_ReqExtras.markers_pass)N)rrrrrrrrrr sr c@sxeZdZdZdeefddZddZddZdd d Z d d Z d dZ dddZ dddZ ddZddZddZdS)rez5Searchable snapshot of distributions on a search pathNcCs i|_||_||_||dS)a!Snapshot distributions available on a search path Any distributions found on `search_path` are added to the environment. `search_path` should be a sequence of ``sys.path`` items. If not supplied, ``sys.path`` is used. `platform` is an optional string specifying the name of the platform that platform-specific distributions must be compatible with. If unspecified, it defaults to the current platform. `python` is an optional string naming the desired version of Python (e.g. ``'3.6'``); it defaults to the current version. You may explicitly set `platform` (and/or `python`) to ``None`` if you wish to map *all* distributions, not just those compatible with the running platform or Python version. N)_distmaprHpythonscan)r search_pathrHr(rrrrszEnvironment.__init__cCs2|jdkp|jdkp|j|jk}|o0t|j|jS)zIs distribution `dist` acceptable for this environment? The distribution must match the platform and python version requirements specified when this environment was created, or False is returned. N)r( py_versionrtrH)rrZ py_compatrrrcan_adds   zEnvironment.can_addcCs|j|j|dS)z"Remove `dist` from the environmentN)r'r8removerrrrr-szEnvironment.removecCs4|dkrtj}|D]}t|D]}||qqdS)adScan `search_path` for distributions usable in this environment Any distributions found are added to the environment. `search_path` should be a sequence of ``sys.path`` items. If not supplied, ``sys.path`` is used. Only distributions conforming to the platform/python version defined at initialization are added. N)rGrrar)rr*rrrrrr)s  zEnvironment.scancCs|}|j|gS)aReturn a newest-to-oldest list of distributions for `project_name` Uses case-insensitive `project_name` comparison, assuming all the project's distributions use their project's name converted to all lowercase as their key. )lowerr'r)rrZdistribution_keyrrr __getitem__szEnvironment.__getitem__cCsL||rH|rH|j|jg}||krH|||jtddddS)zLAdd `dist` if we ``can_add()`` it and it has not already been added hashcmpTr8reverseN) r, has_versionr'rr8rroperator attrgetter)rrrrrrrs  zEnvironment.addFcCsfz||}Wntk r,|s$d}YnX|dk r:|S||jD]}||krD|SqD|||S)aFind distribution best matching `req` and usable on `working_set` This calls the ``find(req)`` method of the `working_set` to see if a suitable distribution is already active. (This may raise ``VersionConflict`` if an unsuitable version of the project is already active in the specified `working_set`.) If a suitable distribution isn't active, this method returns the newest distribution in the environment that meets the ``Requirement`` in `req`. If no suitable distribution is found, and `installer` is supplied, then the result of calling the environment's ``obtain(req, installer)`` method will be returned. N)rrlr8obtain)rrr_rr rrrrrs  zEnvironment.best_matchcCs|dk r||SdS)aObtain a distribution matching `requirement` (e.g. via download) Obtain a distro that matches requirement (e.g. via download). In the base ``Environment`` class, this routine just returns ``installer(requirement)``, unless `installer` is None, in which case None is returned instead. This method is a hook that allows subclasses to attempt other ways of obtaining a distribution before falling back to the `installer` argument.Nr)rZ requirementrrrrr6+s zEnvironment.obtainccs"|jD]}||r |Vq dS)z=Yield the unique project names of the available distributionsN)r'r rr8rrrr7szEnvironment.__iter__cCsVt|tr||n)rZ archive_namenamesZ extract_pathZ target_pathrrrget_cache_paths    zResourceManager.get_cache_pathcCsVtjdkr|tjdsdSt|j}|tj@s>|tj@rRd|}t |t dS)aN If the default extraction path is overridden and set to an insecure location, such as /tmp, it opens up an opportunity for an attacker to replace an extracted file with an unauthorized payload. Warn the user if a known insecure location is used. See Distribute #375 for more details. ntZwindirNz%s is writable by group/others and vulnerable to attack when used with get_resource_filename. Consider a more secure location (set with .set_extraction_path or the PYTHON_EGG_CACHE environment variable).) rrrenvironstatst_modeS_IWOTHS_IWGRPwarningswarn UserWarning)rmodemsgrrrrLs  z,ResourceManager._warn_unsafe_extraction_pathcCs.tjdkr*t|jdBd@}t||dS)a4Perform any platform-specific postprocessing of `tempname` This is where Mac header rewrites should be done; other platforms don't have anything special they should do. Resource providers should call this method ONLY after successfully extracting a compressed resource. They must NOT call it on resources that are already in the filesystem. `tempname` is the current (temporary) name of the file, and `filename` is the name it will be renamed to by the caller after this routine returns. posiximiN)rrrQrRchmod)rZtempnamefilenamerXrrr postprocesss zResourceManager.postprocesscCs|jrtd||_dS)aSet the base path where resources will be extracted to, if needed. If you do not call this routine before any extractions take place, the path defaults to the return value of ``get_default_cache()``. (Which is based on the ``PYTHON_EGG_CACHE`` environment variable, with various platform-specific fallbacks. See that routine's documentation for more details.) Resources are extracted to subdirectories of this path based upon information given by the ``IResourceProvider``. You may set this to a temporary directory, but then you must call ``cleanup_resources()`` to delete the extracted files when done. There is no guarantee that ``cleanup_resources()`` will be able to remove all extracted files. (Note: you may not change the extraction path for a given resource manager once resources have been extracted, unless you first call ``cleanup_resources()``.) z5Can't change extraction path, files already extractedN)r>rLrArrrrrrbs z#ResourceManager.set_extraction_pathFcCsdS)aB Delete all extracted resource files and directories, returning a list of the file and directory names that could not be successfully removed. This function does not have any concurrency protection, so it should generally only be called when the extraction path is a temporary directory exclusive to a single process. This method is not automatically called; you must call it explicitly or register it as an ``atexit`` function if you wish to ensure cleanup of a temporary directory used for extractions. Nr)rforcerrrrc sz!ResourceManager.cleanup_resources)r)F)rrrrrArr\r]rZrYrXr[rIrN staticmethodrLr]rbrcrrrrrges   cCstjdptjddS)z Return the ``PYTHON_EGG_CACHE`` environment variable or a platform-relevant user cache dir for an app named "Python-Eggs". ZPYTHON_EGG_CACHEz Python-Eggs)Zappname)rrPrrZuser_cache_dirrrrrrds  cCstdd|S)zConvert an arbitrary string to a standard distribution name Any runs of non-alphanumeric/. characters are replaced with a single '-'. [^A-Za-z0-9.]+-)resubrrrrrq%scCsJzttj|WStjjk rD|dd}tdd|YSXdS)zB Convert an arbitrary string to a standard version string rrBrarbN)rrrrr rrcrd)rrrrrr-s  cCstdd|S)zConvert an arbitrary string to a standard 'extra' name Any runs of non-alphanumeric characters are replaced with a single '_', and the result is always lowercased. z[^A-Za-z0-9.-]+r)rcrdr.)r!rrrrw9scCs |ddS)z|Convert a project or version name to its filename-escaped form Any '-' characters are currently replaced with '_'. rbrrrrrrrxBsc CsHz t|Wn6tk rB}zd|_d|_|WYSd}~XYnXdS)zo Validate text as a PEP 508 environment marker; return an exception if invalid or False otherwise. NF)rz SyntaxErrorr\lineno)texterrrryJs c CsJztj|}|WStjjk rD}z t|W5d}~XYnXdS)z Evaluate a PEP 508 environment marker. Return a boolean indicating the marker result in this environment. Raise SyntaxError if marker is invalid. This implementation uses the 'pyparsing' module. N)rZmarkersZMarkerr$Z InvalidMarkerre)rgr!r#rhrrrrzXs   c@seZdZdZdZdZdZddZddZddZ d d Z d d Z d dZ ddZ ddZddZddZddZddZddZddZdd Zd!d"Zd#d$Zd%d&Zed'd(Zd)d*ZdS)+rzETry to implement resources and metadata for arbitrary PEP 302 loadersNcCs(t|dd|_tjt|dd|_dS)Nr__file__r)rrrrdirname module_pathrrrrrrnszNullProvider.__init__cCs||j|Sr)_fnrkrrrrrrrrsz"NullProvider.get_resource_filenamecCst|||Sr)ioBytesIOrrnrrrrusz NullProvider.get_resource_streamcCs|||j|Sr)_getrmrkrnrrrrxsz NullProvider.get_resource_stringcCs|||j|Sr)_hasrmrkrrrrrr{szNullProvider.has_resourcecCs||j|Sr)rmegg_inforrrrr_get_metadata_path~szNullProvider._get_metadata_pathcCs |js |jS||}||Sr)rtrvrrrrrrrrrs zNullProvider.has_metadatac Cst|js dS||}||}tjr(|Sz |dWStk rn}z|jd||7_W5d}~XYnXdS)Nrutf-8z in {} file at path: {}) rtrvrqrPY2decodeUnicodeDecodeErrorreasonr)rrrvalueexcrrrrs   zNullProvider.get_metadatacCst||SrrurrurrrrszNullProvider.get_metadata_linescCs|||j|Sr)_isdirrmrkrsrrrr]szNullProvider.resource_isdircCs|jo|||j|Sr)rtrrmrurrrrszNullProvider.metadata_isdircCs|||j|Sr)_listdirrmrkrsrrrr[szNullProvider.resource_listdircCs|jr|||j|SgSr)rtrrmrurrrrszNullProvider.metadata_listdirc Csd|}||s$tdjft||dd}|dd}||j|}||d<tj |rt | }t ||d}t|||n>dd lm}t|d|d|f||<t ||d} t| ||dS) Nzscripts/z') # => <script> do_nasty_stuff() </script> # sanitize_html('Click here for $100') # => Click here for $100 def sanitize_token(self, token): # accommodate filters which use token_type differently token_type = token["type"] if token_type in ("StartTag", "EndTag", "EmptyTag"): name = token["name"] namespace = token["namespace"] if ((namespace, name) in self.allowed_elements or (namespace is None and (namespaces["html"], name) in self.allowed_elements)): return self.allowed_token(token) else: return self.disallowed_token(token) elif token_type == "Comment": pass else: return token def allowed_token(self, token): if "data" in token: attrs = token["data"] attr_names = set(attrs.keys()) # Remove forbidden attributes for to_remove in (attr_names - self.allowed_attributes): del token["data"][to_remove] attr_names.remove(to_remove) # Remove attributes with disallowed URL values for attr in (attr_names & self.attr_val_is_uri): assert attr in attrs # I don't have a clue where this regexp comes from or why it matches those # characters, nor why we call unescape. I just know it's always been here. # Should you be worried by this comment in a sanitizer? Yes. On the other hand, all # this will do is remove *more* than it otherwise would. val_unescaped = re.sub("[`\x00-\x20\x7f-\xa0\\s]+", '', unescape(attrs[attr])).lower() # remove replacement characters from unescaped characters val_unescaped = val_unescaped.replace("\ufffd", "") try: uri = urlparse.urlparse(val_unescaped) except ValueError: uri = None del attrs[attr] if uri and uri.scheme: if uri.scheme not in self.allowed_protocols: del attrs[attr] if uri.scheme == 'data': m = data_content_type.match(uri.path) if not m: del attrs[attr] elif m.group('content_type') not in self.allowed_content_types: del attrs[attr] for attr in self.svg_attr_val_allows_ref: if attr in attrs: attrs[attr] = re.sub(r'url\s*\(\s*[^#\s][^)]+?\)', ' ', unescape(attrs[attr])) if (token["name"] in self.svg_allow_local_href and (namespaces['xlink'], 'href') in attrs and re.search(r'^\s*[^#\s].*', attrs[(namespaces['xlink'], 'href')])): del attrs[(namespaces['xlink'], 'href')] if (None, 'style') in attrs: attrs[(None, 'style')] = self.sanitize_css(attrs[(None, 'style')]) token["data"] = attrs return token def disallowed_token(self, token): token_type = token["type"] if token_type == "EndTag": token["data"] = "" % token["name"] elif token["data"]: assert token_type in ("StartTag", "EmptyTag") attrs = [] for (ns, name), v in token["data"].items(): attrs.append(' %s="%s"' % (name if ns is None else "%s:%s" % (prefixes[ns], name), escape(v))) token["data"] = "<%s%s>" % (token["name"], ''.join(attrs)) else: token["data"] = "<%s>" % token["name"] if token.get("selfClosing"): token["data"] = token["data"][:-1] + "/>" token["type"] = "Characters" del token["name"] return token def sanitize_css(self, style): # disallow urls style = re.compile(r'url\s*\(\s*[^\s)]+?\s*\)\s*').sub(' ', style) # gauntlet if not re.match(r"""^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$""", style): return '' if not re.match(r"^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$", style): return '' clean = [] for prop, value in re.findall(r"([-\w]+)\s*:\s*([^:;]*)", style): if not value: continue if prop.lower() in self.allowed_css_properties: clean.append(prop + ': ' + value + ';') elif prop.split('-')[0].lower() in ['background', 'border', 'margin', 'padding']: for keyword in value.split(): if keyword not in self.allowed_css_keywords and \ not re.match(r"^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$", keyword): # noqa break else: clean.append(prop + ': ' + value + ';') elif prop.lower() in self.allowed_svg_properties: clean.append(prop + ': ' + value + ';') return ' '.join(clean) site-packages/pip/_vendor/html5lib/filters/whitespace.py000064400000002276151464452720017427 0ustar00from __future__ import absolute_import, division, unicode_literals import re from . import base from ..constants import rcdataElements, spaceCharacters spaceCharacters = "".join(spaceCharacters) SPACES_REGEX = re.compile("[%s]+" % spaceCharacters) class Filter(base.Filter): """Collapses whitespace except in pre, textarea, and script elements""" spacePreserveElements = frozenset(["pre", "textarea"] + list(rcdataElements)) def __iter__(self): preserve = 0 for token in base.Filter.__iter__(self): type = token["type"] if type == "StartTag" \ and (preserve or token["name"] in self.spacePreserveElements): preserve += 1 elif type == "EndTag" and preserve: preserve -= 1 elif not preserve and type == "SpaceCharacters" and token["data"]: # Test on token["data"] above to not introduce spaces where there were not token["data"] = " " elif not preserve and type == "Characters": token["data"] = collapse_spaces(token["data"]) yield token def collapse_spaces(text): return SPACES_REGEX.sub(' ', text) site-packages/pip/_vendor/html5lib/filters/__pycache__/base.cpython-38.pyc000064400000001502151464452720022462 0ustar00U .e@s(ddlmZmZmZGdddeZdS))absolute_importdivisionunicode_literalsc@s$eZdZddZddZddZdS)FiltercCs ||_dSN)source)selfrr E/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/base.py__init__szFilter.__init__cCs t|jSr)iterr)rr r r __iter__szFilter.__iter__cCs t|j|Sr)getattrr)rnamer r r __getattr__ szFilter.__getattr__N)__name__ __module__ __qualname__r r rr r r r rsrN)Z __future__rrrobjectrr r r r ssite-packages/pip/_vendor/html5lib/filters/__pycache__/optionaltags.cpython-38.pyc000064400000005247151464452720024266 0ustar00U .e\)@s6ddlmZmZmZddlmZGdddejZdS))absolute_importdivisionunicode_literals)basec@s0eZdZdZddZddZddZdd Zd S) Filterz+Removes optional tags from the token streamccsHd}}|jD] }|dk r&|||fV|}|}q|dk rD||dfVdS)N)source)selfZ previous1Z previous2tokenr M/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/optionaltags.pysliders  z Filter.sliderccsp|D]b\}}}|d}|dkrD|ds<||d||sj|Vq|dkrd||d|sj|Vq|VqdS)NtypeStartTagdatanameEndTag)r is_optional_startis_optional_end)r previousr nextrr r r __iter__szFilter.__iter__cCs|r |dpd}|dkr |dkS|dkrJ|dkr4dS|dkr|ddkSn|d krx|dkr^d S|d krr|dd kSdSnb|d kr|dkr|ddkSd Sn@|dkr|d kr|r|ddkr|ddkrd S|ddkSd Sd S)NrhtmlCommentZSpaceCharactersheadrZEmptyTagTrrbodyFr)ZscriptZstylecolgroupcoltbody)r theadtfoottrr )r tagnamerrrr r r rs6    zFilter.is_optional_startcCs|r |dpd}|dkr |dkS|dkrP|dkr<|d|kS|dkpJ|dkSn|dkr|dkrl|ddkS|d kr|dkp|dkSd Snp|d kr|d kr|dd kS|dkp|dkSn@|dkr|dkr|ddkS|dkp|dkSn|dkr |dkr |ddkS|dkp|dkSn|dkrT|dkr8d S|dkrN|ddkSdSn|dkr|dkrt|ddkS|dkr|dkp|dkSd Snf|dkr|dkr|ddkS|dkp|dkSn2|dkr|dkr|ddkS|dkp|dkSd S)Nr)rrrr)Zlioptgroupr#rrr)Zdtddr&Fpr)ZaddressZarticleZasideZ blockquoteZdatagridZdialogdirZdivZdlZfieldsetZfooterZformZh1Zh2Zh3Zh4Zh5Zh6headerZhrZmenuZnavZolr'ZpreZsectiontableZuloption)r+r%)ZrtZrprT)r!r )r r"r r")ZtdZthr )r r$rrr r r rXs\                     zFilter.is_optional_endN)__name__ __module__ __qualname____doc__r rrrr r r r rs   9rN)Z __future__rrrrrr r r r s site-packages/pip/_vendor/html5lib/filters/__pycache__/optionaltags.cpython-38.opt-1.pyc000064400000005247151464452720025225 0ustar00U .e\)@s6ddlmZmZmZddlmZGdddejZdS))absolute_importdivisionunicode_literals)basec@s0eZdZdZddZddZddZdd Zd S) Filterz+Removes optional tags from the token streamccsHd}}|jD] }|dk r&|||fV|}|}q|dk rD||dfVdS)N)source)selfZ previous1Z previous2tokenr M/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/optionaltags.pysliders  z Filter.sliderccsp|D]b\}}}|d}|dkrD|ds<||d||sj|Vq|dkrd||d|sj|Vq|VqdS)NtypeStartTagdatanameEndTag)r is_optional_startis_optional_end)r previousr nextrr r r __iter__szFilter.__iter__cCs|r |dpd}|dkr |dkS|dkrJ|dkr4dS|dkr|ddkSn|d krx|dkr^d S|d krr|dd kSdSnb|d kr|dkr|ddkSd Sn@|dkr|d kr|r|ddkr|ddkrd S|ddkSd Sd S)NrhtmlCommentZSpaceCharactersheadrZEmptyTagTrrbodyFr)ZscriptZstylecolgroupcoltbody)r theadtfoottrr )r tagnamerrrr r r rs6    zFilter.is_optional_startcCs|r |dpd}|dkr |dkS|dkrP|dkr<|d|kS|dkpJ|dkSn|dkr|dkrl|ddkS|d kr|dkp|dkSd Snp|d kr|d kr|dd kS|dkp|dkSn@|dkr|dkr|ddkS|dkp|dkSn|dkr |dkr |ddkS|dkp|dkSn|dkrT|dkr8d S|dkrN|ddkSdSn|dkr|dkrt|ddkS|dkr|dkp|dkSd Snf|dkr|dkr|ddkS|dkp|dkSn2|dkr|dkr|ddkS|dkp|dkSd S)Nr)rrrr)Zlioptgroupr#rrr)Zdtddr&Fpr)ZaddressZarticleZasideZ blockquoteZdatagridZdialogdirZdivZdlZfieldsetZfooterZformZh1Zh2Zh3Zh4Zh5Zh6headerZhrZmenuZnavZolr'ZpreZsectiontableZuloption)r+r%)ZrtZrprT)r!r )r r"r r")ZtdZthr )r r$rrr r r rXs\                     zFilter.is_optional_endN)__name__ __module__ __qualname____doc__r rrrr r r r rs   9rN)Z __future__rrrrrr r r r s site-packages/pip/_vendor/html5lib/filters/__pycache__/__init__.cpython-38.pyc000064400000000244151464452720023311 0ustar00U .e@sdS)NrrrI/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/__init__.pysite-packages/pip/_vendor/html5lib/filters/__pycache__/sanitizer.cpython-38.opt-1.pyc000064400000037520151464452720024530 0ustar00U .efD@s> ddlmZmZmZddlZddlmZmZddlm Z ddl m Z ddl mZmZd gZeed d fed d fed d fed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed d fed d!fed d"fed d#fed d$fed d%fed d&fed d'fed d(fed d)fed d*fed d+fed d,fed d-fed d.fed d/fed d0fed d1fed d2fed d3fed d4fed d5fed d6fed d7fed d8fed d9fed d:fed d;fed dfed d?fed d@fed dAfed dBfed dCfed dDfed dEfed dFfed dGfed dHfed dIfed dJfed dKfed dLfed dMfed dNfed dOfed dPfed dQfed dRfed dSfed dTfed dUfed dVfed dWfed dXfed dYfed dZfed d[fed d\fed d]fed d^fed d_fed d`fed dafed dbfed dcfed ddfed defed dffed dgfed dhfed difed djfed dkfed dlfed dmfedndofedndpfedndqfedndrfedndsfedndtfedndufedndvfedndwfedndxfedndyfedndzfednd{fednd|fednd}fednd~fedndfedndfedndfedndfedndfedndfedndfedndfedndfedndfedndfedd feddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddffZedddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd d d d d ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7ed8d9fd:dd;d;d;d<d=d>d?d@dAdBdCdDdEdFdddGdHdIdJdKdKdLdMdNdOdOdOdPdQddRdSdTdUdVd6d6edWdXfedWdYfedWdZfd[d\d]d^d_d`dadbdcdddedfdgdhdɐdidːdjdkdldmdndodpdqd@drdސdsdtdudvdwdxdydzd{d|d}d~dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd(dddd,dddddddddÐdĐdŐd6dƐdǐdȐdɐdedWdfedWdfedWdXfedWdfedWdYfedWdfedWdZfed8dfed8d9fed8dfdАdѐdҐdӐfCZedd!dddd ddddd edWdXfed8dff ZedԃZedՃZedփZed׃Zed؃ZedكZedڃZedejZGdd d e jZdS()absolute_importdivisionunicode_literalsN)escapeunescape) urllib_parse)base) namespacesprefixesFilterhtmlaabbrZacronymZaddressZareaZarticleZasideZaudiobZbigZ blockquotebrZbuttonZcanvasZcaptioncentercitecodecolZcolgroupZcommandZdatagridZdatalistZdddelZdetailsZdfnZdialogdirZdivZdlZdtZemz event-sourceZfieldsetZ figcaptionZfigureZfooterfontformheaderZh1Zh2Zh3Zh4Zh5Zh6ZhriZimginputZinsZkeygenZkbdlabelZlegendZlimmapZmenuZmeterZmulticolZnavZnextidZoloutputZoptgroupZoptionpZpreZprogressqsZsampZsectionZselectZsmallZsoundsourceZspacerspanZstrikeZstrongsubZsuptableZtbodyZtdZtextareatimeZtfootZthZtheadZtrZttuZulvarZvideoZmathmlZmactionZmathZmerrorZmfracmiZ mmultiscriptsZmnmoZmoverZmpaddedZmphantomZ mprescriptsZmrootZmrowZmspaceZmsqrtZmstyleZmsubZmsubsupZmsupZmtableZmtdZmtextZmtrZmunderZ munderovernoneZsvganimate animateColor animateMotionanimateTransformZclipPathZcircleZdefsZdescZellipsez font-facezfont-face-namez font-face-srcgZglyphZhkernlinearGradientlinemarkerZmetadataz missing-glyphZmpathpathZpolygonZpolylineradialGradientZrectsetstopZswitchtexttitleZtspanuse)Nr)NZaccept)Nzaccept-charset)NZ accesskey)Naction)NZalign)NZalt)NZ autocomplete)NZ autofocus)NZaxis)N background)NZbalance)NZbgcolor)NZ bgproperties)Nborder)NZ bordercolor)NZbordercolordark)NZbordercolorlight)NZ bottompadding)NZ cellpadding)NZ cellspacing)NZch)NZ challenge)Nchar)NZcharoff)NZchoff)Ncharset)Nchecked)Nr)Nclass)Nclear)Ncolor)NZcols)NZcolspan)NZcompact)NZcontenteditable)NZcontrols)NZcoords)Ndata)NZdatafld)NZ datapagesize)NZdatasrc)NZdatetime)Ndefault)NZdelay)Nr)NZdisabled)NZ draggable)NZdynsrc)NZenctype)Nend)NZface)Nfor)Nr)Nframe)NZ galleryimg)NZgutter)NZheaders)Nheight)NZ hidefocus)NZhidden)NZhigh)Nhref)NZhreflang)NZhspace)NZicon)Nid)NZ inputmode)NZismap)NZkeytype)Nr)NZ leftspacing)Nlang)Nlist)NZlongdesc)NZloop)NZ loopcount)NZloopend)NZ loopstart)NZlow)NZlowsrc)Nmax)NZ maxlength)NZmedia)Nmethod)Nmin)NZmultiple)Nname)NZnohref)NZnoshade)Nnowrap)Nopen)NZoptimumNpattern)NZping)Nz point-size)NZposter)NZpqg)NZpreload)Nprompt)NZ radiogroup)Nreadonly)NZrel)Nz repeat-max)Nz repeat-min)Nreplace)NZrequired)NZrev)NZ rightspacing)NZrows)NZrowspan)NZrules)NZscope)NZselected)Nshape)Nsize)Nr&)Nsrc)Nstart)NstepNstyle)NZsummary)Nsuppress)NZtabindex)Ntarget)Ntemplate)Nr<)NZ toppadding)Ntype)NZ unselectable)NZusemap)Nurn)NZvalign)Nvalue)NZvariable)Nvolume)NZvspace)NZvrml)Nwidth)NZwrapZxmlrO)NZ actiontype)NZ columnalign)NZ columnlines)NZ columnspacing)NZ columnspan)Ndepth)Ndisplay)NZ displaystyle)NZ equalcolumns)NZ equalrows)NZfence)NZ fontstyle)NZ fontweight)NZ linethickness)NZlspace)NZmathbackground)NZ mathcolor)NZ mathvariant)Nmaxsize)NZminsize)Nother)NZrowalign)NZrowlines)NZ rowspacing)NZrspace)NZ scriptlevel)NZ selection)NZ separator)NZstretchyxlinkrMZshowrf)Nz accent-height)N accumulate)NZadditive)NZ alphabetic)Nz arabic-form)NZascent)NZ attributeName)NZ attributeType)NZ baseProfile)NZbbox)NZbegin)NZby)NZcalcMode)Nz cap-heightNz clip-path)Nzcolor-rendering)NZcontent)NZcx)NZcy)Nd)NZdx)NZdy)NZdescent)NZdurNfill)N fill-opacity)N fill-rule)N font-family)N font-size)Nz font-stretch)N font-style)N font-variant)N font-weight)Nfrom)NZfx)NZfy)NZg1)NZg2)Nz glyph-name)NZ gradientUnits)NZhanging)Nz horiz-adv-x)Nzhoriz-origin-x)NZ ideographic)Nk)NZ keyPoints)NZ keySplines)NZkeyTimesNz marker-endNz marker-midNz marker-start)NZ markerHeight)NZ markerUnits)NZ markerWidth)NZ mathematical)Noffset)NZopacity)NZorient)Norigin)Nzoverline-position)Nzoverline-thickness)Nzpanose-1)Nr7)NZ pathLength)NZpoints)NZpreserveAspectRatio)Nr)NZrefX)NZrefY)NZ repeatCount)NZ repeatDur)NZrequiredExtensions)NZrequiredFeatures)NZrestart)Nrotate)Nrx)NZry)NZslope)NZstemh)NZstemv)Nz stop-color)Nz stop-opacity)Nzstrikethrough-position)Nzstrikethrough-thicknessNstroke)Nzstroke-dasharray)Nzstroke-dashoffset)Nstroke-linecap)Nstroke-linejoin)Nzstroke-miterlimit)Nstroke-opacity)N stroke-width)NZsystemLanguage)Nz text-anchor)Nto)NZ transform)NZu1)NZu2)Nzunderline-position)Nzunderline-thickness)NZunicode)Nz unicode-range)Nz units-per-em)Nvalues)Nversion)NZviewBox)NZ visibility)NZwidths)Nx)Nzx-height)NZx1)NZx2ZactuateZarcroleZroler Zspace)Ny)NZy1)NZy2)NZ zoomAndPan) rq)Nz color-profileNcursorrsNfilter)Nr6rrr~)Nmaskr))NZaltGlyph)Nr/)Nr0)Nr1)Nr2r)NZfeImager)Nr4rW)Nr8)NZtextpath)NZtref)Nr9)Nr=).Zazimuthzbackground-colorzborder-bottom-colorzborder-collapsez border-colorzborder-left-colorzborder-right-colorzborder-top-colorrErFr directionrlZ elevationfloatrrwrxryrzr{rLzletter-spacingz line-heightZoverflowpausez pause-afterz pause-beforeZpitchz pitch-rangeZrichnessZspeakz speak-headerz speak-numeralzspeak-punctuationz speech-rateZstressz text-alignztext-decorationz text-indentz unicode-bidizvertical-alignz voice-familyriz white-spacerj)'autoZaquaZblackblockZblueZboldZbothZbottomZbrownrZcollapseZdashedZdottedZfuchsiaZgrayZgreenz !importantZitalicleftZlimeZmaroonZmediumr.ZnavyZnormalrUZoliveZpointerZpurpleZredrightZsolidZsilverZtealtopZ transparentZ underlineZwhiteZyellow)rtrurvrrrrr)Zed2kZftpZhttpZhttpsZircZmailtoZnewsZgopherZnntpZtelnetZwebcalZxmppZcalltoZfeedrgZaimZrsynctagZsshZsftpZrtspZafsrG)z image/pngz image/jpegz image/gifz image/webpz image/bmpz text/plainaL ^ # Match a content type / (?P[-a-zA-Z0-9.]+/[-a-zA-Z0-9.]+) # Match any character set and encoding (?:(?:;charset=(?:[-a-zA-Z0-9]+)(?:;(?:base64))?) |(?:;(?:base64))?(?:;charset=(?:[-a-zA-Z0-9]+))?) # Assume the rest is data ,.* $ c s^eZdZdZeeeeee e e e e f fdd ZddZddZdd Zd d Zd d ZZS)r zISanitizes token stream of XHTML+MathML+SVG and of inline style attributesc sPtt||||_||_||_||_||_||_||_ | |_ | |_ | |_ dS)aSCreates a Filter :arg allowed_elements: set of elements to allow--everything else will be escaped :arg allowed_attributes: set of attributes to allow in elements--everything else will be stripped :arg allowed_css_properties: set of CSS properties to allow--everything else will be stripped :arg allowed_css_keywords: set of CSS keywords to allow--everything else will be stripped :arg allowed_svg_properties: set of SVG properties to allow--everything else will be removed :arg allowed_protocols: set of allowed protocols for URIs :arg allowed_content_types: set of allowed content types for ``data`` URIs. :arg attr_val_is_uri: set of attributes that have URI values--values that have a scheme not listed in ``allowed_protocols`` are removed :arg svg_attr_val_allows_ref: set of SVG attributes that can have references :arg svg_allow_local_href: set of SVG elements that can have local hrefs--these are removed N) superr __init__allowed_elementsallowed_attributesallowed_css_propertiesallowed_css_keywordsallowed_svg_propertiesallowed_protocolsallowed_content_typesattr_val_is_urisvg_attr_val_allows_refsvg_allow_local_href) selfr%rrrrrrrrrr __class__J/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/sanitizer.pyrs+zFilter.__init__ccs*tj|D]}||}|r |Vq dS)N)r r __iter__sanitize_token)rtokenrrrrs zFilter.__iter__cCsp|d}|dkr^|d}|d}||f|jksH|dkrRtd|f|jkrR||S||Sn|dkrhn|SdS)Nrf)ZStartTagEndTagZEmptyTagrT namespacerComment)rr allowed_tokendisallowed_token)rr token_typerTrrrrr s  zFilter.sanitize_tokenc Csd|kr|d}t|}||jD]}|d|=||q(||j@D]}tddt||}| dd}zt |}Wnt k rd}||=YnX|rL|j rL|j |j kr||=|j dkrLt|j}|s||=qL|d|jkrL||=qL|jD]*}||krtddt||||<q|d|jkrvtd d f|krvtd |td d frv|td d f=d |kr||d |d <||d<|S) NrGu [`- - \s]+u�Z content_typezurl\s*\(\s*[^#\s][^)]+?\) rTrorMz ^\s*[^#\s].*ra)r9keysrremoverrer'rlowerr[urlparse ValueErrorZschemerdata_content_typematchr7grouprrrr search sanitize_css) rrattrsZ attr_namesZ to_removeattrZ val_unescapedZurirrrrrsV               zFilter.allowed_tokencCs|d}|dkr"d|d|d<n|drg}|dD]:\\}}}|d|dkrX|ndt||ft|fq:d|dd |f|d<nd |d|d<|d r|ddd d |d<d|d<|d=|S)NrfrzrTrGz %s="%s"z%s:%sz<%s%s>rz<%s>Z selfClosingz/>Z Characters)itemsappendr rjoinget)rrrrnsrTvrrrrQs0 zFilter.disallowed_tokencCstdd|}td|s"dStd|s2dSg}td|D]\}}|sPqB||jkrv||d|dqB|d d d kr|D]}||j krtd |sqq||d|dqB||j krB||d|dqBd |S) Nzurl\s*\(\s*[^\s)]+?\s*\)\s*rz@^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$rz ^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$z([-\w]+)\s*:\s*([^:;]*)z: ;-r)r?r@ZmarginZpaddingz_^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$) rcompiler'rfindallrrrsplitrrr)rrbZcleanZproprhkeywordrrrres*     zFilter.sanitize_css)__name__ __module__ __qualname____doc__rrrrrrrrrrrrrrrr __classcell__rrrrr s"72) Z __future__rrrrZxml.sax.saxutilsrrZpip._vendor.six.movesrrrr Z constantsr r __all__ frozensetrrrrrrrrrrrVERBOSErr rrrrs                                                                                                                                                                   % K   1 *   site-packages/pip/_vendor/html5lib/filters/__pycache__/lint.cpython-38.opt-1.pyc000064400000003505151464452720023462 0ustar00U .e;@shddlmZmZmZddlmZddlmZddlm Z m Z ddlm Z d e Z Gd d d ej Z d S) )absolute_importdivisionunicode_literals) text_type)base) namespaces voidElements)spaceCharacterscs*eZdZdZdfdd ZddZZS)FilterzgLints the token stream for errors If it finds any errors, it'll raise an ``AssertionError``. Tcstt||||_dS)zCreates a Filter :arg source: the source token stream :arg require_matching_tags: whether or not to require matching tags N)superr __init__require_matching_tags)selfsourcer __class__E/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/lint.pyrszFilter.__init__c cs.g}tj|D]}|d}|dkr|d}|d}|rF|tdkrP|tkrPn|dkrl|jrl|||f|dD] \\}}}qxn|dkr|d}|d}|r|tdkr|tkrn|jr|}nX|d kr|d}nF|d kr|d}|d krn,|d kr |d}n|d krn |dkr"n|VqdS)Ntype)StartTagZEmptyTag namespacenameZhtmlrdataZEndTagComment)Z CharactersSpaceCharactersrZDoctypeZEntityZSerializerError) rr __iter__r r rappenditemspop) rZ open_elementstokenrrrvaluestartrrrrrsF   zFilter.__iter__)T)__name__ __module__ __qualname____doc__rr __classcell__rrrrr s r N)Z __future__rrrZpip._vendor.sixrr rZ constantsr r r joinr rrrrs     site-packages/pip/_vendor/html5lib/filters/__pycache__/sanitizer.cpython-38.pyc000064400000037615151464452720023576 0ustar00U .efD@s> ddlmZmZmZddlZddlmZmZddlm Z ddl m Z ddl mZmZd gZeed d fed d fed d fed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed dfed d fed d!fed d"fed d#fed d$fed d%fed d&fed d'fed d(fed d)fed d*fed d+fed d,fed d-fed d.fed d/fed d0fed d1fed d2fed d3fed d4fed d5fed d6fed d7fed d8fed d9fed d:fed d;fed dfed d?fed d@fed dAfed dBfed dCfed dDfed dEfed dFfed dGfed dHfed dIfed dJfed dKfed dLfed dMfed dNfed dOfed dPfed dQfed dRfed dSfed dTfed dUfed dVfed dWfed dXfed dYfed dZfed d[fed d\fed d]fed d^fed d_fed d`fed dafed dbfed dcfed ddfed defed dffed dgfed dhfed difed djfed dkfed dlfed dmfedndofedndpfedndqfedndrfedndsfedndtfedndufedndvfedndwfedndxfedndyfedndzfednd{fednd|fednd}fednd~fedndfedndfedndfedndfedndfedndfedndfedndfedndfedndfedndfedd feddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddffZedddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd d d d d ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7ed8d9fd:dd;d;d;d<d=d>d?d@dAdBdCdDdEdFdddGdHdIdJdKdKdLdMdNdOdOdOdPdQddRdSdTdUdVd6d6edWdXfedWdYfedWdZfd[d\d]d^d_d`dadbdcdddedfdgdhdɐdidːdjdkdldmdndodpdqd@drdސdsdtdudvdwdxdydzd{d|d}d~dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd(dddd,dddddddddÐdĐdŐd6dƐdǐdȐdɐdedWdfedWdfedWdXfedWdfedWdYfedWdfedWdZfed8dfed8d9fed8dfdАdѐdҐdӐfCZedd!dddd ddddd edWdXfed8dff ZedԃZedՃZedփZed׃Zed؃ZedكZedڃZedejZGdd d e jZdS()absolute_importdivisionunicode_literalsN)escapeunescape) urllib_parse)base) namespacesprefixesFilterhtmlaabbrZacronymZaddressZareaZarticleZasideZaudiobZbigZ blockquotebrZbuttonZcanvasZcaptioncentercitecodecolZcolgroupZcommandZdatagridZdatalistZdddelZdetailsZdfnZdialogdirZdivZdlZdtZemz event-sourceZfieldsetZ figcaptionZfigureZfooterfontformheaderZh1Zh2Zh3Zh4Zh5Zh6ZhriZimginputZinsZkeygenZkbdlabelZlegendZlimmapZmenuZmeterZmulticolZnavZnextidZoloutputZoptgroupZoptionpZpreZprogressqsZsampZsectionZselectZsmallZsoundsourceZspacerspanZstrikeZstrongsubZsuptableZtbodyZtdZtextareatimeZtfootZthZtheadZtrZttuZulvarZvideoZmathmlZmactionZmathZmerrorZmfracmiZ mmultiscriptsZmnmoZmoverZmpaddedZmphantomZ mprescriptsZmrootZmrowZmspaceZmsqrtZmstyleZmsubZmsubsupZmsupZmtableZmtdZmtextZmtrZmunderZ munderovernoneZsvganimate animateColor animateMotionanimateTransformZclipPathZcircleZdefsZdescZellipsez font-facezfont-face-namez font-face-srcgZglyphZhkernlinearGradientlinemarkerZmetadataz missing-glyphZmpathpathZpolygonZpolylineradialGradientZrectsetstopZswitchtexttitleZtspanuse)Nr)NZaccept)Nzaccept-charset)NZ accesskey)Naction)NZalign)NZalt)NZ autocomplete)NZ autofocus)NZaxis)N background)NZbalance)NZbgcolor)NZ bgproperties)Nborder)NZ bordercolor)NZbordercolordark)NZbordercolorlight)NZ bottompadding)NZ cellpadding)NZ cellspacing)NZch)NZ challenge)Nchar)NZcharoff)NZchoff)Ncharset)Nchecked)Nr)Nclass)Nclear)Ncolor)NZcols)NZcolspan)NZcompact)NZcontenteditable)NZcontrols)NZcoords)Ndata)NZdatafld)NZ datapagesize)NZdatasrc)NZdatetime)Ndefault)NZdelay)Nr)NZdisabled)NZ draggable)NZdynsrc)NZenctype)Nend)NZface)Nfor)Nr)Nframe)NZ galleryimg)NZgutter)NZheaders)Nheight)NZ hidefocus)NZhidden)NZhigh)Nhref)NZhreflang)NZhspace)NZicon)Nid)NZ inputmode)NZismap)NZkeytype)Nr)NZ leftspacing)Nlang)Nlist)NZlongdesc)NZloop)NZ loopcount)NZloopend)NZ loopstart)NZlow)NZlowsrc)Nmax)NZ maxlength)NZmedia)Nmethod)Nmin)NZmultiple)Nname)NZnohref)NZnoshade)Nnowrap)Nopen)NZoptimumNpattern)NZping)Nz point-size)NZposter)NZpqg)NZpreload)Nprompt)NZ radiogroup)Nreadonly)NZrel)Nz repeat-max)Nz repeat-min)Nreplace)NZrequired)NZrev)NZ rightspacing)NZrows)NZrowspan)NZrules)NZscope)NZselected)Nshape)Nsize)Nr&)Nsrc)Nstart)NstepNstyle)NZsummary)Nsuppress)NZtabindex)Ntarget)Ntemplate)Nr<)NZ toppadding)Ntype)NZ unselectable)NZusemap)Nurn)NZvalign)Nvalue)NZvariable)Nvolume)NZvspace)NZvrml)Nwidth)NZwrapZxmlrO)NZ actiontype)NZ columnalign)NZ columnlines)NZ columnspacing)NZ columnspan)Ndepth)Ndisplay)NZ displaystyle)NZ equalcolumns)NZ equalrows)NZfence)NZ fontstyle)NZ fontweight)NZ linethickness)NZlspace)NZmathbackground)NZ mathcolor)NZ mathvariant)Nmaxsize)NZminsize)Nother)NZrowalign)NZrowlines)NZ rowspacing)NZrspace)NZ scriptlevel)NZ selection)NZ separator)NZstretchyxlinkrMZshowrf)Nz accent-height)N accumulate)NZadditive)NZ alphabetic)Nz arabic-form)NZascent)NZ attributeName)NZ attributeType)NZ baseProfile)NZbbox)NZbegin)NZby)NZcalcMode)Nz cap-heightNz clip-path)Nzcolor-rendering)NZcontent)NZcx)NZcy)Nd)NZdx)NZdy)NZdescent)NZdurNfill)N fill-opacity)N fill-rule)N font-family)N font-size)Nz font-stretch)N font-style)N font-variant)N font-weight)Nfrom)NZfx)NZfy)NZg1)NZg2)Nz glyph-name)NZ gradientUnits)NZhanging)Nz horiz-adv-x)Nzhoriz-origin-x)NZ ideographic)Nk)NZ keyPoints)NZ keySplines)NZkeyTimesNz marker-endNz marker-midNz marker-start)NZ markerHeight)NZ markerUnits)NZ markerWidth)NZ mathematical)Noffset)NZopacity)NZorient)Norigin)Nzoverline-position)Nzoverline-thickness)Nzpanose-1)Nr7)NZ pathLength)NZpoints)NZpreserveAspectRatio)Nr)NZrefX)NZrefY)NZ repeatCount)NZ repeatDur)NZrequiredExtensions)NZrequiredFeatures)NZrestart)Nrotate)Nrx)NZry)NZslope)NZstemh)NZstemv)Nz stop-color)Nz stop-opacity)Nzstrikethrough-position)Nzstrikethrough-thicknessNstroke)Nzstroke-dasharray)Nzstroke-dashoffset)Nstroke-linecap)Nstroke-linejoin)Nzstroke-miterlimit)Nstroke-opacity)N stroke-width)NZsystemLanguage)Nz text-anchor)Nto)NZ transform)NZu1)NZu2)Nzunderline-position)Nzunderline-thickness)NZunicode)Nz unicode-range)Nz units-per-em)Nvalues)Nversion)NZviewBox)NZ visibility)NZwidths)Nx)Nzx-height)NZx1)NZx2ZactuateZarcroleZroler Zspace)Ny)NZy1)NZy2)NZ zoomAndPan) rq)Nz color-profileNcursorrsNfilter)Nr6rrr~)Nmaskr))NZaltGlyph)Nr/)Nr0)Nr1)Nr2r)NZfeImager)Nr4rW)Nr8)NZtextpath)NZtref)Nr9)Nr=).Zazimuthzbackground-colorzborder-bottom-colorzborder-collapsez border-colorzborder-left-colorzborder-right-colorzborder-top-colorrErFr directionrlZ elevationfloatrrwrxryrzr{rLzletter-spacingz line-heightZoverflowpausez pause-afterz pause-beforeZpitchz pitch-rangeZrichnessZspeakz speak-headerz speak-numeralzspeak-punctuationz speech-rateZstressz text-alignztext-decorationz text-indentz unicode-bidizvertical-alignz voice-familyriz white-spacerj)'autoZaquaZblackblockZblueZboldZbothZbottomZbrownrZcollapseZdashedZdottedZfuchsiaZgrayZgreenz !importantZitalicleftZlimeZmaroonZmediumr.ZnavyZnormalrUZoliveZpointerZpurpleZredrightZsolidZsilverZtealtopZ transparentZ underlineZwhiteZyellow)rtrurvrrrrr)Zed2kZftpZhttpZhttpsZircZmailtoZnewsZgopherZnntpZtelnetZwebcalZxmppZcalltoZfeedrgZaimZrsynctagZsshZsftpZrtspZafsrG)z image/pngz image/jpegz image/gifz image/webpz image/bmpz text/plainaL ^ # Match a content type / (?P[-a-zA-Z0-9.]+/[-a-zA-Z0-9.]+) # Match any character set and encoding (?:(?:;charset=(?:[-a-zA-Z0-9]+)(?:;(?:base64))?) |(?:;(?:base64))?(?:;charset=(?:[-a-zA-Z0-9]+))?) # Assume the rest is data ,.* $ c s^eZdZdZeeeeee e e e e f fdd ZddZddZdd Zd d Zd d ZZS)r zISanitizes token stream of XHTML+MathML+SVG and of inline style attributesc sPtt||||_||_||_||_||_||_||_ | |_ | |_ | |_ dS)aSCreates a Filter :arg allowed_elements: set of elements to allow--everything else will be escaped :arg allowed_attributes: set of attributes to allow in elements--everything else will be stripped :arg allowed_css_properties: set of CSS properties to allow--everything else will be stripped :arg allowed_css_keywords: set of CSS keywords to allow--everything else will be stripped :arg allowed_svg_properties: set of SVG properties to allow--everything else will be removed :arg allowed_protocols: set of allowed protocols for URIs :arg allowed_content_types: set of allowed content types for ``data`` URIs. :arg attr_val_is_uri: set of attributes that have URI values--values that have a scheme not listed in ``allowed_protocols`` are removed :arg svg_attr_val_allows_ref: set of SVG attributes that can have references :arg svg_allow_local_href: set of SVG elements that can have local hrefs--these are removed N) superr __init__allowed_elementsallowed_attributesallowed_css_propertiesallowed_css_keywordsallowed_svg_propertiesallowed_protocolsallowed_content_typesattr_val_is_urisvg_attr_val_allows_refsvg_allow_local_href) selfr%rrrrrrrrrr __class__J/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/sanitizer.pyrs+zFilter.__init__ccs*tj|D]}||}|r |Vq dS)N)r r __iter__sanitize_token)rtokenrrrrs zFilter.__iter__cCsp|d}|dkr^|d}|d}||f|jksH|dkrRtd|f|jkrR||S||Sn|dkrhn|SdS)Nrf)StartTagEndTagEmptyTagrT namespacerComment)rr allowed_tokendisallowed_token)rr token_typerTrrrrr s  zFilter.sanitize_tokenc Csd|kr|d}t|}||jD]}|d|=||q(||j@D]}||ks\ttddt|| }| dd}zt |}Wnt k rd}||=YnX|rL|j rL|j |jkr||=|j dkrLt|j}|s||=qL|d|jkrL||=qL|jD]*}||kr tddt||||<q |d|jkrtd d f|krtd |td d fr|td d f=d |kr||d |d <||d<|S) NrGu [`- - \s]+u�Z content_typezurl\s*\(\s*[^#\s][^)]+?\) rTrorMz ^\s*[^#\s].*ra)r9keysrremoverAssertionErrorrer'rlowerr[urlparse ValueErrorZschemerdata_content_typematchr7grouprrrr search sanitize_css) rrattrsZ attr_namesZ to_removeattrZ val_unescapedZurirrrrrsX                zFilter.allowed_tokencCs|d}|dkr"d|d|d<n|dr|dks6tg}|dD]:\\}}}|d|dkrd|ndt||ft|fqFd |dd |f|d<nd |d|d<|d r|ddd d|d<d|d<|d=|S)NrfrzrTrG)rrz %s="%s"z%s:%sz<%s%s>rz<%s>Z selfClosingz/>Z Characters)ritemsappendr rjoinget)rrrrnsrTvrrrrQs 0 zFilter.disallowed_tokencCstdd|}td|s"dStd|s2dSg}td|D]\}}|sPqB||jkrv||d|dqB|d d d kr|D]}||j krtd |sqq||d|dqB||j krB||d|dqBd |S) Nzurl\s*\(\s*[^\s)]+?\s*\)\s*rz@^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$rz ^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$z([-\w]+)\s*:\s*([^:;]*)z: ;-r)r?r@ZmarginZpaddingz_^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$) rcompiler'rfindallrrrsplitrrr)rrbZcleanZproprhkeywordrrrres*     zFilter.sanitize_css)__name__ __module__ __qualname____doc__rrrrrrrrrrrrrrrr __classcell__rrrrr s"72) Z __future__rrrrZxml.sax.saxutilsrrZpip._vendor.six.movesrrrr Z constantsr r __all__ frozensetrrrrrrrrrrrVERBOSErr rrrrs                                                                                                                                                                   % K   1 *   site-packages/pip/_vendor/html5lib/filters/__pycache__/lint.cpython-38.pyc000064400000005046151464452720022525 0ustar00U .e;@shddlmZmZmZddlmZddlmZddlm Z m Z ddlm Z d e Z Gd d d ej Z d S) )absolute_importdivisionunicode_literals) text_type)base) namespaces voidElements)spaceCharacterscs*eZdZdZdfdd ZddZZS)FilterzgLints the token stream for errors If it finds any errors, it'll raise an ``AssertionError``. Tcstt||||_dS)zCreates a Filter :arg source: the source token stream :arg require_matching_tags: whether or not to require matching tags N)superr __init__require_matching_tags)selfsourcer __class__E/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/lint.pyrszFilter.__init__c csDg}tj|D],}|d}|dkrF|d}|d}|dksNt|tsNt|dksZtt|tsht|dksttt|dtst|r|tdkr|tkr|dkstn |d kst|d kr|j r| ||f|d D]`\\}}}|dkst|tst|dkstt|ts$t|dks2tt|tstqn|d kr|d}|d}|dkszt|tszt|dkstt|tst|dkst|r|tdkr|tkrd std d |in"|j r8| }|||fks8tn6|dkr(|d}t|ts8tn|dkrx|d}t|tsJt|dksXt|dkr8| tdks8tn|dkr|d}|dkst|tst|ddkst|tst|ddks8t|ts8tnV|dkrt|dts8tn6|dkr"t|dts8tnd s8tdd|i|VqdS)Ntype)StartTagEmptyTag namespacenamer dataZhtmlrrZEndTagFz.Void element reported as EndTag token: %(tag)stagComment)Z CharactersSpaceCharactersrZDoctypeZpublicIdZsystemIdZEntityZSerializerErrorzUnknown token type: %(type)s)rr __iter__ isinstancerAssertionErrordictr r rappenditemspopstripr ) rZ open_elementstokenrrrvaluestartrrrrr sl            zFilter.__iter__)T)__name__ __module__ __qualname____doc__rr __classcell__rrrrr s r N)Z __future__rrrZpip._vendor.sixrr rZ constantsr r r joinr rrrrs     site-packages/pip/_vendor/html5lib/filters/__pycache__/alphabeticalattributes.cpython-38.opt-1.pyc000064400000002406151464452720027233 0ustar00U .e@sJddlmZmZmZddlmZddlmZddZGdddej Z d S) )absolute_importdivisionunicode_literals)base) OrderedDictcCs|ddpd|ddfS)zReturn an appropriate key for an attribute for sorting Attributes have a namespace that can be either ``None`` or a string. We can't compare the two because they're different types, so we convert ``None`` to an empty string first. rr)attrr r W/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/alphabeticalattributes.py _attr_keysr c@seZdZdZddZdS)Filterz$Alphabetizes attributes for elementsccs\tj|D]J}|ddkrPt}t|dtdD]\}}|||<q6||d<|Vq dS)Ntype)ZStartTagZEmptyTagdata)key)rr __iter__rsorteditemsr )selftokenZattrsnamevaluer r r rs   zFilter.__iter__N)__name__ __module__ __qualname____doc__rr r r r r sr N) Z __future__rrrrr collectionsrr r r r r r s   site-packages/pip/_vendor/html5lib/filters/__pycache__/whitespace.cpython-38.pyc000064400000002464151464452720023714 0ustar00U .e@snddlmZmZmZddlZddlmZddlmZm Z d e Z e de Z Gd d d ej Z d d ZdS) )absolute_importdivisionunicode_literalsN)base)rcdataElementsspaceCharactersz[%s]+c@s,eZdZdZeddgeeZddZdS)FilterzACollapses whitespace except in pre, textarea, and script elementsZpreZtextareaccsd}tj|D]}|d}|dkr@|s6|d|jkr@|d7}nP|dkrV|rV|d8}n:|st|dkrt|drtd |d<n|s|d krt|d|d<|VqdS) NrtypeZStartTagnamerZEndTagZSpaceCharactersdata Z Characters)rr __iter__spacePreserveElementscollapse_spaces)selfZpreservetokenr rK/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/whitespace.pyrs       zFilter.__iter__N) __name__ __module__ __qualname____doc__ frozensetlistrrrrrrrr sr cCs td|S)Nr) SPACES_REGEXsub)textrrrr%sr)Z __future__rrrrer rZ constantsrr joincompilerr rrrrrs  site-packages/pip/_vendor/html5lib/filters/__pycache__/base.cpython-38.opt-1.pyc000064400000001502151464452720023421 0ustar00U .e@s(ddlmZmZmZGdddeZdS))absolute_importdivisionunicode_literalsc@s$eZdZddZddZddZdS)FiltercCs ||_dSN)source)selfrr E/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/base.py__init__szFilter.__init__cCs t|jSr)iterr)rr r r __iter__szFilter.__iter__cCs t|j|Sr)getattrr)rnamer r r __getattr__ szFilter.__getattr__N)__name__ __module__ __qualname__r r rr r r r rsrN)Z __future__rrrobjectrr r r r ssite-packages/pip/_vendor/html5lib/filters/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000244151464452720024250 0ustar00U .e@sdS)NrrrI/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/__init__.pysite-packages/pip/_vendor/html5lib/filters/__pycache__/inject_meta_charset.cpython-38.opt-1.pyc000064400000003460151464452720026507 0ustar00U .e @s6ddlmZmZmZddlmZGdddejZdS))absolute_importdivisionunicode_literals)basec@s eZdZdZddZddZdS)Filterz=Injects ```` tag into head of documentcCstj||||_dS)ztCreates a Filter :arg source: the source token stream :arg encoding: the encoding to set N)rr__init__encoding)selfsourcer r T/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/inject_meta_charset.pyrszFilter.__init__c csd}|jdk}g}tj|D]}|d}|dkrL|ddkrHd}nd|dkrJ|ddkrd }|d D]X\\}}} |dk rqvqv|d kr|j|d ||f<d }qqv|d krv| dkrvd }qv|rd|d krd|j|d d<d }nP|ddkr|sdd|d dVddd|jidVdddVd }qnf|dkr|ddkr|r|dV|sddd|jidV|r|dVqd }d}|dkr||q|VqdS)NZpre_headtypeZStartTagnameheadZin_headZEmptyTagmetaFdatacharsetTz http-equivz content-type)NZcontentztext/html; charset=%s)rrr)NrZEndTag)rrrZ post_head)r rr__iter__loweritemspopappend) r stateZ meta_foundZpendingtokenrZhas_http_equiv_content_type namespacervaluer r r rs^        zFilter.__iter__N)__name__ __module__ __qualname____doc__rrr r r r rs rN)Z __future__rrrrrr r r r s site-packages/pip/_vendor/html5lib/filters/__pycache__/whitespace.cpython-38.opt-1.pyc000064400000002464151464452720024653 0ustar00U .e@snddlmZmZmZddlZddlmZddlmZm Z d e Z e de Z Gd d d ej Z d d ZdS) )absolute_importdivisionunicode_literalsN)base)rcdataElementsspaceCharactersz[%s]+c@s,eZdZdZeddgeeZddZdS)FilterzACollapses whitespace except in pre, textarea, and script elementsZpreZtextareaccsd}tj|D]}|d}|dkr@|s6|d|jkr@|d7}nP|dkrV|rV|d8}n:|st|dkrt|drtd |d<n|s|d krt|d|d<|VqdS) NrtypeZStartTagnamerZEndTagZSpaceCharactersdata Z Characters)rr __iter__spacePreserveElementscollapse_spaces)selfZpreservetokenr rK/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/whitespace.pyrs       zFilter.__iter__N) __name__ __module__ __qualname____doc__ frozensetlistrrrrrrrr sr cCs td|S)Nr) SPACES_REGEXsub)textrrrr%sr)Z __future__rrrrer rZ constantsrr joincompilerr rrrrrs  site-packages/pip/_vendor/html5lib/filters/__pycache__/alphabeticalattributes.cpython-38.pyc000064400000002406151464452720026274 0ustar00U .e@sJddlmZmZmZddlmZddlmZddZGdddej Z d S) )absolute_importdivisionunicode_literals)base) OrderedDictcCs|ddpd|ddfS)zReturn an appropriate key for an attribute for sorting Attributes have a namespace that can be either ``None`` or a string. We can't compare the two because they're different types, so we convert ``None`` to an empty string first. rr)attrr r W/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/alphabeticalattributes.py _attr_keysr c@seZdZdZddZdS)Filterz$Alphabetizes attributes for elementsccs\tj|D]J}|ddkrPt}t|dtdD]\}}|||<q6||d<|Vq dS)Ntype)ZStartTagZEmptyTagdata)key)rr __iter__rsorteditemsr )selftokenZattrsnamevaluer r r rs   zFilter.__iter__N)__name__ __module__ __qualname____doc__rr r r r r sr N) Z __future__rrrrr collectionsrr r r r r r s   site-packages/pip/_vendor/html5lib/filters/__pycache__/inject_meta_charset.cpython-38.pyc000064400000003460151464452720025550 0ustar00U .e @s6ddlmZmZmZddlmZGdddejZdS))absolute_importdivisionunicode_literals)basec@s eZdZdZddZddZdS)Filterz=Injects ```` tag into head of documentcCstj||||_dS)ztCreates a Filter :arg source: the source token stream :arg encoding: the encoding to set N)rr__init__encoding)selfsourcer r T/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/filters/inject_meta_charset.pyrszFilter.__init__c csd}|jdk}g}tj|D]}|d}|dkrL|ddkrHd}nd|dkrJ|ddkrd }|d D]X\\}}} |dk rqvqv|d kr|j|d ||f<d }qqv|d krv| dkrvd }qv|rd|d krd|j|d d<d }nP|ddkr|sdd|d dVddd|jidVdddVd }qnf|dkr|ddkr|r|dV|sddd|jidV|r|dVqd }d}|dkr||q|VqdS)NZpre_headtypeZStartTagnameheadZin_headZEmptyTagmetaFdatacharsetTz http-equivz content-type)NZcontentztext/html; charset=%s)rrr)NrZEndTag)rrrZ post_head)r rr__iter__loweritemspopappend) r stateZ meta_foundZpendingtokenrZhas_http_equiv_content_type namespacervaluer r r rs^        zFilter.__iter__N)__name__ __module__ __qualname____doc__rrr r r r rs rN)Z __future__rrrrrr r r r s site-packages/pip/_vendor/html5lib/filters/inject_meta_charset.py000064400000005601151464452720021261 0ustar00from __future__ import absolute_import, division, unicode_literals from . import base class Filter(base.Filter): """Injects ```` tag into head of document""" def __init__(self, source, encoding): """Creates a Filter :arg source: the source token stream :arg encoding: the encoding to set """ base.Filter.__init__(self, source) self.encoding = encoding def __iter__(self): state = "pre_head" meta_found = (self.encoding is None) pending = [] for token in base.Filter.__iter__(self): type = token["type"] if type == "StartTag": if token["name"].lower() == "head": state = "in_head" elif type == "EmptyTag": if token["name"].lower() == "meta": # replace charset with actual encoding has_http_equiv_content_type = False for (namespace, name), value in token["data"].items(): if namespace is not None: continue elif name.lower() == 'charset': token["data"][(namespace, name)] = self.encoding meta_found = True break elif name == 'http-equiv' and value.lower() == 'content-type': has_http_equiv_content_type = True else: if has_http_equiv_content_type and (None, "content") in token["data"]: token["data"][(None, "content")] = 'text/html; charset=%s' % self.encoding meta_found = True elif token["name"].lower() == "head" and not meta_found: # insert meta into empty head yield {"type": "StartTag", "name": "head", "data": token["data"]} yield {"type": "EmptyTag", "name": "meta", "data": {(None, "charset"): self.encoding}} yield {"type": "EndTag", "name": "head"} meta_found = True continue elif type == "EndTag": if token["name"].lower() == "head" and pending: # insert meta into head (if necessary) and flush pending queue yield pending.pop(0) if not meta_found: yield {"type": "EmptyTag", "name": "meta", "data": {(None, "charset"): self.encoding}} while pending: yield pending.pop(0) meta_found = True state = "post_head" if state == "in_head": pending.append(token) else: yield token site-packages/pip/_vendor/html5lib/filters/lint.py000064400000007073151464452720016241 0ustar00from __future__ import absolute_import, division, unicode_literals from pip._vendor.six import text_type from . import base from ..constants import namespaces, voidElements from ..constants import spaceCharacters spaceCharacters = "".join(spaceCharacters) class Filter(base.Filter): """Lints the token stream for errors If it finds any errors, it'll raise an ``AssertionError``. """ def __init__(self, source, require_matching_tags=True): """Creates a Filter :arg source: the source token stream :arg require_matching_tags: whether or not to require matching tags """ super(Filter, self).__init__(source) self.require_matching_tags = require_matching_tags def __iter__(self): open_elements = [] for token in base.Filter.__iter__(self): type = token["type"] if type in ("StartTag", "EmptyTag"): namespace = token["namespace"] name = token["name"] assert namespace is None or isinstance(namespace, text_type) assert namespace != "" assert isinstance(name, text_type) assert name != "" assert isinstance(token["data"], dict) if (not namespace or namespace == namespaces["html"]) and name in voidElements: assert type == "EmptyTag" else: assert type == "StartTag" if type == "StartTag" and self.require_matching_tags: open_elements.append((namespace, name)) for (namespace, name), value in token["data"].items(): assert namespace is None or isinstance(namespace, text_type) assert namespace != "" assert isinstance(name, text_type) assert name != "" assert isinstance(value, text_type) elif type == "EndTag": namespace = token["namespace"] name = token["name"] assert namespace is None or isinstance(namespace, text_type) assert namespace != "" assert isinstance(name, text_type) assert name != "" if (not namespace or namespace == namespaces["html"]) and name in voidElements: assert False, "Void element reported as EndTag token: %(tag)s" % {"tag": name} elif self.require_matching_tags: start = open_elements.pop() assert start == (namespace, name) elif type == "Comment": data = token["data"] assert isinstance(data, text_type) elif type in ("Characters", "SpaceCharacters"): data = token["data"] assert isinstance(data, text_type) assert data != "" if type == "SpaceCharacters": assert data.strip(spaceCharacters) == "" elif type == "Doctype": name = token["name"] assert name is None or isinstance(name, text_type) assert token["publicId"] is None or isinstance(name, text_type) assert token["systemId"] is None or isinstance(name, text_type) elif type == "Entity": assert isinstance(token["name"], text_type) elif type == "SerializerError": assert isinstance(token["data"], text_type) else: assert False, "Unknown token type: %(type)s" % {"type": type} yield token site-packages/pip/_vendor/html5lib/filters/__init__.py000064400000000000151464452720017011 0ustar00site-packages/pip/_vendor/html5lib/treebuilders/base.py000064400000034363151464452720017230 0ustar00from __future__ import absolute_import, division, unicode_literals from pip._vendor.six import text_type from ..constants import scopingElements, tableInsertModeElements, namespaces # The scope markers are inserted when entering object elements, # marquees, table cells, and table captions, and are used to prevent formatting # from "leaking" into tables, object elements, and marquees. Marker = None listElementsMap = { None: (frozenset(scopingElements), False), "button": (frozenset(scopingElements | set([(namespaces["html"], "button")])), False), "list": (frozenset(scopingElements | set([(namespaces["html"], "ol"), (namespaces["html"], "ul")])), False), "table": (frozenset([(namespaces["html"], "html"), (namespaces["html"], "table")]), False), "select": (frozenset([(namespaces["html"], "optgroup"), (namespaces["html"], "option")]), True) } class Node(object): """Represents an item in the tree""" def __init__(self, name): """Creates a Node :arg name: The tag name associated with the node """ # The tag name assocaited with the node self.name = name # The parent of the current node (or None for the document node) self.parent = None # The value of the current node (applies to text nodes and comments) self.value = None # A dict holding name -> value pairs for attributes of the node self.attributes = {} # A list of child nodes of the current node. This must include all # elements but not necessarily other node types. self.childNodes = [] # A list of miscellaneous flags that can be set on the node. self._flags = [] def __str__(self): attributesStr = " ".join(["%s=\"%s\"" % (name, value) for name, value in self.attributes.items()]) if attributesStr: return "<%s %s>" % (self.name, attributesStr) else: return "<%s>" % (self.name) def __repr__(self): return "<%s>" % (self.name) def appendChild(self, node): """Insert node as a child of the current node :arg node: the node to insert """ raise NotImplementedError def insertText(self, data, insertBefore=None): """Insert data as text in the current node, positioned before the start of node insertBefore or to the end of the node's text. :arg data: the data to insert :arg insertBefore: True if you want to insert the text before the node and False if you want to insert it after the node """ raise NotImplementedError def insertBefore(self, node, refNode): """Insert node as a child of the current node, before refNode in the list of child nodes. Raises ValueError if refNode is not a child of the current node :arg node: the node to insert :arg refNode: the child node to insert the node before """ raise NotImplementedError def removeChild(self, node): """Remove node from the children of the current node :arg node: the child node to remove """ raise NotImplementedError def reparentChildren(self, newParent): """Move all the children of the current node to newParent. This is needed so that trees that don't store text as nodes move the text in the correct way :arg newParent: the node to move all this node's children to """ # XXX - should this method be made more general? for child in self.childNodes: newParent.appendChild(child) self.childNodes = [] def cloneNode(self): """Return a shallow copy of the current node i.e. a node with the same name and attributes but with no parent or child nodes """ raise NotImplementedError def hasContent(self): """Return true if the node has children or text, false otherwise """ raise NotImplementedError class ActiveFormattingElements(list): def append(self, node): equalCount = 0 if node != Marker: for element in self[::-1]: if element == Marker: break if self.nodesEqual(element, node): equalCount += 1 if equalCount == 3: self.remove(element) break list.append(self, node) def nodesEqual(self, node1, node2): if not node1.nameTuple == node2.nameTuple: return False if not node1.attributes == node2.attributes: return False return True class TreeBuilder(object): """Base treebuilder implementation * documentClass - the class to use for the bottommost node of a document * elementClass - the class to use for HTML Elements * commentClass - the class to use for comments * doctypeClass - the class to use for doctypes """ # pylint:disable=not-callable # Document class documentClass = None # The class to use for creating a node elementClass = None # The class to use for creating comments commentClass = None # The class to use for creating doctypes doctypeClass = None # Fragment class fragmentClass = None def __init__(self, namespaceHTMLElements): """Create a TreeBuilder :arg namespaceHTMLElements: whether or not to namespace HTML elements """ if namespaceHTMLElements: self.defaultNamespace = "http://www.w3.org/1999/xhtml" else: self.defaultNamespace = None self.reset() def reset(self): self.openElements = [] self.activeFormattingElements = ActiveFormattingElements() # XXX - rename these to headElement, formElement self.headPointer = None self.formPointer = None self.insertFromTable = False self.document = self.documentClass() def elementInScope(self, target, variant=None): # If we pass a node in we match that. if we pass a string # match any node with that name exactNode = hasattr(target, "nameTuple") if not exactNode: if isinstance(target, text_type): target = (namespaces["html"], target) assert isinstance(target, tuple) listElements, invert = listElementsMap[variant] for node in reversed(self.openElements): if exactNode and node == target: return True elif not exactNode and node.nameTuple == target: return True elif (invert ^ (node.nameTuple in listElements)): return False assert False # We should never reach this point def reconstructActiveFormattingElements(self): # Within this algorithm the order of steps described in the # specification is not quite the same as the order of steps in the # code. It should still do the same though. # Step 1: stop the algorithm when there's nothing to do. if not self.activeFormattingElements: return # Step 2 and step 3: we start with the last element. So i is -1. i = len(self.activeFormattingElements) - 1 entry = self.activeFormattingElements[i] if entry == Marker or entry in self.openElements: return # Step 6 while entry != Marker and entry not in self.openElements: if i == 0: # This will be reset to 0 below i = -1 break i -= 1 # Step 5: let entry be one earlier in the list. entry = self.activeFormattingElements[i] while True: # Step 7 i += 1 # Step 8 entry = self.activeFormattingElements[i] clone = entry.cloneNode() # Mainly to get a new copy of the attributes # Step 9 element = self.insertElement({"type": "StartTag", "name": clone.name, "namespace": clone.namespace, "data": clone.attributes}) # Step 10 self.activeFormattingElements[i] = element # Step 11 if element == self.activeFormattingElements[-1]: break def clearActiveFormattingElements(self): entry = self.activeFormattingElements.pop() while self.activeFormattingElements and entry != Marker: entry = self.activeFormattingElements.pop() def elementInActiveFormattingElements(self, name): """Check if an element exists between the end of the active formatting elements and the last marker. If it does, return it, else return false""" for item in self.activeFormattingElements[::-1]: # Check for Marker first because if it's a Marker it doesn't have a # name attribute. if item == Marker: break elif item.name == name: return item return False def insertRoot(self, token): element = self.createElement(token) self.openElements.append(element) self.document.appendChild(element) def insertDoctype(self, token): name = token["name"] publicId = token["publicId"] systemId = token["systemId"] doctype = self.doctypeClass(name, publicId, systemId) self.document.appendChild(doctype) def insertComment(self, token, parent=None): if parent is None: parent = self.openElements[-1] parent.appendChild(self.commentClass(token["data"])) def createElement(self, token): """Create an element but don't insert it anywhere""" name = token["name"] namespace = token.get("namespace", self.defaultNamespace) element = self.elementClass(name, namespace) element.attributes = token["data"] return element def _getInsertFromTable(self): return self._insertFromTable def _setInsertFromTable(self, value): """Switch the function used to insert an element from the normal one to the misnested table one and back again""" self._insertFromTable = value if value: self.insertElement = self.insertElementTable else: self.insertElement = self.insertElementNormal insertFromTable = property(_getInsertFromTable, _setInsertFromTable) def insertElementNormal(self, token): name = token["name"] assert isinstance(name, text_type), "Element %s not unicode" % name namespace = token.get("namespace", self.defaultNamespace) element = self.elementClass(name, namespace) element.attributes = token["data"] self.openElements[-1].appendChild(element) self.openElements.append(element) return element def insertElementTable(self, token): """Create an element and insert it into the tree""" element = self.createElement(token) if self.openElements[-1].name not in tableInsertModeElements: return self.insertElementNormal(token) else: # We should be in the InTable mode. This means we want to do # special magic element rearranging parent, insertBefore = self.getTableMisnestedNodePosition() if insertBefore is None: parent.appendChild(element) else: parent.insertBefore(element, insertBefore) self.openElements.append(element) return element def insertText(self, data, parent=None): """Insert text data.""" if parent is None: parent = self.openElements[-1] if (not self.insertFromTable or (self.insertFromTable and self.openElements[-1].name not in tableInsertModeElements)): parent.insertText(data) else: # We should be in the InTable mode. This means we want to do # special magic element rearranging parent, insertBefore = self.getTableMisnestedNodePosition() parent.insertText(data, insertBefore) def getTableMisnestedNodePosition(self): """Get the foster parent element, and sibling to insert before (or None) when inserting a misnested table node""" # The foster parent element is the one which comes before the most # recently opened table element # XXX - this is really inelegant lastTable = None fosterParent = None insertBefore = None for elm in self.openElements[::-1]: if elm.name == "table": lastTable = elm break if lastTable: # XXX - we should really check that this parent is actually a # node here if lastTable.parent: fosterParent = lastTable.parent insertBefore = lastTable else: fosterParent = self.openElements[ self.openElements.index(lastTable) - 1] else: fosterParent = self.openElements[0] return fosterParent, insertBefore def generateImpliedEndTags(self, exclude=None): name = self.openElements[-1].name # XXX td, th and tr are not actually needed if (name in frozenset(("dd", "dt", "li", "option", "optgroup", "p", "rp", "rt")) and name != exclude): self.openElements.pop() # XXX This is not entirely what the specification says. We should # investigate it more closely. self.generateImpliedEndTags(exclude) def getDocument(self): """Return the final tree""" return self.document def getFragment(self): """Return the final fragment""" # assert self.innerHTML fragment = self.fragmentClass() self.openElements[0].reparentChildren(fragment) return fragment def testSerializer(self, node): """Serialize the subtree of node in the format required by unit tests :arg node: the node from which to start serializing """ raise NotImplementedError site-packages/pip/_vendor/html5lib/treebuilders/etree.py000064400000030734151464452720017420 0ustar00from __future__ import absolute_import, division, unicode_literals # pylint:disable=protected-access from pip._vendor.six import text_type import re from . import base from .. import _ihatexml from .. import constants from ..constants import namespaces from .._utils import moduleFactoryFactory tag_regexp = re.compile("{([^}]*)}(.*)") def getETreeBuilder(ElementTreeImplementation, fullTree=False): ElementTree = ElementTreeImplementation ElementTreeCommentType = ElementTree.Comment("asd").tag class Element(base.Node): def __init__(self, name, namespace=None): self._name = name self._namespace = namespace self._element = ElementTree.Element(self._getETreeTag(name, namespace)) if namespace is None: self.nameTuple = namespaces["html"], self._name else: self.nameTuple = self._namespace, self._name self.parent = None self._childNodes = [] self._flags = [] def _getETreeTag(self, name, namespace): if namespace is None: etree_tag = name else: etree_tag = "{%s}%s" % (namespace, name) return etree_tag def _setName(self, name): self._name = name self._element.tag = self._getETreeTag(self._name, self._namespace) def _getName(self): return self._name name = property(_getName, _setName) def _setNamespace(self, namespace): self._namespace = namespace self._element.tag = self._getETreeTag(self._name, self._namespace) def _getNamespace(self): return self._namespace namespace = property(_getNamespace, _setNamespace) def _getAttributes(self): return self._element.attrib def _setAttributes(self, attributes): # Delete existing attributes first # XXX - there may be a better way to do this... for key in list(self._element.attrib.keys()): del self._element.attrib[key] for key, value in attributes.items(): if isinstance(key, tuple): name = "{%s}%s" % (key[2], key[1]) else: name = key self._element.set(name, value) attributes = property(_getAttributes, _setAttributes) def _getChildNodes(self): return self._childNodes def _setChildNodes(self, value): del self._element[:] self._childNodes = [] for element in value: self.insertChild(element) childNodes = property(_getChildNodes, _setChildNodes) def hasContent(self): """Return true if the node has children or text""" return bool(self._element.text or len(self._element)) def appendChild(self, node): self._childNodes.append(node) self._element.append(node._element) node.parent = self def insertBefore(self, node, refNode): index = list(self._element).index(refNode._element) self._element.insert(index, node._element) node.parent = self def removeChild(self, node): self._childNodes.remove(node) self._element.remove(node._element) node.parent = None def insertText(self, data, insertBefore=None): if not(len(self._element)): if not self._element.text: self._element.text = "" self._element.text += data elif insertBefore is None: # Insert the text as the tail of the last child element if not self._element[-1].tail: self._element[-1].tail = "" self._element[-1].tail += data else: # Insert the text before the specified node children = list(self._element) index = children.index(insertBefore._element) if index > 0: if not self._element[index - 1].tail: self._element[index - 1].tail = "" self._element[index - 1].tail += data else: if not self._element.text: self._element.text = "" self._element.text += data def cloneNode(self): element = type(self)(self.name, self.namespace) for name, value in self.attributes.items(): element.attributes[name] = value return element def reparentChildren(self, newParent): if newParent.childNodes: newParent.childNodes[-1]._element.tail += self._element.text else: if not newParent._element.text: newParent._element.text = "" if self._element.text is not None: newParent._element.text += self._element.text self._element.text = "" base.Node.reparentChildren(self, newParent) class Comment(Element): def __init__(self, data): # Use the superclass constructor to set all properties on the # wrapper element self._element = ElementTree.Comment(data) self.parent = None self._childNodes = [] self._flags = [] def _getData(self): return self._element.text def _setData(self, value): self._element.text = value data = property(_getData, _setData) class DocumentType(Element): def __init__(self, name, publicId, systemId): Element.__init__(self, "") self._element.text = name self.publicId = publicId self.systemId = systemId def _getPublicId(self): return self._element.get("publicId", "") def _setPublicId(self, value): if value is not None: self._element.set("publicId", value) publicId = property(_getPublicId, _setPublicId) def _getSystemId(self): return self._element.get("systemId", "") def _setSystemId(self, value): if value is not None: self._element.set("systemId", value) systemId = property(_getSystemId, _setSystemId) class Document(Element): def __init__(self): Element.__init__(self, "DOCUMENT_ROOT") class DocumentFragment(Element): def __init__(self): Element.__init__(self, "DOCUMENT_FRAGMENT") def testSerializer(element): rv = [] def serializeElement(element, indent=0): if not(hasattr(element, "tag")): element = element.getroot() if element.tag == "": if element.get("publicId") or element.get("systemId"): publicId = element.get("publicId") or "" systemId = element.get("systemId") or "" rv.append("""""" % (element.text, publicId, systemId)) else: rv.append("" % (element.text,)) elif element.tag == "DOCUMENT_ROOT": rv.append("#document") if element.text is not None: rv.append("|%s\"%s\"" % (' ' * (indent + 2), element.text)) if element.tail is not None: raise TypeError("Document node cannot have tail") if hasattr(element, "attrib") and len(element.attrib): raise TypeError("Document node cannot have attributes") elif element.tag == ElementTreeCommentType: rv.append("|%s" % (' ' * indent, element.text)) else: assert isinstance(element.tag, text_type), \ "Expected unicode, got %s, %s" % (type(element.tag), element.tag) nsmatch = tag_regexp.match(element.tag) if nsmatch is None: name = element.tag else: ns, name = nsmatch.groups() prefix = constants.prefixes[ns] name = "%s %s" % (prefix, name) rv.append("|%s<%s>" % (' ' * indent, name)) if hasattr(element, "attrib"): attributes = [] for name, value in element.attrib.items(): nsmatch = tag_regexp.match(name) if nsmatch is not None: ns, name = nsmatch.groups() prefix = constants.prefixes[ns] attr_string = "%s %s" % (prefix, name) else: attr_string = name attributes.append((attr_string, value)) for name, value in sorted(attributes): rv.append('|%s%s="%s"' % (' ' * (indent + 2), name, value)) if element.text: rv.append("|%s\"%s\"" % (' ' * (indent + 2), element.text)) indent += 2 for child in element: serializeElement(child, indent) if element.tail: rv.append("|%s\"%s\"" % (' ' * (indent - 2), element.tail)) serializeElement(element, 0) return "\n".join(rv) def tostring(element): # pylint:disable=unused-variable """Serialize an element and its child nodes to a string""" rv = [] filter = _ihatexml.InfosetFilter() def serializeElement(element): if isinstance(element, ElementTree.ElementTree): element = element.getroot() if element.tag == "": if element.get("publicId") or element.get("systemId"): publicId = element.get("publicId") or "" systemId = element.get("systemId") or "" rv.append("""""" % (element.text, publicId, systemId)) else: rv.append("" % (element.text,)) elif element.tag == "DOCUMENT_ROOT": if element.text is not None: rv.append(element.text) if element.tail is not None: raise TypeError("Document node cannot have tail") if hasattr(element, "attrib") and len(element.attrib): raise TypeError("Document node cannot have attributes") for child in element: serializeElement(child) elif element.tag == ElementTreeCommentType: rv.append("" % (element.text,)) else: # This is assumed to be an ordinary element if not element.attrib: rv.append("<%s>" % (filter.fromXmlName(element.tag),)) else: attr = " ".join(["%s=\"%s\"" % ( filter.fromXmlName(name), value) for name, value in element.attrib.items()]) rv.append("<%s %s>" % (element.tag, attr)) if element.text: rv.append(element.text) for child in element: serializeElement(child) rv.append("" % (element.tag,)) if element.tail: rv.append(element.tail) serializeElement(element) return "".join(rv) class TreeBuilder(base.TreeBuilder): # pylint:disable=unused-variable documentClass = Document doctypeClass = DocumentType elementClass = Element commentClass = Comment fragmentClass = DocumentFragment implementation = ElementTreeImplementation def testSerializer(self, element): return testSerializer(element) def getDocument(self): if fullTree: return self.document._element else: if self.defaultNamespace is not None: return self.document._element.find( "{%s}html" % self.defaultNamespace) else: return self.document._element.find("html") def getFragment(self): return base.TreeBuilder.getFragment(self)._element return locals() getETreeModule = moduleFactoryFactory(getETreeBuilder) site-packages/pip/_vendor/html5lib/treebuilders/etree_lxml.py000064400000033452151464452720020454 0ustar00"""Module for supporting the lxml.etree library. The idea here is to use as much of the native library as possible, without using fragile hacks like custom element names that break between releases. The downside of this is that we cannot represent all possible trees; specifically the following are known to cause problems: Text or comments as siblings of the root element Docypes with no name When any of these things occur, we emit a DataLossWarning """ from __future__ import absolute_import, division, unicode_literals # pylint:disable=protected-access import warnings import re import sys from . import base from ..constants import DataLossWarning from .. import constants from . import etree as etree_builders from .. import _ihatexml import lxml.etree as etree fullTree = True tag_regexp = re.compile("{([^}]*)}(.*)") comment_type = etree.Comment("asd").tag class DocumentType(object): def __init__(self, name, publicId, systemId): self.name = name self.publicId = publicId self.systemId = systemId class Document(object): def __init__(self): self._elementTree = None self._childNodes = [] def appendChild(self, element): self._elementTree.getroot().addnext(element._element) def _getChildNodes(self): return self._childNodes childNodes = property(_getChildNodes) def testSerializer(element): rv = [] infosetFilter = _ihatexml.InfosetFilter(preventDoubleDashComments=True) def serializeElement(element, indent=0): if not hasattr(element, "tag"): if hasattr(element, "getroot"): # Full tree case rv.append("#document") if element.docinfo.internalDTD: if not (element.docinfo.public_id or element.docinfo.system_url): dtd_str = "" % element.docinfo.root_name else: dtd_str = """""" % ( element.docinfo.root_name, element.docinfo.public_id, element.docinfo.system_url) rv.append("|%s%s" % (' ' * (indent + 2), dtd_str)) next_element = element.getroot() while next_element.getprevious() is not None: next_element = next_element.getprevious() while next_element is not None: serializeElement(next_element, indent + 2) next_element = next_element.getnext() elif isinstance(element, str) or isinstance(element, bytes): # Text in a fragment assert isinstance(element, str) or sys.version_info[0] == 2 rv.append("|%s\"%s\"" % (' ' * indent, element)) else: # Fragment case rv.append("#document-fragment") for next_element in element: serializeElement(next_element, indent + 2) elif element.tag == comment_type: rv.append("|%s" % (' ' * indent, element.text)) if hasattr(element, "tail") and element.tail: rv.append("|%s\"%s\"" % (' ' * indent, element.tail)) else: assert isinstance(element, etree._Element) nsmatch = etree_builders.tag_regexp.match(element.tag) if nsmatch is not None: ns = nsmatch.group(1) tag = nsmatch.group(2) prefix = constants.prefixes[ns] rv.append("|%s<%s %s>" % (' ' * indent, prefix, infosetFilter.fromXmlName(tag))) else: rv.append("|%s<%s>" % (' ' * indent, infosetFilter.fromXmlName(element.tag))) if hasattr(element, "attrib"): attributes = [] for name, value in element.attrib.items(): nsmatch = tag_regexp.match(name) if nsmatch is not None: ns, name = nsmatch.groups() name = infosetFilter.fromXmlName(name) prefix = constants.prefixes[ns] attr_string = "%s %s" % (prefix, name) else: attr_string = infosetFilter.fromXmlName(name) attributes.append((attr_string, value)) for name, value in sorted(attributes): rv.append('|%s%s="%s"' % (' ' * (indent + 2), name, value)) if element.text: rv.append("|%s\"%s\"" % (' ' * (indent + 2), element.text)) indent += 2 for child in element: serializeElement(child, indent) if hasattr(element, "tail") and element.tail: rv.append("|%s\"%s\"" % (' ' * (indent - 2), element.tail)) serializeElement(element, 0) return "\n".join(rv) def tostring(element): """Serialize an element and its child nodes to a string""" rv = [] def serializeElement(element): if not hasattr(element, "tag"): if element.docinfo.internalDTD: if element.docinfo.doctype: dtd_str = element.docinfo.doctype else: dtd_str = "" % element.docinfo.root_name rv.append(dtd_str) serializeElement(element.getroot()) elif element.tag == comment_type: rv.append("" % (element.text,)) else: # This is assumed to be an ordinary element if not element.attrib: rv.append("<%s>" % (element.tag,)) else: attr = " ".join(["%s=\"%s\"" % (name, value) for name, value in element.attrib.items()]) rv.append("<%s %s>" % (element.tag, attr)) if element.text: rv.append(element.text) for child in element: serializeElement(child) rv.append("" % (element.tag,)) if hasattr(element, "tail") and element.tail: rv.append(element.tail) serializeElement(element) return "".join(rv) class TreeBuilder(base.TreeBuilder): documentClass = Document doctypeClass = DocumentType elementClass = None commentClass = None fragmentClass = Document implementation = etree def __init__(self, namespaceHTMLElements, fullTree=False): builder = etree_builders.getETreeModule(etree, fullTree=fullTree) infosetFilter = self.infosetFilter = _ihatexml.InfosetFilter(preventDoubleDashComments=True) self.namespaceHTMLElements = namespaceHTMLElements class Attributes(dict): def __init__(self, element, value=None): if value is None: value = {} self._element = element dict.__init__(self, value) # pylint:disable=non-parent-init-called for key, value in self.items(): if isinstance(key, tuple): name = "{%s}%s" % (key[2], infosetFilter.coerceAttribute(key[1])) else: name = infosetFilter.coerceAttribute(key) self._element._element.attrib[name] = value def __setitem__(self, key, value): dict.__setitem__(self, key, value) if isinstance(key, tuple): name = "{%s}%s" % (key[2], infosetFilter.coerceAttribute(key[1])) else: name = infosetFilter.coerceAttribute(key) self._element._element.attrib[name] = value class Element(builder.Element): def __init__(self, name, namespace): name = infosetFilter.coerceElement(name) builder.Element.__init__(self, name, namespace=namespace) self._attributes = Attributes(self) def _setName(self, name): self._name = infosetFilter.coerceElement(name) self._element.tag = self._getETreeTag( self._name, self._namespace) def _getName(self): return infosetFilter.fromXmlName(self._name) name = property(_getName, _setName) def _getAttributes(self): return self._attributes def _setAttributes(self, attributes): self._attributes = Attributes(self, attributes) attributes = property(_getAttributes, _setAttributes) def insertText(self, data, insertBefore=None): data = infosetFilter.coerceCharacters(data) builder.Element.insertText(self, data, insertBefore) def appendChild(self, child): builder.Element.appendChild(self, child) class Comment(builder.Comment): def __init__(self, data): data = infosetFilter.coerceComment(data) builder.Comment.__init__(self, data) def _setData(self, data): data = infosetFilter.coerceComment(data) self._element.text = data def _getData(self): return self._element.text data = property(_getData, _setData) self.elementClass = Element self.commentClass = Comment # self.fragmentClass = builder.DocumentFragment base.TreeBuilder.__init__(self, namespaceHTMLElements) def reset(self): base.TreeBuilder.reset(self) self.insertComment = self.insertCommentInitial self.initial_comments = [] self.doctype = None def testSerializer(self, element): return testSerializer(element) def getDocument(self): if fullTree: return self.document._elementTree else: return self.document._elementTree.getroot() def getFragment(self): fragment = [] element = self.openElements[0]._element if element.text: fragment.append(element.text) fragment.extend(list(element)) if element.tail: fragment.append(element.tail) return fragment def insertDoctype(self, token): name = token["name"] publicId = token["publicId"] systemId = token["systemId"] if not name: warnings.warn("lxml cannot represent empty doctype", DataLossWarning) self.doctype = None else: coercedName = self.infosetFilter.coerceElement(name) if coercedName != name: warnings.warn("lxml cannot represent non-xml doctype", DataLossWarning) doctype = self.doctypeClass(coercedName, publicId, systemId) self.doctype = doctype def insertCommentInitial(self, data, parent=None): assert parent is None or parent is self.document assert self.document._elementTree is None self.initial_comments.append(data) def insertCommentMain(self, data, parent=None): if (parent == self.document and self.document._elementTree.getroot()[-1].tag == comment_type): warnings.warn("lxml cannot represent adjacent comments beyond the root elements", DataLossWarning) super(TreeBuilder, self).insertComment(data, parent) def insertRoot(self, token): # Because of the way libxml2 works, it doesn't seem to be possible to # alter information like the doctype after the tree has been parsed. # Therefore we need to use the built-in parser to create our initial # tree, after which we can add elements like normal docStr = "" if self.doctype: assert self.doctype.name docStr += "= 0 and sysid.find('"') >= 0: warnings.warn("DOCTYPE system cannot contain single and double quotes", DataLossWarning) sysid = sysid.replace("'", 'U00027') if sysid.find("'") >= 0: docStr += '"%s"' % sysid else: docStr += "'%s'" % sysid else: docStr += "''" docStr += ">" if self.doctype.name != token["name"]: warnings.warn("lxml cannot represent doctype with a different name to the root element", DataLossWarning) docStr += "" root = etree.fromstring(docStr) # Append the initial comments: for comment_token in self.initial_comments: comment = self.commentClass(comment_token["data"]) root.addprevious(comment._element) # Create the root document and add the ElementTree to it self.document = self.documentClass() self.document._elementTree = root.getroottree() # Give the root element the right name name = token["name"] namespace = token.get("namespace", self.defaultNamespace) if namespace is None: etree_tag = name else: etree_tag = "{%s}%s" % (namespace, name) root.tag = etree_tag # Add the root element to the internal child/open data structures root_element = self.elementClass(name, namespace) root_element._element = root self.document._childNodes.append(root_element) self.openElements.append(root_element) # Reset to the default insert comment function self.insertComment = self.insertCommentMain site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/base.cpython-38.pyc000064400000026062151464452720023513 0ustar00U .e8@sddlmZmZmZddlmZddlmZmZm Z dZ e edfe ee e ddfgBdfe ee e dd fe dd fgBdfe e ddfe dd fgdfe e dd fe dd fgdfdZ GdddeZGdddeZGdddeZdS))absolute_importdivisionunicode_literals) text_type)scopingElementstableInsertModeElements namespacesNFhtmlbuttonZolZultableoptgroupoptionT)Nr listr Zselectc@sbeZdZdZddZddZddZdd Zdd d Zd dZ ddZ ddZ ddZ ddZ d S)NodezRepresents an item in the treecCs(||_d|_d|_i|_g|_g|_dS)zRCreates a Node :arg name: The tag name associated with the node N)nameparentvalue attributes childNodesZ_flags)selfrrJ/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/base.py__init__s z Node.__init__cCs:ddd|jD}|r,d|j|fSd|jSdS)N cSsg|]\}}d||fqS)z%s="%s"r).0rrrrr .sz Node.__str__..z<%s %s><%s>)joinritemsr)rZ attributesStrrrr__str__-s  z Node.__str__cCs d|jS)Nr)rrrrr__repr__6sz Node.__repr__cCstdS)z[Insert node as a child of the current node :arg node: the node to insert NNotImplementedErrorrnoderrr appendChild9szNode.appendChildNcCstdS)aBInsert data as text in the current node, positioned before the start of node insertBefore or to the end of the node's text. :arg data: the data to insert :arg insertBefore: True if you want to insert the text before the node and False if you want to insert it after the node Nr#)rdata insertBeforerrr insertTextAs zNode.insertTextcCstdS)aInsert node as a child of the current node, before refNode in the list of child nodes. Raises ValueError if refNode is not a child of the current node :arg node: the node to insert :arg refNode: the child node to insert the node before Nr#)rr&ZrefNoderrrr)Ms zNode.insertBeforecCstdS)zhRemove node from the children of the current node :arg node: the child node to remove Nr#r%rrr removeChildYszNode.removeChildcCs |jD]}||qg|_dS)zMove all the children of the current node to newParent. This is needed so that trees that don't store text as nodes move the text in the correct way :arg newParent: the node to move all this node's children to N)rr')rZ newParentZchildrrrreparentChildrenas  zNode.reparentChildrencCstdS)zReturn a shallow copy of the current node i.e. a node with the same name and attributes but with no parent or child nodes Nr#r!rrr cloneNodenszNode.cloneNodecCstdS)zFReturn true if the node has children or text, false otherwise Nr#r!rrr hasContenttszNode.hasContent)N)__name__ __module__ __qualname____doc__rr r"r'r*r)r+r,r-r.rrrrrs    rc@seZdZddZddZdS)ActiveFormattingElementscCsfd}|tkrV|dddD]:}|tkr*qV|||r>|d7}|dkr||qVqt||dS)Nr)Marker nodesEqualremoverappend)rr&Z equalCountelementrrrr:{s  zActiveFormattingElements.appendcCs$|j|jksdS|j|jks dSdS)NFT) nameTupler)rZnode1Znode2rrrr8s   z#ActiveFormattingElements.nodesEqualN)r/r0r1r:r8rrrrr3zs r3c@seZdZdZdZdZdZdZdZddZ ddZ d+ddZ d d Z d d Z d dZddZddZd,ddZddZddZddZeeeZddZddZd-dd Zd!d"Zd.d#d$Zd%d&Zd'd(Zd)d*ZdS)/ TreeBuilderaBase treebuilder implementation * documentClass - the class to use for the bottommost node of a document * elementClass - the class to use for HTML Elements * commentClass - the class to use for comments * doctypeClass - the class to use for doctypes NcCs|r d|_nd|_|dS)zmCreate a TreeBuilder :arg namespaceHTMLElements: whether or not to namespace HTML elements zhttp://www.w3.org/1999/xhtmlN)defaultNamespacereset)rZnamespaceHTMLElementsrrrrszTreeBuilder.__init__cCs.g|_t|_d|_d|_d|_||_dS)NF) openElementsr3activeFormattingElementsZ headPointerZ formPointerinsertFromTable documentClassdocumentr!rrrr?s zTreeBuilder.resetcCst|d}|s2t|tr$td|f}t|ts2tt|\}}t|jD]>}|r^||kr^dS|sr|j |krrdS||j |kArHdSqHdstdS)Nr<r TF) hasattr isinstancerr tupleAssertionErrorlistElementsMapreversedr@r<)rtargetZvariantZ exactNodeZ listElementsinvertr&rrrelementInScopes     zTreeBuilder.elementInScopecCs|js dSt|jd}|j|}|tks4||jkr8dS|tkrl||jkrl|dkrXd}ql|d8}|j|}q8|d7}|j|}|}|d|j|j|jd}||j|<||jdkrlqqldS)Nr5rr4ZStartTag)typer namespacer() rAlenr7r@r- insertElementrrOr)rientryZcloner;rrr#reconstructActiveFormattingElementss.    z/TreeBuilder.reconstructActiveFormattingElementscCs(|j}|jr$|tkr$|j}q dSN)rApopr7)rrSrrrclearActiveFormattingElementss z)TreeBuilder.clearActiveFormattingElementscCs:|jdddD]$}|tkr"q6q|j|kr|SqdS)zCheck if an element exists between the end of the active formatting elements and the last marker. If it does, return it, else return falseNr4F)rAr7r)rritemrrr!elementInActiveFormattingElements s   z-TreeBuilder.elementInActiveFormattingElementscCs&||}|j||j|dSrU) createElementr@r:rDr')rtokenr;rrr insertRoots  zTreeBuilder.insertRootcCs6|d}|d}|d}||||}|j|dS)NrpublicIdsystemId) doctypeClassrDr')rr[rr]r^Zdoctyperrr insertDoctype s zTreeBuilder.insertDoctypecCs*|dkr|jd}|||ddS)Nr4r()r@r' commentClass)rr[rrrr insertComment(s zTreeBuilder.insertCommentcCs0|d}|d|j}|||}|d|_|S)z.Create an element but don't insert it anywhererrOr()getr> elementClassrrr[rrOr;rrrrZ-s   zTreeBuilder.createElementcCs|jSrU)_insertFromTabler!rrr_getInsertFromTable5szTreeBuilder._getInsertFromTablecCs ||_|r|j|_n|j|_dS)zsSwitch the function used to insert an element from the normal one to the misnested table one and back againN)rfinsertElementTablerQinsertElementNormal)rrrrr_setInsertFromTable8s zTreeBuilder._setInsertFromTablecCsb|d}t|tstd||d|j}|||}|d|_|jd||j ||S)NrzElement %s not unicoderOr(r4) rFrrHrcr>rdrr@r'r:rerrrriCs   zTreeBuilder.insertElementNormalcCs`||}|jdjtkr$||S|\}}|dkrD||n ||||j||S)z-Create an element and insert it into the treer4N) rZr@rrrigetTableMisnestedNodePositionr'r)r:)rr[r;rr)rrrrhMs      zTreeBuilder.insertElementTablecCsV|dkr|jd}|jr.|jr:|jdjtkr:||n|\}}|||dS)zInsert text data.Nr4)r@rBrrr*rk)rr(rr)rrrr*]s     zTreeBuilder.insertTextcCstd}d}d}|jdddD]}|jdkr|}q4q|rb|jrJ|j}|}ql|j|j|d}n |jd}||fS)zsGet the foster parent element, and sibling to insert before (or None) when inserting a misnested table nodeNr4r r5r)r@rrindex)rZ lastTableZ fosterParentr)Zelmrrrrkls   z)TreeBuilder.getTableMisnestedNodePositioncCs8|jdj}|tdkr4||kr4|j||dS)Nr4)ZddZdtZlirr pZrpZrt)r@r frozensetrVgenerateImpliedEndTags)rexcluderrrrros    z"TreeBuilder.generateImpliedEndTagscCs|jS)zReturn the final tree)rDr!rrr getDocumentszTreeBuilder.getDocumentcCs|}|jd||S)zReturn the final fragmentr) fragmentClassr@r,)rZfragmentrrr getFragmentszTreeBuilder.getFragmentcCstdS)zSerialize the subtree of node in the format required by unit tests :arg node: the node from which to start serializing Nr#r%rrrtestSerializerszTreeBuilder.testSerializer)N)N)N)N)r/r0r1r2rCrdrar_rrrr?rMrTrWrYr\r`rbrZrgrjpropertyrBrirhr*rkrorqrsrtrrrrr=s6   .     r=)Z __future__rrrZpip._vendor.sixrZ constantsrrr r7rnsetrIobjectrrr3r=rrrrs0        csite-packages/pip/_vendor/html5lib/treebuilders/__pycache__/etree_lxml.cpython-38.pyc000064400000027000151464452720024732 0ustar00U .e*7@sdZddlmZmZmZddlZddlZddlZddlm Z ddl m Z ddlm Z dd lm Z dd lmZddlm Z d Zed Ze d jZGdddeZGdddeZddZddZGddde jZdS)aModule for supporting the lxml.etree library. The idea here is to use as much of the native library as possible, without using fragile hacks like custom element names that break between releases. The downside of this is that we cannot represent all possible trees; specifically the following are known to cause problems: Text or comments as siblings of the root element Docypes with no name When any of these things occur, we emit a DataLossWarning )absolute_importdivisionunicode_literalsN)base)DataLossWarning) constants)etree) _ihatexmlTz {([^}]*)}(.*)Zasdc@seZdZddZdS) DocumentTypecCs||_||_||_dSN)namepublicIdsystemId)selfrrrrP/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/etree_lxml.py__init__#szDocumentType.__init__N)__name__ __module__ __qualname__rrrrrr "sr c@s,eZdZddZddZddZeeZdS)DocumentcCsd|_g|_dSr ) _elementTree _childNodesrrrrr*szDocument.__init__cCs|j|jdSr )rgetrootZaddnext_elementrelementrrr appendChild.szDocument.appendChildcCs|jSr )rrrrr_getChildNodes1szDocument._getChildNodesN)rrrrr r!propertyZ childNodesrrrrr)srcs6gtjdddfdd |ddS)NTZpreventDoubleDashCommentsrc s,t|ds,t|drd|jjrz|jjsF|jjsFd|jj}nd|jj|jj|jjf}dd|d|f|}|dk r|}q|dk r||d| }qnnt |t st |t rt |t st jd dkstd d||fn$d |D]}||dqn|jtkrd d||jft|d r(|jr(d d||jfnt |tjsttj|j}|dk r|d}|d}tj|}dd|||fndd||jft|drg}|jD]d\} } t| }|dk rj|\}} | } tj|}d|| f} n | } || | fq t |D]&\} } dd|d| | fq|jrڈd d|d|jf|d7}|D]} | |qt|d r(|jr(d d|d|jfdS)Ntagrz #document zz|%s%s rrz|%s"%s"z#document-fragmentz|%stailrz |%s<%s %s>z|%s<%s>attribz%s %sz |%s%s="%s")!hasattrappenddocinfo internalDTDZ public_idZ system_url root_namerZ getpreviousZgetnext isinstancestrbytessys version_infoAssertionErrorr$ comment_typetextr'r Z_Elementetree_builders tag_regexpmatchgroupr prefixes fromXmlNamer(itemsgroupssorted) rindentdtd_strZ next_elementZnsmatchnsr$prefix attributesrvalueZ attr_stringchild infosetFilterrvserializeElementrrrI;s|                       z(testSerializer..serializeElement )r)r InfosetFilterjoinrrrFrtestSerializer7s  F rNcs$gfdd|dS)z4Serialize an element and its child nodes to a stringcst|dsH|jjr:|jjr$|jj}n d|jj}||n|jtkrfd|j fn~|j sd|jfn.d dd|j D}d|j|f|j r|j |D] }|qĈd |jft|d r|j r|j dS) Nr$r%z z<%s>r&cSsg|]\}}d||fqS)z%s="%s"r).0rrDrrr sz6tostring..serializeElement..z<%s %s>zr')r)r+r,doctyper-r*rr$r4r5r(rLr<r')rr@attrrErHrIrrrIs,        z"tostring..serializeElement)rLrMrrSrtostrings rUcszeZdZeZeZdZdZeZ e Z dddZ ddZ ddZd d Zd d Zd dZdddZdfdd ZddZZS) TreeBuilderNFcstjt|dtjdd|_||_GfdddtGfdddj}Gfdd d j }||_ ||_ t j ||dS) N)fullTreeTr#cs&eZdZdfdd ZfddZdS)z(TreeBuilder.__init__..AttributesNcsr|dkr i}||_t|||D]F\}}t|trTd|d|df}n |}||jjj|<q&dSN{%s}%srr)rdictrr<r.tuplecoerceAttributer()rrrDkeyrrGrrrs   z1TreeBuilder.__init__..Attributes.__init__csPt|||t|tr4d|d|df}n |}||jjj|<dSrX)rZ __setitem__r.r[r\rr()rr]rDrr^rrr_s   z4TreeBuilder.__init__..Attributes.__setitem__)N)rrrrr_rr^rr Attributess r`csxeZdZfddZfddZfddZeeeZddZfd d Z eee Z dfd d Z fddZ d S)z%TreeBuilder.__init__..Elementcs*|}jj|||d||_dS)N) namespace) coerceElementElementr _attributes)rrrar`builderrGrrrs z.TreeBuilder.__init__..Element.__init__cs$||_||j|j|j_dSr )rb_nameZ _getETreeTagZ _namespacerr$)rrr^rr_setNames  z.TreeBuilder.__init__..Element._setNamecs |jSr )r;rgrr^rr_getNamesz.TreeBuilder.__init__..Element._getNamecSs|jSr rdrrrr_getAttributessz4TreeBuilder.__init__..Element._getAttributescs|||_dSr rj)rrC)r`rr_setAttributessz4TreeBuilder.__init__..Element._setAttributesNcs|}j|||dSr )ZcoerceCharactersrc insertText)rdataZ insertBeforerfrGrrrms z0TreeBuilder.__init__..Element.insertTextcsj||dSr )rcr )rrE)rfrrr sz1TreeBuilder.__init__..Element.appendChild)N) rrrrrhrir"rrkrlrCrmr rrerrrcs     rccs8eZdZfddZfddZddZeeeZdS)z%TreeBuilder.__init__..Commentcs|}j||dSr ) coerceCommentCommentrrrnrorrrs z.TreeBuilder.__init__..Comment.__init__cs|}||j_dSr )rprr5rrr^rr_setDatas z.TreeBuilder.__init__..Comment._setDatacSs|jjSr )rr5rrrr_getDatasz.TreeBuilder.__init__..Comment._getDataN)rrrrrsrtr"rnrrorrrqs rq)r6ZgetETreeModuler r rKrGnamespaceHTMLElementsrZrcrq elementClass commentClassrrVr)rrurWrcrqrrerrszTreeBuilder.__init__cCs$tj||j|_g|_d|_dSr )rrVresetinsertCommentInitial insertCommentinitial_commentsrQrrrrrxs zTreeBuilder.resetcCst|Sr )rNrrrrrN szTreeBuilder.testSerializercCstr |jjS|jjSdSr )rWdocumentrrrrrr getDocument szTreeBuilder.getDocumentcCsFg}|jdj}|jr"||j|t||jrB||j|S)Nr) openElementsrr5r*extendlistr')rZfragmentrrrr getFragments   zTreeBuilder.getFragmentcCsh|d}|d}|d}|s0tdtd|_n4|j|}||krPtdt||||}||_dS)Nrrrz#lxml cannot represent empty doctypez%lxml cannot represent non-xml doctype)warningswarnrrQrGrb doctypeClass)rtokenrrrZ coercedNamerQrrr insertDoctypes   zTreeBuilder.insertDoctypecCs6|dks||jkst|jjdks&t|j|dSr )r|r3rr{r*rrnparentrrrry,sz TreeBuilder.insertCommentInitialcsB||jkr,|jjdjtkr,tdttt | ||dS)Nz@lxml cannot represent adjacent comments beyond the root elements) r|rrr$r4rrrsuperrVrzr __class__rrinsertCommentMain1s   zTreeBuilder.insertCommentMainc Csd}|jr|jjst|d|jj7}|jjdk s>|jjdk r|d|j|jjpRd7}|jjr|jj}|ddkr|ddkrt dt | dd}|ddkr|d |7}q|d |7}n|d 7}|d 7}|jj|d krt dt |d7}t |}|jD] }||d}||jq ||_||j_|d }|d|j}|dkrj|} n d||f} | |_|||} || _|jj| |j| |j|_dS)NrTz rzGlxml cannot represent doctype with a different name to the root elementz$rnrarY)rQrr3rrrGZ coercePubidfindrrrreplacer Z fromstringr{rwZ addpreviousr documentClassr|Z getroottreergetZdefaultNamespacer$rvrr*r~rrz) rrZdocStrZsysidroot comment_tokenZcommentrraZ etree_tagZ root_elementrrr insertRoot7sP              zTreeBuilder.insertRoot)F)N)N)rrrrrr rrvrwZ fragmentClassr implementationrrxrNr}rrryrr __classcell__rrrrrVs L  rV)__doc__Z __future__rrrrrer1rTrr rr r6r Z lxml.etreerWcompiler7rqr$r4objectr rrNrUrVrrrrs$         O)site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/__init__.cpython-38.pyc000064400000006330151464452720024334 0ustar00U .e@s6dZddlmZmZmZddlmZiZdddZdS) a)A collection of modules for building different kinds of trees from HTML documents. To create a treebuilder for a new type of tree, you need to do implement several things: 1. A set of classes for various types of elements: Document, Doctype, Comment, Element. These must implement the interface of ``base.treebuilders.Node`` (although comment nodes have a different signature for their constructor, see ``treebuilders.etree.Comment``) Textual content may also be implemented as another node type, or not, as your tree implementation requires. 2. A treebuilder object (called ``TreeBuilder`` by convention) that inherits from ``treebuilders.base.TreeBuilder``. This has 4 required attributes: * ``documentClass`` - the class to use for the bottommost node of a document * ``elementClass`` - the class to use for HTML Elements * ``commentClass`` - the class to use for comments * ``doctypeClass`` - the class to use for doctypes It also has one required method: * ``getDocument`` - Returns the root node of the complete document tree 3. If you wish to run the unit tests, you must also create a ``testSerializer`` method on your treebuilder which accepts a node and returns a string containing Node and its children serialized according to the format used in the unittests )absolute_importdivisionunicode_literals) default_etreeNcKs|}|tkr|dkrLddlm}|dkr>> from html5lib.treebuilders import getTreeBuilder >>> builder = getTreeBuilder('etree') dom)rNr)minidomZlxml) etree_lxmletree)r zUnrecognised treebuilder "%s" )lowertreeBuilderCacherZxml.domr Z getDomModuleZ TreeBuilderr r rZgetETreeModule ValueErrorget)ZtreeTypeimplementationkwargsrr r r rN/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/__init__.pygetTreeBuilder's$      r)N) __doc__Z __future__rrrZ_utilsrr rrrrrs site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/etree.cpython-38.opt-1.pyc000064400000026667151464452720024657 0ustar00U .e1@sddlmZmZmZddlmZddlZddlmZddlm Z ddlm Z dd l m Z dd l m Z ed Zdd dZe eZdS))absolute_importdivisionunicode_literals) text_typeN)base) _ihatexml) constants) namespaces)moduleFactoryFactoryz {([^}]*)}(.*)Fc sdjGfdddtjGfdddGfdddGfdd d Gfd d d fd d  fdd}G fdddtj}tS)NZasdcseZdZd$fdd ZddZddZdd ZeeeZd d Z d d Z ee e Z ddZ ddZ ee e ZddZddZeeeZddZddZddZddZd%ddZd d!Zd"d#ZdS)&z getETreeBuilder..ElementNcs^||_||_||||_|dkr:td|jf|_n|j|jf|_d|_g|_g|_ dS)Nhtml) _name _namespaceElement _getETreeTag_elementr Z nameTupleparent _childNodes_flags)selfname namespace ElementTreeK/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/etree.py__init__s z)getETreeBuilder..Element.__init__cSs|dkr|}n d||f}|S)N{%s}%sr)rrrZ etree_tagrrrr#s z-getETreeBuilder..Element._getETreeTagcSs||_||j|j|j_dSN)rrrrtag)rrrrr_setName*sz)getETreeBuilder..Element._setNamecSs|jSr)rrrrr_getName.sz)getETreeBuilder..Element._getNamecSs||_||j|j|j_dSr)rrrrr )rrrrr _setNamespace3sz.getETreeBuilder..Element._setNamespacecSs|jSr)rr"rrr _getNamespace7sz.getETreeBuilder..Element._getNamespacecSs|jjSr)rattribr"rrr_getAttributes<sz/getETreeBuilder..Element._getAttributescSsht|jjD]}|jj|=q|D]:\}}t|trPd|d|df}n|}|j||q(dS)Nrrr)listrr&keysitems isinstancetupleset)r attributeskeyvaluerrrr_setAttributes?s  z/getETreeBuilder..Element._setAttributescSs|jSr)rr"rrr_getChildNodesMsz/getETreeBuilder..Element._getChildNodescSs*|jdd=g|_|D]}||qdSr)rrZ insertChild)rr0elementrrr_setChildNodesPs z/getETreeBuilder..Element._setChildNodescSst|jjpt|jS)z,Return true if the node has children or text)boolrtextlenr"rrr hasContentXsz+getETreeBuilder..Element.hasContentcSs$|j||j|j||_dSr)rappendrrrnoderrr appendChild\s z,getETreeBuilder..Element.appendChildcSs,t|j|j}|j||j||_dSr)r(rindexinsertr)rr;ZrefNoder=rrr insertBeforeasz-getETreeBuilder..Element.insertBeforecSs$|j||j|jd|_dSr)rremoverrr:rrr removeChildfs z,getETreeBuilder..Element.removeChildcSst|js,|jjsd|j_|jj|7_n|dkrb|jdjsLd|jd_|jdj|7_nxt|j}||j}|dkr|j|djsd|j|d_|j|dj|7_n |jjsd|j_|jj|7_dS)Nrr)r7rr6tailr(r=)rdatar?Zchildrenr=rrr insertTextks"     z+getETreeBuilder..Element.insertTextcSs4t||j|j}|jD]\}}||j|<q|Sr)typerrr.r*)rr3rr0rrr cloneNodes z*getETreeBuilder..Element.cloneNodecSsl|jr"|jdjj|jj7_n0|jjs2d|j_|jjdk rR|jj|jj7_d|j_tj||dS)NrCrB) childNodesrrDr6rNodereparentChildren)rZ newParentrrrrKs z1getETreeBuilder..Element.reparentChildren)N)N)__name__ __module__ __qualname__rrr!r#propertyrr$r%rr'r1r.r2r4rIr8r<r?rArFrHrKrrrrrs*      rcs2eZdZfddZddZddZeeeZdS)z getETreeBuilder..Commentcs"||_d|_g|_g|_dSr)Commentrrrr)rrErrrrs z)getETreeBuilder..Comment.__init__cSs|jjSrrr6r"rrr_getDatasz)getETreeBuilder..Comment._getDatacSs ||j_dSrrQrr0rrr_setDatasz)getETreeBuilder..Comment._setDataN)rLrMrNrrRrTrOrErrrrrPs rPcsLeZdZfddZddZddZeeeZddZd d Z eee Z d S) z%getETreeBuilder..DocumentTypecs$|d||j_||_||_dS)N )rrr6publicIdsystemId)rrrVrWrrrrs z.getETreeBuilder..DocumentType.__init__cSs|jddS)NrVrBrgetr"rrr _getPublicIdsz2getETreeBuilder..DocumentType._getPublicIdcSs|dk r|jd|dS)NrVrr-rSrrr _setPublicIdsz2getETreeBuilder..DocumentType._setPublicIdcSs|jddS)NrWrBrYr"rrr _getSystemIdsz2getETreeBuilder..DocumentType._getSystemIdcSs|dk r|jd|dS)NrWr\rSrrr _setSystemIdsz2getETreeBuilder..DocumentType._setSystemIdN) rLrMrNrr[r]rOrVr^r_rWrrXrr DocumentTypes  r`cseZdZfddZdS)z!getETreeBuilder..Documentcs|ddS)N DOCUMENT_ROOTrr"rXrrrsz*getETreeBuilder..Document.__init__NrLrMrNrrrXrrDocumentsrdcseZdZfddZdS)z)getETreeBuilder..DocumentFragmentcs|ddS)NZDOCUMENT_FRAGMENTrbr"rXrrrsz2getETreeBuilder..DocumentFragment.__init__NrcrrXrrDocumentFragmentsrecs*gdfdd |ddS)Nrc snt|ds|}|jdkrz|ds0|drd|dp raz #documentz|%s"%s" rDocument node cannot have tailr&$Document node cannot have attributesz|%sz%s %sz|%s<%s>z |%s%s="%s")hasattrgetrootr rZr9r6rD TypeErrorr7r& tag_regexpmatchgroupsr prefixesr*sorted) r3indentrVrWZnsmatchrnsprefixr.r0Z attr_stringchild)ElementTreeCommentTypervserializeElementrrrxs\                    zAgetETreeBuilder..testSerializer..serializeElement )r)joinr3)rv)rwrxrtestSerializers7 z'getETreeBuilder..testSerializercs2gtfdd|dS)z4Serialize an element and its child nodes to a stringcst|jr|}|jdkr||ds2|drf|dp>d}|dpLd}d|j||fnd|jfn|jdkr|jdk r|j|jdk rtdt |d rt |j rtd |D] }|qn|jkrd |jfn|j sd  |jfn2d fdd|j D}d|j|f|jrd|j|D]}|qhd|jf|jr|jdS)NrUrVrWrBzrfrarhr&riz z<%s>rgcs"g|]\}}d||fqS)z%s="%s") fromXmlName).0rr0)filterrr &s zOgetETreeBuilder..tostring..serializeElement..z<%s %s>z)r+rrkr rZr9r6rDrlrjr7r&r}rzr*)r3rVrWruattr)rrvrrwrxrrrxsD           z;getETreeBuilder..tostring..serializeElementrB)r Z InfosetFilterrzr{)rrv)rrwrxrtostrings -z!getETreeBuilder..tostringcsDeZdZZZZZZZfddZ fddZ ddZ dS)z$getETreeBuilder..TreeBuildercs|Srr)rr3)r|rrr|Asz3getETreeBuilder..TreeBuilder.testSerializercs<r |jjS|jdk r*|jjd|jS|jjdSdS)Nz{%s}htmlr )ZdocumentrZdefaultNamespacefindr")fullTreerr getDocumentDs z0getETreeBuilder..TreeBuilder.getDocumentcSstj|jSr)r TreeBuilder getFragmentrr"rrrrNsz0getETreeBuilder..TreeBuilder.getFragmentN) rLrMrNZ documentClassZ doctypeClassZ elementClassZ commentClassZ fragmentClassimplementationr|rrr)rPrdrer`rElementTreeImplementationrr|rrr9s  r)rPr rrJrlocals)rrrrr) rPrdrer`rrrvrrr|rgetETreeBuilders ~ >6$r)F)Z __future__rrrZpip._vendor.sixrrerBrr r r Z_utilsr compilermrZgetETreeModulerrrrs        Esite-packages/pip/_vendor/html5lib/treebuilders/__pycache__/etree_lxml.cpython-38.opt-1.pyc000064400000026507151464452720025704 0ustar00U .e*7@sdZddlmZmZmZddlZddlZddlZddlm Z ddl m Z ddlm Z dd lm Z dd lmZddlm Z d Zed Ze d jZGdddeZGdddeZddZddZGddde jZdS)aModule for supporting the lxml.etree library. The idea here is to use as much of the native library as possible, without using fragile hacks like custom element names that break between releases. The downside of this is that we cannot represent all possible trees; specifically the following are known to cause problems: Text or comments as siblings of the root element Docypes with no name When any of these things occur, we emit a DataLossWarning )absolute_importdivisionunicode_literalsN)base)DataLossWarning) constants)etree) _ihatexmlTz {([^}]*)}(.*)Zasdc@seZdZddZdS) DocumentTypecCs||_||_||_dSN)namepublicIdsystemId)selfrrrrP/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/etree_lxml.py__init__#szDocumentType.__init__N)__name__ __module__ __qualname__rrrrrr "sr c@s,eZdZddZddZddZeeZdS)DocumentcCsd|_g|_dSr ) _elementTree _childNodesrrrrr*szDocument.__init__cCs|j|jdSr )rgetrootZaddnext_elementrelementrrr appendChild.szDocument.appendChildcCs|jSr )rrrrr_getChildNodes1szDocument._getChildNodesN)rrrrr r!propertyZ childNodesrrrrr)srcs6gtjdddfdd |ddS)NTZpreventDoubleDashCommentsrc st|ds t|drd|jjrz|jjsF|jjsFd|jj}nd|jj|jj|jjf}dd|d|f|}|dk r|}q|dk r||d| }qnNt |t st |t rd d||fn"d |D]}||dqn|j tkr`d d||jft|d r|jrd d||jfntj|j }|dk r|d }|d}tj|}dd|||fndd||j ft|drg}|jD]d\} } t| }|dk r8|\}} | } tj|}d|| f} n | } || | fqt|D]&\} } dd|d| | fq\|jrd d|d|jf|d7}|D]} | |qt|d r|jrd d|d|jfdS)Ntagrz #document zz|%s%s rz|%s"%s"z#document-fragmentz|%stailrz |%s<%s %s>z|%s<%s>attribz%s %sz |%s%s="%s")hasattrappenddocinfo internalDTDZ public_idZ system_url root_namerZ getpreviousZgetnext isinstancestrbytesr$ comment_typetextr'etree_builders tag_regexpmatchgroupr prefixes fromXmlNamer(itemsgroupssorted) rindentdtd_strZ next_elementZnsmatchnsr$prefix attributesrvalueZ attr_stringchild infosetFilterrvserializeElementrrrF;sx                       z(testSerializer..serializeElement )r)r InfosetFilterjoinrrrCrtestSerializer7s  F rKcs$gfdd|dS)z4Serialize an element and its child nodes to a stringcst|dsH|jjr:|jjr$|jj}n d|jj}||n|jtkrfd|j fn~|j sd|jfn.d dd|j D}d|j|f|j r|j |D] }|qĈd |jft|d r|j r|j dS) Nr$r%z z<%s>r&cSsg|]\}}d||fqS)z%s="%s"r).0rrArrr sz6tostring..serializeElement..z<%s %s>zr')r)r+r,doctyper-r*rr$r1r2r(rIr9r')rr=attrrBrErFrrrFs,        z"tostring..serializeElement)rIrJrrPrtostrings rRcszeZdZeZeZdZdZeZ e Z dddZ ddZ ddZd d Zd d Zd dZdddZdfdd ZddZZS) TreeBuilderNFcstjt|dtjdd|_||_GfdddtGfdddj}Gfdd d j }||_ ||_ t j ||dS) N)fullTreeTr#cs&eZdZdfdd ZfddZdS)z(TreeBuilder.__init__..AttributesNcsr|dkr i}||_t|||D]F\}}t|trTd|d|df}n |}||jjj|<q&dSN{%s}%srr)rdictrr9r.tuplecoerceAttributer()rrrAkeyrrDrrrs   z1TreeBuilder.__init__..Attributes.__init__csPt|||t|tr4d|d|df}n |}||jjj|<dSrU)rW __setitem__r.rXrYrr()rrZrArr[rrr\s   z4TreeBuilder.__init__..Attributes.__setitem__)N)rrrrr\rr[rr Attributess r]csxeZdZfddZfddZfddZeeeZddZfd d Z eee Z dfd d Z fddZ d S)z%TreeBuilder.__init__..Elementcs*|}jj|||d||_dS)N) namespace) coerceElementElementr _attributes)rrr^r]builderrDrrrs z.TreeBuilder.__init__..Element.__init__cs$||_||j|j|j_dSr )r__nameZ _getETreeTagZ _namespacerr$)rrr[rr_setNames  z.TreeBuilder.__init__..Element._setNamecs |jSr )r8rdrr[rr_getNamesz.TreeBuilder.__init__..Element._getNamecSs|jSr rarrrr_getAttributessz4TreeBuilder.__init__..Element._getAttributescs|||_dSr rg)rr@)r]rr_setAttributessz4TreeBuilder.__init__..Element._setAttributesNcs|}j|||dSr )ZcoerceCharactersr` insertText)rdataZ insertBeforercrDrrrjs z0TreeBuilder.__init__..Element.insertTextcsj||dSr )r`r )rrB)rcrrr sz1TreeBuilder.__init__..Element.appendChild)N) rrrrrerfr"rrhrir@rjr rrbrrr`s     r`cs8eZdZfddZfddZddZeeeZdS)z%TreeBuilder.__init__..Commentcs|}j||dSr ) coerceCommentCommentrrrkrlrrrs z.TreeBuilder.__init__..Comment.__init__cs|}||j_dSr )rmrr2ror[rr_setDatas z.TreeBuilder.__init__..Comment._setDatacSs|jjSr )rr2rrrr_getDatasz.TreeBuilder.__init__..Comment._getDataN)rrrrrprqr"rkrrlrrrns rn)r3ZgetETreeModuler r rHrDnamespaceHTMLElementsrWr`rn elementClass commentClassrrSr)rrrrTr`rnrrbrrszTreeBuilder.__init__cCs$tj||j|_g|_d|_dSr )rrSresetinsertCommentInitial insertCommentinitial_commentsrNrrrrrus zTreeBuilder.resetcCst|Sr )rKrrrrrK szTreeBuilder.testSerializercCstr |jjS|jjSdSr )rTdocumentrrrrrr getDocument szTreeBuilder.getDocumentcCsFg}|jdj}|jr"||j|t||jrB||j|S)Nr) openElementsrr2r*extendlistr')rZfragmentrrrr getFragments   zTreeBuilder.getFragmentcCsh|d}|d}|d}|s0tdtd|_n4|j|}||krPtdt||||}||_dS)Nrrrz#lxml cannot represent empty doctypez%lxml cannot represent non-xml doctype)warningswarnrrNrDr_ doctypeClass)rtokenrrrZ coercedNamerNrrr insertDoctypes   zTreeBuilder.insertDoctypecCs|j|dSr )rxr*rrkparentrrrrv,sz TreeBuilder.insertCommentInitialcsB||jkr,|jjdjtkr,tdttt | ||dS)Nz@lxml cannot represent adjacent comments beyond the root elements) ryrrr$r1rrrsuperrSrwr __class__rrinsertCommentMain1s   zTreeBuilder.insertCommentMainc Csd}|jr|d|jj7}|jjdk s2|jjdk r|d|j|jjpFd7}|jjr|jj}|ddkr|ddkrtdt | dd}|ddkr|d |7}q|d |7}n|d 7}|d 7}|jj|d krtdt |d7}t |}|j D] }||d}||jq||_||j_|d }|d|j}|dkr^|} n d||f} | |_|||} || _|jj| |j| |j|_dS)NrQz rzGlxml cannot represent doctype with a different name to the root elementz$rkr^rV)rNrrrrDZ coercePubidfindrrrreplacer Z fromstringrxrtZ addpreviousr documentClassryZ getroottreergetZdefaultNamespacer$rsrr*r{rrw) rrZdocStrZsysidroot comment_tokenZcommentrr^Z etree_tagZ root_elementrrr insertRoot7sN             zTreeBuilder.insertRoot)F)N)N)rrrrrr rrsrtZ fragmentClassr implementationrrurKrzr~rrvrr __classcell__rrrrrSs L  rS)__doc__Z __future__rrrrresysrQrr rr r3r Z lxml.etreerTcompiler4rnr$r1objectr rrKrRrSrrrrs$         O)site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/dom.cpython-38.opt-1.pyc000064400000022315151464452720024314 0ustar00U .e"@sddlmZmZmZzddlmZWn ek rDddlmZYnXddlm Z m Z ddl Z ddl m Z ddl mZdd lmZdd lmZd d ZeeZdS) )absolute_importdivisionunicode_literals)MutableMapping)minidomNodeN)base) constants) namespaces)moduleFactoryFactorycsVGdddtGfdddtjGfdddtj}ddtS) Nc@sLeZdZddZddZddZddZd d Zd d Zd dZ ddZ dS)zgetDomBuilder..AttrListcSs ||_dSNelementselfrrI/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/dom.py__init__sz(getDomBuilder..AttrList.__init__cSst|jjSr)iterr attributeskeysrrrr__iter__sz(getDomBuilder..AttrList.__iter__cSs4t|trtn |jj|}||_||jj|<dSr) isinstancetupleNotImplementedErrorr ownerDocumentZcreateAttributevaluer)rnamerattrrrr __setitem__s  z+getDomBuilder..AttrList.__setitem__cSs t|jjSr)lenrrrrrr__len__#sz'getDomBuilder..AttrList.__len__cSst|jjSr)listrritemsrrrrr&&sz%getDomBuilder..AttrList.itemscSst|jjSr)r%rrvaluesrrrrr')sz&getDomBuilder..AttrList.valuescSs"t|trtn|jj|jSdSr)rrrrrrrr rrr __getitem__,s z+getDomBuilder..AttrList.__getitem__cSst|trtn |jj|=dSr)rrrrrr(rrr __delitem__2s z+getDomBuilder..AttrList.__delitem__N) __name__ __module__ __qualname__rrr"r$r&r'r)r*rrrrAttrListsr.cseZdZddZeddZddZddd Zd d Zd d Z ddZ fddZ ddZ ee e Z fddZddZddZeeZdS)z"getDomBuilder..NodeBuildercSstj||j||_dSr)r rrnodeNamerrrrrr9sz+getDomBuilder..NodeBuilder.__init__cSst|jdr|jjpdS)N namespaceURI)hasattrrr0rrrr=s z+getDomBuilder..NodeBuilder.cSs||_|j|jdSr)parentr appendChildrnoderrrr4@sz.getDomBuilder..NodeBuilder.appendChildNcSs4|jj|}|r$|j||jn |j|dSr)rrcreateTextNode insertBeforer4)rdatar8textrrr insertTextDsz-getDomBuilder..NodeBuilder.insertTextcSs|j|j|j||_dSr)rr8r3)rr6ZrefNoderrrr8Ksz/getDomBuilder..NodeBuilder.insertBeforecSs&|jj|jkr|j|jd|_dSr)rZ parentNode removeChildr3r5rrrr<Osz.getDomBuilder..NodeBuilder.removeChildcSs6|jr,|jj}|j||j|qg|_dSr)r hasChildNodesZ firstChildr<r4 childNodes)rZ newParentchildrrrreparentChildrenTs   z3getDomBuilder..NodeBuilder.reparentChildrencs |jSrrr)r.rr getAttributes[sz0getDomBuilder..NodeBuilder.getAttributescSsv|rrt|D]`\}}t|trb|ddk rD|dd|d}n|d}|j|d||q|j||qdS)Nr:rr )r%r&rrrZsetAttributeNSZ setAttribute)rrr rZ qualifiedNamerrr setAttributes^s  z0getDomBuilder..NodeBuilder.setAttributescs|jdS)NF)r cloneNoder NodeBuilderrrrDmsz,getDomBuilder..NodeBuilder.cloneNodecSs |jSr)rr=rrrr hasContentpsz-getDomBuilder..NodeBuilder.hasContentcSs(|jdkrtd|jfS|j|jfSdS)NZhtml) namespacer r rrrr getNameTupless z/getDomBuilder..NodeBuilder.getNameTuple)N)r+r,r-rpropertyrHr4r;r8r<r@rArCrrDrGrIZ nameTupler)r.rFrrrF8s     rFcseZdZfddZfddZdfdd Zfdd Zfd d Zd d ZfddZ ddZ ddZ dddZ Z dZdS)z"getDomBuilder..TreeBuildercsddd|_t|Sr)getDOMImplementationZcreateDocumentdomweakrefproxyr)Domrr documentClass|sz0getDomBuilder..TreeBuilder.documentClasscsR|d}|d}|d}}||||}|j|tkrN|j|_dS)Nr publicIdsystemId)rKZcreateDocumentTypeZdocumentr4rrLr)rtokenr rQrRZdomimplZdoctype)rOrFrr insertDoctypesz0getDomBuilder..TreeBuilder.insertDoctypeNcs6|dkr |jdkr |j|}n|j||}|Sr)ZdefaultNamespacerLZ createElementZcreateElementNS)rr rHr6rErr elementClasssz/getDomBuilder..TreeBuilder.elementClasscs|j|Sr)rLZ createComment)rr9rErr commentClasssz/getDomBuilder..TreeBuilder.commentClasscs|jSr)rLZcreateDocumentFragmentrrErr fragmentClasssz0getDomBuilder..TreeBuilder.fragmentClasscSs|j|jdSr)rLr4rr5rrrr4sz.getDomBuilder..TreeBuilder.appendChildcs|Srrr)testSerializerrrrXsz1getDomBuilder..TreeBuilder.testSerializercSs|jSr)rLrrrr getDocumentsz.getDomBuilder..TreeBuilder.getDocumentcSstj|jSr)r TreeBuilder getFragmentrrrrrr[sz.getDomBuilder..TreeBuilder.getFragmentcSsp|}||krtj|||nNt|jdrXtj|jjkrXt|jj|j_|jj tj|j |j |dS)N_child_node_types) r rZr;r1rLr TEXT_NODEr\r%appendr4r7)rr9r3rrrr;s z-getDomBuilder..TreeBuilder.insertText)N)N)r+r,r-rPrTrUrVrWr4rXrYr[r;implementationr r)rODomImplementationrFrXrrrZ{s      rZcs0|gdfdd |ddS)Nrc s|jtjkr|jrj|js|jrP|jp&d}|jp0d}dd||j||fq~dd||jfndd|fnr|jtjkrdnX|jtjkrdn>|jtj krވdd||j fn|jtj krd d||j fnt |d r6|j dk r6d tj|j |jf}n|j}d d||f|rg}tt|jD]T}|j|}|j}|j}|j } | rd tj| |jf}n|j}|||fqnt|D]&\}}d d|d||fq|d7}|jD]} | |qdS)Nz|%s z|%sz|%sz #documentz#document-fragmentz|%sz|%s"%s"r0z%s %sz|%s<%s>z |%s%s="%s"r )ZnodeTyperZDOCUMENT_TYPE_NODEr rQrRr^Z DOCUMENT_NODEZDOCUMENT_FRAGMENT_NODEZ COMMENT_NODEZ nodeValuer]r1r0r prefixesr/Z hasAttributesranger#ritemrZ localNamesortedr>) rindentrQrRr rir!rnsr?rvserializeElementrrrlsT             z?getDomBuilder..testSerializer..serializeElement )r)Z normalizejoinrrrjrrXs . z%getDomBuilder..testSerializer)rr rrZlocals)r`rZr)r.rOr`rFrXr getDomBuilders $C:6rp)Z __future__rrrZcollections.abcr ImportError collectionsZxml.domrrrMrar r r Z_utilsr rpZ getDomModulerrrrs    _site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/base.cpython-38.opt-1.pyc000064400000025704151464452720024454 0ustar00U .e8@sddlmZmZmZddlmZddlmZmZm Z dZ e edfe ee e ddfgBdfe ee e dd fe dd fgBdfe e ddfe dd fgdfe e dd fe dd fgdfdZ GdddeZGdddeZGdddeZdS))absolute_importdivisionunicode_literals) text_type)scopingElementstableInsertModeElements namespacesNFhtmlbuttonZolZultableoptgroupoptionT)Nr listr Zselectc@sbeZdZdZddZddZddZdd Zdd d Zd dZ ddZ ddZ ddZ ddZ d S)NodezRepresents an item in the treecCs(||_d|_d|_i|_g|_g|_dS)zRCreates a Node :arg name: The tag name associated with the node N)nameparentvalue attributes childNodesZ_flags)selfrrJ/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/base.py__init__s z Node.__init__cCs:ddd|jD}|r,d|j|fSd|jSdS)N cSsg|]\}}d||fqS)z%s="%s"r).0rrrrr .sz Node.__str__..z<%s %s><%s>)joinritemsr)rZ attributesStrrrr__str__-s  z Node.__str__cCs d|jS)Nr)rrrrr__repr__6sz Node.__repr__cCstdS)z[Insert node as a child of the current node :arg node: the node to insert NNotImplementedErrorrnoderrr appendChild9szNode.appendChildNcCstdS)aBInsert data as text in the current node, positioned before the start of node insertBefore or to the end of the node's text. :arg data: the data to insert :arg insertBefore: True if you want to insert the text before the node and False if you want to insert it after the node Nr#)rdata insertBeforerrr insertTextAs zNode.insertTextcCstdS)aInsert node as a child of the current node, before refNode in the list of child nodes. Raises ValueError if refNode is not a child of the current node :arg node: the node to insert :arg refNode: the child node to insert the node before Nr#)rr&ZrefNoderrrr)Ms zNode.insertBeforecCstdS)zhRemove node from the children of the current node :arg node: the child node to remove Nr#r%rrr removeChildYszNode.removeChildcCs |jD]}||qg|_dS)zMove all the children of the current node to newParent. This is needed so that trees that don't store text as nodes move the text in the correct way :arg newParent: the node to move all this node's children to N)rr')rZ newParentZchildrrrreparentChildrenas  zNode.reparentChildrencCstdS)zReturn a shallow copy of the current node i.e. a node with the same name and attributes but with no parent or child nodes Nr#r!rrr cloneNodenszNode.cloneNodecCstdS)zFReturn true if the node has children or text, false otherwise Nr#r!rrr hasContenttszNode.hasContent)N)__name__ __module__ __qualname____doc__rr r"r'r*r)r+r,r-r.rrrrrs    rc@seZdZddZddZdS)ActiveFormattingElementscCsfd}|tkrV|dddD]:}|tkr*qV|||r>|d7}|dkr||qVqt||dS)Nr)Marker nodesEqualremoverappend)rr&Z equalCountelementrrrr:{s  zActiveFormattingElements.appendcCs$|j|jksdS|j|jks dSdS)NFT) nameTupler)rZnode1Znode2rrrr8s   z#ActiveFormattingElements.nodesEqualN)r/r0r1r:r8rrrrr3zs r3c@seZdZdZdZdZdZdZdZddZ ddZ d+ddZ d d Z d d Z d dZddZddZd,ddZddZddZddZeeeZddZddZd-dd Zd!d"Zd.d#d$Zd%d&Zd'd(Zd)d*ZdS)/ TreeBuilderaBase treebuilder implementation * documentClass - the class to use for the bottommost node of a document * elementClass - the class to use for HTML Elements * commentClass - the class to use for comments * doctypeClass - the class to use for doctypes NcCs|r d|_nd|_|dS)zmCreate a TreeBuilder :arg namespaceHTMLElements: whether or not to namespace HTML elements zhttp://www.w3.org/1999/xhtmlN)defaultNamespacereset)rZnamespaceHTMLElementsrrrrszTreeBuilder.__init__cCs.g|_t|_d|_d|_d|_||_dS)NF) openElementsr3activeFormattingElementsZ headPointerZ formPointerinsertFromTable documentClassdocumentr!rrrr?s zTreeBuilder.resetcCs~t|d}|s$t|tr$td|f}t|\}}t|jD]>}|rP||krPdS|sd|j|krddS||j|kAr:dSq:dS)Nr<r TF)hasattr isinstancerr listElementsMapreversedr@r<)rtargetZvariantZ exactNodeZ listElementsinvertr&rrrelementInScopes     zTreeBuilder.elementInScopecCs|js dSt|jd}|j|}|tks4||jkr8dS|tkrl||jkrl|dkrXd}ql|d8}|j|}q8|d7}|j|}|}|d|j|j|jd}||j|<||jdkrlqqldS)Nr5rr4ZStartTag)typer namespacer() rAlenr7r@r- insertElementrrMr)rientryZcloner;rrr#reconstructActiveFormattingElementss.    z/TreeBuilder.reconstructActiveFormattingElementscCs(|j}|jr$|tkr$|j}q dSN)rApopr7)rrQrrrclearActiveFormattingElementss z)TreeBuilder.clearActiveFormattingElementscCs:|jdddD]$}|tkr"q6q|j|kr|SqdS)zCheck if an element exists between the end of the active formatting elements and the last marker. If it does, return it, else return falseNr4F)rAr7r)rritemrrr!elementInActiveFormattingElements s   z-TreeBuilder.elementInActiveFormattingElementscCs&||}|j||j|dSrS) createElementr@r:rDr')rtokenr;rrr insertRoots  zTreeBuilder.insertRootcCs6|d}|d}|d}||||}|j|dS)NrpublicIdsystemId) doctypeClassrDr')rrYrr[r\Zdoctyperrr insertDoctype s zTreeBuilder.insertDoctypecCs*|dkr|jd}|||ddS)Nr4r()r@r' commentClass)rrYrrrr insertComment(s zTreeBuilder.insertCommentcCs0|d}|d|j}|||}|d|_|S)z.Create an element but don't insert it anywhererrMr()getr> elementClassrrrYrrMr;rrrrX-s   zTreeBuilder.createElementcCs|jSrS)_insertFromTabler!rrr_getInsertFromTable5szTreeBuilder._getInsertFromTablecCs ||_|r|j|_n|j|_dS)zsSwitch the function used to insert an element from the normal one to the misnested table one and back againN)rdinsertElementTablerOinsertElementNormal)rrrrr_setInsertFromTable8s zTreeBuilder._setInsertFromTablecCsL|d}|d|j}|||}|d|_|jd||j||S)NrrMr(r4)rar>rbrr@r'r:rcrrrrgCs   zTreeBuilder.insertElementNormalcCs`||}|jdjtkr$||S|\}}|dkrD||n ||||j||S)z-Create an element and insert it into the treer4N) rXr@rrrggetTableMisnestedNodePositionr'r)r:)rrYr;rr)rrrrfMs      zTreeBuilder.insertElementTablecCsV|dkr|jd}|jr.|jr:|jdjtkr:||n|\}}|||dS)zInsert text data.Nr4)r@rBrrr*ri)rr(rr)rrrr*]s     zTreeBuilder.insertTextcCstd}d}d}|jdddD]}|jdkr|}q4q|rb|jrJ|j}|}ql|j|j|d}n |jd}||fS)zsGet the foster parent element, and sibling to insert before (or None) when inserting a misnested table nodeNr4r r5r)r@rrindex)rZ lastTableZ fosterParentr)Zelmrrrrils   z)TreeBuilder.getTableMisnestedNodePositioncCs8|jdj}|tdkr4||kr4|j||dS)Nr4)ZddZdtZlirr pZrpZrt)r@r frozensetrTgenerateImpliedEndTags)rexcluderrrrrms    z"TreeBuilder.generateImpliedEndTagscCs|jS)zReturn the final tree)rDr!rrr getDocumentszTreeBuilder.getDocumentcCs|}|jd||S)zReturn the final fragmentr) fragmentClassr@r,)rZfragmentrrr getFragmentszTreeBuilder.getFragmentcCstdS)zSerialize the subtree of node in the format required by unit tests :arg node: the node from which to start serializing Nr#r%rrrtestSerializerszTreeBuilder.testSerializer)N)N)N)N)r/r0r1r2rCrbr_r]rprr?rKrRrUrWrZr^r`rXrerhpropertyrBrgrfr*rirmrorqrrrrrrr=s6   .     r=)Z __future__rrrZpip._vendor.sixrZ constantsrrr r7rlsetrGobjectrrr3r=rrrrs0        csite-packages/pip/_vendor/html5lib/treebuilders/__pycache__/__init__.cpython-38.opt-1.pyc000064400000006330151464452720025273 0ustar00U .e@s6dZddlmZmZmZddlmZiZdddZdS) a)A collection of modules for building different kinds of trees from HTML documents. To create a treebuilder for a new type of tree, you need to do implement several things: 1. A set of classes for various types of elements: Document, Doctype, Comment, Element. These must implement the interface of ``base.treebuilders.Node`` (although comment nodes have a different signature for their constructor, see ``treebuilders.etree.Comment``) Textual content may also be implemented as another node type, or not, as your tree implementation requires. 2. A treebuilder object (called ``TreeBuilder`` by convention) that inherits from ``treebuilders.base.TreeBuilder``. This has 4 required attributes: * ``documentClass`` - the class to use for the bottommost node of a document * ``elementClass`` - the class to use for HTML Elements * ``commentClass`` - the class to use for comments * ``doctypeClass`` - the class to use for doctypes It also has one required method: * ``getDocument`` - Returns the root node of the complete document tree 3. If you wish to run the unit tests, you must also create a ``testSerializer`` method on your treebuilder which accepts a node and returns a string containing Node and its children serialized according to the format used in the unittests )absolute_importdivisionunicode_literals) default_etreeNcKs|}|tkr|dkrLddlm}|dkr>> from html5lib.treebuilders import getTreeBuilder >>> builder = getTreeBuilder('etree') dom)rNr)minidomZlxml) etree_lxmletree)r zUnrecognised treebuilder "%s" )lowertreeBuilderCacherZxml.domr Z getDomModuleZ TreeBuilderr r rZgetETreeModule ValueErrorget)ZtreeTypeimplementationkwargsrr r r rN/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/__init__.pygetTreeBuilder's$      r)N) __doc__Z __future__rrrZ_utilsrr rrrrrs site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/dom.cpython-38.pyc000064400000022315151464452720023355 0ustar00U .e"@sddlmZmZmZzddlmZWn ek rDddlmZYnXddlm Z m Z ddl Z ddl m Z ddl mZdd lmZdd lmZd d ZeeZdS) )absolute_importdivisionunicode_literals)MutableMapping)minidomNodeN)base) constants) namespaces)moduleFactoryFactorycsVGdddtGfdddtjGfdddtj}ddtS) Nc@sLeZdZddZddZddZddZd d Zd d Zd dZ ddZ dS)zgetDomBuilder..AttrListcSs ||_dSNelementselfrrI/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/dom.py__init__sz(getDomBuilder..AttrList.__init__cSst|jjSr)iterr attributeskeysrrrr__iter__sz(getDomBuilder..AttrList.__iter__cSs4t|trtn |jj|}||_||jj|<dSr) isinstancetupleNotImplementedErrorr ownerDocumentZcreateAttributevaluer)rnamerattrrrr __setitem__s  z+getDomBuilder..AttrList.__setitem__cSs t|jjSr)lenrrrrrr__len__#sz'getDomBuilder..AttrList.__len__cSst|jjSr)listrritemsrrrrr&&sz%getDomBuilder..AttrList.itemscSst|jjSr)r%rrvaluesrrrrr')sz&getDomBuilder..AttrList.valuescSs"t|trtn|jj|jSdSr)rrrrrrrr rrr __getitem__,s z+getDomBuilder..AttrList.__getitem__cSst|trtn |jj|=dSr)rrrrrr(rrr __delitem__2s z+getDomBuilder..AttrList.__delitem__N) __name__ __module__ __qualname__rrr"r$r&r'r)r*rrrrAttrListsr.cseZdZddZeddZddZddd Zd d Zd d Z ddZ fddZ ddZ ee e Z fddZddZddZeeZdS)z"getDomBuilder..NodeBuildercSstj||j||_dSr)r rrnodeNamerrrrrr9sz+getDomBuilder..NodeBuilder.__init__cSst|jdr|jjpdS)N namespaceURI)hasattrrr0rrrr=s z+getDomBuilder..NodeBuilder.cSs||_|j|jdSr)parentr appendChildrnoderrrr4@sz.getDomBuilder..NodeBuilder.appendChildNcSs4|jj|}|r$|j||jn |j|dSr)rrcreateTextNode insertBeforer4)rdatar8textrrr insertTextDsz-getDomBuilder..NodeBuilder.insertTextcSs|j|j|j||_dSr)rr8r3)rr6ZrefNoderrrr8Ksz/getDomBuilder..NodeBuilder.insertBeforecSs&|jj|jkr|j|jd|_dSr)rZ parentNode removeChildr3r5rrrr<Osz.getDomBuilder..NodeBuilder.removeChildcSs6|jr,|jj}|j||j|qg|_dSr)r hasChildNodesZ firstChildr<r4 childNodes)rZ newParentchildrrrreparentChildrenTs   z3getDomBuilder..NodeBuilder.reparentChildrencs |jSrrr)r.rr getAttributes[sz0getDomBuilder..NodeBuilder.getAttributescSsv|rrt|D]`\}}t|trb|ddk rD|dd|d}n|d}|j|d||q|j||qdS)Nr:rr )r%r&rrrZsetAttributeNSZ setAttribute)rrr rZ qualifiedNamerrr setAttributes^s  z0getDomBuilder..NodeBuilder.setAttributescs|jdS)NF)r cloneNoder NodeBuilderrrrDmsz,getDomBuilder..NodeBuilder.cloneNodecSs |jSr)rr=rrrr hasContentpsz-getDomBuilder..NodeBuilder.hasContentcSs(|jdkrtd|jfS|j|jfSdS)NZhtml) namespacer r rrrr getNameTupless z/getDomBuilder..NodeBuilder.getNameTuple)N)r+r,r-rpropertyrHr4r;r8r<r@rArCrrDrGrIZ nameTupler)r.rFrrrF8s     rFcseZdZfddZfddZdfdd Zfdd Zfd d Zd d ZfddZ ddZ ddZ dddZ Z dZdS)z"getDomBuilder..TreeBuildercsddd|_t|Sr)getDOMImplementationZcreateDocumentdomweakrefproxyr)Domrr documentClass|sz0getDomBuilder..TreeBuilder.documentClasscsR|d}|d}|d}}||||}|j|tkrN|j|_dS)Nr publicIdsystemId)rKZcreateDocumentTypeZdocumentr4rrLr)rtokenr rQrRZdomimplZdoctype)rOrFrr insertDoctypesz0getDomBuilder..TreeBuilder.insertDoctypeNcs6|dkr |jdkr |j|}n|j||}|Sr)ZdefaultNamespacerLZ createElementZcreateElementNS)rr rHr6rErr elementClasssz/getDomBuilder..TreeBuilder.elementClasscs|j|Sr)rLZ createComment)rr9rErr commentClasssz/getDomBuilder..TreeBuilder.commentClasscs|jSr)rLZcreateDocumentFragmentrrErr fragmentClasssz0getDomBuilder..TreeBuilder.fragmentClasscSs|j|jdSr)rLr4rr5rrrr4sz.getDomBuilder..TreeBuilder.appendChildcs|Srrr)testSerializerrrrXsz1getDomBuilder..TreeBuilder.testSerializercSs|jSr)rLrrrr getDocumentsz.getDomBuilder..TreeBuilder.getDocumentcSstj|jSr)r TreeBuilder getFragmentrrrrrr[sz.getDomBuilder..TreeBuilder.getFragmentcSsp|}||krtj|||nNt|jdrXtj|jjkrXt|jj|j_|jj tj|j |j |dS)N_child_node_types) r rZr;r1rLr TEXT_NODEr\r%appendr4r7)rr9r3rrrr;s z-getDomBuilder..TreeBuilder.insertText)N)N)r+r,r-rPrTrUrVrWr4rXrYr[r;implementationr r)rODomImplementationrFrXrrrZ{s      rZcs0|gdfdd |ddS)Nrc s|jtjkr|jrj|js|jrP|jp&d}|jp0d}dd||j||fq~dd||jfndd|fnr|jtjkrdnX|jtjkrdn>|jtj krވdd||j fn|jtj krd d||j fnt |d r6|j dk r6d tj|j |jf}n|j}d d||f|rg}tt|jD]T}|j|}|j}|j}|j } | rd tj| |jf}n|j}|||fqnt|D]&\}}d d|d||fq|d7}|jD]} | |qdS)Nz|%s z|%sz|%sz #documentz#document-fragmentz|%sz|%s"%s"r0z%s %sz|%s<%s>z |%s%s="%s"r )ZnodeTyperZDOCUMENT_TYPE_NODEr rQrRr^Z DOCUMENT_NODEZDOCUMENT_FRAGMENT_NODEZ COMMENT_NODEZ nodeValuer]r1r0r prefixesr/Z hasAttributesranger#ritemrZ localNamesortedr>) rindentrQrRr rir!rnsr?rvserializeElementrrrlsT             z?getDomBuilder..testSerializer..serializeElement )r)Z normalizejoinrrrjrrXs . z%getDomBuilder..testSerializer)rr rrZlocals)r`rZr)r.rOr`rFrXr getDomBuilders $C:6rp)Z __future__rrrZcollections.abcr ImportError collectionsZxml.domrrrMrar r r Z_utilsr rpZ getDomModulerrrrs    _site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/etree.cpython-38.pyc000064400000027040151464452720023702 0ustar00U .e1@sddlmZmZmZddlmZddlZddlmZddlm Z ddlm Z dd l m Z dd l m Z ed Zdd dZe eZdS))absolute_importdivisionunicode_literals) text_typeN)base) _ihatexml) constants) namespaces)moduleFactoryFactoryz {([^}]*)}(.*)Fc sdjGfdddtjGfdddGfdddGfdd d Gfd d d fd d  fdd}G fdddtj}tS)NZasdcseZdZd$fdd ZddZddZdd ZeeeZd d Z d d Z ee e Z ddZ ddZ ee e ZddZddZeeeZddZddZddZddZd%ddZd d!Zd"d#ZdS)&z getETreeBuilder..ElementNcs^||_||_||||_|dkr:td|jf|_n|j|jf|_d|_g|_g|_ dS)Nhtml) _name _namespaceElement _getETreeTag_elementr Z nameTupleparent _childNodes_flags)selfname namespace ElementTreeK/usr/lib/python3.8/site-packages/pip/_vendor/html5lib/treebuilders/etree.py__init__s z)getETreeBuilder..Element.__init__cSs|dkr|}n d||f}|S)N{%s}%sr)rrrZ etree_tagrrrr#s z-getETreeBuilder..Element._getETreeTagcSs||_||j|j|j_dSN)rrrrtag)rrrrr_setName*sz)getETreeBuilder..Element._setNamecSs|jSr)rrrrr_getName.sz)getETreeBuilder..Element._getNamecSs||_||j|j|j_dSr)rrrrr )rrrrr _setNamespace3sz.getETreeBuilder..Element._setNamespacecSs|jSr)rr"rrr _getNamespace7sz.getETreeBuilder..Element._getNamespacecSs|jjSr)rattribr"rrr_getAttributes<sz/getETreeBuilder..Element._getAttributescSsht|jjD]}|jj|=q|D]:\}}t|trPd|d|df}n|}|j||q(dS)Nrrr)listrr&keysitems isinstancetupleset)r attributeskeyvaluerrrr_setAttributes?s  z/getETreeBuilder..Element._setAttributescSs|jSr)rr"rrr_getChildNodesMsz/getETreeBuilder..Element._getChildNodescSs*|jdd=g|_|D]}||qdSr)rrZ insertChild)rr0elementrrr_setChildNodesPs z/getETreeBuilder..Element._setChildNodescSst|jjpt|jS)z,Return true if the node has children or text)boolrtextlenr"rrr hasContentXsz+getETreeBuilder..Element.hasContentcSs$|j||j|j||_dSr)rappendrrrnoderrr appendChild\s z,getETreeBuilder..Element.appendChildcSs,t|j|j}|j||j||_dSr)r(rindexinsertr)rr;ZrefNoder=rrr insertBeforeasz-getETreeBuilder..Element.insertBeforecSs$|j||j|jd|_dSr)rremoverrr:rrr removeChildfs z,getETreeBuilder..Element.removeChildcSst|js,|jjsd|j_|jj|7_n|dkrb|jdjsLd|jd_|jdj|7_nxt|j}||j}|dkr|j|djsd|j|d_|j|dj|7_n |jjsd|j_|jj|7_dS)Nrr)r7rr6tailr(r=)rdatar?Zchildrenr=rrr insertTextks"     z+getETreeBuilder..Element.insertTextcSs4t||j|j}|jD]\}}||j|<q|Sr)typerrr.r*)rr3rr0rrr cloneNodes z*getETreeBuilder..Element.cloneNodecSsl|jr"|jdjj|jj7_n0|jjs2d|j_|jjdk rR|jj|jj7_d|j_tj||dS)NrCrB) childNodesrrDr6rNodereparentChildren)rZ newParentrrrrKs z1getETreeBuilder..Element.reparentChildren)N)N)__name__ __module__ __qualname__rrr!r#propertyrr$r%rr'r1r.r2r4rIr8r<r?rArFrHrKrrrrrs*      rcs2eZdZfddZddZddZeeeZdS)z getETreeBuilder..Commentcs"||_d|_g|_g|_dSr)Commentrrrr)rrErrrrs z)getETreeBuilder..Comment.__init__cSs|jjSrrr6r"rrr_getDatasz)getETreeBuilder..Comment._getDatacSs ||j_dSrrQrr0rrr_setDatasz)getETreeBuilder..Comment._setDataN)rLrMrNrrRrTrOrErrrrrPs rPcsLeZdZfddZddZddZeeeZddZd d Z eee Z d S) z%getETreeBuilder..DocumentTypecs$|d||j_||_||_dS)N )rrr6publicIdsystemId)rrrVrWrrrrs z.getETreeBuilder..DocumentType.__init__cSs|jddS)NrVrBrgetr"rrr _getPublicIdsz2getETreeBuilder..DocumentType._getPublicIdcSs|dk r|jd|dS)NrVrr-rSrrr _setPublicIdsz2getETreeBuilder..DocumentType._setPublicIdcSs|jddS)NrWrBrYr"rrr _getSystemIdsz2getETreeBuilder..DocumentType._getSystemIdcSs|dk r|jd|dS)NrWr\rSrrr _setSystemIdsz2getETreeBuilder..DocumentType._setSystemIdN) rLrMrNrr[r]rOrVr^r_rWrrXrr DocumentTypes  r`cseZdZfddZdS)z!getETreeBuilder..Documentcs|ddS)N DOCUMENT_ROOTrr"rXrrrsz*getETreeBuilder..Document.__init__NrLrMrNrrrXrrDocumentsrdcseZdZfddZdS)z)getETreeBuilder..DocumentFragmentcs|ddS)NZDOCUMENT_FRAGMENTrbr"rXrrrsz2getETreeBuilder..DocumentFragment.__init__NrcrrXrrDocumentFragmentsrecs*gdfdd |ddS)Nrc st|ds|}|jdkrz|ds0|drd|dp raz #documentz|%s"%s" rDocument node cannot have tailr&$Document node cannot have attributesz|%szExpected unicode, got %s, %sz%s %sz|%s<%s>z |%s%s="%s")hasattrgetrootr rZr9r6rD TypeErrorr7r&r+rAssertionErrorrG tag_regexpmatchgroupsr prefixesr*sorted) r3indentrVrWZnsmatchrnsprefixr.r0Z attr_stringchild)ElementTreeCommentTypervserializeElementrrrysb                    zAgetETreeBuilder..testSerializer..serializeElement )r)joinr3)rw)rxryrtestSerializers7 z'getETreeBuilder..testSerializercs2gtfdd|dS)z4Serialize an element and its child nodes to a stringcst|jr|}|jdkr||ds2|drf|dp>d}|dpLd}d|j||fnd|jfn|jdkr|jdk r|j|jdk rtdt |d rt |j rtd |D] }|qn|jkrd |jfn|j sd  |jfn2d fdd|j D}d|j|f|jrd|j|D]}|qhd|jf|jr|jdS)NrUrVrWrBzrfrarhr&riz z<%s>rgcs"g|]\}}d||fqS)z%s="%s") fromXmlName).0rr0)filterrr &s zOgetETreeBuilder..tostring..serializeElement..z<%s %s>z)r+rrkr rZr9r6rDrlrjr7r&r~r{r*)r3rVrWrvattr)rrwrrxryrrrysD           z;getETreeBuilder..tostring..serializeElementrB)r Z InfosetFilterr{r|)rrw)rrxryrtostrings -z!getETreeBuilder..tostringcsDeZdZZZZZZZfddZ fddZ ddZ dS)z$getETreeBuilder..TreeBuildercs|Srr)rr3)r}rrr}Asz3getETreeBuilder..TreeBuilder.testSerializercs<r |jjS|jdk r*|jjd|jS|jjdSdS)Nz{%s}htmlr )ZdocumentrZdefaultNamespacefindr")fullTreerr getDocumentDs z0getETreeBuilder..TreeBuilder.getDocumentcSstj|jSr)r TreeBuilder getFragmentrr"rrrrNsz0getETreeBuilder..TreeBuilder.getFragmentN) rLrMrNZ documentClassZ doctypeClassZ elementClassZ commentClassZ fragmentClassimplementationr}rrr)rPrdrer`rElementTreeImplementationrr}rrr9s  r)rPr rrJrlocals)rrrrr) rPrdrer`rrrwrrr}rgetETreeBuilders ~ >6$r)F)Z __future__rrrZpip._vendor.sixrrerBrr r r Z_utilsr compilernrZgetETreeModulerrrrs        Esite-packages/pip/_vendor/html5lib/treebuilders/dom.py000064400000021335151464452720017070 0ustar00from __future__ import absolute_import, division, unicode_literals try: from collections.abc import MutableMapping except ImportError: # Python 2.7 from collections import MutableMapping from xml.dom import minidom, Node import weakref from . import base from .. import constants from ..constants import namespaces from .._utils import moduleFactoryFactory def getDomBuilder(DomImplementation): Dom = DomImplementation class AttrList(MutableMapping): def __init__(self, element): self.element = element def __iter__(self): return iter(self.element.attributes.keys()) def __setitem__(self, name, value): if isinstance(name, tuple): raise NotImplementedError else: attr = self.element.ownerDocument.createAttribute(name) attr.value = value self.element.attributes[name] = attr def __len__(self): return len(self.element.attributes) def items(self): return list(self.element.attributes.items()) def values(self): return list(self.element.attributes.values()) def __getitem__(self, name): if isinstance(name, tuple): raise NotImplementedError else: return self.element.attributes[name].value def __delitem__(self, name): if isinstance(name, tuple): raise NotImplementedError else: del self.element.attributes[name] class NodeBuilder(base.Node): def __init__(self, element): base.Node.__init__(self, element.nodeName) self.element = element namespace = property(lambda self: hasattr(self.element, "namespaceURI") and self.element.namespaceURI or None) def appendChild(self, node): node.parent = self self.element.appendChild(node.element) def insertText(self, data, insertBefore=None): text = self.element.ownerDocument.createTextNode(data) if insertBefore: self.element.insertBefore(text, insertBefore.element) else: self.element.appendChild(text) def insertBefore(self, node, refNode): self.element.insertBefore(node.element, refNode.element) node.parent = self def removeChild(self, node): if node.element.parentNode == self.element: self.element.removeChild(node.element) node.parent = None def reparentChildren(self, newParent): while self.element.hasChildNodes(): child = self.element.firstChild self.element.removeChild(child) newParent.element.appendChild(child) self.childNodes = [] def getAttributes(self): return AttrList(self.element) def setAttributes(self, attributes): if attributes: for name, value in list(attributes.items()): if isinstance(name, tuple): if name[0] is not None: qualifiedName = (name[0] + ":" + name[1]) else: qualifiedName = name[1] self.element.setAttributeNS(name[2], qualifiedName, value) else: self.element.setAttribute( name, value) attributes = property(getAttributes, setAttributes) def cloneNode(self): return NodeBuilder(self.element.cloneNode(False)) def hasContent(self): return self.element.hasChildNodes() def getNameTuple(self): if self.namespace is None: return namespaces["html"], self.name else: return self.namespace, self.name nameTuple = property(getNameTuple) class TreeBuilder(base.TreeBuilder): # pylint:disable=unused-variable def documentClass(self): self.dom = Dom.getDOMImplementation().createDocument(None, None, None) return weakref.proxy(self) def insertDoctype(self, token): name = token["name"] publicId = token["publicId"] systemId = token["systemId"] domimpl = Dom.getDOMImplementation() doctype = domimpl.createDocumentType(name, publicId, systemId) self.document.appendChild(NodeBuilder(doctype)) if Dom == minidom: doctype.ownerDocument = self.dom def elementClass(self, name, namespace=None): if namespace is None and self.defaultNamespace is None: node = self.dom.createElement(name) else: node = self.dom.createElementNS(namespace, name) return NodeBuilder(node) def commentClass(self, data): return NodeBuilder(self.dom.createComment(data)) def fragmentClass(self): return NodeBuilder(self.dom.createDocumentFragment()) def appendChild(self, node): self.dom.appendChild(node.element) def testSerializer(self, element): return testSerializer(element) def getDocument(self): return self.dom def getFragment(self): return base.TreeBuilder.getFragment(self).element def insertText(self, data, parent=None): data = data if parent != self: base.TreeBuilder.insertText(self, data, parent) else: # HACK: allow text nodes as children of the document node if hasattr(self.dom, '_child_node_types'): # pylint:disable=protected-access if Node.TEXT_NODE not in self.dom._child_node_types: self.dom._child_node_types = list(self.dom._child_node_types) self.dom._child_node_types.append(Node.TEXT_NODE) self.dom.appendChild(self.dom.createTextNode(data)) implementation = DomImplementation name = None def testSerializer(element): element.normalize() rv = [] def serializeElement(element, indent=0): if element.nodeType == Node.DOCUMENT_TYPE_NODE: if element.name: if element.publicId or element.systemId: publicId = element.publicId or "" systemId = element.systemId or "" rv.append("""|%s""" % (' ' * indent, element.name, publicId, systemId)) else: rv.append("|%s" % (' ' * indent, element.name)) else: rv.append("|%s" % (' ' * indent,)) elif element.nodeType == Node.DOCUMENT_NODE: rv.append("#document") elif element.nodeType == Node.DOCUMENT_FRAGMENT_NODE: rv.append("#document-fragment") elif element.nodeType == Node.COMMENT_NODE: rv.append("|%s" % (' ' * indent, element.nodeValue)) elif element.nodeType == Node.TEXT_NODE: rv.append("|%s\"%s\"" % (' ' * indent, element.nodeValue)) else: if (hasattr(element, "namespaceURI") and element.namespaceURI is not None): name = "%s %s" % (constants.prefixes[element.namespaceURI], element.nodeName) else: name = element.nodeName rv.append("|%s<%s>" % (' ' * indent, name)) if element.hasAttributes(): attributes = [] for i in range(len(element.attributes)): attr = element.attributes.item(i) name = attr.nodeName value = attr.value ns = attr.namespaceURI if ns: name = "%s %s" % (constants.prefixes[ns], attr.localName) else: name = attr.nodeName attributes.append((name, value)) for name, value in sorted(attributes): rv.append('|%s%s="%s"' % (' ' * (indent + 2), name, value)) indent += 2 for child in element.childNodes: serializeElement(child, indent) serializeElement(element, 0) return "\n".join(rv) return locals() # The actual means to get a module! getDomModule = moduleFactoryFactory(getDomBuilder) site-packages/pip/_vendor/html5lib/treebuilders/__init__.py000064400000007010151464452720020042 0ustar00"""A collection of modules for building different kinds of trees from HTML documents. To create a treebuilder for a new type of tree, you need to do implement several things: 1. A set of classes for various types of elements: Document, Doctype, Comment, Element. These must implement the interface of ``base.treebuilders.Node`` (although comment nodes have a different signature for their constructor, see ``treebuilders.etree.Comment``) Textual content may also be implemented as another node type, or not, as your tree implementation requires. 2. A treebuilder object (called ``TreeBuilder`` by convention) that inherits from ``treebuilders.base.TreeBuilder``. This has 4 required attributes: * ``documentClass`` - the class to use for the bottommost node of a document * ``elementClass`` - the class to use for HTML Elements * ``commentClass`` - the class to use for comments * ``doctypeClass`` - the class to use for doctypes It also has one required method: * ``getDocument`` - Returns the root node of the complete document tree 3. If you wish to run the unit tests, you must also create a ``testSerializer`` method on your treebuilder which accepts a node and returns a string containing Node and its children serialized according to the format used in the unittests """ from __future__ import absolute_import, division, unicode_literals from .._utils import default_etree treeBuilderCache = {} def getTreeBuilder(treeType, implementation=None, **kwargs): """Get a TreeBuilder class for various types of trees with built-in support :arg treeType: the name of the tree type required (case-insensitive). Supported values are: * "dom" - A generic builder for DOM implementations, defaulting to a xml.dom.minidom based implementation. * "etree" - A generic builder for tree implementations exposing an ElementTree-like interface, defaulting to xml.etree.cElementTree if available and xml.etree.ElementTree if not. * "lxml" - A etree-based builder for lxml.etree, handling limitations of lxml's implementation. :arg implementation: (Currently applies to the "etree" and "dom" tree types). A module implementing the tree type e.g. xml.etree.ElementTree or xml.etree.cElementTree. :arg kwargs: Any additional options to pass to the TreeBuilder when creating it. Example: >>> from html5lib.treebuilders import getTreeBuilder >>> builder = getTreeBuilder('etree') """ treeType = treeType.lower() if treeType not in treeBuilderCache: if treeType == "dom": from . import dom # Come up with a sane default (pref. from the stdlib) if implementation is None: from xml.dom import minidom implementation = minidom # NEVER cache here, caching is done in the dom submodule return dom.getDomModule(implementation, **kwargs).TreeBuilder elif treeType == "lxml": from . import etree_lxml treeBuilderCache[treeType] = etree_lxml.TreeBuilder elif treeType == "etree": from . import etree if implementation is None: implementation = default_etree # NEVER cache here, caching is done in the etree submodule return etree.getETreeModule(implementation, **kwargs).TreeBuilder else: raise ValueError("""Unrecognised treebuilder "%s" """ % treeType) return treeBuilderCache.get(treeType) site-packages/pip/_vendor/html5lib/__init__.py000064400000002212151464452720015350 0ustar00""" HTML parsing library based on the `WHATWG HTML specification `_. The parser is designed to be compatible with existing HTML found in the wild and implements well-defined error recovery that is largely compatible with modern desktop web browsers. Example usage:: from pip._vendor import html5lib with open("my_document.html", "rb") as f: tree = html5lib.parse(f) For convenience, this module re-exports the following names: * :func:`~.html5parser.parse` * :func:`~.html5parser.parseFragment` * :class:`~.html5parser.HTMLParser` * :func:`~.treebuilders.getTreeBuilder` * :func:`~.treewalkers.getTreeWalker` * :func:`~.serializer.serialize` """ from __future__ import absolute_import, division, unicode_literals from .html5parser import HTMLParser, parse, parseFragment from .treebuilders import getTreeBuilder from .treewalkers import getTreeWalker from .serializer import serialize __all__ = ["HTMLParser", "parse", "parseFragment", "getTreeBuilder", "getTreeWalker", "serialize"] # this has to be at the top level, see how setup.py parses this #: Distribution version number. __version__ = "1.0.1" site-packages/pip/_vendor/html5lib/constants.py000064400000243076151464452720015644 0ustar00from __future__ import absolute_import, division, unicode_literals import string EOF = None E = { "null-character": "Null character in input stream, replaced with U+FFFD.", "invalid-codepoint": "Invalid codepoint in stream.", "incorrectly-placed-solidus": "Solidus (/) incorrectly placed in tag.", "incorrect-cr-newline-entity": "Incorrect CR newline entity, replaced with LF.", "illegal-windows-1252-entity": "Entity used with illegal number (windows-1252 reference).", "cant-convert-numeric-entity": "Numeric entity couldn't be converted to character " "(codepoint U+%(charAsInt)08x).", "illegal-codepoint-for-numeric-entity": "Numeric entity represents an illegal codepoint: " "U+%(charAsInt)08x.", "numeric-entity-without-semicolon": "Numeric entity didn't end with ';'.", "expected-numeric-entity-but-got-eof": "Numeric entity expected. Got end of file instead.", "expected-numeric-entity": "Numeric entity expected but none found.", "named-entity-without-semicolon": "Named entity didn't end with ';'.", "expected-named-entity": "Named entity expected. Got none.", "attributes-in-end-tag": "End tag contains unexpected attributes.", 'self-closing-flag-on-end-tag': "End tag contains unexpected self-closing flag.", "expected-tag-name-but-got-right-bracket": "Expected tag name. Got '>' instead.", "expected-tag-name-but-got-question-mark": "Expected tag name. Got '?' instead. (HTML doesn't " "support processing instructions.)", "expected-tag-name": "Expected tag name. Got something else instead", "expected-closing-tag-but-got-right-bracket": "Expected closing tag. Got '>' instead. Ignoring ''.", "expected-closing-tag-but-got-eof": "Expected closing tag. Unexpected end of file.", "expected-closing-tag-but-got-char": "Expected closing tag. Unexpected character '%(data)s' found.", "eof-in-tag-name": "Unexpected end of file in the tag name.", "expected-attribute-name-but-got-eof": "Unexpected end of file. Expected attribute name instead.", "eof-in-attribute-name": "Unexpected end of file in attribute name.", "invalid-character-in-attribute-name": "Invalid character in attribute name", "duplicate-attribute": "Dropped duplicate attribute on tag.", "expected-end-of-tag-name-but-got-eof": "Unexpected end of file. Expected = or end of tag.", "expected-attribute-value-but-got-eof": "Unexpected end of file. Expected attribute value.", "expected-attribute-value-but-got-right-bracket": "Expected attribute value. Got '>' instead.", 'equals-in-unquoted-attribute-value': "Unexpected = in unquoted attribute", 'unexpected-character-in-unquoted-attribute-value': "Unexpected character in unquoted attribute", "invalid-character-after-attribute-name": "Unexpected character after attribute name.", "unexpected-character-after-attribute-value": "Unexpected character after attribute value.", "eof-in-attribute-value-double-quote": "Unexpected end of file in attribute value (\").", "eof-in-attribute-value-single-quote": "Unexpected end of file in attribute value (').", "eof-in-attribute-value-no-quotes": "Unexpected end of file in attribute value.", "unexpected-EOF-after-solidus-in-tag": "Unexpected end of file in tag. Expected >", "unexpected-character-after-solidus-in-tag": "Unexpected character after / in tag. Expected >", "expected-dashes-or-doctype": "Expected '--' or 'DOCTYPE'. Not found.", "unexpected-bang-after-double-dash-in-comment": "Unexpected ! after -- in comment", "unexpected-space-after-double-dash-in-comment": "Unexpected space after -- in comment", "incorrect-comment": "Incorrect comment.", "eof-in-comment": "Unexpected end of file in comment.", "eof-in-comment-end-dash": "Unexpected end of file in comment (-)", "unexpected-dash-after-double-dash-in-comment": "Unexpected '-' after '--' found in comment.", "eof-in-comment-double-dash": "Unexpected end of file in comment (--).", "eof-in-comment-end-space-state": "Unexpected end of file in comment.", "eof-in-comment-end-bang-state": "Unexpected end of file in comment.", "unexpected-char-in-comment": "Unexpected character in comment found.", "need-space-after-doctype": "No space after literal string 'DOCTYPE'.", "expected-doctype-name-but-got-right-bracket": "Unexpected > character. Expected DOCTYPE name.", "expected-doctype-name-but-got-eof": "Unexpected end of file. Expected DOCTYPE name.", "eof-in-doctype-name": "Unexpected end of file in DOCTYPE name.", "eof-in-doctype": "Unexpected end of file in DOCTYPE.", "expected-space-or-right-bracket-in-doctype": "Expected space or '>'. Got '%(data)s'", "unexpected-end-of-doctype": "Unexpected end of DOCTYPE.", "unexpected-char-in-doctype": "Unexpected character in DOCTYPE.", "eof-in-innerhtml": "XXX innerHTML EOF", "unexpected-doctype": "Unexpected DOCTYPE. Ignored.", "non-html-root": "html needs to be the first start tag.", "expected-doctype-but-got-eof": "Unexpected End of file. Expected DOCTYPE.", "unknown-doctype": "Erroneous DOCTYPE.", "expected-doctype-but-got-chars": "Unexpected non-space characters. Expected DOCTYPE.", "expected-doctype-but-got-start-tag": "Unexpected start tag (%(name)s). Expected DOCTYPE.", "expected-doctype-but-got-end-tag": "Unexpected end tag (%(name)s). Expected DOCTYPE.", "end-tag-after-implied-root": "Unexpected end tag (%(name)s) after the (implied) root element.", "expected-named-closing-tag-but-got-eof": "Unexpected end of file. Expected end tag (%(name)s).", "two-heads-are-not-better-than-one": "Unexpected start tag head in existing head. Ignored.", "unexpected-end-tag": "Unexpected end tag (%(name)s). Ignored.", "unexpected-start-tag-out-of-my-head": "Unexpected start tag (%(name)s) that can be in head. Moved.", "unexpected-start-tag": "Unexpected start tag (%(name)s).", "missing-end-tag": "Missing end tag (%(name)s).", "missing-end-tags": "Missing end tags (%(name)s).", "unexpected-start-tag-implies-end-tag": "Unexpected start tag (%(startName)s) " "implies end tag (%(endName)s).", "unexpected-start-tag-treated-as": "Unexpected start tag (%(originalName)s). Treated as %(newName)s.", "deprecated-tag": "Unexpected start tag %(name)s. Don't use it!", "unexpected-start-tag-ignored": "Unexpected start tag %(name)s. Ignored.", "expected-one-end-tag-but-got-another": "Unexpected end tag (%(gotName)s). " "Missing end tag (%(expectedName)s).", "end-tag-too-early": "End tag (%(name)s) seen too early. Expected other end tag.", "end-tag-too-early-named": "Unexpected end tag (%(gotName)s). Expected end tag (%(expectedName)s).", "end-tag-too-early-ignored": "End tag (%(name)s) seen too early. Ignored.", "adoption-agency-1.1": "End tag (%(name)s) violates step 1, " "paragraph 1 of the adoption agency algorithm.", "adoption-agency-1.2": "End tag (%(name)s) violates step 1, " "paragraph 2 of the adoption agency algorithm.", "adoption-agency-1.3": "End tag (%(name)s) violates step 1, " "paragraph 3 of the adoption agency algorithm.", "adoption-agency-4.4": "End tag (%(name)s) violates step 4, " "paragraph 4 of the adoption agency algorithm.", "unexpected-end-tag-treated-as": "Unexpected end tag (%(originalName)s). Treated as %(newName)s.", "no-end-tag": "This element (%(name)s) has no end tag.", "unexpected-implied-end-tag-in-table": "Unexpected implied end tag (%(name)s) in the table phase.", "unexpected-implied-end-tag-in-table-body": "Unexpected implied end tag (%(name)s) in the table body phase.", "unexpected-char-implies-table-voodoo": "Unexpected non-space characters in " "table context caused voodoo mode.", "unexpected-hidden-input-in-table": "Unexpected input with type hidden in table context.", "unexpected-form-in-table": "Unexpected form in table context.", "unexpected-start-tag-implies-table-voodoo": "Unexpected start tag (%(name)s) in " "table context caused voodoo mode.", "unexpected-end-tag-implies-table-voodoo": "Unexpected end tag (%(name)s) in " "table context caused voodoo mode.", "unexpected-cell-in-table-body": "Unexpected table cell start tag (%(name)s) " "in the table body phase.", "unexpected-cell-end-tag": "Got table cell end tag (%(name)s) " "while required end tags are missing.", "unexpected-end-tag-in-table-body": "Unexpected end tag (%(name)s) in the table body phase. Ignored.", "unexpected-implied-end-tag-in-table-row": "Unexpected implied end tag (%(name)s) in the table row phase.", "unexpected-end-tag-in-table-row": "Unexpected end tag (%(name)s) in the table row phase. Ignored.", "unexpected-select-in-select": "Unexpected select start tag in the select phase " "treated as select end tag.", "unexpected-input-in-select": "Unexpected input start tag in the select phase.", "unexpected-start-tag-in-select": "Unexpected start tag token (%(name)s in the select phase. " "Ignored.", "unexpected-end-tag-in-select": "Unexpected end tag (%(name)s) in the select phase. Ignored.", "unexpected-table-element-start-tag-in-select-in-table": "Unexpected table element start tag (%(name)s) in the select in table phase.", "unexpected-table-element-end-tag-in-select-in-table": "Unexpected table element end tag (%(name)s) in the select in table phase.", "unexpected-char-after-body": "Unexpected non-space characters in the after body phase.", "unexpected-start-tag-after-body": "Unexpected start tag token (%(name)s)" " in the after body phase.", "unexpected-end-tag-after-body": "Unexpected end tag token (%(name)s)" " in the after body phase.", "unexpected-char-in-frameset": "Unexpected characters in the frameset phase. Characters ignored.", "unexpected-start-tag-in-frameset": "Unexpected start tag token (%(name)s)" " in the frameset phase. Ignored.", "unexpected-frameset-in-frameset-innerhtml": "Unexpected end tag token (frameset) " "in the frameset phase (innerHTML).", "unexpected-end-tag-in-frameset": "Unexpected end tag token (%(name)s)" " in the frameset phase. Ignored.", "unexpected-char-after-frameset": "Unexpected non-space characters in the " "after frameset phase. Ignored.", "unexpected-start-tag-after-frameset": "Unexpected start tag (%(name)s)" " in the after frameset phase. Ignored.", "unexpected-end-tag-after-frameset": "Unexpected end tag (%(name)s)" " in the after frameset phase. Ignored.", "unexpected-end-tag-after-body-innerhtml": "Unexpected end tag after body(innerHtml)", "expected-eof-but-got-char": "Unexpected non-space characters. Expected end of file.", "expected-eof-but-got-start-tag": "Unexpected start tag (%(name)s)" ". Expected end of file.", "expected-eof-but-got-end-tag": "Unexpected end tag (%(name)s)" ". Expected end of file.", "eof-in-table": "Unexpected end of file. Expected table content.", "eof-in-select": "Unexpected end of file. Expected select content.", "eof-in-frameset": "Unexpected end of file. Expected frameset content.", "eof-in-script-in-script": "Unexpected end of file. Expected script content.", "eof-in-foreign-lands": "Unexpected end of file. Expected foreign content", "non-void-element-with-trailing-solidus": "Trailing solidus not allowed on element %(name)s", "unexpected-html-element-in-foreign-content": "Element %(name)s not allowed in a non-html context", "unexpected-end-tag-before-html": "Unexpected end tag (%(name)s) before html.", "unexpected-inhead-noscript-tag": "Element %(name)s not allowed in a inhead-noscript context", "eof-in-head-noscript": "Unexpected end of file. Expected inhead-noscript content", "char-in-head-noscript": "Unexpected non-space character. Expected inhead-noscript content", "XXX-undefined-error": "Undefined error (this sucks and should be fixed)", } namespaces = { "html": "http://www.w3.org/1999/xhtml", "mathml": "http://www.w3.org/1998/Math/MathML", "svg": "http://www.w3.org/2000/svg", "xlink": "http://www.w3.org/1999/xlink", "xml": "http://www.w3.org/XML/1998/namespace", "xmlns": "http://www.w3.org/2000/xmlns/" } scopingElements = frozenset([ (namespaces["html"], "applet"), (namespaces["html"], "caption"), (namespaces["html"], "html"), (namespaces["html"], "marquee"), (namespaces["html"], "object"), (namespaces["html"], "table"), (namespaces["html"], "td"), (namespaces["html"], "th"), (namespaces["mathml"], "mi"), (namespaces["mathml"], "mo"), (namespaces["mathml"], "mn"), (namespaces["mathml"], "ms"), (namespaces["mathml"], "mtext"), (namespaces["mathml"], "annotation-xml"), (namespaces["svg"], "foreignObject"), (namespaces["svg"], "desc"), (namespaces["svg"], "title"), ]) formattingElements = frozenset([ (namespaces["html"], "a"), (namespaces["html"], "b"), (namespaces["html"], "big"), (namespaces["html"], "code"), (namespaces["html"], "em"), (namespaces["html"], "font"), (namespaces["html"], "i"), (namespaces["html"], "nobr"), (namespaces["html"], "s"), (namespaces["html"], "small"), (namespaces["html"], "strike"), (namespaces["html"], "strong"), (namespaces["html"], "tt"), (namespaces["html"], "u") ]) specialElements = frozenset([ (namespaces["html"], "address"), (namespaces["html"], "applet"), (namespaces["html"], "area"), (namespaces["html"], "article"), (namespaces["html"], "aside"), (namespaces["html"], "base"), (namespaces["html"], "basefont"), (namespaces["html"], "bgsound"), (namespaces["html"], "blockquote"), (namespaces["html"], "body"), (namespaces["html"], "br"), (namespaces["html"], "button"), (namespaces["html"], "caption"), (namespaces["html"], "center"), (namespaces["html"], "col"), (namespaces["html"], "colgroup"), (namespaces["html"], "command"), (namespaces["html"], "dd"), (namespaces["html"], "details"), (namespaces["html"], "dir"), (namespaces["html"], "div"), (namespaces["html"], "dl"), (namespaces["html"], "dt"), (namespaces["html"], "embed"), (namespaces["html"], "fieldset"), (namespaces["html"], "figure"), (namespaces["html"], "footer"), (namespaces["html"], "form"), (namespaces["html"], "frame"), (namespaces["html"], "frameset"), (namespaces["html"], "h1"), (namespaces["html"], "h2"), (namespaces["html"], "h3"), (namespaces["html"], "h4"), (namespaces["html"], "h5"), (namespaces["html"], "h6"), (namespaces["html"], "head"), (namespaces["html"], "header"), (namespaces["html"], "hr"), (namespaces["html"], "html"), (namespaces["html"], "iframe"), # Note that image is commented out in the spec as "this isn't an # element that can end up on the stack, so it doesn't matter," (namespaces["html"], "image"), (namespaces["html"], "img"), (namespaces["html"], "input"), (namespaces["html"], "isindex"), (namespaces["html"], "li"), (namespaces["html"], "link"), (namespaces["html"], "listing"), (namespaces["html"], "marquee"), (namespaces["html"], "menu"), (namespaces["html"], "meta"), (namespaces["html"], "nav"), (namespaces["html"], "noembed"), (namespaces["html"], "noframes"), (namespaces["html"], "noscript"), (namespaces["html"], "object"), (namespaces["html"], "ol"), (namespaces["html"], "p"), (namespaces["html"], "param"), (namespaces["html"], "plaintext"), (namespaces["html"], "pre"), (namespaces["html"], "script"), (namespaces["html"], "section"), (namespaces["html"], "select"), (namespaces["html"], "style"), (namespaces["html"], "table"), (namespaces["html"], "tbody"), (namespaces["html"], "td"), (namespaces["html"], "textarea"), (namespaces["html"], "tfoot"), (namespaces["html"], "th"), (namespaces["html"], "thead"), (namespaces["html"], "title"), (namespaces["html"], "tr"), (namespaces["html"], "ul"), (namespaces["html"], "wbr"), (namespaces["html"], "xmp"), (namespaces["svg"], "foreignObject") ]) htmlIntegrationPointElements = frozenset([ (namespaces["mathml"], "annotation-xml"), (namespaces["svg"], "foreignObject"), (namespaces["svg"], "desc"), (namespaces["svg"], "title") ]) mathmlTextIntegrationPointElements = frozenset([ (namespaces["mathml"], "mi"), (namespaces["mathml"], "mo"), (namespaces["mathml"], "mn"), (namespaces["mathml"], "ms"), (namespaces["mathml"], "mtext") ]) adjustSVGAttributes = { "attributename": "attributeName", "attributetype": "attributeType", "basefrequency": "baseFrequency", "baseprofile": "baseProfile", "calcmode": "calcMode", "clippathunits": "clipPathUnits", "contentscripttype": "contentScriptType", "contentstyletype": "contentStyleType", "diffuseconstant": "diffuseConstant", "edgemode": "edgeMode", "externalresourcesrequired": "externalResourcesRequired", "filterres": "filterRes", "filterunits": "filterUnits", "glyphref": "glyphRef", "gradienttransform": "gradientTransform", "gradientunits": "gradientUnits", "kernelmatrix": "kernelMatrix", "kernelunitlength": "kernelUnitLength", "keypoints": "keyPoints", "keysplines": "keySplines", "keytimes": "keyTimes", "lengthadjust": "lengthAdjust", "limitingconeangle": "limitingConeAngle", "markerheight": "markerHeight", "markerunits": "markerUnits", "markerwidth": "markerWidth", "maskcontentunits": "maskContentUnits", "maskunits": "maskUnits", "numoctaves": "numOctaves", "pathlength": "pathLength", "patterncontentunits": "patternContentUnits", "patterntransform": "patternTransform", "patternunits": "patternUnits", "pointsatx": "pointsAtX", "pointsaty": "pointsAtY", "pointsatz": "pointsAtZ", "preservealpha": "preserveAlpha", "preserveaspectratio": "preserveAspectRatio", "primitiveunits": "primitiveUnits", "refx": "refX", "refy": "refY", "repeatcount": "repeatCount", "repeatdur": "repeatDur", "requiredextensions": "requiredExtensions", "requiredfeatures": "requiredFeatures", "specularconstant": "specularConstant", "specularexponent": "specularExponent", "spreadmethod": "spreadMethod", "startoffset": "startOffset", "stddeviation": "stdDeviation", "stitchtiles": "stitchTiles", "surfacescale": "surfaceScale", "systemlanguage": "systemLanguage", "tablevalues": "tableValues", "targetx": "targetX", "targety": "targetY", "textlength": "textLength", "viewbox": "viewBox", "viewtarget": "viewTarget", "xchannelselector": "xChannelSelector", "ychannelselector": "yChannelSelector", "zoomandpan": "zoomAndPan" } adjustMathMLAttributes = {"definitionurl": "definitionURL"} adjustForeignAttributes = { "xlink:actuate": ("xlink", "actuate", namespaces["xlink"]), "xlink:arcrole": ("xlink", "arcrole", namespaces["xlink"]), "xlink:href": ("xlink", "href", namespaces["xlink"]), "xlink:role": ("xlink", "role", namespaces["xlink"]), "xlink:show": ("xlink", "show", namespaces["xlink"]), "xlink:title": ("xlink", "title", namespaces["xlink"]), "xlink:type": ("xlink", "type", namespaces["xlink"]), "xml:base": ("xml", "base", namespaces["xml"]), "xml:lang": ("xml", "lang", namespaces["xml"]), "xml:space": ("xml", "space", namespaces["xml"]), "xmlns": (None, "xmlns", namespaces["xmlns"]), "xmlns:xlink": ("xmlns", "xlink", namespaces["xmlns"]) } unadjustForeignAttributes = dict([((ns, local), qname) for qname, (prefix, local, ns) in adjustForeignAttributes.items()]) spaceCharacters = frozenset([ "\t", "\n", "\u000C", " ", "\r" ]) tableInsertModeElements = frozenset([ "table", "tbody", "tfoot", "thead", "tr" ]) asciiLowercase = frozenset(string.ascii_lowercase) asciiUppercase = frozenset(string.ascii_uppercase) asciiLetters = frozenset(string.ascii_letters) digits = frozenset(string.digits) hexDigits = frozenset(string.hexdigits) asciiUpper2Lower = dict([(ord(c), ord(c.lower())) for c in string.ascii_uppercase]) # Heading elements need to be ordered headingElements = ( "h1", "h2", "h3", "h4", "h5", "h6" ) voidElements = frozenset([ "base", "command", "event-source", "link", "meta", "hr", "br", "img", "embed", "param", "area", "col", "input", "source", "track" ]) cdataElements = frozenset(['title', 'textarea']) rcdataElements = frozenset([ 'style', 'script', 'xmp', 'iframe', 'noembed', 'noframes', 'noscript' ]) booleanAttributes = { "": frozenset(["irrelevant", "itemscope"]), "style": frozenset(["scoped"]), "img": frozenset(["ismap"]), "audio": frozenset(["autoplay", "controls"]), "video": frozenset(["autoplay", "controls"]), "script": frozenset(["defer", "async"]), "details": frozenset(["open"]), "datagrid": frozenset(["multiple", "disabled"]), "command": frozenset(["hidden", "disabled", "checked", "default"]), "hr": frozenset(["noshade"]), "menu": frozenset(["autosubmit"]), "fieldset": frozenset(["disabled", "readonly"]), "option": frozenset(["disabled", "readonly", "selected"]), "optgroup": frozenset(["disabled", "readonly"]), "button": frozenset(["disabled", "autofocus"]), "input": frozenset(["disabled", "readonly", "required", "autofocus", "checked", "ismap"]), "select": frozenset(["disabled", "readonly", "autofocus", "multiple"]), "output": frozenset(["disabled", "readonly"]), "iframe": frozenset(["seamless"]), } # entitiesWindows1252 has to be _ordered_ and needs to have an index. It # therefore can't be a frozenset. entitiesWindows1252 = ( 8364, # 0x80 0x20AC EURO SIGN 65533, # 0x81 UNDEFINED 8218, # 0x82 0x201A SINGLE LOW-9 QUOTATION MARK 402, # 0x83 0x0192 LATIN SMALL LETTER F WITH HOOK 8222, # 0x84 0x201E DOUBLE LOW-9 QUOTATION MARK 8230, # 0x85 0x2026 HORIZONTAL ELLIPSIS 8224, # 0x86 0x2020 DAGGER 8225, # 0x87 0x2021 DOUBLE DAGGER 710, # 0x88 0x02C6 MODIFIER LETTER CIRCUMFLEX ACCENT 8240, # 0x89 0x2030 PER MILLE SIGN 352, # 0x8A 0x0160 LATIN CAPITAL LETTER S WITH CARON 8249, # 0x8B 0x2039 SINGLE LEFT-POINTING ANGLE QUOTATION MARK 338, # 0x8C 0x0152 LATIN CAPITAL LIGATURE OE 65533, # 0x8D UNDEFINED 381, # 0x8E 0x017D LATIN CAPITAL LETTER Z WITH CARON 65533, # 0x8F UNDEFINED 65533, # 0x90 UNDEFINED 8216, # 0x91 0x2018 LEFT SINGLE QUOTATION MARK 8217, # 0x92 0x2019 RIGHT SINGLE QUOTATION MARK 8220, # 0x93 0x201C LEFT DOUBLE QUOTATION MARK 8221, # 0x94 0x201D RIGHT DOUBLE QUOTATION MARK 8226, # 0x95 0x2022 BULLET 8211, # 0x96 0x2013 EN DASH 8212, # 0x97 0x2014 EM DASH 732, # 0x98 0x02DC SMALL TILDE 8482, # 0x99 0x2122 TRADE MARK SIGN 353, # 0x9A 0x0161 LATIN SMALL LETTER S WITH CARON 8250, # 0x9B 0x203A SINGLE RIGHT-POINTING ANGLE QUOTATION MARK 339, # 0x9C 0x0153 LATIN SMALL LIGATURE OE 65533, # 0x9D UNDEFINED 382, # 0x9E 0x017E LATIN SMALL LETTER Z WITH CARON 376 # 0x9F 0x0178 LATIN CAPITAL LETTER Y WITH DIAERESIS ) xmlEntities = frozenset(['lt;', 'gt;', 'amp;', 'apos;', 'quot;']) entities = { "AElig": "\xc6", "AElig;": "\xc6", "AMP": "&", "AMP;": "&", "Aacute": "\xc1", "Aacute;": "\xc1", "Abreve;": "\u0102", "Acirc": "\xc2", "Acirc;": "\xc2", "Acy;": "\u0410", "Afr;": "\U0001d504", "Agrave": "\xc0", "Agrave;": "\xc0", "Alpha;": "\u0391", "Amacr;": "\u0100", "And;": "\u2a53", "Aogon;": "\u0104", "Aopf;": "\U0001d538", "ApplyFunction;": "\u2061", "Aring": "\xc5", "Aring;": "\xc5", "Ascr;": "\U0001d49c", "Assign;": "\u2254", "Atilde": "\xc3", "Atilde;": "\xc3", "Auml": "\xc4", "Auml;": "\xc4", "Backslash;": "\u2216", "Barv;": "\u2ae7", "Barwed;": "\u2306", "Bcy;": "\u0411", "Because;": "\u2235", "Bernoullis;": "\u212c", "Beta;": "\u0392", "Bfr;": "\U0001d505", "Bopf;": "\U0001d539", "Breve;": "\u02d8", "Bscr;": "\u212c", "Bumpeq;": "\u224e", "CHcy;": "\u0427", "COPY": "\xa9", "COPY;": "\xa9", "Cacute;": "\u0106", "Cap;": "\u22d2", "CapitalDifferentialD;": "\u2145", "Cayleys;": "\u212d", "Ccaron;": "\u010c", "Ccedil": "\xc7", "Ccedil;": "\xc7", "Ccirc;": "\u0108", "Cconint;": "\u2230", "Cdot;": "\u010a", "Cedilla;": "\xb8", "CenterDot;": "\xb7", "Cfr;": "\u212d", "Chi;": "\u03a7", "CircleDot;": "\u2299", "CircleMinus;": "\u2296", "CirclePlus;": "\u2295", "CircleTimes;": "\u2297", "ClockwiseContourIntegral;": "\u2232", "CloseCurlyDoubleQuote;": "\u201d", "CloseCurlyQuote;": "\u2019", "Colon;": "\u2237", "Colone;": "\u2a74", "Congruent;": "\u2261", "Conint;": "\u222f", "ContourIntegral;": "\u222e", "Copf;": "\u2102", "Coproduct;": "\u2210", "CounterClockwiseContourIntegral;": "\u2233", "Cross;": "\u2a2f", "Cscr;": "\U0001d49e", "Cup;": "\u22d3", "CupCap;": "\u224d", "DD;": "\u2145", "DDotrahd;": "\u2911", "DJcy;": "\u0402", "DScy;": "\u0405", "DZcy;": "\u040f", "Dagger;": "\u2021", "Darr;": "\u21a1", "Dashv;": "\u2ae4", "Dcaron;": "\u010e", "Dcy;": "\u0414", "Del;": "\u2207", "Delta;": "\u0394", "Dfr;": "\U0001d507", "DiacriticalAcute;": "\xb4", "DiacriticalDot;": "\u02d9", "DiacriticalDoubleAcute;": "\u02dd", "DiacriticalGrave;": "`", "DiacriticalTilde;": "\u02dc", "Diamond;": "\u22c4", "DifferentialD;": "\u2146", "Dopf;": "\U0001d53b", "Dot;": "\xa8", "DotDot;": "\u20dc", "DotEqual;": "\u2250", "DoubleContourIntegral;": "\u222f", "DoubleDot;": "\xa8", "DoubleDownArrow;": "\u21d3", "DoubleLeftArrow;": "\u21d0", "DoubleLeftRightArrow;": "\u21d4", "DoubleLeftTee;": "\u2ae4", "DoubleLongLeftArrow;": "\u27f8", "DoubleLongLeftRightArrow;": "\u27fa", "DoubleLongRightArrow;": "\u27f9", "DoubleRightArrow;": "\u21d2", "DoubleRightTee;": "\u22a8", "DoubleUpArrow;": "\u21d1", "DoubleUpDownArrow;": "\u21d5", "DoubleVerticalBar;": "\u2225", "DownArrow;": "\u2193", "DownArrowBar;": "\u2913", "DownArrowUpArrow;": "\u21f5", "DownBreve;": "\u0311", "DownLeftRightVector;": "\u2950", "DownLeftTeeVector;": "\u295e", "DownLeftVector;": "\u21bd", "DownLeftVectorBar;": "\u2956", "DownRightTeeVector;": "\u295f", "DownRightVector;": "\u21c1", "DownRightVectorBar;": "\u2957", "DownTee;": "\u22a4", "DownTeeArrow;": "\u21a7", "Downarrow;": "\u21d3", "Dscr;": "\U0001d49f", "Dstrok;": "\u0110", "ENG;": "\u014a", "ETH": "\xd0", "ETH;": "\xd0", "Eacute": "\xc9", "Eacute;": "\xc9", "Ecaron;": "\u011a", "Ecirc": "\xca", "Ecirc;": "\xca", "Ecy;": "\u042d", "Edot;": "\u0116", "Efr;": "\U0001d508", "Egrave": "\xc8", "Egrave;": "\xc8", "Element;": "\u2208", "Emacr;": "\u0112", "EmptySmallSquare;": "\u25fb", "EmptyVerySmallSquare;": "\u25ab", "Eogon;": "\u0118", "Eopf;": "\U0001d53c", "Epsilon;": "\u0395", "Equal;": "\u2a75", "EqualTilde;": "\u2242", "Equilibrium;": "\u21cc", "Escr;": "\u2130", "Esim;": "\u2a73", "Eta;": "\u0397", "Euml": "\xcb", "Euml;": "\xcb", "Exists;": "\u2203", "ExponentialE;": "\u2147", "Fcy;": "\u0424", "Ffr;": "\U0001d509", "FilledSmallSquare;": "\u25fc", "FilledVerySmallSquare;": "\u25aa", "Fopf;": "\U0001d53d", "ForAll;": "\u2200", "Fouriertrf;": "\u2131", "Fscr;": "\u2131", "GJcy;": "\u0403", "GT": ">", "GT;": ">", "Gamma;": "\u0393", "Gammad;": "\u03dc", "Gbreve;": "\u011e", "Gcedil;": "\u0122", "Gcirc;": "\u011c", "Gcy;": "\u0413", "Gdot;": "\u0120", "Gfr;": "\U0001d50a", "Gg;": "\u22d9", "Gopf;": "\U0001d53e", "GreaterEqual;": "\u2265", "GreaterEqualLess;": "\u22db", "GreaterFullEqual;": "\u2267", "GreaterGreater;": "\u2aa2", "GreaterLess;": "\u2277", "GreaterSlantEqual;": "\u2a7e", "GreaterTilde;": "\u2273", "Gscr;": "\U0001d4a2", "Gt;": "\u226b", "HARDcy;": "\u042a", "Hacek;": "\u02c7", "Hat;": "^", "Hcirc;": "\u0124", "Hfr;": "\u210c", "HilbertSpace;": "\u210b", "Hopf;": "\u210d", "HorizontalLine;": "\u2500", "Hscr;": "\u210b", "Hstrok;": "\u0126", "HumpDownHump;": "\u224e", "HumpEqual;": "\u224f", "IEcy;": "\u0415", "IJlig;": "\u0132", "IOcy;": "\u0401", "Iacute": "\xcd", "Iacute;": "\xcd", "Icirc": "\xce", "Icirc;": "\xce", "Icy;": "\u0418", "Idot;": "\u0130", "Ifr;": "\u2111", "Igrave": "\xcc", "Igrave;": "\xcc", "Im;": "\u2111", "Imacr;": "\u012a", "ImaginaryI;": "\u2148", "Implies;": "\u21d2", "Int;": "\u222c", "Integral;": "\u222b", "Intersection;": "\u22c2", "InvisibleComma;": "\u2063", "InvisibleTimes;": "\u2062", "Iogon;": "\u012e", "Iopf;": "\U0001d540", "Iota;": "\u0399", "Iscr;": "\u2110", "Itilde;": "\u0128", "Iukcy;": "\u0406", "Iuml": "\xcf", "Iuml;": "\xcf", "Jcirc;": "\u0134", "Jcy;": "\u0419", "Jfr;": "\U0001d50d", "Jopf;": "\U0001d541", "Jscr;": "\U0001d4a5", "Jsercy;": "\u0408", "Jukcy;": "\u0404", "KHcy;": "\u0425", "KJcy;": "\u040c", "Kappa;": "\u039a", "Kcedil;": "\u0136", "Kcy;": "\u041a", "Kfr;": "\U0001d50e", "Kopf;": "\U0001d542", "Kscr;": "\U0001d4a6", "LJcy;": "\u0409", "LT": "<", "LT;": "<", "Lacute;": "\u0139", "Lambda;": "\u039b", "Lang;": "\u27ea", "Laplacetrf;": "\u2112", "Larr;": "\u219e", "Lcaron;": "\u013d", "Lcedil;": "\u013b", "Lcy;": "\u041b", "LeftAngleBracket;": "\u27e8", "LeftArrow;": "\u2190", "LeftArrowBar;": "\u21e4", "LeftArrowRightArrow;": "\u21c6", "LeftCeiling;": "\u2308", "LeftDoubleBracket;": "\u27e6", "LeftDownTeeVector;": "\u2961", "LeftDownVector;": "\u21c3", "LeftDownVectorBar;": "\u2959", "LeftFloor;": "\u230a", "LeftRightArrow;": "\u2194", "LeftRightVector;": "\u294e", "LeftTee;": "\u22a3", "LeftTeeArrow;": "\u21a4", "LeftTeeVector;": "\u295a", "LeftTriangle;": "\u22b2", "LeftTriangleBar;": "\u29cf", "LeftTriangleEqual;": "\u22b4", "LeftUpDownVector;": "\u2951", "LeftUpTeeVector;": "\u2960", "LeftUpVector;": "\u21bf", "LeftUpVectorBar;": "\u2958", "LeftVector;": "\u21bc", "LeftVectorBar;": "\u2952", "Leftarrow;": "\u21d0", "Leftrightarrow;": "\u21d4", "LessEqualGreater;": "\u22da", "LessFullEqual;": "\u2266", "LessGreater;": "\u2276", "LessLess;": "\u2aa1", "LessSlantEqual;": "\u2a7d", "LessTilde;": "\u2272", "Lfr;": "\U0001d50f", "Ll;": "\u22d8", "Lleftarrow;": "\u21da", "Lmidot;": "\u013f", "LongLeftArrow;": "\u27f5", "LongLeftRightArrow;": "\u27f7", "LongRightArrow;": "\u27f6", "Longleftarrow;": "\u27f8", "Longleftrightarrow;": "\u27fa", "Longrightarrow;": "\u27f9", "Lopf;": "\U0001d543", "LowerLeftArrow;": "\u2199", "LowerRightArrow;": "\u2198", "Lscr;": "\u2112", "Lsh;": "\u21b0", "Lstrok;": "\u0141", "Lt;": "\u226a", "Map;": "\u2905", "Mcy;": "\u041c", "MediumSpace;": "\u205f", "Mellintrf;": "\u2133", "Mfr;": "\U0001d510", "MinusPlus;": "\u2213", "Mopf;": "\U0001d544", "Mscr;": "\u2133", "Mu;": "\u039c", "NJcy;": "\u040a", "Nacute;": "\u0143", "Ncaron;": "\u0147", "Ncedil;": "\u0145", "Ncy;": "\u041d", "NegativeMediumSpace;": "\u200b", "NegativeThickSpace;": "\u200b", "NegativeThinSpace;": "\u200b", "NegativeVeryThinSpace;": "\u200b", "NestedGreaterGreater;": "\u226b", "NestedLessLess;": "\u226a", "NewLine;": "\n", "Nfr;": "\U0001d511", "NoBreak;": "\u2060", "NonBreakingSpace;": "\xa0", "Nopf;": "\u2115", "Not;": "\u2aec", "NotCongruent;": "\u2262", "NotCupCap;": "\u226d", "NotDoubleVerticalBar;": "\u2226", "NotElement;": "\u2209", "NotEqual;": "\u2260", "NotEqualTilde;": "\u2242\u0338", "NotExists;": "\u2204", "NotGreater;": "\u226f", "NotGreaterEqual;": "\u2271", "NotGreaterFullEqual;": "\u2267\u0338", "NotGreaterGreater;": "\u226b\u0338", "NotGreaterLess;": "\u2279", "NotGreaterSlantEqual;": "\u2a7e\u0338", "NotGreaterTilde;": "\u2275", "NotHumpDownHump;": "\u224e\u0338", "NotHumpEqual;": "\u224f\u0338", "NotLeftTriangle;": "\u22ea", "NotLeftTriangleBar;": "\u29cf\u0338", "NotLeftTriangleEqual;": "\u22ec", "NotLess;": "\u226e", "NotLessEqual;": "\u2270", "NotLessGreater;": "\u2278", "NotLessLess;": "\u226a\u0338", "NotLessSlantEqual;": "\u2a7d\u0338", "NotLessTilde;": "\u2274", "NotNestedGreaterGreater;": "\u2aa2\u0338", "NotNestedLessLess;": "\u2aa1\u0338", "NotPrecedes;": "\u2280", "NotPrecedesEqual;": "\u2aaf\u0338", "NotPrecedesSlantEqual;": "\u22e0", "NotReverseElement;": "\u220c", "NotRightTriangle;": "\u22eb", "NotRightTriangleBar;": "\u29d0\u0338", "NotRightTriangleEqual;": "\u22ed", "NotSquareSubset;": "\u228f\u0338", "NotSquareSubsetEqual;": "\u22e2", "NotSquareSuperset;": "\u2290\u0338", "NotSquareSupersetEqual;": "\u22e3", "NotSubset;": "\u2282\u20d2", "NotSubsetEqual;": "\u2288", "NotSucceeds;": "\u2281", "NotSucceedsEqual;": "\u2ab0\u0338", "NotSucceedsSlantEqual;": "\u22e1", "NotSucceedsTilde;": "\u227f\u0338", "NotSuperset;": "\u2283\u20d2", "NotSupersetEqual;": "\u2289", "NotTilde;": "\u2241", "NotTildeEqual;": "\u2244", "NotTildeFullEqual;": "\u2247", "NotTildeTilde;": "\u2249", "NotVerticalBar;": "\u2224", "Nscr;": "\U0001d4a9", "Ntilde": "\xd1", "Ntilde;": "\xd1", "Nu;": "\u039d", "OElig;": "\u0152", "Oacute": "\xd3", "Oacute;": "\xd3", "Ocirc": "\xd4", "Ocirc;": "\xd4", "Ocy;": "\u041e", "Odblac;": "\u0150", "Ofr;": "\U0001d512", "Ograve": "\xd2", "Ograve;": "\xd2", "Omacr;": "\u014c", "Omega;": "\u03a9", "Omicron;": "\u039f", "Oopf;": "\U0001d546", "OpenCurlyDoubleQuote;": "\u201c", "OpenCurlyQuote;": "\u2018", "Or;": "\u2a54", "Oscr;": "\U0001d4aa", "Oslash": "\xd8", "Oslash;": "\xd8", "Otilde": "\xd5", "Otilde;": "\xd5", "Otimes;": "\u2a37", "Ouml": "\xd6", "Ouml;": "\xd6", "OverBar;": "\u203e", "OverBrace;": "\u23de", "OverBracket;": "\u23b4", "OverParenthesis;": "\u23dc", "PartialD;": "\u2202", "Pcy;": "\u041f", "Pfr;": "\U0001d513", "Phi;": "\u03a6", "Pi;": "\u03a0", "PlusMinus;": "\xb1", "Poincareplane;": "\u210c", "Popf;": "\u2119", "Pr;": "\u2abb", "Precedes;": "\u227a", "PrecedesEqual;": "\u2aaf", "PrecedesSlantEqual;": "\u227c", "PrecedesTilde;": "\u227e", "Prime;": "\u2033", "Product;": "\u220f", "Proportion;": "\u2237", "Proportional;": "\u221d", "Pscr;": "\U0001d4ab", "Psi;": "\u03a8", "QUOT": "\"", "QUOT;": "\"", "Qfr;": "\U0001d514", "Qopf;": "\u211a", "Qscr;": "\U0001d4ac", "RBarr;": "\u2910", "REG": "\xae", "REG;": "\xae", "Racute;": "\u0154", "Rang;": "\u27eb", "Rarr;": "\u21a0", "Rarrtl;": "\u2916", "Rcaron;": "\u0158", "Rcedil;": "\u0156", "Rcy;": "\u0420", "Re;": "\u211c", "ReverseElement;": "\u220b", "ReverseEquilibrium;": "\u21cb", "ReverseUpEquilibrium;": "\u296f", "Rfr;": "\u211c", "Rho;": "\u03a1", "RightAngleBracket;": "\u27e9", "RightArrow;": "\u2192", "RightArrowBar;": "\u21e5", "RightArrowLeftArrow;": "\u21c4", "RightCeiling;": "\u2309", "RightDoubleBracket;": "\u27e7", "RightDownTeeVector;": "\u295d", "RightDownVector;": "\u21c2", "RightDownVectorBar;": "\u2955", "RightFloor;": "\u230b", "RightTee;": "\u22a2", "RightTeeArrow;": "\u21a6", "RightTeeVector;": "\u295b", "RightTriangle;": "\u22b3", "RightTriangleBar;": "\u29d0", "RightTriangleEqual;": "\u22b5", "RightUpDownVector;": "\u294f", "RightUpTeeVector;": "\u295c", "RightUpVector;": "\u21be", "RightUpVectorBar;": "\u2954", "RightVector;": "\u21c0", "RightVectorBar;": "\u2953", "Rightarrow;": "\u21d2", "Ropf;": "\u211d", "RoundImplies;": "\u2970", "Rrightarrow;": "\u21db", "Rscr;": "\u211b", "Rsh;": "\u21b1", "RuleDelayed;": "\u29f4", "SHCHcy;": "\u0429", "SHcy;": "\u0428", "SOFTcy;": "\u042c", "Sacute;": "\u015a", "Sc;": "\u2abc", "Scaron;": "\u0160", "Scedil;": "\u015e", "Scirc;": "\u015c", "Scy;": "\u0421", "Sfr;": "\U0001d516", "ShortDownArrow;": "\u2193", "ShortLeftArrow;": "\u2190", "ShortRightArrow;": "\u2192", "ShortUpArrow;": "\u2191", "Sigma;": "\u03a3", "SmallCircle;": "\u2218", "Sopf;": "\U0001d54a", "Sqrt;": "\u221a", "Square;": "\u25a1", "SquareIntersection;": "\u2293", "SquareSubset;": "\u228f", "SquareSubsetEqual;": "\u2291", "SquareSuperset;": "\u2290", "SquareSupersetEqual;": "\u2292", "SquareUnion;": "\u2294", "Sscr;": "\U0001d4ae", "Star;": "\u22c6", "Sub;": "\u22d0", "Subset;": "\u22d0", "SubsetEqual;": "\u2286", "Succeeds;": "\u227b", "SucceedsEqual;": "\u2ab0", "SucceedsSlantEqual;": "\u227d", "SucceedsTilde;": "\u227f", "SuchThat;": "\u220b", "Sum;": "\u2211", "Sup;": "\u22d1", "Superset;": "\u2283", "SupersetEqual;": "\u2287", "Supset;": "\u22d1", "THORN": "\xde", "THORN;": "\xde", "TRADE;": "\u2122", "TSHcy;": "\u040b", "TScy;": "\u0426", "Tab;": "\t", "Tau;": "\u03a4", "Tcaron;": "\u0164", "Tcedil;": "\u0162", "Tcy;": "\u0422", "Tfr;": "\U0001d517", "Therefore;": "\u2234", "Theta;": "\u0398", "ThickSpace;": "\u205f\u200a", "ThinSpace;": "\u2009", "Tilde;": "\u223c", "TildeEqual;": "\u2243", "TildeFullEqual;": "\u2245", "TildeTilde;": "\u2248", "Topf;": "\U0001d54b", "TripleDot;": "\u20db", "Tscr;": "\U0001d4af", "Tstrok;": "\u0166", "Uacute": "\xda", "Uacute;": "\xda", "Uarr;": "\u219f", "Uarrocir;": "\u2949", "Ubrcy;": "\u040e", "Ubreve;": "\u016c", "Ucirc": "\xdb", "Ucirc;": "\xdb", "Ucy;": "\u0423", "Udblac;": "\u0170", "Ufr;": "\U0001d518", "Ugrave": "\xd9", "Ugrave;": "\xd9", "Umacr;": "\u016a", "UnderBar;": "_", "UnderBrace;": "\u23df", "UnderBracket;": "\u23b5", "UnderParenthesis;": "\u23dd", "Union;": "\u22c3", "UnionPlus;": "\u228e", "Uogon;": "\u0172", "Uopf;": "\U0001d54c", "UpArrow;": "\u2191", "UpArrowBar;": "\u2912", "UpArrowDownArrow;": "\u21c5", "UpDownArrow;": "\u2195", "UpEquilibrium;": "\u296e", "UpTee;": "\u22a5", "UpTeeArrow;": "\u21a5", "Uparrow;": "\u21d1", "Updownarrow;": "\u21d5", "UpperLeftArrow;": "\u2196", "UpperRightArrow;": "\u2197", "Upsi;": "\u03d2", "Upsilon;": "\u03a5", "Uring;": "\u016e", "Uscr;": "\U0001d4b0", "Utilde;": "\u0168", "Uuml": "\xdc", "Uuml;": "\xdc", "VDash;": "\u22ab", "Vbar;": "\u2aeb", "Vcy;": "\u0412", "Vdash;": "\u22a9", "Vdashl;": "\u2ae6", "Vee;": "\u22c1", "Verbar;": "\u2016", "Vert;": "\u2016", "VerticalBar;": "\u2223", "VerticalLine;": "|", "VerticalSeparator;": "\u2758", "VerticalTilde;": "\u2240", "VeryThinSpace;": "\u200a", "Vfr;": "\U0001d519", "Vopf;": "\U0001d54d", "Vscr;": "\U0001d4b1", "Vvdash;": "\u22aa", "Wcirc;": "\u0174", "Wedge;": "\u22c0", "Wfr;": "\U0001d51a", "Wopf;": "\U0001d54e", "Wscr;": "\U0001d4b2", "Xfr;": "\U0001d51b", "Xi;": "\u039e", "Xopf;": "\U0001d54f", "Xscr;": "\U0001d4b3", "YAcy;": "\u042f", "YIcy;": "\u0407", "YUcy;": "\u042e", "Yacute": "\xdd", "Yacute;": "\xdd", "Ycirc;": "\u0176", "Ycy;": "\u042b", "Yfr;": "\U0001d51c", "Yopf;": "\U0001d550", "Yscr;": "\U0001d4b4", "Yuml;": "\u0178", "ZHcy;": "\u0416", "Zacute;": "\u0179", "Zcaron;": "\u017d", "Zcy;": "\u0417", "Zdot;": "\u017b", "ZeroWidthSpace;": "\u200b", "Zeta;": "\u0396", "Zfr;": "\u2128", "Zopf;": "\u2124", "Zscr;": "\U0001d4b5", "aacute": "\xe1", "aacute;": "\xe1", "abreve;": "\u0103", "ac;": "\u223e", "acE;": "\u223e\u0333", "acd;": "\u223f", "acirc": "\xe2", "acirc;": "\xe2", "acute": "\xb4", "acute;": "\xb4", "acy;": "\u0430", "aelig": "\xe6", "aelig;": "\xe6", "af;": "\u2061", "afr;": "\U0001d51e", "agrave": "\xe0", "agrave;": "\xe0", "alefsym;": "\u2135", "aleph;": "\u2135", "alpha;": "\u03b1", "amacr;": "\u0101", "amalg;": "\u2a3f", "amp": "&", "amp;": "&", "and;": "\u2227", "andand;": "\u2a55", "andd;": "\u2a5c", "andslope;": "\u2a58", "andv;": "\u2a5a", "ang;": "\u2220", "ange;": "\u29a4", "angle;": "\u2220", "angmsd;": "\u2221", "angmsdaa;": "\u29a8", "angmsdab;": "\u29a9", "angmsdac;": "\u29aa", "angmsdad;": "\u29ab", "angmsdae;": "\u29ac", "angmsdaf;": "\u29ad", "angmsdag;": "\u29ae", "angmsdah;": "\u29af", "angrt;": "\u221f", "angrtvb;": "\u22be", "angrtvbd;": "\u299d", "angsph;": "\u2222", "angst;": "\xc5", "angzarr;": "\u237c", "aogon;": "\u0105", "aopf;": "\U0001d552", "ap;": "\u2248", "apE;": "\u2a70", "apacir;": "\u2a6f", "ape;": "\u224a", "apid;": "\u224b", "apos;": "'", "approx;": "\u2248", "approxeq;": "\u224a", "aring": "\xe5", "aring;": "\xe5", "ascr;": "\U0001d4b6", "ast;": "*", "asymp;": "\u2248", "asympeq;": "\u224d", "atilde": "\xe3", "atilde;": "\xe3", "auml": "\xe4", "auml;": "\xe4", "awconint;": "\u2233", "awint;": "\u2a11", "bNot;": "\u2aed", "backcong;": "\u224c", "backepsilon;": "\u03f6", "backprime;": "\u2035", "backsim;": "\u223d", "backsimeq;": "\u22cd", "barvee;": "\u22bd", "barwed;": "\u2305", "barwedge;": "\u2305", "bbrk;": "\u23b5", "bbrktbrk;": "\u23b6", "bcong;": "\u224c", "bcy;": "\u0431", "bdquo;": "\u201e", "becaus;": "\u2235", "because;": "\u2235", "bemptyv;": "\u29b0", "bepsi;": "\u03f6", "bernou;": "\u212c", "beta;": "\u03b2", "beth;": "\u2136", "between;": "\u226c", "bfr;": "\U0001d51f", "bigcap;": "\u22c2", "bigcirc;": "\u25ef", "bigcup;": "\u22c3", "bigodot;": "\u2a00", "bigoplus;": "\u2a01", "bigotimes;": "\u2a02", "bigsqcup;": "\u2a06", "bigstar;": "\u2605", "bigtriangledown;": "\u25bd", "bigtriangleup;": "\u25b3", "biguplus;": "\u2a04", "bigvee;": "\u22c1", "bigwedge;": "\u22c0", "bkarow;": "\u290d", "blacklozenge;": "\u29eb", "blacksquare;": "\u25aa", "blacktriangle;": "\u25b4", "blacktriangledown;": "\u25be", "blacktriangleleft;": "\u25c2", "blacktriangleright;": "\u25b8", "blank;": "\u2423", "blk12;": "\u2592", "blk14;": "\u2591", "blk34;": "\u2593", "block;": "\u2588", "bne;": "=\u20e5", "bnequiv;": "\u2261\u20e5", "bnot;": "\u2310", "bopf;": "\U0001d553", "bot;": "\u22a5", "bottom;": "\u22a5", "bowtie;": "\u22c8", "boxDL;": "\u2557", "boxDR;": "\u2554", "boxDl;": "\u2556", "boxDr;": "\u2553", "boxH;": "\u2550", "boxHD;": "\u2566", "boxHU;": "\u2569", "boxHd;": "\u2564", "boxHu;": "\u2567", "boxUL;": "\u255d", "boxUR;": "\u255a", "boxUl;": "\u255c", "boxUr;": "\u2559", "boxV;": "\u2551", "boxVH;": "\u256c", "boxVL;": "\u2563", "boxVR;": "\u2560", "boxVh;": "\u256b", "boxVl;": "\u2562", "boxVr;": "\u255f", "boxbox;": "\u29c9", "boxdL;": "\u2555", "boxdR;": "\u2552", "boxdl;": "\u2510", "boxdr;": "\u250c", "boxh;": "\u2500", "boxhD;": "\u2565", "boxhU;": "\u2568", "boxhd;": "\u252c", "boxhu;": "\u2534", "boxminus;": "\u229f", "boxplus;": "\u229e", "boxtimes;": "\u22a0", "boxuL;": "\u255b", "boxuR;": "\u2558", "boxul;": "\u2518", "boxur;": "\u2514", "boxv;": "\u2502", "boxvH;": "\u256a", "boxvL;": "\u2561", "boxvR;": "\u255e", "boxvh;": "\u253c", "boxvl;": "\u2524", "boxvr;": "\u251c", "bprime;": "\u2035", "breve;": "\u02d8", "brvbar": "\xa6", "brvbar;": "\xa6", "bscr;": "\U0001d4b7", "bsemi;": "\u204f", "bsim;": "\u223d", "bsime;": "\u22cd", "bsol;": "\\", "bsolb;": "\u29c5", "bsolhsub;": "\u27c8", "bull;": "\u2022", "bullet;": "\u2022", "bump;": "\u224e", "bumpE;": "\u2aae", "bumpe;": "\u224f", "bumpeq;": "\u224f", "cacute;": "\u0107", "cap;": "\u2229", "capand;": "\u2a44", "capbrcup;": "\u2a49", "capcap;": "\u2a4b", "capcup;": "\u2a47", "capdot;": "\u2a40", "caps;": "\u2229\ufe00", "caret;": "\u2041", "caron;": "\u02c7", "ccaps;": "\u2a4d", "ccaron;": "\u010d", "ccedil": "\xe7", "ccedil;": "\xe7", "ccirc;": "\u0109", "ccups;": "\u2a4c", "ccupssm;": "\u2a50", "cdot;": "\u010b", "cedil": "\xb8", "cedil;": "\xb8", "cemptyv;": "\u29b2", "cent": "\xa2", "cent;": "\xa2", "centerdot;": "\xb7", "cfr;": "\U0001d520", "chcy;": "\u0447", "check;": "\u2713", "checkmark;": "\u2713", "chi;": "\u03c7", "cir;": "\u25cb", "cirE;": "\u29c3", "circ;": "\u02c6", "circeq;": "\u2257", "circlearrowleft;": "\u21ba", "circlearrowright;": "\u21bb", "circledR;": "\xae", "circledS;": "\u24c8", "circledast;": "\u229b", "circledcirc;": "\u229a", "circleddash;": "\u229d", "cire;": "\u2257", "cirfnint;": "\u2a10", "cirmid;": "\u2aef", "cirscir;": "\u29c2", "clubs;": "\u2663", "clubsuit;": "\u2663", "colon;": ":", "colone;": "\u2254", "coloneq;": "\u2254", "comma;": ",", "commat;": "@", "comp;": "\u2201", "compfn;": "\u2218", "complement;": "\u2201", "complexes;": "\u2102", "cong;": "\u2245", "congdot;": "\u2a6d", "conint;": "\u222e", "copf;": "\U0001d554", "coprod;": "\u2210", "copy": "\xa9", "copy;": "\xa9", "copysr;": "\u2117", "crarr;": "\u21b5", "cross;": "\u2717", "cscr;": "\U0001d4b8", "csub;": "\u2acf", "csube;": "\u2ad1", "csup;": "\u2ad0", "csupe;": "\u2ad2", "ctdot;": "\u22ef", "cudarrl;": "\u2938", "cudarrr;": "\u2935", "cuepr;": "\u22de", "cuesc;": "\u22df", "cularr;": "\u21b6", "cularrp;": "\u293d", "cup;": "\u222a", "cupbrcap;": "\u2a48", "cupcap;": "\u2a46", "cupcup;": "\u2a4a", "cupdot;": "\u228d", "cupor;": "\u2a45", "cups;": "\u222a\ufe00", "curarr;": "\u21b7", "curarrm;": "\u293c", "curlyeqprec;": "\u22de", "curlyeqsucc;": "\u22df", "curlyvee;": "\u22ce", "curlywedge;": "\u22cf", "curren": "\xa4", "curren;": "\xa4", "curvearrowleft;": "\u21b6", "curvearrowright;": "\u21b7", "cuvee;": "\u22ce", "cuwed;": "\u22cf", "cwconint;": "\u2232", "cwint;": "\u2231", "cylcty;": "\u232d", "dArr;": "\u21d3", "dHar;": "\u2965", "dagger;": "\u2020", "daleth;": "\u2138", "darr;": "\u2193", "dash;": "\u2010", "dashv;": "\u22a3", "dbkarow;": "\u290f", "dblac;": "\u02dd", "dcaron;": "\u010f", "dcy;": "\u0434", "dd;": "\u2146", "ddagger;": "\u2021", "ddarr;": "\u21ca", "ddotseq;": "\u2a77", "deg": "\xb0", "deg;": "\xb0", "delta;": "\u03b4", "demptyv;": "\u29b1", "dfisht;": "\u297f", "dfr;": "\U0001d521", "dharl;": "\u21c3", "dharr;": "\u21c2", "diam;": "\u22c4", "diamond;": "\u22c4", "diamondsuit;": "\u2666", "diams;": "\u2666", "die;": "\xa8", "digamma;": "\u03dd", "disin;": "\u22f2", "div;": "\xf7", "divide": "\xf7", "divide;": "\xf7", "divideontimes;": "\u22c7", "divonx;": "\u22c7", "djcy;": "\u0452", "dlcorn;": "\u231e", "dlcrop;": "\u230d", "dollar;": "$", "dopf;": "\U0001d555", "dot;": "\u02d9", "doteq;": "\u2250", "doteqdot;": "\u2251", "dotminus;": "\u2238", "dotplus;": "\u2214", "dotsquare;": "\u22a1", "doublebarwedge;": "\u2306", "downarrow;": "\u2193", "downdownarrows;": "\u21ca", "downharpoonleft;": "\u21c3", "downharpoonright;": "\u21c2", "drbkarow;": "\u2910", "drcorn;": "\u231f", "drcrop;": "\u230c", "dscr;": "\U0001d4b9", "dscy;": "\u0455", "dsol;": "\u29f6", "dstrok;": "\u0111", "dtdot;": "\u22f1", "dtri;": "\u25bf", "dtrif;": "\u25be", "duarr;": "\u21f5", "duhar;": "\u296f", "dwangle;": "\u29a6", "dzcy;": "\u045f", "dzigrarr;": "\u27ff", "eDDot;": "\u2a77", "eDot;": "\u2251", "eacute": "\xe9", "eacute;": "\xe9", "easter;": "\u2a6e", "ecaron;": "\u011b", "ecir;": "\u2256", "ecirc": "\xea", "ecirc;": "\xea", "ecolon;": "\u2255", "ecy;": "\u044d", "edot;": "\u0117", "ee;": "\u2147", "efDot;": "\u2252", "efr;": "\U0001d522", "eg;": "\u2a9a", "egrave": "\xe8", "egrave;": "\xe8", "egs;": "\u2a96", "egsdot;": "\u2a98", "el;": "\u2a99", "elinters;": "\u23e7", "ell;": "\u2113", "els;": "\u2a95", "elsdot;": "\u2a97", "emacr;": "\u0113", "empty;": "\u2205", "emptyset;": "\u2205", "emptyv;": "\u2205", "emsp13;": "\u2004", "emsp14;": "\u2005", "emsp;": "\u2003", "eng;": "\u014b", "ensp;": "\u2002", "eogon;": "\u0119", "eopf;": "\U0001d556", "epar;": "\u22d5", "eparsl;": "\u29e3", "eplus;": "\u2a71", "epsi;": "\u03b5", "epsilon;": "\u03b5", "epsiv;": "\u03f5", "eqcirc;": "\u2256", "eqcolon;": "\u2255", "eqsim;": "\u2242", "eqslantgtr;": "\u2a96", "eqslantless;": "\u2a95", "equals;": "=", "equest;": "\u225f", "equiv;": "\u2261", "equivDD;": "\u2a78", "eqvparsl;": "\u29e5", "erDot;": "\u2253", "erarr;": "\u2971", "escr;": "\u212f", "esdot;": "\u2250", "esim;": "\u2242", "eta;": "\u03b7", "eth": "\xf0", "eth;": "\xf0", "euml": "\xeb", "euml;": "\xeb", "euro;": "\u20ac", "excl;": "!", "exist;": "\u2203", "expectation;": "\u2130", "exponentiale;": "\u2147", "fallingdotseq;": "\u2252", "fcy;": "\u0444", "female;": "\u2640", "ffilig;": "\ufb03", "fflig;": "\ufb00", "ffllig;": "\ufb04", "ffr;": "\U0001d523", "filig;": "\ufb01", "fjlig;": "fj", "flat;": "\u266d", "fllig;": "\ufb02", "fltns;": "\u25b1", "fnof;": "\u0192", "fopf;": "\U0001d557", "forall;": "\u2200", "fork;": "\u22d4", "forkv;": "\u2ad9", "fpartint;": "\u2a0d", "frac12": "\xbd", "frac12;": "\xbd", "frac13;": "\u2153", "frac14": "\xbc", "frac14;": "\xbc", "frac15;": "\u2155", "frac16;": "\u2159", "frac18;": "\u215b", "frac23;": "\u2154", "frac25;": "\u2156", "frac34": "\xbe", "frac34;": "\xbe", "frac35;": "\u2157", "frac38;": "\u215c", "frac45;": "\u2158", "frac56;": "\u215a", "frac58;": "\u215d", "frac78;": "\u215e", "frasl;": "\u2044", "frown;": "\u2322", "fscr;": "\U0001d4bb", "gE;": "\u2267", "gEl;": "\u2a8c", "gacute;": "\u01f5", "gamma;": "\u03b3", "gammad;": "\u03dd", "gap;": "\u2a86", "gbreve;": "\u011f", "gcirc;": "\u011d", "gcy;": "\u0433", "gdot;": "\u0121", "ge;": "\u2265", "gel;": "\u22db", "geq;": "\u2265", "geqq;": "\u2267", "geqslant;": "\u2a7e", "ges;": "\u2a7e", "gescc;": "\u2aa9", "gesdot;": "\u2a80", "gesdoto;": "\u2a82", "gesdotol;": "\u2a84", "gesl;": "\u22db\ufe00", "gesles;": "\u2a94", "gfr;": "\U0001d524", "gg;": "\u226b", "ggg;": "\u22d9", "gimel;": "\u2137", "gjcy;": "\u0453", "gl;": "\u2277", "glE;": "\u2a92", "gla;": "\u2aa5", "glj;": "\u2aa4", "gnE;": "\u2269", "gnap;": "\u2a8a", "gnapprox;": "\u2a8a", "gne;": "\u2a88", "gneq;": "\u2a88", "gneqq;": "\u2269", "gnsim;": "\u22e7", "gopf;": "\U0001d558", "grave;": "`", "gscr;": "\u210a", "gsim;": "\u2273", "gsime;": "\u2a8e", "gsiml;": "\u2a90", "gt": ">", "gt;": ">", "gtcc;": "\u2aa7", "gtcir;": "\u2a7a", "gtdot;": "\u22d7", "gtlPar;": "\u2995", "gtquest;": "\u2a7c", "gtrapprox;": "\u2a86", "gtrarr;": "\u2978", "gtrdot;": "\u22d7", "gtreqless;": "\u22db", "gtreqqless;": "\u2a8c", "gtrless;": "\u2277", "gtrsim;": "\u2273", "gvertneqq;": "\u2269\ufe00", "gvnE;": "\u2269\ufe00", "hArr;": "\u21d4", "hairsp;": "\u200a", "half;": "\xbd", "hamilt;": "\u210b", "hardcy;": "\u044a", "harr;": "\u2194", "harrcir;": "\u2948", "harrw;": "\u21ad", "hbar;": "\u210f", "hcirc;": "\u0125", "hearts;": "\u2665", "heartsuit;": "\u2665", "hellip;": "\u2026", "hercon;": "\u22b9", "hfr;": "\U0001d525", "hksearow;": "\u2925", "hkswarow;": "\u2926", "hoarr;": "\u21ff", "homtht;": "\u223b", "hookleftarrow;": "\u21a9", "hookrightarrow;": "\u21aa", "hopf;": "\U0001d559", "horbar;": "\u2015", "hscr;": "\U0001d4bd", "hslash;": "\u210f", "hstrok;": "\u0127", "hybull;": "\u2043", "hyphen;": "\u2010", "iacute": "\xed", "iacute;": "\xed", "ic;": "\u2063", "icirc": "\xee", "icirc;": "\xee", "icy;": "\u0438", "iecy;": "\u0435", "iexcl": "\xa1", "iexcl;": "\xa1", "iff;": "\u21d4", "ifr;": "\U0001d526", "igrave": "\xec", "igrave;": "\xec", "ii;": "\u2148", "iiiint;": "\u2a0c", "iiint;": "\u222d", "iinfin;": "\u29dc", "iiota;": "\u2129", "ijlig;": "\u0133", "imacr;": "\u012b", "image;": "\u2111", "imagline;": "\u2110", "imagpart;": "\u2111", "imath;": "\u0131", "imof;": "\u22b7", "imped;": "\u01b5", "in;": "\u2208", "incare;": "\u2105", "infin;": "\u221e", "infintie;": "\u29dd", "inodot;": "\u0131", "int;": "\u222b", "intcal;": "\u22ba", "integers;": "\u2124", "intercal;": "\u22ba", "intlarhk;": "\u2a17", "intprod;": "\u2a3c", "iocy;": "\u0451", "iogon;": "\u012f", "iopf;": "\U0001d55a", "iota;": "\u03b9", "iprod;": "\u2a3c", "iquest": "\xbf", "iquest;": "\xbf", "iscr;": "\U0001d4be", "isin;": "\u2208", "isinE;": "\u22f9", "isindot;": "\u22f5", "isins;": "\u22f4", "isinsv;": "\u22f3", "isinv;": "\u2208", "it;": "\u2062", "itilde;": "\u0129", "iukcy;": "\u0456", "iuml": "\xef", "iuml;": "\xef", "jcirc;": "\u0135", "jcy;": "\u0439", "jfr;": "\U0001d527", "jmath;": "\u0237", "jopf;": "\U0001d55b", "jscr;": "\U0001d4bf", "jsercy;": "\u0458", "jukcy;": "\u0454", "kappa;": "\u03ba", "kappav;": "\u03f0", "kcedil;": "\u0137", "kcy;": "\u043a", "kfr;": "\U0001d528", "kgreen;": "\u0138", "khcy;": "\u0445", "kjcy;": "\u045c", "kopf;": "\U0001d55c", "kscr;": "\U0001d4c0", "lAarr;": "\u21da", "lArr;": "\u21d0", "lAtail;": "\u291b", "lBarr;": "\u290e", "lE;": "\u2266", "lEg;": "\u2a8b", "lHar;": "\u2962", "lacute;": "\u013a", "laemptyv;": "\u29b4", "lagran;": "\u2112", "lambda;": "\u03bb", "lang;": "\u27e8", "langd;": "\u2991", "langle;": "\u27e8", "lap;": "\u2a85", "laquo": "\xab", "laquo;": "\xab", "larr;": "\u2190", "larrb;": "\u21e4", "larrbfs;": "\u291f", "larrfs;": "\u291d", "larrhk;": "\u21a9", "larrlp;": "\u21ab", "larrpl;": "\u2939", "larrsim;": "\u2973", "larrtl;": "\u21a2", "lat;": "\u2aab", "latail;": "\u2919", "late;": "\u2aad", "lates;": "\u2aad\ufe00", "lbarr;": "\u290c", "lbbrk;": "\u2772", "lbrace;": "{", "lbrack;": "[", "lbrke;": "\u298b", "lbrksld;": "\u298f", "lbrkslu;": "\u298d", "lcaron;": "\u013e", "lcedil;": "\u013c", "lceil;": "\u2308", "lcub;": "{", "lcy;": "\u043b", "ldca;": "\u2936", "ldquo;": "\u201c", "ldquor;": "\u201e", "ldrdhar;": "\u2967", "ldrushar;": "\u294b", "ldsh;": "\u21b2", "le;": "\u2264", "leftarrow;": "\u2190", "leftarrowtail;": "\u21a2", "leftharpoondown;": "\u21bd", "leftharpoonup;": "\u21bc", "leftleftarrows;": "\u21c7", "leftrightarrow;": "\u2194", "leftrightarrows;": "\u21c6", "leftrightharpoons;": "\u21cb", "leftrightsquigarrow;": "\u21ad", "leftthreetimes;": "\u22cb", "leg;": "\u22da", "leq;": "\u2264", "leqq;": "\u2266", "leqslant;": "\u2a7d", "les;": "\u2a7d", "lescc;": "\u2aa8", "lesdot;": "\u2a7f", "lesdoto;": "\u2a81", "lesdotor;": "\u2a83", "lesg;": "\u22da\ufe00", "lesges;": "\u2a93", "lessapprox;": "\u2a85", "lessdot;": "\u22d6", "lesseqgtr;": "\u22da", "lesseqqgtr;": "\u2a8b", "lessgtr;": "\u2276", "lesssim;": "\u2272", "lfisht;": "\u297c", "lfloor;": "\u230a", "lfr;": "\U0001d529", "lg;": "\u2276", "lgE;": "\u2a91", "lhard;": "\u21bd", "lharu;": "\u21bc", "lharul;": "\u296a", "lhblk;": "\u2584", "ljcy;": "\u0459", "ll;": "\u226a", "llarr;": "\u21c7", "llcorner;": "\u231e", "llhard;": "\u296b", "lltri;": "\u25fa", "lmidot;": "\u0140", "lmoust;": "\u23b0", "lmoustache;": "\u23b0", "lnE;": "\u2268", "lnap;": "\u2a89", "lnapprox;": "\u2a89", "lne;": "\u2a87", "lneq;": "\u2a87", "lneqq;": "\u2268", "lnsim;": "\u22e6", "loang;": "\u27ec", "loarr;": "\u21fd", "lobrk;": "\u27e6", "longleftarrow;": "\u27f5", "longleftrightarrow;": "\u27f7", "longmapsto;": "\u27fc", "longrightarrow;": "\u27f6", "looparrowleft;": "\u21ab", "looparrowright;": "\u21ac", "lopar;": "\u2985", "lopf;": "\U0001d55d", "loplus;": "\u2a2d", "lotimes;": "\u2a34", "lowast;": "\u2217", "lowbar;": "_", "loz;": "\u25ca", "lozenge;": "\u25ca", "lozf;": "\u29eb", "lpar;": "(", "lparlt;": "\u2993", "lrarr;": "\u21c6", "lrcorner;": "\u231f", "lrhar;": "\u21cb", "lrhard;": "\u296d", "lrm;": "\u200e", "lrtri;": "\u22bf", "lsaquo;": "\u2039", "lscr;": "\U0001d4c1", "lsh;": "\u21b0", "lsim;": "\u2272", "lsime;": "\u2a8d", "lsimg;": "\u2a8f", "lsqb;": "[", "lsquo;": "\u2018", "lsquor;": "\u201a", "lstrok;": "\u0142", "lt": "<", "lt;": "<", "ltcc;": "\u2aa6", "ltcir;": "\u2a79", "ltdot;": "\u22d6", "lthree;": "\u22cb", "ltimes;": "\u22c9", "ltlarr;": "\u2976", "ltquest;": "\u2a7b", "ltrPar;": "\u2996", "ltri;": "\u25c3", "ltrie;": "\u22b4", "ltrif;": "\u25c2", "lurdshar;": "\u294a", "luruhar;": "\u2966", "lvertneqq;": "\u2268\ufe00", "lvnE;": "\u2268\ufe00", "mDDot;": "\u223a", "macr": "\xaf", "macr;": "\xaf", "male;": "\u2642", "malt;": "\u2720", "maltese;": "\u2720", "map;": "\u21a6", "mapsto;": "\u21a6", "mapstodown;": "\u21a7", "mapstoleft;": "\u21a4", "mapstoup;": "\u21a5", "marker;": "\u25ae", "mcomma;": "\u2a29", "mcy;": "\u043c", "mdash;": "\u2014", "measuredangle;": "\u2221", "mfr;": "\U0001d52a", "mho;": "\u2127", "micro": "\xb5", "micro;": "\xb5", "mid;": "\u2223", "midast;": "*", "midcir;": "\u2af0", "middot": "\xb7", "middot;": "\xb7", "minus;": "\u2212", "minusb;": "\u229f", "minusd;": "\u2238", "minusdu;": "\u2a2a", "mlcp;": "\u2adb", "mldr;": "\u2026", "mnplus;": "\u2213", "models;": "\u22a7", "mopf;": "\U0001d55e", "mp;": "\u2213", "mscr;": "\U0001d4c2", "mstpos;": "\u223e", "mu;": "\u03bc", "multimap;": "\u22b8", "mumap;": "\u22b8", "nGg;": "\u22d9\u0338", "nGt;": "\u226b\u20d2", "nGtv;": "\u226b\u0338", "nLeftarrow;": "\u21cd", "nLeftrightarrow;": "\u21ce", "nLl;": "\u22d8\u0338", "nLt;": "\u226a\u20d2", "nLtv;": "\u226a\u0338", "nRightarrow;": "\u21cf", "nVDash;": "\u22af", "nVdash;": "\u22ae", "nabla;": "\u2207", "nacute;": "\u0144", "nang;": "\u2220\u20d2", "nap;": "\u2249", "napE;": "\u2a70\u0338", "napid;": "\u224b\u0338", "napos;": "\u0149", "napprox;": "\u2249", "natur;": "\u266e", "natural;": "\u266e", "naturals;": "\u2115", "nbsp": "\xa0", "nbsp;": "\xa0", "nbump;": "\u224e\u0338", "nbumpe;": "\u224f\u0338", "ncap;": "\u2a43", "ncaron;": "\u0148", "ncedil;": "\u0146", "ncong;": "\u2247", "ncongdot;": "\u2a6d\u0338", "ncup;": "\u2a42", "ncy;": "\u043d", "ndash;": "\u2013", "ne;": "\u2260", "neArr;": "\u21d7", "nearhk;": "\u2924", "nearr;": "\u2197", "nearrow;": "\u2197", "nedot;": "\u2250\u0338", "nequiv;": "\u2262", "nesear;": "\u2928", "nesim;": "\u2242\u0338", "nexist;": "\u2204", "nexists;": "\u2204", "nfr;": "\U0001d52b", "ngE;": "\u2267\u0338", "nge;": "\u2271", "ngeq;": "\u2271", "ngeqq;": "\u2267\u0338", "ngeqslant;": "\u2a7e\u0338", "nges;": "\u2a7e\u0338", "ngsim;": "\u2275", "ngt;": "\u226f", "ngtr;": "\u226f", "nhArr;": "\u21ce", "nharr;": "\u21ae", "nhpar;": "\u2af2", "ni;": "\u220b", "nis;": "\u22fc", "nisd;": "\u22fa", "niv;": "\u220b", "njcy;": "\u045a", "nlArr;": "\u21cd", "nlE;": "\u2266\u0338", "nlarr;": "\u219a", "nldr;": "\u2025", "nle;": "\u2270", "nleftarrow;": "\u219a", "nleftrightarrow;": "\u21ae", "nleq;": "\u2270", "nleqq;": "\u2266\u0338", "nleqslant;": "\u2a7d\u0338", "nles;": "\u2a7d\u0338", "nless;": "\u226e", "nlsim;": "\u2274", "nlt;": "\u226e", "nltri;": "\u22ea", "nltrie;": "\u22ec", "nmid;": "\u2224", "nopf;": "\U0001d55f", "not": "\xac", "not;": "\xac", "notin;": "\u2209", "notinE;": "\u22f9\u0338", "notindot;": "\u22f5\u0338", "notinva;": "\u2209", "notinvb;": "\u22f7", "notinvc;": "\u22f6", "notni;": "\u220c", "notniva;": "\u220c", "notnivb;": "\u22fe", "notnivc;": "\u22fd", "npar;": "\u2226", "nparallel;": "\u2226", "nparsl;": "\u2afd\u20e5", "npart;": "\u2202\u0338", "npolint;": "\u2a14", "npr;": "\u2280", "nprcue;": "\u22e0", "npre;": "\u2aaf\u0338", "nprec;": "\u2280", "npreceq;": "\u2aaf\u0338", "nrArr;": "\u21cf", "nrarr;": "\u219b", "nrarrc;": "\u2933\u0338", "nrarrw;": "\u219d\u0338", "nrightarrow;": "\u219b", "nrtri;": "\u22eb", "nrtrie;": "\u22ed", "nsc;": "\u2281", "nsccue;": "\u22e1", "nsce;": "\u2ab0\u0338", "nscr;": "\U0001d4c3", "nshortmid;": "\u2224", "nshortparallel;": "\u2226", "nsim;": "\u2241", "nsime;": "\u2244", "nsimeq;": "\u2244", "nsmid;": "\u2224", "nspar;": "\u2226", "nsqsube;": "\u22e2", "nsqsupe;": "\u22e3", "nsub;": "\u2284", "nsubE;": "\u2ac5\u0338", "nsube;": "\u2288", "nsubset;": "\u2282\u20d2", "nsubseteq;": "\u2288", "nsubseteqq;": "\u2ac5\u0338", "nsucc;": "\u2281", "nsucceq;": "\u2ab0\u0338", "nsup;": "\u2285", "nsupE;": "\u2ac6\u0338", "nsupe;": "\u2289", "nsupset;": "\u2283\u20d2", "nsupseteq;": "\u2289", "nsupseteqq;": "\u2ac6\u0338", "ntgl;": "\u2279", "ntilde": "\xf1", "ntilde;": "\xf1", "ntlg;": "\u2278", "ntriangleleft;": "\u22ea", "ntrianglelefteq;": "\u22ec", "ntriangleright;": "\u22eb", "ntrianglerighteq;": "\u22ed", "nu;": "\u03bd", "num;": "#", "numero;": "\u2116", "numsp;": "\u2007", "nvDash;": "\u22ad", "nvHarr;": "\u2904", "nvap;": "\u224d\u20d2", "nvdash;": "\u22ac", "nvge;": "\u2265\u20d2", "nvgt;": ">\u20d2", "nvinfin;": "\u29de", "nvlArr;": "\u2902", "nvle;": "\u2264\u20d2", "nvlt;": "<\u20d2", "nvltrie;": "\u22b4\u20d2", "nvrArr;": "\u2903", "nvrtrie;": "\u22b5\u20d2", "nvsim;": "\u223c\u20d2", "nwArr;": "\u21d6", "nwarhk;": "\u2923", "nwarr;": "\u2196", "nwarrow;": "\u2196", "nwnear;": "\u2927", "oS;": "\u24c8", "oacute": "\xf3", "oacute;": "\xf3", "oast;": "\u229b", "ocir;": "\u229a", "ocirc": "\xf4", "ocirc;": "\xf4", "ocy;": "\u043e", "odash;": "\u229d", "odblac;": "\u0151", "odiv;": "\u2a38", "odot;": "\u2299", "odsold;": "\u29bc", "oelig;": "\u0153", "ofcir;": "\u29bf", "ofr;": "\U0001d52c", "ogon;": "\u02db", "ograve": "\xf2", "ograve;": "\xf2", "ogt;": "\u29c1", "ohbar;": "\u29b5", "ohm;": "\u03a9", "oint;": "\u222e", "olarr;": "\u21ba", "olcir;": "\u29be", "olcross;": "\u29bb", "oline;": "\u203e", "olt;": "\u29c0", "omacr;": "\u014d", "omega;": "\u03c9", "omicron;": "\u03bf", "omid;": "\u29b6", "ominus;": "\u2296", "oopf;": "\U0001d560", "opar;": "\u29b7", "operp;": "\u29b9", "oplus;": "\u2295", "or;": "\u2228", "orarr;": "\u21bb", "ord;": "\u2a5d", "order;": "\u2134", "orderof;": "\u2134", "ordf": "\xaa", "ordf;": "\xaa", "ordm": "\xba", "ordm;": "\xba", "origof;": "\u22b6", "oror;": "\u2a56", "orslope;": "\u2a57", "orv;": "\u2a5b", "oscr;": "\u2134", "oslash": "\xf8", "oslash;": "\xf8", "osol;": "\u2298", "otilde": "\xf5", "otilde;": "\xf5", "otimes;": "\u2297", "otimesas;": "\u2a36", "ouml": "\xf6", "ouml;": "\xf6", "ovbar;": "\u233d", "par;": "\u2225", "para": "\xb6", "para;": "\xb6", "parallel;": "\u2225", "parsim;": "\u2af3", "parsl;": "\u2afd", "part;": "\u2202", "pcy;": "\u043f", "percnt;": "%", "period;": ".", "permil;": "\u2030", "perp;": "\u22a5", "pertenk;": "\u2031", "pfr;": "\U0001d52d", "phi;": "\u03c6", "phiv;": "\u03d5", "phmmat;": "\u2133", "phone;": "\u260e", "pi;": "\u03c0", "pitchfork;": "\u22d4", "piv;": "\u03d6", "planck;": "\u210f", "planckh;": "\u210e", "plankv;": "\u210f", "plus;": "+", "plusacir;": "\u2a23", "plusb;": "\u229e", "pluscir;": "\u2a22", "plusdo;": "\u2214", "plusdu;": "\u2a25", "pluse;": "\u2a72", "plusmn": "\xb1", "plusmn;": "\xb1", "plussim;": "\u2a26", "plustwo;": "\u2a27", "pm;": "\xb1", "pointint;": "\u2a15", "popf;": "\U0001d561", "pound": "\xa3", "pound;": "\xa3", "pr;": "\u227a", "prE;": "\u2ab3", "prap;": "\u2ab7", "prcue;": "\u227c", "pre;": "\u2aaf", "prec;": "\u227a", "precapprox;": "\u2ab7", "preccurlyeq;": "\u227c", "preceq;": "\u2aaf", "precnapprox;": "\u2ab9", "precneqq;": "\u2ab5", "precnsim;": "\u22e8", "precsim;": "\u227e", "prime;": "\u2032", "primes;": "\u2119", "prnE;": "\u2ab5", "prnap;": "\u2ab9", "prnsim;": "\u22e8", "prod;": "\u220f", "profalar;": "\u232e", "profline;": "\u2312", "profsurf;": "\u2313", "prop;": "\u221d", "propto;": "\u221d", "prsim;": "\u227e", "prurel;": "\u22b0", "pscr;": "\U0001d4c5", "psi;": "\u03c8", "puncsp;": "\u2008", "qfr;": "\U0001d52e", "qint;": "\u2a0c", "qopf;": "\U0001d562", "qprime;": "\u2057", "qscr;": "\U0001d4c6", "quaternions;": "\u210d", "quatint;": "\u2a16", "quest;": "?", "questeq;": "\u225f", "quot": "\"", "quot;": "\"", "rAarr;": "\u21db", "rArr;": "\u21d2", "rAtail;": "\u291c", "rBarr;": "\u290f", "rHar;": "\u2964", "race;": "\u223d\u0331", "racute;": "\u0155", "radic;": "\u221a", "raemptyv;": "\u29b3", "rang;": "\u27e9", "rangd;": "\u2992", "range;": "\u29a5", "rangle;": "\u27e9", "raquo": "\xbb", "raquo;": "\xbb", "rarr;": "\u2192", "rarrap;": "\u2975", "rarrb;": "\u21e5", "rarrbfs;": "\u2920", "rarrc;": "\u2933", "rarrfs;": "\u291e", "rarrhk;": "\u21aa", "rarrlp;": "\u21ac", "rarrpl;": "\u2945", "rarrsim;": "\u2974", "rarrtl;": "\u21a3", "rarrw;": "\u219d", "ratail;": "\u291a", "ratio;": "\u2236", "rationals;": "\u211a", "rbarr;": "\u290d", "rbbrk;": "\u2773", "rbrace;": "}", "rbrack;": "]", "rbrke;": "\u298c", "rbrksld;": "\u298e", "rbrkslu;": "\u2990", "rcaron;": "\u0159", "rcedil;": "\u0157", "rceil;": "\u2309", "rcub;": "}", "rcy;": "\u0440", "rdca;": "\u2937", "rdldhar;": "\u2969", "rdquo;": "\u201d", "rdquor;": "\u201d", "rdsh;": "\u21b3", "real;": "\u211c", "realine;": "\u211b", "realpart;": "\u211c", "reals;": "\u211d", "rect;": "\u25ad", "reg": "\xae", "reg;": "\xae", "rfisht;": "\u297d", "rfloor;": "\u230b", "rfr;": "\U0001d52f", "rhard;": "\u21c1", "rharu;": "\u21c0", "rharul;": "\u296c", "rho;": "\u03c1", "rhov;": "\u03f1", "rightarrow;": "\u2192", "rightarrowtail;": "\u21a3", "rightharpoondown;": "\u21c1", "rightharpoonup;": "\u21c0", "rightleftarrows;": "\u21c4", "rightleftharpoons;": "\u21cc", "rightrightarrows;": "\u21c9", "rightsquigarrow;": "\u219d", "rightthreetimes;": "\u22cc", "ring;": "\u02da", "risingdotseq;": "\u2253", "rlarr;": "\u21c4", "rlhar;": "\u21cc", "rlm;": "\u200f", "rmoust;": "\u23b1", "rmoustache;": "\u23b1", "rnmid;": "\u2aee", "roang;": "\u27ed", "roarr;": "\u21fe", "robrk;": "\u27e7", "ropar;": "\u2986", "ropf;": "\U0001d563", "roplus;": "\u2a2e", "rotimes;": "\u2a35", "rpar;": ")", "rpargt;": "\u2994", "rppolint;": "\u2a12", "rrarr;": "\u21c9", "rsaquo;": "\u203a", "rscr;": "\U0001d4c7", "rsh;": "\u21b1", "rsqb;": "]", "rsquo;": "\u2019", "rsquor;": "\u2019", "rthree;": "\u22cc", "rtimes;": "\u22ca", "rtri;": "\u25b9", "rtrie;": "\u22b5", "rtrif;": "\u25b8", "rtriltri;": "\u29ce", "ruluhar;": "\u2968", "rx;": "\u211e", "sacute;": "\u015b", "sbquo;": "\u201a", "sc;": "\u227b", "scE;": "\u2ab4", "scap;": "\u2ab8", "scaron;": "\u0161", "sccue;": "\u227d", "sce;": "\u2ab0", "scedil;": "\u015f", "scirc;": "\u015d", "scnE;": "\u2ab6", "scnap;": "\u2aba", "scnsim;": "\u22e9", "scpolint;": "\u2a13", "scsim;": "\u227f", "scy;": "\u0441", "sdot;": "\u22c5", "sdotb;": "\u22a1", "sdote;": "\u2a66", "seArr;": "\u21d8", "searhk;": "\u2925", "searr;": "\u2198", "searrow;": "\u2198", "sect": "\xa7", "sect;": "\xa7", "semi;": ";", "seswar;": "\u2929", "setminus;": "\u2216", "setmn;": "\u2216", "sext;": "\u2736", "sfr;": "\U0001d530", "sfrown;": "\u2322", "sharp;": "\u266f", "shchcy;": "\u0449", "shcy;": "\u0448", "shortmid;": "\u2223", "shortparallel;": "\u2225", "shy": "\xad", "shy;": "\xad", "sigma;": "\u03c3", "sigmaf;": "\u03c2", "sigmav;": "\u03c2", "sim;": "\u223c", "simdot;": "\u2a6a", "sime;": "\u2243", "simeq;": "\u2243", "simg;": "\u2a9e", "simgE;": "\u2aa0", "siml;": "\u2a9d", "simlE;": "\u2a9f", "simne;": "\u2246", "simplus;": "\u2a24", "simrarr;": "\u2972", "slarr;": "\u2190", "smallsetminus;": "\u2216", "smashp;": "\u2a33", "smeparsl;": "\u29e4", "smid;": "\u2223", "smile;": "\u2323", "smt;": "\u2aaa", "smte;": "\u2aac", "smtes;": "\u2aac\ufe00", "softcy;": "\u044c", "sol;": "/", "solb;": "\u29c4", "solbar;": "\u233f", "sopf;": "\U0001d564", "spades;": "\u2660", "spadesuit;": "\u2660", "spar;": "\u2225", "sqcap;": "\u2293", "sqcaps;": "\u2293\ufe00", "sqcup;": "\u2294", "sqcups;": "\u2294\ufe00", "sqsub;": "\u228f", "sqsube;": "\u2291", "sqsubset;": "\u228f", "sqsubseteq;": "\u2291", "sqsup;": "\u2290", "sqsupe;": "\u2292", "sqsupset;": "\u2290", "sqsupseteq;": "\u2292", "squ;": "\u25a1", "square;": "\u25a1", "squarf;": "\u25aa", "squf;": "\u25aa", "srarr;": "\u2192", "sscr;": "\U0001d4c8", "ssetmn;": "\u2216", "ssmile;": "\u2323", "sstarf;": "\u22c6", "star;": "\u2606", "starf;": "\u2605", "straightepsilon;": "\u03f5", "straightphi;": "\u03d5", "strns;": "\xaf", "sub;": "\u2282", "subE;": "\u2ac5", "subdot;": "\u2abd", "sube;": "\u2286", "subedot;": "\u2ac3", "submult;": "\u2ac1", "subnE;": "\u2acb", "subne;": "\u228a", "subplus;": "\u2abf", "subrarr;": "\u2979", "subset;": "\u2282", "subseteq;": "\u2286", "subseteqq;": "\u2ac5", "subsetneq;": "\u228a", "subsetneqq;": "\u2acb", "subsim;": "\u2ac7", "subsub;": "\u2ad5", "subsup;": "\u2ad3", "succ;": "\u227b", "succapprox;": "\u2ab8", "succcurlyeq;": "\u227d", "succeq;": "\u2ab0", "succnapprox;": "\u2aba", "succneqq;": "\u2ab6", "succnsim;": "\u22e9", "succsim;": "\u227f", "sum;": "\u2211", "sung;": "\u266a", "sup1": "\xb9", "sup1;": "\xb9", "sup2": "\xb2", "sup2;": "\xb2", "sup3": "\xb3", "sup3;": "\xb3", "sup;": "\u2283", "supE;": "\u2ac6", "supdot;": "\u2abe", "supdsub;": "\u2ad8", "supe;": "\u2287", "supedot;": "\u2ac4", "suphsol;": "\u27c9", "suphsub;": "\u2ad7", "suplarr;": "\u297b", "supmult;": "\u2ac2", "supnE;": "\u2acc", "supne;": "\u228b", "supplus;": "\u2ac0", "supset;": "\u2283", "supseteq;": "\u2287", "supseteqq;": "\u2ac6", "supsetneq;": "\u228b", "supsetneqq;": "\u2acc", "supsim;": "\u2ac8", "supsub;": "\u2ad4", "supsup;": "\u2ad6", "swArr;": "\u21d9", "swarhk;": "\u2926", "swarr;": "\u2199", "swarrow;": "\u2199", "swnwar;": "\u292a", "szlig": "\xdf", "szlig;": "\xdf", "target;": "\u2316", "tau;": "\u03c4", "tbrk;": "\u23b4", "tcaron;": "\u0165", "tcedil;": "\u0163", "tcy;": "\u0442", "tdot;": "\u20db", "telrec;": "\u2315", "tfr;": "\U0001d531", "there4;": "\u2234", "therefore;": "\u2234", "theta;": "\u03b8", "thetasym;": "\u03d1", "thetav;": "\u03d1", "thickapprox;": "\u2248", "thicksim;": "\u223c", "thinsp;": "\u2009", "thkap;": "\u2248", "thksim;": "\u223c", "thorn": "\xfe", "thorn;": "\xfe", "tilde;": "\u02dc", "times": "\xd7", "times;": "\xd7", "timesb;": "\u22a0", "timesbar;": "\u2a31", "timesd;": "\u2a30", "tint;": "\u222d", "toea;": "\u2928", "top;": "\u22a4", "topbot;": "\u2336", "topcir;": "\u2af1", "topf;": "\U0001d565", "topfork;": "\u2ada", "tosa;": "\u2929", "tprime;": "\u2034", "trade;": "\u2122", "triangle;": "\u25b5", "triangledown;": "\u25bf", "triangleleft;": "\u25c3", "trianglelefteq;": "\u22b4", "triangleq;": "\u225c", "triangleright;": "\u25b9", "trianglerighteq;": "\u22b5", "tridot;": "\u25ec", "trie;": "\u225c", "triminus;": "\u2a3a", "triplus;": "\u2a39", "trisb;": "\u29cd", "tritime;": "\u2a3b", "trpezium;": "\u23e2", "tscr;": "\U0001d4c9", "tscy;": "\u0446", "tshcy;": "\u045b", "tstrok;": "\u0167", "twixt;": "\u226c", "twoheadleftarrow;": "\u219e", "twoheadrightarrow;": "\u21a0", "uArr;": "\u21d1", "uHar;": "\u2963", "uacute": "\xfa", "uacute;": "\xfa", "uarr;": "\u2191", "ubrcy;": "\u045e", "ubreve;": "\u016d", "ucirc": "\xfb", "ucirc;": "\xfb", "ucy;": "\u0443", "udarr;": "\u21c5", "udblac;": "\u0171", "udhar;": "\u296e", "ufisht;": "\u297e", "ufr;": "\U0001d532", "ugrave": "\xf9", "ugrave;": "\xf9", "uharl;": "\u21bf", "uharr;": "\u21be", "uhblk;": "\u2580", "ulcorn;": "\u231c", "ulcorner;": "\u231c", "ulcrop;": "\u230f", "ultri;": "\u25f8", "umacr;": "\u016b", "uml": "\xa8", "uml;": "\xa8", "uogon;": "\u0173", "uopf;": "\U0001d566", "uparrow;": "\u2191", "updownarrow;": "\u2195", "upharpoonleft;": "\u21bf", "upharpoonright;": "\u21be", "uplus;": "\u228e", "upsi;": "\u03c5", "upsih;": "\u03d2", "upsilon;": "\u03c5", "upuparrows;": "\u21c8", "urcorn;": "\u231d", "urcorner;": "\u231d", "urcrop;": "\u230e", "uring;": "\u016f", "urtri;": "\u25f9", "uscr;": "\U0001d4ca", "utdot;": "\u22f0", "utilde;": "\u0169", "utri;": "\u25b5", "utrif;": "\u25b4", "uuarr;": "\u21c8", "uuml": "\xfc", "uuml;": "\xfc", "uwangle;": "\u29a7", "vArr;": "\u21d5", "vBar;": "\u2ae8", "vBarv;": "\u2ae9", "vDash;": "\u22a8", "vangrt;": "\u299c", "varepsilon;": "\u03f5", "varkappa;": "\u03f0", "varnothing;": "\u2205", "varphi;": "\u03d5", "varpi;": "\u03d6", "varpropto;": "\u221d", "varr;": "\u2195", "varrho;": "\u03f1", "varsigma;": "\u03c2", "varsubsetneq;": "\u228a\ufe00", "varsubsetneqq;": "\u2acb\ufe00", "varsupsetneq;": "\u228b\ufe00", "varsupsetneqq;": "\u2acc\ufe00", "vartheta;": "\u03d1", "vartriangleleft;": "\u22b2", "vartriangleright;": "\u22b3", "vcy;": "\u0432", "vdash;": "\u22a2", "vee;": "\u2228", "veebar;": "\u22bb", "veeeq;": "\u225a", "vellip;": "\u22ee", "verbar;": "|", "vert;": "|", "vfr;": "\U0001d533", "vltri;": "\u22b2", "vnsub;": "\u2282\u20d2", "vnsup;": "\u2283\u20d2", "vopf;": "\U0001d567", "vprop;": "\u221d", "vrtri;": "\u22b3", "vscr;": "\U0001d4cb", "vsubnE;": "\u2acb\ufe00", "vsubne;": "\u228a\ufe00", "vsupnE;": "\u2acc\ufe00", "vsupne;": "\u228b\ufe00", "vzigzag;": "\u299a", "wcirc;": "\u0175", "wedbar;": "\u2a5f", "wedge;": "\u2227", "wedgeq;": "\u2259", "weierp;": "\u2118", "wfr;": "\U0001d534", "wopf;": "\U0001d568", "wp;": "\u2118", "wr;": "\u2240", "wreath;": "\u2240", "wscr;": "\U0001d4cc", "xcap;": "\u22c2", "xcirc;": "\u25ef", "xcup;": "\u22c3", "xdtri;": "\u25bd", "xfr;": "\U0001d535", "xhArr;": "\u27fa", "xharr;": "\u27f7", "xi;": "\u03be", "xlArr;": "\u27f8", "xlarr;": "\u27f5", "xmap;": "\u27fc", "xnis;": "\u22fb", "xodot;": "\u2a00", "xopf;": "\U0001d569", "xoplus;": "\u2a01", "xotime;": "\u2a02", "xrArr;": "\u27f9", "xrarr;": "\u27f6", "xscr;": "\U0001d4cd", "xsqcup;": "\u2a06", "xuplus;": "\u2a04", "xutri;": "\u25b3", "xvee;": "\u22c1", "xwedge;": "\u22c0", "yacute": "\xfd", "yacute;": "\xfd", "yacy;": "\u044f", "ycirc;": "\u0177", "ycy;": "\u044b", "yen": "\xa5", "yen;": "\xa5", "yfr;": "\U0001d536", "yicy;": "\u0457", "yopf;": "\U0001d56a", "yscr;": "\U0001d4ce", "yucy;": "\u044e", "yuml": "\xff", "yuml;": "\xff", "zacute;": "\u017a", "zcaron;": "\u017e", "zcy;": "\u0437", "zdot;": "\u017c", "zeetrf;": "\u2128", "zeta;": "\u03b6", "zfr;": "\U0001d537", "zhcy;": "\u0436", "zigrarr;": "\u21dd", "zopf;": "\U0001d56b", "zscr;": "\U0001d4cf", "zwj;": "\u200d", "zwnj;": "\u200c", } replacementCharacters = { 0x0: "\uFFFD", 0x0d: "\u000D", 0x80: "\u20AC", 0x81: "\u0081", 0x82: "\u201A", 0x83: "\u0192", 0x84: "\u201E", 0x85: "\u2026", 0x86: "\u2020", 0x87: "\u2021", 0x88: "\u02C6", 0x89: "\u2030", 0x8A: "\u0160", 0x8B: "\u2039", 0x8C: "\u0152", 0x8D: "\u008D", 0x8E: "\u017D", 0x8F: "\u008F", 0x90: "\u0090", 0x91: "\u2018", 0x92: "\u2019", 0x93: "\u201C", 0x94: "\u201D", 0x95: "\u2022", 0x96: "\u2013", 0x97: "\u2014", 0x98: "\u02DC", 0x99: "\u2122", 0x9A: "\u0161", 0x9B: "\u203A", 0x9C: "\u0153", 0x9D: "\u009D", 0x9E: "\u017E", 0x9F: "\u0178", } tokenTypes = { "Doctype": 0, "Characters": 1, "SpaceCharacters": 2, "StartTag": 3, "EndTag": 4, "EmptyTag": 5, "Comment": 6, "ParseError": 7 } tagTokenTypes = frozenset([tokenTypes["StartTag"], tokenTypes["EndTag"], tokenTypes["EmptyTag"]]) prefixes = dict([(v, k) for k, v in namespaces.items()]) prefixes["http://www.w3.org/1998/Math/MathML"] = "math" class DataLossWarning(UserWarning): """Raised when the current tree is unable to represent the input data""" pass class _ReparseException(Exception): pass site-packages/pip/_vendor/six.py000064400000077304151464452720012712 0ustar00# Copyright (c) 2010-2018 Benjamin Peterson # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. """Utilities for writing code that runs on Python 2 and 3""" from __future__ import absolute_import import functools import itertools import operator import sys import types __author__ = "Benjamin Peterson " __version__ = "1.12.0" # Useful for very coarse version differentiation. PY2 = sys.version_info[0] == 2 PY3 = sys.version_info[0] == 3 PY34 = sys.version_info[0:2] >= (3, 4) if PY3: string_types = str, integer_types = int, class_types = type, text_type = str binary_type = bytes MAXSIZE = sys.maxsize else: string_types = basestring, integer_types = (int, long) class_types = (type, types.ClassType) text_type = unicode binary_type = str if sys.platform.startswith("java"): # Jython always uses 32 bits. MAXSIZE = int((1 << 31) - 1) else: # It's possible to have sizeof(long) != sizeof(Py_ssize_t). class X(object): def __len__(self): return 1 << 31 try: len(X()) except OverflowError: # 32-bit MAXSIZE = int((1 << 31) - 1) else: # 64-bit MAXSIZE = int((1 << 63) - 1) del X def _add_doc(func, doc): """Add documentation to a function.""" func.__doc__ = doc def _import_module(name): """Import module, returning the module after the last dot.""" __import__(name) return sys.modules[name] class _LazyDescr(object): def __init__(self, name): self.name = name def __get__(self, obj, tp): result = self._resolve() setattr(obj, self.name, result) # Invokes __set__. try: # This is a bit ugly, but it avoids running this again by # removing this descriptor. delattr(obj.__class__, self.name) except AttributeError: pass return result class MovedModule(_LazyDescr): def __init__(self, name, old, new=None): super(MovedModule, self).__init__(name) if PY3: if new is None: new = name self.mod = new else: self.mod = old def _resolve(self): return _import_module(self.mod) def __getattr__(self, attr): _module = self._resolve() value = getattr(_module, attr) setattr(self, attr, value) return value class _LazyModule(types.ModuleType): def __init__(self, name): super(_LazyModule, self).__init__(name) self.__doc__ = self.__class__.__doc__ def __dir__(self): attrs = ["__doc__", "__name__"] attrs += [attr.name for attr in self._moved_attributes] return attrs # Subclasses should override this _moved_attributes = [] class MovedAttribute(_LazyDescr): def __init__(self, name, old_mod, new_mod, old_attr=None, new_attr=None): super(MovedAttribute, self).__init__(name) if PY3: if new_mod is None: new_mod = name self.mod = new_mod if new_attr is None: if old_attr is None: new_attr = name else: new_attr = old_attr self.attr = new_attr else: self.mod = old_mod if old_attr is None: old_attr = name self.attr = old_attr def _resolve(self): module = _import_module(self.mod) return getattr(module, self.attr) class _SixMetaPathImporter(object): """ A meta path importer to import six.moves and its submodules. This class implements a PEP302 finder and loader. It should be compatible with Python 2.5 and all existing versions of Python3 """ def __init__(self, six_module_name): self.name = six_module_name self.known_modules = {} def _add_module(self, mod, *fullnames): for fullname in fullnames: self.known_modules[self.name + "." + fullname] = mod def _get_module(self, fullname): return self.known_modules[self.name + "." + fullname] def find_module(self, fullname, path=None): if fullname in self.known_modules: return self return None def __get_module(self, fullname): try: return self.known_modules[fullname] except KeyError: raise ImportError("This loader does not know module " + fullname) def load_module(self, fullname): try: # in case of a reload return sys.modules[fullname] except KeyError: pass mod = self.__get_module(fullname) if isinstance(mod, MovedModule): mod = mod._resolve() else: mod.__loader__ = self sys.modules[fullname] = mod return mod def is_package(self, fullname): """ Return true, if the named module is a package. We need this method to get correct spec objects with Python 3.4 (see PEP451) """ return hasattr(self.__get_module(fullname), "__path__") def get_code(self, fullname): """Return None Required, if is_package is implemented""" self.__get_module(fullname) # eventually raises ImportError return None get_source = get_code # same as get_code _importer = _SixMetaPathImporter(__name__) class _MovedItems(_LazyModule): """Lazy loading of moved objects""" __path__ = [] # mark as package _moved_attributes = [ MovedAttribute("cStringIO", "cStringIO", "io", "StringIO"), MovedAttribute("filter", "itertools", "builtins", "ifilter", "filter"), MovedAttribute("filterfalse", "itertools", "itertools", "ifilterfalse", "filterfalse"), MovedAttribute("input", "__builtin__", "builtins", "raw_input", "input"), MovedAttribute("intern", "__builtin__", "sys"), MovedAttribute("map", "itertools", "builtins", "imap", "map"), MovedAttribute("getcwd", "os", "os", "getcwdu", "getcwd"), MovedAttribute("getcwdb", "os", "os", "getcwd", "getcwdb"), MovedAttribute("getoutput", "commands", "subprocess"), MovedAttribute("range", "__builtin__", "builtins", "xrange", "range"), MovedAttribute("reload_module", "__builtin__", "importlib" if PY34 else "imp", "reload"), MovedAttribute("reduce", "__builtin__", "functools"), MovedAttribute("shlex_quote", "pipes", "shlex", "quote"), MovedAttribute("StringIO", "StringIO", "io"), MovedAttribute("UserDict", "UserDict", "collections"), MovedAttribute("UserList", "UserList", "collections"), MovedAttribute("UserString", "UserString", "collections"), MovedAttribute("xrange", "__builtin__", "builtins", "xrange", "range"), MovedAttribute("zip", "itertools", "builtins", "izip", "zip"), MovedAttribute("zip_longest", "itertools", "itertools", "izip_longest", "zip_longest"), MovedModule("builtins", "__builtin__"), MovedModule("configparser", "ConfigParser"), MovedModule("copyreg", "copy_reg"), MovedModule("dbm_gnu", "gdbm", "dbm.gnu"), MovedModule("_dummy_thread", "dummy_thread", "_dummy_thread"), MovedModule("http_cookiejar", "cookielib", "http.cookiejar"), MovedModule("http_cookies", "Cookie", "http.cookies"), MovedModule("html_entities", "htmlentitydefs", "html.entities"), MovedModule("html_parser", "HTMLParser", "html.parser"), MovedModule("http_client", "httplib", "http.client"), MovedModule("email_mime_base", "email.MIMEBase", "email.mime.base"), MovedModule("email_mime_image", "email.MIMEImage", "email.mime.image"), MovedModule("email_mime_multipart", "email.MIMEMultipart", "email.mime.multipart"), MovedModule("email_mime_nonmultipart", "email.MIMENonMultipart", "email.mime.nonmultipart"), MovedModule("email_mime_text", "email.MIMEText", "email.mime.text"), MovedModule("BaseHTTPServer", "BaseHTTPServer", "http.server"), MovedModule("CGIHTTPServer", "CGIHTTPServer", "http.server"), MovedModule("SimpleHTTPServer", "SimpleHTTPServer", "http.server"), MovedModule("cPickle", "cPickle", "pickle"), MovedModule("queue", "Queue"), MovedModule("reprlib", "repr"), MovedModule("socketserver", "SocketServer"), MovedModule("_thread", "thread", "_thread"), MovedModule("tkinter", "Tkinter"), MovedModule("tkinter_dialog", "Dialog", "tkinter.dialog"), MovedModule("tkinter_filedialog", "FileDialog", "tkinter.filedialog"), MovedModule("tkinter_scrolledtext", "ScrolledText", "tkinter.scrolledtext"), MovedModule("tkinter_simpledialog", "SimpleDialog", "tkinter.simpledialog"), MovedModule("tkinter_tix", "Tix", "tkinter.tix"), MovedModule("tkinter_ttk", "ttk", "tkinter.ttk"), MovedModule("tkinter_constants", "Tkconstants", "tkinter.constants"), MovedModule("tkinter_dnd", "Tkdnd", "tkinter.dnd"), MovedModule("tkinter_colorchooser", "tkColorChooser", "tkinter.colorchooser"), MovedModule("tkinter_commondialog", "tkCommonDialog", "tkinter.commondialog"), MovedModule("tkinter_tkfiledialog", "tkFileDialog", "tkinter.filedialog"), MovedModule("tkinter_font", "tkFont", "tkinter.font"), MovedModule("tkinter_messagebox", "tkMessageBox", "tkinter.messagebox"), MovedModule("tkinter_tksimpledialog", "tkSimpleDialog", "tkinter.simpledialog"), MovedModule("urllib_parse", __name__ + ".moves.urllib_parse", "urllib.parse"), MovedModule("urllib_error", __name__ + ".moves.urllib_error", "urllib.error"), MovedModule("urllib", __name__ + ".moves.urllib", __name__ + ".moves.urllib"), MovedModule("urllib_robotparser", "robotparser", "urllib.robotparser"), MovedModule("xmlrpc_client", "xmlrpclib", "xmlrpc.client"), MovedModule("xmlrpc_server", "SimpleXMLRPCServer", "xmlrpc.server"), ] # Add windows specific modules. if sys.platform == "win32": _moved_attributes += [ MovedModule("winreg", "_winreg"), ] for attr in _moved_attributes: setattr(_MovedItems, attr.name, attr) if isinstance(attr, MovedModule): _importer._add_module(attr, "moves." + attr.name) del attr _MovedItems._moved_attributes = _moved_attributes moves = _MovedItems(__name__ + ".moves") _importer._add_module(moves, "moves") class Module_six_moves_urllib_parse(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_parse""" _urllib_parse_moved_attributes = [ MovedAttribute("ParseResult", "urlparse", "urllib.parse"), MovedAttribute("SplitResult", "urlparse", "urllib.parse"), MovedAttribute("parse_qs", "urlparse", "urllib.parse"), MovedAttribute("parse_qsl", "urlparse", "urllib.parse"), MovedAttribute("urldefrag", "urlparse", "urllib.parse"), MovedAttribute("urljoin", "urlparse", "urllib.parse"), MovedAttribute("urlparse", "urlparse", "urllib.parse"), MovedAttribute("urlsplit", "urlparse", "urllib.parse"), MovedAttribute("urlunparse", "urlparse", "urllib.parse"), MovedAttribute("urlunsplit", "urlparse", "urllib.parse"), MovedAttribute("quote", "urllib", "urllib.parse"), MovedAttribute("quote_plus", "urllib", "urllib.parse"), MovedAttribute("unquote", "urllib", "urllib.parse"), MovedAttribute("unquote_plus", "urllib", "urllib.parse"), MovedAttribute("unquote_to_bytes", "urllib", "urllib.parse", "unquote", "unquote_to_bytes"), MovedAttribute("urlencode", "urllib", "urllib.parse"), MovedAttribute("splitquery", "urllib", "urllib.parse"), MovedAttribute("splittag", "urllib", "urllib.parse"), MovedAttribute("splituser", "urllib", "urllib.parse"), MovedAttribute("splitvalue", "urllib", "urllib.parse"), MovedAttribute("uses_fragment", "urlparse", "urllib.parse"), MovedAttribute("uses_netloc", "urlparse", "urllib.parse"), MovedAttribute("uses_params", "urlparse", "urllib.parse"), MovedAttribute("uses_query", "urlparse", "urllib.parse"), MovedAttribute("uses_relative", "urlparse", "urllib.parse"), ] for attr in _urllib_parse_moved_attributes: setattr(Module_six_moves_urllib_parse, attr.name, attr) del attr Module_six_moves_urllib_parse._moved_attributes = _urllib_parse_moved_attributes _importer._add_module(Module_six_moves_urllib_parse(__name__ + ".moves.urllib_parse"), "moves.urllib_parse", "moves.urllib.parse") class Module_six_moves_urllib_error(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_error""" _urllib_error_moved_attributes = [ MovedAttribute("URLError", "urllib2", "urllib.error"), MovedAttribute("HTTPError", "urllib2", "urllib.error"), MovedAttribute("ContentTooShortError", "urllib", "urllib.error"), ] for attr in _urllib_error_moved_attributes: setattr(Module_six_moves_urllib_error, attr.name, attr) del attr Module_six_moves_urllib_error._moved_attributes = _urllib_error_moved_attributes _importer._add_module(Module_six_moves_urllib_error(__name__ + ".moves.urllib.error"), "moves.urllib_error", "moves.urllib.error") class Module_six_moves_urllib_request(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_request""" _urllib_request_moved_attributes = [ MovedAttribute("urlopen", "urllib2", "urllib.request"), MovedAttribute("install_opener", "urllib2", "urllib.request"), MovedAttribute("build_opener", "urllib2", "urllib.request"), MovedAttribute("pathname2url", "urllib", "urllib.request"), MovedAttribute("url2pathname", "urllib", "urllib.request"), MovedAttribute("getproxies", "urllib", "urllib.request"), MovedAttribute("Request", "urllib2", "urllib.request"), MovedAttribute("OpenerDirector", "urllib2", "urllib.request"), MovedAttribute("HTTPDefaultErrorHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPRedirectHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPCookieProcessor", "urllib2", "urllib.request"), MovedAttribute("ProxyHandler", "urllib2", "urllib.request"), MovedAttribute("BaseHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPPasswordMgr", "urllib2", "urllib.request"), MovedAttribute("HTTPPasswordMgrWithDefaultRealm", "urllib2", "urllib.request"), MovedAttribute("AbstractBasicAuthHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPBasicAuthHandler", "urllib2", "urllib.request"), MovedAttribute("ProxyBasicAuthHandler", "urllib2", "urllib.request"), MovedAttribute("AbstractDigestAuthHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPDigestAuthHandler", "urllib2", "urllib.request"), MovedAttribute("ProxyDigestAuthHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPSHandler", "urllib2", "urllib.request"), MovedAttribute("FileHandler", "urllib2", "urllib.request"), MovedAttribute("FTPHandler", "urllib2", "urllib.request"), MovedAttribute("CacheFTPHandler", "urllib2", "urllib.request"), MovedAttribute("UnknownHandler", "urllib2", "urllib.request"), MovedAttribute("HTTPErrorProcessor", "urllib2", "urllib.request"), MovedAttribute("urlretrieve", "urllib", "urllib.request"), MovedAttribute("urlcleanup", "urllib", "urllib.request"), MovedAttribute("URLopener", "urllib", "urllib.request"), MovedAttribute("FancyURLopener", "urllib", "urllib.request"), MovedAttribute("proxy_bypass", "urllib", "urllib.request"), MovedAttribute("parse_http_list", "urllib2", "urllib.request"), MovedAttribute("parse_keqv_list", "urllib2", "urllib.request"), ] for attr in _urllib_request_moved_attributes: setattr(Module_six_moves_urllib_request, attr.name, attr) del attr Module_six_moves_urllib_request._moved_attributes = _urllib_request_moved_attributes _importer._add_module(Module_six_moves_urllib_request(__name__ + ".moves.urllib.request"), "moves.urllib_request", "moves.urllib.request") class Module_six_moves_urllib_response(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_response""" _urllib_response_moved_attributes = [ MovedAttribute("addbase", "urllib", "urllib.response"), MovedAttribute("addclosehook", "urllib", "urllib.response"), MovedAttribute("addinfo", "urllib", "urllib.response"), MovedAttribute("addinfourl", "urllib", "urllib.response"), ] for attr in _urllib_response_moved_attributes: setattr(Module_six_moves_urllib_response, attr.name, attr) del attr Module_six_moves_urllib_response._moved_attributes = _urllib_response_moved_attributes _importer._add_module(Module_six_moves_urllib_response(__name__ + ".moves.urllib.response"), "moves.urllib_response", "moves.urllib.response") class Module_six_moves_urllib_robotparser(_LazyModule): """Lazy loading of moved objects in six.moves.urllib_robotparser""" _urllib_robotparser_moved_attributes = [ MovedAttribute("RobotFileParser", "robotparser", "urllib.robotparser"), ] for attr in _urllib_robotparser_moved_attributes: setattr(Module_six_moves_urllib_robotparser, attr.name, attr) del attr Module_six_moves_urllib_robotparser._moved_attributes = _urllib_robotparser_moved_attributes _importer._add_module(Module_six_moves_urllib_robotparser(__name__ + ".moves.urllib.robotparser"), "moves.urllib_robotparser", "moves.urllib.robotparser") class Module_six_moves_urllib(types.ModuleType): """Create a six.moves.urllib namespace that resembles the Python 3 namespace""" __path__ = [] # mark as package parse = _importer._get_module("moves.urllib_parse") error = _importer._get_module("moves.urllib_error") request = _importer._get_module("moves.urllib_request") response = _importer._get_module("moves.urllib_response") robotparser = _importer._get_module("moves.urllib_robotparser") def __dir__(self): return ['parse', 'error', 'request', 'response', 'robotparser'] _importer._add_module(Module_six_moves_urllib(__name__ + ".moves.urllib"), "moves.urllib") def add_move(move): """Add an item to six.moves.""" setattr(_MovedItems, move.name, move) def remove_move(name): """Remove item from six.moves.""" try: delattr(_MovedItems, name) except AttributeError: try: del moves.__dict__[name] except KeyError: raise AttributeError("no such move, %r" % (name,)) if PY3: _meth_func = "__func__" _meth_self = "__self__" _func_closure = "__closure__" _func_code = "__code__" _func_defaults = "__defaults__" _func_globals = "__globals__" else: _meth_func = "im_func" _meth_self = "im_self" _func_closure = "func_closure" _func_code = "func_code" _func_defaults = "func_defaults" _func_globals = "func_globals" try: advance_iterator = next except NameError: def advance_iterator(it): return it.next() next = advance_iterator try: callable = callable except NameError: def callable(obj): return any("__call__" in klass.__dict__ for klass in type(obj).__mro__) if PY3: def get_unbound_function(unbound): return unbound create_bound_method = types.MethodType def create_unbound_method(func, cls): return func Iterator = object else: def get_unbound_function(unbound): return unbound.im_func def create_bound_method(func, obj): return types.MethodType(func, obj, obj.__class__) def create_unbound_method(func, cls): return types.MethodType(func, None, cls) class Iterator(object): def next(self): return type(self).__next__(self) callable = callable _add_doc(get_unbound_function, """Get the function out of a possibly unbound function""") get_method_function = operator.attrgetter(_meth_func) get_method_self = operator.attrgetter(_meth_self) get_function_closure = operator.attrgetter(_func_closure) get_function_code = operator.attrgetter(_func_code) get_function_defaults = operator.attrgetter(_func_defaults) get_function_globals = operator.attrgetter(_func_globals) if PY3: def iterkeys(d, **kw): return iter(d.keys(**kw)) def itervalues(d, **kw): return iter(d.values(**kw)) def iteritems(d, **kw): return iter(d.items(**kw)) def iterlists(d, **kw): return iter(d.lists(**kw)) viewkeys = operator.methodcaller("keys") viewvalues = operator.methodcaller("values") viewitems = operator.methodcaller("items") else: def iterkeys(d, **kw): return d.iterkeys(**kw) def itervalues(d, **kw): return d.itervalues(**kw) def iteritems(d, **kw): return d.iteritems(**kw) def iterlists(d, **kw): return d.iterlists(**kw) viewkeys = operator.methodcaller("viewkeys") viewvalues = operator.methodcaller("viewvalues") viewitems = operator.methodcaller("viewitems") _add_doc(iterkeys, "Return an iterator over the keys of a dictionary.") _add_doc(itervalues, "Return an iterator over the values of a dictionary.") _add_doc(iteritems, "Return an iterator over the (key, value) pairs of a dictionary.") _add_doc(iterlists, "Return an iterator over the (key, [values]) pairs of a dictionary.") if PY3: def b(s): return s.encode("latin-1") def u(s): return s unichr = chr import struct int2byte = struct.Struct(">B").pack del struct byte2int = operator.itemgetter(0) indexbytes = operator.getitem iterbytes = iter import io StringIO = io.StringIO BytesIO = io.BytesIO _assertCountEqual = "assertCountEqual" if sys.version_info[1] <= 1: _assertRaisesRegex = "assertRaisesRegexp" _assertRegex = "assertRegexpMatches" else: _assertRaisesRegex = "assertRaisesRegex" _assertRegex = "assertRegex" else: def b(s): return s # Workaround for standalone backslash def u(s): return unicode(s.replace(r'\\', r'\\\\'), "unicode_escape") unichr = unichr int2byte = chr def byte2int(bs): return ord(bs[0]) def indexbytes(buf, i): return ord(buf[i]) iterbytes = functools.partial(itertools.imap, ord) import StringIO StringIO = BytesIO = StringIO.StringIO _assertCountEqual = "assertItemsEqual" _assertRaisesRegex = "assertRaisesRegexp" _assertRegex = "assertRegexpMatches" _add_doc(b, """Byte literal""") _add_doc(u, """Text literal""") def assertCountEqual(self, *args, **kwargs): return getattr(self, _assertCountEqual)(*args, **kwargs) def assertRaisesRegex(self, *args, **kwargs): return getattr(self, _assertRaisesRegex)(*args, **kwargs) def assertRegex(self, *args, **kwargs): return getattr(self, _assertRegex)(*args, **kwargs) if PY3: exec_ = getattr(moves.builtins, "exec") def reraise(tp, value, tb=None): try: if value is None: value = tp() if value.__traceback__ is not tb: raise value.with_traceback(tb) raise value finally: value = None tb = None else: def exec_(_code_, _globs_=None, _locs_=None): """Execute code in a namespace.""" if _globs_ is None: frame = sys._getframe(1) _globs_ = frame.f_globals if _locs_ is None: _locs_ = frame.f_locals del frame elif _locs_ is None: _locs_ = _globs_ exec("""exec _code_ in _globs_, _locs_""") exec_("""def reraise(tp, value, tb=None): try: raise tp, value, tb finally: tb = None """) if sys.version_info[:2] == (3, 2): exec_("""def raise_from(value, from_value): try: if from_value is None: raise value raise value from from_value finally: value = None """) elif sys.version_info[:2] > (3, 2): exec_("""def raise_from(value, from_value): try: raise value from from_value finally: value = None """) else: def raise_from(value, from_value): raise value print_ = getattr(moves.builtins, "print", None) if print_ is None: def print_(*args, **kwargs): """The new-style print function for Python 2.4 and 2.5.""" fp = kwargs.pop("file", sys.stdout) if fp is None: return def write(data): if not isinstance(data, basestring): data = str(data) # If the file has an encoding, encode unicode with it. if (isinstance(fp, file) and isinstance(data, unicode) and fp.encoding is not None): errors = getattr(fp, "errors", None) if errors is None: errors = "strict" data = data.encode(fp.encoding, errors) fp.write(data) want_unicode = False sep = kwargs.pop("sep", None) if sep is not None: if isinstance(sep, unicode): want_unicode = True elif not isinstance(sep, str): raise TypeError("sep must be None or a string") end = kwargs.pop("end", None) if end is not None: if isinstance(end, unicode): want_unicode = True elif not isinstance(end, str): raise TypeError("end must be None or a string") if kwargs: raise TypeError("invalid keyword arguments to print()") if not want_unicode: for arg in args: if isinstance(arg, unicode): want_unicode = True break if want_unicode: newline = unicode("\n") space = unicode(" ") else: newline = "\n" space = " " if sep is None: sep = space if end is None: end = newline for i, arg in enumerate(args): if i: write(sep) write(arg) write(end) if sys.version_info[:2] < (3, 3): _print = print_ def print_(*args, **kwargs): fp = kwargs.get("file", sys.stdout) flush = kwargs.pop("flush", False) _print(*args, **kwargs) if flush and fp is not None: fp.flush() _add_doc(reraise, """Reraise an exception.""") if sys.version_info[0:2] < (3, 4): def wraps(wrapped, assigned=functools.WRAPPER_ASSIGNMENTS, updated=functools.WRAPPER_UPDATES): def wrapper(f): f = functools.wraps(wrapped, assigned, updated)(f) f.__wrapped__ = wrapped return f return wrapper else: wraps = functools.wraps def with_metaclass(meta, *bases): """Create a base class with a metaclass.""" # This requires a bit of explanation: the basic idea is to make a dummy # metaclass for one level of class instantiation that replaces itself with # the actual metaclass. class metaclass(type): def __new__(cls, name, this_bases, d): return meta(name, bases, d) @classmethod def __prepare__(cls, name, this_bases): return meta.__prepare__(name, bases) return type.__new__(metaclass, 'temporary_class', (), {}) def add_metaclass(metaclass): """Class decorator for creating a class with a metaclass.""" def wrapper(cls): orig_vars = cls.__dict__.copy() slots = orig_vars.get('__slots__') if slots is not None: if isinstance(slots, str): slots = [slots] for slots_var in slots: orig_vars.pop(slots_var) orig_vars.pop('__dict__', None) orig_vars.pop('__weakref__', None) if hasattr(cls, '__qualname__'): orig_vars['__qualname__'] = cls.__qualname__ return metaclass(cls.__name__, cls.__bases__, orig_vars) return wrapper def ensure_binary(s, encoding='utf-8', errors='strict'): """Coerce **s** to six.binary_type. For Python 2: - `unicode` -> encoded to `str` - `str` -> `str` For Python 3: - `str` -> encoded to `bytes` - `bytes` -> `bytes` """ if isinstance(s, text_type): return s.encode(encoding, errors) elif isinstance(s, binary_type): return s else: raise TypeError("not expecting type '%s'" % type(s)) def ensure_str(s, encoding='utf-8', errors='strict'): """Coerce *s* to `str`. For Python 2: - `unicode` -> encoded to `str` - `str` -> `str` For Python 3: - `str` -> `str` - `bytes` -> decoded to `str` """ if not isinstance(s, (text_type, binary_type)): raise TypeError("not expecting type '%s'" % type(s)) if PY2 and isinstance(s, text_type): s = s.encode(encoding, errors) elif PY3 and isinstance(s, binary_type): s = s.decode(encoding, errors) return s def ensure_text(s, encoding='utf-8', errors='strict'): """Coerce *s* to six.text_type. For Python 2: - `unicode` -> `unicode` - `str` -> `unicode` For Python 3: - `str` -> `str` - `bytes` -> decoded to `str` """ if isinstance(s, binary_type): return s.decode(encoding, errors) elif isinstance(s, text_type): return s else: raise TypeError("not expecting type '%s'" % type(s)) def python_2_unicode_compatible(klass): """ A decorator that defines __unicode__ and __str__ methods under Python 2. Under Python 3 it does nothing. To support Python 2 and 3 with a single code base, define a __str__ method returning text and apply this decorator to the class. """ if PY2: if '__str__' not in klass.__dict__: raise ValueError("@python_2_unicode_compatible cannot be applied " "to %s because it doesn't define __str__()." % klass.__name__) klass.__unicode__ = klass.__str__ klass.__str__ = lambda self: self.__unicode__().encode('utf-8') return klass # Complete the moves implementation. # This code is at the end of this module to speed up module loading. # Turn this module into a package. __path__ = [] # required for PEP 302 and PEP 451 __package__ = __name__ # see PEP 366 @ReservedAssignment if globals().get("__spec__") is not None: __spec__.submodule_search_locations = [] # PEP 451 @UndefinedVariable # Remove other six meta path importers, since they cause problems. This can # happen if six is removed from sys.modules and then reloaded. (Setuptools does # this for some reason.) if sys.meta_path: for i, importer in enumerate(sys.meta_path): # Here's some real nastiness: Another "instance" of the six module might # be floating around. Therefore, we can't use isinstance() to check for # the six meta path importer, since the other six instance will have # inserted an importer with different class. if (type(importer).__name__ == "_SixMetaPathImporter" and importer.name == __name__): del sys.meta_path[i] break del i, importer # Finally, add the importer to the meta path import hook. sys.meta_path.append(_importer) site-packages/pip/_vendor/pyparsing.py000064400001002454151464452720014116 0ustar00# -*- coding: utf-8 -*- # module pyparsing.py # # Copyright (c) 2003-2019 Paul T. McGuire # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be # included in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, # EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF # MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. # IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY # CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, # TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE # SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # __doc__ = \ """ pyparsing module - Classes and methods to define and execute parsing grammars ============================================================================= The pyparsing module is an alternative approach to creating and executing simple grammars, vs. the traditional lex/yacc approach, or the use of regular expressions. With pyparsing, you don't need to learn a new syntax for defining grammars or matching expressions - the parsing module provides a library of classes that you use to construct the grammar directly in Python. Here is a program to parse "Hello, World!" (or any greeting of the form ``", !"``), built up using :class:`Word`, :class:`Literal`, and :class:`And` elements (the :class:`'+'` operators create :class:`And` expressions, and the strings are auto-converted to :class:`Literal` expressions):: from pip._vendor.pyparsing import Word, alphas # define grammar of a greeting greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) The program outputs the following:: Hello, World! -> ['Hello', ',', 'World', '!'] The Python representation of the grammar is quite readable, owing to the self-explanatory class names, and the use of '+', '|' and '^' operators. The :class:`ParseResults` object returned from :class:`ParserElement.parseString` can be accessed as a nested list, a dictionary, or an object with named attributes. The pyparsing module handles some of the problems that are typically vexing when writing text parsers: - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello , World !", etc.) - quoted strings - embedded comments Getting Started - ----------------- Visit the classes :class:`ParserElement` and :class:`ParseResults` to see the base classes that most other pyparsing classes inherit from. Use the docstrings for examples of how to: - construct literal match expressions from :class:`Literal` and :class:`CaselessLiteral` classes - construct character word-group expressions using the :class:`Word` class - see how to create repetitive expressions using :class:`ZeroOrMore` and :class:`OneOrMore` classes - use :class:`'+'`, :class:`'|'`, :class:`'^'`, and :class:`'&'` operators to combine simple expressions into more complex ones - associate names with your parsed results using :class:`ParserElement.setResultsName` - access the parsed data, which is returned as a :class:`ParseResults` object - find some helpful expression short-cuts like :class:`delimitedList` and :class:`oneOf` - find more useful common expressions in the :class:`pyparsing_common` namespace class """ __version__ = "2.4.2" __versionTime__ = "29 Jul 2019 02:58 UTC" __author__ = "Paul McGuire " import string from weakref import ref as wkref import copy import sys import warnings import re import sre_constants import collections import pprint import traceback import types from datetime import datetime from operator import itemgetter import itertools from functools import wraps try: # Python 3 from itertools import filterfalse except ImportError: from itertools import ifilterfalse as filterfalse try: from _thread import RLock except ImportError: from threading import RLock try: # Python 3 from collections.abc import Iterable from collections.abc import MutableMapping, Mapping except ImportError: # Python 2.7 from collections import Iterable from collections import MutableMapping, Mapping try: from collections import OrderedDict as _OrderedDict except ImportError: try: from ordereddict import OrderedDict as _OrderedDict except ImportError: _OrderedDict = None try: from types import SimpleNamespace except ImportError: class SimpleNamespace: pass # version compatibility configuration __compat__ = SimpleNamespace() __compat__.__doc__ = """ A cross-version compatibility configuration for pyparsing features that will be released in a future version. By setting values in this configuration to True, those features can be enabled in prior versions for compatibility development and testing. - collect_all_And_tokens - flag to enable fix for Issue #63 that fixes erroneous grouping of results names when an And expression is nested within an Or or MatchFirst; set to True to enable bugfix released in pyparsing 2.3.0, or False to preserve pre-2.3.0 handling of named results """ __compat__.collect_all_And_tokens = True __diag__ = SimpleNamespace() __diag__.__doc__ = """ Diagnostic configuration (all default to False) - warn_multiple_tokens_in_named_alternation - flag to enable warnings when a results name is defined on a MatchFirst or Or expression with one or more And subexpressions (only warns if __compat__.collect_all_And_tokens is False) - warn_ungrouped_named_tokens_in_collection - flag to enable warnings when a results name is defined on a containing expression with ungrouped subexpressions that also have results names - warn_name_set_on_empty_Forward - flag to enable warnings whan a Forward is defined with a results name, but has no contents defined - warn_on_multiple_string_args_to_oneof - flag to enable warnings whan oneOf is incorrectly called with multiple str arguments - enable_debug_on_named_expressions - flag to auto-enable debug on all subsequent calls to ParserElement.setName() """ __diag__.warn_multiple_tokens_in_named_alternation = False __diag__.warn_ungrouped_named_tokens_in_collection = False __diag__.warn_name_set_on_empty_Forward = False __diag__.warn_on_multiple_string_args_to_oneof = False __diag__.enable_debug_on_named_expressions = False # ~ sys.stderr.write("testing pyparsing module, version %s, %s\n" % (__version__, __versionTime__)) __all__ = ['__version__', '__versionTime__', '__author__', '__compat__', '__diag__', 'And', 'CaselessKeyword', 'CaselessLiteral', 'CharsNotIn', 'Combine', 'Dict', 'Each', 'Empty', 'FollowedBy', 'Forward', 'GoToColumn', 'Group', 'Keyword', 'LineEnd', 'LineStart', 'Literal', 'PrecededBy', 'MatchFirst', 'NoMatch', 'NotAny', 'OneOrMore', 'OnlyOnce', 'Optional', 'Or', 'ParseBaseException', 'ParseElementEnhance', 'ParseException', 'ParseExpression', 'ParseFatalException', 'ParseResults', 'ParseSyntaxException', 'ParserElement', 'QuotedString', 'RecursiveGrammarException', 'Regex', 'SkipTo', 'StringEnd', 'StringStart', 'Suppress', 'Token', 'TokenConverter', 'White', 'Word', 'WordEnd', 'WordStart', 'ZeroOrMore', 'Char', 'alphanums', 'alphas', 'alphas8bit', 'anyCloseTag', 'anyOpenTag', 'cStyleComment', 'col', 'commaSeparatedList', 'commonHTMLEntity', 'countedArray', 'cppStyleComment', 'dblQuotedString', 'dblSlashComment', 'delimitedList', 'dictOf', 'downcaseTokens', 'empty', 'hexnums', 'htmlComment', 'javaStyleComment', 'line', 'lineEnd', 'lineStart', 'lineno', 'makeHTMLTags', 'makeXMLTags', 'matchOnlyAtCol', 'matchPreviousExpr', 'matchPreviousLiteral', 'nestedExpr', 'nullDebugAction', 'nums', 'oneOf', 'opAssoc', 'operatorPrecedence', 'printables', 'punc8bit', 'pythonStyleComment', 'quotedString', 'removeQuotes', 'replaceHTMLEntity', 'replaceWith', 'restOfLine', 'sglQuotedString', 'srange', 'stringEnd', 'stringStart', 'traceParseAction', 'unicodeString', 'upcaseTokens', 'withAttribute', 'indentedBlock', 'originalTextFor', 'ungroup', 'infixNotation', 'locatedExpr', 'withClass', 'CloseMatch', 'tokenMap', 'pyparsing_common', 'pyparsing_unicode', 'unicode_set', 'conditionAsParseAction', ] system_version = tuple(sys.version_info)[:3] PY_3 = system_version[0] == 3 if PY_3: _MAX_INT = sys.maxsize basestring = str unichr = chr unicode = str _ustr = str # build list of single arg builtins, that can be used as parse actions singleArgBuiltins = [sum, len, sorted, reversed, list, tuple, set, any, all, min, max] else: _MAX_INT = sys.maxint range = xrange def _ustr(obj): """Drop-in replacement for str(obj) that tries to be Unicode friendly. It first tries str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It then < returns the unicode object | encodes it with the default encoding | ... >. """ if isinstance(obj, unicode): return obj try: # If this works, then _ustr(obj) has the same behaviour as str(obj), so # it won't break any existing code. return str(obj) except UnicodeEncodeError: # Else encode it ret = unicode(obj).encode(sys.getdefaultencoding(), 'xmlcharrefreplace') xmlcharref = Regex(r'&#\d+;') xmlcharref.setParseAction(lambda t: '\\u' + hex(int(t[0][2:-1]))[2:]) return xmlcharref.transformString(ret) # build list of single arg builtins, tolerant of Python version, that can be used as parse actions singleArgBuiltins = [] import __builtin__ for fname in "sum len sorted reversed list tuple set any all min max".split(): try: singleArgBuiltins.append(getattr(__builtin__, fname)) except AttributeError: continue _generatorType = type((y for y in range(1))) def _xml_escape(data): """Escape &, <, >, ", ', etc. in a string of data.""" # ampersand must be replaced first from_symbols = '&><"\'' to_symbols = ('&' + s + ';' for s in "amp gt lt quot apos".split()) for from_, to_ in zip(from_symbols, to_symbols): data = data.replace(from_, to_) return data alphas = string.ascii_uppercase + string.ascii_lowercase nums = "0123456789" hexnums = nums + "ABCDEFabcdef" alphanums = alphas + nums _bslash = chr(92) printables = "".join(c for c in string.printable if c not in string.whitespace) def conditionAsParseAction(fn, message=None, fatal=False): msg = message if message is not None else "failed user-defined condition" exc_type = ParseFatalException if fatal else ParseException fn = _trim_arity(fn) @wraps(fn) def pa(s, l, t): if not bool(fn(s, l, t)): raise exc_type(s, l, msg) return pa class ParseBaseException(Exception): """base exception class for all parsing runtime exceptions""" # Performance tuning: we construct a *lot* of these, so keep this # constructor as small and fast as possible def __init__(self, pstr, loc=0, msg=None, elem=None): self.loc = loc if msg is None: self.msg = pstr self.pstr = "" else: self.msg = msg self.pstr = pstr self.parserElement = elem self.args = (pstr, loc, msg) @classmethod def _from_exception(cls, pe): """ internal factory method to simplify creating one type of ParseException from another - avoids having __init__ signature conflicts among subclasses """ return cls(pe.pstr, pe.loc, pe.msg, pe.parserElement) def __getattr__(self, aname): """supported attributes by name are: - lineno - returns the line number of the exception text - col - returns the column number of the exception text - line - returns the line containing the exception text """ if aname == "lineno": return lineno(self.loc, self.pstr) elif aname in ("col", "column"): return col(self.loc, self.pstr) elif aname == "line": return line(self.loc, self.pstr) else: raise AttributeError(aname) def __str__(self): if self.pstr: if self.loc >= len(self.pstr): foundstr = ', found end of text' else: foundstr = (', found %r' % self.pstr[self.loc:self.loc + 1]).replace(r'\\', '\\') else: foundstr = '' return ("%s%s (at char %d), (line:%d, col:%d)" % (self.msg, foundstr, self.loc, self.lineno, self.column)) def __repr__(self): return _ustr(self) def markInputline(self, markerString=">!<"): """Extracts the exception line from the input string, and marks the location of the exception with a special symbol. """ line_str = self.line line_column = self.column - 1 if markerString: line_str = "".join((line_str[:line_column], markerString, line_str[line_column:])) return line_str.strip() def __dir__(self): return "lineno col line".split() + dir(type(self)) class ParseException(ParseBaseException): """ Exception thrown when parse expressions don't match class; supported attributes by name are: - lineno - returns the line number of the exception text - col - returns the column number of the exception text - line - returns the line containing the exception text Example:: try: Word(nums).setName("integer").parseString("ABC") except ParseException as pe: print(pe) print("column: {}".format(pe.col)) prints:: Expected integer (at char 0), (line:1, col:1) column: 1 """ @staticmethod def explain(exc, depth=16): """ Method to take an exception and translate the Python internal traceback into a list of the pyparsing expressions that caused the exception to be raised. Parameters: - exc - exception raised during parsing (need not be a ParseException, in support of Python exceptions that might be raised in a parse action) - depth (default=16) - number of levels back in the stack trace to list expression and function names; if None, the full stack trace names will be listed; if 0, only the failing input line, marker, and exception string will be shown Returns a multi-line string listing the ParserElements and/or function names in the exception's stack trace. Note: the diagnostic output will include string representations of the expressions that failed to parse. These representations will be more helpful if you use `setName` to give identifiable names to your expressions. Otherwise they will use the default string forms, which may be cryptic to read. explain() is only supported under Python 3. """ import inspect if depth is None: depth = sys.getrecursionlimit() ret = [] if isinstance(exc, ParseBaseException): ret.append(exc.line) ret.append(' ' * (exc.col - 1) + '^') ret.append("{0}: {1}".format(type(exc).__name__, exc)) if depth > 0: callers = inspect.getinnerframes(exc.__traceback__, context=depth) seen = set() for i, ff in enumerate(callers[-depth:]): frm = ff[0] f_self = frm.f_locals.get('self', None) if isinstance(f_self, ParserElement): if frm.f_code.co_name not in ('parseImpl', '_parseNoCache'): continue if f_self in seen: continue seen.add(f_self) self_type = type(f_self) ret.append("{0}.{1} - {2}".format(self_type.__module__, self_type.__name__, f_self)) elif f_self is not None: self_type = type(f_self) ret.append("{0}.{1}".format(self_type.__module__, self_type.__name__)) else: code = frm.f_code if code.co_name in ('wrapper', ''): continue ret.append("{0}".format(code.co_name)) depth -= 1 if not depth: break return '\n'.join(ret) class ParseFatalException(ParseBaseException): """user-throwable exception thrown when inconsistent parse content is found; stops all parsing immediately""" pass class ParseSyntaxException(ParseFatalException): """just like :class:`ParseFatalException`, but thrown internally when an :class:`ErrorStop` ('-' operator) indicates that parsing is to stop immediately because an unbacktrackable syntax error has been found. """ pass #~ class ReparseException(ParseBaseException): #~ """Experimental class - parse actions can raise this exception to cause #~ pyparsing to reparse the input string: #~ - with a modified input string, and/or #~ - with a modified start location #~ Set the values of the ReparseException in the constructor, and raise the #~ exception in a parse action to cause pyparsing to use the new string/location. #~ Setting the values as None causes no change to be made. #~ """ #~ def __init_( self, newstring, restartLoc ): #~ self.newParseText = newstring #~ self.reparseLoc = restartLoc class RecursiveGrammarException(Exception): """exception thrown by :class:`ParserElement.validate` if the grammar could be improperly recursive """ def __init__(self, parseElementList): self.parseElementTrace = parseElementList def __str__(self): return "RecursiveGrammarException: %s" % self.parseElementTrace class _ParseResultsWithOffset(object): def __init__(self, p1, p2): self.tup = (p1, p2) def __getitem__(self, i): return self.tup[i] def __repr__(self): return repr(self.tup[0]) def setOffset(self, i): self.tup = (self.tup[0], i) class ParseResults(object): """Structured parse results, to provide multiple means of access to the parsed data: - as a list (``len(results)``) - by list index (``results[0], results[1]``, etc.) - by attribute (``results.`` - see :class:`ParserElement.setResultsName`) Example:: integer = Word(nums) date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: # date_str = integer("year") + '/' + integer("month") + '/' + integer("day") # parseString returns a ParseResults object result = date_str.parseString("1999/12/31") def test(s, fn=repr): print("%s -> %s" % (s, fn(eval(s)))) test("list(result)") test("result[0]") test("result['month']") test("result.day") test("'month' in result") test("'minutes' in result") test("result.dump()", str) prints:: list(result) -> ['1999', '/', '12', '/', '31'] result[0] -> '1999' result['month'] -> '12' result.day -> '31' 'month' in result -> True 'minutes' in result -> False result.dump() -> ['1999', '/', '12', '/', '31'] - day: 31 - month: 12 - year: 1999 """ def __new__(cls, toklist=None, name=None, asList=True, modal=True): if isinstance(toklist, cls): return toklist retobj = object.__new__(cls) retobj.__doinit = True return retobj # Performance tuning: we construct a *lot* of these, so keep this # constructor as small and fast as possible def __init__(self, toklist=None, name=None, asList=True, modal=True, isinstance=isinstance): if self.__doinit: self.__doinit = False self.__name = None self.__parent = None self.__accumNames = {} self.__asList = asList self.__modal = modal if toklist is None: toklist = [] if isinstance(toklist, list): self.__toklist = toklist[:] elif isinstance(toklist, _generatorType): self.__toklist = list(toklist) else: self.__toklist = [toklist] self.__tokdict = dict() if name is not None and name: if not modal: self.__accumNames[name] = 0 if isinstance(name, int): name = _ustr(name) # will always return a str, but use _ustr for consistency self.__name = name if not (isinstance(toklist, (type(None), basestring, list)) and toklist in (None, '', [])): if isinstance(toklist, basestring): toklist = [toklist] if asList: if isinstance(toklist, ParseResults): self[name] = _ParseResultsWithOffset(ParseResults(toklist.__toklist), 0) else: self[name] = _ParseResultsWithOffset(ParseResults(toklist[0]), 0) self[name].__name = name else: try: self[name] = toklist[0] except (KeyError, TypeError, IndexError): self[name] = toklist def __getitem__(self, i): if isinstance(i, (int, slice)): return self.__toklist[i] else: if i not in self.__accumNames: return self.__tokdict[i][-1][0] else: return ParseResults([v[0] for v in self.__tokdict[i]]) def __setitem__(self, k, v, isinstance=isinstance): if isinstance(v, _ParseResultsWithOffset): self.__tokdict[k] = self.__tokdict.get(k, list()) + [v] sub = v[0] elif isinstance(k, (int, slice)): self.__toklist[k] = v sub = v else: self.__tokdict[k] = self.__tokdict.get(k, list()) + [_ParseResultsWithOffset(v, 0)] sub = v if isinstance(sub, ParseResults): sub.__parent = wkref(self) def __delitem__(self, i): if isinstance(i, (int, slice)): mylen = len(self.__toklist) del self.__toklist[i] # convert int to slice if isinstance(i, int): if i < 0: i += mylen i = slice(i, i + 1) # get removed indices removed = list(range(*i.indices(mylen))) removed.reverse() # fixup indices in token dictionary for name, occurrences in self.__tokdict.items(): for j in removed: for k, (value, position) in enumerate(occurrences): occurrences[k] = _ParseResultsWithOffset(value, position - (position > j)) else: del self.__tokdict[i] def __contains__(self, k): return k in self.__tokdict def __len__(self): return len(self.__toklist) def __bool__(self): return (not not self.__toklist) __nonzero__ = __bool__ def __iter__(self): return iter(self.__toklist) def __reversed__(self): return iter(self.__toklist[::-1]) def _iterkeys(self): if hasattr(self.__tokdict, "iterkeys"): return self.__tokdict.iterkeys() else: return iter(self.__tokdict) def _itervalues(self): return (self[k] for k in self._iterkeys()) def _iteritems(self): return ((k, self[k]) for k in self._iterkeys()) if PY_3: keys = _iterkeys """Returns an iterator of all named result keys.""" values = _itervalues """Returns an iterator of all named result values.""" items = _iteritems """Returns an iterator of all named result key-value tuples.""" else: iterkeys = _iterkeys """Returns an iterator of all named result keys (Python 2.x only).""" itervalues = _itervalues """Returns an iterator of all named result values (Python 2.x only).""" iteritems = _iteritems """Returns an iterator of all named result key-value tuples (Python 2.x only).""" def keys(self): """Returns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).""" return list(self.iterkeys()) def values(self): """Returns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).""" return list(self.itervalues()) def items(self): """Returns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).""" return list(self.iteritems()) def haskeys(self): """Since keys() returns an iterator, this method is helpful in bypassing code that looks for the existence of any defined results names.""" return bool(self.__tokdict) def pop(self, *args, **kwargs): """ Removes and returns item at specified index (default= ``last``). Supports both ``list`` and ``dict`` semantics for ``pop()``. If passed no argument or an integer argument, it will use ``list`` semantics and pop tokens from the list of parsed tokens. If passed a non-integer argument (most likely a string), it will use ``dict`` semantics and pop the corresponding value from any defined results names. A second default return value argument is supported, just as in ``dict.pop()``. Example:: def remove_first(tokens): tokens.pop(0) print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321'] label = Word(alphas) patt = label("LABEL") + OneOrMore(Word(nums)) print(patt.parseString("AAB 123 321").dump()) # Use pop() in a parse action to remove named result (note that corresponding value is not # removed from list form of results) def remove_LABEL(tokens): tokens.pop("LABEL") return tokens patt.addParseAction(remove_LABEL) print(patt.parseString("AAB 123 321").dump()) prints:: ['AAB', '123', '321'] - LABEL: AAB ['AAB', '123', '321'] """ if not args: args = [-1] for k, v in kwargs.items(): if k == 'default': args = (args[0], v) else: raise TypeError("pop() got an unexpected keyword argument '%s'" % k) if (isinstance(args[0], int) or len(args) == 1 or args[0] in self): index = args[0] ret = self[index] del self[index] return ret else: defaultvalue = args[1] return defaultvalue def get(self, key, defaultValue=None): """ Returns named result matching the given key, or if there is no such name, then returns the given ``defaultValue`` or ``None`` if no ``defaultValue`` is specified. Similar to ``dict.get()``. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString("1999/12/31") print(result.get("year")) # -> '1999' print(result.get("hour", "not specified")) # -> 'not specified' print(result.get("hour")) # -> None """ if key in self: return self[key] else: return defaultValue def insert(self, index, insStr): """ Inserts new element at location index in the list of parsed tokens. Similar to ``list.insert()``. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to insert the parse location in the front of the parsed results def insert_locn(locn, tokens): tokens.insert(0, locn) print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321'] """ self.__toklist.insert(index, insStr) # fixup indices in token dictionary for name, occurrences in self.__tokdict.items(): for k, (value, position) in enumerate(occurrences): occurrences[k] = _ParseResultsWithOffset(value, position + (position > index)) def append(self, item): """ Add single element to end of ParseResults list of elements. Example:: print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321'] # use a parse action to compute the sum of the parsed integers, and add it to the end def append_sum(tokens): tokens.append(sum(map(int, tokens))) print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444] """ self.__toklist.append(item) def extend(self, itemseq): """ Add sequence of elements to end of ParseResults list of elements. Example:: patt = OneOrMore(Word(alphas)) # use a parse action to append the reverse of the matched strings, to make a palindrome def make_palindrome(tokens): tokens.extend(reversed([t[::-1] for t in tokens])) return ''.join(tokens) print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl' """ if isinstance(itemseq, ParseResults): self.__iadd__(itemseq) else: self.__toklist.extend(itemseq) def clear(self): """ Clear all elements and results names. """ del self.__toklist[:] self.__tokdict.clear() def __getattr__(self, name): try: return self[name] except KeyError: return "" def __add__(self, other): ret = self.copy() ret += other return ret def __iadd__(self, other): if other.__tokdict: offset = len(self.__toklist) addoffset = lambda a: offset if a < 0 else a + offset otheritems = other.__tokdict.items() otherdictitems = [(k, _ParseResultsWithOffset(v[0], addoffset(v[1]))) for k, vlist in otheritems for v in vlist] for k, v in otherdictitems: self[k] = v if isinstance(v[0], ParseResults): v[0].__parent = wkref(self) self.__toklist += other.__toklist self.__accumNames.update(other.__accumNames) return self def __radd__(self, other): if isinstance(other, int) and other == 0: # useful for merging many ParseResults using sum() builtin return self.copy() else: # this may raise a TypeError - so be it return other + self def __repr__(self): return "(%s, %s)" % (repr(self.__toklist), repr(self.__tokdict)) def __str__(self): return '[' + ', '.join(_ustr(i) if isinstance(i, ParseResults) else repr(i) for i in self.__toklist) + ']' def _asStringList(self, sep=''): out = [] for item in self.__toklist: if out and sep: out.append(sep) if isinstance(item, ParseResults): out += item._asStringList() else: out.append(_ustr(item)) return out def asList(self): """ Returns the parse results as a nested list of matching tokens, all converted to strings. Example:: patt = OneOrMore(Word(alphas)) result = patt.parseString("sldkj lsdkj sldkj") # even though the result prints in string-like form, it is actually a pyparsing ParseResults print(type(result), result) # -> ['sldkj', 'lsdkj', 'sldkj'] # Use asList() to create an actual list result_list = result.asList() print(type(result_list), result_list) # -> ['sldkj', 'lsdkj', 'sldkj'] """ return [res.asList() if isinstance(res, ParseResults) else res for res in self.__toklist] def asDict(self): """ Returns the named parse results as a nested dictionary. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString('12/31/1999') print(type(result), repr(result)) # -> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]}) result_dict = result.asDict() print(type(result_dict), repr(result_dict)) # -> {'day': '1999', 'year': '12', 'month': '31'} # even though a ParseResults supports dict-like access, sometime you just need to have a dict import json print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"} """ if PY_3: item_fn = self.items else: item_fn = self.iteritems def toItem(obj): if isinstance(obj, ParseResults): if obj.haskeys(): return obj.asDict() else: return [toItem(v) for v in obj] else: return obj return dict((k, toItem(v)) for k, v in item_fn()) def copy(self): """ Returns a new copy of a :class:`ParseResults` object. """ ret = ParseResults(self.__toklist) ret.__tokdict = dict(self.__tokdict.items()) ret.__parent = self.__parent ret.__accumNames.update(self.__accumNames) ret.__name = self.__name return ret def asXML(self, doctag=None, namedItemsOnly=False, indent="", formatted=True): """ (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names. """ nl = "\n" out = [] namedItems = dict((v[1], k) for (k, vlist) in self.__tokdict.items() for v in vlist) nextLevelIndent = indent + " " # collapse out indents if formatting is not desired if not formatted: indent = "" nextLevelIndent = "" nl = "" selfTag = None if doctag is not None: selfTag = doctag else: if self.__name: selfTag = self.__name if not selfTag: if namedItemsOnly: return "" else: selfTag = "ITEM" out += [nl, indent, "<", selfTag, ">"] for i, res in enumerate(self.__toklist): if isinstance(res, ParseResults): if i in namedItems: out += [res.asXML(namedItems[i], namedItemsOnly and doctag is None, nextLevelIndent, formatted)] else: out += [res.asXML(None, namedItemsOnly and doctag is None, nextLevelIndent, formatted)] else: # individual token, see if there is a name for it resTag = None if i in namedItems: resTag = namedItems[i] if not resTag: if namedItemsOnly: continue else: resTag = "ITEM" xmlBodyText = _xml_escape(_ustr(res)) out += [nl, nextLevelIndent, "<", resTag, ">", xmlBodyText, ""] out += [nl, indent, ""] return "".join(out) def __lookup(self, sub): for k, vlist in self.__tokdict.items(): for v, loc in vlist: if sub is v: return k return None def getName(self): r""" Returns the results name for this token expression. Useful when several different expressions might match at a particular location. Example:: integer = Word(nums) ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d") house_number_expr = Suppress('#') + Word(nums, alphanums) user_data = (Group(house_number_expr)("house_number") | Group(ssn_expr)("ssn") | Group(integer)("age")) user_info = OneOrMore(user_data) result = user_info.parseString("22 111-22-3333 #221B") for item in result: print(item.getName(), ':', item[0]) prints:: age : 22 ssn : 111-22-3333 house_number : 221B """ if self.__name: return self.__name elif self.__parent: par = self.__parent() if par: return par.__lookup(self) else: return None elif (len(self) == 1 and len(self.__tokdict) == 1 and next(iter(self.__tokdict.values()))[0][1] in (0, -1)): return next(iter(self.__tokdict.keys())) else: return None def dump(self, indent='', full=True, include_list=True, _depth=0): """ Diagnostic method for listing out the contents of a :class:`ParseResults`. Accepts an optional ``indent`` argument so that this string can be embedded in a nested display of other data. Example:: integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") result = date_str.parseString('12/31/1999') print(result.dump()) prints:: ['12', '/', '31', '/', '1999'] - day: 1999 - month: 31 - year: 12 """ out = [] NL = '\n' if include_list: out.append(indent + _ustr(self.asList())) else: out.append('') if full: if self.haskeys(): items = sorted((str(k), v) for k, v in self.items()) for k, v in items: if out: out.append(NL) out.append("%s%s- %s: " % (indent, (' ' * _depth), k)) if isinstance(v, ParseResults): if v: out.append(v.dump(indent=indent, full=full, include_list=include_list, _depth=_depth + 1)) else: out.append(_ustr(v)) else: out.append(repr(v)) elif any(isinstance(vv, ParseResults) for vv in self): v = self for i, vv in enumerate(v): if isinstance(vv, ParseResults): out.append("\n%s%s[%d]:\n%s%s%s" % (indent, (' ' * (_depth)), i, indent, (' ' * (_depth + 1)), vv.dump(indent=indent, full=full, include_list=include_list, _depth=_depth + 1))) else: out.append("\n%s%s[%d]:\n%s%s%s" % (indent, (' ' * (_depth)), i, indent, (' ' * (_depth + 1)), _ustr(vv))) return "".join(out) def pprint(self, *args, **kwargs): """ Pretty-printer for parsed results as a list, using the `pprint `_ module. Accepts additional positional or keyword args as defined for `pprint.pprint `_ . Example:: ident = Word(alphas, alphanums) num = Word(nums) func = Forward() term = ident | num | Group('(' + func + ')') func <<= ident + Group(Optional(delimitedList(term))) result = func.parseString("fna a,b,(fnb c,d,200),100") result.pprint(width=40) prints:: ['fna', ['a', 'b', ['(', 'fnb', ['c', 'd', '200'], ')'], '100']] """ pprint.pprint(self.asList(), *args, **kwargs) # add support for pickle protocol def __getstate__(self): return (self.__toklist, (self.__tokdict.copy(), self.__parent is not None and self.__parent() or None, self.__accumNames, self.__name)) def __setstate__(self, state): self.__toklist = state[0] self.__tokdict, par, inAccumNames, self.__name = state[1] self.__accumNames = {} self.__accumNames.update(inAccumNames) if par is not None: self.__parent = wkref(par) else: self.__parent = None def __getnewargs__(self): return self.__toklist, self.__name, self.__asList, self.__modal def __dir__(self): return dir(type(self)) + list(self.keys()) @classmethod def from_dict(cls, other, name=None): """ Helper classmethod to construct a ParseResults from a dict, preserving the name-value relations as results names. If an optional 'name' argument is given, a nested ParseResults will be returned """ def is_iterable(obj): try: iter(obj) except Exception: return False else: if PY_3: return not isinstance(obj, (str, bytes)) else: return not isinstance(obj, basestring) ret = cls([]) for k, v in other.items(): if isinstance(v, Mapping): ret += cls.from_dict(v, name=k) else: ret += cls([v], name=k, asList=is_iterable(v)) if name is not None: ret = cls([ret], name=name) return ret MutableMapping.register(ParseResults) def col (loc, strg): """Returns current column within a string, counting newlines as line separators. The first column is number 1. Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See :class:`ParserElement.parseString` for more information on parsing strings containing ```` s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. """ s = strg return 1 if 0 < loc < len(s) and s[loc-1] == '\n' else loc - s.rfind("\n", 0, loc) def lineno(loc, strg): """Returns current line number within a string, counting newlines as line separators. The first line is number 1. Note - the default parsing behavior is to expand tabs in the input string before starting the parsing process. See :class:`ParserElement.parseString` for more information on parsing strings containing ```` s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. """ return strg.count("\n", 0, loc) + 1 def line(loc, strg): """Returns the line of text containing loc within a string, counting newlines as line separators. """ lastCR = strg.rfind("\n", 0, loc) nextCR = strg.find("\n", loc) if nextCR >= 0: return strg[lastCR + 1:nextCR] else: return strg[lastCR + 1:] def _defaultStartDebugAction(instring, loc, expr): print(("Match " + _ustr(expr) + " at loc " + _ustr(loc) + "(%d,%d)" % (lineno(loc, instring), col(loc, instring)))) def _defaultSuccessDebugAction(instring, startloc, endloc, expr, toks): print("Matched " + _ustr(expr) + " -> " + str(toks.asList())) def _defaultExceptionDebugAction(instring, loc, expr, exc): print("Exception raised:" + _ustr(exc)) def nullDebugAction(*args): """'Do-nothing' debug action, to suppress debugging output during parsing.""" pass # Only works on Python 3.x - nonlocal is toxic to Python 2 installs #~ 'decorator to trim function calls to match the arity of the target' #~ def _trim_arity(func, maxargs=3): #~ if func in singleArgBuiltins: #~ return lambda s,l,t: func(t) #~ limit = 0 #~ foundArity = False #~ def wrapper(*args): #~ nonlocal limit,foundArity #~ while 1: #~ try: #~ ret = func(*args[limit:]) #~ foundArity = True #~ return ret #~ except TypeError: #~ if limit == maxargs or foundArity: #~ raise #~ limit += 1 #~ continue #~ return wrapper # this version is Python 2.x-3.x cross-compatible 'decorator to trim function calls to match the arity of the target' def _trim_arity(func, maxargs=2): if func in singleArgBuiltins: return lambda s, l, t: func(t) limit = [0] foundArity = [False] # traceback return data structure changed in Py3.5 - normalize back to plain tuples if system_version[:2] >= (3, 5): def extract_stack(limit=0): # special handling for Python 3.5.0 - extra deep call stack by 1 offset = -3 if system_version == (3, 5, 0) else -2 frame_summary = traceback.extract_stack(limit=-offset + limit - 1)[offset] return [frame_summary[:2]] def extract_tb(tb, limit=0): frames = traceback.extract_tb(tb, limit=limit) frame_summary = frames[-1] return [frame_summary[:2]] else: extract_stack = traceback.extract_stack extract_tb = traceback.extract_tb # synthesize what would be returned by traceback.extract_stack at the call to # user's parse action 'func', so that we don't incur call penalty at parse time LINE_DIFF = 6 # IF ANY CODE CHANGES, EVEN JUST COMMENTS OR BLANK LINES, BETWEEN THE NEXT LINE AND # THE CALL TO FUNC INSIDE WRAPPER, LINE_DIFF MUST BE MODIFIED!!!! this_line = extract_stack(limit=2)[-1] pa_call_line_synth = (this_line[0], this_line[1] + LINE_DIFF) def wrapper(*args): while 1: try: ret = func(*args[limit[0]:]) foundArity[0] = True return ret except TypeError: # re-raise TypeErrors if they did not come from our arity testing if foundArity[0]: raise else: try: tb = sys.exc_info()[-1] if not extract_tb(tb, limit=2)[-1][:2] == pa_call_line_synth: raise finally: try: del tb except NameError: pass if limit[0] <= maxargs: limit[0] += 1 continue raise # copy func name to wrapper for sensible debug output func_name = "" try: func_name = getattr(func, '__name__', getattr(func, '__class__').__name__) except Exception: func_name = str(func) wrapper.__name__ = func_name return wrapper class ParserElement(object): """Abstract base level parser element class.""" DEFAULT_WHITE_CHARS = " \n\t\r" verbose_stacktrace = False @staticmethod def setDefaultWhitespaceChars(chars): r""" Overrides the default whitespace chars Example:: # default whitespace chars are space, and newline OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def', 'ghi', 'jkl'] # change to just treat newline as significant ParserElement.setDefaultWhitespaceChars(" \t") OneOrMore(Word(alphas)).parseString("abc def\nghi jkl") # -> ['abc', 'def'] """ ParserElement.DEFAULT_WHITE_CHARS = chars @staticmethod def inlineLiteralsUsing(cls): """ Set class to be used for inclusion of string literals into a parser. Example:: # default literal class used is Literal integer = Word(nums) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # change to Suppress ParserElement.inlineLiteralsUsing(Suppress) date_str = integer("year") + '/' + integer("month") + '/' + integer("day") date_str.parseString("1999/12/31") # -> ['1999', '12', '31'] """ ParserElement._literalStringClass = cls def __init__(self, savelist=False): self.parseAction = list() self.failAction = None # ~ self.name = "" # don't define self.name, let subclasses try/except upcall self.strRepr = None self.resultsName = None self.saveAsList = savelist self.skipWhitespace = True self.whiteChars = set(ParserElement.DEFAULT_WHITE_CHARS) self.copyDefaultWhiteChars = True self.mayReturnEmpty = False # used when checking for left-recursion self.keepTabs = False self.ignoreExprs = list() self.debug = False self.streamlined = False self.mayIndexError = True # used to optimize exception handling for subclasses that don't advance parse index self.errmsg = "" self.modalResults = True # used to mark results names as modal (report only last) or cumulative (list all) self.debugActions = (None, None, None) # custom debug actions self.re = None self.callPreparse = True # used to avoid redundant calls to preParse self.callDuringTry = False def copy(self): """ Make a copy of this :class:`ParserElement`. Useful for defining different parse actions for the same parsing pattern, using copies of the original parse element. Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) integerK = integer.copy().addParseAction(lambda toks: toks[0] * 1024) + Suppress("K") integerM = integer.copy().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M") print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M")) prints:: [5120, 100, 655360, 268435456] Equivalent form of ``expr.copy()`` is just ``expr()``:: integerM = integer().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M") """ cpy = copy.copy(self) cpy.parseAction = self.parseAction[:] cpy.ignoreExprs = self.ignoreExprs[:] if self.copyDefaultWhiteChars: cpy.whiteChars = ParserElement.DEFAULT_WHITE_CHARS return cpy def setName(self, name): """ Define name for this expression, makes debugging and exception messages clearer. Example:: Word(nums).parseString("ABC") # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1) Word(nums).setName("integer").parseString("ABC") # -> Exception: Expected integer (at char 0), (line:1, col:1) """ self.name = name self.errmsg = "Expected " + self.name if __diag__.enable_debug_on_named_expressions: self.setDebug() return self def setResultsName(self, name, listAllMatches=False): """ Define name for referencing matching tokens as a nested attribute of the returned parse results. NOTE: this returns a *copy* of the original :class:`ParserElement` object; this is so that the client can define a basic element, such as an integer, and reference it in multiple places with different names. You can also set results names using the abbreviated syntax, ``expr("name")`` in place of ``expr.setResultsName("name")`` - see :class:`__call__`. Example:: date_str = (integer.setResultsName("year") + '/' + integer.setResultsName("month") + '/' + integer.setResultsName("day")) # equivalent form: date_str = integer("year") + '/' + integer("month") + '/' + integer("day") """ return self._setResultsName(name, listAllMatches) def _setResultsName(self, name, listAllMatches=False): newself = self.copy() if name.endswith("*"): name = name[:-1] listAllMatches = True newself.resultsName = name newself.modalResults = not listAllMatches return newself def setBreak(self, breakFlag=True): """Method to invoke the Python pdb debugger when this element is about to be parsed. Set ``breakFlag`` to True to enable, False to disable. """ if breakFlag: _parseMethod = self._parse def breaker(instring, loc, doActions=True, callPreParse=True): import pdb # this call to pdb.set_trace() is intentional, not a checkin error pdb.set_trace() return _parseMethod(instring, loc, doActions, callPreParse) breaker._originalParseMethod = _parseMethod self._parse = breaker else: if hasattr(self._parse, "_originalParseMethod"): self._parse = self._parse._originalParseMethod return self def setParseAction(self, *fns, **kwargs): """ Define one or more actions to perform when successfully matching parse element definition. Parse action fn is a callable method with 0-3 arguments, called as ``fn(s, loc, toks)`` , ``fn(loc, toks)`` , ``fn(toks)`` , or just ``fn()`` , where: - s = the original string being parsed (see note below) - loc = the location of the matching substring - toks = a list of the matched tokens, packaged as a :class:`ParseResults` object If the functions in fns modify the tokens, they can return them as the return value from fn, and the modified list of tokens will replace the original. Otherwise, fn does not need to return any value. If None is passed as the parse action, all previously added parse actions for this expression are cleared. Optional keyword arguments: - callDuringTry = (default= ``False``) indicate if parse action should be run during lookaheads and alternate testing Note: the default parsing behavior is to expand tabs in the input string before starting the parsing process. See :class:`parseString for more information on parsing strings containing ```` s, and suggested methods to maintain a consistent view of the parsed string, the parse location, and line and column positions within the parsed string. Example:: integer = Word(nums) date_str = integer + '/' + integer + '/' + integer date_str.parseString("1999/12/31") # -> ['1999', '/', '12', '/', '31'] # use parse action to convert to ints at parse time integer = Word(nums).setParseAction(lambda toks: int(toks[0])) date_str = integer + '/' + integer + '/' + integer # note that integer fields are now ints, not strings date_str.parseString("1999/12/31") # -> [1999, '/', 12, '/', 31] """ if list(fns) == [None,]: self.parseAction = [] else: if not all(callable(fn) for fn in fns): raise TypeError("parse actions must be callable") self.parseAction = list(map(_trim_arity, list(fns))) self.callDuringTry = kwargs.get("callDuringTry", False) return self def addParseAction(self, *fns, **kwargs): """ Add one or more parse actions to expression's list of parse actions. See :class:`setParseAction`. See examples in :class:`copy`. """ self.parseAction += list(map(_trim_arity, list(fns))) self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False) return self def addCondition(self, *fns, **kwargs): """Add a boolean predicate function to expression's list of parse actions. See :class:`setParseAction` for function call signatures. Unlike ``setParseAction``, functions passed to ``addCondition`` need to return boolean success/fail of the condition. Optional keyword arguments: - message = define a custom message to be used in the raised exception - fatal = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException Example:: integer = Word(nums).setParseAction(lambda toks: int(toks[0])) year_int = integer.copy() year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later") date_str = year_int + '/' + integer + '/' + integer result = date_str.parseString("1999/12/31") # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1) """ for fn in fns: self.parseAction.append(conditionAsParseAction(fn, message=kwargs.get('message'), fatal=kwargs.get('fatal', False))) self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False) return self def setFailAction(self, fn): """Define action to perform if parsing fails at this expression. Fail acton fn is a callable function that takes the arguments ``fn(s, loc, expr, err)`` where: - s = string being parsed - loc = location where expression match was attempted and failed - expr = the parse expression that failed - err = the exception thrown The function returns no value. It may throw :class:`ParseFatalException` if it is desired to stop parsing immediately.""" self.failAction = fn return self def _skipIgnorables(self, instring, loc): exprsFound = True while exprsFound: exprsFound = False for e in self.ignoreExprs: try: while 1: loc, dummy = e._parse(instring, loc) exprsFound = True except ParseException: pass return loc def preParse(self, instring, loc): if self.ignoreExprs: loc = self._skipIgnorables(instring, loc) if self.skipWhitespace: wt = self.whiteChars instrlen = len(instring) while loc < instrlen and instring[loc] in wt: loc += 1 return loc def parseImpl(self, instring, loc, doActions=True): return loc, [] def postParse(self, instring, loc, tokenlist): return tokenlist # ~ @profile def _parseNoCache(self, instring, loc, doActions=True, callPreParse=True): TRY, MATCH, FAIL = 0, 1, 2 debugging = (self.debug) # and doActions) if debugging or self.failAction: # ~ print ("Match", self, "at loc", loc, "(%d, %d)" % (lineno(loc, instring), col(loc, instring))) if self.debugActions[TRY]: self.debugActions[TRY](instring, loc, self) try: if callPreParse and self.callPreparse: preloc = self.preParse(instring, loc) else: preloc = loc tokensStart = preloc if self.mayIndexError or preloc >= len(instring): try: loc, tokens = self.parseImpl(instring, preloc, doActions) except IndexError: raise ParseException(instring, len(instring), self.errmsg, self) else: loc, tokens = self.parseImpl(instring, preloc, doActions) except Exception as err: # ~ print ("Exception raised:", err) if self.debugActions[FAIL]: self.debugActions[FAIL](instring, tokensStart, self, err) if self.failAction: self.failAction(instring, tokensStart, self, err) raise else: if callPreParse and self.callPreparse: preloc = self.preParse(instring, loc) else: preloc = loc tokensStart = preloc if self.mayIndexError or preloc >= len(instring): try: loc, tokens = self.parseImpl(instring, preloc, doActions) except IndexError: raise ParseException(instring, len(instring), self.errmsg, self) else: loc, tokens = self.parseImpl(instring, preloc, doActions) tokens = self.postParse(instring, loc, tokens) retTokens = ParseResults(tokens, self.resultsName, asList=self.saveAsList, modal=self.modalResults) if self.parseAction and (doActions or self.callDuringTry): if debugging: try: for fn in self.parseAction: try: tokens = fn(instring, tokensStart, retTokens) except IndexError as parse_action_exc: exc = ParseException("exception raised in parse action") exc.__cause__ = parse_action_exc raise exc if tokens is not None and tokens is not retTokens: retTokens = ParseResults(tokens, self.resultsName, asList=self.saveAsList and isinstance(tokens, (ParseResults, list)), modal=self.modalResults) except Exception as err: # ~ print "Exception raised in user parse action:", err if self.debugActions[FAIL]: self.debugActions[FAIL](instring, tokensStart, self, err) raise else: for fn in self.parseAction: try: tokens = fn(instring, tokensStart, retTokens) except IndexError as parse_action_exc: exc = ParseException("exception raised in parse action") exc.__cause__ = parse_action_exc raise exc if tokens is not None and tokens is not retTokens: retTokens = ParseResults(tokens, self.resultsName, asList=self.saveAsList and isinstance(tokens, (ParseResults, list)), modal=self.modalResults) if debugging: # ~ print ("Matched", self, "->", retTokens.asList()) if self.debugActions[MATCH]: self.debugActions[MATCH](instring, tokensStart, loc, self, retTokens) return loc, retTokens def tryParse(self, instring, loc): try: return self._parse(instring, loc, doActions=False)[0] except ParseFatalException: raise ParseException(instring, loc, self.errmsg, self) def canParseNext(self, instring, loc): try: self.tryParse(instring, loc) except (ParseException, IndexError): return False else: return True class _UnboundedCache(object): def __init__(self): cache = {} self.not_in_cache = not_in_cache = object() def get(self, key): return cache.get(key, not_in_cache) def set(self, key, value): cache[key] = value def clear(self): cache.clear() def cache_len(self): return len(cache) self.get = types.MethodType(get, self) self.set = types.MethodType(set, self) self.clear = types.MethodType(clear, self) self.__len__ = types.MethodType(cache_len, self) if _OrderedDict is not None: class _FifoCache(object): def __init__(self, size): self.not_in_cache = not_in_cache = object() cache = _OrderedDict() def get(self, key): return cache.get(key, not_in_cache) def set(self, key, value): cache[key] = value while len(cache) > size: try: cache.popitem(False) except KeyError: pass def clear(self): cache.clear() def cache_len(self): return len(cache) self.get = types.MethodType(get, self) self.set = types.MethodType(set, self) self.clear = types.MethodType(clear, self) self.__len__ = types.MethodType(cache_len, self) else: class _FifoCache(object): def __init__(self, size): self.not_in_cache = not_in_cache = object() cache = {} key_fifo = collections.deque([], size) def get(self, key): return cache.get(key, not_in_cache) def set(self, key, value): cache[key] = value while len(key_fifo) > size: cache.pop(key_fifo.popleft(), None) key_fifo.append(key) def clear(self): cache.clear() key_fifo.clear() def cache_len(self): return len(cache) self.get = types.MethodType(get, self) self.set = types.MethodType(set, self) self.clear = types.MethodType(clear, self) self.__len__ = types.MethodType(cache_len, self) # argument cache for optimizing repeated calls when backtracking through recursive expressions packrat_cache = {} # this is set later by enabledPackrat(); this is here so that resetCache() doesn't fail packrat_cache_lock = RLock() packrat_cache_stats = [0, 0] # this method gets repeatedly called during backtracking with the same arguments - # we can cache these arguments and save ourselves the trouble of re-parsing the contained expression def _parseCache(self, instring, loc, doActions=True, callPreParse=True): HIT, MISS = 0, 1 lookup = (self, instring, loc, callPreParse, doActions) with ParserElement.packrat_cache_lock: cache = ParserElement.packrat_cache value = cache.get(lookup) if value is cache.not_in_cache: ParserElement.packrat_cache_stats[MISS] += 1 try: value = self._parseNoCache(instring, loc, doActions, callPreParse) except ParseBaseException as pe: # cache a copy of the exception, without the traceback cache.set(lookup, pe.__class__(*pe.args)) raise else: cache.set(lookup, (value[0], value[1].copy())) return value else: ParserElement.packrat_cache_stats[HIT] += 1 if isinstance(value, Exception): raise value return value[0], value[1].copy() _parse = _parseNoCache @staticmethod def resetCache(): ParserElement.packrat_cache.clear() ParserElement.packrat_cache_stats[:] = [0] * len(ParserElement.packrat_cache_stats) _packratEnabled = False @staticmethod def enablePackrat(cache_size_limit=128): """Enables "packrat" parsing, which adds memoizing to the parsing logic. Repeated parse attempts at the same string location (which happens often in many complex grammars) can immediately return a cached value, instead of re-executing parsing/validating code. Memoizing is done of both valid results and parsing exceptions. Parameters: - cache_size_limit - (default= ``128``) - if an integer value is provided will limit the size of the packrat cache; if None is passed, then the cache size will be unbounded; if 0 is passed, the cache will be effectively disabled. This speedup may break existing programs that use parse actions that have side-effects. For this reason, packrat parsing is disabled when you first import pyparsing. To activate the packrat feature, your program must call the class method :class:`ParserElement.enablePackrat`. For best results, call ``enablePackrat()`` immediately after importing pyparsing. Example:: from pip._vendor import pyparsing pyparsing.ParserElement.enablePackrat() """ if not ParserElement._packratEnabled: ParserElement._packratEnabled = True if cache_size_limit is None: ParserElement.packrat_cache = ParserElement._UnboundedCache() else: ParserElement.packrat_cache = ParserElement._FifoCache(cache_size_limit) ParserElement._parse = ParserElement._parseCache def parseString(self, instring, parseAll=False): """ Execute the parse expression with the given string. This is the main interface to the client code, once the complete expression has been built. Returns the parsed data as a :class:`ParseResults` object, which may be accessed as a list, or as a dict or object with attributes if the given parser includes results names. If you want the grammar to require that the entire input string be successfully parsed, then set ``parseAll`` to True (equivalent to ending the grammar with ``StringEnd()``). Note: ``parseString`` implicitly calls ``expandtabs()`` on the input string, in order to report proper column numbers in parse actions. If the input string contains tabs and the grammar uses parse actions that use the ``loc`` argument to index into the string being parsed, you can ensure you have a consistent view of the input string by: - calling ``parseWithTabs`` on your grammar before calling ``parseString`` (see :class:`parseWithTabs`) - define your parse action using the full ``(s, loc, toks)`` signature, and reference the input string using the parse action's ``s`` argument - explictly expand the tabs in your input string before calling ``parseString`` Example:: Word('a').parseString('aaaaabaaa') # -> ['aaaaa'] Word('a').parseString('aaaaabaaa', parseAll=True) # -> Exception: Expected end of text """ ParserElement.resetCache() if not self.streamlined: self.streamline() # ~ self.saveAsList = True for e in self.ignoreExprs: e.streamline() if not self.keepTabs: instring = instring.expandtabs() try: loc, tokens = self._parse(instring, 0) if parseAll: loc = self.preParse(instring, loc) se = Empty() + StringEnd() se._parse(instring, loc) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc else: return tokens def scanString(self, instring, maxMatches=_MAX_INT, overlap=False): """ Scan the input string for expression matches. Each match will return the matching tokens, start location, and end location. May be called with optional ``maxMatches`` argument, to clip scanning after 'n' matches are found. If ``overlap`` is specified, then overlapping matches will be reported. Note that the start and end locations are reported relative to the string being parsed. See :class:`parseString` for more information on parsing strings with embedded tabs. Example:: source = "sldjf123lsdjjkf345sldkjf879lkjsfd987" print(source) for tokens, start, end in Word(alphas).scanString(source): print(' '*start + '^'*(end-start)) print(' '*start + tokens[0]) prints:: sldjf123lsdjjkf345sldkjf879lkjsfd987 ^^^^^ sldjf ^^^^^^^ lsdjjkf ^^^^^^ sldkjf ^^^^^^ lkjsfd """ if not self.streamlined: self.streamline() for e in self.ignoreExprs: e.streamline() if not self.keepTabs: instring = _ustr(instring).expandtabs() instrlen = len(instring) loc = 0 preparseFn = self.preParse parseFn = self._parse ParserElement.resetCache() matches = 0 try: while loc <= instrlen and matches < maxMatches: try: preloc = preparseFn(instring, loc) nextLoc, tokens = parseFn(instring, preloc, callPreParse=False) except ParseException: loc = preloc + 1 else: if nextLoc > loc: matches += 1 yield tokens, preloc, nextLoc if overlap: nextloc = preparseFn(instring, loc) if nextloc > loc: loc = nextLoc else: loc += 1 else: loc = nextLoc else: loc = preloc + 1 except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def transformString(self, instring): """ Extension to :class:`scanString`, to modify matching text with modified tokens that may be returned from a parse action. To use ``transformString``, define a grammar and attach a parse action to it that modifies the returned token list. Invoking ``transformString()`` on a target string will then scan for matches, and replace the matched text patterns according to the logic in the parse action. ``transformString()`` returns the resulting transformed string. Example:: wd = Word(alphas) wd.setParseAction(lambda toks: toks[0].title()) print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york.")) prints:: Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York. """ out = [] lastE = 0 # force preservation of s, to minimize unwanted transformation of string, and to # keep string locs straight between transformString and scanString self.keepTabs = True try: for t, s, e in self.scanString(instring): out.append(instring[lastE:s]) if t: if isinstance(t, ParseResults): out += t.asList() elif isinstance(t, list): out += t else: out.append(t) lastE = e out.append(instring[lastE:]) out = [o for o in out if o] return "".join(map(_ustr, _flatten(out))) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def searchString(self, instring, maxMatches=_MAX_INT): """ Another extension to :class:`scanString`, simplifying the access to the tokens found to match the given parse expression. May be called with optional ``maxMatches`` argument, to clip searching after 'n' matches are found. Example:: # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters cap_word = Word(alphas.upper(), alphas.lower()) print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")) # the sum() builtin can be used to merge results into a single ParseResults object print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))) prints:: [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']] ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity'] """ try: return ParseResults([t for t, s, e in self.scanString(instring, maxMatches)]) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def split(self, instring, maxsplit=_MAX_INT, includeSeparators=False): """ Generator method to split a string using the given expression as a separator. May be called with optional ``maxsplit`` argument, to limit the number of splits; and the optional ``includeSeparators`` argument (default= ``False``), if the separating matching text should be included in the split results. Example:: punc = oneOf(list(".,;:/-!?")) print(list(punc.split("This, this?, this sentence, is badly punctuated!"))) prints:: ['This', ' this', '', ' this sentence', ' is badly punctuated', ''] """ splits = 0 last = 0 for t, s, e in self.scanString(instring, maxMatches=maxsplit): yield instring[last:s] if includeSeparators: yield t[0] last = e yield instring[last:] def __add__(self, other): """ Implementation of + operator - returns :class:`And`. Adding strings to a ParserElement converts them to :class:`Literal`s by default. Example:: greet = Word(alphas) + "," + Word(alphas) + "!" hello = "Hello, World!" print (hello, "->", greet.parseString(hello)) prints:: Hello, World! -> ['Hello', ',', 'World', '!'] ``...`` may be used as a parse expression as a short form of :class:`SkipTo`. Literal('start') + ... + Literal('end') is equivalent to: Literal('start') + SkipTo('end')("_skipped*") + Literal('end') Note that the skipped text is returned with '_skipped' as a results name, and to support having multiple skips in the same parser, the value returned is a list of all skipped text. """ if other is Ellipsis: return _PendingSkip(self) if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return And([self, other]) def __radd__(self, other): """ Implementation of + operator when left operand is not a :class:`ParserElement` """ if other is Ellipsis: return SkipTo(self)("_skipped*") + self if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other + self def __sub__(self, other): """ Implementation of - operator, returns :class:`And` with error stop """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return self + And._ErrorStop() + other def __rsub__(self, other): """ Implementation of - operator when left operand is not a :class:`ParserElement` """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other - self def __mul__(self, other): """ Implementation of * operator, allows use of ``expr * 3`` in place of ``expr + expr + expr``. Expressions may also me multiplied by a 2-integer tuple, similar to ``{min, max}`` multipliers in regular expressions. Tuples may also include ``None`` as in: - ``expr*(n, None)`` or ``expr*(n, )`` is equivalent to ``expr*n + ZeroOrMore(expr)`` (read as "at least n instances of ``expr``") - ``expr*(None, n)`` is equivalent to ``expr*(0, n)`` (read as "0 to n instances of ``expr``") - ``expr*(None, None)`` is equivalent to ``ZeroOrMore(expr)`` - ``expr*(1, None)`` is equivalent to ``OneOrMore(expr)`` Note that ``expr*(None, n)`` does not raise an exception if more than n exprs exist in the input stream; that is, ``expr*(None, n)`` does not enforce a maximum number of expr occurrences. If this behavior is desired, then write ``expr*(None, n) + ~expr`` """ if other is Ellipsis: other = (0, None) elif isinstance(other, tuple) and other[:1] == (Ellipsis,): other = ((0, ) + other[1:] + (None,))[:2] if isinstance(other, int): minElements, optElements = other, 0 elif isinstance(other, tuple): other = tuple(o if o is not Ellipsis else None for o in other) other = (other + (None, None))[:2] if other[0] is None: other = (0, other[1]) if isinstance(other[0], int) and other[1] is None: if other[0] == 0: return ZeroOrMore(self) if other[0] == 1: return OneOrMore(self) else: return self * other[0] + ZeroOrMore(self) elif isinstance(other[0], int) and isinstance(other[1], int): minElements, optElements = other optElements -= minElements else: raise TypeError("cannot multiply 'ParserElement' and ('%s', '%s') objects", type(other[0]), type(other[1])) else: raise TypeError("cannot multiply 'ParserElement' and '%s' objects", type(other)) if minElements < 0: raise ValueError("cannot multiply ParserElement by negative value") if optElements < 0: raise ValueError("second tuple value must be greater or equal to first tuple value") if minElements == optElements == 0: raise ValueError("cannot multiply ParserElement by 0 or (0, 0)") if optElements: def makeOptionalList(n): if n > 1: return Optional(self + makeOptionalList(n - 1)) else: return Optional(self) if minElements: if minElements == 1: ret = self + makeOptionalList(optElements) else: ret = And([self] * minElements) + makeOptionalList(optElements) else: ret = makeOptionalList(optElements) else: if minElements == 1: ret = self else: ret = And([self] * minElements) return ret def __rmul__(self, other): return self.__mul__(other) def __or__(self, other): """ Implementation of | operator - returns :class:`MatchFirst` """ if other is Ellipsis: return _PendingSkip(self, must_skip=True) if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return MatchFirst([self, other]) def __ror__(self, other): """ Implementation of | operator when left operand is not a :class:`ParserElement` """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other | self def __xor__(self, other): """ Implementation of ^ operator - returns :class:`Or` """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return Or([self, other]) def __rxor__(self, other): """ Implementation of ^ operator when left operand is not a :class:`ParserElement` """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other ^ self def __and__(self, other): """ Implementation of & operator - returns :class:`Each` """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return Each([self, other]) def __rand__(self, other): """ Implementation of & operator when left operand is not a :class:`ParserElement` """ if isinstance(other, basestring): other = self._literalStringClass(other) if not isinstance(other, ParserElement): warnings.warn("Cannot combine element of type %s with ParserElement" % type(other), SyntaxWarning, stacklevel=2) return None return other & self def __invert__(self): """ Implementation of ~ operator - returns :class:`NotAny` """ return NotAny(self) def __iter__(self): # must implement __iter__ to override legacy use of sequential access to __getitem__ to # iterate over a sequence raise TypeError('%r object is not iterable' % self.__class__.__name__) def __getitem__(self, key): """ use ``[]`` indexing notation as a short form for expression repetition: - ``expr[n]`` is equivalent to ``expr*n`` - ``expr[m, n]`` is equivalent to ``expr*(m, n)`` - ``expr[n, ...]`` or ``expr[n,]`` is equivalent to ``expr*n + ZeroOrMore(expr)`` (read as "at least n instances of ``expr``") - ``expr[..., n]`` is equivalent to ``expr*(0, n)`` (read as "0 to n instances of ``expr``") - ``expr[...]`` and ``expr[0, ...]`` are equivalent to ``ZeroOrMore(expr)`` - ``expr[1, ...]`` is equivalent to ``OneOrMore(expr)`` ``None`` may be used in place of ``...``. Note that ``expr[..., n]`` and ``expr[m, n]``do not raise an exception if more than ``n`` ``expr``s exist in the input stream. If this behavior is desired, then write ``expr[..., n] + ~expr``. """ # convert single arg keys to tuples try: if isinstance(key, str): key = (key,) iter(key) except TypeError: key = (key, key) if len(key) > 2: warnings.warn("only 1 or 2 index arguments supported ({0}{1})".format(key[:5], '... [{0}]'.format(len(key)) if len(key) > 5 else '')) # clip to 2 elements ret = self * tuple(key[:2]) return ret def __call__(self, name=None): """ Shortcut for :class:`setResultsName`, with ``listAllMatches=False``. If ``name`` is given with a trailing ``'*'`` character, then ``listAllMatches`` will be passed as ``True``. If ``name` is omitted, same as calling :class:`copy`. Example:: # these are equivalent userdata = Word(alphas).setResultsName("name") + Word(nums + "-").setResultsName("socsecno") userdata = Word(alphas)("name") + Word(nums + "-")("socsecno") """ if name is not None: return self._setResultsName(name) else: return self.copy() def suppress(self): """ Suppresses the output of this :class:`ParserElement`; useful to keep punctuation from cluttering up returned output. """ return Suppress(self) def leaveWhitespace(self): """ Disables the skipping of whitespace before matching the characters in the :class:`ParserElement`'s defined pattern. This is normally only used internally by the pyparsing module, but may be needed in some whitespace-sensitive grammars. """ self.skipWhitespace = False return self def setWhitespaceChars(self, chars): """ Overrides the default whitespace chars """ self.skipWhitespace = True self.whiteChars = chars self.copyDefaultWhiteChars = False return self def parseWithTabs(self): """ Overrides default behavior to expand ````s to spaces before parsing the input string. Must be called before ``parseString`` when the input grammar contains elements that match ```` characters. """ self.keepTabs = True return self def ignore(self, other): """ Define expression to be ignored (e.g., comments) while doing pattern matching; may be called repeatedly, to define multiple comment or other ignorable patterns. Example:: patt = OneOrMore(Word(alphas)) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj'] patt.ignore(cStyleComment) patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd'] """ if isinstance(other, basestring): other = Suppress(other) if isinstance(other, Suppress): if other not in self.ignoreExprs: self.ignoreExprs.append(other) else: self.ignoreExprs.append(Suppress(other.copy())) return self def setDebugActions(self, startAction, successAction, exceptionAction): """ Enable display of debugging messages while doing pattern matching. """ self.debugActions = (startAction or _defaultStartDebugAction, successAction or _defaultSuccessDebugAction, exceptionAction or _defaultExceptionDebugAction) self.debug = True return self def setDebug(self, flag=True): """ Enable display of debugging messages while doing pattern matching. Set ``flag`` to True to enable, False to disable. Example:: wd = Word(alphas).setName("alphaword") integer = Word(nums).setName("numword") term = wd | integer # turn on debugging for wd wd.setDebug() OneOrMore(term).parseString("abc 123 xyz 890") prints:: Match alphaword at loc 0(1,1) Matched alphaword -> ['abc'] Match alphaword at loc 3(1,4) Exception raised:Expected alphaword (at char 4), (line:1, col:5) Match alphaword at loc 7(1,8) Matched alphaword -> ['xyz'] Match alphaword at loc 11(1,12) Exception raised:Expected alphaword (at char 12), (line:1, col:13) Match alphaword at loc 15(1,16) Exception raised:Expected alphaword (at char 15), (line:1, col:16) The output shown is that produced by the default debug actions - custom debug actions can be specified using :class:`setDebugActions`. Prior to attempting to match the ``wd`` expression, the debugging message ``"Match at loc (,)"`` is shown. Then if the parse succeeds, a ``"Matched"`` message is shown, or an ``"Exception raised"`` message is shown. Also note the use of :class:`setName` to assign a human-readable name to the expression, which makes debugging and exception messages easier to understand - for instance, the default name created for the :class:`Word` expression without calling ``setName`` is ``"W:(ABCD...)"``. """ if flag: self.setDebugActions(_defaultStartDebugAction, _defaultSuccessDebugAction, _defaultExceptionDebugAction) else: self.debug = False return self def __str__(self): return self.name def __repr__(self): return _ustr(self) def streamline(self): self.streamlined = True self.strRepr = None return self def checkRecursion(self, parseElementList): pass def validate(self, validateTrace=None): """ Check defined expressions for valid structure, check for infinite recursive definitions. """ self.checkRecursion([]) def parseFile(self, file_or_filename, parseAll=False): """ Execute the parse expression on the given file or filename. If a filename is specified (instead of a file object), the entire file is opened, read, and closed before parsing. """ try: file_contents = file_or_filename.read() except AttributeError: with open(file_or_filename, "r") as f: file_contents = f.read() try: return self.parseString(file_contents, parseAll) except ParseBaseException as exc: if ParserElement.verbose_stacktrace: raise else: # catch and re-raise exception from here, clears out pyparsing internal stack trace raise exc def __eq__(self, other): if isinstance(other, ParserElement): if PY_3: self is other or super(ParserElement, self).__eq__(other) else: return self is other or vars(self) == vars(other) elif isinstance(other, basestring): return self.matches(other) else: return super(ParserElement, self) == other def __ne__(self, other): return not (self == other) def __hash__(self): return id(self) def __req__(self, other): return self == other def __rne__(self, other): return not (self == other) def matches(self, testString, parseAll=True): """ Method for quick testing of a parser against a test string. Good for simple inline microtests of sub expressions while building up larger parser. Parameters: - testString - to test against this expression for a match - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests Example:: expr = Word(nums) assert expr.matches("100") """ try: self.parseString(_ustr(testString), parseAll=parseAll) return True except ParseBaseException: return False def runTests(self, tests, parseAll=True, comment='#', fullDump=True, printResults=True, failureTests=False, postParse=None, file=None): """ Execute the parse expression on a series of test strings, showing each test, the parsed results or where the parse failed. Quick and easy way to run a parse expression against a list of sample strings. Parameters: - tests - a list of separate test strings, or a multiline string of test strings - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests - comment - (default= ``'#'``) - expression for indicating embedded comments in the test string; pass None to disable comment filtering - fullDump - (default= ``True``) - dump results as list followed by results names in nested outline; if False, only dump nested list - printResults - (default= ``True``) prints test output to stdout - failureTests - (default= ``False``) indicates if these tests are expected to fail parsing - postParse - (default= ``None``) optional callback for successful parse results; called as `fn(test_string, parse_results)` and returns a string to be added to the test output - file - (default=``None``) optional file-like object to which test output will be written; if None, will default to ``sys.stdout`` Returns: a (success, results) tuple, where success indicates that all tests succeeded (or failed if ``failureTests`` is True), and the results contain a list of lines of each test's output Example:: number_expr = pyparsing_common.number.copy() result = number_expr.runTests(''' # unsigned integer 100 # negative integer -100 # float with scientific notation 6.02e23 # integer with scientific notation 1e-12 ''') print("Success" if result[0] else "Failed!") result = number_expr.runTests(''' # stray character 100Z # missing leading digit before '.' -.100 # too many '.' 3.14.159 ''', failureTests=True) print("Success" if result[0] else "Failed!") prints:: # unsigned integer 100 [100] # negative integer -100 [-100] # float with scientific notation 6.02e23 [6.02e+23] # integer with scientific notation 1e-12 [1e-12] Success # stray character 100Z ^ FAIL: Expected end of text (at char 3), (line:1, col:4) # missing leading digit before '.' -.100 ^ FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1) # too many '.' 3.14.159 ^ FAIL: Expected end of text (at char 4), (line:1, col:5) Success Each test string must be on a single line. If you want to test a string that spans multiple lines, create a test like this:: expr.runTest(r"this is a test\\n of strings that spans \\n 3 lines") (Note that this is a raw string literal, you must include the leading 'r'.) """ if isinstance(tests, basestring): tests = list(map(str.strip, tests.rstrip().splitlines())) if isinstance(comment, basestring): comment = Literal(comment) if file is None: file = sys.stdout print_ = file.write allResults = [] comments = [] success = True NL = Literal(r'\n').addParseAction(replaceWith('\n')).ignore(quotedString) BOM = u'\ufeff' for t in tests: if comment is not None and comment.matches(t, False) or comments and not t: comments.append(t) continue if not t: continue out = ['\n'.join(comments), t] comments = [] try: # convert newline marks to actual newlines, and strip leading BOM if present t = NL.transformString(t.lstrip(BOM)) result = self.parseString(t, parseAll=parseAll) except ParseBaseException as pe: fatal = "(FATAL)" if isinstance(pe, ParseFatalException) else "" if '\n' in t: out.append(line(pe.loc, t)) out.append(' ' * (col(pe.loc, t) - 1) + '^' + fatal) else: out.append(' ' * pe.loc + '^' + fatal) out.append("FAIL: " + str(pe)) success = success and failureTests result = pe except Exception as exc: out.append("FAIL-EXCEPTION: " + str(exc)) success = success and failureTests result = exc else: success = success and not failureTests if postParse is not None: try: pp_value = postParse(t, result) if pp_value is not None: if isinstance(pp_value, ParseResults): out.append(pp_value.dump()) else: out.append(str(pp_value)) else: out.append(result.dump()) except Exception as e: out.append(result.dump(full=fullDump)) out.append("{0} failed: {1}: {2}".format(postParse.__name__, type(e).__name__, e)) else: out.append(result.dump(full=fullDump)) if printResults: if fullDump: out.append('') print_('\n'.join(out)) allResults.append((t, result)) return success, allResults class _PendingSkip(ParserElement): # internal placeholder class to hold a place were '...' is added to a parser element, # once another ParserElement is added, this placeholder will be replaced with a SkipTo def __init__(self, expr, must_skip=False): super(_PendingSkip, self).__init__() self.strRepr = str(expr + Empty()).replace('Empty', '...') self.name = self.strRepr self.anchor = expr self.must_skip = must_skip def __add__(self, other): skipper = SkipTo(other).setName("...")("_skipped*") if self.must_skip: def must_skip(t): if not t._skipped or t._skipped.asList() == ['']: del t[0] t.pop("_skipped", None) def show_skip(t): if t._skipped.asList()[-1:] == ['']: skipped = t.pop('_skipped') t['_skipped'] = 'missing <' + repr(self.anchor) + '>' return (self.anchor + skipper().addParseAction(must_skip) | skipper().addParseAction(show_skip)) + other return self.anchor + skipper + other def __repr__(self): return self.strRepr def parseImpl(self, *args): raise Exception("use of `...` expression without following SkipTo target expression") class Token(ParserElement): """Abstract :class:`ParserElement` subclass, for defining atomic matching patterns. """ def __init__(self): super(Token, self).__init__(savelist=False) class Empty(Token): """An empty token, will always match. """ def __init__(self): super(Empty, self).__init__() self.name = "Empty" self.mayReturnEmpty = True self.mayIndexError = False class NoMatch(Token): """A token that will never match. """ def __init__(self): super(NoMatch, self).__init__() self.name = "NoMatch" self.mayReturnEmpty = True self.mayIndexError = False self.errmsg = "Unmatchable token" def parseImpl(self, instring, loc, doActions=True): raise ParseException(instring, loc, self.errmsg, self) class Literal(Token): """Token to exactly match a specified string. Example:: Literal('blah').parseString('blah') # -> ['blah'] Literal('blah').parseString('blahfooblah') # -> ['blah'] Literal('blah').parseString('bla') # -> Exception: Expected "blah" For case-insensitive matching, use :class:`CaselessLiteral`. For keyword matching (force word break before and after the matched string), use :class:`Keyword` or :class:`CaselessKeyword`. """ def __init__(self, matchString): super(Literal, self).__init__() self.match = matchString self.matchLen = len(matchString) try: self.firstMatchChar = matchString[0] except IndexError: warnings.warn("null string passed to Literal; use Empty() instead", SyntaxWarning, stacklevel=2) self.__class__ = Empty self.name = '"%s"' % _ustr(self.match) self.errmsg = "Expected " + self.name self.mayReturnEmpty = False self.mayIndexError = False # Performance tuning: modify __class__ to select # a parseImpl optimized for single-character check if self.matchLen == 1 and type(self) is Literal: self.__class__ = _SingleCharLiteral def parseImpl(self, instring, loc, doActions=True): if instring[loc] == self.firstMatchChar and instring.startswith(self.match, loc): return loc + self.matchLen, self.match raise ParseException(instring, loc, self.errmsg, self) class _SingleCharLiteral(Literal): def parseImpl(self, instring, loc, doActions=True): if instring[loc] == self.firstMatchChar: return loc + 1, self.match raise ParseException(instring, loc, self.errmsg, self) _L = Literal ParserElement._literalStringClass = Literal class Keyword(Token): """Token to exactly match a specified string as a keyword, that is, it must be immediately followed by a non-keyword character. Compare with :class:`Literal`: - ``Literal("if")`` will match the leading ``'if'`` in ``'ifAndOnlyIf'``. - ``Keyword("if")`` will not; it will only match the leading ``'if'`` in ``'if x=1'``, or ``'if(y==2)'`` Accepts two optional constructor arguments in addition to the keyword string: - ``identChars`` is a string of characters that would be valid identifier characters, defaulting to all alphanumerics + "_" and "$" - ``caseless`` allows case-insensitive matching, default is ``False``. Example:: Keyword("start").parseString("start") # -> ['start'] Keyword("start").parseString("starting") # -> Exception For case-insensitive matching, use :class:`CaselessKeyword`. """ DEFAULT_KEYWORD_CHARS = alphanums + "_$" def __init__(self, matchString, identChars=None, caseless=False): super(Keyword, self).__init__() if identChars is None: identChars = Keyword.DEFAULT_KEYWORD_CHARS self.match = matchString self.matchLen = len(matchString) try: self.firstMatchChar = matchString[0] except IndexError: warnings.warn("null string passed to Keyword; use Empty() instead", SyntaxWarning, stacklevel=2) self.name = '"%s"' % self.match self.errmsg = "Expected " + self.name self.mayReturnEmpty = False self.mayIndexError = False self.caseless = caseless if caseless: self.caselessmatch = matchString.upper() identChars = identChars.upper() self.identChars = set(identChars) def parseImpl(self, instring, loc, doActions=True): if self.caseless: if ((instring[loc:loc + self.matchLen].upper() == self.caselessmatch) and (loc >= len(instring) - self.matchLen or instring[loc + self.matchLen].upper() not in self.identChars) and (loc == 0 or instring[loc - 1].upper() not in self.identChars)): return loc + self.matchLen, self.match else: if instring[loc] == self.firstMatchChar: if ((self.matchLen == 1 or instring.startswith(self.match, loc)) and (loc >= len(instring) - self.matchLen or instring[loc + self.matchLen] not in self.identChars) and (loc == 0 or instring[loc - 1] not in self.identChars)): return loc + self.matchLen, self.match raise ParseException(instring, loc, self.errmsg, self) def copy(self): c = super(Keyword, self).copy() c.identChars = Keyword.DEFAULT_KEYWORD_CHARS return c @staticmethod def setDefaultKeywordChars(chars): """Overrides the default Keyword chars """ Keyword.DEFAULT_KEYWORD_CHARS = chars class CaselessLiteral(Literal): """Token to match a specified string, ignoring case of letters. Note: the matched results will always be in the case of the given match string, NOT the case of the input text. Example:: OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD'] (Contrast with example for :class:`CaselessKeyword`.) """ def __init__(self, matchString): super(CaselessLiteral, self).__init__(matchString.upper()) # Preserve the defining literal. self.returnString = matchString self.name = "'%s'" % self.returnString self.errmsg = "Expected " + self.name def parseImpl(self, instring, loc, doActions=True): if instring[loc:loc + self.matchLen].upper() == self.match: return loc + self.matchLen, self.returnString raise ParseException(instring, loc, self.errmsg, self) class CaselessKeyword(Keyword): """ Caseless version of :class:`Keyword`. Example:: OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD'] (Contrast with example for :class:`CaselessLiteral`.) """ def __init__(self, matchString, identChars=None): super(CaselessKeyword, self).__init__(matchString, identChars, caseless=True) class CloseMatch(Token): """A variation on :class:`Literal` which matches "close" matches, that is, strings with at most 'n' mismatching characters. :class:`CloseMatch` takes parameters: - ``match_string`` - string to be matched - ``maxMismatches`` - (``default=1``) maximum number of mismatches allowed to count as a match The results from a successful parse will contain the matched text from the input string and the following named results: - ``mismatches`` - a list of the positions within the match_string where mismatches were found - ``original`` - the original match_string used to compare against the input string If ``mismatches`` is an empty list, then the match was an exact match. Example:: patt = CloseMatch("ATCATCGAATGGA") patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']}) patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1) # exact match patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']}) # close match allowing up to 2 mismatches patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2) patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']}) """ def __init__(self, match_string, maxMismatches=1): super(CloseMatch, self).__init__() self.name = match_string self.match_string = match_string self.maxMismatches = maxMismatches self.errmsg = "Expected %r (with up to %d mismatches)" % (self.match_string, self.maxMismatches) self.mayIndexError = False self.mayReturnEmpty = False def parseImpl(self, instring, loc, doActions=True): start = loc instrlen = len(instring) maxloc = start + len(self.match_string) if maxloc <= instrlen: match_string = self.match_string match_stringloc = 0 mismatches = [] maxMismatches = self.maxMismatches for match_stringloc, s_m in enumerate(zip(instring[loc:maxloc], match_string)): src, mat = s_m if src != mat: mismatches.append(match_stringloc) if len(mismatches) > maxMismatches: break else: loc = match_stringloc + 1 results = ParseResults([instring[start:loc]]) results['original'] = match_string results['mismatches'] = mismatches return loc, results raise ParseException(instring, loc, self.errmsg, self) class Word(Token): """Token for matching words composed of allowed character sets. Defined with string containing all allowed initial characters, an optional string containing allowed body characters (if omitted, defaults to the initial character set), and an optional minimum, maximum, and/or exact length. The default value for ``min`` is 1 (a minimum value < 1 is not valid); the default values for ``max`` and ``exact`` are 0, meaning no maximum or exact length restriction. An optional ``excludeChars`` parameter can list characters that might be found in the input ``bodyChars`` string; useful to define a word of all printables except for one or two characters, for instance. :class:`srange` is useful for defining custom character set strings for defining ``Word`` expressions, using range notation from regular expression character sets. A common mistake is to use :class:`Word` to match a specific literal string, as in ``Word("Address")``. Remember that :class:`Word` uses the string argument to define *sets* of matchable characters. This expression would match "Add", "AAA", "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an exact literal string, use :class:`Literal` or :class:`Keyword`. pyparsing includes helper strings for building Words: - :class:`alphas` - :class:`nums` - :class:`alphanums` - :class:`hexnums` - :class:`alphas8bit` (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.) - :class:`punc8bit` (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.) - :class:`printables` (any non-whitespace character) Example:: # a word composed of digits integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9")) # a word with a leading capital, and zero or more lowercase capital_word = Word(alphas.upper(), alphas.lower()) # hostnames are alphanumeric, with leading alpha, and '-' hostname = Word(alphas, alphanums + '-') # roman numeral (not a strict parser, accepts invalid mix of characters) roman = Word("IVXLCDM") # any string of non-whitespace characters, except for ',' csv_value = Word(printables, excludeChars=",") """ def __init__(self, initChars, bodyChars=None, min=1, max=0, exact=0, asKeyword=False, excludeChars=None): super(Word, self).__init__() if excludeChars: excludeChars = set(excludeChars) initChars = ''.join(c for c in initChars if c not in excludeChars) if bodyChars: bodyChars = ''.join(c for c in bodyChars if c not in excludeChars) self.initCharsOrig = initChars self.initChars = set(initChars) if bodyChars: self.bodyCharsOrig = bodyChars self.bodyChars = set(bodyChars) else: self.bodyCharsOrig = initChars self.bodyChars = set(initChars) self.maxSpecified = max > 0 if min < 1: raise ValueError("cannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permitted") self.minLen = min if max > 0: self.maxLen = max else: self.maxLen = _MAX_INT if exact > 0: self.maxLen = exact self.minLen = exact self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayIndexError = False self.asKeyword = asKeyword if ' ' not in self.initCharsOrig + self.bodyCharsOrig and (min == 1 and max == 0 and exact == 0): if self.bodyCharsOrig == self.initCharsOrig: self.reString = "[%s]+" % _escapeRegexRangeChars(self.initCharsOrig) elif len(self.initCharsOrig) == 1: self.reString = "%s[%s]*" % (re.escape(self.initCharsOrig), _escapeRegexRangeChars(self.bodyCharsOrig),) else: self.reString = "[%s][%s]*" % (_escapeRegexRangeChars(self.initCharsOrig), _escapeRegexRangeChars(self.bodyCharsOrig),) if self.asKeyword: self.reString = r"\b" + self.reString + r"\b" try: self.re = re.compile(self.reString) except Exception: self.re = None else: self.re_match = self.re.match self.__class__ = _WordRegex def parseImpl(self, instring, loc, doActions=True): if instring[loc] not in self.initChars: raise ParseException(instring, loc, self.errmsg, self) start = loc loc += 1 instrlen = len(instring) bodychars = self.bodyChars maxloc = start + self.maxLen maxloc = min(maxloc, instrlen) while loc < maxloc and instring[loc] in bodychars: loc += 1 throwException = False if loc - start < self.minLen: throwException = True elif self.maxSpecified and loc < instrlen and instring[loc] in bodychars: throwException = True elif self.asKeyword: if (start > 0 and instring[start - 1] in bodychars or loc < instrlen and instring[loc] in bodychars): throwException = True if throwException: raise ParseException(instring, loc, self.errmsg, self) return loc, instring[start:loc] def __str__(self): try: return super(Word, self).__str__() except Exception: pass if self.strRepr is None: def charsAsStr(s): if len(s) > 4: return s[:4] + "..." else: return s if self.initCharsOrig != self.bodyCharsOrig: self.strRepr = "W:(%s, %s)" % (charsAsStr(self.initCharsOrig), charsAsStr(self.bodyCharsOrig)) else: self.strRepr = "W:(%s)" % charsAsStr(self.initCharsOrig) return self.strRepr class _WordRegex(Word): def parseImpl(self, instring, loc, doActions=True): result = self.re_match(instring, loc) if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() return loc, result.group() class Char(_WordRegex): """A short-cut class for defining ``Word(characters, exact=1)``, when defining a match of any single character in a string of characters. """ def __init__(self, charset, asKeyword=False, excludeChars=None): super(Char, self).__init__(charset, exact=1, asKeyword=asKeyword, excludeChars=excludeChars) self.reString = "[%s]" % _escapeRegexRangeChars(''.join(self.initChars)) if asKeyword: self.reString = r"\b%s\b" % self.reString self.re = re.compile(self.reString) self.re_match = self.re.match class Regex(Token): r"""Token for matching strings that match a given regular expression. Defined with string specifying the regular expression in a form recognized by the stdlib Python `re module `_. If the given regex contains named groups (defined using ``(?P...)``), these will be preserved as named parse results. Example:: realnum = Regex(r"[+-]?\d+\.\d*") date = Regex(r'(?P\d{4})-(?P\d\d?)-(?P\d\d?)') # ref: https://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression roman = Regex(r"M{0,4}(CM|CD|D?{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})") """ compiledREtype = type(re.compile("[A-Z]")) def __init__(self, pattern, flags=0, asGroupList=False, asMatch=False): """The parameters ``pattern`` and ``flags`` are passed to the ``re.compile()`` function as-is. See the Python `re module `_ module for an explanation of the acceptable patterns and flags. """ super(Regex, self).__init__() if isinstance(pattern, basestring): if not pattern: warnings.warn("null string passed to Regex; use Empty() instead", SyntaxWarning, stacklevel=2) self.pattern = pattern self.flags = flags try: self.re = re.compile(self.pattern, self.flags) self.reString = self.pattern except sre_constants.error: warnings.warn("invalid pattern (%s) passed to Regex" % pattern, SyntaxWarning, stacklevel=2) raise elif isinstance(pattern, Regex.compiledREtype): self.re = pattern self.pattern = self.reString = str(pattern) self.flags = flags else: raise ValueError("Regex may only be constructed with a string or a compiled RE object") self.re_match = self.re.match self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayIndexError = False self.mayReturnEmpty = True self.asGroupList = asGroupList self.asMatch = asMatch if self.asGroupList: self.parseImpl = self.parseImplAsGroupList if self.asMatch: self.parseImpl = self.parseImplAsMatch def parseImpl(self, instring, loc, doActions=True): result = self.re_match(instring, loc) if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() ret = ParseResults(result.group()) d = result.groupdict() if d: for k, v in d.items(): ret[k] = v return loc, ret def parseImplAsGroupList(self, instring, loc, doActions=True): result = self.re_match(instring, loc) if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() ret = result.groups() return loc, ret def parseImplAsMatch(self, instring, loc, doActions=True): result = self.re_match(instring, loc) if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() ret = result return loc, ret def __str__(self): try: return super(Regex, self).__str__() except Exception: pass if self.strRepr is None: self.strRepr = "Re:(%s)" % repr(self.pattern) return self.strRepr def sub(self, repl): r""" Return Regex with an attached parse action to transform the parsed result as if called using `re.sub(expr, repl, string) `_. Example:: make_html = Regex(r"(\w+):(.*?):").sub(r"<\1>\2") print(make_html.transformString("h1:main title:")) # prints "

main title

" """ if self.asGroupList: warnings.warn("cannot use sub() with Regex(asGroupList=True)", SyntaxWarning, stacklevel=2) raise SyntaxError() if self.asMatch and callable(repl): warnings.warn("cannot use sub() with a callable with Regex(asMatch=True)", SyntaxWarning, stacklevel=2) raise SyntaxError() if self.asMatch: def pa(tokens): return tokens[0].expand(repl) else: def pa(tokens): return self.re.sub(repl, tokens[0]) return self.addParseAction(pa) class QuotedString(Token): r""" Token for matching strings that are delimited by quoting characters. Defined with the following parameters: - quoteChar - string of one or more characters defining the quote delimiting string - escChar - character to escape quotes, typically backslash (default= ``None``) - escQuote - special quote sequence to escape an embedded quote string (such as SQL's ``""`` to escape an embedded ``"``) (default= ``None``) - multiline - boolean indicating whether quotes can span multiple lines (default= ``False``) - unquoteResults - boolean indicating whether the matched text should be unquoted (default= ``True``) - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default= ``None`` => same as quoteChar) - convertWhitespaceEscapes - convert escaped whitespace (``'\t'``, ``'\n'``, etc.) to actual whitespace (default= ``True``) Example:: qs = QuotedString('"') print(qs.searchString('lsjdf "This is the quote" sldjf')) complex_qs = QuotedString('{{', endQuoteChar='}}') print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf')) sql_qs = QuotedString('"', escQuote='""') print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf')) prints:: [['This is the quote']] [['This is the "quote"']] [['This is the quote with "embedded" quotes']] """ def __init__(self, quoteChar, escChar=None, escQuote=None, multiline=False, unquoteResults=True, endQuoteChar=None, convertWhitespaceEscapes=True): super(QuotedString, self).__init__() # remove white space from quote chars - wont work anyway quoteChar = quoteChar.strip() if not quoteChar: warnings.warn("quoteChar cannot be the empty string", SyntaxWarning, stacklevel=2) raise SyntaxError() if endQuoteChar is None: endQuoteChar = quoteChar else: endQuoteChar = endQuoteChar.strip() if not endQuoteChar: warnings.warn("endQuoteChar cannot be the empty string", SyntaxWarning, stacklevel=2) raise SyntaxError() self.quoteChar = quoteChar self.quoteCharLen = len(quoteChar) self.firstQuoteChar = quoteChar[0] self.endQuoteChar = endQuoteChar self.endQuoteCharLen = len(endQuoteChar) self.escChar = escChar self.escQuote = escQuote self.unquoteResults = unquoteResults self.convertWhitespaceEscapes = convertWhitespaceEscapes if multiline: self.flags = re.MULTILINE | re.DOTALL self.pattern = r'%s(?:[^%s%s]' % (re.escape(self.quoteChar), _escapeRegexRangeChars(self.endQuoteChar[0]), (escChar is not None and _escapeRegexRangeChars(escChar) or '')) else: self.flags = 0 self.pattern = r'%s(?:[^%s\n\r%s]' % (re.escape(self.quoteChar), _escapeRegexRangeChars(self.endQuoteChar[0]), (escChar is not None and _escapeRegexRangeChars(escChar) or '')) if len(self.endQuoteChar) > 1: self.pattern += ( '|(?:' + ')|(?:'.join("%s[^%s]" % (re.escape(self.endQuoteChar[:i]), _escapeRegexRangeChars(self.endQuoteChar[i])) for i in range(len(self.endQuoteChar) - 1, 0, -1)) + ')') if escQuote: self.pattern += (r'|(?:%s)' % re.escape(escQuote)) if escChar: self.pattern += (r'|(?:%s.)' % re.escape(escChar)) self.escCharReplacePattern = re.escape(self.escChar) + "(.)" self.pattern += (r')*%s' % re.escape(self.endQuoteChar)) try: self.re = re.compile(self.pattern, self.flags) self.reString = self.pattern self.re_match = self.re.match except sre_constants.error: warnings.warn("invalid pattern (%s) passed to Regex" % self.pattern, SyntaxWarning, stacklevel=2) raise self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayIndexError = False self.mayReturnEmpty = True def parseImpl(self, instring, loc, doActions=True): result = instring[loc] == self.firstQuoteChar and self.re_match(instring, loc) or None if not result: raise ParseException(instring, loc, self.errmsg, self) loc = result.end() ret = result.group() if self.unquoteResults: # strip off quotes ret = ret[self.quoteCharLen: -self.endQuoteCharLen] if isinstance(ret, basestring): # replace escaped whitespace if '\\' in ret and self.convertWhitespaceEscapes: ws_map = { r'\t': '\t', r'\n': '\n', r'\f': '\f', r'\r': '\r', } for wslit, wschar in ws_map.items(): ret = ret.replace(wslit, wschar) # replace escaped characters if self.escChar: ret = re.sub(self.escCharReplacePattern, r"\g<1>", ret) # replace escaped quotes if self.escQuote: ret = ret.replace(self.escQuote, self.endQuoteChar) return loc, ret def __str__(self): try: return super(QuotedString, self).__str__() except Exception: pass if self.strRepr is None: self.strRepr = "quoted string, starting with %s ending with %s" % (self.quoteChar, self.endQuoteChar) return self.strRepr class CharsNotIn(Token): """Token for matching words composed of characters *not* in a given set (will include whitespace in matched characters if not listed in the provided exclusion set - see example). Defined with string containing all disallowed characters, and an optional minimum, maximum, and/or exact length. The default value for ``min`` is 1 (a minimum value < 1 is not valid); the default values for ``max`` and ``exact`` are 0, meaning no maximum or exact length restriction. Example:: # define a comma-separated-value as anything that is not a ',' csv_value = CharsNotIn(',') print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213")) prints:: ['dkls', 'lsdkjf', 's12 34', '@!#', '213'] """ def __init__(self, notChars, min=1, max=0, exact=0): super(CharsNotIn, self).__init__() self.skipWhitespace = False self.notChars = notChars if min < 1: raise ValueError("cannot specify a minimum length < 1; use " "Optional(CharsNotIn()) if zero-length char group is permitted") self.minLen = min if max > 0: self.maxLen = max else: self.maxLen = _MAX_INT if exact > 0: self.maxLen = exact self.minLen = exact self.name = _ustr(self) self.errmsg = "Expected " + self.name self.mayReturnEmpty = (self.minLen == 0) self.mayIndexError = False def parseImpl(self, instring, loc, doActions=True): if instring[loc] in self.notChars: raise ParseException(instring, loc, self.errmsg, self) start = loc loc += 1 notchars = self.notChars maxlen = min(start + self.maxLen, len(instring)) while loc < maxlen and instring[loc] not in notchars: loc += 1 if loc - start < self.minLen: raise ParseException(instring, loc, self.errmsg, self) return loc, instring[start:loc] def __str__(self): try: return super(CharsNotIn, self).__str__() except Exception: pass if self.strRepr is None: if len(self.notChars) > 4: self.strRepr = "!W:(%s...)" % self.notChars[:4] else: self.strRepr = "!W:(%s)" % self.notChars return self.strRepr class White(Token): """Special matching class for matching whitespace. Normally, whitespace is ignored by pyparsing grammars. This class is included when some whitespace structures are significant. Define with a string containing the whitespace characters to be matched; default is ``" \\t\\r\\n"``. Also takes optional ``min``, ``max``, and ``exact`` arguments, as defined for the :class:`Word` class. """ whiteStrs = { ' ' : '', '\t': '', '\n': '', '\r': '', '\f': '', 'u\00A0': '', 'u\1680': '', 'u\180E': '', 'u\2000': '', 'u\2001': '', 'u\2002': '', 'u\2003': '', 'u\2004': '', 'u\2005': '', 'u\2006': '', 'u\2007': '', 'u\2008': '', 'u\2009': '', 'u\200A': '', 'u\200B': '', 'u\202F': '', 'u\205F': '', 'u\3000': '', } def __init__(self, ws=" \t\r\n", min=1, max=0, exact=0): super(White, self).__init__() self.matchWhite = ws self.setWhitespaceChars("".join(c for c in self.whiteChars if c not in self.matchWhite)) # ~ self.leaveWhitespace() self.name = ("".join(White.whiteStrs[c] for c in self.matchWhite)) self.mayReturnEmpty = True self.errmsg = "Expected " + self.name self.minLen = min if max > 0: self.maxLen = max else: self.maxLen = _MAX_INT if exact > 0: self.maxLen = exact self.minLen = exact def parseImpl(self, instring, loc, doActions=True): if instring[loc] not in self.matchWhite: raise ParseException(instring, loc, self.errmsg, self) start = loc loc += 1 maxloc = start + self.maxLen maxloc = min(maxloc, len(instring)) while loc < maxloc and instring[loc] in self.matchWhite: loc += 1 if loc - start < self.minLen: raise ParseException(instring, loc, self.errmsg, self) return loc, instring[start:loc] class _PositionToken(Token): def __init__(self): super(_PositionToken, self).__init__() self.name = self.__class__.__name__ self.mayReturnEmpty = True self.mayIndexError = False class GoToColumn(_PositionToken): """Token to advance to a specific column of input text; useful for tabular report scraping. """ def __init__(self, colno): super(GoToColumn, self).__init__() self.col = colno def preParse(self, instring, loc): if col(loc, instring) != self.col: instrlen = len(instring) if self.ignoreExprs: loc = self._skipIgnorables(instring, loc) while loc < instrlen and instring[loc].isspace() and col(loc, instring) != self.col: loc += 1 return loc def parseImpl(self, instring, loc, doActions=True): thiscol = col(loc, instring) if thiscol > self.col: raise ParseException(instring, loc, "Text not in expected column", self) newloc = loc + self.col - thiscol ret = instring[loc: newloc] return newloc, ret class LineStart(_PositionToken): r"""Matches if current position is at the beginning of a line within the parse string Example:: test = '''\ AAA this line AAA and this line AAA but not this one B AAA and definitely not this one ''' for t in (LineStart() + 'AAA' + restOfLine).searchString(test): print(t) prints:: ['AAA', ' this line'] ['AAA', ' and this line'] """ def __init__(self): super(LineStart, self).__init__() self.errmsg = "Expected start of line" def parseImpl(self, instring, loc, doActions=True): if col(loc, instring) == 1: return loc, [] raise ParseException(instring, loc, self.errmsg, self) class LineEnd(_PositionToken): """Matches if current position is at the end of a line within the parse string """ def __init__(self): super(LineEnd, self).__init__() self.setWhitespaceChars(ParserElement.DEFAULT_WHITE_CHARS.replace("\n", "")) self.errmsg = "Expected end of line" def parseImpl(self, instring, loc, doActions=True): if loc < len(instring): if instring[loc] == "\n": return loc + 1, "\n" else: raise ParseException(instring, loc, self.errmsg, self) elif loc == len(instring): return loc + 1, [] else: raise ParseException(instring, loc, self.errmsg, self) class StringStart(_PositionToken): """Matches if current position is at the beginning of the parse string """ def __init__(self): super(StringStart, self).__init__() self.errmsg = "Expected start of text" def parseImpl(self, instring, loc, doActions=True): if loc != 0: # see if entire string up to here is just whitespace and ignoreables if loc != self.preParse(instring, 0): raise ParseException(instring, loc, self.errmsg, self) return loc, [] class StringEnd(_PositionToken): """Matches if current position is at the end of the parse string """ def __init__(self): super(StringEnd, self).__init__() self.errmsg = "Expected end of text" def parseImpl(self, instring, loc, doActions=True): if loc < len(instring): raise ParseException(instring, loc, self.errmsg, self) elif loc == len(instring): return loc + 1, [] elif loc > len(instring): return loc, [] else: raise ParseException(instring, loc, self.errmsg, self) class WordStart(_PositionToken): """Matches if the current position is at the beginning of a Word, and is not preceded by any character in a given set of ``wordChars`` (default= ``printables``). To emulate the ``\b`` behavior of regular expressions, use ``WordStart(alphanums)``. ``WordStart`` will also match at the beginning of the string being parsed, or at the beginning of a line. """ def __init__(self, wordChars=printables): super(WordStart, self).__init__() self.wordChars = set(wordChars) self.errmsg = "Not at the start of a word" def parseImpl(self, instring, loc, doActions=True): if loc != 0: if (instring[loc - 1] in self.wordChars or instring[loc] not in self.wordChars): raise ParseException(instring, loc, self.errmsg, self) return loc, [] class WordEnd(_PositionToken): """Matches if the current position is at the end of a Word, and is not followed by any character in a given set of ``wordChars`` (default= ``printables``). To emulate the ``\b`` behavior of regular expressions, use ``WordEnd(alphanums)``. ``WordEnd`` will also match at the end of the string being parsed, or at the end of a line. """ def __init__(self, wordChars=printables): super(WordEnd, self).__init__() self.wordChars = set(wordChars) self.skipWhitespace = False self.errmsg = "Not at the end of a word" def parseImpl(self, instring, loc, doActions=True): instrlen = len(instring) if instrlen > 0 and loc < instrlen: if (instring[loc] in self.wordChars or instring[loc - 1] not in self.wordChars): raise ParseException(instring, loc, self.errmsg, self) return loc, [] class ParseExpression(ParserElement): """Abstract subclass of ParserElement, for combining and post-processing parsed tokens. """ def __init__(self, exprs, savelist=False): super(ParseExpression, self).__init__(savelist) if isinstance(exprs, _generatorType): exprs = list(exprs) if isinstance(exprs, basestring): self.exprs = [self._literalStringClass(exprs)] elif isinstance(exprs, ParserElement): self.exprs = [exprs] elif isinstance(exprs, Iterable): exprs = list(exprs) # if sequence of strings provided, wrap with Literal if any(isinstance(expr, basestring) for expr in exprs): exprs = (self._literalStringClass(e) if isinstance(e, basestring) else e for e in exprs) self.exprs = list(exprs) else: try: self.exprs = list(exprs) except TypeError: self.exprs = [exprs] self.callPreparse = False def append(self, other): self.exprs.append(other) self.strRepr = None return self def leaveWhitespace(self): """Extends ``leaveWhitespace`` defined in base class, and also invokes ``leaveWhitespace`` on all contained expressions.""" self.skipWhitespace = False self.exprs = [e.copy() for e in self.exprs] for e in self.exprs: e.leaveWhitespace() return self def ignore(self, other): if isinstance(other, Suppress): if other not in self.ignoreExprs: super(ParseExpression, self).ignore(other) for e in self.exprs: e.ignore(self.ignoreExprs[-1]) else: super(ParseExpression, self).ignore(other) for e in self.exprs: e.ignore(self.ignoreExprs[-1]) return self def __str__(self): try: return super(ParseExpression, self).__str__() except Exception: pass if self.strRepr is None: self.strRepr = "%s:(%s)" % (self.__class__.__name__, _ustr(self.exprs)) return self.strRepr def streamline(self): super(ParseExpression, self).streamline() for e in self.exprs: e.streamline() # collapse nested And's of the form And(And(And(a, b), c), d) to And(a, b, c, d) # but only if there are no parse actions or resultsNames on the nested And's # (likewise for Or's and MatchFirst's) if len(self.exprs) == 2: other = self.exprs[0] if (isinstance(other, self.__class__) and not other.parseAction and other.resultsName is None and not other.debug): self.exprs = other.exprs[:] + [self.exprs[1]] self.strRepr = None self.mayReturnEmpty |= other.mayReturnEmpty self.mayIndexError |= other.mayIndexError other = self.exprs[-1] if (isinstance(other, self.__class__) and not other.parseAction and other.resultsName is None and not other.debug): self.exprs = self.exprs[:-1] + other.exprs[:] self.strRepr = None self.mayReturnEmpty |= other.mayReturnEmpty self.mayIndexError |= other.mayIndexError self.errmsg = "Expected " + _ustr(self) return self def validate(self, validateTrace=None): tmp = (validateTrace if validateTrace is not None else [])[:] + [self] for e in self.exprs: e.validate(tmp) self.checkRecursion([]) def copy(self): ret = super(ParseExpression, self).copy() ret.exprs = [e.copy() for e in self.exprs] return ret def _setResultsName(self, name, listAllMatches=False): if __diag__.warn_ungrouped_named_tokens_in_collection: for e in self.exprs: if isinstance(e, ParserElement) and e.resultsName: warnings.warn("{0}: setting results name {1!r} on {2} expression " "collides with {3!r} on contained expression".format("warn_ungrouped_named_tokens_in_collection", name, type(self).__name__, e.resultsName), stacklevel=3) return super(ParseExpression, self)._setResultsName(name, listAllMatches) class And(ParseExpression): """ Requires all given :class:`ParseExpression` s to be found in the given order. Expressions may be separated by whitespace. May be constructed using the ``'+'`` operator. May also be constructed using the ``'-'`` operator, which will suppress backtracking. Example:: integer = Word(nums) name_expr = OneOrMore(Word(alphas)) expr = And([integer("id"), name_expr("name"), integer("age")]) # more easily written as: expr = integer("id") + name_expr("name") + integer("age") """ class _ErrorStop(Empty): def __init__(self, *args, **kwargs): super(And._ErrorStop, self).__init__(*args, **kwargs) self.name = '-' self.leaveWhitespace() def __init__(self, exprs, savelist=True): if exprs and Ellipsis in exprs: tmp = [] for i, expr in enumerate(exprs): if expr is Ellipsis: if i < len(exprs) - 1: skipto_arg = (Empty() + exprs[i + 1]).exprs[-1] tmp.append(SkipTo(skipto_arg)("_skipped*")) else: raise Exception("cannot construct And with sequence ending in ...") else: tmp.append(expr) exprs[:] = tmp super(And, self).__init__(exprs, savelist) self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs) self.setWhitespaceChars(self.exprs[0].whiteChars) self.skipWhitespace = self.exprs[0].skipWhitespace self.callPreparse = True def streamline(self): # collapse any _PendingSkip's if self.exprs: if any(isinstance(e, ParseExpression) and e.exprs and isinstance(e.exprs[-1], _PendingSkip) for e in self.exprs[:-1]): for i, e in enumerate(self.exprs[:-1]): if e is None: continue if (isinstance(e, ParseExpression) and e.exprs and isinstance(e.exprs[-1], _PendingSkip)): e.exprs[-1] = e.exprs[-1] + self.exprs[i + 1] self.exprs[i + 1] = None self.exprs = [e for e in self.exprs if e is not None] super(And, self).streamline() self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs) return self def parseImpl(self, instring, loc, doActions=True): # pass False as last arg to _parse for first element, since we already # pre-parsed the string as part of our And pre-parsing loc, resultlist = self.exprs[0]._parse(instring, loc, doActions, callPreParse=False) errorStop = False for e in self.exprs[1:]: if isinstance(e, And._ErrorStop): errorStop = True continue if errorStop: try: loc, exprtokens = e._parse(instring, loc, doActions) except ParseSyntaxException: raise except ParseBaseException as pe: pe.__traceback__ = None raise ParseSyntaxException._from_exception(pe) except IndexError: raise ParseSyntaxException(instring, len(instring), self.errmsg, self) else: loc, exprtokens = e._parse(instring, loc, doActions) if exprtokens or exprtokens.haskeys(): resultlist += exprtokens return loc, resultlist def __iadd__(self, other): if isinstance(other, basestring): other = self._literalStringClass(other) return self.append(other) # And([self, other]) def checkRecursion(self, parseElementList): subRecCheckList = parseElementList[:] + [self] for e in self.exprs: e.checkRecursion(subRecCheckList) if not e.mayReturnEmpty: break def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "{" + " ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr class Or(ParseExpression): """Requires that at least one :class:`ParseExpression` is found. If two expressions match, the expression that matches the longest string will be used. May be constructed using the ``'^'`` operator. Example:: # construct Or using '^' operator number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) prints:: [['123'], ['3.1416'], ['789']] """ def __init__(self, exprs, savelist=False): super(Or, self).__init__(exprs, savelist) if self.exprs: self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs) else: self.mayReturnEmpty = True def streamline(self): super(Or, self).streamline() if __compat__.collect_all_And_tokens: self.saveAsList = any(e.saveAsList for e in self.exprs) return self def parseImpl(self, instring, loc, doActions=True): maxExcLoc = -1 maxException = None matches = [] for e in self.exprs: try: loc2 = e.tryParse(instring, loc) except ParseException as err: err.__traceback__ = None if err.loc > maxExcLoc: maxException = err maxExcLoc = err.loc except IndexError: if len(instring) > maxExcLoc: maxException = ParseException(instring, len(instring), e.errmsg, self) maxExcLoc = len(instring) else: # save match among all matches, to retry longest to shortest matches.append((loc2, e)) if matches: # re-evaluate all matches in descending order of length of match, in case attached actions # might change whether or how much they match of the input. matches.sort(key=itemgetter(0), reverse=True) if not doActions: # no further conditions or parse actions to change the selection of # alternative, so the first match will be the best match best_expr = matches[0][1] return best_expr._parse(instring, loc, doActions) longest = -1, None for loc1, expr1 in matches: if loc1 <= longest[0]: # already have a longer match than this one will deliver, we are done return longest try: loc2, toks = expr1._parse(instring, loc, doActions) except ParseException as err: err.__traceback__ = None if err.loc > maxExcLoc: maxException = err maxExcLoc = err.loc else: if loc2 >= loc1: return loc2, toks # didn't match as much as before elif loc2 > longest[0]: longest = loc2, toks if longest != (-1, None): return longest if maxException is not None: maxException.msg = self.errmsg raise maxException else: raise ParseException(instring, loc, "no defined alternatives to match", self) def __ixor__(self, other): if isinstance(other, basestring): other = self._literalStringClass(other) return self.append(other) # Or([self, other]) def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "{" + " ^ ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr def checkRecursion(self, parseElementList): subRecCheckList = parseElementList[:] + [self] for e in self.exprs: e.checkRecursion(subRecCheckList) def _setResultsName(self, name, listAllMatches=False): if (not __compat__.collect_all_And_tokens and __diag__.warn_multiple_tokens_in_named_alternation): if any(isinstance(e, And) for e in self.exprs): warnings.warn("{0}: setting results name {1!r} on {2} expression " "may only return a single token for an And alternative, " "in future will return the full list of tokens".format( "warn_multiple_tokens_in_named_alternation", name, type(self).__name__), stacklevel=3) return super(Or, self)._setResultsName(name, listAllMatches) class MatchFirst(ParseExpression): """Requires that at least one :class:`ParseExpression` is found. If two expressions match, the first one listed is the one that will match. May be constructed using the ``'|'`` operator. Example:: # construct MatchFirst using '|' operator # watch the order of expressions to match number = Word(nums) | Combine(Word(nums) + '.' + Word(nums)) print(number.searchString("123 3.1416 789")) # Fail! -> [['123'], ['3'], ['1416'], ['789']] # put more selective expression first number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums) print(number.searchString("123 3.1416 789")) # Better -> [['123'], ['3.1416'], ['789']] """ def __init__(self, exprs, savelist=False): super(MatchFirst, self).__init__(exprs, savelist) if self.exprs: self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs) else: self.mayReturnEmpty = True def streamline(self): super(MatchFirst, self).streamline() if __compat__.collect_all_And_tokens: self.saveAsList = any(e.saveAsList for e in self.exprs) return self def parseImpl(self, instring, loc, doActions=True): maxExcLoc = -1 maxException = None for e in self.exprs: try: ret = e._parse(instring, loc, doActions) return ret except ParseException as err: if err.loc > maxExcLoc: maxException = err maxExcLoc = err.loc except IndexError: if len(instring) > maxExcLoc: maxException = ParseException(instring, len(instring), e.errmsg, self) maxExcLoc = len(instring) # only got here if no expression matched, raise exception for match that made it the furthest else: if maxException is not None: maxException.msg = self.errmsg raise maxException else: raise ParseException(instring, loc, "no defined alternatives to match", self) def __ior__(self, other): if isinstance(other, basestring): other = self._literalStringClass(other) return self.append(other) # MatchFirst([self, other]) def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "{" + " | ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr def checkRecursion(self, parseElementList): subRecCheckList = parseElementList[:] + [self] for e in self.exprs: e.checkRecursion(subRecCheckList) def _setResultsName(self, name, listAllMatches=False): if (not __compat__.collect_all_And_tokens and __diag__.warn_multiple_tokens_in_named_alternation): if any(isinstance(e, And) for e in self.exprs): warnings.warn("{0}: setting results name {1!r} on {2} expression " "may only return a single token for an And alternative, " "in future will return the full list of tokens".format( "warn_multiple_tokens_in_named_alternation", name, type(self).__name__), stacklevel=3) return super(MatchFirst, self)._setResultsName(name, listAllMatches) class Each(ParseExpression): """Requires all given :class:`ParseExpression` s to be found, but in any order. Expressions may be separated by whitespace. May be constructed using the ``'&'`` operator. Example:: color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN") shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON") integer = Word(nums) shape_attr = "shape:" + shape_type("shape") posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn") color_attr = "color:" + color("color") size_attr = "size:" + integer("size") # use Each (using operator '&') to accept attributes in any order # (shape and posn are required, color and size are optional) shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr) shape_spec.runTests(''' shape: SQUARE color: BLACK posn: 100, 120 shape: CIRCLE size: 50 color: BLUE posn: 50,80 color:GREEN size:20 shape:TRIANGLE posn:20,40 ''' ) prints:: shape: SQUARE color: BLACK posn: 100, 120 ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']] - color: BLACK - posn: ['100', ',', '120'] - x: 100 - y: 120 - shape: SQUARE shape: CIRCLE size: 50 color: BLUE posn: 50,80 ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']] - color: BLUE - posn: ['50', ',', '80'] - x: 50 - y: 80 - shape: CIRCLE - size: 50 color: GREEN size: 20 shape: TRIANGLE posn: 20,40 ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']] - color: GREEN - posn: ['20', ',', '40'] - x: 20 - y: 40 - shape: TRIANGLE - size: 20 """ def __init__(self, exprs, savelist=True): super(Each, self).__init__(exprs, savelist) self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs) self.skipWhitespace = True self.initExprGroups = True self.saveAsList = True def streamline(self): super(Each, self).streamline() self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs) return self def parseImpl(self, instring, loc, doActions=True): if self.initExprGroups: self.opt1map = dict((id(e.expr), e) for e in self.exprs if isinstance(e, Optional)) opt1 = [e.expr for e in self.exprs if isinstance(e, Optional)] opt2 = [e for e in self.exprs if e.mayReturnEmpty and not isinstance(e, Optional)] self.optionals = opt1 + opt2 self.multioptionals = [e.expr for e in self.exprs if isinstance(e, ZeroOrMore)] self.multirequired = [e.expr for e in self.exprs if isinstance(e, OneOrMore)] self.required = [e for e in self.exprs if not isinstance(e, (Optional, ZeroOrMore, OneOrMore))] self.required += self.multirequired self.initExprGroups = False tmpLoc = loc tmpReqd = self.required[:] tmpOpt = self.optionals[:] matchOrder = [] keepMatching = True while keepMatching: tmpExprs = tmpReqd + tmpOpt + self.multioptionals + self.multirequired failed = [] for e in tmpExprs: try: tmpLoc = e.tryParse(instring, tmpLoc) except ParseException: failed.append(e) else: matchOrder.append(self.opt1map.get(id(e), e)) if e in tmpReqd: tmpReqd.remove(e) elif e in tmpOpt: tmpOpt.remove(e) if len(failed) == len(tmpExprs): keepMatching = False if tmpReqd: missing = ", ".join(_ustr(e) for e in tmpReqd) raise ParseException(instring, loc, "Missing one or more required elements (%s)" % missing) # add any unmatched Optionals, in case they have default values defined matchOrder += [e for e in self.exprs if isinstance(e, Optional) and e.expr in tmpOpt] resultlist = [] for e in matchOrder: loc, results = e._parse(instring, loc, doActions) resultlist.append(results) finalResults = sum(resultlist, ParseResults([])) return loc, finalResults def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "{" + " & ".join(_ustr(e) for e in self.exprs) + "}" return self.strRepr def checkRecursion(self, parseElementList): subRecCheckList = parseElementList[:] + [self] for e in self.exprs: e.checkRecursion(subRecCheckList) class ParseElementEnhance(ParserElement): """Abstract subclass of :class:`ParserElement`, for combining and post-processing parsed tokens. """ def __init__(self, expr, savelist=False): super(ParseElementEnhance, self).__init__(savelist) if isinstance(expr, basestring): if issubclass(self._literalStringClass, Token): expr = self._literalStringClass(expr) else: expr = self._literalStringClass(Literal(expr)) self.expr = expr self.strRepr = None if expr is not None: self.mayIndexError = expr.mayIndexError self.mayReturnEmpty = expr.mayReturnEmpty self.setWhitespaceChars(expr.whiteChars) self.skipWhitespace = expr.skipWhitespace self.saveAsList = expr.saveAsList self.callPreparse = expr.callPreparse self.ignoreExprs.extend(expr.ignoreExprs) def parseImpl(self, instring, loc, doActions=True): if self.expr is not None: return self.expr._parse(instring, loc, doActions, callPreParse=False) else: raise ParseException("", loc, self.errmsg, self) def leaveWhitespace(self): self.skipWhitespace = False self.expr = self.expr.copy() if self.expr is not None: self.expr.leaveWhitespace() return self def ignore(self, other): if isinstance(other, Suppress): if other not in self.ignoreExprs: super(ParseElementEnhance, self).ignore(other) if self.expr is not None: self.expr.ignore(self.ignoreExprs[-1]) else: super(ParseElementEnhance, self).ignore(other) if self.expr is not None: self.expr.ignore(self.ignoreExprs[-1]) return self def streamline(self): super(ParseElementEnhance, self).streamline() if self.expr is not None: self.expr.streamline() return self def checkRecursion(self, parseElementList): if self in parseElementList: raise RecursiveGrammarException(parseElementList + [self]) subRecCheckList = parseElementList[:] + [self] if self.expr is not None: self.expr.checkRecursion(subRecCheckList) def validate(self, validateTrace=None): if validateTrace is None: validateTrace = [] tmp = validateTrace[:] + [self] if self.expr is not None: self.expr.validate(tmp) self.checkRecursion([]) def __str__(self): try: return super(ParseElementEnhance, self).__str__() except Exception: pass if self.strRepr is None and self.expr is not None: self.strRepr = "%s:(%s)" % (self.__class__.__name__, _ustr(self.expr)) return self.strRepr class FollowedBy(ParseElementEnhance): """Lookahead matching of the given parse expression. ``FollowedBy`` does *not* advance the parsing position within the input string, it only verifies that the specified parse expression matches at the current position. ``FollowedBy`` always returns a null token list. If any results names are defined in the lookahead expression, those *will* be returned for access by name. Example:: # use FollowedBy to match a label only if it is followed by a ':' data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint() prints:: [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']] """ def __init__(self, expr): super(FollowedBy, self).__init__(expr) self.mayReturnEmpty = True def parseImpl(self, instring, loc, doActions=True): # by using self._expr.parse and deleting the contents of the returned ParseResults list # we keep any named results that were defined in the FollowedBy expression _, ret = self.expr._parse(instring, loc, doActions=doActions) del ret[:] return loc, ret class PrecededBy(ParseElementEnhance): """Lookbehind matching of the given parse expression. ``PrecededBy`` does not advance the parsing position within the input string, it only verifies that the specified parse expression matches prior to the current position. ``PrecededBy`` always returns a null token list, but if a results name is defined on the given expression, it is returned. Parameters: - expr - expression that must match prior to the current parse location - retreat - (default= ``None``) - (int) maximum number of characters to lookbehind prior to the current parse location If the lookbehind expression is a string, Literal, Keyword, or a Word or CharsNotIn with a specified exact or maximum length, then the retreat parameter is not required. Otherwise, retreat must be specified to give a maximum number of characters to look back from the current parse position for a lookbehind match. Example:: # VB-style variable names with type prefixes int_var = PrecededBy("#") + pyparsing_common.identifier str_var = PrecededBy("$") + pyparsing_common.identifier """ def __init__(self, expr, retreat=None): super(PrecededBy, self).__init__(expr) self.expr = self.expr().leaveWhitespace() self.mayReturnEmpty = True self.mayIndexError = False self.exact = False if isinstance(expr, str): retreat = len(expr) self.exact = True elif isinstance(expr, (Literal, Keyword)): retreat = expr.matchLen self.exact = True elif isinstance(expr, (Word, CharsNotIn)) and expr.maxLen != _MAX_INT: retreat = expr.maxLen self.exact = True elif isinstance(expr, _PositionToken): retreat = 0 self.exact = True self.retreat = retreat self.errmsg = "not preceded by " + str(expr) self.skipWhitespace = False def parseImpl(self, instring, loc=0, doActions=True): if self.exact: if loc < self.retreat: raise ParseException(instring, loc, self.errmsg) start = loc - self.retreat _, ret = self.expr._parse(instring, start) else: # retreat specified a maximum lookbehind window, iterate test_expr = self.expr + StringEnd() instring_slice = instring[:loc] last_expr = ParseException(instring, loc, self.errmsg) for offset in range(1, min(loc, self.retreat + 1)): try: _, ret = test_expr._parse(instring_slice, loc - offset) except ParseBaseException as pbe: last_expr = pbe else: break else: raise last_expr # return empty list of tokens, but preserve any defined results names del ret[:] return loc, ret class NotAny(ParseElementEnhance): """Lookahead to disallow matching with the given parse expression. ``NotAny`` does *not* advance the parsing position within the input string, it only verifies that the specified parse expression does *not* match at the current position. Also, ``NotAny`` does *not* skip over leading whitespace. ``NotAny`` always returns a null token list. May be constructed using the '~' operator. Example:: AND, OR, NOT = map(CaselessKeyword, "AND OR NOT".split()) # take care not to mistake keywords for identifiers ident = ~(AND | OR | NOT) + Word(alphas) boolean_term = Optional(NOT) + ident # very crude boolean expression - to support parenthesis groups and # operation hierarchy, use infixNotation boolean_expr = boolean_term + ZeroOrMore((AND | OR) + boolean_term) # integers that are followed by "." are actually floats integer = Word(nums) + ~Char(".") """ def __init__(self, expr): super(NotAny, self).__init__(expr) # ~ self.leaveWhitespace() self.skipWhitespace = False # do NOT use self.leaveWhitespace(), don't want to propagate to exprs self.mayReturnEmpty = True self.errmsg = "Found unwanted token, " + _ustr(self.expr) def parseImpl(self, instring, loc, doActions=True): if self.expr.canParseNext(instring, loc): raise ParseException(instring, loc, self.errmsg, self) return loc, [] def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "~{" + _ustr(self.expr) + "}" return self.strRepr class _MultipleMatch(ParseElementEnhance): def __init__(self, expr, stopOn=None): super(_MultipleMatch, self).__init__(expr) self.saveAsList = True ender = stopOn if isinstance(ender, basestring): ender = self._literalStringClass(ender) self.stopOn(ender) def stopOn(self, ender): if isinstance(ender, basestring): ender = self._literalStringClass(ender) self.not_ender = ~ender if ender is not None else None return self def parseImpl(self, instring, loc, doActions=True): self_expr_parse = self.expr._parse self_skip_ignorables = self._skipIgnorables check_ender = self.not_ender is not None if check_ender: try_not_ender = self.not_ender.tryParse # must be at least one (but first see if we are the stopOn sentinel; # if so, fail) if check_ender: try_not_ender(instring, loc) loc, tokens = self_expr_parse(instring, loc, doActions, callPreParse=False) try: hasIgnoreExprs = (not not self.ignoreExprs) while 1: if check_ender: try_not_ender(instring, loc) if hasIgnoreExprs: preloc = self_skip_ignorables(instring, loc) else: preloc = loc loc, tmptokens = self_expr_parse(instring, preloc, doActions) if tmptokens or tmptokens.haskeys(): tokens += tmptokens except (ParseException, IndexError): pass return loc, tokens def _setResultsName(self, name, listAllMatches=False): if __diag__.warn_ungrouped_named_tokens_in_collection: for e in [self.expr] + getattr(self.expr, 'exprs', []): if isinstance(e, ParserElement) and e.resultsName: warnings.warn("{0}: setting results name {1!r} on {2} expression " "collides with {3!r} on contained expression".format("warn_ungrouped_named_tokens_in_collection", name, type(self).__name__, e.resultsName), stacklevel=3) return super(_MultipleMatch, self)._setResultsName(name, listAllMatches) class OneOrMore(_MultipleMatch): """Repetition of one or more of the given expression. Parameters: - expr - expression that must match one or more times - stopOn - (default= ``None``) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: BLACK" OneOrMore(attr_expr).parseString(text).pprint() # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']] # use stopOn attribute for OneOrMore to avoid reading label string as part of the data attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']] # could also be written as (attr_expr * (1,)).parseString(text).pprint() """ def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "{" + _ustr(self.expr) + "}..." return self.strRepr class ZeroOrMore(_MultipleMatch): """Optional repetition of zero or more of the given expression. Parameters: - expr - expression that must match zero or more times - stopOn - (default= ``None``) - expression for a terminating sentinel (only required if the sentinel would ordinarily match the repetition expression) Example: similar to :class:`OneOrMore` """ def __init__(self, expr, stopOn=None): super(ZeroOrMore, self).__init__(expr, stopOn=stopOn) self.mayReturnEmpty = True def parseImpl(self, instring, loc, doActions=True): try: return super(ZeroOrMore, self).parseImpl(instring, loc, doActions) except (ParseException, IndexError): return loc, [] def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "[" + _ustr(self.expr) + "]..." return self.strRepr class _NullToken(object): def __bool__(self): return False __nonzero__ = __bool__ def __str__(self): return "" class Optional(ParseElementEnhance): """Optional matching of the given expression. Parameters: - expr - expression that must match zero or more times - default (optional) - value to be returned if the optional expression is not found. Example:: # US postal code can be a 5-digit zip, plus optional 4-digit qualifier zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4))) zip.runTests(''' # traditional ZIP code 12345 # ZIP+4 form 12101-0001 # invalid ZIP 98765- ''') prints:: # traditional ZIP code 12345 ['12345'] # ZIP+4 form 12101-0001 ['12101-0001'] # invalid ZIP 98765- ^ FAIL: Expected end of text (at char 5), (line:1, col:6) """ __optionalNotMatched = _NullToken() def __init__(self, expr, default=__optionalNotMatched): super(Optional, self).__init__(expr, savelist=False) self.saveAsList = self.expr.saveAsList self.defaultValue = default self.mayReturnEmpty = True def parseImpl(self, instring, loc, doActions=True): try: loc, tokens = self.expr._parse(instring, loc, doActions, callPreParse=False) except (ParseException, IndexError): if self.defaultValue is not self.__optionalNotMatched: if self.expr.resultsName: tokens = ParseResults([self.defaultValue]) tokens[self.expr.resultsName] = self.defaultValue else: tokens = [self.defaultValue] else: tokens = [] return loc, tokens def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is None: self.strRepr = "[" + _ustr(self.expr) + "]" return self.strRepr class SkipTo(ParseElementEnhance): """Token for skipping over all undefined text until the matched expression is found. Parameters: - expr - target expression marking the end of the data to be skipped - include - (default= ``False``) if True, the target expression is also parsed (the skipped text and target expression are returned as a 2-element list). - ignore - (default= ``None``) used to define grammars (typically quoted strings and comments) that might contain false matches to the target expression - failOn - (default= ``None``) define expressions that are not allowed to be included in the skipped test; if found before the target expression is found, the SkipTo is not a match Example:: report = ''' Outstanding Issues Report - 1 Jan 2000 # | Severity | Description | Days Open -----+----------+-------------------------------------------+----------- 101 | Critical | Intermittent system crash | 6 94 | Cosmetic | Spelling error on Login ('log|n') | 14 79 | Minor | System slow when running too many reports | 47 ''' integer = Word(nums) SEP = Suppress('|') # use SkipTo to simply match everything up until the next SEP # - ignore quoted strings, so that a '|' character inside a quoted string does not match # - parse action will call token.strip() for each matched token, i.e., the description body string_data = SkipTo(SEP, ignore=quotedString) string_data.setParseAction(tokenMap(str.strip)) ticket_expr = (integer("issue_num") + SEP + string_data("sev") + SEP + string_data("desc") + SEP + integer("days_open")) for tkt in ticket_expr.searchString(report): print tkt.dump() prints:: ['101', 'Critical', 'Intermittent system crash', '6'] - days_open: 6 - desc: Intermittent system crash - issue_num: 101 - sev: Critical ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14'] - days_open: 14 - desc: Spelling error on Login ('log|n') - issue_num: 94 - sev: Cosmetic ['79', 'Minor', 'System slow when running too many reports', '47'] - days_open: 47 - desc: System slow when running too many reports - issue_num: 79 - sev: Minor """ def __init__(self, other, include=False, ignore=None, failOn=None): super(SkipTo, self).__init__(other) self.ignoreExpr = ignore self.mayReturnEmpty = True self.mayIndexError = False self.includeMatch = include self.saveAsList = False if isinstance(failOn, basestring): self.failOn = self._literalStringClass(failOn) else: self.failOn = failOn self.errmsg = "No match found for " + _ustr(self.expr) def parseImpl(self, instring, loc, doActions=True): startloc = loc instrlen = len(instring) expr = self.expr expr_parse = self.expr._parse self_failOn_canParseNext = self.failOn.canParseNext if self.failOn is not None else None self_ignoreExpr_tryParse = self.ignoreExpr.tryParse if self.ignoreExpr is not None else None tmploc = loc while tmploc <= instrlen: if self_failOn_canParseNext is not None: # break if failOn expression matches if self_failOn_canParseNext(instring, tmploc): break if self_ignoreExpr_tryParse is not None: # advance past ignore expressions while 1: try: tmploc = self_ignoreExpr_tryParse(instring, tmploc) except ParseBaseException: break try: expr_parse(instring, tmploc, doActions=False, callPreParse=False) except (ParseException, IndexError): # no match, advance loc in string tmploc += 1 else: # matched skipto expr, done break else: # ran off the end of the input string without matching skipto expr, fail raise ParseException(instring, loc, self.errmsg, self) # build up return values loc = tmploc skiptext = instring[startloc:loc] skipresult = ParseResults(skiptext) if self.includeMatch: loc, mat = expr_parse(instring, loc, doActions, callPreParse=False) skipresult += mat return loc, skipresult class Forward(ParseElementEnhance): """Forward declaration of an expression to be defined later - used for recursive grammars, such as algebraic infix notation. When the expression is known, it is assigned to the ``Forward`` variable using the '<<' operator. Note: take care when assigning to ``Forward`` not to overlook precedence of operators. Specifically, '|' has a lower precedence than '<<', so that:: fwdExpr << a | b | c will actually be evaluated as:: (fwdExpr << a) | b | c thereby leaving b and c out as parseable alternatives. It is recommended that you explicitly group the values inserted into the ``Forward``:: fwdExpr << (a | b | c) Converting to use the '<<=' operator instead will avoid this problem. See :class:`ParseResults.pprint` for an example of a recursive parser created using ``Forward``. """ def __init__(self, other=None): super(Forward, self).__init__(other, savelist=False) def __lshift__(self, other): if isinstance(other, basestring): other = self._literalStringClass(other) self.expr = other self.strRepr = None self.mayIndexError = self.expr.mayIndexError self.mayReturnEmpty = self.expr.mayReturnEmpty self.setWhitespaceChars(self.expr.whiteChars) self.skipWhitespace = self.expr.skipWhitespace self.saveAsList = self.expr.saveAsList self.ignoreExprs.extend(self.expr.ignoreExprs) return self def __ilshift__(self, other): return self << other def leaveWhitespace(self): self.skipWhitespace = False return self def streamline(self): if not self.streamlined: self.streamlined = True if self.expr is not None: self.expr.streamline() return self def validate(self, validateTrace=None): if validateTrace is None: validateTrace = [] if self not in validateTrace: tmp = validateTrace[:] + [self] if self.expr is not None: self.expr.validate(tmp) self.checkRecursion([]) def __str__(self): if hasattr(self, "name"): return self.name if self.strRepr is not None: return self.strRepr # Avoid infinite recursion by setting a temporary strRepr self.strRepr = ": ..." # Use the string representation of main expression. retString = '...' try: if self.expr is not None: retString = _ustr(self.expr)[:1000] else: retString = "None" finally: self.strRepr = self.__class__.__name__ + ": " + retString return self.strRepr def copy(self): if self.expr is not None: return super(Forward, self).copy() else: ret = Forward() ret <<= self return ret def _setResultsName(self, name, listAllMatches=False): if __diag__.warn_name_set_on_empty_Forward: if self.expr is None: warnings.warn("{0}: setting results name {0!r} on {1} expression " "that has no contained expression".format("warn_name_set_on_empty_Forward", name, type(self).__name__), stacklevel=3) return super(Forward, self)._setResultsName(name, listAllMatches) class TokenConverter(ParseElementEnhance): """ Abstract subclass of :class:`ParseExpression`, for converting parsed results. """ def __init__(self, expr, savelist=False): super(TokenConverter, self).__init__(expr) # , savelist) self.saveAsList = False class Combine(TokenConverter): """Converter to concatenate all matching tokens to a single string. By default, the matching patterns must also be contiguous in the input string; this can be disabled by specifying ``'adjacent=False'`` in the constructor. Example:: real = Word(nums) + '.' + Word(nums) print(real.parseString('3.1416')) # -> ['3', '.', '1416'] # will also erroneously match the following print(real.parseString('3. 1416')) # -> ['3', '.', '1416'] real = Combine(Word(nums) + '.' + Word(nums)) print(real.parseString('3.1416')) # -> ['3.1416'] # no match when there are internal spaces print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...) """ def __init__(self, expr, joinString="", adjacent=True): super(Combine, self).__init__(expr) # suppress whitespace-stripping in contained parse expressions, but re-enable it on the Combine itself if adjacent: self.leaveWhitespace() self.adjacent = adjacent self.skipWhitespace = True self.joinString = joinString self.callPreparse = True def ignore(self, other): if self.adjacent: ParserElement.ignore(self, other) else: super(Combine, self).ignore(other) return self def postParse(self, instring, loc, tokenlist): retToks = tokenlist.copy() del retToks[:] retToks += ParseResults(["".join(tokenlist._asStringList(self.joinString))], modal=self.modalResults) if self.resultsName and retToks.haskeys(): return [retToks] else: return retToks class Group(TokenConverter): """Converter to return the matched tokens as a list - useful for returning tokens of :class:`ZeroOrMore` and :class:`OneOrMore` expressions. Example:: ident = Word(alphas) num = Word(nums) term = ident | num func = ident + Optional(delimitedList(term)) print(func.parseString("fn a, b, 100")) # -> ['fn', 'a', 'b', '100'] func = ident + Group(Optional(delimitedList(term))) print(func.parseString("fn a, b, 100")) # -> ['fn', ['a', 'b', '100']] """ def __init__(self, expr): super(Group, self).__init__(expr) self.saveAsList = True def postParse(self, instring, loc, tokenlist): return [tokenlist] class Dict(TokenConverter): """Converter to return a repetitive expression as a list, but also as a dictionary. Each element can also be referenced using the first token in the expression as its key. Useful for tabular report scraping when the first column can be used as a item key. Example:: data_word = Word(alphas) label = data_word + FollowedBy(':') attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join)) text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) # print attributes as plain groups print(OneOrMore(attr_expr).parseString(text).dump()) # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names result = Dict(OneOrMore(Group(attr_expr))).parseString(text) print(result.dump()) # access named fields as dict entries, or output as dict print(result['shape']) print(result.asDict()) prints:: ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap'] [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'} See more examples at :class:`ParseResults` of accessing fields by results name. """ def __init__(self, expr): super(Dict, self).__init__(expr) self.saveAsList = True def postParse(self, instring, loc, tokenlist): for i, tok in enumerate(tokenlist): if len(tok) == 0: continue ikey = tok[0] if isinstance(ikey, int): ikey = _ustr(tok[0]).strip() if len(tok) == 1: tokenlist[ikey] = _ParseResultsWithOffset("", i) elif len(tok) == 2 and not isinstance(tok[1], ParseResults): tokenlist[ikey] = _ParseResultsWithOffset(tok[1], i) else: dictvalue = tok.copy() # ParseResults(i) del dictvalue[0] if len(dictvalue) != 1 or (isinstance(dictvalue, ParseResults) and dictvalue.haskeys()): tokenlist[ikey] = _ParseResultsWithOffset(dictvalue, i) else: tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0], i) if self.resultsName: return [tokenlist] else: return tokenlist class Suppress(TokenConverter): """Converter for ignoring the results of a parsed expression. Example:: source = "a, b, c,d" wd = Word(alphas) wd_list1 = wd + ZeroOrMore(',' + wd) print(wd_list1.parseString(source)) # often, delimiters that are useful during parsing are just in the # way afterward - use Suppress to keep them out of the parsed output wd_list2 = wd + ZeroOrMore(Suppress(',') + wd) print(wd_list2.parseString(source)) prints:: ['a', ',', 'b', ',', 'c', ',', 'd'] ['a', 'b', 'c', 'd'] (See also :class:`delimitedList`.) """ def postParse(self, instring, loc, tokenlist): return [] def suppress(self): return self class OnlyOnce(object): """Wrapper for parse actions, to ensure they are only called once. """ def __init__(self, methodCall): self.callable = _trim_arity(methodCall) self.called = False def __call__(self, s, l, t): if not self.called: results = self.callable(s, l, t) self.called = True return results raise ParseException(s, l, "") def reset(self): self.called = False def traceParseAction(f): """Decorator for debugging parse actions. When the parse action is called, this decorator will print ``">> entering method-name(line:, , )"``. When the parse action completes, the decorator will print ``"<<"`` followed by the returned value, or any exception that the parse action raised. Example:: wd = Word(alphas) @traceParseAction def remove_duplicate_chars(tokens): return ''.join(sorted(set(''.join(tokens)))) wds = OneOrMore(wd).setParseAction(remove_duplicate_chars) print(wds.parseString("slkdjs sld sldd sdlf sdljf")) prints:: >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {})) < 3: thisFunc = paArgs[0].__class__.__name__ + '.' + thisFunc sys.stderr.write(">>entering %s(line: '%s', %d, %r)\n" % (thisFunc, line(l, s), l, t)) try: ret = f(*paArgs) except Exception as exc: sys.stderr.write("< ['aa', 'bb', 'cc'] delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE'] """ dlName = _ustr(expr) + " [" + _ustr(delim) + " " + _ustr(expr) + "]..." if combine: return Combine(expr + ZeroOrMore(delim + expr)).setName(dlName) else: return (expr + ZeroOrMore(Suppress(delim) + expr)).setName(dlName) def countedArray(expr, intExpr=None): """Helper to define a counted list of expressions. This helper defines a pattern of the form:: integer expr expr expr... where the leading integer tells how many expr expressions follow. The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed. If ``intExpr`` is specified, it should be a pyparsing expression that produces an integer value. Example:: countedArray(Word(alphas)).parseString('2 ab cd ef') # -> ['ab', 'cd'] # in this parser, the leading integer value is given in binary, # '10' indicating that 2 values are in the array binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2)) countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef') # -> ['ab', 'cd'] """ arrayExpr = Forward() def countFieldParseAction(s, l, t): n = t[0] arrayExpr << (n and Group(And([expr] * n)) or Group(empty)) return [] if intExpr is None: intExpr = Word(nums).setParseAction(lambda t: int(t[0])) else: intExpr = intExpr.copy() intExpr.setName("arrayLen") intExpr.addParseAction(countFieldParseAction, callDuringTry=True) return (intExpr + arrayExpr).setName('(len) ' + _ustr(expr) + '...') def _flatten(L): ret = [] for i in L: if isinstance(i, list): ret.extend(_flatten(i)) else: ret.append(i) return ret def matchPreviousLiteral(expr): """Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousLiteral(first) matchExpr = first + ":" + second will match ``"1:1"``, but not ``"1:2"``. Because this matches a previous literal, will also match the leading ``"1:1"`` in ``"1:10"``. If this is not desired, use :class:`matchPreviousExpr`. Do *not* use with packrat parsing enabled. """ rep = Forward() def copyTokenToRepeater(s, l, t): if t: if len(t) == 1: rep << t[0] else: # flatten t tokens tflat = _flatten(t.asList()) rep << And(Literal(tt) for tt in tflat) else: rep << Empty() expr.addParseAction(copyTokenToRepeater, callDuringTry=True) rep.setName('(prev) ' + _ustr(expr)) return rep def matchPreviousExpr(expr): """Helper to define an expression that is indirectly defined from the tokens matched in a previous expression, that is, it looks for a 'repeat' of a previous expression. For example:: first = Word(nums) second = matchPreviousExpr(first) matchExpr = first + ":" + second will match ``"1:1"``, but not ``"1:2"``. Because this matches by expressions, will *not* match the leading ``"1:1"`` in ``"1:10"``; the expressions are evaluated first, and then compared, so ``"1"`` is compared with ``"10"``. Do *not* use with packrat parsing enabled. """ rep = Forward() e2 = expr.copy() rep <<= e2 def copyTokenToRepeater(s, l, t): matchTokens = _flatten(t.asList()) def mustMatchTheseTokens(s, l, t): theseTokens = _flatten(t.asList()) if theseTokens != matchTokens: raise ParseException('', 0, '') rep.setParseAction(mustMatchTheseTokens, callDuringTry=True) expr.addParseAction(copyTokenToRepeater, callDuringTry=True) rep.setName('(prev) ' + _ustr(expr)) return rep def _escapeRegexRangeChars(s): # ~ escape these chars: ^-] for c in r"\^-]": s = s.replace(c, _bslash + c) s = s.replace("\n", r"\n") s = s.replace("\t", r"\t") return _ustr(s) def oneOf(strs, caseless=False, useRegex=True, asKeyword=False): """Helper to quickly define a set of alternative Literals, and makes sure to do longest-first testing when there is a conflict, regardless of the input order, but returns a :class:`MatchFirst` for best performance. Parameters: - strs - a string of space-delimited literals, or a collection of string literals - caseless - (default= ``False``) - treat all literals as caseless - useRegex - (default= ``True``) - as an optimization, will generate a Regex object; otherwise, will generate a :class:`MatchFirst` object (if ``caseless=True`` or ``asKeyword=True``, or if creating a :class:`Regex` raises an exception) - asKeyword - (default=``False``) - enforce Keyword-style matching on the generated expressions Example:: comp_oper = oneOf("< = > <= >= !=") var = Word(alphas) number = Word(nums) term = var | number comparison_expr = term + comp_oper + term print(comparison_expr.searchString("B = 12 AA=23 B<=AA AA>12")) prints:: [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']] """ if isinstance(caseless, basestring): warnings.warn("More than one string argument passed to oneOf, pass " "choices as a list or space-delimited string", stacklevel=2) if caseless: isequal = (lambda a, b: a.upper() == b.upper()) masks = (lambda a, b: b.upper().startswith(a.upper())) parseElementClass = CaselessKeyword if asKeyword else CaselessLiteral else: isequal = (lambda a, b: a == b) masks = (lambda a, b: b.startswith(a)) parseElementClass = Keyword if asKeyword else Literal symbols = [] if isinstance(strs, basestring): symbols = strs.split() elif isinstance(strs, Iterable): symbols = list(strs) else: warnings.warn("Invalid argument to oneOf, expected string or iterable", SyntaxWarning, stacklevel=2) if not symbols: return NoMatch() if not asKeyword: # if not producing keywords, need to reorder to take care to avoid masking # longer choices with shorter ones i = 0 while i < len(symbols) - 1: cur = symbols[i] for j, other in enumerate(symbols[i + 1:]): if isequal(other, cur): del symbols[i + j + 1] break elif masks(cur, other): del symbols[i + j + 1] symbols.insert(i, other) break else: i += 1 if not (caseless or asKeyword) and useRegex: # ~ print (strs, "->", "|".join([_escapeRegexChars(sym) for sym in symbols])) try: if len(symbols) == len("".join(symbols)): return Regex("[%s]" % "".join(_escapeRegexRangeChars(sym) for sym in symbols)).setName(' | '.join(symbols)) else: return Regex("|".join(re.escape(sym) for sym in symbols)).setName(' | '.join(symbols)) except Exception: warnings.warn("Exception creating Regex for oneOf, building MatchFirst", SyntaxWarning, stacklevel=2) # last resort, just use MatchFirst return MatchFirst(parseElementClass(sym) for sym in symbols).setName(' | '.join(symbols)) def dictOf(key, value): """Helper to easily and clearly define a dictionary by specifying the respective patterns for the key and value. Takes care of defining the :class:`Dict`, :class:`ZeroOrMore`, and :class:`Group` tokens in the proper order. The key pattern can include delimiting markers or punctuation, as long as they are suppressed, thereby leaving the significant key text. The value pattern can include named results, so that the :class:`Dict` results can include named token fields. Example:: text = "shape: SQUARE posn: upper left color: light blue texture: burlap" attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)) print(OneOrMore(attr_expr).parseString(text).dump()) attr_label = label attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join) # similar to Dict, but simpler call format result = dictOf(attr_label, attr_value).parseString(text) print(result.dump()) print(result['shape']) print(result.shape) # object attribute access works too print(result.asDict()) prints:: [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']] - color: light blue - posn: upper left - shape: SQUARE - texture: burlap SQUARE SQUARE {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'} """ return Dict(OneOrMore(Group(key + value))) def originalTextFor(expr, asString=True): """Helper to return the original, untokenized text for a given expression. Useful to restore the parsed fields of an HTML start tag into the raw tag text itself, or to revert separate tokens with intervening whitespace back to the original matching input text. By default, returns astring containing the original parsed text. If the optional ``asString`` argument is passed as ``False``, then the return value is a :class:`ParseResults` containing any results names that were originally matched, and a single token containing the original matched text from the input string. So if the expression passed to :class:`originalTextFor` contains expressions with defined results names, you must set ``asString`` to ``False`` if you want to preserve those results name values. Example:: src = "this is test bold text normal text " for tag in ("b", "i"): opener, closer = makeHTMLTags(tag) patt = originalTextFor(opener + SkipTo(closer) + closer) print(patt.searchString(src)[0]) prints:: [' bold text '] ['text'] """ locMarker = Empty().setParseAction(lambda s, loc, t: loc) endlocMarker = locMarker.copy() endlocMarker.callPreparse = False matchExpr = locMarker("_original_start") + expr + endlocMarker("_original_end") if asString: extractText = lambda s, l, t: s[t._original_start: t._original_end] else: def extractText(s, l, t): t[:] = [s[t.pop('_original_start'):t.pop('_original_end')]] matchExpr.setParseAction(extractText) matchExpr.ignoreExprs = expr.ignoreExprs return matchExpr def ungroup(expr): """Helper to undo pyparsing's default grouping of And expressions, even if all but one are non-empty. """ return TokenConverter(expr).addParseAction(lambda t: t[0]) def locatedExpr(expr): """Helper to decorate a returned token with its starting and ending locations in the input string. This helper adds the following results names: - locn_start = location where matched expression begins - locn_end = location where matched expression ends - value = the actual parsed results Be careful if the input text contains ```` characters, you may want to call :class:`ParserElement.parseWithTabs` Example:: wd = Word(alphas) for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"): print(match) prints:: [[0, 'ljsdf', 5]] [[8, 'lksdjjf', 15]] [[18, 'lkkjj', 23]] """ locator = Empty().setParseAction(lambda s, l, t: l) return Group(locator("locn_start") + expr("value") + locator.copy().leaveWhitespace()("locn_end")) # convenience constants for positional expressions empty = Empty().setName("empty") lineStart = LineStart().setName("lineStart") lineEnd = LineEnd().setName("lineEnd") stringStart = StringStart().setName("stringStart") stringEnd = StringEnd().setName("stringEnd") _escapedPunc = Word(_bslash, r"\[]-*.$+^?()~ ", exact=2).setParseAction(lambda s, l, t: t[0][1]) _escapedHexChar = Regex(r"\\0?[xX][0-9a-fA-F]+").setParseAction(lambda s, l, t: unichr(int(t[0].lstrip(r'\0x'), 16))) _escapedOctChar = Regex(r"\\0[0-7]+").setParseAction(lambda s, l, t: unichr(int(t[0][1:], 8))) _singleChar = _escapedPunc | _escapedHexChar | _escapedOctChar | CharsNotIn(r'\]', exact=1) _charRange = Group(_singleChar + Suppress("-") + _singleChar) _reBracketExpr = Literal("[") + Optional("^").setResultsName("negate") + Group(OneOrMore(_charRange | _singleChar)).setResultsName("body") + "]" def srange(s): r"""Helper to easily define string ranges for use in Word construction. Borrows syntax from regexp '[]' string range definitions:: srange("[0-9]") -> "0123456789" srange("[a-z]") -> "abcdefghijklmnopqrstuvwxyz" srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_" The input string must be enclosed in []'s, and the returned string is the expanded character set joined into a single string. The values enclosed in the []'s may be: - a single character - an escaped character with a leading backslash (such as ``\-`` or ``\]``) - an escaped hex character with a leading ``'\x'`` (``\x21``, which is a ``'!'`` character) (``\0x##`` is also supported for backwards compatibility) - an escaped octal character with a leading ``'\0'`` (``\041``, which is a ``'!'`` character) - a range of any of the above, separated by a dash (``'a-z'``, etc.) - any combination of the above (``'aeiouy'``, ``'a-zA-Z0-9_$'``, etc.) """ _expanded = lambda p: p if not isinstance(p, ParseResults) else ''.join(unichr(c) for c in range(ord(p[0]), ord(p[1]) + 1)) try: return "".join(_expanded(part) for part in _reBracketExpr.parseString(s).body) except Exception: return "" def matchOnlyAtCol(n): """Helper method for defining parse actions that require matching at a specific column in the input text. """ def verifyCol(strg, locn, toks): if col(locn, strg) != n: raise ParseException(strg, locn, "matched token not at column %d" % n) return verifyCol def replaceWith(replStr): """Helper method for common parse actions that simply return a literal value. Especially useful when used with :class:`transformString` (). Example:: num = Word(nums).setParseAction(lambda toks: int(toks[0])) na = oneOf("N/A NA").setParseAction(replaceWith(math.nan)) term = na | num OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234] """ return lambda s, l, t: [replStr] def removeQuotes(s, l, t): """Helper parse action for removing quotation marks from parsed quoted strings. Example:: # by default, quotation marks are included in parsed results quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"] # use removeQuotes to strip quotation marks from parsed results quotedString.setParseAction(removeQuotes) quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"] """ return t[0][1:-1] def tokenMap(func, *args): """Helper to define a parse action by mapping a function to all elements of a ParseResults list. If any additional args are passed, they are forwarded to the given function as additional arguments after the token, as in ``hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))``, which will convert the parsed data to an integer using base 16. Example (compare the last to example in :class:`ParserElement.transformString`:: hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16)) hex_ints.runTests(''' 00 11 22 aa FF 0a 0d 1a ''') upperword = Word(alphas).setParseAction(tokenMap(str.upper)) OneOrMore(upperword).runTests(''' my kingdom for a horse ''') wd = Word(alphas).setParseAction(tokenMap(str.title)) OneOrMore(wd).setParseAction(' '.join).runTests(''' now is the winter of our discontent made glorious summer by this sun of york ''') prints:: 00 11 22 aa FF 0a 0d 1a [0, 17, 34, 170, 255, 10, 13, 26] my kingdom for a horse ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE'] now is the winter of our discontent made glorious summer by this sun of york ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York'] """ def pa(s, l, t): return [func(tokn, *args) for tokn in t] try: func_name = getattr(func, '__name__', getattr(func, '__class__').__name__) except Exception: func_name = str(func) pa.__name__ = func_name return pa upcaseTokens = tokenMap(lambda t: _ustr(t).upper()) """(Deprecated) Helper parse action to convert tokens to upper case. Deprecated in favor of :class:`pyparsing_common.upcaseTokens`""" downcaseTokens = tokenMap(lambda t: _ustr(t).lower()) """(Deprecated) Helper parse action to convert tokens to lower case. Deprecated in favor of :class:`pyparsing_common.downcaseTokens`""" def _makeTags(tagStr, xml, suppress_LT=Suppress("<"), suppress_GT=Suppress(">")): """Internal helper to construct opening and closing tag expressions, given a tag name""" if isinstance(tagStr, basestring): resname = tagStr tagStr = Keyword(tagStr, caseless=not xml) else: resname = tagStr.name tagAttrName = Word(alphas, alphanums + "_-:") if xml: tagAttrValue = dblQuotedString.copy().setParseAction(removeQuotes) openTag = (suppress_LT + tagStr("tag") + Dict(ZeroOrMore(Group(tagAttrName + Suppress("=") + tagAttrValue))) + Optional("/", default=[False])("empty").setParseAction(lambda s, l, t: t[0] == '/') + suppress_GT) else: tagAttrValue = quotedString.copy().setParseAction(removeQuotes) | Word(printables, excludeChars=">") openTag = (suppress_LT + tagStr("tag") + Dict(ZeroOrMore(Group(tagAttrName.setParseAction(downcaseTokens) + Optional(Suppress("=") + tagAttrValue)))) + Optional("/", default=[False])("empty").setParseAction(lambda s, l, t: t[0] == '/') + suppress_GT) closeTag = Combine(_L("", adjacent=False) openTag.setName("<%s>" % resname) # add start results name in parse action now that ungrouped names are not reported at two levels openTag.addParseAction(lambda t: t.__setitem__("start" + "".join(resname.replace(":", " ").title().split()), t.copy())) closeTag = closeTag("end" + "".join(resname.replace(":", " ").title().split())).setName("" % resname) openTag.tag = resname closeTag.tag = resname openTag.tag_body = SkipTo(closeTag()) return openTag, closeTag def makeHTMLTags(tagStr): """Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values. Example:: text = 'More info at the pyparsing wiki page' # makeHTMLTags returns pyparsing expressions for the opening and # closing tags as a 2-tuple a, a_end = makeHTMLTags("A") link_expr = a + SkipTo(a_end)("link_text") + a_end for link in link_expr.searchString(text): # attributes in the tag (like "href" shown here) are # also accessible as named results print(link.link_text, '->', link.href) prints:: pyparsing -> https://github.com/pyparsing/pyparsing/wiki """ return _makeTags(tagStr, False) def makeXMLTags(tagStr): """Helper to construct opening and closing tag expressions for XML, given a tag name. Matches tags only in the given upper/lower case. Example: similar to :class:`makeHTMLTags` """ return _makeTags(tagStr, True) def withAttribute(*args, **attrDict): """Helper to create a validating parse action to be used with start tags created with :class:`makeXMLTags` or :class:`makeHTMLTags`. Use ``withAttribute`` to qualify a starting tag with a required attribute value, to avoid false matches on common tags such as ```` or ``
``. Call ``withAttribute`` with a series of attribute names and values. Specify the list of filter attributes names and values as: - keyword arguments, as in ``(align="right")``, or - as an explicit dict with ``**`` operator, when an attribute name is also a Python reserved word, as in ``**{"class":"Customer", "align":"right"}`` - a list of name-value tuples, as in ``(("ns1:class", "Customer"), ("ns2:align", "right"))`` For attribute names with a namespace prefix, you must use the second form. Attribute names are matched insensitive to upper/lower case. If just testing for ``class`` (with or without a namespace), use :class:`withClass`. To verify that the attribute exists, but without specifying a value, pass ``withAttribute.ANY_VALUE`` as the value. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this has no type
''' div,div_end = makeHTMLTags("div") # only match div tag having a type attribute with value "grid" div_grid = div().setParseAction(withAttribute(type="grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) # construct a match with any div tag having a type attribute, regardless of the value div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 """ if args: attrs = args[:] else: attrs = attrDict.items() attrs = [(k, v) for k, v in attrs] def pa(s, l, tokens): for attrName, attrValue in attrs: if attrName not in tokens: raise ParseException(s, l, "no matching attribute " + attrName) if attrValue != withAttribute.ANY_VALUE and tokens[attrName] != attrValue: raise ParseException(s, l, "attribute '%s' has value '%s', must be '%s'" % (attrName, tokens[attrName], attrValue)) return pa withAttribute.ANY_VALUE = object() def withClass(classname, namespace=''): """Simplified version of :class:`withAttribute` when matching on a div class - made difficult because ``class`` is a reserved word in Python. Example:: html = '''
Some text
1 4 0 1 0
1,3 2,3 1,1
this <div> has no class
''' div,div_end = makeHTMLTags("div") div_grid = div().setParseAction(withClass("grid")) grid_expr = div_grid + SkipTo(div | div_end)("body") for grid_header in grid_expr.searchString(html): print(grid_header.body) div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE)) div_expr = div_any_type + SkipTo(div | div_end)("body") for div_header in div_expr.searchString(html): print(div_header.body) prints:: 1 4 0 1 0 1 4 0 1 0 1,3 2,3 1,1 """ classattr = "%s:class" % namespace if namespace else "class" return withAttribute(**{classattr: classname}) opAssoc = SimpleNamespace() opAssoc.LEFT = object() opAssoc.RIGHT = object() def infixNotation(baseExpr, opList, lpar=Suppress('('), rpar=Suppress(')')): """Helper method for constructing grammars of expressions made up of operators working in a precedence hierarchy. Operators may be unary or binary, left- or right-associative. Parse actions can also be attached to operator expressions. The generated parser will also recognize the use of parentheses to override operator precedences (see example below). Note: if you define a deep operator list, you may see performance issues when using infixNotation. See :class:`ParserElement.enablePackrat` for a mechanism to potentially improve your parser performance. Parameters: - baseExpr - expression representing the most basic element for the nested - opList - list of tuples, one for each operator precedence level in the expression grammar; each tuple is of the form ``(opExpr, numTerms, rightLeftAssoc, parseAction)``, where: - opExpr is the pyparsing expression for the operator; may also be a string, which will be converted to a Literal; if numTerms is 3, opExpr is a tuple of two expressions, for the two operators separating the 3 terms - numTerms is the number of terms for this operator (must be 1, 2, or 3) - rightLeftAssoc is the indicator whether the operator is right or left associative, using the pyparsing-defined constants ``opAssoc.RIGHT`` and ``opAssoc.LEFT``. - parseAction is the parse action to be associated with expressions matching this operator expression (the parse action tuple member may be omitted); if the parse action is passed a tuple or list of functions, this is equivalent to calling ``setParseAction(*fn)`` (:class:`ParserElement.setParseAction`) - lpar - expression for matching left-parentheses (default= ``Suppress('(')``) - rpar - expression for matching right-parentheses (default= ``Suppress(')')``) Example:: # simple example of four-function arithmetic with ints and # variable names integer = pyparsing_common.signed_integer varname = pyparsing_common.identifier arith_expr = infixNotation(integer | varname, [ ('-', 1, opAssoc.RIGHT), (oneOf('* /'), 2, opAssoc.LEFT), (oneOf('+ -'), 2, opAssoc.LEFT), ]) arith_expr.runTests(''' 5+3*6 (5+3)*6 -2--11 ''', fullDump=False) prints:: 5+3*6 [[5, '+', [3, '*', 6]]] (5+3)*6 [[[5, '+', 3], '*', 6]] -2--11 [[['-', 2], '-', ['-', 11]]] """ # captive version of FollowedBy that does not do parse actions or capture results names class _FB(FollowedBy): def parseImpl(self, instring, loc, doActions=True): self.expr.tryParse(instring, loc) return loc, [] ret = Forward() lastExpr = baseExpr | (lpar + ret + rpar) for i, operDef in enumerate(opList): opExpr, arity, rightLeftAssoc, pa = (operDef + (None, ))[:4] termName = "%s term" % opExpr if arity < 3 else "%s%s term" % opExpr if arity == 3: if opExpr is None or len(opExpr) != 2: raise ValueError( "if numterms=3, opExpr must be a tuple or list of two expressions") opExpr1, opExpr2 = opExpr thisExpr = Forward().setName(termName) if rightLeftAssoc == opAssoc.LEFT: if arity == 1: matchExpr = _FB(lastExpr + opExpr) + Group(lastExpr + OneOrMore(opExpr)) elif arity == 2: if opExpr is not None: matchExpr = _FB(lastExpr + opExpr + lastExpr) + Group(lastExpr + OneOrMore(opExpr + lastExpr)) else: matchExpr = _FB(lastExpr + lastExpr) + Group(lastExpr + OneOrMore(lastExpr)) elif arity == 3: matchExpr = (_FB(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr) + Group(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr)) else: raise ValueError("operator must be unary (1), binary (2), or ternary (3)") elif rightLeftAssoc == opAssoc.RIGHT: if arity == 1: # try to avoid LR with this extra test if not isinstance(opExpr, Optional): opExpr = Optional(opExpr) matchExpr = _FB(opExpr.expr + thisExpr) + Group(opExpr + thisExpr) elif arity == 2: if opExpr is not None: matchExpr = _FB(lastExpr + opExpr + thisExpr) + Group(lastExpr + OneOrMore(opExpr + thisExpr)) else: matchExpr = _FB(lastExpr + thisExpr) + Group(lastExpr + OneOrMore(thisExpr)) elif arity == 3: matchExpr = (_FB(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr) + Group(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr)) else: raise ValueError("operator must be unary (1), binary (2), or ternary (3)") else: raise ValueError("operator must indicate right or left associativity") if pa: if isinstance(pa, (tuple, list)): matchExpr.setParseAction(*pa) else: matchExpr.setParseAction(pa) thisExpr <<= (matchExpr.setName(termName) | lastExpr) lastExpr = thisExpr ret <<= lastExpr return ret operatorPrecedence = infixNotation """(Deprecated) Former name of :class:`infixNotation`, will be dropped in a future release.""" dblQuotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*') + '"').setName("string enclosed in double quotes") sglQuotedString = Combine(Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*") + "'").setName("string enclosed in single quotes") quotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*') + '"' | Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*") + "'").setName("quotedString using single or double quotes") unicodeString = Combine(_L('u') + quotedString.copy()).setName("unicode string literal") def nestedExpr(opener="(", closer=")", content=None, ignoreExpr=quotedString.copy()): """Helper method for defining nested lists enclosed in opening and closing delimiters ("(" and ")" are the default). Parameters: - opener - opening character for a nested list (default= ``"("``); can also be a pyparsing expression - closer - closing character for a nested list (default= ``")"``); can also be a pyparsing expression - content - expression for items within the nested lists (default= ``None``) - ignoreExpr - expression for ignoring opening and closing delimiters (default= :class:`quotedString`) If an expression is not provided for the content argument, the nested expression will capture all whitespace-delimited content between delimiters as a list of separate values. Use the ``ignoreExpr`` argument to define expressions that may contain opening or closing characters that should not be treated as opening or closing characters for nesting, such as quotedString or a comment expression. Specify multiple expressions using an :class:`Or` or :class:`MatchFirst`. The default is :class:`quotedString`, but if no expressions are to be ignored, then pass ``None`` for this argument. Example:: data_type = oneOf("void int short long char float double") decl_data_type = Combine(data_type + Optional(Word('*'))) ident = Word(alphas+'_', alphanums+'_') number = pyparsing_common.number arg = Group(decl_data_type + ident) LPAR, RPAR = map(Suppress, "()") code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment)) c_function = (decl_data_type("type") + ident("name") + LPAR + Optional(delimitedList(arg), [])("args") + RPAR + code_body("body")) c_function.ignore(cStyleComment) source_code = ''' int is_odd(int x) { return (x%2); } int dec_to_hex(char hchar) { if (hchar >= '0' && hchar <= '9') { return (ord(hchar)-ord('0')); } else { return (10+ord(hchar)-ord('A')); } } ''' for func in c_function.searchString(source_code): print("%(name)s (%(type)s) args: %(args)s" % func) prints:: is_odd (int) args: [['int', 'x']] dec_to_hex (int) args: [['char', 'hchar']] """ if opener == closer: raise ValueError("opening and closing strings cannot be the same") if content is None: if isinstance(opener, basestring) and isinstance(closer, basestring): if len(opener) == 1 and len(closer) == 1: if ignoreExpr is not None: content = (Combine(OneOrMore(~ignoreExpr + CharsNotIn(opener + closer + ParserElement.DEFAULT_WHITE_CHARS, exact=1) ) ).setParseAction(lambda t: t[0].strip())) else: content = (empty.copy() + CharsNotIn(opener + closer + ParserElement.DEFAULT_WHITE_CHARS ).setParseAction(lambda t: t[0].strip())) else: if ignoreExpr is not None: content = (Combine(OneOrMore(~ignoreExpr + ~Literal(opener) + ~Literal(closer) + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS, exact=1)) ).setParseAction(lambda t: t[0].strip())) else: content = (Combine(OneOrMore(~Literal(opener) + ~Literal(closer) + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS, exact=1)) ).setParseAction(lambda t: t[0].strip())) else: raise ValueError("opening and closing arguments must be strings if no content expression is given") ret = Forward() if ignoreExpr is not None: ret <<= Group(Suppress(opener) + ZeroOrMore(ignoreExpr | ret | content) + Suppress(closer)) else: ret <<= Group(Suppress(opener) + ZeroOrMore(ret | content) + Suppress(closer)) ret.setName('nested %s%s expression' % (opener, closer)) return ret def indentedBlock(blockStatementExpr, indentStack, indent=True): """Helper method for defining space-delimited indentation blocks, such as those used to define block statements in Python source code. Parameters: - blockStatementExpr - expression defining syntax of statement that is repeated within the indented block - indentStack - list created by caller to manage indentation stack (multiple statementWithIndentedBlock expressions within a single grammar should share a common indentStack) - indent - boolean indicating whether block must be indented beyond the current level; set to False for block of left-most statements (default= ``True``) A valid block must contain at least one ``blockStatement``. Example:: data = ''' def A(z): A1 B = 100 G = A2 A2 A3 B def BB(a,b,c): BB1 def BBA(): bba1 bba2 bba3 C D def spam(x,y): def eggs(z): pass ''' indentStack = [1] stmt = Forward() identifier = Word(alphas, alphanums) funcDecl = ("def" + identifier + Group("(" + Optional(delimitedList(identifier)) + ")") + ":") func_body = indentedBlock(stmt, indentStack) funcDef = Group(funcDecl + func_body) rvalue = Forward() funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")") rvalue << (funcCall | identifier | Word(nums)) assignment = Group(identifier + "=" + rvalue) stmt << (funcDef | assignment | identifier) module_body = OneOrMore(stmt) parseTree = module_body.parseString(data) parseTree.pprint() prints:: [['def', 'A', ['(', 'z', ')'], ':', [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]], 'B', ['def', 'BB', ['(', 'a', 'b', 'c', ')'], ':', [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]], 'C', 'D', ['def', 'spam', ['(', 'x', 'y', ')'], ':', [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] """ backup_stack = indentStack[:] def reset_stack(): indentStack[:] = backup_stack def checkPeerIndent(s, l, t): if l >= len(s): return curCol = col(l, s) if curCol != indentStack[-1]: if curCol > indentStack[-1]: raise ParseException(s, l, "illegal nesting") raise ParseException(s, l, "not a peer entry") def checkSubIndent(s, l, t): curCol = col(l, s) if curCol > indentStack[-1]: indentStack.append(curCol) else: raise ParseException(s, l, "not a subentry") def checkUnindent(s, l, t): if l >= len(s): return curCol = col(l, s) if not(indentStack and curCol in indentStack): raise ParseException(s, l, "not an unindent") if curCol < indentStack[-1]: indentStack.pop() NL = OneOrMore(LineEnd().setWhitespaceChars("\t ").suppress()) INDENT = (Empty() + Empty().setParseAction(checkSubIndent)).setName('INDENT') PEER = Empty().setParseAction(checkPeerIndent).setName('') UNDENT = Empty().setParseAction(checkUnindent).setName('UNINDENT') if indent: smExpr = Group(Optional(NL) + INDENT + OneOrMore(PEER + Group(blockStatementExpr) + Optional(NL)) + UNDENT) else: smExpr = Group(Optional(NL) + OneOrMore(PEER + Group(blockStatementExpr) + Optional(NL)) + UNDENT) smExpr.setFailAction(lambda a, b, c, d: reset_stack()) blockStatementExpr.ignore(_bslash + LineEnd()) return smExpr.setName('indented block') alphas8bit = srange(r"[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]") punc8bit = srange(r"[\0xa1-\0xbf\0xd7\0xf7]") anyOpenTag, anyCloseTag = makeHTMLTags(Word(alphas, alphanums + "_:").setName('any tag')) _htmlEntityMap = dict(zip("gt lt amp nbsp quot apos".split(), '><& "\'')) commonHTMLEntity = Regex('&(?P' + '|'.join(_htmlEntityMap.keys()) +");").setName("common HTML entity") def replaceHTMLEntity(t): """Helper parser action to replace common HTML entities with their special characters""" return _htmlEntityMap.get(t.entity) # it's easy to get these comment structures wrong - they're very common, so may as well make them available cStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/').setName("C style comment") "Comment of the form ``/* ... */``" htmlComment = Regex(r"").setName("HTML comment") "Comment of the form ````" restOfLine = Regex(r".*").leaveWhitespace().setName("rest of line") dblSlashComment = Regex(r"//(?:\\\n|[^\n])*").setName("// comment") "Comment of the form ``// ... (to end of line)``" cppStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/' | dblSlashComment).setName("C++ style comment") "Comment of either form :class:`cStyleComment` or :class:`dblSlashComment`" javaStyleComment = cppStyleComment "Same as :class:`cppStyleComment`" pythonStyleComment = Regex(r"#.*").setName("Python style comment") "Comment of the form ``# ... (to end of line)``" _commasepitem = Combine(OneOrMore(Word(printables, excludeChars=',') + Optional(Word(" \t") + ~Literal(",") + ~LineEnd()))).streamline().setName("commaItem") commaSeparatedList = delimitedList(Optional(quotedString.copy() | _commasepitem, default="")).setName("commaSeparatedList") """(Deprecated) Predefined expression of 1 or more printable words or quoted strings, separated by commas. This expression is deprecated in favor of :class:`pyparsing_common.comma_separated_list`. """ # some other useful expressions - using lower-case class name since we are really using this as a namespace class pyparsing_common: """Here are some common low-level expressions that may be useful in jump-starting parser development: - numeric forms (:class:`integers`, :class:`reals`, :class:`scientific notation`) - common :class:`programming identifiers` - network addresses (:class:`MAC`, :class:`IPv4`, :class:`IPv6`) - ISO8601 :class:`dates` and :class:`datetime` - :class:`UUID` - :class:`comma-separated list` Parse actions: - :class:`convertToInteger` - :class:`convertToFloat` - :class:`convertToDate` - :class:`convertToDatetime` - :class:`stripHTMLTags` - :class:`upcaseTokens` - :class:`downcaseTokens` Example:: pyparsing_common.number.runTests(''' # any int or real number, returned as the appropriate type 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.fnumber.runTests(''' # any int or real number, returned as float 100 -100 +100 3.14159 6.02e23 1e-12 ''') pyparsing_common.hex_integer.runTests(''' # hex numbers 100 FF ''') pyparsing_common.fraction.runTests(''' # fractions 1/2 -3/4 ''') pyparsing_common.mixed_integer.runTests(''' # mixed fractions 1 1/2 -3/4 1-3/4 ''') import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(''' # uuid 12345678-1234-5678-1234-567812345678 ''') prints:: # any int or real number, returned as the appropriate type 100 [100] -100 [-100] +100 [100] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # any int or real number, returned as float 100 [100.0] -100 [-100.0] +100 [100.0] 3.14159 [3.14159] 6.02e23 [6.02e+23] 1e-12 [1e-12] # hex numbers 100 [256] FF [255] # fractions 1/2 [0.5] -3/4 [-0.75] # mixed fractions 1 [1] 1/2 [0.5] -3/4 [-0.75] 1-3/4 [1.75] # uuid 12345678-1234-5678-1234-567812345678 [UUID('12345678-1234-5678-1234-567812345678')] """ convertToInteger = tokenMap(int) """ Parse action for converting parsed integers to Python int """ convertToFloat = tokenMap(float) """ Parse action for converting parsed numbers to Python float """ integer = Word(nums).setName("integer").setParseAction(convertToInteger) """expression that parses an unsigned integer, returns an int""" hex_integer = Word(hexnums).setName("hex integer").setParseAction(tokenMap(int, 16)) """expression that parses a hexadecimal integer, returns an int""" signed_integer = Regex(r'[+-]?\d+').setName("signed integer").setParseAction(convertToInteger) """expression that parses an integer with optional leading sign, returns an int""" fraction = (signed_integer().setParseAction(convertToFloat) + '/' + signed_integer().setParseAction(convertToFloat)).setName("fraction") """fractional expression of an integer divided by an integer, returns a float""" fraction.addParseAction(lambda t: t[0]/t[-1]) mixed_integer = (fraction | signed_integer + Optional(Optional('-').suppress() + fraction)).setName("fraction or mixed integer-fraction") """mixed integer of the form 'integer - fraction', with optional leading integer, returns float""" mixed_integer.addParseAction(sum) real = Regex(r'[+-]?(:?\d+\.\d*|\.\d+)').setName("real number").setParseAction(convertToFloat) """expression that parses a floating point number and returns a float""" sci_real = Regex(r'[+-]?(:?\d+(:?[eE][+-]?\d+)|(:?\d+\.\d*|\.\d+)(:?[eE][+-]?\d+)?)').setName("real number with scientific notation").setParseAction(convertToFloat) """expression that parses a floating point number with optional scientific notation and returns a float""" # streamlining this expression makes the docs nicer-looking number = (sci_real | real | signed_integer).streamline() """any numeric expression, returns the corresponding Python type""" fnumber = Regex(r'[+-]?\d+\.?\d*([eE][+-]?\d+)?').setName("fnumber").setParseAction(convertToFloat) """any int or real number, returned as float""" identifier = Word(alphas + '_', alphanums + '_').setName("identifier") """typical code identifier (leading alpha or '_', followed by 0 or more alphas, nums, or '_')""" ipv4_address = Regex(r'(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}').setName("IPv4 address") "IPv4 address (``0.0.0.0 - 255.255.255.255``)" _ipv6_part = Regex(r'[0-9a-fA-F]{1,4}').setName("hex_integer") _full_ipv6_address = (_ipv6_part + (':' + _ipv6_part) * 7).setName("full IPv6 address") _short_ipv6_address = (Optional(_ipv6_part + (':' + _ipv6_part) * (0, 6)) + "::" + Optional(_ipv6_part + (':' + _ipv6_part) * (0, 6)) ).setName("short IPv6 address") _short_ipv6_address.addCondition(lambda t: sum(1 for tt in t if pyparsing_common._ipv6_part.matches(tt)) < 8) _mixed_ipv6_address = ("::ffff:" + ipv4_address).setName("mixed IPv6 address") ipv6_address = Combine((_full_ipv6_address | _mixed_ipv6_address | _short_ipv6_address).setName("IPv6 address")).setName("IPv6 address") "IPv6 address (long, short, or mixed form)" mac_address = Regex(r'[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}').setName("MAC address") "MAC address xx:xx:xx:xx:xx (may also have '-' or '.' delimiters)" @staticmethod def convertToDate(fmt="%Y-%m-%d"): """ Helper to create a parse action for converting parsed date string to Python datetime.date Params - - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%d"``) Example:: date_expr = pyparsing_common.iso8601_date.copy() date_expr.setParseAction(pyparsing_common.convertToDate()) print(date_expr.parseString("1999-12-31")) prints:: [datetime.date(1999, 12, 31)] """ def cvt_fn(s, l, t): try: return datetime.strptime(t[0], fmt).date() except ValueError as ve: raise ParseException(s, l, str(ve)) return cvt_fn @staticmethod def convertToDatetime(fmt="%Y-%m-%dT%H:%M:%S.%f"): """Helper to create a parse action for converting parsed datetime string to Python datetime.datetime Params - - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%dT%H:%M:%S.%f"``) Example:: dt_expr = pyparsing_common.iso8601_datetime.copy() dt_expr.setParseAction(pyparsing_common.convertToDatetime()) print(dt_expr.parseString("1999-12-31T23:59:59.999")) prints:: [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)] """ def cvt_fn(s, l, t): try: return datetime.strptime(t[0], fmt) except ValueError as ve: raise ParseException(s, l, str(ve)) return cvt_fn iso8601_date = Regex(r'(?P\d{4})(?:-(?P\d\d)(?:-(?P\d\d))?)?').setName("ISO8601 date") "ISO8601 date (``yyyy-mm-dd``)" iso8601_datetime = Regex(r'(?P\d{4})-(?P\d\d)-(?P\d\d)[T ](?P\d\d):(?P\d\d)(:(?P\d\d(\.\d*)?)?)?(?PZ|[+-]\d\d:?\d\d)?').setName("ISO8601 datetime") "ISO8601 datetime (``yyyy-mm-ddThh:mm:ss.s(Z|+-00:00)``) - trailing seconds, milliseconds, and timezone optional; accepts separating ``'T'`` or ``' '``" uuid = Regex(r'[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}').setName("UUID") "UUID (``xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx``)" _html_stripper = anyOpenTag.suppress() | anyCloseTag.suppress() @staticmethod def stripHTMLTags(s, l, tokens): """Parse action to remove HTML tags from web page HTML source Example:: # strip HTML links from normal text text = 'More info at the
pyparsing wiki page' td, td_end = makeHTMLTags("TD") table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end print(table_text.parseString(text).body) Prints:: More info at the pyparsing wiki page """ return pyparsing_common._html_stripper.transformString(tokens[0]) _commasepitem = Combine(OneOrMore(~Literal(",") + ~LineEnd() + Word(printables, excludeChars=',') + Optional(White(" \t")))).streamline().setName("commaItem") comma_separated_list = delimitedList(Optional(quotedString.copy() | _commasepitem, default='') ).setName("comma separated list") """Predefined expression of 1 or more printable words or quoted strings, separated by commas.""" upcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).upper())) """Parse action to convert tokens to upper case.""" downcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).lower())) """Parse action to convert tokens to lower case.""" class _lazyclassproperty(object): def __init__(self, fn): self.fn = fn self.__doc__ = fn.__doc__ self.__name__ = fn.__name__ def __get__(self, obj, cls): if cls is None: cls = type(obj) if not hasattr(cls, '_intern') or any(cls._intern is getattr(superclass, '_intern', []) for superclass in cls.__mro__[1:]): cls._intern = {} attrname = self.fn.__name__ if attrname not in cls._intern: cls._intern[attrname] = self.fn(cls) return cls._intern[attrname] class unicode_set(object): """ A set of Unicode characters, for language-specific strings for ``alphas``, ``nums``, ``alphanums``, and ``printables``. A unicode_set is defined by a list of ranges in the Unicode character set, in a class attribute ``_ranges``, such as:: _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),] A unicode set can also be defined using multiple inheritance of other unicode sets:: class CJK(Chinese, Japanese, Korean): pass """ _ranges = [] @classmethod def _get_chars_for_ranges(cls): ret = [] for cc in cls.__mro__: if cc is unicode_set: break for rr in cc._ranges: ret.extend(range(rr[0], rr[-1] + 1)) return [unichr(c) for c in sorted(set(ret))] @_lazyclassproperty def printables(cls): "all non-whitespace characters in this range" return u''.join(filterfalse(unicode.isspace, cls._get_chars_for_ranges())) @_lazyclassproperty def alphas(cls): "all alphabetic characters in this range" return u''.join(filter(unicode.isalpha, cls._get_chars_for_ranges())) @_lazyclassproperty def nums(cls): "all numeric digit characters in this range" return u''.join(filter(unicode.isdigit, cls._get_chars_for_ranges())) @_lazyclassproperty def alphanums(cls): "all alphanumeric characters in this range" return cls.alphas + cls.nums class pyparsing_unicode(unicode_set): """ A namespace class for defining common language unicode_sets. """ _ranges = [(32, sys.maxunicode)] class Latin1(unicode_set): "Unicode set for Latin-1 Unicode Character Range" _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),] class LatinA(unicode_set): "Unicode set for Latin-A Unicode Character Range" _ranges = [(0x0100, 0x017f),] class LatinB(unicode_set): "Unicode set for Latin-B Unicode Character Range" _ranges = [(0x0180, 0x024f),] class Greek(unicode_set): "Unicode set for Greek Unicode Character Ranges" _ranges = [ (0x0370, 0x03ff), (0x1f00, 0x1f15), (0x1f18, 0x1f1d), (0x1f20, 0x1f45), (0x1f48, 0x1f4d), (0x1f50, 0x1f57), (0x1f59,), (0x1f5b,), (0x1f5d,), (0x1f5f, 0x1f7d), (0x1f80, 0x1fb4), (0x1fb6, 0x1fc4), (0x1fc6, 0x1fd3), (0x1fd6, 0x1fdb), (0x1fdd, 0x1fef), (0x1ff2, 0x1ff4), (0x1ff6, 0x1ffe), ] class Cyrillic(unicode_set): "Unicode set for Cyrillic Unicode Character Range" _ranges = [(0x0400, 0x04ff)] class Chinese(unicode_set): "Unicode set for Chinese Unicode Character Range" _ranges = [(0x4e00, 0x9fff), (0x3000, 0x303f),] class Japanese(unicode_set): "Unicode set for Japanese Unicode Character Range, combining Kanji, Hiragana, and Katakana ranges" _ranges = [] class Kanji(unicode_set): "Unicode set for Kanji Unicode Character Range" _ranges = [(0x4E00, 0x9Fbf), (0x3000, 0x303f),] class Hiragana(unicode_set): "Unicode set for Hiragana Unicode Character Range" _ranges = [(0x3040, 0x309f),] class Katakana(unicode_set): "Unicode set for Katakana Unicode Character Range" _ranges = [(0x30a0, 0x30ff),] class Korean(unicode_set): "Unicode set for Korean Unicode Character Range" _ranges = [(0xac00, 0xd7af), (0x1100, 0x11ff), (0x3130, 0x318f), (0xa960, 0xa97f), (0xd7b0, 0xd7ff), (0x3000, 0x303f),] class CJK(Chinese, Japanese, Korean): "Unicode set for combined Chinese, Japanese, and Korean (CJK) Unicode Character Range" pass class Thai(unicode_set): "Unicode set for Thai Unicode Character Range" _ranges = [(0x0e01, 0x0e3a), (0x0e3f, 0x0e5b),] class Arabic(unicode_set): "Unicode set for Arabic Unicode Character Range" _ranges = [(0x0600, 0x061b), (0x061e, 0x06ff), (0x0700, 0x077f),] class Hebrew(unicode_set): "Unicode set for Hebrew Unicode Character Range" _ranges = [(0x0590, 0x05ff),] class Devanagari(unicode_set): "Unicode set for Devanagari Unicode Character Range" _ranges = [(0x0900, 0x097f), (0xa8e0, 0xa8ff)] pyparsing_unicode.Japanese._ranges = (pyparsing_unicode.Japanese.Kanji._ranges + pyparsing_unicode.Japanese.Hiragana._ranges + pyparsing_unicode.Japanese.Katakana._ranges) # define ranges in language character sets if PY_3: setattr(pyparsing_unicode, u"العربية", pyparsing_unicode.Arabic) setattr(pyparsing_unicode, u"中文", pyparsing_unicode.Chinese) setattr(pyparsing_unicode, u"кириллица", pyparsing_unicode.Cyrillic) setattr(pyparsing_unicode, u"Ελληνικά", pyparsing_unicode.Greek) setattr(pyparsing_unicode, u"עִברִית", pyparsing_unicode.Hebrew) setattr(pyparsing_unicode, u"日本語", pyparsing_unicode.Japanese) setattr(pyparsing_unicode.Japanese, u"漢字", pyparsing_unicode.Japanese.Kanji) setattr(pyparsing_unicode.Japanese, u"カタカナ", pyparsing_unicode.Japanese.Katakana) setattr(pyparsing_unicode.Japanese, u"ひらがな", pyparsing_unicode.Japanese.Hiragana) setattr(pyparsing_unicode, u"한국어", pyparsing_unicode.Korean) setattr(pyparsing_unicode, u"ไทย", pyparsing_unicode.Thai) setattr(pyparsing_unicode, u"देवनागरी", pyparsing_unicode.Devanagari) if __name__ == "__main__": selectToken = CaselessLiteral("select") fromToken = CaselessLiteral("from") ident = Word(alphas, alphanums + "_$") columnName = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens) columnNameList = Group(delimitedList(columnName)).setName("columns") columnSpec = ('*' | columnNameList) tableName = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens) tableNameList = Group(delimitedList(tableName)).setName("tables") simpleSQL = selectToken("command") + columnSpec("columns") + fromToken + tableNameList("tables") # demo runTests method, including embedded comments in test string simpleSQL.runTests(""" # '*' as column list and dotted table name select * from SYS.XYZZY # caseless match on "SELECT", and casts back to "select" SELECT * from XYZZY, ABC # list of column names, and mixed case SELECT keyword Select AA,BB,CC from Sys.dual # multiple tables Select A, B, C from Sys.dual, Table2 # invalid SELECT keyword - should fail Xelect A, B, C from Sys.dual # incomplete command - should fail Select # invalid column name - should fail Select ^^^ frox Sys.dual """) pyparsing_common.number.runTests(""" 100 -100 +100 3.14159 6.02e23 1e-12 """) # any int or real number, returned as float pyparsing_common.fnumber.runTests(""" 100 -100 +100 3.14159 6.02e23 1e-12 """) pyparsing_common.hex_integer.runTests(""" 100 FF """) import uuid pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID)) pyparsing_common.uuid.runTests(""" 12345678-1234-5678-1234-567812345678 """) site-packages/pip/_vendor/contextlib2.py000064400000041023151464452720014331 0ustar00"""contextlib2 - backports and enhancements to the contextlib module""" import abc import sys import warnings from collections import deque from functools import wraps __all__ = ["contextmanager", "closing", "nullcontext", "AbstractContextManager", "ContextDecorator", "ExitStack", "redirect_stdout", "redirect_stderr", "suppress"] # Backwards compatibility __all__ += ["ContextStack"] # Backport abc.ABC if sys.version_info[:2] >= (3, 4): _abc_ABC = abc.ABC else: _abc_ABC = abc.ABCMeta('ABC', (object,), {'__slots__': ()}) # Backport classic class MRO def _classic_mro(C, result): if C in result: return result.append(C) for B in C.__bases__: _classic_mro(B, result) return result # Backport _collections_abc._check_methods def _check_methods(C, *methods): try: mro = C.__mro__ except AttributeError: mro = tuple(_classic_mro(C, [])) for method in methods: for B in mro: if method in B.__dict__: if B.__dict__[method] is None: return NotImplemented break else: return NotImplemented return True class AbstractContextManager(_abc_ABC): """An abstract base class for context managers.""" def __enter__(self): """Return `self` upon entering the runtime context.""" return self @abc.abstractmethod def __exit__(self, exc_type, exc_value, traceback): """Raise any exception triggered within the runtime context.""" return None @classmethod def __subclasshook__(cls, C): """Check whether subclass is considered a subclass of this ABC.""" if cls is AbstractContextManager: return _check_methods(C, "__enter__", "__exit__") return NotImplemented class ContextDecorator(object): """A base class or mixin that enables context managers to work as decorators.""" def refresh_cm(self): """Returns the context manager used to actually wrap the call to the decorated function. The default implementation just returns *self*. Overriding this method allows otherwise one-shot context managers like _GeneratorContextManager to support use as decorators via implicit recreation. DEPRECATED: refresh_cm was never added to the standard library's ContextDecorator API """ warnings.warn("refresh_cm was never added to the standard library", DeprecationWarning) return self._recreate_cm() def _recreate_cm(self): """Return a recreated instance of self. Allows an otherwise one-shot context manager like _GeneratorContextManager to support use as a decorator via implicit recreation. This is a private interface just for _GeneratorContextManager. See issue #11647 for details. """ return self def __call__(self, func): @wraps(func) def inner(*args, **kwds): with self._recreate_cm(): return func(*args, **kwds) return inner class _GeneratorContextManager(ContextDecorator): """Helper for @contextmanager decorator.""" def __init__(self, func, args, kwds): self.gen = func(*args, **kwds) self.func, self.args, self.kwds = func, args, kwds # Issue 19330: ensure context manager instances have good docstrings doc = getattr(func, "__doc__", None) if doc is None: doc = type(self).__doc__ self.__doc__ = doc # Unfortunately, this still doesn't provide good help output when # inspecting the created context manager instances, since pydoc # currently bypasses the instance docstring and shows the docstring # for the class instead. # See http://bugs.python.org/issue19404 for more details. def _recreate_cm(self): # _GCM instances are one-shot context managers, so the # CM must be recreated each time a decorated function is # called return self.__class__(self.func, self.args, self.kwds) def __enter__(self): try: return next(self.gen) except StopIteration: raise RuntimeError("generator didn't yield") def __exit__(self, type, value, traceback): if type is None: try: next(self.gen) except StopIteration: return else: raise RuntimeError("generator didn't stop") else: if value is None: # Need to force instantiation so we can reliably # tell if we get the same exception back value = type() try: self.gen.throw(type, value, traceback) raise RuntimeError("generator didn't stop after throw()") except StopIteration as exc: # Suppress StopIteration *unless* it's the same exception that # was passed to throw(). This prevents a StopIteration # raised inside the "with" statement from being suppressed. return exc is not value except RuntimeError as exc: # Don't re-raise the passed in exception if exc is value: return False # Likewise, avoid suppressing if a StopIteration exception # was passed to throw() and later wrapped into a RuntimeError # (see PEP 479). if _HAVE_EXCEPTION_CHAINING and exc.__cause__ is value: return False raise except: # only re-raise if it's *not* the exception that was # passed to throw(), because __exit__() must not raise # an exception unless __exit__() itself failed. But throw() # has to raise the exception to signal propagation, so this # fixes the impedance mismatch between the throw() protocol # and the __exit__() protocol. # if sys.exc_info()[1] is not value: raise def contextmanager(func): """@contextmanager decorator. Typical usage: @contextmanager def some_generator(): try: yield finally: This makes this: with some_generator() as : equivalent to this: try: = finally: """ @wraps(func) def helper(*args, **kwds): return _GeneratorContextManager(func, args, kwds) return helper class closing(object): """Context to automatically close something at the end of a block. Code like this: with closing(.open()) as f: is equivalent to this: f = .open() try: finally: f.close() """ def __init__(self, thing): self.thing = thing def __enter__(self): return self.thing def __exit__(self, *exc_info): self.thing.close() class _RedirectStream(object): _stream = None def __init__(self, new_target): self._new_target = new_target # We use a list of old targets to make this CM re-entrant self._old_targets = [] def __enter__(self): self._old_targets.append(getattr(sys, self._stream)) setattr(sys, self._stream, self._new_target) return self._new_target def __exit__(self, exctype, excinst, exctb): setattr(sys, self._stream, self._old_targets.pop()) class redirect_stdout(_RedirectStream): """Context manager for temporarily redirecting stdout to another file. # How to send help() to stderr with redirect_stdout(sys.stderr): help(dir) # How to write help() to a file with open('help.txt', 'w') as f: with redirect_stdout(f): help(pow) """ _stream = "stdout" class redirect_stderr(_RedirectStream): """Context manager for temporarily redirecting stderr to another file.""" _stream = "stderr" class suppress(object): """Context manager to suppress specified exceptions After the exception is suppressed, execution proceeds with the next statement following the with statement. with suppress(FileNotFoundError): os.remove(somefile) # Execution still resumes here if the file was already removed """ def __init__(self, *exceptions): self._exceptions = exceptions def __enter__(self): pass def __exit__(self, exctype, excinst, exctb): # Unlike isinstance and issubclass, CPython exception handling # currently only looks at the concrete type hierarchy (ignoring # the instance and subclass checking hooks). While Guido considers # that a bug rather than a feature, it's a fairly hard one to fix # due to various internal implementation details. suppress provides # the simpler issubclass based semantics, rather than trying to # exactly reproduce the limitations of the CPython interpreter. # # See http://bugs.python.org/issue12029 for more details return exctype is not None and issubclass(exctype, self._exceptions) # Context manipulation is Python 3 only _HAVE_EXCEPTION_CHAINING = sys.version_info[0] >= 3 if _HAVE_EXCEPTION_CHAINING: def _make_context_fixer(frame_exc): def _fix_exception_context(new_exc, old_exc): # Context may not be correct, so find the end of the chain while 1: exc_context = new_exc.__context__ if exc_context is old_exc: # Context is already set correctly (see issue 20317) return if exc_context is None or exc_context is frame_exc: break new_exc = exc_context # Change the end of the chain to point to the exception # we expect it to reference new_exc.__context__ = old_exc return _fix_exception_context def _reraise_with_existing_context(exc_details): try: # bare "raise exc_details[1]" replaces our carefully # set-up context fixed_ctx = exc_details[1].__context__ raise exc_details[1] except BaseException: exc_details[1].__context__ = fixed_ctx raise else: # No exception context in Python 2 def _make_context_fixer(frame_exc): return lambda new_exc, old_exc: None # Use 3 argument raise in Python 2, # but use exec to avoid SyntaxError in Python 3 def _reraise_with_existing_context(exc_details): exc_type, exc_value, exc_tb = exc_details exec("raise exc_type, exc_value, exc_tb") # Handle old-style classes if they exist try: from types import InstanceType except ImportError: # Python 3 doesn't have old-style classes _get_type = type else: # Need to handle old-style context managers on Python 2 def _get_type(obj): obj_type = type(obj) if obj_type is InstanceType: return obj.__class__ # Old-style class return obj_type # New-style class # Inspired by discussions on http://bugs.python.org/issue13585 class ExitStack(object): """Context manager for dynamic management of a stack of exit callbacks For example: with ExitStack() as stack: files = [stack.enter_context(open(fname)) for fname in filenames] # All opened files will automatically be closed at the end of # the with statement, even if attempts to open files later # in the list raise an exception """ def __init__(self): self._exit_callbacks = deque() def pop_all(self): """Preserve the context stack by transferring it to a new instance""" new_stack = type(self)() new_stack._exit_callbacks = self._exit_callbacks self._exit_callbacks = deque() return new_stack def _push_cm_exit(self, cm, cm_exit): """Helper to correctly register callbacks to __exit__ methods""" def _exit_wrapper(*exc_details): return cm_exit(cm, *exc_details) _exit_wrapper.__self__ = cm self.push(_exit_wrapper) def push(self, exit): """Registers a callback with the standard __exit__ method signature Can suppress exceptions the same way __exit__ methods can. Also accepts any object with an __exit__ method (registering a call to the method instead of the object itself) """ # We use an unbound method rather than a bound method to follow # the standard lookup behaviour for special methods _cb_type = _get_type(exit) try: exit_method = _cb_type.__exit__ except AttributeError: # Not a context manager, so assume its a callable self._exit_callbacks.append(exit) else: self._push_cm_exit(exit, exit_method) return exit # Allow use as a decorator def callback(self, callback, *args, **kwds): """Registers an arbitrary callback and arguments. Cannot suppress exceptions. """ def _exit_wrapper(exc_type, exc, tb): callback(*args, **kwds) # We changed the signature, so using @wraps is not appropriate, but # setting __wrapped__ may still help with introspection _exit_wrapper.__wrapped__ = callback self.push(_exit_wrapper) return callback # Allow use as a decorator def enter_context(self, cm): """Enters the supplied context manager If successful, also pushes its __exit__ method as a callback and returns the result of the __enter__ method. """ # We look up the special methods on the type to match the with statement _cm_type = _get_type(cm) _exit = _cm_type.__exit__ result = _cm_type.__enter__(cm) self._push_cm_exit(cm, _exit) return result def close(self): """Immediately unwind the context stack""" self.__exit__(None, None, None) def __enter__(self): return self def __exit__(self, *exc_details): received_exc = exc_details[0] is not None # We manipulate the exception state so it behaves as though # we were actually nesting multiple with statements frame_exc = sys.exc_info()[1] _fix_exception_context = _make_context_fixer(frame_exc) # Callbacks are invoked in LIFO order to match the behaviour of # nested context managers suppressed_exc = False pending_raise = False while self._exit_callbacks: cb = self._exit_callbacks.pop() try: if cb(*exc_details): suppressed_exc = True pending_raise = False exc_details = (None, None, None) except: new_exc_details = sys.exc_info() # simulate the stack of exceptions by setting the context _fix_exception_context(new_exc_details[1], exc_details[1]) pending_raise = True exc_details = new_exc_details if pending_raise: _reraise_with_existing_context(exc_details) return received_exc and suppressed_exc # Preserve backwards compatibility class ContextStack(ExitStack): """Backwards compatibility alias for ExitStack""" def __init__(self): warnings.warn("ContextStack has been renamed to ExitStack", DeprecationWarning) super(ContextStack, self).__init__() def register_exit(self, callback): return self.push(callback) def register(self, callback, *args, **kwds): return self.callback(callback, *args, **kwds) def preserve(self): return self.pop_all() class nullcontext(AbstractContextManager): """Context manager that does no additional processing. Used as a stand-in for a normal context manager, when a particular block of code is only sometimes used with a normal context manager: cm = optional_cm if condition else nullcontext() with cm: # Perform operation, using optional_cm if condition is True """ def __init__(self, enter_result=None): self.enter_result = enter_result def __enter__(self): return self.enter_result def __exit__(self, *excinfo): pass site-packages/pip/_vendor/pep517/check.py000064400000013475151464452720014204 0ustar00"""Check a project and backend by attempting to build using PEP 517 hooks. """ import argparse import logging import os from os.path import isfile, join as pjoin from toml import TomlDecodeError, load as toml_load import shutil from subprocess import CalledProcessError import sys import tarfile from tempfile import mkdtemp import zipfile from .colorlog import enable_colourful_output from .envbuild import BuildEnvironment from .wrappers import Pep517HookCaller log = logging.getLogger(__name__) def check_build_sdist(hooks, build_sys_requires): with BuildEnvironment() as env: try: env.pip_install(build_sys_requires) log.info('Installed static build dependencies') except CalledProcessError: log.error('Failed to install static build dependencies') return False try: reqs = hooks.get_requires_for_build_sdist({}) log.info('Got build requires: %s', reqs) except Exception: log.error('Failure in get_requires_for_build_sdist', exc_info=True) return False try: env.pip_install(reqs) log.info('Installed dynamic build dependencies') except CalledProcessError: log.error('Failed to install dynamic build dependencies') return False td = mkdtemp() log.info('Trying to build sdist in %s', td) try: try: filename = hooks.build_sdist(td, {}) log.info('build_sdist returned %r', filename) except Exception: log.info('Failure in build_sdist', exc_info=True) return False if not filename.endswith('.tar.gz'): log.error( "Filename %s doesn't have .tar.gz extension", filename) return False path = pjoin(td, filename) if isfile(path): log.info("Output file %s exists", path) else: log.error("Output file %s does not exist", path) return False if tarfile.is_tarfile(path): log.info("Output file is a tar file") else: log.error("Output file is not a tar file") return False finally: shutil.rmtree(td) return True def check_build_wheel(hooks, build_sys_requires): with BuildEnvironment() as env: try: env.pip_install(build_sys_requires) log.info('Installed static build dependencies') except CalledProcessError: log.error('Failed to install static build dependencies') return False try: reqs = hooks.get_requires_for_build_wheel({}) log.info('Got build requires: %s', reqs) except Exception: log.error('Failure in get_requires_for_build_sdist', exc_info=True) return False try: env.pip_install(reqs) log.info('Installed dynamic build dependencies') except CalledProcessError: log.error('Failed to install dynamic build dependencies') return False td = mkdtemp() log.info('Trying to build wheel in %s', td) try: try: filename = hooks.build_wheel(td, {}) log.info('build_wheel returned %r', filename) except Exception: log.info('Failure in build_wheel', exc_info=True) return False if not filename.endswith('.whl'): log.error("Filename %s doesn't have .whl extension", filename) return False path = pjoin(td, filename) if isfile(path): log.info("Output file %s exists", path) else: log.error("Output file %s does not exist", path) return False if zipfile.is_zipfile(path): log.info("Output file is a zip file") else: log.error("Output file is not a zip file") return False finally: shutil.rmtree(td) return True def check(source_dir): pyproject = pjoin(source_dir, 'pyproject.toml') if isfile(pyproject): log.info('Found pyproject.toml') else: log.error('Missing pyproject.toml') return False try: with open(pyproject) as f: pyproject_data = toml_load(f) # Ensure the mandatory data can be loaded buildsys = pyproject_data['build-system'] requires = buildsys['requires'] backend = buildsys['build-backend'] backend_path = buildsys.get('backend-path') log.info('Loaded pyproject.toml') except (TomlDecodeError, KeyError): log.error("Invalid pyproject.toml", exc_info=True) return False hooks = Pep517HookCaller(source_dir, backend, backend_path) sdist_ok = check_build_sdist(hooks, requires) wheel_ok = check_build_wheel(hooks, requires) if not sdist_ok: log.warning('Sdist checks failed; scroll up to see') if not wheel_ok: log.warning('Wheel checks failed') return sdist_ok def main(argv=None): ap = argparse.ArgumentParser() ap.add_argument( 'source_dir', help="A directory containing pyproject.toml") args = ap.parse_args(argv) enable_colourful_output() ok = check(args.source_dir) if ok: print(ansi('Checks passed', 'green')) else: print(ansi('Checks failed', 'red')) sys.exit(1) ansi_codes = { 'reset': '\x1b[0m', 'bold': '\x1b[1m', 'red': '\x1b[31m', 'green': '\x1b[32m', } def ansi(s, attr): if os.name != 'nt' and sys.stdout.isatty(): return ansi_codes[attr] + str(s) + ansi_codes['reset'] else: return str(s) if __name__ == '__main__': main() site-packages/pip/_vendor/pep517/meta.py000064400000004637151464452720014055 0ustar00"""Build metadata for a project using PEP 517 hooks. """ import argparse import logging import os import shutil import functools try: import importlib.metadata as imp_meta except ImportError: import importlib_metadata as imp_meta try: from zipfile import Path except ImportError: from zipp import Path from .envbuild import BuildEnvironment from .wrappers import Pep517HookCaller, quiet_subprocess_runner from .dirtools import tempdir, mkdir_p, dir_to_zipfile from .build import validate_system, load_system, compat_system log = logging.getLogger(__name__) def _prep_meta(hooks, env, dest): reqs = hooks.get_requires_for_build_wheel({}) log.info('Got build requires: %s', reqs) env.pip_install(reqs) log.info('Installed dynamic build dependencies') with tempdir() as td: log.info('Trying to build metadata in %s', td) filename = hooks.prepare_metadata_for_build_wheel(td, {}) source = os.path.join(td, filename) shutil.move(source, os.path.join(dest, os.path.basename(filename))) def build(source_dir='.', dest=None, system=None): system = system or load_system(source_dir) dest = os.path.join(source_dir, dest or 'dist') mkdir_p(dest) validate_system(system) hooks = Pep517HookCaller( source_dir, system['build-backend'], system.get('backend-path') ) with hooks.subprocess_runner(quiet_subprocess_runner): with BuildEnvironment() as env: env.pip_install(system['requires']) _prep_meta(hooks, env, dest) def build_as_zip(builder=build): with tempdir() as out_dir: builder(dest=out_dir) return dir_to_zipfile(out_dir) def load(root): """ Given a source directory (root) of a package, return an importlib.metadata.Distribution object with metadata build from that package. """ root = os.path.expanduser(root) system = compat_system(root) builder = functools.partial(build, source_dir=root, system=system) path = Path(build_as_zip(builder)) return imp_meta.PathDistribution(path) parser = argparse.ArgumentParser() parser.add_argument( 'source_dir', help="A directory containing pyproject.toml", ) parser.add_argument( '--out-dir', '-o', help="Destination in which to save the builds relative to source dir", ) def main(): args = parser.parse_args() build(args.source_dir, args.out_dir) if __name__ == '__main__': main() site-packages/pip/_vendor/pep517/__pycache__/wrappers.cpython-38.opt-1.pyc000064400000023636151464452720022217 0ustar00U .eg)@s ddlZddlmZddlZddlmZmZmZddl Z ddl m Z m Z m Z ddlZddlmZddlmZeeeedZed d ZGd d d eZGd ddeZGdddeZGdddeZdddZdddZddZGdddeZGdddej Z!dS)N)contextmanager)dirnameabspathjoin) check_call check_outputSTDOUT)mkdtemp)compatz_in_process.pyc cs"t}z |VW5t|XdSN)r shutilZrmtree)tdr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/wrappers.pytempdirs rc@seZdZdZddZdS)BackendUnavailablezEWill be raised if the backend cannot be imported in the hook process.cCs ||_dSr  tracebackselfrrrr__init__szBackendUnavailable.__init__N__name__ __module__ __qualname____doc__rrrrrrsrc@seZdZdZddZdS)BackendInvalidz)Will be raised if the backend is invalid.cCs||_||_||_dSr  backend_name backend_pathmessage)rrr r!rrrr szBackendInvalid.__init__Nrrrrrrsrcs eZdZdZfddZZS) HookMissingz Will be raised on missing hooks.cstt||||_dSr )superr"r hook_name)rr$ __class__rrr(szHookMissing.__init__)rrrrr __classcell__rrr%rr"&sr"c@seZdZdZddZdS)UnsupportedOperationzDMay be raised by build_sdist if the backend indicates that it can't.cCs ||_dSr rrrrrr/szUnsupportedOperation.__init__Nrrrrrr(-sr(cCs*tj}|r||t|||ddS)z5The default method of calling the wrapper subprocess.)cwdenvN)osenvironcopyupdatercmdr) extra_environr*rrrdefault_subprocess_runner3s  r2cCs,tj}|r||t|||tddS)zDA method of calling the wrapper subprocess while suppressing output.)r)r*stderrN)r+r,r-r.rrr/rrrquiet_subprocess_runner<s  r4cCsntj|rtdtj|}tjtj||}tj|}tj|}tj||g|krjtd|S)zNormalise and check a backend path. Ensure that the requested backend path is specified as a relative path, and resolves to a location under the given source tree. Return an absolute version of the requested path. zpaths must be relativez paths must be inside source tree) r+pathisabs ValueErrorrnormpathrnormcase commonprefix)Z source_treeZ requestedZ abs_sourceZ abs_requestedZ norm_sourceZnorm_requestedrrrnorm_and_checkEs    r;c@s`eZdZdZdddZeddZdddZdd d Zdd d Z dddZ dddZ ddZ dS)Pep517HookCalleraeA wrapper around a source directory to be built with a PEP 517 backend. source_dir : The path to the source directory, containing pyproject.toml. build_backend : The build backend spec, as per PEP 517, from pyproject.toml. backend_path : The backend path, as per PEP 517, from pyproject.toml. runner : A callable that invokes the wrapper subprocess. The 'runner', if provided, must expect the following: cmd : a list of strings representing the command and arguments to execute, as would be passed to e.g. 'subprocess.check_call'. cwd : a string representing the working directory that must be used for the subprocess. Corresponds to the provided source_dir. extra_environ : a dict mapping environment variable names to values which must be set for the subprocess execution. NcsB|dkr t}t|_|_|r2fdd|D}|_|_dS)Ncsg|]}tj|qSr)r; source_dir).0prrr {sz-Pep517HookCaller.__init__..)r2rr= build_backendr _subprocess_runner)rr=rBr runnerrr@rrns  zPep517HookCaller.__init__ccs|j}||_dV||_dS)z\A context manager for temporarily overriding the default subprocess runner. N)rC)rrDprevrrrsubprocess_runnersz"Pep517HookCaller.subprocess_runnercCs|dd|iS)aGIdentify packages required for building a wheel Returns a list of dependency specifications, e.g.: ["wheel >= 0.25", "setuptools"] This does not include requirements specified in pyproject.toml. It returns the result of calling the equivalently named hook in a subprocess. get_requires_for_build_wheelconfig_settings _call_hookrrHrrrrGs z-Pep517HookCaller.get_requires_for_build_wheelTcCs|dt|||dS)aqPrepare a *.dist-info folder with metadata for this project. Returns the name of the newly created folder. If the build backend defines a hook with this name, it will be called in a subprocess. If not, the backend will be asked to build a wheel, and the dist-info extracted from that (unless _allow_fallback is False). prepare_metadata_for_build_wheel)metadata_directoryrH_allow_fallbackrJr)rrMrHrNrrrrLs z1Pep517HookCaller.prepare_metadata_for_build_wheelcCs(|dk rt|}|dt|||dS)avBuild a wheel from this project. Returns the name of the newly created file. In general, this will call the 'build_wheel' hook in the backend. However, if that was previously called by 'prepare_metadata_for_build_wheel', and the same metadata_directory is used, the previously built wheel will be copied to wheel_directory. N build_wheel)wheel_directoryrHrM)rrJ)rrQrHrMrrrrPs zPep517HookCaller.build_wheelcCs|dd|iS)a<Identify packages required for building a wheel Returns a list of dependency specifications, e.g.: ["setuptools >= 26"] This does not include requirements specified in pyproject.toml. It returns the result of calling the equivalently named hook in a subprocess. get_requires_for_build_sdistrHrIrKrrrrRs z-Pep517HookCaller.get_requires_for_build_sdistcCs|dt||dS)zBuild an sdist from this project. Returns the name of the newly created file. This calls the 'build_sdist' backend hook in a subprocess. build_sdist)sdist_directoryrHrO)rrTrHrrrrSszPep517HookCaller.build_sdistc Cs@tjddkr|jd}n|j}d|i}|jrbtj|j}tjddkrZ|t}||d<t }d|i}t j |t |ddd|j tjt||g|j|d t t |d }|d rt|d d |drt|d d |drt|j|j|dd d|dr"t||dW5QRSQRXdS)NrASCIIZPEP517_BUILD_BACKENDZPEP517_BACKEND_PATHkwargsz input.json)indent)r)r1z output.jsonZ unsupportedrZ no_backendZbackend_invalidZ backend_errorrZ hook_missingZ return_val)sys version_inforBencoder r+pathseprgetfilesystemencodingrr Z write_jsonpjoinrC executable_in_proc_scriptr=Z read_jsongetr(rrr") rr$rWrBr1r rZ hook_inputdatarrrrJsB      zPep517HookCaller._call_hook)NN)N)NT)NN)N)N) rrrrrrrFrGrLrPrRrSrJrrrrr<]s       r<c@s<eZdZdZddZddZeddZdd Zd d Z d S) LoggerWrapperzd Read messages from a pipe and redirect them to a logger (see python's logging module). cCsHtj|d|_||_||_t\|_|_ t |j|_ | dS)NT) threadingThreadrZdaemonloggerlevelr+pipeZfd_readfd_writefdopenreaderstart)rrgrhrrrrs zLoggerWrapper.__init__cCs|jSr )rjr@rrrfilenoszLoggerWrapper.filenocCs|tjr|ddS|S)N)endswithr+linesep)msgrrrremove_newline!szLoggerWrapper.remove_newlinecCs |jD]}|||qdSr )rl_writers)rlinerrrrun%s zLoggerWrapper.runcCs|j|j|dSr )rglogrh)rr!rrrrt)szLoggerWrapper._writeN) rrrrrrn staticmethodrsrvrtrrrrrd s  rd)NN)NN)"re contextlibrr+Zos.pathrrrr_r subprocessrrrrZZtempfiler rYr __file__rar Exceptionrrr"r(r2r4r;objectr<rfrdrrrrs*     /site-packages/pip/_vendor/pep517/__pycache__/compat.cpython-38.pyc000064400000001765151464452720020677 0ustar00U .e @sjdZddlZddlZejddkr4ddZddZnddZd dZzeZWnek rdeZYnXdS) zPython 2/3 compatibilityNc Ks.t|ddd}tj||f|W5QRXdS)Nwutf-8encodingopenjsondumpobjpathkwargsfr=/usr/lib/python3.8/site-packages/pip/_vendor/pep517/compat.py write_json src Cs0t|ddd}t|W5QRSQRXdS)Nrrrrr loadr rrrr read_jsonsrc Ks2t|d}tj||fddi|W5QRXdS)Nwbrrrr rrrrs c Cs,t|d}t|W5QRSQRXdS)Nrbrrrrrrs ) __doc__r sys version_inforrFileNotFoundError NameErrorIOErrorrrrrs site-packages/pip/_vendor/pep517/__pycache__/check.cpython-38.pyc000064400000011240151464452720020456 0ustar00U .e=@sdZddlZddlZddlZddlmZmZddlm Z m Z ddl Z ddl mZddlZddlZddlmZddlZddlmZdd lmZdd lmZeeZd d Zd dZddZdddZ dddddZ!ddZ"edkre dS)zHCheck a project and backend by attempting to build using PEP 517 hooks. N)isfilejoin)TomlDecodeErrorload)CalledProcessError)mkdtemp)enable_colourful_output)BuildEnvironment)Pep517HookCallerc Cst}z||tdWn*tk rLtdYW5QRdSXz|i}td|Wn.tk rtjdddYW5QRdSXz||tdWn*tk rtd YW5QRdSXt}td |zz| |i}td |Wn4tk r@tjd ddYWW5QRdSX| d sltd|WW5QRdSt ||}t|rtd|ntd|WFW5QRdSt|rtdntdWW5QRdSW5t |XW5QRdSQRXdS)N#Installed static build dependencies+Failed to install static build dependenciesFGot build requires: %s'Failure in get_requires_for_build_sdistTexc_info$Installed dynamic build dependencies,Failed to install dynamic build dependencieszTrying to build sdist in %szbuild_sdist returned %rzFailure in build_sdistz.tar.gzz*Filename %s doesn't have .tar.gz extensionOutput file %s existsOutput file %s does not existzOutput file is a tar filezOutput file is not a tar file)r pip_installloginforerrorZget_requires_for_build_sdist ExceptionrshutilrmtreeZ build_sdistendswithpjoinrtarfileZ is_tarfilehooksZbuild_sys_requiresenvZreqsZtdfilenamepathr%s6      98" site-packages/pip/_vendor/pep517/__pycache__/_in_process.cpython-38.pyc000064400000016405151464452720021714 0ustar00U .ep@s0dZddlmZddlmZddlZddlZddlmZddlZddl Z ddl Z ddl Z ddl Z Gddde ZGdd d e ZGd d d e Zd d ZddZddZddZdZddZddZddZd(ddZddZGdd d e ZGd!d"d"e Zd#d$Zddddd$hZd%d&Ze d'kr,edS))agThis is invoked in a subprocess to call the build backend hooks. It expects: - Command line args: hook_name, control_dir - Environment variables: PEP517_BUILD_BACKEND=entry.point:spec PEP517_BACKEND_PATH=paths (separated with os.pathsep) - control_dir/input.json: - {"kwargs": {...}} Results: - control_dir/output.json - {"return_val": ...} )glob) import_moduleN)joinc@seZdZdZddZdS)BackendUnavailablez&Raised if we cannot import the backendcCs ||_dSN tracebackselfrr B/usr/lib/python3.8/site-packages/pip/_vendor/pep517/_in_process.py__init__szBackendUnavailable.__init__N__name__ __module__ __qualname____doc__r r r r r rsrc@seZdZdZddZdS)BackendInvalidz Raised if the backend is invalidcCs ||_dSr)message)r rr r r r %szBackendInvalid.__init__Nrr r r r r#src@seZdZdZdS) HookMissingzARaised if a hook is missing and we are not executing the fallbackNrrrrr r r r r)srcCs<tjtj|}tjtj|}tj||g|kS)z5Test if a file is located within the given directory.)ospathnormcaseabspath commonprefix)filenameZ directoryr r r contained_in-srcstjd}|r*|tj}|tjdd<tjd}|d\}}}z t|Wn t k rpt t YnX|rt fdd|Dstd|r|d D]}t|qS) zFind and load the build backendZPEP517_BACKEND_PATHNrZPEP517_BUILD_BACKEND:c3s|]}tj|VqdSr)r__file__).0robjr r Dsz!_build_backend..z(Backend was not loaded from backend-path.)renvirongetsplitpathsepsysr partitionr ImportErrorrr format_excanyrgetattr)Z backend_pathZextra_pathitemsZepZmod_path_Zobj_pathZ path_partr r!r _build_backend4s&      r0cCs6t}z |j}Wntk r(gYSX||SdSzfInvoke the optional get_requires_for_build_wheel hook Returns [] if the hook is not defined. N)r0get_requires_for_build_wheelAttributeErrorconfig_settingsbackendhookr r r r2Ps   r2cCsJt}z |j}Wn*tk r:|s*tt|||YSX|||SdS)zInvoke optional prepare_metadata_for_build_wheel Implements a fallback by building a wheel if the hook isn't defined, unless _allow_fallback is False in which case HookMissing is raised. N)r0 prepare_metadata_for_build_wheelr3r_get_wheel_metadata_from_wheel)metadata_directoryr5Z_allow_fallbackr6r7r r r r8^s  r8ZPEP517_ALREADY_BUILT_WHEELcCs@g}|D]}td|}|r ||q |r4|StddS)z6Identify the .dist-info folder inside a wheel ZipFile.z[^/\\]+-[^/\\]+\.dist-info/z#No .dist-info folder found in wheelN)Znamelistrematchappend Exception)Zwhl_zipresrmr r r _dist_info_filests   rAc Csddlm}|||}ttj|tdW5QRXtj||}||}t|}|j ||dW5QRX|d ddS)zBuild a wheel and extract the metadata from it. Fallback for when the build backend does not define the 'get_wheel_metadata' hook. r)ZipFilewb)rmembers/) ZzipfilerB build_wheelopenrrrWHEEL_BUILT_MARKERrAZ extractallr')r6r:r5rBZ whl_basenameZwhl_fileZzipfZ dist_infor r r r9s    r9cCsl|sdStj|}tjt|ts*dSttj|d}|sLtddSt |dkrdtddS|dS)zHCheck for a wheel already built during the get_wheel_metadata hook. Nz*.whlz+Found wheel built marker, but no .whl fileszHFound multiple .whl files; unspecified behaviour. Will call build_wheel.r) rrdirnameisfilepjoinrHrrprintlen)r:Zmetadata_parentZ whl_filesr r r _find_already_built_wheels  rOcCs4t|}|r$t||tj|St|||S)zInvoke the mandatory build_wheel hook. If a wheel was already built in the prepare_metadata_for_build_wheel fallback, this will copy it rather than rebuilding the wheel. )rOshutilZcopy2rrbasenamer0rF)Zwheel_directoryr5r:Z prebuilt_whlr r r rFs   rFcCs6t}z |j}Wntk r(gYSX||SdSr1)r0get_requires_for_build_sdistr3r4r r r rRs   rRc@seZdZdZdS)_DummyExceptionz(Nothing should ever raise this exceptionNrr r r r rSsrSc@seZdZdZddZdS)GotUnsupportedOperationz9For internal use when backend raises UnsupportedOperationcCs ||_dSrrr r r r r sz GotUnsupportedOperation.__init__Nrr r r r rTsrTc CsBt}z|||WSt|dtk r<ttYnXdS)z&Invoke the mandatory build_sdist hook.UnsupportedOperationN)r0 build_sdistr.rSrTrr,)Zsdist_directoryr5r6r r r rVs rVc CsNttjdkrtdtjd}tjd}|tkrBtd|t|}tt|d}ddd}z|f|d |d <Wnt k r}zd |d <|j |d <W5d}~XYnt k r}zd |d<|j |d<W5d}~XYnRt k r}zd |d<|j |d <W5d}~XYntk r2d |d<YnXtj|t|ddddS)Nz"Needs args: hook_name, control_dirrIzUnknown hook: %sz input.jsonF) unsupported return_valkwargsrZTZ no_backendrZbackend_invalidZ backend_errorrYZ hook_missingz output.json)indent)rNr)argvexit HOOK_NAMESglobalscompatZ read_jsonrLrrrrrTrZ write_json)Z hook_nameZ control_dirr7Z hook_inputZjson_outer r r mains.     rc__main__)N)!rr importlibrrZos.pathrrLr;rPr)rrar>rrrrr0r2r8rHrAr9rOrFrRrSrTrVr_rcrr r r r sF        site-packages/pip/_vendor/pep517/__pycache__/__init__.cpython-38.pyc000064400000000366151464452720021147 0ustar00U .eT@s dZdZdS)z6Wrappers to build Python packages using PEP 517 hooks z0.7.0N)__doc__ __version__rr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/__init__.pyssite-packages/pip/_vendor/pep517/__pycache__/meta.cpython-38.pyc000064400000005406151464452720020336 0ustar00U .e @sBdZddlZddlZddlZddlZddlZzddlmZWne k rXddl ZYnXzddl m Z Wn e k rddl m Z YnXddlmZddlmZmZddlmZmZmZddlmZmZmZeeZd d Zdd d ZefddZddZ e!Z"e"j#ddde"j#ddddddZ$edkr>e$dS)z2Build metadata for a project using PEP 517 hooks. N)Path)BuildEnvironment)Pep517HookCallerquiet_subprocess_runner)tempdirmkdir_pdir_to_zipfile)validate_system load_system compat_systemc Cs|i}td|||tdtJ}td|||i}tj||}t |tj|tj |W5QRXdS)NzGot build requires: %sz$Installed dynamic build dependencieszTrying to build metadata in %s) Zget_requires_for_build_wheelloginfo pip_installrZ prepare_metadata_for_build_wheelospathjoinshutilZmovebasename)hooksenvdestZreqsZtdfilenamesourcer;/usr/lib/python3.8/site-packages/pip/_vendor/pep517/meta.py _prep_metas      r.c Cs|p t|}tj||pd}t|t|t||d|d}|t 2t }| |dt |||W5QRXW5QRXdS)NZdistz build-backendz backend-pathZrequires) r rrrrr rgetZsubprocess_runnerrrrr) source_dirrsystemrrrrrbuild)s  r!c Cs0t }||dt|W5QRSQRXdS)N)r)rr )builderout_dirrrr build_as_zip8s r$cCs:tj|}t|}tjt||d}tt|}t |S)z Given a source directory (root) of a package, return an importlib.metadata.Distribution object with metadata build from that package. )rr ) rr expanduserr functoolspartialr!rr$imp_metaZPathDistribution)rootr r"rrrrload>s   r*rz%A directory containing pyproject.toml)helpz --out-dirz-oz>Destination in which to save the builds relative to source dircCst}t|j|jdS)N)parser parse_argsr!rr#)argsrrrmainVsr/__main__)rNN)%__doc__argparseZloggingrrr&importlib.metadataZmetadatar( ImportErrorZimportlib_metadataZzipfilerZzippZenvbuildrZwrappersrrZdirtoolsrrr r!r r r Z getLogger__name__r rr$r*ArgumentParserr, add_argumentr/rrrrsF      site-packages/pip/_vendor/pep517/__pycache__/_in_process.cpython-38.opt-1.pyc000064400000016405151464452720022653 0ustar00U .ep@s0dZddlmZddlmZddlZddlZddlmZddlZddl Z ddl Z ddl Z ddl Z Gddde ZGdd d e ZGd d d e Zd d ZddZddZddZdZddZddZddZd(ddZddZGdd d e ZGd!d"d"e Zd#d$Zddddd$hZd%d&Ze d'kr,edS))agThis is invoked in a subprocess to call the build backend hooks. It expects: - Command line args: hook_name, control_dir - Environment variables: PEP517_BUILD_BACKEND=entry.point:spec PEP517_BACKEND_PATH=paths (separated with os.pathsep) - control_dir/input.json: - {"kwargs": {...}} Results: - control_dir/output.json - {"return_val": ...} )glob) import_moduleN)joinc@seZdZdZddZdS)BackendUnavailablez&Raised if we cannot import the backendcCs ||_dSN tracebackselfrr B/usr/lib/python3.8/site-packages/pip/_vendor/pep517/_in_process.py__init__szBackendUnavailable.__init__N__name__ __module__ __qualname____doc__r r r r r rsrc@seZdZdZddZdS)BackendInvalidz Raised if the backend is invalidcCs ||_dSr)message)r rr r r r %szBackendInvalid.__init__Nrr r r r r#src@seZdZdZdS) HookMissingzARaised if a hook is missing and we are not executing the fallbackNrrrrr r r r r)srcCs<tjtj|}tjtj|}tj||g|kS)z5Test if a file is located within the given directory.)ospathnormcaseabspath commonprefix)filenameZ directoryr r r contained_in-srcstjd}|r*|tj}|tjdd<tjd}|d\}}}z t|Wn t k rpt t YnX|rt fdd|Dstd|r|d D]}t|qS) zFind and load the build backendZPEP517_BACKEND_PATHNrZPEP517_BUILD_BACKEND:c3s|]}tj|VqdSr)r__file__).0robjr r Dsz!_build_backend..z(Backend was not loaded from backend-path.)renvirongetsplitpathsepsysr partitionr ImportErrorrr format_excanyrgetattr)Z backend_pathZextra_pathitemsZepZmod_path_Zobj_pathZ path_partr r!r _build_backend4s&      r0cCs6t}z |j}Wntk r(gYSX||SdSzfInvoke the optional get_requires_for_build_wheel hook Returns [] if the hook is not defined. N)r0get_requires_for_build_wheelAttributeErrorconfig_settingsbackendhookr r r r2Ps   r2cCsJt}z |j}Wn*tk r:|s*tt|||YSX|||SdS)zInvoke optional prepare_metadata_for_build_wheel Implements a fallback by building a wheel if the hook isn't defined, unless _allow_fallback is False in which case HookMissing is raised. N)r0 prepare_metadata_for_build_wheelr3r_get_wheel_metadata_from_wheel)metadata_directoryr5Z_allow_fallbackr6r7r r r r8^s  r8ZPEP517_ALREADY_BUILT_WHEELcCs@g}|D]}td|}|r ||q |r4|StddS)z6Identify the .dist-info folder inside a wheel ZipFile.z[^/\\]+-[^/\\]+\.dist-info/z#No .dist-info folder found in wheelN)Znamelistrematchappend Exception)Zwhl_zipresrmr r r _dist_info_filests   rAc Csddlm}|||}ttj|tdW5QRXtj||}||}t|}|j ||dW5QRX|d ddS)zBuild a wheel and extract the metadata from it. Fallback for when the build backend does not define the 'get_wheel_metadata' hook. r)ZipFilewb)rmembers/) ZzipfilerB build_wheelopenrrrWHEEL_BUILT_MARKERrAZ extractallr')r6r:r5rBZ whl_basenameZwhl_fileZzipfZ dist_infor r r r9s    r9cCsl|sdStj|}tjt|ts*dSttj|d}|sLtddSt |dkrdtddS|dS)zHCheck for a wheel already built during the get_wheel_metadata hook. Nz*.whlz+Found wheel built marker, but no .whl fileszHFound multiple .whl files; unspecified behaviour. Will call build_wheel.r) rrdirnameisfilepjoinrHrrprintlen)r:Zmetadata_parentZ whl_filesr r r _find_already_built_wheels  rOcCs4t|}|r$t||tj|St|||S)zInvoke the mandatory build_wheel hook. If a wheel was already built in the prepare_metadata_for_build_wheel fallback, this will copy it rather than rebuilding the wheel. )rOshutilZcopy2rrbasenamer0rF)Zwheel_directoryr5r:Z prebuilt_whlr r r rFs   rFcCs6t}z |j}Wntk r(gYSX||SdSr1)r0get_requires_for_build_sdistr3r4r r r rRs   rRc@seZdZdZdS)_DummyExceptionz(Nothing should ever raise this exceptionNrr r r r rSsrSc@seZdZdZddZdS)GotUnsupportedOperationz9For internal use when backend raises UnsupportedOperationcCs ||_dSrrr r r r r sz GotUnsupportedOperation.__init__Nrr r r r rTsrTc CsBt}z|||WSt|dtk r<ttYnXdS)z&Invoke the mandatory build_sdist hook.UnsupportedOperationN)r0 build_sdistr.rSrTrr,)Zsdist_directoryr5r6r r r rVs rVc CsNttjdkrtdtjd}tjd}|tkrBtd|t|}tt|d}ddd}z|f|d |d <Wnt k r}zd |d <|j |d <W5d}~XYnt k r}zd |d<|j |d<W5d}~XYnRt k r}zd |d<|j |d <W5d}~XYntk r2d |d<YnXtj|t|ddddS)Nz"Needs args: hook_name, control_dirrIzUnknown hook: %sz input.jsonF) unsupported return_valkwargsrZTZ no_backendrZbackend_invalidZ backend_errorrYZ hook_missingz output.json)indent)rNr)argvexit HOOK_NAMESglobalscompatZ read_jsonrLrrrrrTrZ write_json)Z hook_nameZ control_dirr7Z hook_inputZjson_outer r r mains.     rc__main__)N)!rr importlibrrZos.pathrrLr;rPr)rrar>rrrrr0r2r8rHrAr9rOrFrRrSrTrVr_rcrr r r r sF        site-packages/pip/_vendor/pep517/__pycache__/wrappers.cpython-38.pyc000064400000023636151464452720021260 0ustar00U .eg)@s ddlZddlmZddlZddlmZmZmZddl Z ddl m Z m Z m Z ddlZddlmZddlmZeeeedZed d ZGd d d eZGd ddeZGdddeZGdddeZdddZdddZddZGdddeZGdddej Z!dS)N)contextmanager)dirnameabspathjoin) check_call check_outputSTDOUT)mkdtemp)compatz_in_process.pyc cs"t}z |VW5t|XdSN)r shutilZrmtree)tdr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/wrappers.pytempdirs rc@seZdZdZddZdS)BackendUnavailablezEWill be raised if the backend cannot be imported in the hook process.cCs ||_dSr  tracebackselfrrrr__init__szBackendUnavailable.__init__N__name__ __module__ __qualname____doc__rrrrrrsrc@seZdZdZddZdS)BackendInvalidz)Will be raised if the backend is invalid.cCs||_||_||_dSr  backend_name backend_pathmessage)rrr r!rrrr szBackendInvalid.__init__Nrrrrrrsrcs eZdZdZfddZZS) HookMissingz Will be raised on missing hooks.cstt||||_dSr )superr"r hook_name)rr$ __class__rrr(szHookMissing.__init__)rrrrr __classcell__rrr%rr"&sr"c@seZdZdZddZdS)UnsupportedOperationzDMay be raised by build_sdist if the backend indicates that it can't.cCs ||_dSr rrrrrr/szUnsupportedOperation.__init__Nrrrrrr(-sr(cCs*tj}|r||t|||ddS)z5The default method of calling the wrapper subprocess.)cwdenvN)osenvironcopyupdatercmdr) extra_environr*rrrdefault_subprocess_runner3s  r2cCs,tj}|r||t|||tddS)zDA method of calling the wrapper subprocess while suppressing output.)r)r*stderrN)r+r,r-r.rrr/rrrquiet_subprocess_runner<s  r4cCsntj|rtdtj|}tjtj||}tj|}tj|}tj||g|krjtd|S)zNormalise and check a backend path. Ensure that the requested backend path is specified as a relative path, and resolves to a location under the given source tree. Return an absolute version of the requested path. zpaths must be relativez paths must be inside source tree) r+pathisabs ValueErrorrnormpathrnormcase commonprefix)Z source_treeZ requestedZ abs_sourceZ abs_requestedZ norm_sourceZnorm_requestedrrrnorm_and_checkEs    r;c@s`eZdZdZdddZeddZdddZdd d Zdd d Z dddZ dddZ ddZ dS)Pep517HookCalleraeA wrapper around a source directory to be built with a PEP 517 backend. source_dir : The path to the source directory, containing pyproject.toml. build_backend : The build backend spec, as per PEP 517, from pyproject.toml. backend_path : The backend path, as per PEP 517, from pyproject.toml. runner : A callable that invokes the wrapper subprocess. The 'runner', if provided, must expect the following: cmd : a list of strings representing the command and arguments to execute, as would be passed to e.g. 'subprocess.check_call'. cwd : a string representing the working directory that must be used for the subprocess. Corresponds to the provided source_dir. extra_environ : a dict mapping environment variable names to values which must be set for the subprocess execution. NcsB|dkr t}t|_|_|r2fdd|D}|_|_dS)Ncsg|]}tj|qSr)r; source_dir).0prrr {sz-Pep517HookCaller.__init__..)r2rr= build_backendr _subprocess_runner)rr=rBr runnerrr@rrns  zPep517HookCaller.__init__ccs|j}||_dV||_dS)z\A context manager for temporarily overriding the default subprocess runner. N)rC)rrDprevrrrsubprocess_runnersz"Pep517HookCaller.subprocess_runnercCs|dd|iS)aGIdentify packages required for building a wheel Returns a list of dependency specifications, e.g.: ["wheel >= 0.25", "setuptools"] This does not include requirements specified in pyproject.toml. It returns the result of calling the equivalently named hook in a subprocess. get_requires_for_build_wheelconfig_settings _call_hookrrHrrrrGs z-Pep517HookCaller.get_requires_for_build_wheelTcCs|dt|||dS)aqPrepare a *.dist-info folder with metadata for this project. Returns the name of the newly created folder. If the build backend defines a hook with this name, it will be called in a subprocess. If not, the backend will be asked to build a wheel, and the dist-info extracted from that (unless _allow_fallback is False). prepare_metadata_for_build_wheel)metadata_directoryrH_allow_fallbackrJr)rrMrHrNrrrrLs z1Pep517HookCaller.prepare_metadata_for_build_wheelcCs(|dk rt|}|dt|||dS)avBuild a wheel from this project. Returns the name of the newly created file. In general, this will call the 'build_wheel' hook in the backend. However, if that was previously called by 'prepare_metadata_for_build_wheel', and the same metadata_directory is used, the previously built wheel will be copied to wheel_directory. N build_wheel)wheel_directoryrHrM)rrJ)rrQrHrMrrrrPs zPep517HookCaller.build_wheelcCs|dd|iS)a<Identify packages required for building a wheel Returns a list of dependency specifications, e.g.: ["setuptools >= 26"] This does not include requirements specified in pyproject.toml. It returns the result of calling the equivalently named hook in a subprocess. get_requires_for_build_sdistrHrIrKrrrrRs z-Pep517HookCaller.get_requires_for_build_sdistcCs|dt||dS)zBuild an sdist from this project. Returns the name of the newly created file. This calls the 'build_sdist' backend hook in a subprocess. build_sdist)sdist_directoryrHrO)rrTrHrrrrSszPep517HookCaller.build_sdistc Cs@tjddkr|jd}n|j}d|i}|jrbtj|j}tjddkrZ|t}||d<t }d|i}t j |t |ddd|j tjt||g|j|d t t |d }|d rt|d d |drt|d d |drt|j|j|dd d|dr"t||dW5QRSQRXdS)NrASCIIZPEP517_BUILD_BACKENDZPEP517_BACKEND_PATHkwargsz input.json)indent)r)r1z output.jsonZ unsupportedrZ no_backendZbackend_invalidZ backend_errorrZ hook_missingZ return_val)sys version_inforBencoder r+pathseprgetfilesystemencodingrr Z write_jsonpjoinrC executable_in_proc_scriptr=Z read_jsongetr(rrr") rr$rWrBr1r rZ hook_inputdatarrrrJsB      zPep517HookCaller._call_hook)NN)N)NT)NN)N)N) rrrrrrrFrGrLrPrRrSrJrrrrr<]s       r<c@s<eZdZdZddZddZeddZdd Zd d Z d S) LoggerWrapperzd Read messages from a pipe and redirect them to a logger (see python's logging module). cCsHtj|d|_||_||_t\|_|_ t |j|_ | dS)NT) threadingThreadrZdaemonloggerlevelr+pipeZfd_readfd_writefdopenreaderstart)rrgrhrrrrs zLoggerWrapper.__init__cCs|jSr )rjr@rrrfilenoszLoggerWrapper.filenocCs|tjr|ddS|S)N)endswithr+linesep)msgrrrremove_newline!szLoggerWrapper.remove_newlinecCs |jD]}|||qdSr )rl_writers)rlinerrrrun%s zLoggerWrapper.runcCs|j|j|dSr )rglogrh)rr!rrrrt)szLoggerWrapper._writeN) rrrrrrn staticmethodrsrvrtrrrrrd s  rd)NN)NN)"re contextlibrr+Zos.pathrrrr_r subprocessrrrrZZtempfiler rYr __file__rar Exceptionrrr"r(r2r4r;objectr<rfrdrrrrs*     /site-packages/pip/_vendor/pep517/__pycache__/envbuild.cpython-38.opt-1.pyc000064400000010442151464452720022153 0ustar00U .e@sdZddlZddlZddlZddlZddlmZddlZddlm Z ddl m Z ddl m Z mZeeZdd ZGd d d eZdd d ZdddZdS)zIBuild wheels/sdists by installing build deps to a temporary environment. N) check_call) get_paths)mkdtemp)Pep517HookCaller LoggerWrapperc CsHttj|d}t|}W5QRX|d}|d|d|dfS)Nzpyproject.tomlz build-systemrequiresz build-backendz backend-path)openospathjointomlloadget) source_dirfZpyproject_dataZbuildsysr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/envbuild.py_load_pyprojectsrc@s6eZdZdZdZd ddZddZdd Zd d ZdS) BuildEnvironmentzContext manager to install build deps in a simple temporary environment Based on code I wrote for pip, which is MIT licensed. NTcCs ||_dS)N)_cleanup)selfZcleanuprrr__init__9szBuildEnvironment.__init__cCstdd|_td|jtjdd|_tjdd|_tj dkrHdnd}t ||j|jdd }|d }|jr|tj |jtjd<n|tj tj tjd<|d |d kr|d }n|d tj |d }|jr|tj |jtjd<n |tjd<|S) Nzpep517-build-env-)prefixzTemporary build environment: %sPATH PYTHONPATHntZ posix_prefix)baseZplatbase)varsscriptsZpurelibZplatlib) rr loginfor environr save_pathsave_pythonpathnamerpathsepdefpath)rZinstall_schemeZ install_dirsrZlib_dirsrrr __enter__<s0      zBuildEnvironment.__enter__cCsT|sdStd|tjddddd|jgt|}t|tttj tttj ddS) zAInstall dependencies into this env by calling pip in a subprocessNzCalling pip to install %sz-mZpipZinstallz--ignore-installedz--prefix)stdoutstderr) r r!sys executabler listrrloggingINFOZERROR)rreqscmdrrr pip_install\s$   zBuildEnvironment.pip_installcCs~|jo|jdk otj|j}|r.t|j|jdkrHtjddn |jtjd<|j dkrntjddn |j tjd<dS)Nrr) rr r isdirshutilZrmtreer#r"popr$)rexc_typeZexc_valZexc_tbZ needs_cleanuprrr__exit__js     zBuildEnvironment.__exit__)T) __name__ __module__ __qualname____doc__r rr(r2r7rrrrrs   rc Csn|dkr i}t|\}}}t|||}t8}||||}|||||W5QRSQRXdS)amBuild a wheel from a source directory using PEP 517 hooks. :param str source_dir: Source directory containing pyproject.toml :param str wheel_dir: Target directory to create wheel in :param dict config_settings: Options to pass to build backend This is a blocking function which will run pip in a subprocess to install build requirements. N)rrrr2Zget_requires_for_build_wheel build_wheel) rZ wheel_dirconfig_settingsrbackend backend_pathhooksenvr0rrrr<~s     r<c Csn|dkr i}t|\}}}t|||}t8}||||}|||||W5QRSQRXdS)amBuild an sdist from a source directory using PEP 517 hooks. :param str source_dir: Source directory containing pyproject.toml :param str sdist_dir: Target directory to place sdist in :param dict config_settings: Options to pass to build backend This is a blocking function which will run pip in a subprocess to install build requirements. N)rrrr2Zget_requires_for_build_sdist build_sdist) rZ sdist_dirr=rr>r?r@rAr0rrrrBs     rB)N)N)r;r r.r r4 subprocessrr+Z sysconfigrZtempfilerZwrappersrrZ getLoggerr8r robjectrr<rBrrrrs     a site-packages/pip/_vendor/pep517/__pycache__/check.cpython-38.opt-1.pyc000064400000011240151464452720021415 0ustar00U .e=@sdZddlZddlZddlZddlmZmZddlm Z m Z ddl Z ddl mZddlZddlZddlmZddlZddlmZdd lmZdd lmZeeZd d Zd dZddZdddZ dddddZ!ddZ"edkre dS)zHCheck a project and backend by attempting to build using PEP 517 hooks. N)isfilejoin)TomlDecodeErrorload)CalledProcessError)mkdtemp)enable_colourful_output)BuildEnvironment)Pep517HookCallerc Cst}z||tdWn*tk rLtdYW5QRdSXz|i}td|Wn.tk rtjdddYW5QRdSXz||tdWn*tk rtd YW5QRdSXt}td |zz| |i}td |Wn4tk r@tjd ddYWW5QRdSX| d sltd|WW5QRdSt ||}t|rtd|ntd|WFW5QRdSt|rtdntdWW5QRdSW5t |XW5QRdSQRXdS)N#Installed static build dependencies+Failed to install static build dependenciesFGot build requires: %s'Failure in get_requires_for_build_sdistTexc_info$Installed dynamic build dependencies,Failed to install dynamic build dependencieszTrying to build sdist in %szbuild_sdist returned %rzFailure in build_sdistz.tar.gzz*Filename %s doesn't have .tar.gz extensionOutput file %s existsOutput file %s does not existzOutput file is a tar filezOutput file is not a tar file)r pip_installloginforerrorZget_requires_for_build_sdist ExceptionrshutilrmtreeZ build_sdistendswithpjoinrtarfileZ is_tarfilehooksZbuild_sys_requiresenvZreqsZtdfilenamepathr%s6      98" site-packages/pip/_vendor/pep517/__pycache__/build.cpython-38.pyc000064400000006406151464452720020510 0ustar00U .e @sdZddlZddlZddlZddlZddlZddlmZddlm Z ddl m Z m Z ddl mZeeZdd Zd d Zd d ZddZd"ddZeZejdddejdddddejdddddejdddddd Zed!kreedS)#z%Build a project using PEP 517 hooks. N)BuildEnvironment)Pep517HookCaller)tempdirmkdir_p)FileNotFoundErrorcCs4ddh}|t|ks0dj|t|d}t|dS)z7 Ensure build system has the requisite fields. requires build-backendz"Missing required fields: {missing})ZmissingN)setformat ValueError)systemZrequiredmessagerDestination in which to save the builds relative to source dircCsPttd|js|jsdnd|js$|js(dndf}|D]}t|j||jq6dS)NZsdistr)listfilterr,Zbinaryr*rZout_dir)argsZdistsr(rrrmainps r6__main__)NN)__doc__argparseZloggingrrr$ZenvbuildrZwrappersrZdirtoolsrrcompatrZ getLogger__name__r!rrrr-r*ArgumentParserparser add_argumentr6 parse_argsrrrrsP        site-packages/pip/_vendor/pep517/__pycache__/meta.cpython-38.opt-1.pyc000064400000005406151464452720021275 0ustar00U .e @sBdZddlZddlZddlZddlZddlZzddlmZWne k rXddl ZYnXzddl m Z Wn e k rddl m Z YnXddlmZddlmZmZddlmZmZmZddlmZmZmZeeZd d Zdd d ZefddZddZ e!Z"e"j#ddde"j#ddddddZ$edkr>e$dS)z2Build metadata for a project using PEP 517 hooks. N)Path)BuildEnvironment)Pep517HookCallerquiet_subprocess_runner)tempdirmkdir_pdir_to_zipfile)validate_system load_system compat_systemc Cs|i}td|||tdtJ}td|||i}tj||}t |tj|tj |W5QRXdS)NzGot build requires: %sz$Installed dynamic build dependencieszTrying to build metadata in %s) Zget_requires_for_build_wheelloginfo pip_installrZ prepare_metadata_for_build_wheelospathjoinshutilZmovebasename)hooksenvdestZreqsZtdfilenamesourcer;/usr/lib/python3.8/site-packages/pip/_vendor/pep517/meta.py _prep_metas      r.c Cs|p t|}tj||pd}t|t|t||d|d}|t 2t }| |dt |||W5QRXW5QRXdS)NZdistz build-backendz backend-pathZrequires) r rrrrr rgetZsubprocess_runnerrrrr) source_dirrsystemrrrrrbuild)s  r!c Cs0t }||dt|W5QRSQRXdS)N)r)rr )builderout_dirrrr build_as_zip8s r$cCs:tj|}t|}tjt||d}tt|}t |S)z Given a source directory (root) of a package, return an importlib.metadata.Distribution object with metadata build from that package. )rr ) rr expanduserr functoolspartialr!rr$imp_metaZPathDistribution)rootr r"rrrrload>s   r*rz%A directory containing pyproject.toml)helpz --out-dirz-oz>Destination in which to save the builds relative to source dircCst}t|j|jdS)N)parser parse_argsr!rr#)argsrrrmainVsr/__main__)rNN)%__doc__argparseZloggingrrr&importlib.metadataZmetadatar( ImportErrorZimportlib_metadataZzipfilerZzippZenvbuildrZwrappersrrZdirtoolsrrr r!r r r Z getLogger__name__r rr$r*ArgumentParserr, add_argumentr/rrrrsF      site-packages/pip/_vendor/pep517/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000366151464452720022106 0ustar00U .eT@s dZdZdS)z6Wrappers to build Python packages using PEP 517 hooks z0.7.0N)__doc__ __version__rr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/__init__.pyssite-packages/pip/_vendor/pep517/__pycache__/dirtools.cpython-38.pyc000064400000002417151464452720021246 0ustar00U .ei@sZddlZddlZddlZddlZddlZddlZddlZejddZddZ ddZ dS)Nc cs$t}z |VW5t|XdS)z2Create a temporary directory in a context manager.N)tempfileZmkdtempshutilZrmtree)Ztdr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/dirtools.pytempdir s rc OsDztj||WStk r>}z|jtjkr.W5d}~XYnXdS)zWLike `mkdir`, but does not raise an exception if the directory already exists. N)osmkdirOSErrorerrnoZEEXIST)argskwargsexcrrrmkdir_ps  rcCst}t|d}t|D]r\}}}|D]0}tj||}tj||}| |ddq,|D],}tj||}tj||}| ||qbq|S)z0Construct an in-memory zip file for a directory.w/) ioBytesIOzipfileZZipFilerwalkpathjoinrelpathZwritestrwrite)rootbufferZzip_filedirsfilesrZfs_pathZrel_pathrrrdir_to_zipfiles r) rr contextlibrrr rcontextmanagerrrrrrrrs  site-packages/pip/_vendor/pep517/__pycache__/envbuild.cpython-38.pyc000064400000010442151464452720021214 0ustar00U .e@sdZddlZddlZddlZddlZddlmZddlZddlm Z ddl m Z ddl m Z mZeeZdd ZGd d d eZdd d ZdddZdS)zIBuild wheels/sdists by installing build deps to a temporary environment. N) check_call) get_paths)mkdtemp)Pep517HookCaller LoggerWrapperc CsHttj|d}t|}W5QRX|d}|d|d|dfS)Nzpyproject.tomlz build-systemrequiresz build-backendz backend-path)openospathjointomlloadget) source_dirfZpyproject_dataZbuildsysr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/envbuild.py_load_pyprojectsrc@s6eZdZdZdZd ddZddZdd Zd d ZdS) BuildEnvironmentzContext manager to install build deps in a simple temporary environment Based on code I wrote for pip, which is MIT licensed. NTcCs ||_dS)N)_cleanup)selfZcleanuprrr__init__9szBuildEnvironment.__init__cCstdd|_td|jtjdd|_tjdd|_tj dkrHdnd}t ||j|jdd }|d }|jr|tj |jtjd<n|tj tj tjd<|d |d kr|d }n|d tj |d }|jr|tj |jtjd<n |tjd<|S) Nzpep517-build-env-)prefixzTemporary build environment: %sPATH PYTHONPATHntZ posix_prefix)baseZplatbase)varsscriptsZpurelibZplatlib) rr loginfor environr save_pathsave_pythonpathnamerpathsepdefpath)rZinstall_schemeZ install_dirsrZlib_dirsrrr __enter__<s0      zBuildEnvironment.__enter__cCsT|sdStd|tjddddd|jgt|}t|tttj tttj ddS) zAInstall dependencies into this env by calling pip in a subprocessNzCalling pip to install %sz-mZpipZinstallz--ignore-installedz--prefix)stdoutstderr) r r!sys executabler listrrloggingINFOZERROR)rreqscmdrrr pip_install\s$   zBuildEnvironment.pip_installcCs~|jo|jdk otj|j}|r.t|j|jdkrHtjddn |jtjd<|j dkrntjddn |j tjd<dS)Nrr) rr r isdirshutilZrmtreer#r"popr$)rexc_typeZexc_valZexc_tbZ needs_cleanuprrr__exit__js     zBuildEnvironment.__exit__)T) __name__ __module__ __qualname____doc__r rr(r2r7rrrrrs   rc Csn|dkr i}t|\}}}t|||}t8}||||}|||||W5QRSQRXdS)amBuild a wheel from a source directory using PEP 517 hooks. :param str source_dir: Source directory containing pyproject.toml :param str wheel_dir: Target directory to create wheel in :param dict config_settings: Options to pass to build backend This is a blocking function which will run pip in a subprocess to install build requirements. N)rrrr2Zget_requires_for_build_wheel build_wheel) rZ wheel_dirconfig_settingsrbackend backend_pathhooksenvr0rrrr<~s     r<c Csn|dkr i}t|\}}}t|||}t8}||||}|||||W5QRSQRXdS)amBuild an sdist from a source directory using PEP 517 hooks. :param str source_dir: Source directory containing pyproject.toml :param str sdist_dir: Target directory to place sdist in :param dict config_settings: Options to pass to build backend This is a blocking function which will run pip in a subprocess to install build requirements. N)rrrr2Zget_requires_for_build_sdist build_sdist) rZ sdist_dirr=rr>r?r@rAr0rrrrBs     rB)N)N)r;r r.r r4 subprocessrr+Z sysconfigrZtempfilerZwrappersrrZ getLoggerr8r robjectrr<rBrrrrs     a site-packages/pip/_vendor/pep517/__pycache__/dirtools.cpython-38.opt-1.pyc000064400000002417151464452720022205 0ustar00U .ei@sZddlZddlZddlZddlZddlZddlZddlZejddZddZ ddZ dS)Nc cs$t}z |VW5t|XdS)z2Create a temporary directory in a context manager.N)tempfileZmkdtempshutilZrmtree)Ztdr?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/dirtools.pytempdir s rc OsDztj||WStk r>}z|jtjkr.W5d}~XYnXdS)zWLike `mkdir`, but does not raise an exception if the directory already exists. N)osmkdirOSErrorerrnoZEEXIST)argskwargsexcrrrmkdir_ps  rcCst}t|d}t|D]r\}}}|D]0}tj||}tj||}| |ddq,|D],}tj||}tj||}| ||qbq|S)z0Construct an in-memory zip file for a directory.w/) ioBytesIOzipfileZZipFilerwalkpathjoinrelpathZwritestrwrite)rootbufferZzip_filedirsfilesrZfs_pathZrel_pathrrrdir_to_zipfiles r) rr contextlibrrr rcontextmanagerrrrrrrrs  site-packages/pip/_vendor/pep517/__pycache__/compat.cpython-38.opt-1.pyc000064400000001765151464452720021636 0ustar00U .e @sjdZddlZddlZejddkr4ddZddZnddZd dZzeZWnek rdeZYnXdS) zPython 2/3 compatibilityNc Ks.t|ddd}tj||f|W5QRXdS)Nwutf-8encodingopenjsondumpobjpathkwargsfr=/usr/lib/python3.8/site-packages/pip/_vendor/pep517/compat.py write_json src Cs0t|ddd}t|W5QRSQRXdS)Nrrrrr loadr rrrr read_jsonsrc Ks2t|d}tj||fddi|W5QRXdS)Nwbrrrr rrrrs c Cs,t|d}t|W5QRSQRXdS)Nrbrrrrrrs ) __doc__r sys version_inforrFileNotFoundError NameErrorIOErrorrrrrs site-packages/pip/_vendor/pep517/__pycache__/colorlog.cpython-38.opt-1.pyc000064400000005544151464452720022172 0ustar00U .e@sfdZddlZddlZz ddlZWnek r8dZYnXddZGdddejZejfddZ dS) zONicer log formatting with colours. Code copied from Tornado, Apache licensed. NcCsVd}trRttjdrRtjrRzttddkr:d}Wntk rPYnX|S)NFisattyZcolorsrT)curseshasattrsysstderrrZ setuptermZtigetnum Exception)colorr ?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/colorlog.py_stderr_supports_colorsr c@s>eZdZdZejdejdejdejdiZ d ddZ d d Z dS) LogFormatterz&Log formatter with colour support TNcCstjj||di|_|rtrtdp6tdp6d}dtjkrNdkr\nn t |d}|j D] \}}t t ||d|j|<qft tdd|_ t}|d |_tn d|_ d |_d S) aO :arg bool color: Enables color support. :arg string fmt: Log message format. It will be applied to the attributes dict of log records. The text between ``%(color)s`` and ``%(end_color)s`` will be colored depending on the level if color support is on. :arg dict colors: color mappings from logging level to terminal color code :arg string datefmt: Datetime format. Used for formatting ``(asctime)`` placeholder in ``prefix_fmt``. .. versionchanged:: 3.2 Added ``fmt`` and ``datefmt`` arguments. )datefmtZsetafZsetf)rr)rr rasciiZsgr0rFN)logging Formatter__init___colorsr rZtigetstrr version_infostrDEFAULT_COLORSitemsZtparm_normalZinitscrZgetmaxyx termwidthZendwin)selfrrZfg_colorlevelnocodeZscrr r r r2s*      zLogFormatter.__init__cCst|j}dj|jd|jd}|t||jkrJd|j|t|}nd}|j|jkrn|j|j}|j}nd}}|j||||S)Nz{initial}-{name}r)initialname z r) lenmessageformatZ levelnamer"rrrr)rrecordZmlenZ right_textZspaceZ start_colorZ end_colorr r r formatMessage]s    zLogFormatter.formatMessage)TN) __name__ __module__ __qualname____doc__rINFOZWARNINGZERRORZCRITICALrrr(r r r r r (s +r cCs0t}|ttj|tj|dS)N)rZ StreamHandlerZ setFormatterr rootZ addHandlerZsetLevel)levelZhandlerr r r enable_colourful_outputos  r0) r,rrr ImportErrorr rr r-r0r r r r s   Gsite-packages/pip/_vendor/pep517/__pycache__/colorlog.cpython-38.pyc000064400000005544151464452720021233 0ustar00U .e@sfdZddlZddlZz ddlZWnek r8dZYnXddZGdddejZejfddZ dS) zONicer log formatting with colours. Code copied from Tornado, Apache licensed. NcCsVd}trRttjdrRtjrRzttddkr:d}Wntk rPYnX|S)NFisattyZcolorsrT)curseshasattrsysstderrrZ setuptermZtigetnum Exception)colorr ?/usr/lib/python3.8/site-packages/pip/_vendor/pep517/colorlog.py_stderr_supports_colorsr c@s>eZdZdZejdejdejdejdiZ d ddZ d d Z dS) LogFormatterz&Log formatter with colour support TNcCstjj||di|_|rtrtdp6tdp6d}dtjkrNdkr\nn t |d}|j D] \}}t t ||d|j|<qft tdd|_ t}|d |_tn d|_ d |_d S) aO :arg bool color: Enables color support. :arg string fmt: Log message format. It will be applied to the attributes dict of log records. The text between ``%(color)s`` and ``%(end_color)s`` will be colored depending on the level if color support is on. :arg dict colors: color mappings from logging level to terminal color code :arg string datefmt: Datetime format. Used for formatting ``(asctime)`` placeholder in ``prefix_fmt``. .. versionchanged:: 3.2 Added ``fmt`` and ``datefmt`` arguments. )datefmtZsetafZsetf)rr)rr rasciiZsgr0rFN)logging Formatter__init___colorsr rZtigetstrr version_infostrDEFAULT_COLORSitemsZtparm_normalZinitscrZgetmaxyx termwidthZendwin)selfrrZfg_colorlevelnocodeZscrr r r r2s*      zLogFormatter.__init__cCst|j}dj|jd|jd}|t||jkrJd|j|t|}nd}|j|jkrn|j|j}|j}nd}}|j||||S)Nz{initial}-{name}r)initialname z r) lenmessageformatZ levelnamer"rrrr)rrecordZmlenZ right_textZspaceZ start_colorZ end_colorr r r formatMessage]s    zLogFormatter.formatMessage)TN) __name__ __module__ __qualname____doc__rINFOZWARNINGZERRORZCRITICALrrr(r r r r r (s +r cCs0t}|ttj|tj|dS)N)rZ StreamHandlerZ setFormatterr rootZ addHandlerZsetLevel)levelZhandlerr r r enable_colourful_outputos  r0) r,rrr ImportErrorr rr r-r0r r r r s   Gsite-packages/pip/_vendor/pep517/__pycache__/build.cpython-38.opt-1.pyc000064400000006406151464452720021447 0ustar00U .e @sdZddlZddlZddlZddlZddlZddlmZddlm Z ddl m Z m Z ddl mZeeZdd Zd d Zd d ZddZd"ddZeZejdddejdddddejdddddejdddddd Zed!kreedS)#z%Build a project using PEP 517 hooks. N)BuildEnvironment)Pep517HookCaller)tempdirmkdir_p)FileNotFoundErrorcCs4ddh}|t|ks0dj|t|d}t|dS)z7 Ensure build system has the requisite fields. requires build-backendz"Missing required fields: {missing})ZmissingN)setformat ValueError)systemZrequiredmessagerDestination in which to save the builds relative to source dircCsPttd|js|jsdnd|js$|js(dndf}|D]}t|j||jq6dS)NZsdistr)listfilterr,Zbinaryr*rZout_dir)argsZdistsr(rrrmainps r6__main__)NN)__doc__argparseZloggingrrr$ZenvbuildrZwrappersrZdirtoolsrrcompatrZ getLogger__name__r!rrrr-r*ArgumentParserparser add_argumentr6 parse_argsrrrrsP        site-packages/pip/_vendor/pep517/envbuild.py000064400000013610151464452720014726 0ustar00"""Build wheels/sdists by installing build deps to a temporary environment. """ import os import logging import toml import shutil from subprocess import check_call import sys from sysconfig import get_paths from tempfile import mkdtemp from .wrappers import Pep517HookCaller, LoggerWrapper log = logging.getLogger(__name__) def _load_pyproject(source_dir): with open(os.path.join(source_dir, 'pyproject.toml')) as f: pyproject_data = toml.load(f) buildsys = pyproject_data['build-system'] return ( buildsys['requires'], buildsys['build-backend'], buildsys.get('backend-path'), ) class BuildEnvironment(object): """Context manager to install build deps in a simple temporary environment Based on code I wrote for pip, which is MIT licensed. """ # Copyright (c) 2008-2016 The pip developers (see AUTHORS.txt file) # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be # included in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, # EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF # MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE # LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION # OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. path = None def __init__(self, cleanup=True): self._cleanup = cleanup def __enter__(self): self.path = mkdtemp(prefix='pep517-build-env-') log.info('Temporary build environment: %s', self.path) self.save_path = os.environ.get('PATH', None) self.save_pythonpath = os.environ.get('PYTHONPATH', None) install_scheme = 'nt' if (os.name == 'nt') else 'posix_prefix' install_dirs = get_paths(install_scheme, vars={ 'base': self.path, 'platbase': self.path, }) scripts = install_dirs['scripts'] if self.save_path: os.environ['PATH'] = scripts + os.pathsep + self.save_path else: os.environ['PATH'] = scripts + os.pathsep + os.defpath if install_dirs['purelib'] == install_dirs['platlib']: lib_dirs = install_dirs['purelib'] else: lib_dirs = install_dirs['purelib'] + os.pathsep + \ install_dirs['platlib'] if self.save_pythonpath: os.environ['PYTHONPATH'] = lib_dirs + os.pathsep + \ self.save_pythonpath else: os.environ['PYTHONPATH'] = lib_dirs return self def pip_install(self, reqs): """Install dependencies into this env by calling pip in a subprocess""" if not reqs: return log.info('Calling pip to install %s', reqs) cmd = [ sys.executable, '-m', 'pip', 'install', '--ignore-installed', '--prefix', self.path] + list(reqs) check_call( cmd, stdout=LoggerWrapper(log, logging.INFO), stderr=LoggerWrapper(log, logging.ERROR), ) def __exit__(self, exc_type, exc_val, exc_tb): needs_cleanup = ( self._cleanup and self.path is not None and os.path.isdir(self.path) ) if needs_cleanup: shutil.rmtree(self.path) if self.save_path is None: os.environ.pop('PATH', None) else: os.environ['PATH'] = self.save_path if self.save_pythonpath is None: os.environ.pop('PYTHONPATH', None) else: os.environ['PYTHONPATH'] = self.save_pythonpath def build_wheel(source_dir, wheel_dir, config_settings=None): """Build a wheel from a source directory using PEP 517 hooks. :param str source_dir: Source directory containing pyproject.toml :param str wheel_dir: Target directory to create wheel in :param dict config_settings: Options to pass to build backend This is a blocking function which will run pip in a subprocess to install build requirements. """ if config_settings is None: config_settings = {} requires, backend, backend_path = _load_pyproject(source_dir) hooks = Pep517HookCaller(source_dir, backend, backend_path) with BuildEnvironment() as env: env.pip_install(requires) reqs = hooks.get_requires_for_build_wheel(config_settings) env.pip_install(reqs) return hooks.build_wheel(wheel_dir, config_settings) def build_sdist(source_dir, sdist_dir, config_settings=None): """Build an sdist from a source directory using PEP 517 hooks. :param str source_dir: Source directory containing pyproject.toml :param str sdist_dir: Target directory to place sdist in :param dict config_settings: Options to pass to build backend This is a blocking function which will run pip in a subprocess to install build requirements. """ if config_settings is None: config_settings = {} requires, backend, backend_path = _load_pyproject(source_dir) hooks = Pep517HookCaller(source_dir, backend, backend_path) with BuildEnvironment() as env: env.pip_install(requires) reqs = hooks.get_requires_for_build_sdist(config_settings) env.pip_install(reqs) return hooks.build_sdist(sdist_dir, config_settings) site-packages/pip/_vendor/pep517/wrappers.py000064400000024547151464452720014774 0ustar00import threading from contextlib import contextmanager import os from os.path import dirname, abspath, join as pjoin import shutil from subprocess import check_call, check_output, STDOUT import sys from tempfile import mkdtemp from . import compat _in_proc_script = pjoin(dirname(abspath(__file__)), '_in_process.py') @contextmanager def tempdir(): td = mkdtemp() try: yield td finally: shutil.rmtree(td) class BackendUnavailable(Exception): """Will be raised if the backend cannot be imported in the hook process.""" def __init__(self, traceback): self.traceback = traceback class BackendInvalid(Exception): """Will be raised if the backend is invalid.""" def __init__(self, backend_name, backend_path, message): self.backend_name = backend_name self.backend_path = backend_path self.message = message class HookMissing(Exception): """Will be raised on missing hooks.""" def __init__(self, hook_name): super(HookMissing, self).__init__(hook_name) self.hook_name = hook_name class UnsupportedOperation(Exception): """May be raised by build_sdist if the backend indicates that it can't.""" def __init__(self, traceback): self.traceback = traceback def default_subprocess_runner(cmd, cwd=None, extra_environ=None): """The default method of calling the wrapper subprocess.""" env = os.environ.copy() if extra_environ: env.update(extra_environ) check_call(cmd, cwd=cwd, env=env) def quiet_subprocess_runner(cmd, cwd=None, extra_environ=None): """A method of calling the wrapper subprocess while suppressing output.""" env = os.environ.copy() if extra_environ: env.update(extra_environ) check_output(cmd, cwd=cwd, env=env, stderr=STDOUT) def norm_and_check(source_tree, requested): """Normalise and check a backend path. Ensure that the requested backend path is specified as a relative path, and resolves to a location under the given source tree. Return an absolute version of the requested path. """ if os.path.isabs(requested): raise ValueError("paths must be relative") abs_source = os.path.abspath(source_tree) abs_requested = os.path.normpath(os.path.join(abs_source, requested)) # We have to use commonprefix for Python 2.7 compatibility. So we # normalise case to avoid problems because commonprefix is a character # based comparison :-( norm_source = os.path.normcase(abs_source) norm_requested = os.path.normcase(abs_requested) if os.path.commonprefix([norm_source, norm_requested]) != norm_source: raise ValueError("paths must be inside source tree") return abs_requested class Pep517HookCaller(object): """A wrapper around a source directory to be built with a PEP 517 backend. source_dir : The path to the source directory, containing pyproject.toml. build_backend : The build backend spec, as per PEP 517, from pyproject.toml. backend_path : The backend path, as per PEP 517, from pyproject.toml. runner : A callable that invokes the wrapper subprocess. The 'runner', if provided, must expect the following: cmd : a list of strings representing the command and arguments to execute, as would be passed to e.g. 'subprocess.check_call'. cwd : a string representing the working directory that must be used for the subprocess. Corresponds to the provided source_dir. extra_environ : a dict mapping environment variable names to values which must be set for the subprocess execution. """ def __init__( self, source_dir, build_backend, backend_path=None, runner=None, ): if runner is None: runner = default_subprocess_runner self.source_dir = abspath(source_dir) self.build_backend = build_backend if backend_path: backend_path = [ norm_and_check(self.source_dir, p) for p in backend_path ] self.backend_path = backend_path self._subprocess_runner = runner # TODO: Is this over-engineered? Maybe frontends only need to # set this when creating the wrapper, not on every call. @contextmanager def subprocess_runner(self, runner): """A context manager for temporarily overriding the default subprocess runner. """ prev = self._subprocess_runner self._subprocess_runner = runner yield self._subprocess_runner = prev def get_requires_for_build_wheel(self, config_settings=None): """Identify packages required for building a wheel Returns a list of dependency specifications, e.g.: ["wheel >= 0.25", "setuptools"] This does not include requirements specified in pyproject.toml. It returns the result of calling the equivalently named hook in a subprocess. """ return self._call_hook('get_requires_for_build_wheel', { 'config_settings': config_settings }) def prepare_metadata_for_build_wheel( self, metadata_directory, config_settings=None, _allow_fallback=True): """Prepare a *.dist-info folder with metadata for this project. Returns the name of the newly created folder. If the build backend defines a hook with this name, it will be called in a subprocess. If not, the backend will be asked to build a wheel, and the dist-info extracted from that (unless _allow_fallback is False). """ return self._call_hook('prepare_metadata_for_build_wheel', { 'metadata_directory': abspath(metadata_directory), 'config_settings': config_settings, '_allow_fallback': _allow_fallback, }) def build_wheel( self, wheel_directory, config_settings=None, metadata_directory=None): """Build a wheel from this project. Returns the name of the newly created file. In general, this will call the 'build_wheel' hook in the backend. However, if that was previously called by 'prepare_metadata_for_build_wheel', and the same metadata_directory is used, the previously built wheel will be copied to wheel_directory. """ if metadata_directory is not None: metadata_directory = abspath(metadata_directory) return self._call_hook('build_wheel', { 'wheel_directory': abspath(wheel_directory), 'config_settings': config_settings, 'metadata_directory': metadata_directory, }) def get_requires_for_build_sdist(self, config_settings=None): """Identify packages required for building a wheel Returns a list of dependency specifications, e.g.: ["setuptools >= 26"] This does not include requirements specified in pyproject.toml. It returns the result of calling the equivalently named hook in a subprocess. """ return self._call_hook('get_requires_for_build_sdist', { 'config_settings': config_settings }) def build_sdist(self, sdist_directory, config_settings=None): """Build an sdist from this project. Returns the name of the newly created file. This calls the 'build_sdist' backend hook in a subprocess. """ return self._call_hook('build_sdist', { 'sdist_directory': abspath(sdist_directory), 'config_settings': config_settings, }) def _call_hook(self, hook_name, kwargs): # On Python 2, pytoml returns Unicode values (which is correct) but the # environment passed to check_call needs to contain string values. We # convert here by encoding using ASCII (the backend can only contain # letters, digits and _, . and : characters, and will be used as a # Python identifier, so non-ASCII content is wrong on Python 2 in # any case). # For backend_path, we use sys.getfilesystemencoding. if sys.version_info[0] == 2: build_backend = self.build_backend.encode('ASCII') else: build_backend = self.build_backend extra_environ = {'PEP517_BUILD_BACKEND': build_backend} if self.backend_path: backend_path = os.pathsep.join(self.backend_path) if sys.version_info[0] == 2: backend_path = backend_path.encode(sys.getfilesystemencoding()) extra_environ['PEP517_BACKEND_PATH'] = backend_path with tempdir() as td: hook_input = {'kwargs': kwargs} compat.write_json(hook_input, pjoin(td, 'input.json'), indent=2) # Run the hook in a subprocess self._subprocess_runner( [sys.executable, _in_proc_script, hook_name, td], cwd=self.source_dir, extra_environ=extra_environ ) data = compat.read_json(pjoin(td, 'output.json')) if data.get('unsupported'): raise UnsupportedOperation(data.get('traceback', '')) if data.get('no_backend'): raise BackendUnavailable(data.get('traceback', '')) if data.get('backend_invalid'): raise BackendInvalid( backend_name=self.build_backend, backend_path=self.backend_path, message=data.get('backend_error', '') ) if data.get('hook_missing'): raise HookMissing(hook_name) return data['return_val'] class LoggerWrapper(threading.Thread): """ Read messages from a pipe and redirect them to a logger (see python's logging module). """ def __init__(self, logger, level): threading.Thread.__init__(self) self.daemon = True self.logger = logger self.level = level # create the pipe and reader self.fd_read, self.fd_write = os.pipe() self.reader = os.fdopen(self.fd_read) self.start() def fileno(self): return self.fd_write @staticmethod def remove_newline(msg): return msg[:-1] if msg.endswith(os.linesep) else msg def run(self): for line in self.reader: self._write(self.remove_newline(line)) def _write(self, message): self.logger.log(self.level, message) site-packages/pip/_vendor/pep517/compat.py000064400000001414151464452720014400 0ustar00"""Python 2/3 compatibility""" import json import sys # Handle reading and writing JSON in UTF-8, on Python 3 and 2. if sys.version_info[0] >= 3: # Python 3 def write_json(obj, path, **kwargs): with open(path, 'w', encoding='utf-8') as f: json.dump(obj, f, **kwargs) def read_json(path): with open(path, 'r', encoding='utf-8') as f: return json.load(f) else: # Python 2 def write_json(obj, path, **kwargs): with open(path, 'wb') as f: json.dump(obj, f, encoding='utf-8', **kwargs) def read_json(path): with open(path, 'rb') as f: return json.load(f) # FileNotFoundError try: FileNotFoundError = FileNotFoundError except NameError: FileNotFoundError = IOError site-packages/pip/_vendor/pep517/colorlog.py000064400000010002151464452720014726 0ustar00"""Nicer log formatting with colours. Code copied from Tornado, Apache licensed. """ # Copyright 2012 Facebook # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import logging import sys try: import curses except ImportError: curses = None def _stderr_supports_color(): color = False if curses and hasattr(sys.stderr, 'isatty') and sys.stderr.isatty(): try: curses.setupterm() if curses.tigetnum("colors") > 0: color = True except Exception: pass return color class LogFormatter(logging.Formatter): """Log formatter with colour support """ DEFAULT_COLORS = { logging.INFO: 2, # Green logging.WARNING: 3, # Yellow logging.ERROR: 1, # Red logging.CRITICAL: 1, } def __init__(self, color=True, datefmt=None): r""" :arg bool color: Enables color support. :arg string fmt: Log message format. It will be applied to the attributes dict of log records. The text between ``%(color)s`` and ``%(end_color)s`` will be colored depending on the level if color support is on. :arg dict colors: color mappings from logging level to terminal color code :arg string datefmt: Datetime format. Used for formatting ``(asctime)`` placeholder in ``prefix_fmt``. .. versionchanged:: 3.2 Added ``fmt`` and ``datefmt`` arguments. """ logging.Formatter.__init__(self, datefmt=datefmt) self._colors = {} if color and _stderr_supports_color(): # The curses module has some str/bytes confusion in # python3. Until version 3.2.3, most methods return # bytes, but only accept strings. In addition, we want to # output these strings with the logging module, which # works with unicode strings. The explicit calls to # unicode() below are harmless in python2 but will do the # right conversion in python 3. fg_color = (curses.tigetstr("setaf") or curses.tigetstr("setf") or "") if (3, 0) < sys.version_info < (3, 2, 3): fg_color = str(fg_color, "ascii") for levelno, code in self.DEFAULT_COLORS.items(): self._colors[levelno] = str( curses.tparm(fg_color, code), "ascii") self._normal = str(curses.tigetstr("sgr0"), "ascii") scr = curses.initscr() self.termwidth = scr.getmaxyx()[1] curses.endwin() else: self._normal = '' # Default width is usually 80, but too wide is # worse than too narrow self.termwidth = 70 def formatMessage(self, record): mlen = len(record.message) right_text = '{initial}-{name}'.format(initial=record.levelname[0], name=record.name) if mlen + len(right_text) < self.termwidth: space = ' ' * (self.termwidth - (mlen + len(right_text))) else: space = ' ' if record.levelno in self._colors: start_color = self._colors[record.levelno] end_color = self._normal else: start_color = end_color = '' return record.message + space + start_color + right_text + end_color def enable_colourful_output(level=logging.INFO): handler = logging.StreamHandler() handler.setFormatter(LogFormatter()) logging.root.addHandler(handler) logging.root.setLevel(level) site-packages/pip/_vendor/pep517/build.py000064400000006366151464452720014227 0ustar00"""Build a project using PEP 517 hooks. """ import argparse import logging import os import toml import shutil from .envbuild import BuildEnvironment from .wrappers import Pep517HookCaller from .dirtools import tempdir, mkdir_p from .compat import FileNotFoundError log = logging.getLogger(__name__) def validate_system(system): """ Ensure build system has the requisite fields. """ required = {'requires', 'build-backend'} if not (required <= set(system)): message = "Missing required fields: {missing}".format( missing=required-set(system), ) raise ValueError(message) def load_system(source_dir): """ Load the build system from a source dir (pyproject.toml). """ pyproject = os.path.join(source_dir, 'pyproject.toml') with open(pyproject) as f: pyproject_data = toml.load(f) return pyproject_data['build-system'] def compat_system(source_dir): """ Given a source dir, attempt to get a build system backend and requirements from pyproject.toml. Fallback to setuptools but only if the file was not found or a build system was not indicated. """ try: system = load_system(source_dir) except (FileNotFoundError, KeyError): system = {} system.setdefault( 'build-backend', 'setuptools.build_meta:__legacy__', ) system.setdefault('requires', ['setuptools', 'wheel']) return system def _do_build(hooks, env, dist, dest): get_requires_name = 'get_requires_for_build_{dist}'.format(**locals()) get_requires = getattr(hooks, get_requires_name) reqs = get_requires({}) log.info('Got build requires: %s', reqs) env.pip_install(reqs) log.info('Installed dynamic build dependencies') with tempdir() as td: log.info('Trying to build %s in %s', dist, td) build_name = 'build_{dist}'.format(**locals()) build = getattr(hooks, build_name) filename = build(td, {}) source = os.path.join(td, filename) shutil.move(source, os.path.join(dest, os.path.basename(filename))) def build(source_dir, dist, dest=None, system=None): system = system or load_system(source_dir) dest = os.path.join(source_dir, dest or 'dist') mkdir_p(dest) validate_system(system) hooks = Pep517HookCaller( source_dir, system['build-backend'], system.get('backend-path') ) with BuildEnvironment() as env: env.pip_install(system['requires']) _do_build(hooks, env, dist, dest) parser = argparse.ArgumentParser() parser.add_argument( 'source_dir', help="A directory containing pyproject.toml", ) parser.add_argument( '--binary', '-b', action='store_true', default=False, ) parser.add_argument( '--source', '-s', action='store_true', default=False, ) parser.add_argument( '--out-dir', '-o', help="Destination in which to save the builds relative to source dir", ) def main(args): # determine which dists to build dists = list(filter(None, ( 'sdist' if args.source or not args.binary else None, 'wheel' if args.binary or not args.source else None, ))) for dist in dists: build(args.source_dir, dist, args.out_dir) if __name__ == '__main__': main(parser.parse_args()) site-packages/pip/_vendor/pep517/dirtools.py000064400000002151151464452720014753 0ustar00import os import io import contextlib import tempfile import shutil import errno import zipfile @contextlib.contextmanager def tempdir(): """Create a temporary directory in a context manager.""" td = tempfile.mkdtemp() try: yield td finally: shutil.rmtree(td) def mkdir_p(*args, **kwargs): """Like `mkdir`, but does not raise an exception if the directory already exists. """ try: return os.mkdir(*args, **kwargs) except OSError as exc: if exc.errno != errno.EEXIST: raise def dir_to_zipfile(root): """Construct an in-memory zip file for a directory.""" buffer = io.BytesIO() zip_file = zipfile.ZipFile(buffer, 'w') for root, dirs, files in os.walk(root): for path in dirs: fs_path = os.path.join(root, path) rel_path = os.path.relpath(fs_path, root) zip_file.writestr(rel_path + '/', '') for path in files: fs_path = os.path.join(root, path) rel_path = os.path.relpath(fs_path, root) zip_file.write(fs_path, rel_path) return zip_file site-packages/pip/_vendor/pep517/_in_process.py000064400000017160151464452720015425 0ustar00"""This is invoked in a subprocess to call the build backend hooks. It expects: - Command line args: hook_name, control_dir - Environment variables: PEP517_BUILD_BACKEND=entry.point:spec PEP517_BACKEND_PATH=paths (separated with os.pathsep) - control_dir/input.json: - {"kwargs": {...}} Results: - control_dir/output.json - {"return_val": ...} """ from glob import glob from importlib import import_module import os import os.path from os.path import join as pjoin import re import shutil import sys import traceback # This is run as a script, not a module, so it can't do a relative import import compat class BackendUnavailable(Exception): """Raised if we cannot import the backend""" def __init__(self, traceback): self.traceback = traceback class BackendInvalid(Exception): """Raised if the backend is invalid""" def __init__(self, message): self.message = message class HookMissing(Exception): """Raised if a hook is missing and we are not executing the fallback""" def contained_in(filename, directory): """Test if a file is located within the given directory.""" filename = os.path.normcase(os.path.abspath(filename)) directory = os.path.normcase(os.path.abspath(directory)) return os.path.commonprefix([filename, directory]) == directory def _build_backend(): """Find and load the build backend""" # Add in-tree backend directories to the front of sys.path. backend_path = os.environ.get('PEP517_BACKEND_PATH') if backend_path: extra_pathitems = backend_path.split(os.pathsep) sys.path[:0] = extra_pathitems ep = os.environ['PEP517_BUILD_BACKEND'] mod_path, _, obj_path = ep.partition(':') try: obj = import_module(mod_path) except ImportError: raise BackendUnavailable(traceback.format_exc()) if backend_path: if not any( contained_in(obj.__file__, path) for path in extra_pathitems ): raise BackendInvalid("Backend was not loaded from backend-path") if obj_path: for path_part in obj_path.split('.'): obj = getattr(obj, path_part) return obj def get_requires_for_build_wheel(config_settings): """Invoke the optional get_requires_for_build_wheel hook Returns [] if the hook is not defined. """ backend = _build_backend() try: hook = backend.get_requires_for_build_wheel except AttributeError: return [] else: return hook(config_settings) def prepare_metadata_for_build_wheel( metadata_directory, config_settings, _allow_fallback): """Invoke optional prepare_metadata_for_build_wheel Implements a fallback by building a wheel if the hook isn't defined, unless _allow_fallback is False in which case HookMissing is raised. """ backend = _build_backend() try: hook = backend.prepare_metadata_for_build_wheel except AttributeError: if not _allow_fallback: raise HookMissing() return _get_wheel_metadata_from_wheel(backend, metadata_directory, config_settings) else: return hook(metadata_directory, config_settings) WHEEL_BUILT_MARKER = 'PEP517_ALREADY_BUILT_WHEEL' def _dist_info_files(whl_zip): """Identify the .dist-info folder inside a wheel ZipFile.""" res = [] for path in whl_zip.namelist(): m = re.match(r'[^/\\]+-[^/\\]+\.dist-info/', path) if m: res.append(path) if res: return res raise Exception("No .dist-info folder found in wheel") def _get_wheel_metadata_from_wheel( backend, metadata_directory, config_settings): """Build a wheel and extract the metadata from it. Fallback for when the build backend does not define the 'get_wheel_metadata' hook. """ from zipfile import ZipFile whl_basename = backend.build_wheel(metadata_directory, config_settings) with open(os.path.join(metadata_directory, WHEEL_BUILT_MARKER), 'wb'): pass # Touch marker file whl_file = os.path.join(metadata_directory, whl_basename) with ZipFile(whl_file) as zipf: dist_info = _dist_info_files(zipf) zipf.extractall(path=metadata_directory, members=dist_info) return dist_info[0].split('/')[0] def _find_already_built_wheel(metadata_directory): """Check for a wheel already built during the get_wheel_metadata hook. """ if not metadata_directory: return None metadata_parent = os.path.dirname(metadata_directory) if not os.path.isfile(pjoin(metadata_parent, WHEEL_BUILT_MARKER)): return None whl_files = glob(os.path.join(metadata_parent, '*.whl')) if not whl_files: print('Found wheel built marker, but no .whl files') return None if len(whl_files) > 1: print('Found multiple .whl files; unspecified behaviour. ' 'Will call build_wheel.') return None # Exactly one .whl file return whl_files[0] def build_wheel(wheel_directory, config_settings, metadata_directory=None): """Invoke the mandatory build_wheel hook. If a wheel was already built in the prepare_metadata_for_build_wheel fallback, this will copy it rather than rebuilding the wheel. """ prebuilt_whl = _find_already_built_wheel(metadata_directory) if prebuilt_whl: shutil.copy2(prebuilt_whl, wheel_directory) return os.path.basename(prebuilt_whl) return _build_backend().build_wheel(wheel_directory, config_settings, metadata_directory) def get_requires_for_build_sdist(config_settings): """Invoke the optional get_requires_for_build_wheel hook Returns [] if the hook is not defined. """ backend = _build_backend() try: hook = backend.get_requires_for_build_sdist except AttributeError: return [] else: return hook(config_settings) class _DummyException(Exception): """Nothing should ever raise this exception""" class GotUnsupportedOperation(Exception): """For internal use when backend raises UnsupportedOperation""" def __init__(self, traceback): self.traceback = traceback def build_sdist(sdist_directory, config_settings): """Invoke the mandatory build_sdist hook.""" backend = _build_backend() try: return backend.build_sdist(sdist_directory, config_settings) except getattr(backend, 'UnsupportedOperation', _DummyException): raise GotUnsupportedOperation(traceback.format_exc()) HOOK_NAMES = { 'get_requires_for_build_wheel', 'prepare_metadata_for_build_wheel', 'build_wheel', 'get_requires_for_build_sdist', 'build_sdist', } def main(): if len(sys.argv) < 3: sys.exit("Needs args: hook_name, control_dir") hook_name = sys.argv[1] control_dir = sys.argv[2] if hook_name not in HOOK_NAMES: sys.exit("Unknown hook: %s" % hook_name) hook = globals()[hook_name] hook_input = compat.read_json(pjoin(control_dir, 'input.json')) json_out = {'unsupported': False, 'return_val': None} try: json_out['return_val'] = hook(**hook_input['kwargs']) except BackendUnavailable as e: json_out['no_backend'] = True json_out['traceback'] = e.traceback except BackendInvalid as e: json_out['backend_invalid'] = True json_out['backend_error'] = e.message except GotUnsupportedOperation as e: json_out['unsupported'] = True json_out['traceback'] = e.traceback except HookMissing: json_out['hook_missing'] = True compat.write_json(json_out, pjoin(control_dir, 'output.json'), indent=2) if __name__ == '__main__': main() site-packages/pip/_vendor/pep517/__init__.py000064400000000124151464452720014651 0ustar00"""Wrappers to build Python packages using PEP 517 hooks """ __version__ = '0.7.0' site-packages/pip/_vendor/certifi/__main__.py000064400000000065151464452720015242 0ustar00from pip._vendor.certifi import where print(where()) site-packages/pip/_vendor/certifi/__pycache__/core.cpython-38.pyc000064400000000621151464452720020736 0ustar00U .e@sdZddlZddZdS)zU certifi.py ~~~~~~~~~~ This module returns the installation location of cacert.pem. NcCsdS)Nz /etc/pki/tls/certs/ca-bundle.crtrrrssite-packages/pip/_vendor/certifi/__pycache__/__init__.cpython-38.pyc000064400000000334151464452720021546 0ustar00U .e4@sddlmZdZdS))wherez 2019.09.11N)Zcorer __version__rr@/usr/lib/python3.8/site-packages/pip/_vendor/certifi/__init__.pys site-packages/pip/_vendor/certifi/__pycache__/core.cpython-38.opt-1.pyc000064400000000621151464452720021675 0ustar00U .e@sdZddlZddZdS)zU certifi.py ~~~~~~~~~~ This module returns the installation location of cacert.pem. NcCsdS)Nz /etc/pki/tls/certs/ca-bundle.crtrrrssite-packages/pip/_vendor/certifi/__pycache__/__main__.cpython-38.opt-1.pyc000064400000000337151464452720022471 0ustar00U .e5@sddlmZeedS))whereN)Zpip._vendor.certifirprintrr@/usr/lib/python3.8/site-packages/pip/_vendor/certifi/__main__.pys site-packages/pip/_vendor/certifi/__pycache__/__init__.cpython-38.opt-1.pyc000064400000000334151464452720022505 0ustar00U .e4@sddlmZdZdS))wherez 2019.09.11N)Zcorer __version__rr@/usr/lib/python3.8/site-packages/pip/_vendor/certifi/__init__.pys site-packages/pip/_vendor/certifi/__pycache__/__main__.cpython-38.pyc000064400000000337151464452720021532 0ustar00U .e5@sddlmZeedS))whereN)Zpip._vendor.certifirprintrr@/usr/lib/python3.8/site-packages/pip/_vendor/certifi/__main__.pys site-packages/pip/_vendor/certifi/__init__.py000064400000000064151464452720015260 0ustar00from .core import where __version__ = "2019.09.11" site-packages/pip/_vendor/certifi/core.py000064400000000274151464452720014454 0ustar00# -*- coding: utf-8 -*- """ certifi.py ~~~~~~~~~~ This module returns the installation location of cacert.pem. """ import os def where(): return '/etc/pki/tls/certs/ca-bundle.crt' site-packages/pip/_vendor/retrying.py000064400000023364151464452720013747 0ustar00## Copyright 2013-2014 Ray Holder ## ## Licensed under the Apache License, Version 2.0 (the "License"); ## you may not use this file except in compliance with the License. ## You may obtain a copy of the License at ## ## http://www.apache.org/licenses/LICENSE-2.0 ## ## Unless required by applicable law or agreed to in writing, software ## distributed under the License is distributed on an "AS IS" BASIS, ## WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. ## See the License for the specific language governing permissions and ## limitations under the License. import random from pip._vendor import six import sys import time import traceback # sys.maxint / 2, since Python 3.2 doesn't have a sys.maxint... MAX_WAIT = 1073741823 def retry(*dargs, **dkw): """ Decorator function that instantiates the Retrying object @param *dargs: positional arguments passed to Retrying object @param **dkw: keyword arguments passed to the Retrying object """ # support both @retry and @retry() as valid syntax if len(dargs) == 1 and callable(dargs[0]): def wrap_simple(f): @six.wraps(f) def wrapped_f(*args, **kw): return Retrying().call(f, *args, **kw) return wrapped_f return wrap_simple(dargs[0]) else: def wrap(f): @six.wraps(f) def wrapped_f(*args, **kw): return Retrying(*dargs, **dkw).call(f, *args, **kw) return wrapped_f return wrap class Retrying(object): def __init__(self, stop=None, wait=None, stop_max_attempt_number=None, stop_max_delay=None, wait_fixed=None, wait_random_min=None, wait_random_max=None, wait_incrementing_start=None, wait_incrementing_increment=None, wait_exponential_multiplier=None, wait_exponential_max=None, retry_on_exception=None, retry_on_result=None, wrap_exception=False, stop_func=None, wait_func=None, wait_jitter_max=None): self._stop_max_attempt_number = 5 if stop_max_attempt_number is None else stop_max_attempt_number self._stop_max_delay = 100 if stop_max_delay is None else stop_max_delay self._wait_fixed = 1000 if wait_fixed is None else wait_fixed self._wait_random_min = 0 if wait_random_min is None else wait_random_min self._wait_random_max = 1000 if wait_random_max is None else wait_random_max self._wait_incrementing_start = 0 if wait_incrementing_start is None else wait_incrementing_start self._wait_incrementing_increment = 100 if wait_incrementing_increment is None else wait_incrementing_increment self._wait_exponential_multiplier = 1 if wait_exponential_multiplier is None else wait_exponential_multiplier self._wait_exponential_max = MAX_WAIT if wait_exponential_max is None else wait_exponential_max self._wait_jitter_max = 0 if wait_jitter_max is None else wait_jitter_max # TODO add chaining of stop behaviors # stop behavior stop_funcs = [] if stop_max_attempt_number is not None: stop_funcs.append(self.stop_after_attempt) if stop_max_delay is not None: stop_funcs.append(self.stop_after_delay) if stop_func is not None: self.stop = stop_func elif stop is None: self.stop = lambda attempts, delay: any(f(attempts, delay) for f in stop_funcs) else: self.stop = getattr(self, stop) # TODO add chaining of wait behaviors # wait behavior wait_funcs = [lambda *args, **kwargs: 0] if wait_fixed is not None: wait_funcs.append(self.fixed_sleep) if wait_random_min is not None or wait_random_max is not None: wait_funcs.append(self.random_sleep) if wait_incrementing_start is not None or wait_incrementing_increment is not None: wait_funcs.append(self.incrementing_sleep) if wait_exponential_multiplier is not None or wait_exponential_max is not None: wait_funcs.append(self.exponential_sleep) if wait_func is not None: self.wait = wait_func elif wait is None: self.wait = lambda attempts, delay: max(f(attempts, delay) for f in wait_funcs) else: self.wait = getattr(self, wait) # retry on exception filter if retry_on_exception is None: self._retry_on_exception = self.always_reject else: self._retry_on_exception = retry_on_exception # TODO simplify retrying by Exception types # retry on result filter if retry_on_result is None: self._retry_on_result = self.never_reject else: self._retry_on_result = retry_on_result self._wrap_exception = wrap_exception def stop_after_attempt(self, previous_attempt_number, delay_since_first_attempt_ms): """Stop after the previous attempt >= stop_max_attempt_number.""" return previous_attempt_number >= self._stop_max_attempt_number def stop_after_delay(self, previous_attempt_number, delay_since_first_attempt_ms): """Stop after the time from the first attempt >= stop_max_delay.""" return delay_since_first_attempt_ms >= self._stop_max_delay def no_sleep(self, previous_attempt_number, delay_since_first_attempt_ms): """Don't sleep at all before retrying.""" return 0 def fixed_sleep(self, previous_attempt_number, delay_since_first_attempt_ms): """Sleep a fixed amount of time between each retry.""" return self._wait_fixed def random_sleep(self, previous_attempt_number, delay_since_first_attempt_ms): """Sleep a random amount of time between wait_random_min and wait_random_max""" return random.randint(self._wait_random_min, self._wait_random_max) def incrementing_sleep(self, previous_attempt_number, delay_since_first_attempt_ms): """ Sleep an incremental amount of time after each attempt, starting at wait_incrementing_start and incrementing by wait_incrementing_increment """ result = self._wait_incrementing_start + (self._wait_incrementing_increment * (previous_attempt_number - 1)) if result < 0: result = 0 return result def exponential_sleep(self, previous_attempt_number, delay_since_first_attempt_ms): exp = 2 ** previous_attempt_number result = self._wait_exponential_multiplier * exp if result > self._wait_exponential_max: result = self._wait_exponential_max if result < 0: result = 0 return result def never_reject(self, result): return False def always_reject(self, result): return True def should_reject(self, attempt): reject = False if attempt.has_exception: reject |= self._retry_on_exception(attempt.value[1]) else: reject |= self._retry_on_result(attempt.value) return reject def call(self, fn, *args, **kwargs): start_time = int(round(time.time() * 1000)) attempt_number = 1 while True: try: attempt = Attempt(fn(*args, **kwargs), attempt_number, False) except: tb = sys.exc_info() attempt = Attempt(tb, attempt_number, True) if not self.should_reject(attempt): return attempt.get(self._wrap_exception) delay_since_first_attempt_ms = int(round(time.time() * 1000)) - start_time if self.stop(attempt_number, delay_since_first_attempt_ms): if not self._wrap_exception and attempt.has_exception: # get() on an attempt with an exception should cause it to be raised, but raise just in case raise attempt.get() else: raise RetryError(attempt) else: sleep = self.wait(attempt_number, delay_since_first_attempt_ms) if self._wait_jitter_max: jitter = random.random() * self._wait_jitter_max sleep = sleep + max(0, jitter) time.sleep(sleep / 1000.0) attempt_number += 1 class Attempt(object): """ An Attempt encapsulates a call to a target function that may end as a normal return value from the function or an Exception depending on what occurred during the execution. """ def __init__(self, value, attempt_number, has_exception): self.value = value self.attempt_number = attempt_number self.has_exception = has_exception def get(self, wrap_exception=False): """ Return the return value of this Attempt instance or raise an Exception. If wrap_exception is true, this Attempt is wrapped inside of a RetryError before being raised. """ if self.has_exception: if wrap_exception: raise RetryError(self) else: six.reraise(self.value[0], self.value[1], self.value[2]) else: return self.value def __repr__(self): if self.has_exception: return "Attempts: {0}, Error:\n{1}".format(self.attempt_number, "".join(traceback.format_tb(self.value[2]))) else: return "Attempts: {0}, Value: {1}".format(self.attempt_number, self.value) class RetryError(Exception): """ A RetryError encapsulates the last Attempt instance right before giving up. """ def __init__(self, last_attempt): self.last_attempt = last_attempt def __str__(self): return "RetryError[{0}]".format(self.last_attempt) site-packages/pip/_vendor/__init__.py000064400000011061151464452720013632 0ustar00""" pip._vendor is for vendoring dependencies of pip to prevent needing pip to depend on something external. Files inside of pip._vendor should be considered immutable and should only be updated to versions from upstream. """ from __future__ import absolute_import import glob import os.path import sys # Downstream redistributors which have debundled our dependencies should also # patch this value to be true. This will trigger the additional patching # to cause things like "six" to be available as pip. DEBUNDLED = False # By default, look in this directory for a bunch of .whl files which we will # add to the beginning of sys.path before attempting to import anything. This # is done to support downstream re-distributors like Debian and Fedora who # wish to create their own Wheels for our dependencies to aid in debundling. WHEEL_DIR = os.path.abspath(os.path.dirname(__file__)) # Define a small helper function to alias our vendored modules to the real ones # if the vendored ones do not exist. This idea of this was taken from # https://github.com/kennethreitz/requests/pull/2567. def vendored(modulename): vendored_name = "{0}.{1}".format(__name__, modulename) try: __import__(modulename, globals(), locals(), level=0) except ImportError: # We can just silently allow import failures to pass here. If we # got to this point it means that ``import pip._vendor.whatever`` # failed and so did ``import whatever``. Since we're importing this # upfront in an attempt to alias imports, not erroring here will # just mean we get a regular import error whenever pip *actually* # tries to import one of these modules to use it, which actually # gives us a better error message than we would have otherwise # gotten. pass else: sys.modules[vendored_name] = sys.modules[modulename] base, head = vendored_name.rsplit(".", 1) setattr(sys.modules[base], head, sys.modules[modulename]) # If we're operating in a debundled setup, then we want to go ahead and trigger # the aliasing of our vendored libraries as well as looking for wheels to add # to our sys.path. This will cause all of this code to be a no-op typically # however downstream redistributors can enable it in a consistent way across # all platforms. if DEBUNDLED: # Actually look inside of WHEEL_DIR to find .whl files and add them to the # front of our sys.path. sys.path[:] = glob.glob(os.path.join(WHEEL_DIR, "*.whl")) + sys.path # Actually alias all of our vendored dependencies. vendored("cachecontrol") vendored("colorama") vendored("contextlib2") vendored("distlib") vendored("distro") vendored("html5lib") vendored("six") vendored("six.moves") vendored("six.moves.urllib") vendored("six.moves.urllib.parse") vendored("packaging") vendored("packaging.version") vendored("packaging.specifiers") vendored("pep517") vendored("pkg_resources") vendored("progress") vendored("pytoml") vendored("retrying") vendored("requests") vendored("requests.exceptions") vendored("requests.packages") vendored("requests.packages.urllib3") vendored("requests.packages.urllib3._collections") vendored("requests.packages.urllib3.connection") vendored("requests.packages.urllib3.connectionpool") vendored("requests.packages.urllib3.contrib") vendored("requests.packages.urllib3.contrib.ntlmpool") vendored("requests.packages.urllib3.contrib.pyopenssl") vendored("requests.packages.urllib3.exceptions") vendored("requests.packages.urllib3.fields") vendored("requests.packages.urllib3.filepost") vendored("requests.packages.urllib3.packages") vendored("requests.packages.urllib3.packages.ordered_dict") vendored("requests.packages.urllib3.packages.six") vendored("requests.packages.urllib3.packages.ssl_match_hostname") vendored("requests.packages.urllib3.packages.ssl_match_hostname." "_implementation") vendored("requests.packages.urllib3.poolmanager") vendored("requests.packages.urllib3.request") vendored("requests.packages.urllib3.response") vendored("requests.packages.urllib3.util") vendored("requests.packages.urllib3.util.connection") vendored("requests.packages.urllib3.util.request") vendored("requests.packages.urllib3.util.response") vendored("requests.packages.urllib3.util.retry") vendored("requests.packages.urllib3.util.ssl_") vendored("requests.packages.urllib3.util.timeout") vendored("requests.packages.urllib3.util.url") vendored("urllib3") site-packages/pip/_vendor/requests/models.py000064400000102743151464452720015241 0ustar00# -*- coding: utf-8 -*- """ requests.models ~~~~~~~~~~~~~~~ This module contains the primary objects that power Requests. """ import datetime import sys # Import encoding now, to avoid implicit import later. # Implicit import within threads may cause LookupError when standard library is in a ZIP, # such as in Embedded Python. See https://github.com/requests/requests/issues/3578. import encodings.idna from pip._vendor.urllib3.fields import RequestField from pip._vendor.urllib3.filepost import encode_multipart_formdata from pip._vendor.urllib3.util import parse_url from pip._vendor.urllib3.exceptions import ( DecodeError, ReadTimeoutError, ProtocolError, LocationParseError) from io import UnsupportedOperation from .hooks import default_hooks from .structures import CaseInsensitiveDict from .auth import HTTPBasicAuth from .cookies import cookiejar_from_dict, get_cookie_header, _copy_cookie_jar from .exceptions import ( HTTPError, MissingSchema, InvalidURL, ChunkedEncodingError, ContentDecodingError, ConnectionError, StreamConsumedError) from ._internal_utils import to_native_string, unicode_is_ascii from .utils import ( guess_filename, get_auth_from_url, requote_uri, stream_decode_response_unicode, to_key_val_list, parse_header_links, iter_slices, guess_json_utf, super_len, check_header_validity) from .compat import ( Callable, Mapping, cookielib, urlunparse, urlsplit, urlencode, str, bytes, is_py2, chardet, builtin_str, basestring) from .compat import json as complexjson from .status_codes import codes #: The set of HTTP status codes that indicate an automatically #: processable redirect. REDIRECT_STATI = ( codes.moved, # 301 codes.found, # 302 codes.other, # 303 codes.temporary_redirect, # 307 codes.permanent_redirect, # 308 ) DEFAULT_REDIRECT_LIMIT = 30 CONTENT_CHUNK_SIZE = 10 * 1024 ITER_CHUNK_SIZE = 512 class RequestEncodingMixin(object): @property def path_url(self): """Build the path URL to use.""" url = [] p = urlsplit(self.url) path = p.path if not path: path = '/' url.append(path) query = p.query if query: url.append('?') url.append(query) return ''.join(url) @staticmethod def _encode_params(data): """Encode parameters in a piece of data. Will successfully encode parameters when passed as a dict or a list of 2-tuples. Order is retained if data is a list of 2-tuples but arbitrary if parameters are supplied as a dict. """ if isinstance(data, (str, bytes)): return data elif hasattr(data, 'read'): return data elif hasattr(data, '__iter__'): result = [] for k, vs in to_key_val_list(data): if isinstance(vs, basestring) or not hasattr(vs, '__iter__'): vs = [vs] for v in vs: if v is not None: result.append( (k.encode('utf-8') if isinstance(k, str) else k, v.encode('utf-8') if isinstance(v, str) else v)) return urlencode(result, doseq=True) else: return data @staticmethod def _encode_files(files, data): """Build the body for a multipart/form-data request. Will successfully encode files when passed as a dict or a list of tuples. Order is retained if data is a list of tuples but arbitrary if parameters are supplied as a dict. The tuples may be 2-tuples (filename, fileobj), 3-tuples (filename, fileobj, contentype) or 4-tuples (filename, fileobj, contentype, custom_headers). """ if (not files): raise ValueError("Files must be provided.") elif isinstance(data, basestring): raise ValueError("Data must not be a string.") new_fields = [] fields = to_key_val_list(data or {}) files = to_key_val_list(files or {}) for field, val in fields: if isinstance(val, basestring) or not hasattr(val, '__iter__'): val = [val] for v in val: if v is not None: # Don't call str() on bytestrings: in Py3 it all goes wrong. if not isinstance(v, bytes): v = str(v) new_fields.append( (field.decode('utf-8') if isinstance(field, bytes) else field, v.encode('utf-8') if isinstance(v, str) else v)) for (k, v) in files: # support for explicit filename ft = None fh = None if isinstance(v, (tuple, list)): if len(v) == 2: fn, fp = v elif len(v) == 3: fn, fp, ft = v else: fn, fp, ft, fh = v else: fn = guess_filename(v) or k fp = v if isinstance(fp, (str, bytes, bytearray)): fdata = fp elif hasattr(fp, 'read'): fdata = fp.read() elif fp is None: continue else: fdata = fp rf = RequestField(name=k, data=fdata, filename=fn, headers=fh) rf.make_multipart(content_type=ft) new_fields.append(rf) body, content_type = encode_multipart_formdata(new_fields) return body, content_type class RequestHooksMixin(object): def register_hook(self, event, hook): """Properly register a hook.""" if event not in self.hooks: raise ValueError('Unsupported event specified, with event name "%s"' % (event)) if isinstance(hook, Callable): self.hooks[event].append(hook) elif hasattr(hook, '__iter__'): self.hooks[event].extend(h for h in hook if isinstance(h, Callable)) def deregister_hook(self, event, hook): """Deregister a previously registered hook. Returns True if the hook existed, False if not. """ try: self.hooks[event].remove(hook) return True except ValueError: return False class Request(RequestHooksMixin): """A user-created :class:`Request ` object. Used to prepare a :class:`PreparedRequest `, which is sent to the server. :param method: HTTP method to use. :param url: URL to send. :param headers: dictionary of headers to send. :param files: dictionary of {filename: fileobject} files to multipart upload. :param data: the body to attach to the request. If a dictionary or list of tuples ``[(key, value)]`` is provided, form-encoding will take place. :param json: json for the body to attach to the request (if files or data is not specified). :param params: URL parameters to append to the URL. If a dictionary or list of tuples ``[(key, value)]`` is provided, form-encoding will take place. :param auth: Auth handler or (user, pass) tuple. :param cookies: dictionary or CookieJar of cookies to attach to this request. :param hooks: dictionary of callback hooks, for internal usage. Usage:: >>> import requests >>> req = requests.Request('GET', 'https://httpbin.org/get') >>> req.prepare() """ def __init__(self, method=None, url=None, headers=None, files=None, data=None, params=None, auth=None, cookies=None, hooks=None, json=None): # Default empty dicts for dict params. data = [] if data is None else data files = [] if files is None else files headers = {} if headers is None else headers params = {} if params is None else params hooks = {} if hooks is None else hooks self.hooks = default_hooks() for (k, v) in list(hooks.items()): self.register_hook(event=k, hook=v) self.method = method self.url = url self.headers = headers self.files = files self.data = data self.json = json self.params = params self.auth = auth self.cookies = cookies def __repr__(self): return '' % (self.method) def prepare(self): """Constructs a :class:`PreparedRequest ` for transmission and returns it.""" p = PreparedRequest() p.prepare( method=self.method, url=self.url, headers=self.headers, files=self.files, data=self.data, json=self.json, params=self.params, auth=self.auth, cookies=self.cookies, hooks=self.hooks, ) return p class PreparedRequest(RequestEncodingMixin, RequestHooksMixin): """The fully mutable :class:`PreparedRequest ` object, containing the exact bytes that will be sent to the server. Generated from either a :class:`Request ` object or manually. Usage:: >>> import requests >>> req = requests.Request('GET', 'https://httpbin.org/get') >>> r = req.prepare() >>> s = requests.Session() >>> s.send(r) """ def __init__(self): #: HTTP verb to send to the server. self.method = None #: HTTP URL to send the request to. self.url = None #: dictionary of HTTP headers. self.headers = None # The `CookieJar` used to create the Cookie header will be stored here # after prepare_cookies is called self._cookies = None #: request body to send to the server. self.body = None #: dictionary of callback hooks, for internal usage. self.hooks = default_hooks() #: integer denoting starting position of a readable file-like body. self._body_position = None def prepare(self, method=None, url=None, headers=None, files=None, data=None, params=None, auth=None, cookies=None, hooks=None, json=None): """Prepares the entire request with the given parameters.""" self.prepare_method(method) self.prepare_url(url, params) self.prepare_headers(headers) self.prepare_cookies(cookies) self.prepare_body(data, files, json) self.prepare_auth(auth, url) # Note that prepare_auth must be last to enable authentication schemes # such as OAuth to work on a fully prepared request. # This MUST go after prepare_auth. Authenticators could add a hook self.prepare_hooks(hooks) def __repr__(self): return '' % (self.method) def copy(self): p = PreparedRequest() p.method = self.method p.url = self.url p.headers = self.headers.copy() if self.headers is not None else None p._cookies = _copy_cookie_jar(self._cookies) p.body = self.body p.hooks = self.hooks p._body_position = self._body_position return p def prepare_method(self, method): """Prepares the given HTTP method.""" self.method = method if self.method is not None: self.method = to_native_string(self.method.upper()) @staticmethod def _get_idna_encoded_host(host): from pip._vendor import idna try: host = idna.encode(host, uts46=True).decode('utf-8') except idna.IDNAError: raise UnicodeError return host def prepare_url(self, url, params): """Prepares the given HTTP URL.""" #: Accept objects that have string representations. #: We're unable to blindly call unicode/str functions #: as this will include the bytestring indicator (b'') #: on python 3.x. #: https://github.com/requests/requests/pull/2238 if isinstance(url, bytes): url = url.decode('utf8') else: url = unicode(url) if is_py2 else str(url) # Remove leading whitespaces from url url = url.lstrip() # Don't do any URL preparation for non-HTTP schemes like `mailto`, # `data` etc to work around exceptions from `url_parse`, which # handles RFC 3986 only. if ':' in url and not url.lower().startswith('http'): self.url = url return # Support for unicode domain names and paths. try: scheme, auth, host, port, path, query, fragment = parse_url(url) except LocationParseError as e: raise InvalidURL(*e.args) if not scheme: error = ("Invalid URL {0!r}: No schema supplied. Perhaps you meant http://{0}?") error = error.format(to_native_string(url, 'utf8')) raise MissingSchema(error) if not host: raise InvalidURL("Invalid URL %r: No host supplied" % url) # In general, we want to try IDNA encoding the hostname if the string contains # non-ASCII characters. This allows users to automatically get the correct IDNA # behaviour. For strings containing only ASCII characters, we need to also verify # it doesn't start with a wildcard (*), before allowing the unencoded hostname. if not unicode_is_ascii(host): try: host = self._get_idna_encoded_host(host) except UnicodeError: raise InvalidURL('URL has an invalid label.') elif host.startswith(u'*'): raise InvalidURL('URL has an invalid label.') # Carefully reconstruct the network location netloc = auth or '' if netloc: netloc += '@' netloc += host if port: netloc += ':' + str(port) # Bare domains aren't valid URLs. if not path: path = '/' if is_py2: if isinstance(scheme, str): scheme = scheme.encode('utf-8') if isinstance(netloc, str): netloc = netloc.encode('utf-8') if isinstance(path, str): path = path.encode('utf-8') if isinstance(query, str): query = query.encode('utf-8') if isinstance(fragment, str): fragment = fragment.encode('utf-8') if isinstance(params, (str, bytes)): params = to_native_string(params) enc_params = self._encode_params(params) if enc_params: if query: query = '%s&%s' % (query, enc_params) else: query = enc_params url = requote_uri(urlunparse([scheme, netloc, path, None, query, fragment])) self.url = url def prepare_headers(self, headers): """Prepares the given HTTP headers.""" self.headers = CaseInsensitiveDict() if headers: for header in headers.items(): # Raise exception on invalid header value. check_header_validity(header) name, value = header self.headers[to_native_string(name)] = value def prepare_body(self, data, files, json=None): """Prepares the given HTTP body data.""" # Check if file, fo, generator, iterator. # If not, run through normal process. # Nottin' on you. body = None content_type = None if not data and json is not None: # urllib3 requires a bytes-like body. Python 2's json.dumps # provides this natively, but Python 3 gives a Unicode string. content_type = 'application/json' body = complexjson.dumps(json) if not isinstance(body, bytes): body = body.encode('utf-8') is_stream = all([ hasattr(data, '__iter__'), not isinstance(data, (basestring, list, tuple, Mapping)) ]) try: length = super_len(data) except (TypeError, AttributeError, UnsupportedOperation): length = None if is_stream: body = data if getattr(body, 'tell', None) is not None: # Record the current file position before reading. # This will allow us to rewind a file in the event # of a redirect. try: self._body_position = body.tell() except (IOError, OSError): # This differentiates from None, allowing us to catch # a failed `tell()` later when trying to rewind the body self._body_position = object() if files: raise NotImplementedError('Streamed bodies and files are mutually exclusive.') if length: self.headers['Content-Length'] = builtin_str(length) else: self.headers['Transfer-Encoding'] = 'chunked' else: # Multi-part file uploads. if files: (body, content_type) = self._encode_files(files, data) else: if data: body = self._encode_params(data) if isinstance(data, basestring) or hasattr(data, 'read'): content_type = None else: content_type = 'application/x-www-form-urlencoded' self.prepare_content_length(body) # Add content-type if it wasn't explicitly provided. if content_type and ('content-type' not in self.headers): self.headers['Content-Type'] = content_type self.body = body def prepare_content_length(self, body): """Prepare Content-Length header based on request method and body""" if body is not None: length = super_len(body) if length: # If length exists, set it. Otherwise, we fallback # to Transfer-Encoding: chunked. self.headers['Content-Length'] = builtin_str(length) elif self.method not in ('GET', 'HEAD') and self.headers.get('Content-Length') is None: # Set Content-Length to 0 for methods that can have a body # but don't provide one. (i.e. not GET or HEAD) self.headers['Content-Length'] = '0' def prepare_auth(self, auth, url=''): """Prepares the given HTTP auth data.""" # If no Auth is explicitly provided, extract it from the URL first. if auth is None: url_auth = get_auth_from_url(self.url) auth = url_auth if any(url_auth) else None if auth: if isinstance(auth, tuple) and len(auth) == 2: # special-case basic HTTP auth auth = HTTPBasicAuth(*auth) # Allow auth to make its changes. r = auth(self) # Update self to reflect the auth changes. self.__dict__.update(r.__dict__) # Recompute Content-Length self.prepare_content_length(self.body) def prepare_cookies(self, cookies): """Prepares the given HTTP cookie data. This function eventually generates a ``Cookie`` header from the given cookies using cookielib. Due to cookielib's design, the header will not be regenerated if it already exists, meaning this function can only be called once for the life of the :class:`PreparedRequest ` object. Any subsequent calls to ``prepare_cookies`` will have no actual effect, unless the "Cookie" header is removed beforehand. """ if isinstance(cookies, cookielib.CookieJar): self._cookies = cookies else: self._cookies = cookiejar_from_dict(cookies) cookie_header = get_cookie_header(self._cookies, self) if cookie_header is not None: self.headers['Cookie'] = cookie_header def prepare_hooks(self, hooks): """Prepares the given hooks.""" # hooks can be passed as None to the prepare method and to this # method. To prevent iterating over None, simply use an empty list # if hooks is False-y hooks = hooks or [] for event in hooks: self.register_hook(event, hooks[event]) class Response(object): """The :class:`Response ` object, which contains a server's response to an HTTP request. """ __attrs__ = [ '_content', 'status_code', 'headers', 'url', 'history', 'encoding', 'reason', 'cookies', 'elapsed', 'request' ] def __init__(self): self._content = False self._content_consumed = False self._next = None #: Integer Code of responded HTTP Status, e.g. 404 or 200. self.status_code = None #: Case-insensitive Dictionary of Response Headers. #: For example, ``headers['content-encoding']`` will return the #: value of a ``'Content-Encoding'`` response header. self.headers = CaseInsensitiveDict() #: File-like object representation of response (for advanced usage). #: Use of ``raw`` requires that ``stream=True`` be set on the request. # This requirement does not apply for use internally to Requests. self.raw = None #: Final URL location of Response. self.url = None #: Encoding to decode with when accessing r.text. self.encoding = None #: A list of :class:`Response ` objects from #: the history of the Request. Any redirect responses will end #: up here. The list is sorted from the oldest to the most recent request. self.history = [] #: Textual reason of responded HTTP Status, e.g. "Not Found" or "OK". self.reason = None #: A CookieJar of Cookies the server sent back. self.cookies = cookiejar_from_dict({}) #: The amount of time elapsed between sending the request #: and the arrival of the response (as a timedelta). #: This property specifically measures the time taken between sending #: the first byte of the request and finishing parsing the headers. It #: is therefore unaffected by consuming the response content or the #: value of the ``stream`` keyword argument. self.elapsed = datetime.timedelta(0) #: The :class:`PreparedRequest ` object to which this #: is a response. self.request = None def __enter__(self): return self def __exit__(self, *args): self.close() def __getstate__(self): # Consume everything; accessing the content attribute makes # sure the content has been fully read. if not self._content_consumed: self.content return {attr: getattr(self, attr, None) for attr in self.__attrs__} def __setstate__(self, state): for name, value in state.items(): setattr(self, name, value) # pickled objects do not have .raw setattr(self, '_content_consumed', True) setattr(self, 'raw', None) def __repr__(self): return '' % (self.status_code) def __bool__(self): """Returns True if :attr:`status_code` is less than 400. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code, is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. """ return self.ok def __nonzero__(self): """Returns True if :attr:`status_code` is less than 400. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code, is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. """ return self.ok def __iter__(self): """Allows you to use a response as an iterator.""" return self.iter_content(128) @property def ok(self): """Returns True if :attr:`status_code` is less than 400, False if not. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. """ try: self.raise_for_status() except HTTPError: return False return True @property def is_redirect(self): """True if this Response is a well-formed HTTP redirect that could have been processed automatically (by :meth:`Session.resolve_redirects`). """ return ('location' in self.headers and self.status_code in REDIRECT_STATI) @property def is_permanent_redirect(self): """True if this Response one of the permanent versions of redirect.""" return ('location' in self.headers and self.status_code in (codes.moved_permanently, codes.permanent_redirect)) @property def next(self): """Returns a PreparedRequest for the next request in a redirect chain, if there is one.""" return self._next @property def apparent_encoding(self): """The apparent encoding, provided by the chardet library.""" return chardet.detect(self.content)['encoding'] def iter_content(self, chunk_size=1, decode_unicode=False): """Iterates over the response data. When stream=True is set on the request, this avoids reading the content at once into memory for large responses. The chunk size is the number of bytes it should read into memory. This is not necessarily the length of each item returned as decoding can take place. chunk_size must be of type int or None. A value of None will function differently depending on the value of `stream`. stream=True will read data as it arrives in whatever size the chunks are received. If stream=False, data is returned as a single chunk. If decode_unicode is True, content will be decoded using the best available encoding based on the response. """ def generate(): # Special case for urllib3. if hasattr(self.raw, 'stream'): try: for chunk in self.raw.stream(chunk_size, decode_content=True): yield chunk except ProtocolError as e: raise ChunkedEncodingError(e) except DecodeError as e: raise ContentDecodingError(e) except ReadTimeoutError as e: raise ConnectionError(e) else: # Standard file-like object. while True: chunk = self.raw.read(chunk_size) if not chunk: break yield chunk self._content_consumed = True if self._content_consumed and isinstance(self._content, bool): raise StreamConsumedError() elif chunk_size is not None and not isinstance(chunk_size, int): raise TypeError("chunk_size must be an int, it is instead a %s." % type(chunk_size)) # simulate reading small chunks of the content reused_chunks = iter_slices(self._content, chunk_size) stream_chunks = generate() chunks = reused_chunks if self._content_consumed else stream_chunks if decode_unicode: chunks = stream_decode_response_unicode(chunks, self) return chunks def iter_lines(self, chunk_size=ITER_CHUNK_SIZE, decode_unicode=False, delimiter=None): """Iterates over the response data, one line at a time. When stream=True is set on the request, this avoids reading the content at once into memory for large responses. .. note:: This method is not reentrant safe. """ pending = None for chunk in self.iter_content(chunk_size=chunk_size, decode_unicode=decode_unicode): if pending is not None: chunk = pending + chunk if delimiter: lines = chunk.split(delimiter) else: lines = chunk.splitlines() if lines and lines[-1] and chunk and lines[-1][-1] == chunk[-1]: pending = lines.pop() else: pending = None for line in lines: yield line if pending is not None: yield pending @property def content(self): """Content of the response, in bytes.""" if self._content is False: # Read the contents. if self._content_consumed: raise RuntimeError( 'The content for this response was already consumed') if self.status_code == 0 or self.raw is None: self._content = None else: self._content = b''.join(self.iter_content(CONTENT_CHUNK_SIZE)) or b'' self._content_consumed = True # don't need to release the connection; that's been handled by urllib3 # since we exhausted the data. return self._content @property def text(self): """Content of the response, in unicode. If Response.encoding is None, encoding will be guessed using ``chardet``. The encoding of the response content is determined based solely on HTTP headers, following RFC 2616 to the letter. If you can take advantage of non-HTTP knowledge to make a better guess at the encoding, you should set ``r.encoding`` appropriately before accessing this property. """ # Try charset from content-type content = None encoding = self.encoding if not self.content: return str('') # Fallback to auto-detected encoding. if self.encoding is None: encoding = self.apparent_encoding # Decode unicode from given encoding. try: content = str(self.content, encoding, errors='replace') except (LookupError, TypeError): # A LookupError is raised if the encoding was not found which could # indicate a misspelling or similar mistake. # # A TypeError can be raised if encoding is None # # So we try blindly encoding. content = str(self.content, errors='replace') return content def json(self, **kwargs): r"""Returns the json-encoded content of a response, if any. :param \*\*kwargs: Optional arguments that ``json.loads`` takes. :raises ValueError: If the response body does not contain valid json. """ if not self.encoding and self.content and len(self.content) > 3: # No encoding set. JSON RFC 4627 section 3 states we should expect # UTF-8, -16 or -32. Detect which one to use; If the detection or # decoding fails, fall back to `self.text` (using chardet to make # a best guess). encoding = guess_json_utf(self.content) if encoding is not None: try: return complexjson.loads( self.content.decode(encoding), **kwargs ) except UnicodeDecodeError: # Wrong UTF codec detected; usually because it's not UTF-8 # but some other 8-bit codec. This is an RFC violation, # and the server didn't bother to tell us what codec *was* # used. pass return complexjson.loads(self.text, **kwargs) @property def links(self): """Returns the parsed header links of the response, if any.""" header = self.headers.get('link') # l = MultiDict() l = {} if header: links = parse_header_links(header) for link in links: key = link.get('rel') or link.get('url') l[key] = link return l def raise_for_status(self): """Raises stored :class:`HTTPError`, if one occurred.""" http_error_msg = '' if isinstance(self.reason, bytes): # We attempt to decode utf-8 first because some servers # choose to localize their reason strings. If the string # isn't utf-8, we fall back to iso-8859-1 for all other # encodings. (See PR #3538) try: reason = self.reason.decode('utf-8') except UnicodeDecodeError: reason = self.reason.decode('iso-8859-1') else: reason = self.reason if 400 <= self.status_code < 500: http_error_msg = u'%s Client Error: %s for url: %s' % (self.status_code, reason, self.url) elif 500 <= self.status_code < 600: http_error_msg = u'%s Server Error: %s for url: %s' % (self.status_code, reason, self.url) if http_error_msg: raise HTTPError(http_error_msg, response=self) def close(self): """Releases the connection back to the pool. Once this method has been called the underlying ``raw`` object must not be accessed again. *Note: Should not normally need to be called explicitly.* """ if not self._content_consumed: self.raw.close() release_conn = getattr(self.raw, 'release_conn', None) if release_conn is not None: release_conn() site-packages/pip/_vendor/requests/_internal_utils.py000064400000002110151464452720017134 0ustar00# -*- coding: utf-8 -*- """ requests._internal_utils ~~~~~~~~~~~~~~ Provides utility functions that are consumed internally by Requests which depend on extremely few external helpers (such as compat) """ from .compat import is_py2, builtin_str, str def to_native_string(string, encoding='ascii'): """Given a string object, regardless of type, returns a representation of that string in the native string type, encoding and decoding where necessary. This assumes ASCII unless told otherwise. """ if isinstance(string, builtin_str): out = string else: if is_py2: out = string.encode(encoding) else: out = string.decode(encoding) return out def unicode_is_ascii(u_string): """Determine if unicode string only contains ASCII characters. :param str u_string: unicode string to check. Must be unicode and not Python 2 `str`. :rtype: bool """ assert isinstance(u_string, str) try: u_string.encode('ascii') return True except UnicodeEncodeError: return False site-packages/pip/_vendor/requests/utils.py000064400000072541151464452720015120 0ustar00# -*- coding: utf-8 -*- """ requests.utils ~~~~~~~~~~~~~~ This module provides utility functions that are used within Requests that are also useful for external consumption. """ import codecs import contextlib import io import os import re import socket import struct import sys import tempfile import warnings import zipfile from .__version__ import __version__ from . import certs # to_native_string is unused here, but imported here for backwards compatibility from ._internal_utils import to_native_string from .compat import parse_http_list as _parse_list_header from .compat import ( quote, urlparse, bytes, str, OrderedDict, unquote, getproxies, proxy_bypass, urlunparse, basestring, integer_types, is_py3, proxy_bypass_environment, getproxies_environment, Mapping) from .cookies import cookiejar_from_dict from .structures import CaseInsensitiveDict from .exceptions import ( InvalidURL, InvalidHeader, FileModeWarning, UnrewindableBodyError) NETRC_FILES = ('.netrc', '_netrc') DEFAULT_CA_BUNDLE_PATH = certs.where() DEFAULT_PORTS = {'http': 80, 'https': 443} if sys.platform == 'win32': # provide a proxy_bypass version on Windows without DNS lookups def proxy_bypass_registry(host): try: if is_py3: import winreg else: import _winreg as winreg except ImportError: return False try: internetSettings = winreg.OpenKey(winreg.HKEY_CURRENT_USER, r'Software\Microsoft\Windows\CurrentVersion\Internet Settings') # ProxyEnable could be REG_SZ or REG_DWORD, normalizing it proxyEnable = int(winreg.QueryValueEx(internetSettings, 'ProxyEnable')[0]) # ProxyOverride is almost always a string proxyOverride = winreg.QueryValueEx(internetSettings, 'ProxyOverride')[0] except OSError: return False if not proxyEnable or not proxyOverride: return False # make a check value list from the registry entry: replace the # '' string by the localhost entry and the corresponding # canonical entry. proxyOverride = proxyOverride.split(';') # now check if we match one of the registry values. for test in proxyOverride: if test == '': if '.' not in host: return True test = test.replace(".", r"\.") # mask dots test = test.replace("*", r".*") # change glob sequence test = test.replace("?", r".") # change glob char if re.match(test, host, re.I): return True return False def proxy_bypass(host): # noqa """Return True, if the host should be bypassed. Checks proxy settings gathered from the environment, if specified, or the registry. """ if getproxies_environment(): return proxy_bypass_environment(host) else: return proxy_bypass_registry(host) def dict_to_sequence(d): """Returns an internal sequence dictionary update.""" if hasattr(d, 'items'): d = d.items() return d def super_len(o): total_length = None current_position = 0 if hasattr(o, '__len__'): total_length = len(o) elif hasattr(o, 'len'): total_length = o.len elif hasattr(o, 'fileno'): try: fileno = o.fileno() except io.UnsupportedOperation: pass else: total_length = os.fstat(fileno).st_size # Having used fstat to determine the file length, we need to # confirm that this file was opened up in binary mode. if 'b' not in o.mode: warnings.warn(( "Requests has determined the content-length for this " "request using the binary size of the file: however, the " "file has been opened in text mode (i.e. without the 'b' " "flag in the mode). This may lead to an incorrect " "content-length. In Requests 3.0, support will be removed " "for files in text mode."), FileModeWarning ) if hasattr(o, 'tell'): try: current_position = o.tell() except (OSError, IOError): # This can happen in some weird situations, such as when the file # is actually a special file descriptor like stdin. In this # instance, we don't know what the length is, so set it to zero and # let requests chunk it instead. if total_length is not None: current_position = total_length else: if hasattr(o, 'seek') and total_length is None: # StringIO and BytesIO have seek but no useable fileno try: # seek to end of file o.seek(0, 2) total_length = o.tell() # seek back to current position to support # partially read file-like objects o.seek(current_position or 0) except (OSError, IOError): total_length = 0 if total_length is None: total_length = 0 return max(0, total_length - current_position) def get_netrc_auth(url, raise_errors=False): """Returns the Requests tuple auth for a given url from netrc.""" try: from netrc import netrc, NetrcParseError netrc_path = None for f in NETRC_FILES: try: loc = os.path.expanduser('~/{}'.format(f)) except KeyError: # os.path.expanduser can fail when $HOME is undefined and # getpwuid fails. See https://bugs.python.org/issue20164 & # https://github.com/requests/requests/issues/1846 return if os.path.exists(loc): netrc_path = loc break # Abort early if there isn't one. if netrc_path is None: return ri = urlparse(url) # Strip port numbers from netloc. This weird `if...encode`` dance is # used for Python 3.2, which doesn't support unicode literals. splitstr = b':' if isinstance(url, str): splitstr = splitstr.decode('ascii') host = ri.netloc.split(splitstr)[0] try: _netrc = netrc(netrc_path).authenticators(host) if _netrc: # Return with login / password login_i = (0 if _netrc[0] else 1) return (_netrc[login_i], _netrc[2]) except (NetrcParseError, IOError): # If there was a parsing error or a permissions issue reading the file, # we'll just skip netrc auth unless explicitly asked to raise errors. if raise_errors: raise # AppEngine hackiness. except (ImportError, AttributeError): pass def guess_filename(obj): """Tries to guess the filename of the given object.""" name = getattr(obj, 'name', None) if (name and isinstance(name, basestring) and name[0] != '<' and name[-1] != '>'): return os.path.basename(name) def extract_zipped_paths(path): """Replace nonexistent paths that look like they refer to a member of a zip archive with the location of an extracted copy of the target, or else just return the provided path unchanged. """ if os.path.exists(path): # this is already a valid path, no need to do anything further return path # find the first valid part of the provided path and treat that as a zip archive # assume the rest of the path is the name of a member in the archive archive, member = os.path.split(path) while archive and not os.path.exists(archive): archive, prefix = os.path.split(archive) member = '/'.join([prefix, member]) if not zipfile.is_zipfile(archive): return path zip_file = zipfile.ZipFile(archive) if member not in zip_file.namelist(): return path # we have a valid zip archive and a valid member of that archive tmp = tempfile.gettempdir() extracted_path = os.path.join(tmp, *member.split('/')) if not os.path.exists(extracted_path): extracted_path = zip_file.extract(member, path=tmp) return extracted_path def from_key_val_list(value): """Take an object and test to see if it can be represented as a dictionary. Unless it can not be represented as such, return an OrderedDict, e.g., :: >>> from_key_val_list([('key', 'val')]) OrderedDict([('key', 'val')]) >>> from_key_val_list('string') ValueError: cannot encode objects that are not 2-tuples >>> from_key_val_list({'key': 'val'}) OrderedDict([('key', 'val')]) :rtype: OrderedDict """ if value is None: return None if isinstance(value, (str, bytes, bool, int)): raise ValueError('cannot encode objects that are not 2-tuples') return OrderedDict(value) def to_key_val_list(value): """Take an object and test to see if it can be represented as a dictionary. If it can be, return a list of tuples, e.g., :: >>> to_key_val_list([('key', 'val')]) [('key', 'val')] >>> to_key_val_list({'key': 'val'}) [('key', 'val')] >>> to_key_val_list('string') ValueError: cannot encode objects that are not 2-tuples. :rtype: list """ if value is None: return None if isinstance(value, (str, bytes, bool, int)): raise ValueError('cannot encode objects that are not 2-tuples') if isinstance(value, Mapping): value = value.items() return list(value) # From mitsuhiko/werkzeug (used with permission). def parse_list_header(value): """Parse lists as described by RFC 2068 Section 2. In particular, parse comma-separated lists where the elements of the list may include quoted-strings. A quoted-string could contain a comma. A non-quoted string could have quotes in the middle. Quotes are removed automatically after parsing. It basically works like :func:`parse_set_header` just that items may appear multiple times and case sensitivity is preserved. The return value is a standard :class:`list`: >>> parse_list_header('token, "quoted value"') ['token', 'quoted value'] To create a header from the :class:`list` again, use the :func:`dump_header` function. :param value: a string with a list header. :return: :class:`list` :rtype: list """ result = [] for item in _parse_list_header(value): if item[:1] == item[-1:] == '"': item = unquote_header_value(item[1:-1]) result.append(item) return result # From mitsuhiko/werkzeug (used with permission). def parse_dict_header(value): """Parse lists of key, value pairs as described by RFC 2068 Section 2 and convert them into a python dict: >>> d = parse_dict_header('foo="is a fish", bar="as well"') >>> type(d) is dict True >>> sorted(d.items()) [('bar', 'as well'), ('foo', 'is a fish')] If there is no value for a key it will be `None`: >>> parse_dict_header('key_without_value') {'key_without_value': None} To create a header from the :class:`dict` again, use the :func:`dump_header` function. :param value: a string with a dict header. :return: :class:`dict` :rtype: dict """ result = {} for item in _parse_list_header(value): if '=' not in item: result[item] = None continue name, value = item.split('=', 1) if value[:1] == value[-1:] == '"': value = unquote_header_value(value[1:-1]) result[name] = value return result # From mitsuhiko/werkzeug (used with permission). def unquote_header_value(value, is_filename=False): r"""Unquotes a header value. (Reversal of :func:`quote_header_value`). This does not use the real unquoting but what browsers are actually using for quoting. :param value: the header value to unquote. :rtype: str """ if value and value[0] == value[-1] == '"': # this is not the real unquoting, but fixing this so that the # RFC is met will result in bugs with internet explorer and # probably some other browsers as well. IE for example is # uploading files with "C:\foo\bar.txt" as filename value = value[1:-1] # if this is a filename and the starting characters look like # a UNC path, then just return the value without quotes. Using the # replace sequence below on a UNC path has the effect of turning # the leading double slash into a single slash and then # _fix_ie_filename() doesn't work correctly. See #458. if not is_filename or value[:2] != '\\\\': return value.replace('\\\\', '\\').replace('\\"', '"') return value def dict_from_cookiejar(cj): """Returns a key/value dictionary from a CookieJar. :param cj: CookieJar object to extract cookies from. :rtype: dict """ cookie_dict = {} for cookie in cj: cookie_dict[cookie.name] = cookie.value return cookie_dict def add_dict_to_cookiejar(cj, cookie_dict): """Returns a CookieJar from a key/value dictionary. :param cj: CookieJar to insert cookies into. :param cookie_dict: Dict of key/values to insert into CookieJar. :rtype: CookieJar """ return cookiejar_from_dict(cookie_dict, cj) def get_encodings_from_content(content): """Returns encodings from given content string. :param content: bytestring to extract encodings from. """ warnings.warn(( 'In requests 3.0, get_encodings_from_content will be removed. For ' 'more information, please see the discussion on issue #2266. (This' ' warning should only appear once.)'), DeprecationWarning) charset_re = re.compile(r']', flags=re.I) pragma_re = re.compile(r']', flags=re.I) xml_re = re.compile(r'^<\?xml.*?encoding=["\']*(.+?)["\'>]') return (charset_re.findall(content) + pragma_re.findall(content) + xml_re.findall(content)) def _parse_content_type_header(header): """Returns content type and parameters from given header :param header: string :return: tuple containing content type and dictionary of parameters """ tokens = header.split(';') content_type, params = tokens[0].strip(), tokens[1:] params_dict = {} items_to_strip = "\"' " for param in params: param = param.strip() if param: key, value = param, True index_of_equals = param.find("=") if index_of_equals != -1: key = param[:index_of_equals].strip(items_to_strip) value = param[index_of_equals + 1:].strip(items_to_strip) params_dict[key.lower()] = value return content_type, params_dict def get_encoding_from_headers(headers): """Returns encodings from given HTTP Header Dict. :param headers: dictionary to extract encoding from. :rtype: str """ content_type = headers.get('content-type') if not content_type: return None content_type, params = _parse_content_type_header(content_type) if 'charset' in params: return params['charset'].strip("'\"") if 'text' in content_type: return 'ISO-8859-1' def stream_decode_response_unicode(iterator, r): """Stream decodes a iterator.""" if r.encoding is None: for item in iterator: yield item return decoder = codecs.getincrementaldecoder(r.encoding)(errors='replace') for chunk in iterator: rv = decoder.decode(chunk) if rv: yield rv rv = decoder.decode(b'', final=True) if rv: yield rv def iter_slices(string, slice_length): """Iterate over slices of a string.""" pos = 0 if slice_length is None or slice_length <= 0: slice_length = len(string) while pos < len(string): yield string[pos:pos + slice_length] pos += slice_length def get_unicode_from_response(r): """Returns the requested content back in unicode. :param r: Response object to get unicode content from. Tried: 1. charset from content-type 2. fall back and replace all unicode characters :rtype: str """ warnings.warn(( 'In requests 3.0, get_unicode_from_response will be removed. For ' 'more information, please see the discussion on issue #2266. (This' ' warning should only appear once.)'), DeprecationWarning) tried_encodings = [] # Try charset from content-type encoding = get_encoding_from_headers(r.headers) if encoding: try: return str(r.content, encoding) except UnicodeError: tried_encodings.append(encoding) # Fall back: try: return str(r.content, encoding, errors='replace') except TypeError: return r.content # The unreserved URI characters (RFC 3986) UNRESERVED_SET = frozenset( "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz" + "0123456789-._~") def unquote_unreserved(uri): """Un-escape any percent-escape sequences in a URI that are unreserved characters. This leaves all reserved, illegal and non-ASCII bytes encoded. :rtype: str """ parts = uri.split('%') for i in range(1, len(parts)): h = parts[i][0:2] if len(h) == 2 and h.isalnum(): try: c = chr(int(h, 16)) except ValueError: raise InvalidURL("Invalid percent-escape sequence: '%s'" % h) if c in UNRESERVED_SET: parts[i] = c + parts[i][2:] else: parts[i] = '%' + parts[i] else: parts[i] = '%' + parts[i] return ''.join(parts) def requote_uri(uri): """Re-quote the given URI. This function passes the given URI through an unquote/quote cycle to ensure that it is fully and consistently quoted. :rtype: str """ safe_with_percent = "!#$%&'()*+,/:;=?@[]~" safe_without_percent = "!#$&'()*+,/:;=?@[]~" try: # Unquote only the unreserved characters # Then quote only illegal characters (do not quote reserved, # unreserved, or '%') return quote(unquote_unreserved(uri), safe=safe_with_percent) except InvalidURL: # We couldn't unquote the given URI, so let's try quoting it, but # there may be unquoted '%'s in the URI. We need to make sure they're # properly quoted so they do not cause issues elsewhere. return quote(uri, safe=safe_without_percent) def address_in_network(ip, net): """This function allows you to check if an IP belongs to a network subnet Example: returns True if ip = 192.168.1.1 and net = 192.168.1.0/24 returns False if ip = 192.168.1.1 and net = 192.168.100.0/24 :rtype: bool """ ipaddr = struct.unpack('=L', socket.inet_aton(ip))[0] netaddr, bits = net.split('/') netmask = struct.unpack('=L', socket.inet_aton(dotted_netmask(int(bits))))[0] network = struct.unpack('=L', socket.inet_aton(netaddr))[0] & netmask return (ipaddr & netmask) == (network & netmask) def dotted_netmask(mask): """Converts mask from /xx format to xxx.xxx.xxx.xxx Example: if mask is 24 function returns 255.255.255.0 :rtype: str """ bits = 0xffffffff ^ (1 << 32 - mask) - 1 return socket.inet_ntoa(struct.pack('>I', bits)) def is_ipv4_address(string_ip): """ :rtype: bool """ try: socket.inet_aton(string_ip) except socket.error: return False return True def is_valid_cidr(string_network): """ Very simple check of the cidr format in no_proxy variable. :rtype: bool """ if string_network.count('/') == 1: try: mask = int(string_network.split('/')[1]) except ValueError: return False if mask < 1 or mask > 32: return False try: socket.inet_aton(string_network.split('/')[0]) except socket.error: return False else: return False return True @contextlib.contextmanager def set_environ(env_name, value): """Set the environment variable 'env_name' to 'value' Save previous value, yield, and then restore the previous value stored in the environment variable 'env_name'. If 'value' is None, do nothing""" value_changed = value is not None if value_changed: old_value = os.environ.get(env_name) os.environ[env_name] = value try: yield finally: if value_changed: if old_value is None: del os.environ[env_name] else: os.environ[env_name] = old_value def should_bypass_proxies(url, no_proxy): """ Returns whether we should bypass proxies or not. :rtype: bool """ # Prioritize lowercase environment variables over uppercase # to keep a consistent behaviour with other http projects (curl, wget). get_proxy = lambda k: os.environ.get(k) or os.environ.get(k.upper()) # First check whether no_proxy is defined. If it is, check that the URL # we're getting isn't in the no_proxy list. no_proxy_arg = no_proxy if no_proxy is None: no_proxy = get_proxy('no_proxy') parsed = urlparse(url) if parsed.hostname is None: # URLs don't always have hostnames, e.g. file:/// urls. return True if no_proxy: # We need to check whether we match here. We need to see if we match # the end of the hostname, both with and without the port. no_proxy = ( host for host in no_proxy.replace(' ', '').split(',') if host ) if is_ipv4_address(parsed.hostname): for proxy_ip in no_proxy: if is_valid_cidr(proxy_ip): if address_in_network(parsed.hostname, proxy_ip): return True elif parsed.hostname == proxy_ip: # If no_proxy ip was defined in plain IP notation instead of cidr notation & # matches the IP of the index return True else: host_with_port = parsed.hostname if parsed.port: host_with_port += ':{}'.format(parsed.port) for host in no_proxy: if parsed.hostname.endswith(host) or host_with_port.endswith(host): # The URL does match something in no_proxy, so we don't want # to apply the proxies on this URL. return True with set_environ('no_proxy', no_proxy_arg): # parsed.hostname can be `None` in cases such as a file URI. try: bypass = proxy_bypass(parsed.hostname) except (TypeError, socket.gaierror): bypass = False if bypass: return True return False def get_environ_proxies(url, no_proxy=None): """ Return a dict of environment proxies. :rtype: dict """ if should_bypass_proxies(url, no_proxy=no_proxy): return {} else: return getproxies() def select_proxy(url, proxies): """Select a proxy for the url, if applicable. :param url: The url being for the request :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs """ proxies = proxies or {} urlparts = urlparse(url) if urlparts.hostname is None: return proxies.get(urlparts.scheme, proxies.get('all')) proxy_keys = [ urlparts.scheme + '://' + urlparts.hostname, urlparts.scheme, 'all://' + urlparts.hostname, 'all', ] proxy = None for proxy_key in proxy_keys: if proxy_key in proxies: proxy = proxies[proxy_key] break return proxy def default_user_agent(name="python-requests"): """ Return a string representing the default user agent. :rtype: str """ return '%s/%s' % (name, __version__) def default_headers(): """ :rtype: requests.structures.CaseInsensitiveDict """ return CaseInsensitiveDict({ 'User-Agent': default_user_agent(), 'Accept-Encoding': ', '.join(('gzip', 'deflate')), 'Accept': '*/*', 'Connection': 'keep-alive', }) def parse_header_links(value): """Return a list of parsed link headers proxies. i.e. Link: ; rel=front; type="image/jpeg",; rel=back;type="image/jpeg" :rtype: list """ links = [] replace_chars = ' \'"' value = value.strip(replace_chars) if not value: return links for val in re.split(', *<', value): try: url, params = val.split(';', 1) except ValueError: url, params = val, '' link = {'url': url.strip('<> \'"')} for param in params.split(';'): try: key, value = param.split('=') except ValueError: break link[key.strip(replace_chars)] = value.strip(replace_chars) links.append(link) return links # Null bytes; no need to recreate these on each call to guess_json_utf _null = '\x00'.encode('ascii') # encoding to ASCII for Python 3 _null2 = _null * 2 _null3 = _null * 3 def guess_json_utf(data): """ :rtype: str """ # JSON always starts with two ASCII characters, so detection is as # easy as counting the nulls and from their location and count # determine the encoding. Also detect a BOM, if present. sample = data[:4] if sample in (codecs.BOM_UTF32_LE, codecs.BOM_UTF32_BE): return 'utf-32' # BOM included if sample[:3] == codecs.BOM_UTF8: return 'utf-8-sig' # BOM included, MS style (discouraged) if sample[:2] in (codecs.BOM_UTF16_LE, codecs.BOM_UTF16_BE): return 'utf-16' # BOM included nullcount = sample.count(_null) if nullcount == 0: return 'utf-8' if nullcount == 2: if sample[::2] == _null2: # 1st and 3rd are null return 'utf-16-be' if sample[1::2] == _null2: # 2nd and 4th are null return 'utf-16-le' # Did not detect 2 valid UTF-16 ascii-range characters if nullcount == 3: if sample[:3] == _null3: return 'utf-32-be' if sample[1:] == _null3: return 'utf-32-le' # Did not detect a valid UTF-32 ascii-range character return None def prepend_scheme_if_needed(url, new_scheme): """Given a URL that may or may not have a scheme, prepend the given scheme. Does not replace a present scheme with the one provided as an argument. :rtype: str """ scheme, netloc, path, params, query, fragment = urlparse(url, new_scheme) # urlparse is a finicky beast, and sometimes decides that there isn't a # netloc present. Assume that it's being over-cautious, and switch netloc # and path if urlparse decided there was no netloc. if not netloc: netloc, path = path, netloc return urlunparse((scheme, netloc, path, params, query, fragment)) def get_auth_from_url(url): """Given a url with authentication components, extract them into a tuple of username,password. :rtype: (str,str) """ parsed = urlparse(url) try: auth = (unquote(parsed.username), unquote(parsed.password)) except (AttributeError, TypeError): auth = ('', '') return auth # Moved outside of function to avoid recompile every call _CLEAN_HEADER_REGEX_BYTE = re.compile(b'^\\S[^\\r\\n]*$|^$') _CLEAN_HEADER_REGEX_STR = re.compile(r'^\S[^\r\n]*$|^$') def check_header_validity(header): """Verifies that header value is a string which doesn't contain leading whitespace or return characters. This prevents unintended header injection. :param header: tuple, in the format (name, value). """ name, value = header if isinstance(value, bytes): pat = _CLEAN_HEADER_REGEX_BYTE else: pat = _CLEAN_HEADER_REGEX_STR try: if not pat.match(value): raise InvalidHeader("Invalid return character or leading space in header: %s" % name) except TypeError: raise InvalidHeader("Value for header {%s: %s} must be of type str or " "bytes, not %s" % (name, value, type(value))) def urldefragauth(url): """ Given a url remove the fragment and the authentication part. :rtype: str """ scheme, netloc, path, params, query, fragment = urlparse(url) # see func:`prepend_scheme_if_needed` if not netloc: netloc, path = path, netloc netloc = netloc.rsplit('@', 1)[-1] return urlunparse((scheme, netloc, path, params, query, '')) def rewind_body(prepared_request): """Move file pointer back to its recorded starting position so it can be read again on redirect. """ body_seek = getattr(prepared_request.body, 'seek', None) if body_seek is not None and isinstance(prepared_request._body_position, integer_types): try: body_seek(prepared_request._body_position) except (IOError, OSError): raise UnrewindableBodyError("An error occurred when rewinding request " "body for redirect.") else: raise UnrewindableBodyError("Unable to rewind request body for redirect.") site-packages/pip/_vendor/requests/adapters.py000064400000052054151464452720015560 0ustar00# -*- coding: utf-8 -*- """ requests.adapters ~~~~~~~~~~~~~~~~~ This module contains the transport adapters that Requests uses to define and maintain connections. """ import os.path import socket from pip._vendor.urllib3.poolmanager import PoolManager, proxy_from_url from pip._vendor.urllib3.response import HTTPResponse from pip._vendor.urllib3.util import parse_url from pip._vendor.urllib3.util import Timeout as TimeoutSauce from pip._vendor.urllib3.util.retry import Retry from pip._vendor.urllib3.exceptions import ClosedPoolError from pip._vendor.urllib3.exceptions import ConnectTimeoutError from pip._vendor.urllib3.exceptions import HTTPError as _HTTPError from pip._vendor.urllib3.exceptions import MaxRetryError from pip._vendor.urllib3.exceptions import NewConnectionError from pip._vendor.urllib3.exceptions import ProxyError as _ProxyError from pip._vendor.urllib3.exceptions import ProtocolError from pip._vendor.urllib3.exceptions import ReadTimeoutError from pip._vendor.urllib3.exceptions import SSLError as _SSLError from pip._vendor.urllib3.exceptions import ResponseError from pip._vendor.urllib3.exceptions import LocationValueError from .models import Response from .compat import urlparse, basestring from .utils import (DEFAULT_CA_BUNDLE_PATH, extract_zipped_paths, get_encoding_from_headers, prepend_scheme_if_needed, get_auth_from_url, urldefragauth, select_proxy) from .structures import CaseInsensitiveDict from .cookies import extract_cookies_to_jar from .exceptions import (ConnectionError, ConnectTimeout, ReadTimeout, SSLError, ProxyError, RetryError, InvalidSchema, InvalidProxyURL, InvalidURL) from .auth import _basic_auth_str try: from pip._vendor.urllib3.contrib.socks import SOCKSProxyManager except ImportError: def SOCKSProxyManager(*args, **kwargs): raise InvalidSchema("Missing dependencies for SOCKS support.") DEFAULT_POOLBLOCK = False DEFAULT_POOLSIZE = 10 DEFAULT_RETRIES = 0 DEFAULT_POOL_TIMEOUT = None class BaseAdapter(object): """The Base Transport Adapter""" def __init__(self): super(BaseAdapter, self).__init__() def send(self, request, stream=False, timeout=None, verify=True, cert=None, proxies=None): """Sends PreparedRequest object. Returns Response object. :param request: The :class:`PreparedRequest ` being sent. :param stream: (optional) Whether to stream the request content. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: (optional) Any user-provided SSL certificate to be trusted. :param proxies: (optional) The proxies dictionary to apply to the request. """ raise NotImplementedError def close(self): """Cleans up adapter specific items.""" raise NotImplementedError class HTTPAdapter(BaseAdapter): """The built-in HTTP Adapter for urllib3. Provides a general-case interface for Requests sessions to contact HTTP and HTTPS urls by implementing the Transport Adapter interface. This class will usually be created by the :class:`Session ` class under the covers. :param pool_connections: The number of urllib3 connection pools to cache. :param pool_maxsize: The maximum number of connections to save in the pool. :param max_retries: The maximum number of retries each connection should attempt. Note, this applies only to failed DNS lookups, socket connections and connection timeouts, never to requests where data has made it to the server. By default, Requests does not retry failed connections. If you need granular control over the conditions under which we retry a request, import urllib3's ``Retry`` class and pass that instead. :param pool_block: Whether the connection pool should block for connections. Usage:: >>> import requests >>> s = requests.Session() >>> a = requests.adapters.HTTPAdapter(max_retries=3) >>> s.mount('http://', a) """ __attrs__ = ['max_retries', 'config', '_pool_connections', '_pool_maxsize', '_pool_block'] def __init__(self, pool_connections=DEFAULT_POOLSIZE, pool_maxsize=DEFAULT_POOLSIZE, max_retries=DEFAULT_RETRIES, pool_block=DEFAULT_POOLBLOCK): if max_retries == DEFAULT_RETRIES: self.max_retries = Retry(0, read=False) else: self.max_retries = Retry.from_int(max_retries) self.config = {} self.proxy_manager = {} super(HTTPAdapter, self).__init__() self._pool_connections = pool_connections self._pool_maxsize = pool_maxsize self._pool_block = pool_block self.init_poolmanager(pool_connections, pool_maxsize, block=pool_block) def __getstate__(self): return {attr: getattr(self, attr, None) for attr in self.__attrs__} def __setstate__(self, state): # Can't handle by adding 'proxy_manager' to self.__attrs__ because # self.poolmanager uses a lambda function, which isn't pickleable. self.proxy_manager = {} self.config = {} for attr, value in state.items(): setattr(self, attr, value) self.init_poolmanager(self._pool_connections, self._pool_maxsize, block=self._pool_block) def init_poolmanager(self, connections, maxsize, block=DEFAULT_POOLBLOCK, **pool_kwargs): """Initializes a urllib3 PoolManager. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param connections: The number of urllib3 connection pools to cache. :param maxsize: The maximum number of connections to save in the pool. :param block: Block when no free connections are available. :param pool_kwargs: Extra keyword arguments used to initialize the Pool Manager. """ # save these values for pickling self._pool_connections = connections self._pool_maxsize = maxsize self._pool_block = block self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, strict=True, **pool_kwargs) def proxy_manager_for(self, proxy, **proxy_kwargs): """Return urllib3 ProxyManager for the given proxy. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param proxy: The proxy to return a urllib3 ProxyManager for. :param proxy_kwargs: Extra keyword arguments used to configure the Proxy Manager. :returns: ProxyManager :rtype: urllib3.ProxyManager """ if proxy in self.proxy_manager: manager = self.proxy_manager[proxy] elif proxy.lower().startswith('socks'): username, password = get_auth_from_url(proxy) manager = self.proxy_manager[proxy] = SOCKSProxyManager( proxy, username=username, password=password, num_pools=self._pool_connections, maxsize=self._pool_maxsize, block=self._pool_block, **proxy_kwargs ) else: proxy_headers = self.proxy_headers(proxy) manager = self.proxy_manager[proxy] = proxy_from_url( proxy, proxy_headers=proxy_headers, num_pools=self._pool_connections, maxsize=self._pool_maxsize, block=self._pool_block, **proxy_kwargs) return manager def cert_verify(self, conn, url, verify, cert): """Verify a SSL certificate. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param conn: The urllib3 connection object associated with the cert. :param url: The requested URL. :param verify: Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: The SSL certificate to verify. """ if url.lower().startswith('https') and verify: cert_loc = None # Allow self-specified cert location. if verify is not True: cert_loc = verify if not cert_loc: cert_loc = extract_zipped_paths(DEFAULT_CA_BUNDLE_PATH) if not cert_loc or not os.path.exists(cert_loc): raise IOError("Could not find a suitable TLS CA certificate bundle, " "invalid path: {}".format(cert_loc)) conn.cert_reqs = 'CERT_REQUIRED' if not os.path.isdir(cert_loc): conn.ca_certs = cert_loc else: conn.ca_cert_dir = cert_loc else: conn.cert_reqs = 'CERT_NONE' conn.ca_certs = None conn.ca_cert_dir = None if cert: if not isinstance(cert, basestring): conn.cert_file = cert[0] conn.key_file = cert[1] else: conn.cert_file = cert conn.key_file = None if conn.cert_file and not os.path.exists(conn.cert_file): raise IOError("Could not find the TLS certificate file, " "invalid path: {}".format(conn.cert_file)) if conn.key_file and not os.path.exists(conn.key_file): raise IOError("Could not find the TLS key file, " "invalid path: {}".format(conn.key_file)) def build_response(self, req, resp): """Builds a :class:`Response ` object from a urllib3 response. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter ` :param req: The :class:`PreparedRequest ` used to generate the response. :param resp: The urllib3 response object. :rtype: requests.Response """ response = Response() # Fallback to None if there's no status_code, for whatever reason. response.status_code = getattr(resp, 'status', None) # Make headers case-insensitive. response.headers = CaseInsensitiveDict(getattr(resp, 'headers', {})) # Set encoding. response.encoding = get_encoding_from_headers(response.headers) response.raw = resp response.reason = response.raw.reason if isinstance(req.url, bytes): response.url = req.url.decode('utf-8') else: response.url = req.url # Add new cookies from the server. extract_cookies_to_jar(response.cookies, req, resp) # Give the Response some context. response.request = req response.connection = self return response def get_connection(self, url, proxies=None): """Returns a urllib3 connection for the given URL. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param url: The URL to connect to. :param proxies: (optional) A Requests-style dictionary of proxies used on this request. :rtype: urllib3.ConnectionPool """ proxy = select_proxy(url, proxies) if proxy: proxy = prepend_scheme_if_needed(proxy, 'http') proxy_url = parse_url(proxy) if not proxy_url.host: raise InvalidProxyURL("Please check proxy URL. It is malformed" " and could be missing the host.") proxy_manager = self.proxy_manager_for(proxy) conn = proxy_manager.connection_from_url(url) else: # Only scheme should be lower case parsed = urlparse(url) url = parsed.geturl() conn = self.poolmanager.connection_from_url(url) return conn def close(self): """Disposes of any internal state. Currently, this closes the PoolManager and any active ProxyManager, which closes any pooled connections. """ self.poolmanager.clear() for proxy in self.proxy_manager.values(): proxy.clear() def request_url(self, request, proxies): """Obtain the url to use when making the final request. If the message is being sent through a HTTP proxy, the full URL has to be used. Otherwise, we should only use the path portion of the URL. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param request: The :class:`PreparedRequest ` being sent. :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs. :rtype: str """ proxy = select_proxy(request.url, proxies) scheme = urlparse(request.url).scheme is_proxied_http_request = (proxy and scheme != 'https') using_socks_proxy = False if proxy: proxy_scheme = urlparse(proxy).scheme.lower() using_socks_proxy = proxy_scheme.startswith('socks') url = request.path_url if is_proxied_http_request and not using_socks_proxy: url = urldefragauth(request.url) return url def add_headers(self, request, **kwargs): """Add any headers needed by the connection. As of v2.0 this does nothing by default, but is left for overriding by users that subclass the :class:`HTTPAdapter `. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param request: The :class:`PreparedRequest ` to add headers to. :param kwargs: The keyword arguments from the call to send(). """ pass def proxy_headers(self, proxy): """Returns a dictionary of the headers to add to any request sent through a proxy. This works with urllib3 magic to ensure that they are correctly sent to the proxy, rather than in a tunnelled request if CONNECT is being used. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param proxy: The url of the proxy being used for this request. :rtype: dict """ headers = {} username, password = get_auth_from_url(proxy) if username: headers['Proxy-Authorization'] = _basic_auth_str(username, password) return headers def send(self, request, stream=False, timeout=None, verify=True, cert=None, proxies=None): """Sends PreparedRequest object. Returns Response object. :param request: The :class:`PreparedRequest ` being sent. :param stream: (optional) Whether to stream the request content. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple or urllib3 Timeout object :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: (optional) Any user-provided SSL certificate to be trusted. :param proxies: (optional) The proxies dictionary to apply to the request. :rtype: requests.Response """ try: conn = self.get_connection(request.url, proxies) except LocationValueError as e: raise InvalidURL(e, request=request) self.cert_verify(conn, request.url, verify, cert) url = self.request_url(request, proxies) self.add_headers(request, stream=stream, timeout=timeout, verify=verify, cert=cert, proxies=proxies) chunked = not (request.body is None or 'Content-Length' in request.headers) if isinstance(timeout, tuple): try: connect, read = timeout timeout = TimeoutSauce(connect=connect, read=read) except ValueError as e: # this may raise a string formatting error. err = ("Invalid timeout {}. Pass a (connect, read) " "timeout tuple, or a single float to set " "both timeouts to the same value".format(timeout)) raise ValueError(err) elif isinstance(timeout, TimeoutSauce): pass else: timeout = TimeoutSauce(connect=timeout, read=timeout) try: if not chunked: resp = conn.urlopen( method=request.method, url=url, body=request.body, headers=request.headers, redirect=False, assert_same_host=False, preload_content=False, decode_content=False, retries=self.max_retries, timeout=timeout ) # Send the request. else: if hasattr(conn, 'proxy_pool'): conn = conn.proxy_pool low_conn = conn._get_conn(timeout=DEFAULT_POOL_TIMEOUT) try: low_conn.putrequest(request.method, url, skip_accept_encoding=True) for header, value in request.headers.items(): low_conn.putheader(header, value) low_conn.endheaders() for i in request.body: low_conn.send(hex(len(i))[2:].encode('utf-8')) low_conn.send(b'\r\n') low_conn.send(i) low_conn.send(b'\r\n') low_conn.send(b'0\r\n\r\n') # Receive the response from the server try: # For Python 2.7, use buffering of HTTP responses r = low_conn.getresponse(buffering=True) except TypeError: # For compatibility with Python 3.3+ r = low_conn.getresponse() resp = HTTPResponse.from_httplib( r, pool=conn, connection=low_conn, preload_content=False, decode_content=False ) except: # If we hit any problems here, clean up the connection. # Then, reraise so that we can handle the actual exception. low_conn.close() raise except (ProtocolError, socket.error) as err: raise ConnectionError(err, request=request) except MaxRetryError as e: if isinstance(e.reason, ConnectTimeoutError): # TODO: Remove this in 3.0.0: see #2811 if not isinstance(e.reason, NewConnectionError): raise ConnectTimeout(e, request=request) if isinstance(e.reason, ResponseError): raise RetryError(e, request=request) if isinstance(e.reason, _ProxyError): raise ProxyError(e, request=request) if isinstance(e.reason, _SSLError): # This branch is for urllib3 v1.22 and later. raise SSLError(e, request=request) raise ConnectionError(e, request=request) except ClosedPoolError as e: raise ConnectionError(e, request=request) except _ProxyError as e: raise ProxyError(e) except (_SSLError, _HTTPError) as e: if isinstance(e, _SSLError): # This branch is for urllib3 versions earlier than v1.22 raise SSLError(e, request=request) elif isinstance(e, ReadTimeoutError): raise ReadTimeout(e, request=request) else: raise return self.build_response(request, resp) site-packages/pip/_vendor/requests/sessions.py000064400000071224151464452720015623 0ustar00# -*- coding: utf-8 -*- """ requests.session ~~~~~~~~~~~~~~~~ This module provides a Session object to manage and persist settings across requests (cookies, auth, proxies). """ import os import sys import time from datetime import timedelta from .auth import _basic_auth_str from .compat import cookielib, is_py3, OrderedDict, urljoin, urlparse, Mapping from .cookies import ( cookiejar_from_dict, extract_cookies_to_jar, RequestsCookieJar, merge_cookies) from .models import Request, PreparedRequest, DEFAULT_REDIRECT_LIMIT from .hooks import default_hooks, dispatch_hook from ._internal_utils import to_native_string from .utils import to_key_val_list, default_headers, DEFAULT_PORTS from .exceptions import ( TooManyRedirects, InvalidSchema, ChunkedEncodingError, ContentDecodingError) from .structures import CaseInsensitiveDict from .adapters import HTTPAdapter from .utils import ( requote_uri, get_environ_proxies, get_netrc_auth, should_bypass_proxies, get_auth_from_url, rewind_body ) from .status_codes import codes # formerly defined here, reexposed here for backward compatibility from .models import REDIRECT_STATI # Preferred clock, based on which one is more accurate on a given system. if sys.platform == 'win32': try: # Python 3.4+ preferred_clock = time.perf_counter except AttributeError: # Earlier than Python 3. preferred_clock = time.clock else: preferred_clock = time.time def merge_setting(request_setting, session_setting, dict_class=OrderedDict): """Determines appropriate setting for a given request, taking into account the explicit setting on that request, and the setting in the session. If a setting is a dictionary, they will be merged together using `dict_class` """ if session_setting is None: return request_setting if request_setting is None: return session_setting # Bypass if not a dictionary (e.g. verify) if not ( isinstance(session_setting, Mapping) and isinstance(request_setting, Mapping) ): return request_setting merged_setting = dict_class(to_key_val_list(session_setting)) merged_setting.update(to_key_val_list(request_setting)) # Remove keys that are set to None. Extract keys first to avoid altering # the dictionary during iteration. none_keys = [k for (k, v) in merged_setting.items() if v is None] for key in none_keys: del merged_setting[key] return merged_setting def merge_hooks(request_hooks, session_hooks, dict_class=OrderedDict): """Properly merges both requests and session hooks. This is necessary because when request_hooks == {'response': []}, the merge breaks Session hooks entirely. """ if session_hooks is None or session_hooks.get('response') == []: return request_hooks if request_hooks is None or request_hooks.get('response') == []: return session_hooks return merge_setting(request_hooks, session_hooks, dict_class) class SessionRedirectMixin(object): def get_redirect_target(self, resp): """Receives a Response. Returns a redirect URI or ``None``""" # Due to the nature of how requests processes redirects this method will # be called at least once upon the original response and at least twice # on each subsequent redirect response (if any). # If a custom mixin is used to handle this logic, it may be advantageous # to cache the redirect location onto the response object as a private # attribute. if resp.is_redirect: location = resp.headers['location'] # Currently the underlying http module on py3 decode headers # in latin1, but empirical evidence suggests that latin1 is very # rarely used with non-ASCII characters in HTTP headers. # It is more likely to get UTF8 header rather than latin1. # This causes incorrect handling of UTF8 encoded location headers. # To solve this, we re-encode the location in latin1. if is_py3: location = location.encode('latin1') return to_native_string(location, 'utf8') return None def should_strip_auth(self, old_url, new_url): """Decide whether Authorization header should be removed when redirecting""" old_parsed = urlparse(old_url) new_parsed = urlparse(new_url) if old_parsed.hostname != new_parsed.hostname: return True # Special case: allow http -> https redirect when using the standard # ports. This isn't specified by RFC 7235, but is kept to avoid # breaking backwards compatibility with older versions of requests # that allowed any redirects on the same host. if (old_parsed.scheme == 'http' and old_parsed.port in (80, None) and new_parsed.scheme == 'https' and new_parsed.port in (443, None)): return False # Handle default port usage corresponding to scheme. changed_port = old_parsed.port != new_parsed.port changed_scheme = old_parsed.scheme != new_parsed.scheme default_port = (DEFAULT_PORTS.get(old_parsed.scheme, None), None) if (not changed_scheme and old_parsed.port in default_port and new_parsed.port in default_port): return False # Standard case: root URI must match return changed_port or changed_scheme def resolve_redirects(self, resp, req, stream=False, timeout=None, verify=True, cert=None, proxies=None, yield_requests=False, **adapter_kwargs): """Receives a Response. Returns a generator of Responses or Requests.""" hist = [] # keep track of history url = self.get_redirect_target(resp) previous_fragment = urlparse(req.url).fragment while url: prepared_request = req.copy() # Update history and keep track of redirects. # resp.history must ignore the original request in this loop hist.append(resp) resp.history = hist[1:] try: resp.content # Consume socket so it can be released except (ChunkedEncodingError, ContentDecodingError, RuntimeError): resp.raw.read(decode_content=False) if len(resp.history) >= self.max_redirects: raise TooManyRedirects('Exceeded %s redirects.' % self.max_redirects, response=resp) # Release the connection back into the pool. resp.close() # Handle redirection without scheme (see: RFC 1808 Section 4) if url.startswith('//'): parsed_rurl = urlparse(resp.url) url = '%s:%s' % (to_native_string(parsed_rurl.scheme), url) # Normalize url case and attach previous fragment if needed (RFC 7231 7.1.2) parsed = urlparse(url) if parsed.fragment == '' and previous_fragment: parsed = parsed._replace(fragment=previous_fragment) elif parsed.fragment: previous_fragment = parsed.fragment url = parsed.geturl() # Facilitate relative 'location' headers, as allowed by RFC 7231. # (e.g. '/path/to/resource' instead of 'http://domain.tld/path/to/resource') # Compliant with RFC3986, we percent encode the url. if not parsed.netloc: url = urljoin(resp.url, requote_uri(url)) else: url = requote_uri(url) prepared_request.url = to_native_string(url) self.rebuild_method(prepared_request, resp) # https://github.com/requests/requests/issues/1084 if resp.status_code not in (codes.temporary_redirect, codes.permanent_redirect): # https://github.com/requests/requests/issues/3490 purged_headers = ('Content-Length', 'Content-Type', 'Transfer-Encoding') for header in purged_headers: prepared_request.headers.pop(header, None) prepared_request.body = None headers = prepared_request.headers try: del headers['Cookie'] except KeyError: pass # Extract any cookies sent on the response to the cookiejar # in the new request. Because we've mutated our copied prepared # request, use the old one that we haven't yet touched. extract_cookies_to_jar(prepared_request._cookies, req, resp.raw) merge_cookies(prepared_request._cookies, self.cookies) prepared_request.prepare_cookies(prepared_request._cookies) # Rebuild auth and proxy information. proxies = self.rebuild_proxies(prepared_request, proxies) self.rebuild_auth(prepared_request, resp) # A failed tell() sets `_body_position` to `object()`. This non-None # value ensures `rewindable` will be True, allowing us to raise an # UnrewindableBodyError, instead of hanging the connection. rewindable = ( prepared_request._body_position is not None and ('Content-Length' in headers or 'Transfer-Encoding' in headers) ) # Attempt to rewind consumed file-like object. if rewindable: rewind_body(prepared_request) # Override the original request. req = prepared_request if yield_requests: yield req else: resp = self.send( req, stream=stream, timeout=timeout, verify=verify, cert=cert, proxies=proxies, allow_redirects=False, **adapter_kwargs ) extract_cookies_to_jar(self.cookies, prepared_request, resp.raw) # extract redirect url, if any, for the next loop url = self.get_redirect_target(resp) yield resp def rebuild_auth(self, prepared_request, response): """When being redirected we may want to strip authentication from the request to avoid leaking credentials. This method intelligently removes and reapplies authentication where possible to avoid credential loss. """ headers = prepared_request.headers url = prepared_request.url if 'Authorization' in headers and self.should_strip_auth(response.request.url, url): # If we get redirected to a new host, we should strip out any # authentication headers. del headers['Authorization'] # .netrc might have more auth for us on our new host. new_auth = get_netrc_auth(url) if self.trust_env else None if new_auth is not None: prepared_request.prepare_auth(new_auth) return def rebuild_proxies(self, prepared_request, proxies): """This method re-evaluates the proxy configuration by considering the environment variables. If we are redirected to a URL covered by NO_PROXY, we strip the proxy configuration. Otherwise, we set missing proxy keys for this URL (in case they were stripped by a previous redirect). This method also replaces the Proxy-Authorization header where necessary. :rtype: dict """ proxies = proxies if proxies is not None else {} headers = prepared_request.headers url = prepared_request.url scheme = urlparse(url).scheme new_proxies = proxies.copy() no_proxy = proxies.get('no_proxy') bypass_proxy = should_bypass_proxies(url, no_proxy=no_proxy) if self.trust_env and not bypass_proxy: environ_proxies = get_environ_proxies(url, no_proxy=no_proxy) proxy = environ_proxies.get(scheme, environ_proxies.get('all')) if proxy: new_proxies.setdefault(scheme, proxy) if 'Proxy-Authorization' in headers: del headers['Proxy-Authorization'] try: username, password = get_auth_from_url(new_proxies[scheme]) except KeyError: username, password = None, None if username and password: headers['Proxy-Authorization'] = _basic_auth_str(username, password) return new_proxies def rebuild_method(self, prepared_request, response): """When being redirected we may want to change the method of the request based on certain specs or browser behavior. """ method = prepared_request.method # https://tools.ietf.org/html/rfc7231#section-6.4.4 if response.status_code == codes.see_other and method != 'HEAD': method = 'GET' # Do what the browsers do, despite standards... # First, turn 302s into GETs. if response.status_code == codes.found and method != 'HEAD': method = 'GET' # Second, if a POST is responded to with a 301, turn it into a GET. # This bizarre behaviour is explained in Issue 1704. if response.status_code == codes.moved and method == 'POST': method = 'GET' prepared_request.method = method class Session(SessionRedirectMixin): """A Requests session. Provides cookie persistence, connection-pooling, and configuration. Basic Usage:: >>> import requests >>> s = requests.Session() >>> s.get('https://httpbin.org/get') Or as a context manager:: >>> with requests.Session() as s: >>> s.get('https://httpbin.org/get') """ __attrs__ = [ 'headers', 'cookies', 'auth', 'proxies', 'hooks', 'params', 'verify', 'cert', 'prefetch', 'adapters', 'stream', 'trust_env', 'max_redirects', ] def __init__(self): #: A case-insensitive dictionary of headers to be sent on each #: :class:`Request ` sent from this #: :class:`Session `. self.headers = default_headers() #: Default Authentication tuple or object to attach to #: :class:`Request `. self.auth = None #: Dictionary mapping protocol or protocol and host to the URL of the proxy #: (e.g. {'http': 'foo.bar:3128', 'http://host.name': 'foo.bar:4012'}) to #: be used on each :class:`Request `. self.proxies = {} #: Event-handling hooks. self.hooks = default_hooks() #: Dictionary of querystring data to attach to each #: :class:`Request `. The dictionary values may be lists for #: representing multivalued query parameters. self.params = {} #: Stream response content default. self.stream = False #: SSL Verification default. self.verify = True #: SSL client certificate default, if String, path to ssl client #: cert file (.pem). If Tuple, ('cert', 'key') pair. self.cert = None #: Maximum number of redirects allowed. If the request exceeds this #: limit, a :class:`TooManyRedirects` exception is raised. #: This defaults to requests.models.DEFAULT_REDIRECT_LIMIT, which is #: 30. self.max_redirects = DEFAULT_REDIRECT_LIMIT #: Trust environment settings for proxy configuration, default #: authentication and similar. self.trust_env = True #: A CookieJar containing all currently outstanding cookies set on this #: session. By default it is a #: :class:`RequestsCookieJar `, but #: may be any other ``cookielib.CookieJar`` compatible object. self.cookies = cookiejar_from_dict({}) # Default connection adapters. self.adapters = OrderedDict() self.mount('https://', HTTPAdapter()) self.mount('http://', HTTPAdapter()) def __enter__(self): return self def __exit__(self, *args): self.close() def prepare_request(self, request): """Constructs a :class:`PreparedRequest ` for transmission and returns it. The :class:`PreparedRequest` has settings merged from the :class:`Request ` instance and those of the :class:`Session`. :param request: :class:`Request` instance to prepare with this session's settings. :rtype: requests.PreparedRequest """ cookies = request.cookies or {} # Bootstrap CookieJar. if not isinstance(cookies, cookielib.CookieJar): cookies = cookiejar_from_dict(cookies) # Merge with session cookies merged_cookies = merge_cookies( merge_cookies(RequestsCookieJar(), self.cookies), cookies) # Set environment's basic authentication if not explicitly set. auth = request.auth if self.trust_env and not auth and not self.auth: auth = get_netrc_auth(request.url) p = PreparedRequest() p.prepare( method=request.method.upper(), url=request.url, files=request.files, data=request.data, json=request.json, headers=merge_setting(request.headers, self.headers, dict_class=CaseInsensitiveDict), params=merge_setting(request.params, self.params), auth=merge_setting(auth, self.auth), cookies=merged_cookies, hooks=merge_hooks(request.hooks, self.hooks), ) return p def request(self, method, url, params=None, data=None, headers=None, cookies=None, files=None, auth=None, timeout=None, allow_redirects=True, proxies=None, hooks=None, stream=None, verify=None, cert=None, json=None): """Constructs a :class:`Request `, prepares it and sends it. Returns :class:`Response ` object. :param method: method for the new :class:`Request` object. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary or bytes to be sent in the query string for the :class:`Request`. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json to send in the body of the :class:`Request`. :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`. :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`. :param files: (optional) Dictionary of ``'filename': file-like-objects`` for multipart encoding upload. :param auth: (optional) Auth tuple or callable to enable Basic/Digest/Custom HTTP Auth. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param allow_redirects: (optional) Set to True by default. :type allow_redirects: bool :param proxies: (optional) Dictionary mapping protocol or protocol and hostname to the URL of the proxy. :param stream: (optional) whether to immediately download the response content. Defaults to ``False``. :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use. Defaults to ``True``. :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair. :rtype: requests.Response """ # Create the Request. req = Request( method=method.upper(), url=url, headers=headers, files=files, data=data or {}, json=json, params=params or {}, auth=auth, cookies=cookies, hooks=hooks, ) prep = self.prepare_request(req) proxies = proxies or {} settings = self.merge_environment_settings( prep.url, proxies, stream, verify, cert ) # Send the request. send_kwargs = { 'timeout': timeout, 'allow_redirects': allow_redirects, } send_kwargs.update(settings) resp = self.send(prep, **send_kwargs) return resp def get(self, url, **kwargs): r"""Sends a GET request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ kwargs.setdefault('allow_redirects', True) return self.request('GET', url, **kwargs) def options(self, url, **kwargs): r"""Sends a OPTIONS request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ kwargs.setdefault('allow_redirects', True) return self.request('OPTIONS', url, **kwargs) def head(self, url, **kwargs): r"""Sends a HEAD request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ kwargs.setdefault('allow_redirects', False) return self.request('HEAD', url, **kwargs) def post(self, url, data=None, json=None, **kwargs): r"""Sends a POST request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ return self.request('POST', url, data=data, json=json, **kwargs) def put(self, url, data=None, **kwargs): r"""Sends a PUT request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ return self.request('PUT', url, data=data, **kwargs) def patch(self, url, data=None, **kwargs): r"""Sends a PATCH request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ return self.request('PATCH', url, data=data, **kwargs) def delete(self, url, **kwargs): r"""Sends a DELETE request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response """ return self.request('DELETE', url, **kwargs) def send(self, request, **kwargs): """Send a given PreparedRequest. :rtype: requests.Response """ # Set defaults that the hooks can utilize to ensure they always have # the correct parameters to reproduce the previous request. kwargs.setdefault('stream', self.stream) kwargs.setdefault('verify', self.verify) kwargs.setdefault('cert', self.cert) kwargs.setdefault('proxies', self.proxies) # It's possible that users might accidentally send a Request object. # Guard against that specific failure case. if isinstance(request, Request): raise ValueError('You can only send PreparedRequests.') # Set up variables needed for resolve_redirects and dispatching of hooks allow_redirects = kwargs.pop('allow_redirects', True) stream = kwargs.get('stream') hooks = request.hooks # Get the appropriate adapter to use adapter = self.get_adapter(url=request.url) # Start time (approximately) of the request start = preferred_clock() # Send the request r = adapter.send(request, **kwargs) # Total elapsed time of the request (approximately) elapsed = preferred_clock() - start r.elapsed = timedelta(seconds=elapsed) # Response manipulation hooks r = dispatch_hook('response', hooks, r, **kwargs) # Persist cookies if r.history: # If the hooks create history then we want those cookies too for resp in r.history: extract_cookies_to_jar(self.cookies, resp.request, resp.raw) extract_cookies_to_jar(self.cookies, request, r.raw) # Redirect resolving generator. gen = self.resolve_redirects(r, request, **kwargs) # Resolve redirects if allowed. history = [resp for resp in gen] if allow_redirects else [] # Shuffle things around if there's history. if history: # Insert the first (original) request at the start history.insert(0, r) # Get the last request made r = history.pop() r.history = history # If redirects aren't being followed, store the response on the Request for Response.next(). if not allow_redirects: try: r._next = next(self.resolve_redirects(r, request, yield_requests=True, **kwargs)) except StopIteration: pass if not stream: r.content return r def merge_environment_settings(self, url, proxies, stream, verify, cert): """ Check the environment and merge it with some settings. :rtype: dict """ # Gather clues from the surrounding environment. if self.trust_env: # Set environment's proxies. no_proxy = proxies.get('no_proxy') if proxies is not None else None env_proxies = get_environ_proxies(url, no_proxy=no_proxy) for (k, v) in env_proxies.items(): proxies.setdefault(k, v) # Look for requests environment configuration and be compatible # with cURL. if verify is True or verify is None: verify = (os.environ.get('REQUESTS_CA_BUNDLE') or os.environ.get('CURL_CA_BUNDLE')) # Merge all the kwargs. proxies = merge_setting(proxies, self.proxies) stream = merge_setting(stream, self.stream) verify = merge_setting(verify, self.verify) cert = merge_setting(cert, self.cert) return {'verify': verify, 'proxies': proxies, 'stream': stream, 'cert': cert} def get_adapter(self, url): """ Returns the appropriate connection adapter for the given URL. :rtype: requests.adapters.BaseAdapter """ for (prefix, adapter) in self.adapters.items(): if url.lower().startswith(prefix.lower()): return adapter # Nothing matches :-/ raise InvalidSchema("No connection adapters were found for '%s'" % url) def close(self): """Closes all adapters and as such the session""" for v in self.adapters.values(): v.close() def mount(self, prefix, adapter): """Registers a connection adapter to a prefix. Adapters are sorted in descending order by prefix length. """ self.adapters[prefix] = adapter keys_to_move = [k for k in self.adapters if len(k) < len(prefix)] for key in keys_to_move: self.adapters[key] = self.adapters.pop(key) def __getstate__(self): state = {attr: getattr(self, attr, None) for attr in self.__attrs__} return state def __setstate__(self, state): for attr, value in state.items(): setattr(self, attr, value) def session(): """ Returns a :class:`Session` for context-management. .. deprecated:: 1.0.0 This method has been deprecated since version 1.0.0 and is only kept for backwards compatibility. New code should use :class:`~requests.sessions.Session` to create a session. This may be removed at a future date. :rtype: Session """ return Session() site-packages/pip/_vendor/requests/auth.py000064400000023736151464452720014723 0ustar00# -*- coding: utf-8 -*- """ requests.auth ~~~~~~~~~~~~~ This module contains the authentication handlers for Requests. """ import os import re import time import hashlib import threading import warnings from base64 import b64encode from .compat import urlparse, str, basestring from .cookies import extract_cookies_to_jar from ._internal_utils import to_native_string from .utils import parse_dict_header CONTENT_TYPE_FORM_URLENCODED = 'application/x-www-form-urlencoded' CONTENT_TYPE_MULTI_PART = 'multipart/form-data' def _basic_auth_str(username, password): """Returns a Basic Auth string.""" # "I want us to put a big-ol' comment on top of it that # says that this behaviour is dumb but we need to preserve # it because people are relying on it." # - Lukasa # # These are here solely to maintain backwards compatibility # for things like ints. This will be removed in 3.0.0. if not isinstance(username, basestring): warnings.warn( "Non-string usernames will no longer be supported in Requests " "3.0.0. Please convert the object you've passed in ({!r}) to " "a string or bytes object in the near future to avoid " "problems.".format(username), category=DeprecationWarning, ) username = str(username) if not isinstance(password, basestring): warnings.warn( "Non-string passwords will no longer be supported in Requests " "3.0.0. Please convert the object you've passed in ({!r}) to " "a string or bytes object in the near future to avoid " "problems.".format(password), category=DeprecationWarning, ) password = str(password) # -- End Removal -- if isinstance(username, str): username = username.encode('latin1') if isinstance(password, str): password = password.encode('latin1') authstr = 'Basic ' + to_native_string( b64encode(b':'.join((username, password))).strip() ) return authstr class AuthBase(object): """Base class that all auth implementations derive from""" def __call__(self, r): raise NotImplementedError('Auth hooks must be callable.') class HTTPBasicAuth(AuthBase): """Attaches HTTP Basic Authentication to the given Request object.""" def __init__(self, username, password): self.username = username self.password = password def __eq__(self, other): return all([ self.username == getattr(other, 'username', None), self.password == getattr(other, 'password', None) ]) def __ne__(self, other): return not self == other def __call__(self, r): r.headers['Authorization'] = _basic_auth_str(self.username, self.password) return r class HTTPProxyAuth(HTTPBasicAuth): """Attaches HTTP Proxy Authentication to a given Request object.""" def __call__(self, r): r.headers['Proxy-Authorization'] = _basic_auth_str(self.username, self.password) return r class HTTPDigestAuth(AuthBase): """Attaches HTTP Digest Authentication to the given Request object.""" def __init__(self, username, password): self.username = username self.password = password # Keep state in per-thread local storage self._thread_local = threading.local() def init_per_thread_state(self): # Ensure state is initialized just once per-thread if not hasattr(self._thread_local, 'init'): self._thread_local.init = True self._thread_local.last_nonce = '' self._thread_local.nonce_count = 0 self._thread_local.chal = {} self._thread_local.pos = None self._thread_local.num_401_calls = None def build_digest_header(self, method, url): """ :rtype: str """ realm = self._thread_local.chal['realm'] nonce = self._thread_local.chal['nonce'] qop = self._thread_local.chal.get('qop') algorithm = self._thread_local.chal.get('algorithm') opaque = self._thread_local.chal.get('opaque') hash_utf8 = None if algorithm is None: _algorithm = 'MD5' else: _algorithm = algorithm.upper() # lambdas assume digest modules are imported at the top level if _algorithm == 'MD5' or _algorithm == 'MD5-SESS': def md5_utf8(x): if isinstance(x, str): x = x.encode('utf-8') return hashlib.md5(x).hexdigest() hash_utf8 = md5_utf8 elif _algorithm == 'SHA': def sha_utf8(x): if isinstance(x, str): x = x.encode('utf-8') return hashlib.sha1(x).hexdigest() hash_utf8 = sha_utf8 elif _algorithm == 'SHA-256': def sha256_utf8(x): if isinstance(x, str): x = x.encode('utf-8') return hashlib.sha256(x).hexdigest() hash_utf8 = sha256_utf8 elif _algorithm == 'SHA-512': def sha512_utf8(x): if isinstance(x, str): x = x.encode('utf-8') return hashlib.sha512(x).hexdigest() hash_utf8 = sha512_utf8 KD = lambda s, d: hash_utf8("%s:%s" % (s, d)) if hash_utf8 is None: return None # XXX not implemented yet entdig = None p_parsed = urlparse(url) #: path is request-uri defined in RFC 2616 which should not be empty path = p_parsed.path or "/" if p_parsed.query: path += '?' + p_parsed.query A1 = '%s:%s:%s' % (self.username, realm, self.password) A2 = '%s:%s' % (method, path) HA1 = hash_utf8(A1) HA2 = hash_utf8(A2) if nonce == self._thread_local.last_nonce: self._thread_local.nonce_count += 1 else: self._thread_local.nonce_count = 1 ncvalue = '%08x' % self._thread_local.nonce_count s = str(self._thread_local.nonce_count).encode('utf-8') s += nonce.encode('utf-8') s += time.ctime().encode('utf-8') s += os.urandom(8) cnonce = (hashlib.sha1(s).hexdigest()[:16]) if _algorithm == 'MD5-SESS': HA1 = hash_utf8('%s:%s:%s' % (HA1, nonce, cnonce)) if not qop: respdig = KD(HA1, "%s:%s" % (nonce, HA2)) elif qop == 'auth' or 'auth' in qop.split(','): noncebit = "%s:%s:%s:%s:%s" % ( nonce, ncvalue, cnonce, 'auth', HA2 ) respdig = KD(HA1, noncebit) else: # XXX handle auth-int. return None self._thread_local.last_nonce = nonce # XXX should the partial digests be encoded too? base = 'username="%s", realm="%s", nonce="%s", uri="%s", ' \ 'response="%s"' % (self.username, realm, nonce, path, respdig) if opaque: base += ', opaque="%s"' % opaque if algorithm: base += ', algorithm="%s"' % algorithm if entdig: base += ', digest="%s"' % entdig if qop: base += ', qop="auth", nc=%s, cnonce="%s"' % (ncvalue, cnonce) return 'Digest %s' % (base) def handle_redirect(self, r, **kwargs): """Reset num_401_calls counter on redirects.""" if r.is_redirect: self._thread_local.num_401_calls = 1 def handle_401(self, r, **kwargs): """ Takes the given response and tries digest-auth, if needed. :rtype: requests.Response """ # If response is not 4xx, do not auth # See https://github.com/requests/requests/issues/3772 if not 400 <= r.status_code < 500: self._thread_local.num_401_calls = 1 return r if self._thread_local.pos is not None: # Rewind the file position indicator of the body to where # it was to resend the request. r.request.body.seek(self._thread_local.pos) s_auth = r.headers.get('www-authenticate', '') if 'digest' in s_auth.lower() and self._thread_local.num_401_calls < 2: self._thread_local.num_401_calls += 1 pat = re.compile(r'digest ', flags=re.IGNORECASE) self._thread_local.chal = parse_dict_header(pat.sub('', s_auth, count=1)) # Consume content and release the original connection # to allow our new request to reuse the same one. r.content r.close() prep = r.request.copy() extract_cookies_to_jar(prep._cookies, r.request, r.raw) prep.prepare_cookies(prep._cookies) prep.headers['Authorization'] = self.build_digest_header( prep.method, prep.url) _r = r.connection.send(prep, **kwargs) _r.history.append(r) _r.request = prep return _r self._thread_local.num_401_calls = 1 return r def __call__(self, r): # Initialize per-thread state, if needed self.init_per_thread_state() # If we have a saved nonce, skip the 401 if self._thread_local.last_nonce: r.headers['Authorization'] = self.build_digest_header(r.method, r.url) try: self._thread_local.pos = r.body.tell() except AttributeError: # In the case of HTTPDigestAuth being reused and the body of # the previous request was a file-like object, pos has the # file position of the previous body. Ensure it's set to # None. self._thread_local.pos = None r.register_hook('response', self.handle_401) r.register_hook('response', self.handle_redirect) self._thread_local.num_401_calls = 1 return r def __eq__(self, other): return all([ self.username == getattr(other, 'username', None), self.password == getattr(other, 'password', None) ]) def __ne__(self, other): return not self == other site-packages/pip/_vendor/requests/__pycache__/models.cpython-38.pyc000064400000056524151464452720021534 0ustar00U .e@sdZddlZddlZddlZddlmZddlmZddl m Z ddl m Z m Z mZmZddlmZdd lmZdd lmZdd lmZdd lmZmZmZdd lmZmZmZm Z m!Z!m"Z"m#Z#ddl$m%Z%m&Z&ddl'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1ddl2m3Z3m4Z4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;mZ>ddl2m?Z@ddlAmBZBeBjCeBjDeBjEeBjFeBjGfZHdZIdZJdZKGdddeLZMGdddeLZNGdddeNZOGdddeMeNZPGdddeLZQdS) z` requests.models ~~~~~~~~~~~~~~~ This module contains the primary objects that power Requests. N) RequestField)encode_multipart_formdata) parse_url) DecodeErrorReadTimeoutError ProtocolErrorLocationParseError)UnsupportedOperation) default_hooks)CaseInsensitiveDict) HTTPBasicAuth)cookiejar_from_dictget_cookie_header_copy_cookie_jar) HTTPError MissingSchema InvalidURLChunkedEncodingErrorContentDecodingErrorConnectionErrorStreamConsumedError)to_native_stringunicode_is_ascii) guess_filenameget_auth_from_url requote_uristream_decode_response_unicodeto_key_val_listparse_header_links iter_slicesguess_json_utf super_lencheck_header_validity) CallableMapping cookielib urlunparseurlsplit urlencodestrbytesis_py2chardet builtin_str basestring)json)codesi(ic@s0eZdZeddZeddZeddZdS)RequestEncodingMixincCsNg}t|j}|j}|sd}|||j}|rD|d||d|S)zBuild the path URL to use./?)r(urlpathappendqueryjoin)selfr7pr8r:r>?/usr/lib/python3.8/site-packages/pip/_vendor/requests/models.pypath_url=s    zRequestEncodingMixin.path_urlcCst|ttfr|St|dr |St|drg}t|D]j\}}t|tsRt|dsX|g}|D]B}|dk r\|t|tr|dn|t|tr|dn|fq\q6t|ddS|SdS)zEncode parameters in a piece of data. Will successfully encode parameters when passed as a dict or a list of 2-tuples. Order is retained if data is a list of 2-tuples but arbitrary if parameters are supplied as a dict. read__iter__Nutf-8T)Zdoseq) isinstancer*r+hasattrrr/r9encoder))dataresultkZvsvr>r>r?_encode_paramsRs$    z#RequestEncodingMixin._encode_paramscCs|stdnt|tr tdg}t|p,i}t|p8i}|D]|\}}t|ts\t|dsb|g}|D]T}|dk rft|tst|}|t|tr|dn|t|tr| dn|fqfq@|D]\}}d}d} t|t t fr t |dkr|\} } n&t |dkr|\} } }n |\} } }} nt |p,|} |} t| tttfrJ| } n(t| dr`| } n| dkrnqn| } t|| | | d } | j|d || qt|\}}||fS) aBuild the body for a multipart/form-data request. Will successfully encode files when passed as a dict or a list of tuples. Order is retained if data is a list of tuples but arbitrary if parameters are supplied as a dict. The tuples may be 2-tuples (filename, fileobj), 3-tuples (filename, fileobj, contentype) or 4-tuples (filename, fileobj, contentype, custom_headers). zFiles must be provided.zData must not be a string.rBNrCrA)namerGfilenameheaders) content_type) ValueErrorrDr/rrEr+r*r9decoderFtuplelistlenr bytearrayrArZmake_multipartr)filesrGZ new_fieldsZfieldsZfieldvalrJrIZftZfhfnfpZfdataZrfbodyrQr>r>r? _encode_filesmsT                 z"RequestEncodingMixin._encode_filesN)__name__ __module__ __qualname__propertyr@ staticmethodrKr]r>r>r>r?r3<s   r3c@seZdZddZddZdS)RequestHooksMixincCsZ||jkrtd|t|tr2|j||n$t|drV|j|dd|DdS)zProperly register a hook.z1Unsupported event specified, with event name "%s"rBcss|]}t|tr|VqdSN)rDr$).0hr>r>r? s z2RequestHooksMixin.register_hook..N)hooksrRrDr$r9rEextendr<eventhookr>r>r? register_hooks     zRequestHooksMixin.register_hookcCs2z|j||WdStk r,YdSXdS)ziDeregister a previously registered hook. Returns True if the hook existed, False if not. TFN)rhremoverRrjr>r>r?deregister_hooks z!RequestHooksMixin.deregister_hookN)r^r_r`rmror>r>r>r?rcs rcc @s*eZdZdZd ddZddZddZdS) Requesta{A user-created :class:`Request ` object. Used to prepare a :class:`PreparedRequest `, which is sent to the server. :param method: HTTP method to use. :param url: URL to send. :param headers: dictionary of headers to send. :param files: dictionary of {filename: fileobject} files to multipart upload. :param data: the body to attach to the request. If a dictionary or list of tuples ``[(key, value)]`` is provided, form-encoding will take place. :param json: json for the body to attach to the request (if files or data is not specified). :param params: URL parameters to append to the URL. If a dictionary or list of tuples ``[(key, value)]`` is provided, form-encoding will take place. :param auth: Auth handler or (user, pass) tuple. :param cookies: dictionary or CookieJar of cookies to attach to this request. :param hooks: dictionary of callback hooks, for internal usage. Usage:: >>> import requests >>> req = requests.Request('GET', 'https://httpbin.org/get') >>> req.prepare() Nc Cs|dkr gn|}|dkrgn|}|dkr,in|}|dkrr>r?__init__s"zRequest.__init__cCs d|jS)Nzrrr<r>r>r?__repr__szRequest.__repr__c Cs<t}|j|j|j|j|j|j|j|j|j |j |j d |S)zXConstructs a :class:`PreparedRequest ` for transmission and returns it.) rrr7rPrXrGr0rsrtrurh) PreparedRequestpreparerrr7rPrXrGr0rsrtrurhr<r=r>r>r?r{s zRequest.prepare) NNNNNNNNNN)r^r_r`__doc__rvryr{r>r>r>r?rps rpc @seZdZdZddZdddZddZd d Zd d Ze d dZ ddZ ddZ dddZ ddZd ddZddZddZdS)!rzaThe fully mutable :class:`PreparedRequest ` object, containing the exact bytes that will be sent to the server. Generated from either a :class:`Request ` object or manually. Usage:: >>> import requests >>> req = requests.Request('GET', 'https://httpbin.org/get') >>> r = req.prepare() >>> s = requests.Session() >>> s.send(r) cCs0d|_d|_d|_d|_d|_t|_d|_dSrd)rrr7rP_cookiesr\r rh_body_positionrxr>r>r?rv"szPreparedRequest.__init__Nc CsR||||||||||||| ||||| dS)z6Prepares the entire request with the given parameters.N)prepare_method prepare_urlprepare_headersprepare_cookies prepare_body prepare_auth prepare_hooks) r<rrr7rPrXrGrsrtrurhr0r>r>r?r{3s     zPreparedRequest.preparecCs d|jS)Nzrwrxr>r>r?ryEszPreparedRequest.__repr__cCsXt}|j|_|j|_|jdk r*|jnd|_t|j|_|j|_|j|_|j |_ |Srd) rzrrr7rPcopyrr~r\rhrr|r>r>r?rHs zPreparedRequest.copycCs$||_|jdk r t|j|_dS)zPrepares the given HTTP method.N)rrrupper)r<rrr>r>r?rSs zPreparedRequest.prepare_methodcCsDddlm}z|j|ddd}Wn|jk r>tYnX|S)Nr)idnaT)Zuts46rC)Z pip._vendorrrFrSZ IDNAError UnicodeError)hostrr>r>r?_get_idna_encoded_hostYs   z&PreparedRequest._get_idna_encoded_hostc Cs,t|tr|d}ntr"t|nt|}|}d|krR|dsR||_ dSzt |\}}}}}}} Wn,t k r} zt | j W5d} ~ XYnX|sd} | t|d} t| |st d|t|sz||}Wntk rt dYnXn|drt d|pd } | r.| d 7} | |7} |rL| dt|7} |sVd }trt|trr|d }t| tr| d } t|tr|d }t|tr|d }t| tr| d } t|ttfrt|}||} | r |rd || f}n| }tt|| |d|| g}||_ dS)zPrepares the given HTTP URL.utf8:ZhttpNzDInvalid URL {0!r}: No schema supplied. Perhaps you meant http://{0}?z Invalid URL %r: No host suppliedzURL has an invalid label.*r6@r4rCz%s&%s)rDr+rSr,Zunicoder*lstriplower startswithr7rrrargsformatrrrrrrFrKrr')r<r7rsZschemertrZportr8r:ZfragmenteerrorZnetlocZ enc_paramsr>r>r?rcsh                 zPreparedRequest.prepare_urlcCs<t|_|r8|D]"}t||\}}||jt|<qdS)z Prepares the given HTTP headers.N)r rPrqr#r)r<rPheaderrNvaluer>r>r?rs  zPreparedRequest.prepare_headersc Csrd}d}|s6|dk r6d}t|}t|ts6|d}tt|dt|ttt t f g}z t |}Wnt t tfk rd}YnX|r|}t|dddk rz||_Wn ttfk rt|_YnX|rtd|rt||jd<n d|jd <np|r|||\}}n2|rB||}t|ts8t|d r>d}nd }|||rhd |jkrh||jd <||_dS)z"Prepares the given HTTP body data.Nzapplication/jsonrCrBtellz1Streamed bodies and files are mutually exclusive.Content-LengthZchunkedzTransfer-EncodingrAz!application/x-www-form-urlencodedz content-typez Content-Type) complexjsondumpsrDr+rFallrEr/rUrTr%r" TypeErrorAttributeErrorr getattrrrIOErrorOSErrorobjectNotImplementedErrorr.rPr]rKprepare_content_lengthr\)r<rGrXr0r\rQZ is_streamlengthr>r>r?rsL          zPreparedRequest.prepare_bodycCsL|dk r$t|}|rHt||jd<n$|jdkrH|jddkrHd|jd<dS)z>Prepare Content-Length header based on request method and bodyNr)ZGETZHEAD0)r"r.rPrrget)r<r\rr>r>r?r s z&PreparedRequest.prepare_content_lengthr6cCsj|dkr"t|j}t|r|nd}|rft|trDt|dkrDt|}||}|j|j| |j dS)z"Prepares the given HTTP auth data.NrL) rr7anyrDrTrVr __dict__updaterr\)r<rtr7Zurl_authrr>r>r?rs zPreparedRequest.prepare_authcCs@t|tjr||_n t||_t|j|}|dk r<||jd<dS)aPrepares the given HTTP cookie data. This function eventually generates a ``Cookie`` header from the given cookies using cookielib. Due to cookielib's design, the header will not be regenerated if it already exists, meaning this function can only be called once for the life of the :class:`PreparedRequest ` object. Any subsequent calls to ``prepare_cookies`` will have no actual effect, unless the "Cookie" header is removed beforehand. NZCookie)rDr&Z CookieJarr~rrrP)r<ruZ cookie_headerr>r>r?r,s   zPreparedRequest.prepare_cookiescCs&|pg}|D]}||||q dS)zPrepares the given hooks.N)rm)r<rhrkr>r>r?r@szPreparedRequest.prepare_hooks) NNNNNNNNNN)N)r6)r^r_r`r}rvr{ryrrrbrrrrrrrrr>r>r>r?rzs2   V E rzc @seZdZdZdddddddd d d g Zd d ZddZddZddZddZ ddZ ddZ ddZ ddZ eddZed d!Zed"d#Zed$d%Zed&d'Zd;d*d+Zed)d,fd-d.Zed/d0Zed1d2Zd3d4Zed5d6Zd7d8Zd9d:Zd,S)<ResponsezhThe :class:`Response ` object, which contains a server's response to an HTTP request. _content status_coderPr7historyencodingreasonruelapsedrequestcCs^d|_d|_d|_d|_t|_d|_d|_d|_g|_ d|_ t i|_ t d|_d|_dS)NFr)r_content_consumed_nextrr rPrawr7rrrrrudatetimeZ timedeltarrrxr>r>r?rvTs  zResponse.__init__cCs|Srdr>rxr>r>r? __enter__szResponse.__enter__cGs |dSrd)close)r<rr>r>r?__exit__szResponse.__exit__cs js jfddjDS)Ncsi|]}|t|dqSrd)r)reattrrxr>r? sz)Response.__getstate__..)rcontent __attrs__rxr>rxr? __getstate__szResponse.__getstate__cCs:|D]\}}t|||qt|ddt|dddS)NrTr)rqsetattr)r<staterNrr>r>r? __setstate__s zResponse.__setstate__cCs d|jS)Nz)rrxr>r>r?ryszResponse.__repr__cCs|jSakReturns True if :attr:`status_code` is less than 400. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code, is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. okrxr>r>r?__bool__szResponse.__bool__cCs|jSrrrxr>r>r? __nonzero__szResponse.__nonzero__cCs |dS)z,Allows you to use a response as an iterator.) iter_contentrxr>r>r?rBszResponse.__iter__cCs(z |Wntk r"YdSXdS)axReturns True if :attr:`status_code` is less than 400, False if not. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. FT)raise_for_statusrrxr>r>r?rs  z Response.okcCsd|jko|jtkS)zTrue if this Response is a well-formed HTTP redirect that could have been processed automatically (by :meth:`Session.resolve_redirects`). location)rPrREDIRECT_STATIrxr>r>r? is_redirectszResponse.is_redirectcCsd|jko|jtjtjfkS)z@True if this Response one of the permanent versions of redirect.r)rPrr1Zmoved_permanentlypermanent_redirectrxr>r>r?is_permanent_redirectszResponse.is_permanent_redirectcCs|jS)zTReturns a PreparedRequest for the next request in a redirect chain, if there is one.)rrxr>r>r?nextsz Response.nextcCst|jdS)z7The apparent encoding, provided by the chardet library.r)r-Zdetectrrxr>r>r?apparent_encodingszResponse.apparent_encodingr Fcs|fdd}jr(tjtr(tn"dk rJttsJtdttj}|}jrf|n|}|rxt |}|S)aIterates over the response data. When stream=True is set on the request, this avoids reading the content at once into memory for large responses. The chunk size is the number of bytes it should read into memory. This is not necessarily the length of each item returned as decoding can take place. chunk_size must be of type int or None. A value of None will function differently depending on the value of `stream`. stream=True will read data as it arrives in whatever size the chunks are received. If stream=False, data is returned as a single chunk. If decode_unicode is True, content will be decoded using the best available encoding based on the response. c 3stjdrz jjddD] }|VqWqtk rV}z t|W5d}~XYqtk r~}z t|W5d}~XYqtk r}z t|W5d}~XYqXnj }|sq|Vqd_ dS)NstreamT)Zdecode_content) rErrrrrrrrrAr)chunkr chunk_sizer<r>r?generates   z'Response.iter_content..generateNz.chunk_size must be an int, it is instead a %s.) rrDrboolrintrtyper r)r<rdecode_unicoderZ reused_chunksZ stream_chunksZchunksr>rr?rs  zResponse.iter_contentNccsd}|j||dD]n}|dk r&||}|r6||}n|}|rl|drl|rl|dd|dkrl|}nd}|D] }|Vqtq|dk r|VdS)zIterates over the response data, one line at a time. When stream=True is set on the request, this avoids reading the content at once into memory for large responses. .. note:: This method is not reentrant safe. N)rr)rsplit splitlinespop)r<rrZ delimiterZpendingrlinesliner>r>r? iter_liness $  zResponse.iter_linescCsV|jdkrJ|jrtd|jdks,|jdkr4d|_nd|tpFd|_d|_|jS)z"Content of the response, in bytes.Fz2The content for this response was already consumedrNT)rr RuntimeErrorrrr;rCONTENT_CHUNK_SIZErxr>r>r?r/s zResponse.contentc Cshd}|j}|jstdS|jdkr(|j}zt|j|dd}Wn&ttfk rbt|jdd}YnX|S)aContent of the response, in unicode. If Response.encoding is None, encoding will be guessed using ``chardet``. The encoding of the response content is determined based solely on HTTP headers, following RFC 2616 to the letter. If you can take advantage of non-HTTP knowledge to make a better guess at the encoding, you should set ``r.encoding`` appropriately before accessing this property. Nr6replace)errors)rrr*r LookupErrorr)r<rrr>r>r?textCs z Response.textcKsj|jsZ|jrZt|jdkrZt|j}|dk rZztj|j|f|WStk rXYnXtj|jf|S)zReturns the json-encoded content of a response, if any. :param \*\*kwargs: Optional arguments that ``json.loads`` takes. :raises ValueError: If the response body does not contain valid json. rMN) rrrVr!rloadsrSUnicodeDecodeErrorr)r<kwargsrr>r>r?r0is  z Response.jsoncCsF|jd}i}|rBt|}|D] }|dp6|d}|||<q |S)z8Returns the parsed header links of the response, if any.linkZrelr7)rPrr)r<rllinksrkeyr>r>r?rs  zResponse.linkscCsd}t|jtrDz|jd}WqJtk r@|jd}YqJXn|j}d|jkr`dkrxnnd|j||jf}n,d|jkrdkrnnd|j||jf}|rt||d d S) z2Raises stored :class:`HTTPError`, if one occurred.r6rCz iso-8859-1iiz%s Client Error: %s for url: %siXz%s Server Error: %s for url: %s)ZresponseN)rDrr+rSrrr7r)r<Zhttp_error_msgrr>r>r?rs zResponse.raise_for_statuscCs0|js|jt|jdd}|dk r,|dS)zReleases the connection back to the pool. Once this method has been called the underlying ``raw`` object must not be accessed again. *Note: Should not normally need to be called explicitly.* release_connN)rrrr)r<rr>r>r?rs  zResponse.close)r F)r^r_r`r}rrvrrrrryrrrBrarrrrrrITER_CHUNK_SIZErrrr0rrrr>r>r>r?rJsT/        7  % r)Rr}rsysZencodings.idnaZ encodingsZpip._vendor.urllib3.fieldsrZpip._vendor.urllib3.filepostrZpip._vendor.urllib3.utilrZpip._vendor.urllib3.exceptionsrrrrior rhr Z structuresr rtr rurrr exceptionsrrrrrrrZ_internal_utilsrrZutilsrrrrrrr r!r"r#compatr$r%r&r'r(r)r*r+r,r-r.r/r0rZ status_codesr1ZmovedfoundotherZtemporary_redirectrrZDEFAULT_REDIRECT_LIMITrrrr3rcrprzrr>r>r>r?sD       $08  rJ<site-packages/pip/_vendor/requests/__pycache__/help.cpython-38.pyc000064400000005206151464452720021170 0ustar00U .e @sdZddlmZddlZddlZddlZddlZddlmZddlm Z ddlm Z ddl m Z zdd lmZWn ek rdZdZdZYnXddlZddlZd d Zd d ZddZedkredS)z'Module containing bug report helper(s).)print_functionN)idna)urllib3)chardet) __version__) pyopensslcCst}|dkrt}nj|dkr\dtjjtjjtjjf}tjjdkrd |tjjg}n(|dkrnt}n|dkrt}nd}||d S) aReturn a dict with the Python implementation and version. Provide both the name and the version of the Python implementation currently running. For example, on CPython 2.7.5 it will return {'name': 'CPython', 'version': '2.7.5'}. This function works best on CPython and PyPy: in particular, it probably doesn't work for Jython or IronPython. Future investigation should be done to work out the correct shape of the code for those platforms. ZCPythonZPyPyz%s.%s.%sfinalZJythonZ IronPythonUnknown)nameversion) platformZpython_implementationZpython_versionsysZpypy_version_infomajorminormicro releaseleveljoin)implementationZimplementation_versionr=/usr/lib/python3.8/site-packages/pip/_vendor/requests/help.py_implementations&     rc Cszttd}Wntk r4ddd}YnXt}dtji}dtji}ddd}trrtjdtj j d}dt t ddi}dt t ddi}tj }d|dk rd|ndi}|||tdk |||||dtid S) z&Generate information for a bug report.)systemreleaser r Nr )r Zopenssl_versionz%xr) rr system_sslZusing_pyopensslZ pyOpenSSLrr cryptographyrZrequests)rrrIOErrorrrrrOpenSSLZSSLZOPENSSL_VERSION_NUMBERgetattrrrsslrrequests_version) Z platform_infoZimplementation_infoZ urllib3_infoZ chardet_infoZpyopenssl_infoZcryptography_infoZ idna_inforZsystem_ssl_inforrrinfo;sR       r"cCsttjtddddS)z)Pretty-print the bug information as JSON.T)Z sort_keysindentN)printjsondumpsr"rrrrmainqsr(__main__)__doc__Z __future__rr&rrr Z pip._vendorrrrr rr!Zpip._vendor.urllib3.contribr ImportErrorrrrr"r(__name__rrrrs,      !6site-packages/pip/_vendor/requests/__pycache__/compat.cpython-38.pyc000064400000003050151464452720021516 0ustar00U .e@sdZddlmZddlZejZeddkZeddkZddlZerddl m Z m Z m Z m Z mZmZmZmZmZddlmZmZmZmZmZddlmZddlZdd lmZdd lmZdd lmZm Z m!Z!m"Z"e#Z$e#Z%e&Z#e'Z'e(e)e*fZ+e(e)fZ,nerdd l-mZmZmZmZmZm Z m Z m Z m Z mZdd l.mZmZmZmZmZddl/m0Zdd l1mZdd l2mZddlm"Z"ddl3mZm Z m!Z!e#Z$e#Z#e%Z%e#e%fZ'e(e*fZ+e(fZ,dS)zq requests.compat ~~~~~~~~~~~~~~~ This module handles import compatibility issues between Python 2 and Python 3. )chardetN) quoteunquote quote_plus unquote_plus urlencode getproxies proxy_bypassproxy_bypass_environmentgetproxies_environment)urlparse urlunparseurljoinurlsplit urldefrag)parse_http_list)Morsel)StringIO)CallableMappingMutableMapping OrderedDict) rrrrr rrrrr)rr r r r ) cookiejar)r)rrr)4__doc__Z pip._vendorrsys version_infoZ_verZis_py2Zis_py3ZjsonZurllibrrrrr r r r r rrrrrZurllib2rZ cookielibZCookierr collectionsrrrrstrZ builtin_strbytesZunicodeZ basestringintZlongfloatZ numeric_typesZ integer_typesZ urllib.parseZurllib.requestZhttprZ http.cookiesioZcollections.abcr$r$?/usr/lib/python3.8/site-packages/pip/_vendor/requests/compat.pysD   ,     0    site-packages/pip/_vendor/requests/__pycache__/certs.cpython-38.pyc000064400000001125151464452720021354 0ustar00U .e@s&dZddlmZedkr"eedS)uF requests.certs ~~~~~~~~~~~~~~ This module returns the preferred default CA certificate bundle. There is only one — the one from the certifi package. If you are packaging Requests, e.g., for a Linux distribution or a managed environment, you can change the definition of where() to return a separately packaged CA bundle. )where__main__N)__doc__Zpip._vendor.certifir__name__printrr>/usr/lib/python3.8/site-packages/pip/_vendor/requests/certs.pys site-packages/pip/_vendor/requests/__pycache__/certs.cpython-38.opt-1.pyc000064400000001125151464452720022313 0ustar00U .e@s&dZddlmZedkr"eedS)uF requests.certs ~~~~~~~~~~~~~~ This module returns the preferred default CA certificate bundle. There is only one — the one from the certifi package. If you are packaging Requests, e.g., for a Linux distribution or a managed environment, you can change the definition of where() to return a separately packaged CA bundle. )where__main__N)__doc__Zpip._vendor.certifir__name__printrr>/usr/lib/python3.8/site-packages/pip/_vendor/requests/certs.pys site-packages/pip/_vendor/requests/__pycache__/cookies.cpython-38.pyc000064400000044556151464452720021707 0ustar00U .eG@sdZddlZddlZddlZddlmZddlmZmZm Z m Z m Z z ddl Z Wne k rlddlZ YnXGdddeZGdd d eZd d Zd d ZdddZGdddeZGdddeje ZddZddZddZd ddZddZdS)!z requests.cookies ~~~~~~~~~~~~~~~~ Compatibility code to be able to use `cookielib.CookieJar` with requests. requests.utils imports from here, so be careful with imports. N)to_native_string) cookieliburlparse urlunparseMorselMutableMappingc@seZdZdZddZddZddZdd Zd d Zd d Z ddZ dddZ ddZ ddZ ddZeddZeddZeddZdS) MockRequestaWraps a `requests.Request` to mimic a `urllib2.Request`. The code in `cookielib.CookieJar` expects this interface in order to correctly manage cookie policies, i.e., determine whether a cookie can be set, given the domains of the request and the cookie. The original request object is read-only. The client is responsible for collecting the new headers via `get_new_headers()` and interpreting them appropriately. You probably want `get_cookie_header`, defined below. cCs ||_i|_t|jjj|_dSN)_r _new_headersrurlschemetype)selfrequestr@/usr/lib/python3.8/site-packages/pip/_vendor/requests/cookies.py__init__%szMockRequest.__init__cCs|jSr )rrrrrget_type*szMockRequest.get_typecCst|jjjSr )rr r Znetlocrrrrget_host-szMockRequest.get_hostcCs|Sr rrrrrget_origin_req_host0szMockRequest.get_origin_req_hostcCsT|jjds|jjSt|jjddd}t|jj}t|j||j|j |j |j gS)NZHostzutf-8)encoding) r headersgetr rrrrpathZparamsZqueryZfragment)rhostZparsedrrr get_full_url3s zMockRequest.get_full_urlcCsdSNTrrrrris_unverifiableAszMockRequest.is_unverifiablecCs||jjkp||jkSr )r rr rnamerrr has_headerDszMockRequest.has_headerNcCs|jj||j||Sr )r rrr )rr#defaultrrr get_headerGszMockRequest.get_headercCs tddS)zMcookielib has no legitimate use for this method; add it back if you find one.z=Cookie headers should be added with add_unredirected_header()N)NotImplementedError)rkeyvalrrr add_headerJszMockRequest.add_headercCs||j|<dSr r rr#valuerrradd_unredirected_headerNsz#MockRequest.add_unredirected_headercCs|jSr r+rrrrget_new_headersQszMockRequest.get_new_headerscCs|Sr )r!rrrr unverifiableTszMockRequest.unverifiablecCs|Sr )rrrrrorigin_req_hostXszMockRequest.origin_req_hostcCs|Sr rrrrrr\szMockRequest.host)N)__name__ __module__ __qualname____doc__rrrrrr!r$r&r*r.r/propertyr0r1rrrrrr s$    r c@s(eZdZdZddZddZddZdS) MockResponsezWraps a `httplib.HTTPMessage` to mimic a `urllib.addinfourl`. ...what? Basically, expose the parsed HTTP headers from the server response the way `cookielib` expects to see them. cCs ||_dS)zMake a MockResponse for `cookielib` to read. :param headers: a httplib.HTTPMessage or analogous carrying the headers N_headers)rrrrrrhszMockResponse.__init__cCs|jSr r8rrrrinfooszMockResponse.infocCs|j|dSr )r9 getheadersr"rrrr;rszMockResponse.getheadersN)r2r3r4r5rr:r;rrrrr7asr7cCs8t|dr|jsdSt|}t|jj}|||dS)zExtract the cookies from the response into a CookieJar. :param jar: cookielib.CookieJar (not necessarily a RequestsCookieJar) :param request: our own requests.Request object :param response: urllib3.HTTPResponse object _original_responseN)hasattrr<r r7msgZextract_cookies)jarrZresponseZreqresrrrextract_cookies_to_jarvs  rAcCs t|}|||dS)zj Produce an appropriate Cookie header string to be sent with `request`, or None. :rtype: str Cookie)r Zadd_cookie_headerr/r)r?rrrrrget_cookie_headers rDcCszg}|D]N}|j|krq|dk r,||jkr,q|dk r@||jkr@q||j|j|jfq|D]\}}}||||q\dS)zkUnsets a cookie by name, by default over all domains and paths. Wraps CookieJar.clear(), is O(n). N)r#domainrappendclear) cookiejarr#rErZ clearablescookierrrremove_cookie_by_names rJc@seZdZdZdS)CookieConflictErrorzThere are two cookies that meet the criteria specified in the cookie jar. Use .get and .set and include domain and path args in order to be more specific. N)r2r3r4r5rrrrrKsrKcseZdZdZd3ddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ ddZd4ddZfddZddZdd Zd!d"Zfd#d$Zfd%d&Zd5d'd(Zd6d)d*Zd+d,Zd-d.Zd/d0Zd1d2ZZS)7RequestsCookieJaraCompatibility class; is a cookielib.CookieJar, but exposes a dict interface. This is the CookieJar we create by default for requests and sessions that don't specify one, since some clients may expect response.cookies and session.cookies to support dict operations. Requests does not use the dict interface internally; it's just for compatibility with external client code. All requests code should work out of the box with externally provided instances of ``CookieJar``, e.g. ``LWPCookieJar`` and ``FileCookieJar``. Unlike a regular CookieJar, this class is pickleable. .. warning:: dictionary operations that are normally O(1) may be O(n). NcCs.z||||WStk r(|YSXdS)zDict-like get() that also supports optional domain and path args in order to resolve naming collisions from using one cookie jar over multiple domains. .. warning:: operation is O(n), not O(1). N)_find_no_duplicatesKeyError)rr#r%rErrrrrszRequestsCookieJar.getcKsX|dkr(t|||d|dddSt|tr|j|kr |dk rRtd||j}q |rb|Std|||fdS)aBoth ``__get_item__`` and ``get`` call this function: it's never used elsewhere in Requests. :param name: a string containing name of cookie :param domain: (optional) string containing domain of cookie :param path: (optional) string containing path of cookie :raises KeyError: if cookie is not found :raises CookieConflictError: if there are multiple cookies that match name and optionally domain and path :return: cookie.value Nz(There are multiple cookies with name, %rrw)rVr#rErrKr-rN)rr#rErZtoReturnrIrrrrMxs    z%RequestsCookieJar._find_no_duplicatescCs|j}|d|S)4Unlike a normal CookieJar, this class is pickleable. _cookies_lock)__dict__rtpoprstaterrr __getstate__s  zRequestsCookieJar.__getstate__cCs$|j|d|jkr t|_dS)ryrzN)r{ru threadingRLockrzr}rrr __setstate__s  zRequestsCookieJar.__setstate__cCs"t}|||||S)z(Return a copy of this RequestsCookieJar.)rLZ set_policy get_policyru)rZnew_cjrrrrts zRequestsCookieJar.copycCs|jS)z&Return the CookiePolicy instance used.)Z_policyrrrrrszRequestsCookieJar.get_policy)NNN)NN)NN)NN)r2r3r4r5rrUrXrZr[r\r]r^rarcrdrergrjrkrlrRrurxrMrrrtr __classcell__rrrhrrLs2            rLcCsN|dkr dSt|dr|St|}||D]}|t|q4|S)Nrt)r=rtrGrR)r?Znew_jarrIrrr_copy_cookie_jars  rcKsd||ddddddddddidd }t|t|}|rLd }t|t|||t|d |d <t|d |d <|d d|d<t|d|d<tjf|S)zMake a cookie from underspecified parameters. By default, the pair of `name` and `value` will be set for the domain '' and sent on every request (this is sometimes called a "supercookie"). rNro/FTHttpOnly) versionr#r-portrErsecureexpiresdiscardcomment comment_urlrestrfc2109z4create_cookie() got unexpected keyword arguments: %srZport_specifiedrEZdomain_specified.Zdomain_initial_dotrZpath_specified)rU TypeErrorrYruboolrmrrB)r#r-rSresultZbadargserrrrrrQs0 rQcCsd}|drPzttt|d}Wqrtk rLtd|dYqrXn"|drrd}tt|d|}t|dt|dd|d||j |d dd |d idt|d |j |d pdd S)zBConvert a Morsel object into a Cookie containing the one k/v pair.Nzmax-agezmax-age: %s must be integerrz%a, %d-%b-%Y %H:%M:%S GMTrFrErrZhttponlyrrr) rrrrErr#rrrrrr-r) inttime ValueErrorrcalendarZtimegmstrptimerQrr(r-)ZmorselrZ time_templaterrrrPs4    rPTcCsR|dkrt}|dk rNdd|D}|D]$}|s8||kr(|t|||q(|S)aCReturns a CookieJar from a key/value dictionary. :param cookie_dict: Dict of key/values to insert into CookieJar. :param cookiejar: (optional) A cookiejar to add the cookies to. :param overwrite: (optional) If False, will not replace cookies already in the jar with new ones. :rtype: CookieJar NcSsg|] }|jqSr)r#).0rIrrr sz'cookiejar_from_dict..)rLrRrQ)Z cookie_dictrH overwriteZnames_from_jarr#rrrcookiejar_from_dicts  rcCsvt|tjstdt|tr.t||dd}nDt|tjrrz||Wn(tk rp|D]}||q\YnX|S)zAdd cookies to cookiejar and returns a merged CookieJar. :param cookiejar: CookieJar object to add the cookies to. :param cookies: Dictionary or CookieJar object to be added. :rtype: CookieJar z!You can only merge into CookieJarF)rHr) rOrrsrdictrruAttributeErrorrR)rHZcookiesZ cookie_in_jarrrr merge_cookiess   r)NN)NT)r5rtrrZ_internal_utilsrcompatrrrrrr ImportErrorZdummy_threadingobjectr r7rArDrJ RuntimeErrorrKrsrLrrQrPrrrrrrs.   H $ site-packages/pip/_vendor/requests/__pycache__/hooks.cpython-38.opt-1.pyc000064400000001674151464452720022327 0ustar00U .e@sdZdgZddZddZdS)z requests.hooks ~~~~~~~~~~~~~~ This module provides the capabilities for the Requests hooks system. Available hooks: ``response``: The response generated from a Request. ZresponsecCsddtDS)NcSsi|] }|gqSr).0Zeventrr>/usr/lib/python3.8/site-packages/pip/_vendor/requests/hooks.py sz!default_hooks..)HOOKSrrrr default_hookssrcKsL|pi}||}|rHt|dr&|g}|D]}||f|}|dk r*|}q*|S)z6Dispatches a hook dictionary on a given piece of data.__call__N)gethasattr)keyZhooksZ hook_datakwargshookZ _hook_datarrr dispatch_hooks   r N)__doc__rrr rrrrs site-packages/pip/_vendor/requests/__pycache__/__init__.cpython-38.pyc000064400000006613151464452720022002 0ustar00U .e @sdZddlmZddlmZddlZddlmZddZd d Zzeej ej Wn0e e fk re d ej ej eYnXdd lmZesz,dd lmZeddlm ZeeWnek rYnXddlmZededdl mZmZmZm Z ddl mZmZmZmZddl m Z m!Z!ddl"m#Z#ddl"m$Z$ddl%m&Z&m'Z'm(Z(ddl)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1ddl2m3Z3m4Z4ddl5m6Z6ddlm7Z7m8Z8m9Z9m:Z:m;Z;mZ>m?Z?ddl@Z@ddl@mAZAe@BeCDeAejde=dddS)a Requests HTTP Library ~~~~~~~~~~~~~~~~~~~~~ Requests is an HTTP library, written in Python, for human beings. Basic GET usage: >>> import requests >>> r = requests.get('https://www.python.org') >>> r.status_code 200 >>> 'Python is a programming language' in r.content True ... or POST: >>> payload = dict(key1='value1', key2='value2') >>> r = requests.post('https://httpbin.org/post', data=payload) >>> print(r.text) { ... "form": { "key2": "value2", "key1": "value1" }, ... } The other HTTP methods are supported - see `requests.api`. Full documentation is at . :copyright: (c) 2017 by Kenneth Reitz. :license: Apache 2.0, see LICENSE for more details. )urllib3)chardetN)RequestsDependencyWarningcCs|d}|dgkstt|dkr.|d|\}}}t|t|t|}}}|dks`t|dkslt|dksxt|ddd\}}}t|t|t|}}}|dkst|dkst|dkstdS) N.Zdev0r)splitAssertionErrorlenappendint)Zurllib3_versionZchardet_versionmajorminorpatchrA/usr/lib/python3.8/site-packages/pip/_vendor/requests/__init__.pycheck_compatibility1s         rcCsXzttt|d}Wntk r.YdSX|dddgkrTd|}t|tdS)Nrrr z4Old version of cryptography ({}) may cause slowdown.) listmaprr ValueErrorformatwarningswarnr)cryptography_versionZwarningrrr_check_cryptographyJs rz?urllib3 ({}) or chardet ({}) doesn't match a supported version!)WINDOWS) pyopenssl) __version__)DependencyWarningignore) __title____description____url__r") __build__ __author____author_email__ __license__) __copyright____cake__)utils)packages)RequestResponsePreparedRequest)requestgetheadpostrputdeleteoptions)sessionSession)codes) RequestExceptionTimeout URLRequiredTooManyRedirects HTTPErrorConnectionErrorFileModeWarningConnectTimeout ReadTimeout) NullHandlerdefaultT)r)E__doc__Z pip._vendorrrr exceptionsrrrr"r rrrZpip._internal.utils.compatr Zpip._vendor.urllib3.contribr!Zinject_into_urllib3Z cryptographyr ImportErrorZpip._vendor.urllib3.exceptionsr# simplefilterr%r&r'r(r)r*r+r,r-r.r/Zmodelsr0r1r2Zapir3r4r5r6rr7r8r9Zsessionsr:r;Z status_codesr<r=r>r?r@rArBrCrDrEZloggingrFZ getLogger__name__Z addHandlerrrrrsP#             ( , site-packages/pip/_vendor/requests/__pycache__/packages.cpython-38.pyc000064400000000722151464452720022014 0ustar00U .e@svddlZdD]dZdeZeeee<eejD]>ZeeksJeedr0ee ddZ ejeejde <q0q dS)N)Zurllib3ZidnaZchardetz pip._vendor..zpip._vendor.requests.packages.) syspackageZvendored_package __import__localslistmodulesmod startswithlenZunprefixed_modr r A/usr/lib/python3.8/site-packages/pip/_vendor/requests/packages.pyssite-packages/pip/_vendor/requests/__pycache__/status_codes.cpython-38.pyc000064400000010070151464452720022733 0ustar00U .e!E@sdaddlmZddddddd d d d d ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGDZedHdIZdJdKZedLS)Ma The ``codes`` object defines a mapping from common names for HTTP statuses to their numerical codes, accessible either as attributes or as dictionary items. >>> requests.codes['temporary_redirect'] 307 >>> requests.codes.teapot 418 >>> requests.codes['\o/'] 200 Some codes have multiple names, and both upper- and lower-case versions of the names are allowed. For example, ``codes.ok``, ``codes.OK``, and ``codes.okay`` all correspond to the HTTP status code 200. ) LookupDict)continue)Zswitching_protocols)Z processing)Z checkpoint)Z uri_too_longZrequest_uri_too_long)okZokayZall_okZall_okayZall_goodz\o/u✓)Zcreated)Zaccepted)Znon_authoritative_infoZnon_authoritative_information)Z no_content)Z reset_contentreset)Zpartial_contentpartial)Z multi_statusZmultiple_statusZ multi_statiZmultiple_stati)Zalready_reported)Zim_used)Zmultiple_choices)Zmoved_permanentlyZmovedz\o-)found)Z see_otherother)Z not_modified)Z use_proxy)Z switch_proxy)Ztemporary_redirectZtemporary_movedZ temporary)Zpermanent_redirectZresume_incompleteZresume)Z bad_requestZbad)Z unauthorized)Zpayment_requiredZpayment)Z forbidden)Z not_foundz-o-)Zmethod_not_allowedZ not_allowed)Znot_acceptable)Zproxy_authentication_requiredZ proxy_authZproxy_authentication)Zrequest_timeoutZtimeout)Zconflict)Zgone)Zlength_required)Zprecondition_failed precondition)Zrequest_entity_too_large)Zrequest_uri_too_large)Zunsupported_media_typeZunsupported_mediaZ media_type)Zrequested_range_not_satisfiableZrequested_rangeZrange_not_satisfiable)Zexpectation_failed)Z im_a_teapotZteapotZ i_am_a_teapot)Zmisdirected_request)Zunprocessable_entityZ unprocessable)locked)Zfailed_dependencyZ dependency)Zunordered_collectionZ unordered)Zupgrade_requiredZupgrade)Zprecondition_requiredr )Ztoo_many_requestsZtoo_many)Zheader_fields_too_largeZfields_too_large)Z no_responseZnone)Z retry_withZretry)Z$blocked_by_windows_parental_controlsZparental_controls)Zunavailable_for_legal_reasonsZ legal_reasons)Zclient_closed_request)Zinternal_server_errorZ server_errorz/o\u✗)Znot_implemented)Z bad_gateway)Zservice_unavailableZ unavailable)Zgateway_timeout)Zhttp_version_not_supportedZ http_version)Zvariant_also_negotiates)Zinsufficient_storage)Zbandwidth_limit_exceededZ bandwidth)Z not_extended)Znetwork_authentication_requiredZ network_authZnetwork_authentication)Ddefgzi,i-i.i/i0i1i2i3i4iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiZ status_codes)namecs~tD]8\}}|D]*}tt|||dstt||qqddtdk rvtddfddttDndadS)N)\/cSs$dddt|D}d||fS)Nz, css|]}d|VqdS)z``%s``N).0nrrE/usr/lib/python3.8/site-packages/pip/_vendor/requests/status_codes.py psz%_init..doc..z* %d: %s)join_codes)codenamesrrr docosz_init..doc c3s|]}|VqdS)Nr)rr$r&rr r!usz_init..) r#itemssetattrcodes startswithupper__doc__r"sorted)r$Ztitlestitlerr(r _iniths  r1N)r.Z structuresrr#r+r1rrrr s P site-packages/pip/_vendor/requests/__pycache__/adapters.cpython-38.opt-1.pyc000064400000041063151464452720023003 0ustar00U .e,T@sdZddlZddlZddlmZmZddlmZddl m Z ddl m Z ddl mZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZddlmZddlmZddlmZddlmZddlmZddlmZddl m!Z!m"Z"ddl#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0m1Z1m2Z2mZmZm3Z3m4Z4m5Z5m6Z6ddl7m8Z8zddl9m:Z:Wne;k rddZ:YnXdZdZ?Gd d!d!e@ZAGd"d#d#eAZBdS)$z requests.adapters ~~~~~~~~~~~~~~~~~ This module contains the transport adapters that Requests uses to define and maintain connections. N) PoolManagerproxy_from_url) HTTPResponse) parse_url)Timeout)Retry)ClosedPoolError)ConnectTimeoutError) HTTPError) MaxRetryError)NewConnectionError) ProxyError) ProtocolError)ReadTimeoutError)SSLError) ResponseError)LocationValueError)Response)urlparse basestring)DEFAULT_CA_BUNDLE_PATHextract_zipped_pathsget_encoding_from_headersprepend_scheme_if_neededget_auth_from_url urldefragauth select_proxy)CaseInsensitiveDict)extract_cookies_to_jar) ConnectionErrorConnectTimeout ReadTimeoutrr RetryError InvalidSchemaInvalidProxyURL InvalidURL)_basic_auth_str)SOCKSProxyManagercOs tddS)Nz'Missing dependencies for SOCKS support.)r$)argskwargsr+A/usr/lib/python3.8/site-packages/pip/_vendor/requests/adapters.pyr(.sr(F cs2eZdZdZfddZd ddZd d ZZS) BaseAdapterzThe Base Transport Adaptercstt|dSN)superr.__init__self __class__r+r,r1:szBaseAdapter.__init__FNTcCstdS)aCSends PreparedRequest object. Returns Response object. :param request: The :class:`PreparedRequest ` being sent. :param stream: (optional) Whether to stream the request content. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: (optional) Any user-provided SSL certificate to be trusted. :param proxies: (optional) The proxies dictionary to apply to the request. NNotImplementedError)r3requeststreamtimeoutverifycertproxiesr+r+r,send=szBaseAdapter.sendcCstdS)z!Cleans up adapter specific items.Nr6r2r+r+r,closeOszBaseAdapter.close)FNTNN)__name__ __module__ __qualname____doc__r1r>r? __classcell__r+r+r4r,r.7s  r.cseZdZdZdddddgZeeeeffdd Zd d Z d d Z efd dZ ddZ ddZ ddZd$ddZddZddZddZddZd%d"d#ZZS)& HTTPAdapteraThe built-in HTTP Adapter for urllib3. Provides a general-case interface for Requests sessions to contact HTTP and HTTPS urls by implementing the Transport Adapter interface. This class will usually be created by the :class:`Session ` class under the covers. :param pool_connections: The number of urllib3 connection pools to cache. :param pool_maxsize: The maximum number of connections to save in the pool. :param max_retries: The maximum number of retries each connection should attempt. Note, this applies only to failed DNS lookups, socket connections and connection timeouts, never to requests where data has made it to the server. By default, Requests does not retry failed connections. If you need granular control over the conditions under which we retry a request, import urllib3's ``Retry`` class and pass that instead. :param pool_block: Whether the connection pool should block for connections. Usage:: >>> import requests >>> s = requests.Session() >>> a = requests.adapters.HTTPAdapter(max_retries=3) >>> s.mount('http://', a) max_retriesconfig_pool_connections _pool_maxsize _pool_blockcsd|tkrtddd|_n t||_i|_i|_tt|||_ ||_ ||_ |j |||ddS)NrF)readblock) DEFAULT_RETRIESrrFZfrom_intrG proxy_managerr0rEr1rHrIrJinit_poolmanager)r3Zpool_connectionsZ pool_maxsizerFZ pool_blockr4r+r,r1qs zHTTPAdapter.__init__csfddjDS)Ncsi|]}|t|dqSr/)getattr).0attrr2r+r, sz,HTTPAdapter.__getstate__..) __attrs__r2r+r2r, __getstate__szHTTPAdapter.__getstate__cCsDi|_i|_|D]\}}t|||q|j|j|j|jddS)NrL)rOrGitemssetattrrPrHrIrJ)r3staterSvaluer+r+r, __setstate__s zHTTPAdapter.__setstate__cKs0||_||_||_tf|||dd||_dS)aInitializes a urllib3 PoolManager. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param connections: The number of urllib3 connection pools to cache. :param maxsize: The maximum number of connections to save in the pool. :param block: Block when no free connections are available. :param pool_kwargs: Extra keyword arguments used to initialize the Pool Manager. T) num_poolsmaxsizerMstrictN)rHrIrJr poolmanager)r3Z connectionsr]rMZ pool_kwargsr+r+r,rPs zHTTPAdapter.init_poolmanagercKs||jkr|j|}n||dr^t|\}}t|f|||j|j|jd|}|j|<n4||}t |f||j|j|jd|}|j|<|S)aReturn urllib3 ProxyManager for the given proxy. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param proxy: The proxy to return a urllib3 ProxyManager for. :param proxy_kwargs: Extra keyword arguments used to configure the Proxy Manager. :returns: ProxyManager :rtype: urllib3.ProxyManager socks)usernamepasswordr\r]rM) proxy_headersr\r]rM) rOlower startswithrr(rHrIrJrcr)r3proxyZ proxy_kwargsZmanagerrarbrcr+r+r,proxy_manager_fors6    zHTTPAdapter.proxy_manager_forcCs|drn|rnd}|dk r"|}|s.tt}|r>tj|sLtd|d|_ tj |sf||_ q||_ nd|_ d|_ d|_ |rt |ts|d|_|d|_n ||_d|_|jrtj|jstd |j|jrtj|jstd |jdS) aAVerify a SSL certificate. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param conn: The urllib3 connection object associated with the cert. :param url: The requested URL. :param verify: Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: The SSL certificate to verify. httpsNTzECould not find a suitable TLS CA certificate bundle, invalid path: {}Z CERT_REQUIREDZ CERT_NONErrz9Could not find the TLS certificate file, invalid path: {}z1Could not find the TLS key file, invalid path: {})rdrerrospathexistsIOErrorformatZ cert_reqsisdirZca_certsZ ca_cert_dir isinstancerZ cert_fileZkey_file)r3connurlr;r<Zcert_locr+r+r, cert_verifys>     zHTTPAdapter.cert_verifycCst}t|dd|_tt|di|_t|j|_||_|jj|_t |j t r^|j d|_ n|j |_ t |j||||_||_|S)aBuilds a :class:`Response ` object from a urllib3 response. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter ` :param req: The :class:`PreparedRequest ` used to generate the response. :param resp: The urllib3 response object. :rtype: requests.Response ZstatusNheadersutf-8)rrQZ status_coderrsrencodingrawreasonrorqbytesdecodercookiesr8 connection)r3ZreqrespZresponser+r+r,build_responses    zHTTPAdapter.build_responseNcCsdt||}|rDt|d}t|}|js.td||}||}nt|}|}|j |}|S)aReturns a urllib3 connection for the given URL. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param url: The URL to connect to. :param proxies: (optional) A Requests-style dictionary of proxies used on this request. :rtype: urllib3.ConnectionPool ZhttpzFPlease check proxy URL. It is malformed and could be missing the host.) rrrZhostr%rgZconnection_from_urlrZgeturlr_)r3rqr=rfZ proxy_urlrOrpZparsedr+r+r,get_connection$s     zHTTPAdapter.get_connectioncCs&|j|jD] }|qdS)zDisposes of any internal state. Currently, this closes the PoolManager and any active ProxyManager, which closes any pooled connections. N)r_clearrOvalues)r3rfr+r+r,r??s zHTTPAdapter.closec Cs`t|j|}t|jj}|o"|dk}d}|rDt|j}|d}|j}|r\|s\t|j}|S)a?Obtain the url to use when making the final request. If the message is being sent through a HTTP proxy, the full URL has to be used. Otherwise, we should only use the path portion of the URL. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param request: The :class:`PreparedRequest ` being sent. :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs. :rtype: str rhFr`)rrqrschemerdreZpath_urlr) r3r8r=rfrZis_proxied_http_requestZusing_socks_proxyZ proxy_schemerqr+r+r, request_urlIs     zHTTPAdapter.request_urlcKsdS)a"Add any headers needed by the connection. As of v2.0 this does nothing by default, but is left for overriding by users that subclass the :class:`HTTPAdapter `. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param request: The :class:`PreparedRequest ` to add headers to. :param kwargs: The keyword arguments from the call to send(). Nr+)r3r8r*r+r+r, add_headersfs zHTTPAdapter.add_headerscCs&i}t|\}}|r"t|||d<|S)aReturns a dictionary of the headers to add to any request sent through a proxy. This works with urllib3 magic to ensure that they are correctly sent to the proxy, rather than in a tunnelled request if CONNECT is being used. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param proxy: The url of the proxy being used for this request. :rtype: dict zProxy-Authorization)rr')r3rfrsrarbr+r+r,rcts  zHTTPAdapter.proxy_headersFTc Csz||j|}Wn.tk r@}zt||dW5d}~XYnX|||j|||||} |j||||||d|jdkpd|jk } t |t rz|\} } t | | d}Wqt k r}zd |} t | W5d}~XYqXnt |t rn t ||d}zF| s2|j|j| |j|jdddd|j|d }nt|d rD|j}|jtd }z|j|j| d d |jD]\}}|||qn||jD]D}|tt|d dd|d|||dq|dz|jd d}Wntk r|}YnXtj |||ddd}Wn|!YnXWnt"t#j$fk r~} zt%| |dW5d} ~ XYnZt&k r"}zt |j't(rt |j't)st*||dt |j't+rt,||dt |j't-rt.||dt |j't/rt0||dt%||dW5d}~XYnt1k rP}zt%||dW5d}~XYnt-k rz}z t.|W5d}~XYn^t/t2fk r}z:t |t/rt0||dnt |t3rt4||dnW5d}~XYnX|5||S)aSends PreparedRequest object. Returns Response object. :param request: The :class:`PreparedRequest ` being sent. :param stream: (optional) Whether to stream the request content. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple or urllib3 Timeout object :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: (optional) Any user-provided SSL certificate to be trusted. :param proxies: (optional) The proxies dictionary to apply to the request. :rtype: requests.Response )r8N)r9r:r;r<r=zContent-Length)connectrKzrInvalid timeout {}. Pass a (connect, read) timeout tuple, or a single float to set both timeouts to the same valueF) methodrqbodyrsZredirectZassert_same_hostpreload_contentdecode_contentZretriesr: proxy_pool)r:T)Zskip_accept_encodingrts s0 ) buffering)Zpoolr{rr)6r~rqrr&rrrrrrsrotuple TimeoutSauce ValueErrorrmZurlopenrrFhasattrrZ _get_connDEFAULT_POOL_TIMEOUTZ putrequestrWZ putheaderZ endheadersr>hexlenencodeZ getresponse TypeErrorrZ from_httplibr?rsocketerrorr r rwr r r!rr# _ProxyErrorr _SSLErrorrr _HTTPErrorrr"r})r3r8r9r:r;r<r=rperqZchunkedrrKerrr|Zlow_connheaderrZirr+r+r,r>s                   zHTTPAdapter.send)N)FNTNN)r@rArBrCrUDEFAULT_POOLSIZErNDEFAULT_POOLBLOCKr1rVr[rPrgrrr}r~r?rrrcr>rDr+r+r4r,rETs* %4%  rE)CrCZos.pathrirZpip._vendor.urllib3.poolmanagerrrZpip._vendor.urllib3.responserZpip._vendor.urllib3.utilrrrZpip._vendor.urllib3.util.retryrZpip._vendor.urllib3.exceptionsrr r rr r r rrrrrrrZmodelsrcompatrrZutilsrrrrrrrZ structuresrrzr exceptionsr r!r"r#r$r%r&Zauthr'Z!pip._vendor.urllib3.contrib.socksr( ImportErrorrrrNrobjectr.rEr+r+r+r,sF                $  , site-packages/pip/_vendor/requests/__pycache__/auth.cpython-38.opt-1.pyc000064400000020145151464452720022137 0ustar00U .e'@sdZddlZddlZddlZddlZddlZddlZddlmZddl m Z m Z m Z ddl mZddlmZddlmZd Zd Zd d ZGd ddeZGdddeZGdddeZGdddeZdS)z] requests.auth ~~~~~~~~~~~~~ This module contains the authentication handlers for Requests. N) b64encode)urlparsestr basestring)extract_cookies_to_jar)to_native_string)parse_dict_headerz!application/x-www-form-urlencodedzmultipart/form-datacCst|ts&tjd|tdt|}t|tsLtjd|tdt|}t|tr`|d}t|trt|d}dtt d ||f }|S)zReturns a Basic Auth string.zNon-string usernames will no longer be supported in Requests 3.0.0. Please convert the object you've passed in ({!r}) to a string or bytes object in the near future to avoid problems.)categoryzNon-string passwords will no longer be supported in Requests 3.0.0. Please convert the object you've passed in ({!r}) to a string or bytes object in the near future to avoid problems.latin1zBasic :) isinstancerwarningswarnformatDeprecationWarningrencoderrjoinstrip)usernamepasswordZauthstrr=/usr/lib/python3.8/site-packages/pip/_vendor/requests/auth.py_basic_auth_strs0      rc@seZdZdZddZdS)AuthBasez4Base class that all auth implementations derive fromcCs tddS)NzAuth hooks must be callable.)NotImplementedErrorselfrrrr__call__KszAuthBase.__call__N__name__ __module__ __qualname____doc__rrrrrrHsrc@s0eZdZdZddZddZddZdd Zd S) HTTPBasicAuthz?Attaches HTTP Basic Authentication to the given Request object.cCs||_||_dSN)rrrrrrrr__init__RszHTTPBasicAuth.__init__cCs(t|jt|ddk|jt|ddkgSNrrallrgetattrrrotherrrr__eq__VszHTTPBasicAuth.__eq__cCs ||k Sr&rr-rrr__ne__\szHTTPBasicAuth.__ne__cCst|j|j|jd<|S)N Authorizationrrrheadersrrrrr_szHTTPBasicAuth.__call__N)r!r"r#r$r(r/r0rrrrrr%Os r%c@seZdZdZddZdS) HTTPProxyAuthz=Attaches HTTP Proxy Authentication to a given Request object.cCst|j|j|jd<|S)NzProxy-Authorizationr2rrrrrgszHTTPProxyAuth.__call__Nr rrrrr4dsr4c@sPeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ dS)HTTPDigestAuthz@Attaches HTTP Digest Authentication to the given Request object.cCs||_||_t|_dSr&)rr threadingZlocal _thread_localr'rrrr(oszHTTPDigestAuth.__init__cCs@t|jds.md5_utf8ZSHAcSs"t|tr|d}t|SrF)r rrrHsha1rIrJrrrsha_utf8s  z4HTTPDigestAuth.build_digest_header..sha_utf8zSHA-256cSs"t|tr|d}t|SrF)r rrrHZsha256rIrJrrr sha256_utf8s  z7HTTPDigestAuth.build_digest_header..sha256_utf8zSHA-512cSs"t|tr|d}t|SrF)r rrrHZsha512rIrJrrr sha512_utf8s  z7HTTPDigestAuth.build_digest_header..sha512_utf8csd||fS)N%s:%sr)sdZ hash_utf8rrz4HTTPDigestAuth.build_digest_header../?z%s:%s:%srQrz%08xrGZauth,z%s:%s:%s:%s:%sz>username="%s", realm="%s", nonce="%s", uri="%s", response="%s"z , opaque="%s"z, algorithm="%s"z , digest="%s"z , qop="auth", nc=%s, cnonce="%s"z Digest %s)r7r=getupperrpathZqueryrrr;r<rrtimectimeosurandomrHrMrIsplit)rmethodurlrArBrCrDrEZ _algorithmrLrNrOrPZKDZentdigZp_parsedr^ZA1ZA2ZHA1ZHA2ZncvaluerRZcnonceZrespdigZnoncebitbaserrTrbuild_digest_headers           z"HTTPDigestAuth.build_digest_headercKs|jrd|j_dS)z)Reset num_401_calls counter on redirects.rN)Z is_redirectr7r?)rrkwargsrrrhandle_redirectszHTTPDigestAuth.handle_redirectcKs"d|jkrdks$nd|j_|S|jjdk rB|jj|jj|jdd}d| kr|jjdkr|jjd7_t j d t j d }t |jd|dd |j_|j||j}t|j|j|j||j||j|j|jd <|jj|f|}|j|||_|Sd|j_|S) zo Takes the given response and tries digest-auth, if needed. :rtype: requests.Response iirNzwww-authenticater9Zdigestzdigest )flags)countr1)Z status_coder7r?r>Zrequestbodyseekr3r\lowerrecompile IGNORECASEr subr=ZcontentclosecopyrZ_cookiesrawZprepare_cookiesrgrdreZ connectionsendhistoryappend)rrrhZs_authZpatZprepZ_rrrr handle_401s2      zHTTPDigestAuth.handle_401cCs~||jjr&||j|j|jd<z|j|j_ Wnt k rTd|j_ YnX| d|j | d|j d|j_|S)Nr1Zresponser)r@r7r;rgrdrer3rmtellr>AttributeErrorZ register_hookrzrir?rrrrrszHTTPDigestAuth.__call__cCs(t|jt|ddk|jt|ddkgSr)r*r-rrrr/*szHTTPDigestAuth.__eq__cCs ||k Sr&rr-rrrr00szHTTPDigestAuth.__ne__N) r!r"r#r$r(r@rgrirzrr/r0rrrrr5ls f,r5)r$rarpr_rHr6rbase64rcompatrrrZcookiesrZ_internal_utilsrZutilsr ZCONTENT_TYPE_FORM_URLENCODEDZCONTENT_TYPE_MULTI_PARTrobjectrr%r4r5rrrrs$    ,site-packages/pip/_vendor/requests/__pycache__/_internal_utils.cpython-38.opt-1.pyc000064400000002332151464452720024367 0ustar00U .eH@s.dZddlmZmZmZd ddZddZdS) z requests._internal_utils ~~~~~~~~~~~~~~ Provides utility functions that are consumed internally by Requests which depend on extremely few external helpers (such as compat) )is_py2 builtin_strstrasciicCs.t|tr|}ntr ||}n ||}|S)zGiven a string object, regardless of type, returns a representation of that string in the native string type, encoding and decoding where necessary. This assumes ASCII unless told otherwise. ) isinstancerrencodedecode)stringencodingoutr H/usr/lib/python3.8/site-packages/pip/_vendor/requests/_internal_utils.pyto_native_strings    rcCs,z|dWdStk r&YdSXdS)zDetermine if unicode string only contains ASCII characters. :param str u_string: unicode string to check. Must be unicode and not Python 2 `str`. :rtype: bool rTFN)rUnicodeEncodeError)Zu_stringr r r unicode_is_asciis  rN)r)__doc__compatrrrrrr r r r s site-packages/pip/_vendor/requests/__pycache__/exceptions.cpython-38.pyc000064400000012130151464452720022413 0ustar00U .e} @sdZddlmZGdddeZGdddeZGdddeZGd d d eZGd d d eZGd ddeZ Gdddee Z Gddde Z GdddeZ GdddeZ GdddeeZGdddeeZGdddeeZGdddeeZGdd d eZGd!d"d"eZGd#d$d$eeZGd%d&d&eeZGd'd(d(eZGd)d*d*eZGd+d,d,eZGd-d.d.eeZGd/d0d0eZd1S)2z` requests.exceptions ~~~~~~~~~~~~~~~~~~~ This module contains the set of Requests' exceptions. ) HTTPErrorcs eZdZdZfddZZS)RequestExceptionzTThere was an ambiguous exception that occurred while handling your request. csX|dd}||_|dd|_|dk rB|jsBt|drB|jj|_tt|j||dS)zBInitialize RequestException with `request` and `response` objects.responseNrequest)poprrhasattrsuperr__init__)selfargskwargsr __class__C/usr/lib/python3.8/site-packages/pip/_vendor/requests/exceptions.pyr s  zRequestException.__init__)__name__ __module__ __qualname____doc__r __classcell__rrr rr src@seZdZdZdS)rzAn HTTP error occurred.Nrrrrrrrrrsrc@seZdZdZdS)ConnectionErrorzA Connection error occurred.Nrrrrrr src@seZdZdZdS) ProxyErrorzA proxy error occurred.Nrrrrrr$src@seZdZdZdS)SSLErrorzAn SSL error occurred.Nrrrrrr(src@seZdZdZdS)TimeoutzThe request timed out. Catching this error will catch both :exc:`~requests.exceptions.ConnectTimeout` and :exc:`~requests.exceptions.ReadTimeout` errors. Nrrrrrr,src@seZdZdZdS)ConnectTimeoutzThe request timed out while trying to connect to the remote server. Requests that produced this error are safe to retry. Nrrrrrr5src@seZdZdZdS) ReadTimeoutz@The server did not send any data in the allotted amount of time.Nrrrrrr<src@seZdZdZdS) URLRequiredz*A valid URL is required to make a request.Nrrrrrr@src@seZdZdZdS)TooManyRedirectszToo many redirects.NrrrrrrDsrc@seZdZdZdS) MissingSchemaz/The URL schema (e.g. http or https) is missing.NrrrrrrHsrc@seZdZdZdS) InvalidSchemaz"See defaults.py for valid schemas.Nrrrrrr Lsr c@seZdZdZdS) InvalidURLz%The URL provided was somehow invalid.Nrrrrrr!Psr!c@seZdZdZdS) InvalidHeaderz.The header value provided was somehow invalid.Nrrrrrr"Tsr"c@seZdZdZdS)InvalidProxyURLz"The proxy URL provided is invalid.Nrrrrrr#Xsr#c@seZdZdZdS)ChunkedEncodingErrorz?The server declared chunked encoding but sent an invalid chunk.Nrrrrrr$\sr$c@seZdZdZdS)ContentDecodingErrorz!Failed to decode response contentNrrrrrr%`sr%c@seZdZdZdS)StreamConsumedErrorz2The content for this response was already consumedNrrrrrr&dsr&c@seZdZdZdS) RetryErrorzCustom retries logic failedNrrrrrr'hsr'c@seZdZdZdS)UnrewindableBodyErrorz:Requests encountered an error when trying to rewind a bodyNrrrrrr(lsr(c@seZdZdZdS)RequestsWarningzBase warning for Requests.Nrrrrrr)rsr)c@seZdZdZdS)FileModeWarningzJA file was opened in text mode, but Requests determined its binary length.Nrrrrrr*wsr*c@seZdZdZdS)RequestsDependencyWarningz@An imported dependency doesn't match the expected version range.Nrrrrrr+|sr+N)rZpip._vendor.urllib3.exceptionsrZ BaseHTTPErrorIOErrorrrrrrrrrr ValueErrorrr r!r"r#r$r% TypeErrorr&r'r(Warningr)DeprecationWarningr*r+rrrrs0  site-packages/pip/_vendor/requests/__pycache__/adapters.cpython-38.pyc000064400000041063151464452720022044 0ustar00U .e,T@sdZddlZddlZddlmZmZddlmZddl m Z ddl m Z ddl mZddlmZdd lmZdd lmZdd lmZdd lmZdd lmZddlmZddlmZddlmZddlmZddlmZddlmZddl m!Z!m"Z"ddl#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0m1Z1m2Z2mZmZm3Z3m4Z4m5Z5m6Z6ddl7m8Z8zddl9m:Z:Wne;k rddZ:YnXdZdZ?Gd d!d!e@ZAGd"d#d#eAZBdS)$z requests.adapters ~~~~~~~~~~~~~~~~~ This module contains the transport adapters that Requests uses to define and maintain connections. N) PoolManagerproxy_from_url) HTTPResponse) parse_url)Timeout)Retry)ClosedPoolError)ConnectTimeoutError) HTTPError) MaxRetryError)NewConnectionError) ProxyError) ProtocolError)ReadTimeoutError)SSLError) ResponseError)LocationValueError)Response)urlparse basestring)DEFAULT_CA_BUNDLE_PATHextract_zipped_pathsget_encoding_from_headersprepend_scheme_if_neededget_auth_from_url urldefragauth select_proxy)CaseInsensitiveDict)extract_cookies_to_jar) ConnectionErrorConnectTimeout ReadTimeoutrr RetryError InvalidSchemaInvalidProxyURL InvalidURL)_basic_auth_str)SOCKSProxyManagercOs tddS)Nz'Missing dependencies for SOCKS support.)r$)argskwargsr+A/usr/lib/python3.8/site-packages/pip/_vendor/requests/adapters.pyr(.sr(F cs2eZdZdZfddZd ddZd d ZZS) BaseAdapterzThe Base Transport Adaptercstt|dSN)superr.__init__self __class__r+r,r1:szBaseAdapter.__init__FNTcCstdS)aCSends PreparedRequest object. Returns Response object. :param request: The :class:`PreparedRequest ` being sent. :param stream: (optional) Whether to stream the request content. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: (optional) Any user-provided SSL certificate to be trusted. :param proxies: (optional) The proxies dictionary to apply to the request. NNotImplementedError)r3requeststreamtimeoutverifycertproxiesr+r+r,send=szBaseAdapter.sendcCstdS)z!Cleans up adapter specific items.Nr6r2r+r+r,closeOszBaseAdapter.close)FNTNN)__name__ __module__ __qualname____doc__r1r>r? __classcell__r+r+r4r,r.7s  r.cseZdZdZdddddgZeeeeffdd Zd d Z d d Z efd dZ ddZ ddZ ddZd$ddZddZddZddZddZd%d"d#ZZS)& HTTPAdapteraThe built-in HTTP Adapter for urllib3. Provides a general-case interface for Requests sessions to contact HTTP and HTTPS urls by implementing the Transport Adapter interface. This class will usually be created by the :class:`Session ` class under the covers. :param pool_connections: The number of urllib3 connection pools to cache. :param pool_maxsize: The maximum number of connections to save in the pool. :param max_retries: The maximum number of retries each connection should attempt. Note, this applies only to failed DNS lookups, socket connections and connection timeouts, never to requests where data has made it to the server. By default, Requests does not retry failed connections. If you need granular control over the conditions under which we retry a request, import urllib3's ``Retry`` class and pass that instead. :param pool_block: Whether the connection pool should block for connections. Usage:: >>> import requests >>> s = requests.Session() >>> a = requests.adapters.HTTPAdapter(max_retries=3) >>> s.mount('http://', a) max_retriesconfig_pool_connections _pool_maxsize _pool_blockcsd|tkrtddd|_n t||_i|_i|_tt|||_ ||_ ||_ |j |||ddS)NrF)readblock) DEFAULT_RETRIESrrFZfrom_intrG proxy_managerr0rEr1rHrIrJinit_poolmanager)r3Zpool_connectionsZ pool_maxsizerFZ pool_blockr4r+r,r1qs zHTTPAdapter.__init__csfddjDS)Ncsi|]}|t|dqSr/)getattr).0attrr2r+r, sz,HTTPAdapter.__getstate__..) __attrs__r2r+r2r, __getstate__szHTTPAdapter.__getstate__cCsDi|_i|_|D]\}}t|||q|j|j|j|jddS)NrL)rOrGitemssetattrrPrHrIrJ)r3staterSvaluer+r+r, __setstate__s zHTTPAdapter.__setstate__cKs0||_||_||_tf|||dd||_dS)aInitializes a urllib3 PoolManager. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param connections: The number of urllib3 connection pools to cache. :param maxsize: The maximum number of connections to save in the pool. :param block: Block when no free connections are available. :param pool_kwargs: Extra keyword arguments used to initialize the Pool Manager. T) num_poolsmaxsizerMstrictN)rHrIrJr poolmanager)r3Z connectionsr]rMZ pool_kwargsr+r+r,rPs zHTTPAdapter.init_poolmanagercKs||jkr|j|}n||dr^t|\}}t|f|||j|j|jd|}|j|<n4||}t |f||j|j|jd|}|j|<|S)aReturn urllib3 ProxyManager for the given proxy. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param proxy: The proxy to return a urllib3 ProxyManager for. :param proxy_kwargs: Extra keyword arguments used to configure the Proxy Manager. :returns: ProxyManager :rtype: urllib3.ProxyManager socks)usernamepasswordr\r]rM) proxy_headersr\r]rM) rOlower startswithrr(rHrIrJrcr)r3proxyZ proxy_kwargsZmanagerrarbrcr+r+r,proxy_manager_fors6    zHTTPAdapter.proxy_manager_forcCs|drn|rnd}|dk r"|}|s.tt}|r>tj|sLtd|d|_ tj |sf||_ q||_ nd|_ d|_ d|_ |rt |ts|d|_|d|_n ||_d|_|jrtj|jstd |j|jrtj|jstd |jdS) aAVerify a SSL certificate. This method should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param conn: The urllib3 connection object associated with the cert. :param url: The requested URL. :param verify: Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: The SSL certificate to verify. httpsNTzECould not find a suitable TLS CA certificate bundle, invalid path: {}Z CERT_REQUIREDZ CERT_NONErrz9Could not find the TLS certificate file, invalid path: {}z1Could not find the TLS key file, invalid path: {})rdrerrospathexistsIOErrorformatZ cert_reqsisdirZca_certsZ ca_cert_dir isinstancerZ cert_fileZkey_file)r3connurlr;r<Zcert_locr+r+r, cert_verifys>     zHTTPAdapter.cert_verifycCst}t|dd|_tt|di|_t|j|_||_|jj|_t |j t r^|j d|_ n|j |_ t |j||||_||_|S)aBuilds a :class:`Response ` object from a urllib3 response. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter ` :param req: The :class:`PreparedRequest ` used to generate the response. :param resp: The urllib3 response object. :rtype: requests.Response ZstatusNheadersutf-8)rrQZ status_coderrsrencodingrawreasonrorqbytesdecodercookiesr8 connection)r3ZreqrespZresponser+r+r,build_responses    zHTTPAdapter.build_responseNcCsdt||}|rDt|d}t|}|js.td||}||}nt|}|}|j |}|S)aReturns a urllib3 connection for the given URL. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param url: The URL to connect to. :param proxies: (optional) A Requests-style dictionary of proxies used on this request. :rtype: urllib3.ConnectionPool ZhttpzFPlease check proxy URL. It is malformed and could be missing the host.) rrrZhostr%rgZconnection_from_urlrZgeturlr_)r3rqr=rfZ proxy_urlrOrpZparsedr+r+r,get_connection$s     zHTTPAdapter.get_connectioncCs&|j|jD] }|qdS)zDisposes of any internal state. Currently, this closes the PoolManager and any active ProxyManager, which closes any pooled connections. N)r_clearrOvalues)r3rfr+r+r,r??s zHTTPAdapter.closec Cs`t|j|}t|jj}|o"|dk}d}|rDt|j}|d}|j}|r\|s\t|j}|S)a?Obtain the url to use when making the final request. If the message is being sent through a HTTP proxy, the full URL has to be used. Otherwise, we should only use the path portion of the URL. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param request: The :class:`PreparedRequest ` being sent. :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs. :rtype: str rhFr`)rrqrschemerdreZpath_urlr) r3r8r=rfrZis_proxied_http_requestZusing_socks_proxyZ proxy_schemerqr+r+r, request_urlIs     zHTTPAdapter.request_urlcKsdS)a"Add any headers needed by the connection. As of v2.0 this does nothing by default, but is left for overriding by users that subclass the :class:`HTTPAdapter `. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param request: The :class:`PreparedRequest ` to add headers to. :param kwargs: The keyword arguments from the call to send(). Nr+)r3r8r*r+r+r, add_headersfs zHTTPAdapter.add_headerscCs&i}t|\}}|r"t|||d<|S)aReturns a dictionary of the headers to add to any request sent through a proxy. This works with urllib3 magic to ensure that they are correctly sent to the proxy, rather than in a tunnelled request if CONNECT is being used. This should not be called from user code, and is only exposed for use when subclassing the :class:`HTTPAdapter `. :param proxy: The url of the proxy being used for this request. :rtype: dict zProxy-Authorization)rr')r3rfrsrarbr+r+r,rcts  zHTTPAdapter.proxy_headersFTc Csz||j|}Wn.tk r@}zt||dW5d}~XYnX|||j|||||} |j||||||d|jdkpd|jk } t |t rz|\} } t | | d}Wqt k r}zd |} t | W5d}~XYqXnt |t rn t ||d}zF| s2|j|j| |j|jdddd|j|d }nt|d rD|j}|jtd }z|j|j| d d |jD]\}}|||qn||jD]D}|tt|d dd|d|||dq|dz|jd d}Wntk r|}YnXtj |||ddd}Wn|!YnXWnt"t#j$fk r~} zt%| |dW5d} ~ XYnZt&k r"}zt |j't(rt |j't)st*||dt |j't+rt,||dt |j't-rt.||dt |j't/rt0||dt%||dW5d}~XYnt1k rP}zt%||dW5d}~XYnt-k rz}z t.|W5d}~XYn^t/t2fk r}z:t |t/rt0||dnt |t3rt4||dnW5d}~XYnX|5||S)aSends PreparedRequest object. Returns Response object. :param request: The :class:`PreparedRequest ` being sent. :param stream: (optional) Whether to stream the request content. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple or urllib3 Timeout object :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use :param cert: (optional) Any user-provided SSL certificate to be trusted. :param proxies: (optional) The proxies dictionary to apply to the request. :rtype: requests.Response )r8N)r9r:r;r<r=zContent-Length)connectrKzrInvalid timeout {}. Pass a (connect, read) timeout tuple, or a single float to set both timeouts to the same valueF) methodrqbodyrsZredirectZassert_same_hostpreload_contentdecode_contentZretriesr: proxy_pool)r:T)Zskip_accept_encodingrts s0 ) buffering)Zpoolr{rr)6r~rqrr&rrrrrrsrotuple TimeoutSauce ValueErrorrmZurlopenrrFhasattrrZ _get_connDEFAULT_POOL_TIMEOUTZ putrequestrWZ putheaderZ endheadersr>hexlenencodeZ getresponse TypeErrorrZ from_httplibr?rsocketerrorr r rwr r r!rr# _ProxyErrorr _SSLErrorrr _HTTPErrorrr"r})r3r8r9r:r;r<r=rperqZchunkedrrKerrr|Zlow_connheaderrZirr+r+r,r>s                   zHTTPAdapter.send)N)FNTNN)r@rArBrCrUDEFAULT_POOLSIZErNDEFAULT_POOLBLOCKr1rVr[rPrgrrr}r~r?rrrcr>rDr+r+r4r,rETs* %4%  rE)CrCZos.pathrirZpip._vendor.urllib3.poolmanagerrrZpip._vendor.urllib3.responserZpip._vendor.urllib3.utilrrrZpip._vendor.urllib3.util.retryrZpip._vendor.urllib3.exceptionsrr r rr r r rrrrrrrZmodelsrcompatrrZutilsrrrrrrrZ structuresrrzr exceptionsr r!r"r#r$r%r&Zauthr'Z!pip._vendor.urllib3.contrib.socksr( ImportErrorrrrNrobjectr.rEr+r+r+r,sF                $  , site-packages/pip/_vendor/requests/__pycache__/structures.cpython-38.pyc000064400000010452151464452720022462 0ustar00U .e @s<dZddlmZmZmZGdddeZGdddeZdS)zO requests.structures ~~~~~~~~~~~~~~~~~~~ Data structures that power Requests. ) OrderedDictMappingMutableMappingc@sbeZdZdZdddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ dS)CaseInsensitiveDictaA case-insensitive ``dict``-like object. Implements all methods and operations of ``MutableMapping`` as well as dict's ``copy``. Also provides ``lower_items``. All keys are expected to be strings. The structure remembers the case of the last key to be set, and ``iter(instance)``, ``keys()``, ``items()``, ``iterkeys()``, and ``iteritems()`` will contain case-sensitive keys. However, querying and contains testing is case insensitive:: cid = CaseInsensitiveDict() cid['Accept'] = 'application/json' cid['aCCEPT'] == 'application/json' # True list(cid) == ['Accept'] # True For example, ``headers['content-encoding']`` will return the value of a ``'Content-Encoding'`` response header, regardless of how the header name was originally stored. If the constructor, ``.update``, or equality comparison operations are given keys that have equal ``.lower()``s, the behavior is undefined. NcKs&t|_|dkri}|j|f|dSN)r_storeupdate)selfdatakwargsr C/usr/lib/python3.8/site-packages/pip/_vendor/requests/structures.py__init__(szCaseInsensitiveDict.__init__cCs||f|j|<dSrrlower)r keyvaluer r r __setitem__.szCaseInsensitiveDict.__setitem__cCs|j|dS)Nrrr rr r r __getitem__3szCaseInsensitiveDict.__getitem__cCs|j|=dSrrrr r r __delitem__6szCaseInsensitiveDict.__delitem__cCsdd|jDS)Ncss|]\}}|VqdSrr ).0ZcasedkeyZ mappedvaluer r r :sz/CaseInsensitiveDict.__iter__..)rvaluesr r r r __iter__9szCaseInsensitiveDict.__iter__cCs t|jSr)lenrrr r r __len__<szCaseInsensitiveDict.__len__cCsdd|jDS)z.Like iteritems(), but with all lowercase keys.css|]\}}||dfVqdS)rNr )rZlowerkeyZkeyvalr r r rAsz2CaseInsensitiveDict.lower_items..)ritemsrr r r lower_items?szCaseInsensitiveDict.lower_itemscCs0t|trt|}ntSt|t|kSr) isinstancerrNotImplementeddictr)r otherr r r __eq__Gs  zCaseInsensitiveDict.__eq__cCst|jSr)rrrrr r r copyPszCaseInsensitiveDict.copycCstt|Sr)strr"rrr r r __repr__SszCaseInsensitiveDict.__repr__)N)__name__ __module__ __qualname____doc__rrrrrrrr$r%r'r r r r r s  rcs<eZdZdZd fdd ZddZddZd d d ZZS) LookupDictzDictionary lookup object.Ncs||_tt|dSr)namesuperr,r)r r- __class__r r rZszLookupDict.__init__cCs d|jS)Nz )r-rr r r r'^szLookupDict.__repr__cCs|j|dSr__dict__getrr r r raszLookupDict.__getitem__cCs|j||Srr1)r rdefaultr r r r3fszLookupDict.get)N)N) r(r)r*r+rr'rr3 __classcell__r r r/r r,Ws r,N)r+compatrrrrr"r,r r r r sJsite-packages/pip/_vendor/requests/__pycache__/api.cpython-38.opt-1.pyc000064400000014507151464452720021754 0ustar00U .e@s\dZddlmZddZdddZdd Zd d Zdd d ZdddZdddZ ddZ dS)z requests.api ~~~~~~~~~~~~ This module implements the Requests API. :copyright: (c) 2012 by Kenneth Reitz. :license: Apache2, see LICENSE for more details. )sessionsc Ks6t$}|jf||d|W5QRSQRXdS)a Constructs and sends a :class:`Request `. :param method: method for the new :class:`Request` object. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary, list of tuples or bytes to send in the query string for the :class:`Request`. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) A JSON serializable Python object to send in the body of the :class:`Request`. :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`. :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`. :param files: (optional) Dictionary of ``'name': file-like-objects`` (or ``{'name': file-tuple}``) for multipart encoding upload. ``file-tuple`` can be a 2-tuple ``('filename', fileobj)``, 3-tuple ``('filename', fileobj, 'content_type')`` or a 4-tuple ``('filename', fileobj, 'content_type', custom_headers)``, where ``'content-type'`` is a string defining the content type of the given file and ``custom_headers`` a dict-like object containing additional headers to add for the file. :param auth: (optional) Auth tuple to enable Basic/Digest/Custom HTTP Auth. :param timeout: (optional) How many seconds to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param allow_redirects: (optional) Boolean. Enable/disable GET/OPTIONS/POST/PUT/PATCH/DELETE/HEAD redirection. Defaults to ``True``. :type allow_redirects: bool :param proxies: (optional) Dictionary mapping protocol to the URL of the proxy. :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use. Defaults to ``True``. :param stream: (optional) if ``False``, the response content will be immediately downloaded. :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair. :return: :class:`Response ` object :rtype: requests.Response Usage:: >>> import requests >>> req = requests.request('GET', 'https://httpbin.org/get') )methodurlN)rZSessionrequest)rrkwargsZsessionr` object :rtype: requests.Response allow_redirectsTgetparams setdefaultr)rr rrrrr ?s r cKs|ddtd|f|S)zSends an OPTIONS request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response r Toptionsr rrrrrrNs rcKs|ddtd|f|S)zSends a HEAD request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response r Fheadr rrrrr[s rcKstd|f||d|S)aSends a POST request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response post)datajsonr)rrrrrrrrhs rcKstd|fd|i|S)aSends a PUT request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response putrrrrrrrrrws rcKstd|fd|i|S)aSends a PATCH request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response patchrrrrrrrs rcKstd|f|S)zSends a DELETE request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response deleterrrrrrs r)N)NN)N)N) __doc__rrr rrrrrrrrrrs /     site-packages/pip/_vendor/requests/__pycache__/sessions.cpython-38.pyc000064400000046062151464452720022113 0ustar00U .er@sdZddlZddlZddlZddlmZddlmZddlm Z m Z m Z m Z m Z mZddlmZmZmZmZddlmZmZmZdd lmZmZdd lmZdd lmZmZm Z dd l!m"Z"m#Z#m$Z$m%Z%dd l&m'Z'ddl(m)Z)ddlm*Z*m+Z+m,Z,m-Z-m.Z.m/Z/ddl0m1Z1ddlm2Z2ej3dkrNz ej4Z5Wne6k rJej7Z5YnXnejZ5e fddZ8e fddZ9Gddde:Z;Gddde;ZIsz!merge_setting..) isinstancer rupdateitems)Zrequest_settingZsession_setting dict_classZmerged_settingZ none_keyskeyr'r'r+ merge_setting1s r2cCs@|dks|dgkr|S|dks0|dgkr4|St|||S)zProperly merges both requests and session hooks. This is necessary because when request_hooks == {'response': []}, the merge breaks Session hooks entirely. Nresponse)getr2)Z request_hooksZ session_hooksr0r'r'r+ merge_hooksPs r5c@s>eZdZddZddZddd Zd d Zd d ZddZdS)SessionRedirectMixincCs,|jr(|jd}tr|d}t|dSdS)z7Receives a Response. Returns a redirect URI or ``None``locationlatin1utf8N)Z is_redirectheadersrencoder)selfrespr7r'r'r+get_redirect_targetas    z(SessionRedirectMixin.get_redirect_targetcCst|}t|}|j|jkr dS|jdkrL|jdkrL|jdkrL|jdkrLdS|j|jk}|j|jk}t|jddf}|s|j|kr|j|krdS|p|S)zFDecide whether Authorization header should be removed when redirectingTZhttp)PNZhttps)iNFN)r ZhostnameschemeZportrr4)r<Zold_urlZnew_urlZ old_parsedZ new_parsedZ changed_portZchanged_schemeZ default_portr'r'r+should_strip_authvs$   z&SessionRedirectMixin.should_strip_authFNTc  ksZg} ||} t|jj} | rV|} | || dd|_z |jWn(tt t fk rr|j j ddYnXt |j|jkrtd|j|d|| drt|j}dt|j| f} t| }|jd kr| r|j| d }n |jr|j} |} |jst|jt| } nt| } t| | _|| ||jtjtjfkrld }|D]}| j|dqPd| _ | j}z |d =Wnt!k rYnXt"| j#||j t$| j#|j%| &| j#|'| |}|(| || j)dk od |kpd|k}|rt*| | }|r|Vq|j+|f|||||dd| }t"|j%| |j ||} |VqdS)zBReceives a Response. Returns a generator of Responses or Requests.rNF)Zdecode_contentzExceeded %s redirects.)r3z//z%s:%s)fragment)Content-Lengthz Content-TypeTransfer-EncodingZCookierDrE)streamtimeoutverifycertproxiesallow_redirects),r>r urlrCcopyappendhistorycontentrr RuntimeErrorrawreadlen max_redirectsrclose startswithrr@_replaceZgeturlZnetlocrrrebuild_method status_coder$Ztemporary_redirectZpermanent_redirectr:popZbodyKeyErrorr Z_cookiesrcookiesZprepare_cookiesrebuild_proxies rebuild_authZ_body_positionr#send)r<r=reqrFrGrHrIrJyield_requestsZadapter_kwargsZhistrLZprevious_fragmentprepared_requestZ parsed_rurlZparsedZpurged_headersheaderr:Z rewindabler'r'r+resolve_redirectss               z&SessionRedirectMixin.resolve_redirectscCsR|j}|j}d|kr*||jj|r*|d=|jr8t|nd}|dk rN||dS)zWhen being redirected we may want to strip authentication from the request to avoid leaking credentials. This method intelligently removes and reapplies authentication where possible to avoid credential loss. Z AuthorizationN)r:rLrArequest trust_envr Z prepare_auth)r<rcr3r:rLZnew_authr'r'r+r_s z!SessionRedirectMixin.rebuild_authc Cs|dk r |ni}|j}|j}t|j}|}|d}t||d}|jr||s|t||d} | || d} | r|| || d|kr|d=zt ||\} } Wnt k rd\} } YnX| r| rt | | |d<|S)aThis method re-evaluates the proxy configuration by considering the environment variables. If we are redirected to a URL covered by NO_PROXY, we strip the proxy configuration. Otherwise, we set missing proxy keys for this URL (in case they were stripped by a previous redirect). This method also replaces the Proxy-Authorization header where necessary. :rtype: dict Nno_proxyrhallzProxy-Authorization)NN) r:rLr r@rMr4r!rgr setdefaultr"r\r) r<rcrJr:rLr@Z new_proxiesrhZ bypass_proxyZenviron_proxiesproxyZusernameZpasswordr'r'r+r^s*       z$SessionRedirectMixin.rebuild_proxiescCsX|j}|jtjkr|dkrd}|jtjkr6|dkr6d}|jtjkrN|dkrNd}||_dS)zWhen being redirected we may want to change the method of the request based on certain specs or browser behavior. HEADGETPOSTN)methodrZr$Z see_otherfoundZmoved)r<rcr3rpr'r'r+rY=sz#SessionRedirectMixin.rebuild_method)FNTNNF) __name__ __module__ __qualname__r>rArer_r^rYr'r'r'r+r6_s q)r6c@seZdZdZdddddddd d d d d dg ZddZddZddZddZd7ddZ ddZ ddZ dd Z d8d!d"Z d9d#d$Zd:d%d&Zd'd(Zd)d*Zd+d,Zd-d.Zd/d0Zd1d2Zd3d4Zd5d6ZdS);SessionaA Requests session. Provides cookie persistence, connection-pooling, and configuration. Basic Usage:: >>> import requests >>> s = requests.Session() >>> s.get('https://httpbin.org/get') Or as a context manager:: >>> with requests.Session() as s: >>> s.get('https://httpbin.org/get') r:r]authrJhooksparamsrHrIZprefetchadaptersrFrgrUcCsrt|_d|_i|_t|_i|_d|_d|_d|_ t |_ d|_ t i|_t|_|dt|dtdS)NFTzhttps://zhttp://)rr:rvrJrrwrxrFrHrIrrUrgr r]rrymountrr<r'r'r+__init__ms zSession.__init__cCs|Sr&r'r{r'r'r+ __enter__szSession.__enter__cGs |dSr&)rV)r<argsr'r'r+__exit__szSession.__exit__c Cs|jpi}t|tjst|}ttt|j|}|j}|jrR|sR|jsRt |j }t }|j |j |j |j|j|jt|j|jtdt|j|jt||j|t|j|jd |S)aConstructs a :class:`PreparedRequest ` for transmission and returns it. The :class:`PreparedRequest` has settings merged from the :class:`Request ` instance and those of the :class:`Session`. :param request: :class:`Request` instance to prepare with this session's settings. :rtype: requests.PreparedRequest )r0) rprLfilesdatajsonr:rxrvr]rw)r]r-rZ CookieJarr rr rvrgr rLrZpreparerpupperrrrr2r:rrxr5rw)r<rfr]Zmerged_cookiesrvpr'r'r+prepare_requests0        zSession.prepare_requestNTc Cstt|||||pi||pi||| d }||}| p8i} ||j| | ||}| | d}|||j|f|}|S)aConstructs a :class:`Request `, prepares it and sends it. Returns :class:`Response ` object. :param method: method for the new :class:`Request` object. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary or bytes to be sent in the query string for the :class:`Request`. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json to send in the body of the :class:`Request`. :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`. :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`. :param files: (optional) Dictionary of ``'filename': file-like-objects`` for multipart encoding upload. :param auth: (optional) Auth tuple or callable to enable Basic/Digest/Custom HTTP Auth. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param allow_redirects: (optional) Set to True by default. :type allow_redirects: bool :param proxies: (optional) Dictionary mapping protocol or protocol and hostname to the URL of the proxy. :param stream: (optional) whether to immediately download the response content. Defaults to ``False``. :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use. Defaults to ``True``. :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair. :rtype: requests.Response ) rprLr:rrrrxrvr]rw)rGrK)rrrmerge_environment_settingsrLr.r`)r<rprLrxrr:r]rrvrGrKrJrwrFrHrIrraZprepZsettingsZ send_kwargsr=r'r'r+rfs6)  zSession.requestcKs|dd|jd|f|S)zSends a GET request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response rKTrnrkrfr<rLkwargsr'r'r+r4s z Session.getcKs|dd|jd|f|S)zSends a OPTIONS request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response rKTZOPTIONSrrr'r'r+options$s zSession.optionscKs|dd|jd|f|S)zSends a HEAD request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response rKFrmrrr'r'r+head/s z Session.headcKs|jd|f||d|S)aSends a POST request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ro)rrrf)r<rLrrrr'r'r+post:s z Session.postcKs|jd|fd|i|S)auSends a PUT request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ZPUTrrr<rLrrr'r'r+putGs z Session.putcKs|jd|fd|i|S)awSends a PATCH request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ZPATCHrrrr'r'r+patchSs z Session.patchcKs|jd|f|S)zSends a DELETE request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ZDELETErrr'r'r+delete_szSession.deletec Ksz|d|j|d|j|d|j|d|jt|trJtd|dd}| d}|j }|j |j d}t }|j|f|}t |} t| d |_td ||f|}|jr|jD]} t|j| j| jqt|j||j|j||f|} |rd d | Dng} | r*| d || }| |_|sjz"t|j||fddi||_Wntk rhYnX|sv|j|S)zISend a given PreparedRequest. :rtype: requests.Response rFrHrIrJz#You can only send PreparedRequests.rKT)rL)Zsecondsr3cSsg|]}|qSr'r')r(r=r'r'r+r,sz Session.send..rrb)rkrFrHrIrJr-r ValueErrorr[r4rw get_adapterrLpreferred_clockr`relapsedrrOr r]rfrRreinsertnextZ_next StopIterationrP) r<rfrrKrFrwadapterstartrrr=genrOr'r'r+r`isB       "z Session.sendc Cs|jrn|dk r|dnd}t||d}|D]\}} ||| q0|dksV|dkrntjdpltjd}t||j}t||j }t||j }t||j }||||dS)z^ Check the environment and merge it with some settings. :rtype: dict NrhriTZREQUESTS_CA_BUNDLEZCURL_CA_BUNDLE)rHrJrFrI) rgr4rr/rkosenvironr2rJrFrHrI) r<rLrJrFrHrIrhZ env_proxiesr)r*r'r'r+rs        z"Session.merge_environment_settingscCs>|jD]"\}}||r |Sq td|dS)z~ Returns the appropriate connection adapter for the given URL. :rtype: requests.adapters.BaseAdapter z*No connection adapters were found for '%s'N)ryr/lowerrWr)r<rLprefixrr'r'r+rs zSession.get_adaptercCs|jD] }|q dS)z+Closes all adapters and as such the sessionN)ryvaluesrV)r<r*r'r'r+rVsz Session.closecs>||j<fdd|jD}|D]}|j||j|<q"dS)zwRegisters a connection adapter to a prefix. Adapters are sorted in descending order by prefix length. cs g|]}t|tkr|qSr')rT)r(r)rr'r+r,sz!Session.mount..N)ryr[)r<rrZ keys_to_mover1r'rr+rzs z Session.mountcsfddjD}|S)Ncsi|]}|t|dqSr&)getattr)r(attrr{r'r+ sz(Session.__getstate__..) __attrs__)r<stater'r{r+ __getstate__szSession.__getstate__cCs"|D]\}}t|||qdSr&)r/setattr)r<rrvaluer'r'r+ __setstate__szSession.__setstate__)NNNNNNNTNNNNNN)NN)N)N)rrrsrt__doc__rr|r}rrrfr4rrrrrrr`rrrVrzrrr'r'r'r+ruTs`7) G    I rucCstS)aZ Returns a :class:`Session` for context-management. .. deprecated:: 1.0.0 This method has been deprecated since version 1.0.0 and is only kept for backwards compatibility. New code should use :class:`~requests.sessions.Session` to create a session. This may be removed at a future date. :rtype: Session )rur'r'r'r+sessions r)>rrsystimeZdatetimerrvrcompatrrrrr r r]r r r rZmodelsrrrrwrrZ_internal_utilsrZutilsrrr exceptionsrrrrZ structuresrryrrrr r!r"r#Z status_codesr$r%platform perf_counterrAttributeErrorZclockr2r5objectr6rurr'r'r'r+s@             v%site-packages/pip/_vendor/requests/__pycache__/sessions.cpython-38.opt-1.pyc000064400000046062151464452720023052 0ustar00U .er@sdZddlZddlZddlZddlmZddlmZddlm Z m Z m Z m Z m Z mZddlmZmZmZmZddlmZmZmZdd lmZmZdd lmZdd lmZmZm Z dd l!m"Z"m#Z#m$Z$m%Z%dd l&m'Z'ddl(m)Z)ddlm*Z*m+Z+m,Z,m-Z-m.Z.m/Z/ddl0m1Z1ddlm2Z2ej3dkrNz ej4Z5Wne6k rJej7Z5YnXnejZ5e fddZ8e fddZ9Gddde:Z;Gddde;ZIsz!merge_setting..) isinstancer rupdateitems)Zrequest_settingZsession_setting dict_classZmerged_settingZ none_keyskeyr'r'r+ merge_setting1s r2cCs@|dks|dgkr|S|dks0|dgkr4|St|||S)zProperly merges both requests and session hooks. This is necessary because when request_hooks == {'response': []}, the merge breaks Session hooks entirely. Nresponse)getr2)Z request_hooksZ session_hooksr0r'r'r+ merge_hooksPs r5c@s>eZdZddZddZddd Zd d Zd d ZddZdS)SessionRedirectMixincCs,|jr(|jd}tr|d}t|dSdS)z7Receives a Response. Returns a redirect URI or ``None``locationlatin1utf8N)Z is_redirectheadersrencoder)selfrespr7r'r'r+get_redirect_targetas    z(SessionRedirectMixin.get_redirect_targetcCst|}t|}|j|jkr dS|jdkrL|jdkrL|jdkrL|jdkrLdS|j|jk}|j|jk}t|jddf}|s|j|kr|j|krdS|p|S)zFDecide whether Authorization header should be removed when redirectingTZhttp)PNZhttps)iNFN)r ZhostnameschemeZportrr4)r<Zold_urlZnew_urlZ old_parsedZ new_parsedZ changed_portZchanged_schemeZ default_portr'r'r+should_strip_authvs$   z&SessionRedirectMixin.should_strip_authFNTc  ksZg} ||} t|jj} | rV|} | || dd|_z |jWn(tt t fk rr|j j ddYnXt |j|jkrtd|j|d|| drt|j}dt|j| f} t| }|jd kr| r|j| d }n |jr|j} |} |jst|jt| } nt| } t| | _|| ||jtjtjfkrld }|D]}| j|dqPd| _ | j}z |d =Wnt!k rYnXt"| j#||j t$| j#|j%| &| j#|'| |}|(| || j)dk od |kpd|k}|rt*| | }|r|Vq|j+|f|||||dd| }t"|j%| |j ||} |VqdS)zBReceives a Response. Returns a generator of Responses or Requests.rNF)Zdecode_contentzExceeded %s redirects.)r3z//z%s:%s)fragment)Content-Lengthz Content-TypeTransfer-EncodingZCookierDrE)streamtimeoutverifycertproxiesallow_redirects),r>r urlrCcopyappendhistorycontentrr RuntimeErrorrawreadlen max_redirectsrclose startswithrr@_replaceZgeturlZnetlocrrrebuild_method status_coder$Ztemporary_redirectZpermanent_redirectr:popZbodyKeyErrorr Z_cookiesrcookiesZprepare_cookiesrebuild_proxies rebuild_authZ_body_positionr#send)r<r=reqrFrGrHrIrJyield_requestsZadapter_kwargsZhistrLZprevious_fragmentprepared_requestZ parsed_rurlZparsedZpurged_headersheaderr:Z rewindabler'r'r+resolve_redirectss               z&SessionRedirectMixin.resolve_redirectscCsR|j}|j}d|kr*||jj|r*|d=|jr8t|nd}|dk rN||dS)zWhen being redirected we may want to strip authentication from the request to avoid leaking credentials. This method intelligently removes and reapplies authentication where possible to avoid credential loss. Z AuthorizationN)r:rLrArequest trust_envr Z prepare_auth)r<rcr3r:rLZnew_authr'r'r+r_s z!SessionRedirectMixin.rebuild_authc Cs|dk r |ni}|j}|j}t|j}|}|d}t||d}|jr||s|t||d} | || d} | r|| || d|kr|d=zt ||\} } Wnt k rd\} } YnX| r| rt | | |d<|S)aThis method re-evaluates the proxy configuration by considering the environment variables. If we are redirected to a URL covered by NO_PROXY, we strip the proxy configuration. Otherwise, we set missing proxy keys for this URL (in case they were stripped by a previous redirect). This method also replaces the Proxy-Authorization header where necessary. :rtype: dict Nno_proxyrhallzProxy-Authorization)NN) r:rLr r@rMr4r!rgr setdefaultr"r\r) r<rcrJr:rLr@Z new_proxiesrhZ bypass_proxyZenviron_proxiesproxyZusernameZpasswordr'r'r+r^s*       z$SessionRedirectMixin.rebuild_proxiescCsX|j}|jtjkr|dkrd}|jtjkr6|dkr6d}|jtjkrN|dkrNd}||_dS)zWhen being redirected we may want to change the method of the request based on certain specs or browser behavior. HEADGETPOSTN)methodrZr$Z see_otherfoundZmoved)r<rcr3rpr'r'r+rY=sz#SessionRedirectMixin.rebuild_method)FNTNNF) __name__ __module__ __qualname__r>rArer_r^rYr'r'r'r+r6_s q)r6c@seZdZdZdddddddd d d d d dg ZddZddZddZddZd7ddZ ddZ ddZ dd Z d8d!d"Z d9d#d$Zd:d%d&Zd'd(Zd)d*Zd+d,Zd-d.Zd/d0Zd1d2Zd3d4Zd5d6ZdS);SessionaA Requests session. Provides cookie persistence, connection-pooling, and configuration. Basic Usage:: >>> import requests >>> s = requests.Session() >>> s.get('https://httpbin.org/get') Or as a context manager:: >>> with requests.Session() as s: >>> s.get('https://httpbin.org/get') r:r]authrJhooksparamsrHrIZprefetchadaptersrFrgrUcCsrt|_d|_i|_t|_i|_d|_d|_d|_ t |_ d|_ t i|_t|_|dt|dtdS)NFTzhttps://zhttp://)rr:rvrJrrwrxrFrHrIrrUrgr r]rrymountrr<r'r'r+__init__ms zSession.__init__cCs|Sr&r'r{r'r'r+ __enter__szSession.__enter__cGs |dSr&)rV)r<argsr'r'r+__exit__szSession.__exit__c Cs|jpi}t|tjst|}ttt|j|}|j}|jrR|sR|jsRt |j }t }|j |j |j |j|j|jt|j|jtdt|j|jt||j|t|j|jd |S)aConstructs a :class:`PreparedRequest ` for transmission and returns it. The :class:`PreparedRequest` has settings merged from the :class:`Request ` instance and those of the :class:`Session`. :param request: :class:`Request` instance to prepare with this session's settings. :rtype: requests.PreparedRequest )r0) rprLfilesdatajsonr:rxrvr]rw)r]r-rZ CookieJarr rr rvrgr rLrZpreparerpupperrrrr2r:rrxr5rw)r<rfr]Zmerged_cookiesrvpr'r'r+prepare_requests0        zSession.prepare_requestNTc Cstt|||||pi||pi||| d }||}| p8i} ||j| | ||}| | d}|||j|f|}|S)aConstructs a :class:`Request `, prepares it and sends it. Returns :class:`Response ` object. :param method: method for the new :class:`Request` object. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary or bytes to be sent in the query string for the :class:`Request`. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json to send in the body of the :class:`Request`. :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`. :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`. :param files: (optional) Dictionary of ``'filename': file-like-objects`` for multipart encoding upload. :param auth: (optional) Auth tuple or callable to enable Basic/Digest/Custom HTTP Auth. :param timeout: (optional) How long to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param allow_redirects: (optional) Set to True by default. :type allow_redirects: bool :param proxies: (optional) Dictionary mapping protocol or protocol and hostname to the URL of the proxy. :param stream: (optional) whether to immediately download the response content. Defaults to ``False``. :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use. Defaults to ``True``. :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair. :rtype: requests.Response ) rprLr:rrrrxrvr]rw)rGrK)rrrmerge_environment_settingsrLr.r`)r<rprLrxrr:r]rrvrGrKrJrwrFrHrIrraZprepZsettingsZ send_kwargsr=r'r'r+rfs6)  zSession.requestcKs|dd|jd|f|S)zSends a GET request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response rKTrnrkrfr<rLkwargsr'r'r+r4s z Session.getcKs|dd|jd|f|S)zSends a OPTIONS request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response rKTZOPTIONSrrr'r'r+options$s zSession.optionscKs|dd|jd|f|S)zSends a HEAD request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response rKFrmrrr'r'r+head/s z Session.headcKs|jd|f||d|S)aSends a POST request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ro)rrrf)r<rLrrrr'r'r+post:s z Session.postcKs|jd|fd|i|S)auSends a PUT request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ZPUTrrr<rLrrr'r'r+putGs z Session.putcKs|jd|fd|i|S)awSends a PATCH request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ZPATCHrrrr'r'r+patchSs z Session.patchcKs|jd|f|S)zSends a DELETE request. Returns :class:`Response` object. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :rtype: requests.Response ZDELETErrr'r'r+delete_szSession.deletec Ksz|d|j|d|j|d|j|d|jt|trJtd|dd}| d}|j }|j |j d}t }|j|f|}t |} t| d |_td ||f|}|jr|jD]} t|j| j| jqt|j||j|j||f|} |rd d | Dng} | r*| d || }| |_|sjz"t|j||fddi||_Wntk rhYnX|sv|j|S)zISend a given PreparedRequest. :rtype: requests.Response rFrHrIrJz#You can only send PreparedRequests.rKT)rL)Zsecondsr3cSsg|]}|qSr'r')r(r=r'r'r+r,sz Session.send..rrb)rkrFrHrIrJr-r ValueErrorr[r4rw get_adapterrLpreferred_clockr`relapsedrrOr r]rfrRreinsertnextZ_next StopIterationrP) r<rfrrKrFrwadapterstartrrr=genrOr'r'r+r`isB       "z Session.sendc Cs|jrn|dk r|dnd}t||d}|D]\}} ||| q0|dksV|dkrntjdpltjd}t||j}t||j }t||j }t||j }||||dS)z^ Check the environment and merge it with some settings. :rtype: dict NrhriTZREQUESTS_CA_BUNDLEZCURL_CA_BUNDLE)rHrJrFrI) rgr4rr/rkosenvironr2rJrFrHrI) r<rLrJrFrHrIrhZ env_proxiesr)r*r'r'r+rs        z"Session.merge_environment_settingscCs>|jD]"\}}||r |Sq td|dS)z~ Returns the appropriate connection adapter for the given URL. :rtype: requests.adapters.BaseAdapter z*No connection adapters were found for '%s'N)ryr/lowerrWr)r<rLprefixrr'r'r+rs zSession.get_adaptercCs|jD] }|q dS)z+Closes all adapters and as such the sessionN)ryvaluesrV)r<r*r'r'r+rVsz Session.closecs>||j<fdd|jD}|D]}|j||j|<q"dS)zwRegisters a connection adapter to a prefix. Adapters are sorted in descending order by prefix length. cs g|]}t|tkr|qSr')rT)r(r)rr'r+r,sz!Session.mount..N)ryr[)r<rrZ keys_to_mover1r'rr+rzs z Session.mountcsfddjD}|S)Ncsi|]}|t|dqSr&)getattr)r(attrr{r'r+ sz(Session.__getstate__..) __attrs__)r<stater'r{r+ __getstate__szSession.__getstate__cCs"|D]\}}t|||qdSr&)r/setattr)r<rrvaluer'r'r+ __setstate__szSession.__setstate__)NNNNNNNTNNNNNN)NN)N)N)rrrsrt__doc__rr|r}rrrfr4rrrrrrr`rrrVrzrrr'r'r'r+ruTs`7) G    I rucCstS)aZ Returns a :class:`Session` for context-management. .. deprecated:: 1.0.0 This method has been deprecated since version 1.0.0 and is only kept for backwards compatibility. New code should use :class:`~requests.sessions.Session` to create a session. This may be removed at a future date. :rtype: Session )rur'r'r'r+sessions r)>rrsystimeZdatetimerrvrcompatrrrrr r r]r r r rZmodelsrrrrwrrZ_internal_utilsrZutilsrrr exceptionsrrrrZ structuresrryrrrr r!r"r#Z status_codesr$r%platform perf_counterrAttributeErrorZclockr2r5objectr6rurr'r'r'r+s@             v%site-packages/pip/_vendor/requests/__pycache__/structures.cpython-38.opt-1.pyc000064400000010452151464452720023421 0ustar00U .e @s<dZddlmZmZmZGdddeZGdddeZdS)zO requests.structures ~~~~~~~~~~~~~~~~~~~ Data structures that power Requests. ) OrderedDictMappingMutableMappingc@sbeZdZdZdddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ dS)CaseInsensitiveDictaA case-insensitive ``dict``-like object. Implements all methods and operations of ``MutableMapping`` as well as dict's ``copy``. Also provides ``lower_items``. All keys are expected to be strings. The structure remembers the case of the last key to be set, and ``iter(instance)``, ``keys()``, ``items()``, ``iterkeys()``, and ``iteritems()`` will contain case-sensitive keys. However, querying and contains testing is case insensitive:: cid = CaseInsensitiveDict() cid['Accept'] = 'application/json' cid['aCCEPT'] == 'application/json' # True list(cid) == ['Accept'] # True For example, ``headers['content-encoding']`` will return the value of a ``'Content-Encoding'`` response header, regardless of how the header name was originally stored. If the constructor, ``.update``, or equality comparison operations are given keys that have equal ``.lower()``s, the behavior is undefined. NcKs&t|_|dkri}|j|f|dSN)r_storeupdate)selfdatakwargsr C/usr/lib/python3.8/site-packages/pip/_vendor/requests/structures.py__init__(szCaseInsensitiveDict.__init__cCs||f|j|<dSrrlower)r keyvaluer r r __setitem__.szCaseInsensitiveDict.__setitem__cCs|j|dS)Nrrr rr r r __getitem__3szCaseInsensitiveDict.__getitem__cCs|j|=dSrrrr r r __delitem__6szCaseInsensitiveDict.__delitem__cCsdd|jDS)Ncss|]\}}|VqdSrr ).0ZcasedkeyZ mappedvaluer r r :sz/CaseInsensitiveDict.__iter__..)rvaluesr r r r __iter__9szCaseInsensitiveDict.__iter__cCs t|jSr)lenrrr r r __len__<szCaseInsensitiveDict.__len__cCsdd|jDS)z.Like iteritems(), but with all lowercase keys.css|]\}}||dfVqdS)rNr )rZlowerkeyZkeyvalr r r rAsz2CaseInsensitiveDict.lower_items..)ritemsrr r r lower_items?szCaseInsensitiveDict.lower_itemscCs0t|trt|}ntSt|t|kSr) isinstancerrNotImplementeddictr)r otherr r r __eq__Gs  zCaseInsensitiveDict.__eq__cCst|jSr)rrrrr r r copyPszCaseInsensitiveDict.copycCstt|Sr)strr"rrr r r __repr__SszCaseInsensitiveDict.__repr__)N)__name__ __module__ __qualname____doc__rrrrrrrr$r%r'r r r r r s  rcs<eZdZdZd fdd ZddZddZd d d ZZS) LookupDictzDictionary lookup object.Ncs||_tt|dSr)namesuperr,r)r r- __class__r r rZszLookupDict.__init__cCs d|jS)Nz )r-rr r r r'^szLookupDict.__repr__cCs|j|dSr__dict__getrr r r raszLookupDict.__getitem__cCs|j||Srr1)r rdefaultr r r r3fszLookupDict.get)N)N) r(r)r*r+rr'rr3 __classcell__r r r/r r,Ws r,N)r+compatrrrrr"r,r r r r sJsite-packages/pip/_vendor/requests/__pycache__/cookies.cpython-38.opt-1.pyc000064400000044556151464452720022646 0ustar00U .eG@sdZddlZddlZddlZddlmZddlmZmZm Z m Z m Z z ddl Z Wne k rlddlZ YnXGdddeZGdd d eZd d Zd d ZdddZGdddeZGdddeje ZddZddZddZd ddZddZdS)!z requests.cookies ~~~~~~~~~~~~~~~~ Compatibility code to be able to use `cookielib.CookieJar` with requests. requests.utils imports from here, so be careful with imports. N)to_native_string) cookieliburlparse urlunparseMorselMutableMappingc@seZdZdZddZddZddZdd Zd d Zd d Z ddZ dddZ ddZ ddZ ddZeddZeddZeddZdS) MockRequestaWraps a `requests.Request` to mimic a `urllib2.Request`. The code in `cookielib.CookieJar` expects this interface in order to correctly manage cookie policies, i.e., determine whether a cookie can be set, given the domains of the request and the cookie. The original request object is read-only. The client is responsible for collecting the new headers via `get_new_headers()` and interpreting them appropriately. You probably want `get_cookie_header`, defined below. cCs ||_i|_t|jjj|_dSN)_r _new_headersrurlschemetype)selfrequestr@/usr/lib/python3.8/site-packages/pip/_vendor/requests/cookies.py__init__%szMockRequest.__init__cCs|jSr )rrrrrget_type*szMockRequest.get_typecCst|jjjSr )rr r Znetlocrrrrget_host-szMockRequest.get_hostcCs|Sr rrrrrget_origin_req_host0szMockRequest.get_origin_req_hostcCsT|jjds|jjSt|jjddd}t|jj}t|j||j|j |j |j gS)NZHostzutf-8)encoding) r headersgetr rrrrpathZparamsZqueryZfragment)rhostZparsedrrr get_full_url3s zMockRequest.get_full_urlcCsdSNTrrrrris_unverifiableAszMockRequest.is_unverifiablecCs||jjkp||jkSr )r rr rnamerrr has_headerDszMockRequest.has_headerNcCs|jj||j||Sr )r rrr )rr#defaultrrr get_headerGszMockRequest.get_headercCs tddS)zMcookielib has no legitimate use for this method; add it back if you find one.z=Cookie headers should be added with add_unredirected_header()N)NotImplementedError)rkeyvalrrr add_headerJszMockRequest.add_headercCs||j|<dSr r rr#valuerrradd_unredirected_headerNsz#MockRequest.add_unredirected_headercCs|jSr r+rrrrget_new_headersQszMockRequest.get_new_headerscCs|Sr )r!rrrr unverifiableTszMockRequest.unverifiablecCs|Sr )rrrrrorigin_req_hostXszMockRequest.origin_req_hostcCs|Sr rrrrrr\szMockRequest.host)N)__name__ __module__ __qualname____doc__rrrrrr!r$r&r*r.r/propertyr0r1rrrrrr s$    r c@s(eZdZdZddZddZddZdS) MockResponsezWraps a `httplib.HTTPMessage` to mimic a `urllib.addinfourl`. ...what? Basically, expose the parsed HTTP headers from the server response the way `cookielib` expects to see them. cCs ||_dS)zMake a MockResponse for `cookielib` to read. :param headers: a httplib.HTTPMessage or analogous carrying the headers N_headers)rrrrrrhszMockResponse.__init__cCs|jSr r8rrrrinfooszMockResponse.infocCs|j|dSr )r9 getheadersr"rrrr;rszMockResponse.getheadersN)r2r3r4r5rr:r;rrrrr7asr7cCs8t|dr|jsdSt|}t|jj}|||dS)zExtract the cookies from the response into a CookieJar. :param jar: cookielib.CookieJar (not necessarily a RequestsCookieJar) :param request: our own requests.Request object :param response: urllib3.HTTPResponse object _original_responseN)hasattrr<r r7msgZextract_cookies)jarrZresponseZreqresrrrextract_cookies_to_jarvs  rAcCs t|}|||dS)zj Produce an appropriate Cookie header string to be sent with `request`, or None. :rtype: str Cookie)r Zadd_cookie_headerr/r)r?rrrrrget_cookie_headers rDcCszg}|D]N}|j|krq|dk r,||jkr,q|dk r@||jkr@q||j|j|jfq|D]\}}}||||q\dS)zkUnsets a cookie by name, by default over all domains and paths. Wraps CookieJar.clear(), is O(n). N)r#domainrappendclear) cookiejarr#rErZ clearablescookierrrremove_cookie_by_names rJc@seZdZdZdS)CookieConflictErrorzThere are two cookies that meet the criteria specified in the cookie jar. Use .get and .set and include domain and path args in order to be more specific. N)r2r3r4r5rrrrrKsrKcseZdZdZd3ddZddZddZd d Zd d Zd dZ ddZ ddZ ddZ ddZ ddZd4ddZfddZddZdd Zd!d"Zfd#d$Zfd%d&Zd5d'd(Zd6d)d*Zd+d,Zd-d.Zd/d0Zd1d2ZZS)7RequestsCookieJaraCompatibility class; is a cookielib.CookieJar, but exposes a dict interface. This is the CookieJar we create by default for requests and sessions that don't specify one, since some clients may expect response.cookies and session.cookies to support dict operations. Requests does not use the dict interface internally; it's just for compatibility with external client code. All requests code should work out of the box with externally provided instances of ``CookieJar``, e.g. ``LWPCookieJar`` and ``FileCookieJar``. Unlike a regular CookieJar, this class is pickleable. .. warning:: dictionary operations that are normally O(1) may be O(n). NcCs.z||||WStk r(|YSXdS)zDict-like get() that also supports optional domain and path args in order to resolve naming collisions from using one cookie jar over multiple domains. .. warning:: operation is O(n), not O(1). N)_find_no_duplicatesKeyError)rr#r%rErrrrrszRequestsCookieJar.getcKsX|dkr(t|||d|dddSt|tr|j|kr |dk rRtd||j}q |rb|Std|||fdS)aBoth ``__get_item__`` and ``get`` call this function: it's never used elsewhere in Requests. :param name: a string containing name of cookie :param domain: (optional) string containing domain of cookie :param path: (optional) string containing path of cookie :raises KeyError: if cookie is not found :raises CookieConflictError: if there are multiple cookies that match name and optionally domain and path :return: cookie.value Nz(There are multiple cookies with name, %rrw)rVr#rErrKr-rN)rr#rErZtoReturnrIrrrrMxs    z%RequestsCookieJar._find_no_duplicatescCs|j}|d|S)4Unlike a normal CookieJar, this class is pickleable. _cookies_lock)__dict__rtpoprstaterrr __getstate__s  zRequestsCookieJar.__getstate__cCs$|j|d|jkr t|_dS)ryrzN)r{ru threadingRLockrzr}rrr __setstate__s  zRequestsCookieJar.__setstate__cCs"t}|||||S)z(Return a copy of this RequestsCookieJar.)rLZ set_policy get_policyru)rZnew_cjrrrrts zRequestsCookieJar.copycCs|jS)z&Return the CookiePolicy instance used.)Z_policyrrrrrszRequestsCookieJar.get_policy)NNN)NN)NN)NN)r2r3r4r5rrUrXrZr[r\r]r^rarcrdrergrjrkrlrRrurxrMrrrtr __classcell__rrrhrrLs2            rLcCsN|dkr dSt|dr|St|}||D]}|t|q4|S)Nrt)r=rtrGrR)r?Znew_jarrIrrr_copy_cookie_jars  rcKsd||ddddddddddidd }t|t|}|rLd }t|t|||t|d |d <t|d |d <|d d|d<t|d|d<tjf|S)zMake a cookie from underspecified parameters. By default, the pair of `name` and `value` will be set for the domain '' and sent on every request (this is sometimes called a "supercookie"). rNro/FTHttpOnly) versionr#r-portrErsecureexpiresdiscardcomment comment_urlrestrfc2109z4create_cookie() got unexpected keyword arguments: %srZport_specifiedrEZdomain_specified.Zdomain_initial_dotrZpath_specified)rU TypeErrorrYruboolrmrrB)r#r-rSresultZbadargserrrrrrQs0 rQcCsd}|drPzttt|d}Wqrtk rLtd|dYqrXn"|drrd}tt|d|}t|dt|dd|d||j |d dd |d idt|d |j |d pdd S)zBConvert a Morsel object into a Cookie containing the one k/v pair.Nzmax-agezmax-age: %s must be integerrz%a, %d-%b-%Y %H:%M:%S GMTrFrErrZhttponlyrrr) rrrrErr#rrrrrr-r) inttime ValueErrorrcalendarZtimegmstrptimerQrr(r-)ZmorselrZ time_templaterrrrPs4    rPTcCsR|dkrt}|dk rNdd|D}|D]$}|s8||kr(|t|||q(|S)aCReturns a CookieJar from a key/value dictionary. :param cookie_dict: Dict of key/values to insert into CookieJar. :param cookiejar: (optional) A cookiejar to add the cookies to. :param overwrite: (optional) If False, will not replace cookies already in the jar with new ones. :rtype: CookieJar NcSsg|] }|jqSr)r#).0rIrrr sz'cookiejar_from_dict..)rLrRrQ)Z cookie_dictrH overwriteZnames_from_jarr#rrrcookiejar_from_dicts  rcCsvt|tjstdt|tr.t||dd}nDt|tjrrz||Wn(tk rp|D]}||q\YnX|S)zAdd cookies to cookiejar and returns a merged CookieJar. :param cookiejar: CookieJar object to add the cookies to. :param cookies: Dictionary or CookieJar object to be added. :rtype: CookieJar z!You can only merge into CookieJarF)rHr) rOrrsrdictrruAttributeErrorrR)rHZcookiesZ cookie_in_jarrrr merge_cookiess   r)NN)NT)r5rtrrZ_internal_utilsrcompatrrrrrr ImportErrorZdummy_threadingobjectr r7rArDrJ RuntimeErrorrKrsrLrrQrPrrrrrrs.   H $ site-packages/pip/_vendor/requests/__pycache__/api.cpython-38.pyc000064400000014507151464452720021015 0ustar00U .e@s\dZddlmZddZdddZdd Zd d Zdd d ZdddZdddZ ddZ dS)z requests.api ~~~~~~~~~~~~ This module implements the Requests API. :copyright: (c) 2012 by Kenneth Reitz. :license: Apache2, see LICENSE for more details. )sessionsc Ks6t$}|jf||d|W5QRSQRXdS)a Constructs and sends a :class:`Request `. :param method: method for the new :class:`Request` object. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary, list of tuples or bytes to send in the query string for the :class:`Request`. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) A JSON serializable Python object to send in the body of the :class:`Request`. :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`. :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`. :param files: (optional) Dictionary of ``'name': file-like-objects`` (or ``{'name': file-tuple}``) for multipart encoding upload. ``file-tuple`` can be a 2-tuple ``('filename', fileobj)``, 3-tuple ``('filename', fileobj, 'content_type')`` or a 4-tuple ``('filename', fileobj, 'content_type', custom_headers)``, where ``'content-type'`` is a string defining the content type of the given file and ``custom_headers`` a dict-like object containing additional headers to add for the file. :param auth: (optional) Auth tuple to enable Basic/Digest/Custom HTTP Auth. :param timeout: (optional) How many seconds to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param allow_redirects: (optional) Boolean. Enable/disable GET/OPTIONS/POST/PUT/PATCH/DELETE/HEAD redirection. Defaults to ``True``. :type allow_redirects: bool :param proxies: (optional) Dictionary mapping protocol to the URL of the proxy. :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use. Defaults to ``True``. :param stream: (optional) if ``False``, the response content will be immediately downloaded. :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair. :return: :class:`Response ` object :rtype: requests.Response Usage:: >>> import requests >>> req = requests.request('GET', 'https://httpbin.org/get') )methodurlN)rZSessionrequest)rrkwargsZsessionr` object :rtype: requests.Response allow_redirectsTgetparams setdefaultr)rr rrrrr ?s r cKs|ddtd|f|S)zSends an OPTIONS request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response r Toptionsr rrrrrrNs rcKs|ddtd|f|S)zSends a HEAD request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response r Fheadr rrrrr[s rcKstd|f||d|S)aSends a POST request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response post)datajsonr)rrrrrrrrhs rcKstd|fd|i|S)aSends a PUT request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response putrrrrrrrrrws rcKstd|fd|i|S)aSends a PATCH request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response patchrrrrrrrs rcKstd|f|S)zSends a DELETE request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response deleterrrrrrs r)N)NN)N)N) __doc__rrr rrrrrrrrrrs /     site-packages/pip/_vendor/requests/__pycache__/_internal_utils.cpython-38.pyc000064400000002404151464452720023430 0ustar00U .eH@s.dZddlmZmZmZd ddZddZdS) z requests._internal_utils ~~~~~~~~~~~~~~ Provides utility functions that are consumed internally by Requests which depend on extremely few external helpers (such as compat) )is_py2 builtin_strstrasciicCs.t|tr|}ntr ||}n ||}|S)zGiven a string object, regardless of type, returns a representation of that string in the native string type, encoding and decoding where necessary. This assumes ASCII unless told otherwise. ) isinstancerrencodedecode)stringencodingoutr H/usr/lib/python3.8/site-packages/pip/_vendor/requests/_internal_utils.pyto_native_strings    rcCs:t|tstz|dWdStk r4YdSXdS)zDetermine if unicode string only contains ASCII characters. :param str u_string: unicode string to check. Must be unicode and not Python 2 `str`. :rtype: bool rTFN)rrAssertionErrorrUnicodeEncodeError)Zu_stringr r r unicode_is_asciis  rN)r)__doc__compatrrrrrr r r r s site-packages/pip/_vendor/requests/__pycache__/auth.cpython-38.pyc000064400000020145151464452720021200 0ustar00U .e'@sdZddlZddlZddlZddlZddlZddlZddlmZddl m Z m Z m Z ddl mZddlmZddlmZd Zd Zd d ZGd ddeZGdddeZGdddeZGdddeZdS)z] requests.auth ~~~~~~~~~~~~~ This module contains the authentication handlers for Requests. N) b64encode)urlparsestr basestring)extract_cookies_to_jar)to_native_string)parse_dict_headerz!application/x-www-form-urlencodedzmultipart/form-datacCst|ts&tjd|tdt|}t|tsLtjd|tdt|}t|tr`|d}t|trt|d}dtt d ||f }|S)zReturns a Basic Auth string.zNon-string usernames will no longer be supported in Requests 3.0.0. Please convert the object you've passed in ({!r}) to a string or bytes object in the near future to avoid problems.)categoryzNon-string passwords will no longer be supported in Requests 3.0.0. Please convert the object you've passed in ({!r}) to a string or bytes object in the near future to avoid problems.latin1zBasic :) isinstancerwarningswarnformatDeprecationWarningrencoderrjoinstrip)usernamepasswordZauthstrr=/usr/lib/python3.8/site-packages/pip/_vendor/requests/auth.py_basic_auth_strs0      rc@seZdZdZddZdS)AuthBasez4Base class that all auth implementations derive fromcCs tddS)NzAuth hooks must be callable.)NotImplementedErrorselfrrrr__call__KszAuthBase.__call__N__name__ __module__ __qualname____doc__rrrrrrHsrc@s0eZdZdZddZddZddZdd Zd S) HTTPBasicAuthz?Attaches HTTP Basic Authentication to the given Request object.cCs||_||_dSN)rrrrrrrr__init__RszHTTPBasicAuth.__init__cCs(t|jt|ddk|jt|ddkgSNrrallrgetattrrrotherrrr__eq__VszHTTPBasicAuth.__eq__cCs ||k Sr&rr-rrr__ne__\szHTTPBasicAuth.__ne__cCst|j|j|jd<|S)N Authorizationrrrheadersrrrrr_szHTTPBasicAuth.__call__N)r!r"r#r$r(r/r0rrrrrr%Os r%c@seZdZdZddZdS) HTTPProxyAuthz=Attaches HTTP Proxy Authentication to a given Request object.cCst|j|j|jd<|S)NzProxy-Authorizationr2rrrrrgszHTTPProxyAuth.__call__Nr rrrrr4dsr4c@sPeZdZdZddZddZddZdd Zd d Zd d Z ddZ ddZ dS)HTTPDigestAuthz@Attaches HTTP Digest Authentication to the given Request object.cCs||_||_t|_dSr&)rr threadingZlocal _thread_localr'rrrr(oszHTTPDigestAuth.__init__cCs@t|jds.md5_utf8ZSHAcSs"t|tr|d}t|SrF)r rrrHsha1rIrJrrrsha_utf8s  z4HTTPDigestAuth.build_digest_header..sha_utf8zSHA-256cSs"t|tr|d}t|SrF)r rrrHZsha256rIrJrrr sha256_utf8s  z7HTTPDigestAuth.build_digest_header..sha256_utf8zSHA-512cSs"t|tr|d}t|SrF)r rrrHZsha512rIrJrrr sha512_utf8s  z7HTTPDigestAuth.build_digest_header..sha512_utf8csd||fS)N%s:%sr)sdZ hash_utf8rrz4HTTPDigestAuth.build_digest_header../?z%s:%s:%srQrz%08xrGZauth,z%s:%s:%s:%s:%sz>username="%s", realm="%s", nonce="%s", uri="%s", response="%s"z , opaque="%s"z, algorithm="%s"z , digest="%s"z , qop="auth", nc=%s, cnonce="%s"z Digest %s)r7r=getupperrpathZqueryrrr;r<rrtimectimeosurandomrHrMrIsplit)rmethodurlrArBrCrDrEZ _algorithmrLrNrOrPZKDZentdigZp_parsedr^ZA1ZA2ZHA1ZHA2ZncvaluerRZcnonceZrespdigZnoncebitbaserrTrbuild_digest_headers           z"HTTPDigestAuth.build_digest_headercKs|jrd|j_dS)z)Reset num_401_calls counter on redirects.rN)Z is_redirectr7r?)rrkwargsrrrhandle_redirectszHTTPDigestAuth.handle_redirectcKs"d|jkrdks$nd|j_|S|jjdk rB|jj|jj|jdd}d| kr|jjdkr|jjd7_t j d t j d }t |jd|dd |j_|j||j}t|j|j|j||j||j|j|jd <|jj|f|}|j|||_|Sd|j_|S) zo Takes the given response and tries digest-auth, if needed. :rtype: requests.Response iirNzwww-authenticater9Zdigestzdigest )flags)countr1)Z status_coder7r?r>Zrequestbodyseekr3r\lowerrecompile IGNORECASEr subr=ZcontentclosecopyrZ_cookiesrawZprepare_cookiesrgrdreZ connectionsendhistoryappend)rrrhZs_authZpatZprepZ_rrrr handle_401s2      zHTTPDigestAuth.handle_401cCs~||jjr&||j|j|jd<z|j|j_ Wnt k rTd|j_ YnX| d|j | d|j d|j_|S)Nr1Zresponser)r@r7r;rgrdrer3rmtellr>AttributeErrorZ register_hookrzrir?rrrrrszHTTPDigestAuth.__call__cCs(t|jt|ddk|jt|ddkgSr)r*r-rrrr/*szHTTPDigestAuth.__eq__cCs ||k Sr&rr-rrrr00szHTTPDigestAuth.__ne__N) r!r"r#r$r(r@rgrirzrr/r0rrrrr5ls f,r5)r$rarpr_rHr6rbase64rcompatrrrZcookiesrZ_internal_utilsrZutilsr ZCONTENT_TYPE_FORM_URLENCODEDZCONTENT_TYPE_MULTI_PARTrobjectrr%r4r5rrrrs$    ,site-packages/pip/_vendor/requests/__pycache__/utils.cpython-38.pyc000064400000053223151464452720021402 0ustar00U .eau@sdZddlZddlZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl Z ddl m Z ddl mZddlmZddlmZddlmZmZmZmZmZmZmZmZmZmZmZmZm Z m!Z!m"Z"dd l#m$Z$dd l%m&Z&dd l'm(Z(m)Z)m*Z*m+Z+d Z,e-Z.d ddZ/ej0dkr2ddZ1ddZddZ2ddZ3dhddZ4ddZ5ddZ6d d!Z7d"d#Z8d$d%Z9d&d'Z:did(d)Z;d*d+Zd0d1Z?d2d3Z@d4d5ZAd6d7ZBd8d9ZCeDd:ZEd;d<ZFd=d>ZGd?d@ZHdAdBZIdCdDZJdEdFZKejLdGdHZMdIdJZNdjdKdLZOdMdNZPdkdPdQZQdRdSZRdTdUZSdVTdWZUeUdXZVeUdYZWdZd[ZXd\d]ZYd^d_ZZe[d`Z\e[daZ]dbdcZ^dddeZ_dfdgZ`dS)lz requests.utils ~~~~~~~~~~~~~~ This module provides utility functions that are used within Requests that are also useful for external consumption. N __version__)certs)to_native_string)parse_http_list)quoteurlparsebytesstr OrderedDictunquote getproxies proxy_bypass urlunparse basestring integer_typesis_py3proxy_bypass_environmentgetproxies_environmentMappingcookiejar_from_dict)CaseInsensitiveDict) InvalidURL InvalidHeaderFileModeWarningUnrewindableBodyError)z.netrc_netrcPi)ZhttpZhttpsZwin32cCsztrddl}nddl}Wntk r0YdSXz6||jd}t||dd}||dd}Wntk r~YdSX|r|sdS| d}|D]T}|dkrd|krd S| dd }| d d }| d d}t ||t j rd SqdS)NrFz;Software\Microsoft\Windows\CurrentVersion\Internet SettingsZ ProxyEnableZ ProxyOverride;z.Tz\.*z.*?)rwinreg_winreg ImportErrorOpenKeyHKEY_CURRENT_USERintZ QueryValueExOSErrorsplitreplacerematchI)hostr$ZinternetSettingsZ proxyEnableZ proxyOverrideZtestr1>/usr/lib/python3.8/site-packages/pip/_vendor/requests/utils.pyproxy_bypass_registry/sF      r3cCstrt|St|SdS)zReturn True, if the host should be bypassed. Checks proxy settings gathered from the environment, if specified, or the registry. N)rrr3)r0r1r1r2rVsrcCst|dr|}|S)z/Returns an internal sequence dictionary update.items)hasattrr4)dr1r1r2dict_to_sequencebs r7c Cs4d}d}t|drt|}nbt|dr.|j}nPt|dr~z |}Wntjk rZYn$Xt|j}d|jkr~t dt t|drz | }Wn$t tfk r|dk r|}Yn^Xt|dr|dkrz&|dd | }||pdWnt tfk rd}YnX|dkr&d}td||S) Nr__len__lenfilenoba%Requests has determined the content-length for this request using the binary size of the file: however, the file has been opened in text mode (i.e. without the 'b' flag in the mode). This may lead to an incorrect content-length. In Requests 3.0, support will be removed for files in text mode.tellseek)r5r9r:ioUnsupportedOperationosfstatst_sizemodewarningswarnrr<r*IOErrorr=max)oZ total_lengthZcurrent_positionr:r1r1r2 super_lenksB             rJFc Cs&zddlm}m}d}tD]J}ztjd|}Wntk rPYWdSXtj|r|}qhq|dkrvWdSt |}d}t |t r| d}|j |d} z:||| } | r| drdnd} | | | dfWWSWn"|tfk r|rYnXWnttfk r YnXdS) z;Returns the Requests tuple auth for a given url from netrc.r)netrcNetrcParseErrorNz~/{}:asciirr>)rKrL NETRC_FILESrApath expanduserformatKeyErrorexistsr isinstancer decodenetlocr+ZauthenticatorsrGr&AttributeError) urlZ raise_errorsrKrLZ netrc_pathfZlocZriZsplitstrr0rZlogin_ir1r1r2get_netrc_auths8     r[cCsBt|dd}|r>t|tr>|ddkr>|ddkr>tj|SdS)z0Tries to guess the filename of the given object.nameNr<>)getattrrUrrArPbasename)objr\r1r1r2guess_filenames   rccCstj|r|Stj|\}}|rPtj|sPtj|\}}d||g}q t|s^|St|}||krx|St }tjj|f|d}tj|s|j ||d}|S)zReplace nonexistent paths that look like they refer to a member of a zip archive with the location of an extracted copy of the target, or else just return the provided path unchanged. /)rP) rArPrTr+joinzipfileZ is_zipfileZZipFileZnamelisttempfileZ gettempdirextract)rParchivememberprefixZzip_fileZtmpZextracted_pathr1r1r2extract_zipped_pathss      rlcCs.|dkr dSt|ttttfr&tdt|S)aTake an object and test to see if it can be represented as a dictionary. Unless it can not be represented as such, return an OrderedDict, e.g., :: >>> from_key_val_list([('key', 'val')]) OrderedDict([('key', 'val')]) >>> from_key_val_list('string') ValueError: cannot encode objects that are not 2-tuples >>> from_key_val_list({'key': 'val'}) OrderedDict([('key', 'val')]) :rtype: OrderedDict N+cannot encode objects that are not 2-tuples)rUr r boolr) ValueErrorr valuer1r1r2from_key_val_lists rrcCs@|dkr dSt|ttttfr&tdt|tr8|}t|S)aTake an object and test to see if it can be represented as a dictionary. If it can be, return a list of tuples, e.g., :: >>> to_key_val_list([('key', 'val')]) [('key', 'val')] >>> to_key_val_list({'key': 'val'}) [('key', 'val')] >>> to_key_val_list('string') ValueError: cannot encode objects that are not 2-tuples. :rtype: list Nrm) rUr r rnr)rorr4listrpr1r1r2to_key_val_lists rtcCsXg}t|D]F}|dd|ddkr4dkrHnnt|dd}||q |S)aParse lists as described by RFC 2068 Section 2. In particular, parse comma-separated lists where the elements of the list may include quoted-strings. A quoted-string could contain a comma. A non-quoted string could have quotes in the middle. Quotes are removed automatically after parsing. It basically works like :func:`parse_set_header` just that items may appear multiple times and case sensitivity is preserved. The return value is a standard :class:`list`: >>> parse_list_header('token, "quoted value"') ['token', 'quoted value'] To create a header from the :class:`list` again, use the :func:`dump_header` function. :param value: a string with a list header. :return: :class:`list` :rtype: list Nrr^")_parse_list_headerunquote_header_valueappend)rqresultitemr1r1r2parse_list_header8s  ( r{cCsxi}t|D]f}d|kr"d||<q |dd\}}|dd|ddkrVdkrjnnt|dd}|||<q |S)a^Parse lists of key, value pairs as described by RFC 2068 Section 2 and convert them into a python dict: >>> d = parse_dict_header('foo="is a fish", bar="as well"') >>> type(d) is dict True >>> sorted(d.items()) [('bar', 'as well'), ('foo', 'is a fish')] If there is no value for a key it will be `None`: >>> parse_dict_header('key_without_value') {'key_without_value': None} To create a header from the :class:`dict` again, use the :func:`dump_header` function. :param value: a string with a dict header. :return: :class:`dict` :rtype: dict =Nrr^ru)rvr+rw)rqryrzr\r1r1r2parse_dict_headerXs ( r}cCs\|rX|d|dkr dkrXnn4|dd}|rD|dddkrX|ddd dS|S) zUnquotes a header value. (Reversal of :func:`quote_header_value`). This does not use the real unquoting but what browsers are actually using for quoting. :param value: the header value to unquote. :rtype: str rr^rurNr>z\\\z\")r,)rqZ is_filenamer1r1r2rw{s $ rwcCsi}|D]}|j||j<q|S)zReturns a key/value dictionary from a CookieJar. :param cj: CookieJar object to extract cookies from. :rtype: dict )rqr\)cj cookie_dictZcookier1r1r2dict_from_cookiejarsrcCs t||S)zReturns a CookieJar from a key/value dictionary. :param cj: CookieJar to insert cookies into. :param cookie_dict: Dict of key/values to insert into CookieJar. :rtype: CookieJar r)rrr1r1r2add_dict_to_cookiejarsrcCsTtdttjdtjd}tjdtjd}td}||||||S)zlReturns encodings from given content string. :param content: bytestring to extract encodings from. zIn requests 3.0, get_encodings_from_content will be removed. For more information, please see the discussion on issue #2266. (This warning should only appear once.)z!])flagsz+]z$^<\?xml.*?encoding=["\']*(.+?)["\'>])rErFDeprecationWarningr-compiler/findall)contentZ charset_reZ pragma_reZxml_rer1r1r2get_encodings_from_contents rc Cs|d}|d|dd}}i}d}|D]`}|}|r0|d}}|d} | dkr|d| |}|| dd|}|||<q0||fS) zReturns content type and parameters from given header :param header: string :return: tuple containing content type and dictionary of parameters r rrNz"' Tr|r^)r+stripfindlower) headertokens content_typeparamsZ params_dictZitems_to_stripparamkeyrqZindex_of_equalsr1r1r2_parse_content_type_headers   rcCsD|d}|sdSt|\}}d|kr4|ddSd|kr@dSdS)z}Returns encodings from given HTTP Header Dict. :param headers: dictionary to extract encoding from. :rtype: str z content-typeNcharsetz'"textz ISO-8859-1)getrr)headersrrr1r1r2get_encoding_from_headerss  rccsj|jdkr|D] }|VqdSt|jdd}|D]}||}|r4|Vq4|jddd}|rf|VdS)zStream decodes a iterator.Nr,errorsT)final)encodingcodecsgetincrementaldecoderrV)iteratorrrzdecoderchunkrvr1r1r2stream_decode_response_unicodes  rccsHd}|dks|dkrt|}|t|krD||||V||7}qdS)z Iterate over slices of a string.rN)r9)stringZ slice_lengthposr1r1r2 iter_slicess  rcCs~tdtg}t|j}|rLzt|j|WStk rJ||YnXzt|j|ddWSt k rx|jYSXdS)zReturns the requested content back in unicode. :param r: Response object to get unicode content from. Tried: 1. charset from content-type 2. fall back and replace all unicode characters :rtype: str zIn requests 3.0, get_unicode_from_response will be removed. For more information, please see the discussion on issue #2266. (This warning should only appear once.)r,rN) rErFrrrr r UnicodeErrorrx TypeError)rZtried_encodingsrr1r1r2get_unicode_from_response s  rzBABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-._~c Cs|d}tdt|D]}||dd}t|dkr|rztt|d}Wn tk rrtd|YnX|tkr|||dd||<qd||||<qd||||<qd |S) zUn-escape any percent-escape sequences in a URI that are unreserved characters. This leaves all reserved, illegal and non-ASCII bytes encoded. :rtype: str %rrr>z%Invalid percent-escape sequence: '%s'N) r+ranger9isalnumchrr)rorUNRESERVED_SETre)uripartsihcr1r1r2unquote_unreserved5s rcCs@d}d}ztt||dWStk r:t||dYSXdS)zRe-quote the given URI. This function passes the given URI through an unquote/quote cycle to ensure that it is fully and consistently quoted. :rtype: str z!#$%&'()*+,/:;=?@[]~z!#$&'()*+,/:;=?@[]~)ZsafeN)rrr)rZsafe_with_percentZsafe_without_percentr1r1r2 requote_uriMs rcCsltdt|d}|d\}}tdttt|d}tdt|d|@}||@||@kS)zThis function allows you to check if an IP belongs to a network subnet Example: returns True if ip = 192.168.1.1 and net = 192.168.1.0/24 returns False if ip = 192.168.1.1 and net = 192.168.100.0/24 :rtype: bool z=Lrrd)structunpacksocket inet_atonr+dotted_netmaskr))ZipZnetZipaddrZnetaddrbitsZnetmaskZnetworkr1r1r2address_in_networkcs rcCs&ddd|>dA}ttd|S)zConverts mask from /xx format to xxx.xxx.xxx.xxx Example: if mask is 24 function returns 255.255.255.0 :rtype: str lr z>I)rZ inet_ntoarpack)maskrr1r1r2rrsrcCs,zt|Wntjk r&YdSXdS)z :rtype: bool FT)rrerror)Z string_ipr1r1r2is_ipv4_address}s rcCs|ddkrzt|dd}Wntk r:YdSX|dksL|dkrPdSzt|ddWqtjk rYdSXndSdS)zV Very simple check of the cidr format in no_proxy variable. :rtype: bool rdrFrrT)countr)r+rorrr)Zstring_networkrr1r1r2 is_valid_cidrs rc csT|dk }|r"tj|}|tj|<z dVW5|rN|dkrDtj|=n |tj|<XdS)zSet the environment variable 'env_name' to 'value' Save previous value, yield, and then restore the previous value stored in the environment variable 'env_name'. If 'value' is None, do nothingN)rAenvironr)Zenv_namerqZ value_changed old_valuer1r1r2 set_environs    rc Cs*dd}|}|dkr|d}t|}|jdkr2dS|rdd|dd d D}t|jr|D]0}t|rt|j|rdSq`|j|kr`dSq`nB|j}|jr|d |j7}|D] }|j |s| |rdSqt d|6zt |j}Wn t t jfk rd }YnXW5QRX|r&dSd S) zL Returns whether we should bypass proxies or not. :rtype: bool cSstj|ptj|SN)rArrupper)kr1r1r2rz'should_bypass_proxies..Nno_proxyTcss|]}|r|VqdSrr1).0r0r1r1r2 sz(should_bypass_proxies.. r,z:{}F)r hostnamer,r+rrrZportrRendswithrrrrZgaierror) rYrZ get_proxyZ no_proxy_argparsedZproxy_ipZhost_with_portr0Zbypassr1r1r2should_bypass_proxiess@      rcCst||driStSdS)zA Return a dict of environment proxies. :rtype: dict )rN)rr)rYrr1r1r2get_environ_proxiess rcCst|pi}t|}|jdkr.||j|dS|jd|j|jd|jdg}d}|D]}||krV||}qpqV|S)zSelect a proxy for the url, if applicable. :param url: The url being for the request :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs Nallz://zall://)r rrscheme)rYZproxiesZurlpartsZ proxy_keysproxyZ proxy_keyr1r1r2 select_proxys rpython-requestscCs d|tfS)zO Return a string representing the default user agent. :rtype: str z%s/%sr)r\r1r1r2default_user_agentsrcCsttdddddS)z9 :rtype: requests.structures.CaseInsensitiveDict z, )ZgzipZdeflatez*/*z keep-alive)z User-AgentzAccept-EncodingZAcceptZ Connection)rrrer1r1r1r2default_headers"s rc Csg}d}||}|s|Std|D]}z|dd\}}Wntk r\|d}}YnXd|di}|dD]F}z|d\}}Wntk rYqYnX|||||<qv||q&|S) zReturn a list of parsed link headers proxies. i.e. Link: ; rel=front; type="image/jpeg",; rel=back;type="image/jpeg" :rtype: list z '"z, * '"r|)rr-r+rorx) rqZlinksZ replace_charsvalrYrlinkrrr1r1r2parse_header_links.s&   rrNr>cCs|dd}|tjtjfkr dS|ddtjkr6dS|ddtjtjfkrRdS|t}|dkrhd S|dkr|dddtkrd S|d ddtkrd S|dkr|ddt krd S|d dt krdSdS)z :rtype: str Nzutf-32rz utf-8-sigr>zutf-16rzutf-8z utf-16-berz utf-16-lez utf-32-bez utf-32-le) r BOM_UTF32_LE BOM_UTF32_BEBOM_UTF8 BOM_UTF16_LE BOM_UTF16_BEr_null_null2_null3)dataZsampleZ nullcountr1r1r2guess_json_utfYs*  rcCs8t||\}}}}}}|s$||}}t||||||fS)zGiven a URL that may or may not have a scheme, prepend the given scheme. Does not replace a present scheme with the one provided as an argument. :rtype: str )r r)rYZ new_schemerrWrPrqueryfragmentr1r1r2prepend_scheme_if_neededys rc CsBt|}zt|jt|jf}Wnttfk r<d}YnX|S)z{Given a url with authentication components, extract them into a tuple of username,password. :rtype: (str,str) )rr)r r ZusernameZpasswordrXr)rYrZauthr1r1r2get_auth_from_urls  rs^\S[^\r\n]*$|^$z^\S[^\r\n]*$|^$c Csf|\}}t|trt}nt}z||s4td|Wn*tk r`td||t|fYnXdS)zVerifies that header value is a string which doesn't contain leading whitespace or return characters. This prevents unintended header injection. :param header: tuple, in the format (name, value). z7Invalid return character or leading space in header: %sz>Value for header {%s: %s} must be of type str or bytes, not %sN)rUr _CLEAN_HEADER_REGEX_BYTE_CLEAN_HEADER_REGEX_STRr.rrtype)rr\rqZpatr1r1r2check_header_validitys   rcCsFt|\}}}}}}|s"||}}|ddd}t|||||dfS)zW Given a url remove the fragment and the authentication part. :rtype: str @rr^r)r rsplitr)rYrrWrPrrrr1r1r2 urldefragauths  rc Cs`t|jdd}|dk rTt|jtrTz||jWq\ttfk rPtdYq\XntddS)zfMove file pointer back to its recorded starting position so it can be read again on redirect. r=Nz;An error occurred when rewinding request body for redirect.z+Unable to rewind request body for redirect.)r`ZbodyrUZ_body_positionrrGr*r)Zprepared_requestZ body_seekr1r1r2 rewind_bodysr)F)F)N)r)a__doc__r contextlibr?rAr-rrsysrgrErfrrrZ_internal_utilsrcompatrrvrr r r r r rrrrrrrrrZcookiesrZ structuresr exceptionsrrrrrOwhereZDEFAULT_CA_BUNDLE_PATHZ DEFAULT_PORTSplatformr3r7rJr[rcrlrrrtr{r}rwrrrrrrrr frozensetrrrrrrrcontextmanagerrrrrrrrencoderrrrrrrrrrrrr1r1r1r2s    D    '  = 3  #   %   =   &    site-packages/pip/_vendor/requests/__pycache__/__init__.cpython-38.opt-1.pyc000064400000006420151464452720022735 0ustar00U .e @sdZddlmZddlmZddlZddlmZddZd d Zzeej ej Wn0e e fk re d ej ej eYnXdd lmZesz,dd lmZeddlm ZeeWnek rYnXddlmZededdl mZmZmZm Z ddl mZmZmZmZddl m Z m!Z!ddl"m#Z#ddl"m$Z$ddl%m&Z&m'Z'm(Z(ddl)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1ddl2m3Z3m4Z4ddl5m6Z6ddlm7Z7m8Z8m9Z9m:Z:m;Z;mZ>m?Z?ddl@Z@ddl@mAZAe@BeCDeAejde=dddS)a Requests HTTP Library ~~~~~~~~~~~~~~~~~~~~~ Requests is an HTTP library, written in Python, for human beings. Basic GET usage: >>> import requests >>> r = requests.get('https://www.python.org') >>> r.status_code 200 >>> 'Python is a programming language' in r.content True ... or POST: >>> payload = dict(key1='value1', key2='value2') >>> r = requests.post('https://httpbin.org/post', data=payload) >>> print(r.text) { ... "form": { "key2": "value2", "key1": "value1" }, ... } The other HTTP methods are supported - see `requests.api`. Full documentation is at . :copyright: (c) 2017 by Kenneth Reitz. :license: Apache 2.0, see LICENSE for more details. )urllib3)chardetN)RequestsDependencyWarningcCs~|d}t|dkr |d|\}}}t|t|t|}}}|ddd\}}}t|t|t|}}}dS)N.0)splitlenappendint)Zurllib3_versionZchardet_versionmajorminorpatchrA/usr/lib/python3.8/site-packages/pip/_vendor/requests/__init__.pycheck_compatibility1s    rcCsXzttt|d}Wntk r.YdSX|dddgkrTd|}t|tdS)Nrrr z4Old version of cryptography ({}) may cause slowdown.) listmapr r ValueErrorformatwarningswarnr)cryptography_versionZwarningrrr_check_cryptographyJs rz?urllib3 ({}) or chardet ({}) doesn't match a supported version!)WINDOWS) pyopenssl) __version__)DependencyWarningignore) __title____description____url__r) __build__ __author____author_email__ __license__) __copyright____cake__)utils)packages)RequestResponsePreparedRequest)requestgetheadpostrputdeleteoptions)sessionSession)codes) RequestExceptionTimeout URLRequiredTooManyRedirects HTTPErrorConnectionErrorFileModeWarningConnectTimeout ReadTimeout) NullHandlerdefaultT)r )E__doc__Z pip._vendorrrr exceptionsrrrrAssertionErrorrrrZpip._internal.utils.compatrZpip._vendor.urllib3.contribrZinject_into_urllib3Z cryptographyr ImportErrorZpip._vendor.urllib3.exceptionsr simplefilterr"r#r$r%r&r'r(r)r*r+r,Zmodelsr-r.r/Zapir0r1r2r3rr4r5r6Zsessionsr7r8Z status_codesr9r:r;r<r=r>r?r@rArBZloggingrCZ getLogger__name__Z addHandlerrrrrsP#             ( , site-packages/pip/_vendor/requests/__pycache__/exceptions.cpython-38.opt-1.pyc000064400000012130151464452720023352 0ustar00U .e} @sdZddlmZGdddeZGdddeZGdddeZGd d d eZGd d d eZGd ddeZ Gdddee Z Gddde Z GdddeZ GdddeZ GdddeeZGdddeeZGdddeeZGdddeeZGdd d eZGd!d"d"eZGd#d$d$eeZGd%d&d&eeZGd'd(d(eZGd)d*d*eZGd+d,d,eZGd-d.d.eeZGd/d0d0eZd1S)2z` requests.exceptions ~~~~~~~~~~~~~~~~~~~ This module contains the set of Requests' exceptions. ) HTTPErrorcs eZdZdZfddZZS)RequestExceptionzTThere was an ambiguous exception that occurred while handling your request. csX|dd}||_|dd|_|dk rB|jsBt|drB|jj|_tt|j||dS)zBInitialize RequestException with `request` and `response` objects.responseNrequest)poprrhasattrsuperr__init__)selfargskwargsr __class__C/usr/lib/python3.8/site-packages/pip/_vendor/requests/exceptions.pyr s  zRequestException.__init__)__name__ __module__ __qualname____doc__r __classcell__rrr rr src@seZdZdZdS)rzAn HTTP error occurred.Nrrrrrrrrrsrc@seZdZdZdS)ConnectionErrorzA Connection error occurred.Nrrrrrr src@seZdZdZdS) ProxyErrorzA proxy error occurred.Nrrrrrr$src@seZdZdZdS)SSLErrorzAn SSL error occurred.Nrrrrrr(src@seZdZdZdS)TimeoutzThe request timed out. Catching this error will catch both :exc:`~requests.exceptions.ConnectTimeout` and :exc:`~requests.exceptions.ReadTimeout` errors. Nrrrrrr,src@seZdZdZdS)ConnectTimeoutzThe request timed out while trying to connect to the remote server. Requests that produced this error are safe to retry. Nrrrrrr5src@seZdZdZdS) ReadTimeoutz@The server did not send any data in the allotted amount of time.Nrrrrrr<src@seZdZdZdS) URLRequiredz*A valid URL is required to make a request.Nrrrrrr@src@seZdZdZdS)TooManyRedirectszToo many redirects.NrrrrrrDsrc@seZdZdZdS) MissingSchemaz/The URL schema (e.g. http or https) is missing.NrrrrrrHsrc@seZdZdZdS) InvalidSchemaz"See defaults.py for valid schemas.Nrrrrrr Lsr c@seZdZdZdS) InvalidURLz%The URL provided was somehow invalid.Nrrrrrr!Psr!c@seZdZdZdS) InvalidHeaderz.The header value provided was somehow invalid.Nrrrrrr"Tsr"c@seZdZdZdS)InvalidProxyURLz"The proxy URL provided is invalid.Nrrrrrr#Xsr#c@seZdZdZdS)ChunkedEncodingErrorz?The server declared chunked encoding but sent an invalid chunk.Nrrrrrr$\sr$c@seZdZdZdS)ContentDecodingErrorz!Failed to decode response contentNrrrrrr%`sr%c@seZdZdZdS)StreamConsumedErrorz2The content for this response was already consumedNrrrrrr&dsr&c@seZdZdZdS) RetryErrorzCustom retries logic failedNrrrrrr'hsr'c@seZdZdZdS)UnrewindableBodyErrorz:Requests encountered an error when trying to rewind a bodyNrrrrrr(lsr(c@seZdZdZdS)RequestsWarningzBase warning for Requests.Nrrrrrr)rsr)c@seZdZdZdS)FileModeWarningzJA file was opened in text mode, but Requests determined its binary length.Nrrrrrr*wsr*c@seZdZdZdS)RequestsDependencyWarningz@An imported dependency doesn't match the expected version range.Nrrrrrr+|sr+N)rZpip._vendor.urllib3.exceptionsrZ BaseHTTPErrorIOErrorrrrrrrrrr ValueErrorrr r!r"r#r$r% TypeErrorr&r'r(Warningr)DeprecationWarningr*r+rrrrs0  site-packages/pip/_vendor/requests/__pycache__/__version__.cpython-38.opt-1.pyc000064400000001002151464452720023446 0ustar00U .e@s,dZdZdZdZdZdZdZdZdZd Z d S) ZrequestszPython HTTP for Humans.zhttp://python-requests.orgz2.22.0i"z Kenneth Reitzzme@kennethreitz.orgz Apache 2.0zCopyright 2019 Kenneth Reitzu ✨ 🍰 ✨N) Z __title__Z__description__Z__url__ __version__Z __build__ __author__Z__author_email__Z __license__Z __copyright__Z__cake__rrD/usr/lib/python3.8/site-packages/pip/_vendor/requests/__version__.pyssite-packages/pip/_vendor/requests/__pycache__/packages.cpython-38.opt-1.pyc000064400000000722151464452720022753 0ustar00U .e@svddlZdD]dZdeZeeee<eejD]>ZeeksJeedr0ee ddZ ejeejde <q0q dS)N)Zurllib3ZidnaZchardetz pip._vendor..zpip._vendor.requests.packages.) syspackageZvendored_package __import__localslistmodulesmod startswithlenZunprefixed_modr r A/usr/lib/python3.8/site-packages/pip/_vendor/requests/packages.pyssite-packages/pip/_vendor/requests/__pycache__/compat.cpython-38.opt-1.pyc000064400000003050151464452720022455 0ustar00U .e@sdZddlmZddlZejZeddkZeddkZddlZerddl m Z m Z m Z m Z mZmZmZmZmZddlmZmZmZmZmZddlmZddlZdd lmZdd lmZdd lmZm Z m!Z!m"Z"e#Z$e#Z%e&Z#e'Z'e(e)e*fZ+e(e)fZ,nerdd l-mZmZmZmZmZm Z m Z m Z m Z mZdd l.mZmZmZmZmZddl/m0Zdd l1mZdd l2mZddlm"Z"ddl3mZm Z m!Z!e#Z$e#Z#e%Z%e#e%fZ'e(e*fZ+e(fZ,dS)zq requests.compat ~~~~~~~~~~~~~~~ This module handles import compatibility issues between Python 2 and Python 3. )chardetN) quoteunquote quote_plus unquote_plus urlencode getproxies proxy_bypassproxy_bypass_environmentgetproxies_environment)urlparse urlunparseurljoinurlsplit urldefrag)parse_http_list)Morsel)StringIO)CallableMappingMutableMapping OrderedDict) rrrrr rrrrr)rr r r r ) cookiejar)r)rrr)4__doc__Z pip._vendorrsys version_infoZ_verZis_py2Zis_py3ZjsonZurllibrrrrr r r r r rrrrrZurllib2rZ cookielibZCookierr collectionsrrrrstrZ builtin_strbytesZunicodeZ basestringintZlongfloatZ numeric_typesZ integer_typesZ urllib.parseZurllib.requestZhttprZ http.cookiesioZcollections.abcr$r$?/usr/lib/python3.8/site-packages/pip/_vendor/requests/compat.pysD   ,     0    site-packages/pip/_vendor/requests/__pycache__/status_codes.cpython-38.opt-1.pyc000064400000010070151464452720023672 0ustar00U .e!E@sdaddlmZddddddd d d d d ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGDZedHdIZdJdKZedLS)Ma The ``codes`` object defines a mapping from common names for HTTP statuses to their numerical codes, accessible either as attributes or as dictionary items. >>> requests.codes['temporary_redirect'] 307 >>> requests.codes.teapot 418 >>> requests.codes['\o/'] 200 Some codes have multiple names, and both upper- and lower-case versions of the names are allowed. For example, ``codes.ok``, ``codes.OK``, and ``codes.okay`` all correspond to the HTTP status code 200. ) LookupDict)continue)Zswitching_protocols)Z processing)Z checkpoint)Z uri_too_longZrequest_uri_too_long)okZokayZall_okZall_okayZall_goodz\o/u✓)Zcreated)Zaccepted)Znon_authoritative_infoZnon_authoritative_information)Z no_content)Z reset_contentreset)Zpartial_contentpartial)Z multi_statusZmultiple_statusZ multi_statiZmultiple_stati)Zalready_reported)Zim_used)Zmultiple_choices)Zmoved_permanentlyZmovedz\o-)found)Z see_otherother)Z not_modified)Z use_proxy)Z switch_proxy)Ztemporary_redirectZtemporary_movedZ temporary)Zpermanent_redirectZresume_incompleteZresume)Z bad_requestZbad)Z unauthorized)Zpayment_requiredZpayment)Z forbidden)Z not_foundz-o-)Zmethod_not_allowedZ not_allowed)Znot_acceptable)Zproxy_authentication_requiredZ proxy_authZproxy_authentication)Zrequest_timeoutZtimeout)Zconflict)Zgone)Zlength_required)Zprecondition_failed precondition)Zrequest_entity_too_large)Zrequest_uri_too_large)Zunsupported_media_typeZunsupported_mediaZ media_type)Zrequested_range_not_satisfiableZrequested_rangeZrange_not_satisfiable)Zexpectation_failed)Z im_a_teapotZteapotZ i_am_a_teapot)Zmisdirected_request)Zunprocessable_entityZ unprocessable)locked)Zfailed_dependencyZ dependency)Zunordered_collectionZ unordered)Zupgrade_requiredZupgrade)Zprecondition_requiredr )Ztoo_many_requestsZtoo_many)Zheader_fields_too_largeZfields_too_large)Z no_responseZnone)Z retry_withZretry)Z$blocked_by_windows_parental_controlsZparental_controls)Zunavailable_for_legal_reasonsZ legal_reasons)Zclient_closed_request)Zinternal_server_errorZ server_errorz/o\u✗)Znot_implemented)Z bad_gateway)Zservice_unavailableZ unavailable)Zgateway_timeout)Zhttp_version_not_supportedZ http_version)Zvariant_also_negotiates)Zinsufficient_storage)Zbandwidth_limit_exceededZ bandwidth)Z not_extended)Znetwork_authentication_requiredZ network_authZnetwork_authentication)Ddefgzi,i-i.i/i0i1i2i3i4iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiZ status_codes)namecs~tD]8\}}|D]*}tt|||dstt||qqddtdk rvtddfddttDndadS)N)\/cSs$dddt|D}d||fS)Nz, css|]}d|VqdS)z``%s``N).0nrrE/usr/lib/python3.8/site-packages/pip/_vendor/requests/status_codes.py psz%_init..doc..z* %d: %s)join_codes)codenamesrrr docosz_init..doc c3s|]}|VqdS)Nr)rr$r&rr r!usz_init..) r#itemssetattrcodes startswithupper__doc__r"sorted)r$Ztitlestitlerr(r _iniths  r1N)r.Z structuresrr#r+r1rrrr s P site-packages/pip/_vendor/requests/__pycache__/help.cpython-38.opt-1.pyc000064400000005206151464452720022127 0ustar00U .e @sdZddlmZddlZddlZddlZddlZddlmZddlm Z ddlm Z ddl m Z zdd lmZWn ek rdZdZdZYnXddlZddlZd d Zd d ZddZedkredS)z'Module containing bug report helper(s).)print_functionN)idna)urllib3)chardet) __version__) pyopensslcCst}|dkrt}nj|dkr\dtjjtjjtjjf}tjjdkrd |tjjg}n(|dkrnt}n|dkrt}nd}||d S) aReturn a dict with the Python implementation and version. Provide both the name and the version of the Python implementation currently running. For example, on CPython 2.7.5 it will return {'name': 'CPython', 'version': '2.7.5'}. This function works best on CPython and PyPy: in particular, it probably doesn't work for Jython or IronPython. Future investigation should be done to work out the correct shape of the code for those platforms. ZCPythonZPyPyz%s.%s.%sfinalZJythonZ IronPythonUnknown)nameversion) platformZpython_implementationZpython_versionsysZpypy_version_infomajorminormicro releaseleveljoin)implementationZimplementation_versionr=/usr/lib/python3.8/site-packages/pip/_vendor/requests/help.py_implementations&     rc Cszttd}Wntk r4ddd}YnXt}dtji}dtji}ddd}trrtjdtj j d}dt t ddi}dt t ddi}tj }d|dk rd|ndi}|||tdk |||||dtid S) z&Generate information for a bug report.)systemreleaser r Nr )r Zopenssl_versionz%xr) rr system_sslZusing_pyopensslZ pyOpenSSLrr cryptographyrZrequests)rrrIOErrorrrrrOpenSSLZSSLZOPENSSL_VERSION_NUMBERgetattrrrsslrrequests_version) Z platform_infoZimplementation_infoZ urllib3_infoZ chardet_infoZpyopenssl_infoZcryptography_infoZ idna_inforZsystem_ssl_inforrrinfo;sR       r"cCsttjtddddS)z)Pretty-print the bug information as JSON.T)Z sort_keysindentN)printjsondumpsr"rrrrmainqsr(__main__)__doc__Z __future__rr&rrr Z pip._vendorrrrr rr!Zpip._vendor.urllib3.contribr ImportErrorrrrr"r(__name__rrrrs,      !6site-packages/pip/_vendor/requests/__pycache__/hooks.cpython-38.pyc000064400000001674151464452720021370 0ustar00U .e@sdZdgZddZddZdS)z requests.hooks ~~~~~~~~~~~~~~ This module provides the capabilities for the Requests hooks system. Available hooks: ``response``: The response generated from a Request. ZresponsecCsddtDS)NcSsi|] }|gqSr).0Zeventrr>/usr/lib/python3.8/site-packages/pip/_vendor/requests/hooks.py sz!default_hooks..)HOOKSrrrr default_hookssrcKsL|pi}||}|rHt|dr&|g}|D]}||f|}|dk r*|}q*|S)z6Dispatches a hook dictionary on a given piece of data.__call__N)gethasattr)keyZhooksZ hook_datakwargshookZ _hook_datarrr dispatch_hooks   r N)__doc__rrr rrrrs site-packages/pip/_vendor/requests/__pycache__/__version__.cpython-38.pyc000064400000001002151464452720022507 0ustar00U .e@s,dZdZdZdZdZdZdZdZdZd Z d S) ZrequestszPython HTTP for Humans.zhttp://python-requests.orgz2.22.0i"z Kenneth Reitzzme@kennethreitz.orgz Apache 2.0zCopyright 2019 Kenneth Reitzu ✨ 🍰 ✨N) Z __title__Z__description__Z__url__ __version__Z __build__ __author__Z__author_email__Z __license__Z __copyright__Z__cake__rrD/usr/lib/python3.8/site-packages/pip/_vendor/requests/__version__.pyssite-packages/pip/_vendor/requests/__pycache__/utils.cpython-38.opt-1.pyc000064400000053223151464452720022341 0ustar00U .eau@sdZddlZddlZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl Z ddl m Z ddl mZddlmZddlmZddlmZmZmZmZmZmZmZmZmZmZmZmZm Z m!Z!m"Z"dd l#m$Z$dd l%m&Z&dd l'm(Z(m)Z)m*Z*m+Z+d Z,e-Z.d ddZ/ej0dkr2ddZ1ddZddZ2ddZ3dhddZ4ddZ5ddZ6d d!Z7d"d#Z8d$d%Z9d&d'Z:did(d)Z;d*d+Zd0d1Z?d2d3Z@d4d5ZAd6d7ZBd8d9ZCeDd:ZEd;d<ZFd=d>ZGd?d@ZHdAdBZIdCdDZJdEdFZKejLdGdHZMdIdJZNdjdKdLZOdMdNZPdkdPdQZQdRdSZRdTdUZSdVTdWZUeUdXZVeUdYZWdZd[ZXd\d]ZYd^d_ZZe[d`Z\e[daZ]dbdcZ^dddeZ_dfdgZ`dS)lz requests.utils ~~~~~~~~~~~~~~ This module provides utility functions that are used within Requests that are also useful for external consumption. N __version__)certs)to_native_string)parse_http_list)quoteurlparsebytesstr OrderedDictunquote getproxies proxy_bypass urlunparse basestring integer_typesis_py3proxy_bypass_environmentgetproxies_environmentMappingcookiejar_from_dict)CaseInsensitiveDict) InvalidURL InvalidHeaderFileModeWarningUnrewindableBodyError)z.netrc_netrcPi)ZhttpZhttpsZwin32cCsztrddl}nddl}Wntk r0YdSXz6||jd}t||dd}||dd}Wntk r~YdSX|r|sdS| d}|D]T}|dkrd|krd S| dd }| d d }| d d}t ||t j rd SqdS)NrFz;Software\Microsoft\Windows\CurrentVersion\Internet SettingsZ ProxyEnableZ ProxyOverride;z.Tz\.*z.*?)rwinreg_winreg ImportErrorOpenKeyHKEY_CURRENT_USERintZ QueryValueExOSErrorsplitreplacerematchI)hostr$ZinternetSettingsZ proxyEnableZ proxyOverrideZtestr1>/usr/lib/python3.8/site-packages/pip/_vendor/requests/utils.pyproxy_bypass_registry/sF      r3cCstrt|St|SdS)zReturn True, if the host should be bypassed. Checks proxy settings gathered from the environment, if specified, or the registry. N)rrr3)r0r1r1r2rVsrcCst|dr|}|S)z/Returns an internal sequence dictionary update.items)hasattrr4)dr1r1r2dict_to_sequencebs r7c Cs4d}d}t|drt|}nbt|dr.|j}nPt|dr~z |}Wntjk rZYn$Xt|j}d|jkr~t dt t|drz | }Wn$t tfk r|dk r|}Yn^Xt|dr|dkrz&|dd | }||pdWnt tfk rd}YnX|dkr&d}td||S) Nr__len__lenfilenoba%Requests has determined the content-length for this request using the binary size of the file: however, the file has been opened in text mode (i.e. without the 'b' flag in the mode). This may lead to an incorrect content-length. In Requests 3.0, support will be removed for files in text mode.tellseek)r5r9r:ioUnsupportedOperationosfstatst_sizemodewarningswarnrr<r*IOErrorr=max)oZ total_lengthZcurrent_positionr:r1r1r2 super_lenksB             rJFc Cs&zddlm}m}d}tD]J}ztjd|}Wntk rPYWdSXtj|r|}qhq|dkrvWdSt |}d}t |t r| d}|j |d} z:||| } | r| drdnd} | | | dfWWSWn"|tfk r|rYnXWnttfk r YnXdS) z;Returns the Requests tuple auth for a given url from netrc.r)netrcNetrcParseErrorNz~/{}:asciirr>)rKrL NETRC_FILESrApath expanduserformatKeyErrorexistsr isinstancer decodenetlocr+ZauthenticatorsrGr&AttributeError) urlZ raise_errorsrKrLZ netrc_pathfZlocZriZsplitstrr0rZlogin_ir1r1r2get_netrc_auths8     r[cCsBt|dd}|r>t|tr>|ddkr>|ddkr>tj|SdS)z0Tries to guess the filename of the given object.nameNr<>)getattrrUrrArPbasename)objr\r1r1r2guess_filenames   rccCstj|r|Stj|\}}|rPtj|sPtj|\}}d||g}q t|s^|St|}||krx|St }tjj|f|d}tj|s|j ||d}|S)zReplace nonexistent paths that look like they refer to a member of a zip archive with the location of an extracted copy of the target, or else just return the provided path unchanged. /)rP) rArPrTr+joinzipfileZ is_zipfileZZipFileZnamelisttempfileZ gettempdirextract)rParchivememberprefixZzip_fileZtmpZextracted_pathr1r1r2extract_zipped_pathss      rlcCs.|dkr dSt|ttttfr&tdt|S)aTake an object and test to see if it can be represented as a dictionary. Unless it can not be represented as such, return an OrderedDict, e.g., :: >>> from_key_val_list([('key', 'val')]) OrderedDict([('key', 'val')]) >>> from_key_val_list('string') ValueError: cannot encode objects that are not 2-tuples >>> from_key_val_list({'key': 'val'}) OrderedDict([('key', 'val')]) :rtype: OrderedDict N+cannot encode objects that are not 2-tuples)rUr r boolr) ValueErrorr valuer1r1r2from_key_val_lists rrcCs@|dkr dSt|ttttfr&tdt|tr8|}t|S)aTake an object and test to see if it can be represented as a dictionary. If it can be, return a list of tuples, e.g., :: >>> to_key_val_list([('key', 'val')]) [('key', 'val')] >>> to_key_val_list({'key': 'val'}) [('key', 'val')] >>> to_key_val_list('string') ValueError: cannot encode objects that are not 2-tuples. :rtype: list Nrm) rUr r rnr)rorr4listrpr1r1r2to_key_val_lists rtcCsXg}t|D]F}|dd|ddkr4dkrHnnt|dd}||q |S)aParse lists as described by RFC 2068 Section 2. In particular, parse comma-separated lists where the elements of the list may include quoted-strings. A quoted-string could contain a comma. A non-quoted string could have quotes in the middle. Quotes are removed automatically after parsing. It basically works like :func:`parse_set_header` just that items may appear multiple times and case sensitivity is preserved. The return value is a standard :class:`list`: >>> parse_list_header('token, "quoted value"') ['token', 'quoted value'] To create a header from the :class:`list` again, use the :func:`dump_header` function. :param value: a string with a list header. :return: :class:`list` :rtype: list Nrr^")_parse_list_headerunquote_header_valueappend)rqresultitemr1r1r2parse_list_header8s  ( r{cCsxi}t|D]f}d|kr"d||<q |dd\}}|dd|ddkrVdkrjnnt|dd}|||<q |S)a^Parse lists of key, value pairs as described by RFC 2068 Section 2 and convert them into a python dict: >>> d = parse_dict_header('foo="is a fish", bar="as well"') >>> type(d) is dict True >>> sorted(d.items()) [('bar', 'as well'), ('foo', 'is a fish')] If there is no value for a key it will be `None`: >>> parse_dict_header('key_without_value') {'key_without_value': None} To create a header from the :class:`dict` again, use the :func:`dump_header` function. :param value: a string with a dict header. :return: :class:`dict` :rtype: dict =Nrr^ru)rvr+rw)rqryrzr\r1r1r2parse_dict_headerXs ( r}cCs\|rX|d|dkr dkrXnn4|dd}|rD|dddkrX|ddd dS|S) zUnquotes a header value. (Reversal of :func:`quote_header_value`). This does not use the real unquoting but what browsers are actually using for quoting. :param value: the header value to unquote. :rtype: str rr^rurNr>z\\\z\")r,)rqZ is_filenamer1r1r2rw{s $ rwcCsi}|D]}|j||j<q|S)zReturns a key/value dictionary from a CookieJar. :param cj: CookieJar object to extract cookies from. :rtype: dict )rqr\)cj cookie_dictZcookier1r1r2dict_from_cookiejarsrcCs t||S)zReturns a CookieJar from a key/value dictionary. :param cj: CookieJar to insert cookies into. :param cookie_dict: Dict of key/values to insert into CookieJar. :rtype: CookieJar r)rrr1r1r2add_dict_to_cookiejarsrcCsTtdttjdtjd}tjdtjd}td}||||||S)zlReturns encodings from given content string. :param content: bytestring to extract encodings from. zIn requests 3.0, get_encodings_from_content will be removed. For more information, please see the discussion on issue #2266. (This warning should only appear once.)z!])flagsz+]z$^<\?xml.*?encoding=["\']*(.+?)["\'>])rErFDeprecationWarningr-compiler/findall)contentZ charset_reZ pragma_reZxml_rer1r1r2get_encodings_from_contents rc Cs|d}|d|dd}}i}d}|D]`}|}|r0|d}}|d} | dkr|d| |}|| dd|}|||<q0||fS) zReturns content type and parameters from given header :param header: string :return: tuple containing content type and dictionary of parameters r rrNz"' Tr|r^)r+stripfindlower) headertokens content_typeparamsZ params_dictZitems_to_stripparamkeyrqZindex_of_equalsr1r1r2_parse_content_type_headers   rcCsD|d}|sdSt|\}}d|kr4|ddSd|kr@dSdS)z}Returns encodings from given HTTP Header Dict. :param headers: dictionary to extract encoding from. :rtype: str z content-typeNcharsetz'"textz ISO-8859-1)getrr)headersrrr1r1r2get_encoding_from_headerss  rccsj|jdkr|D] }|VqdSt|jdd}|D]}||}|r4|Vq4|jddd}|rf|VdS)zStream decodes a iterator.Nr,errorsT)final)encodingcodecsgetincrementaldecoderrV)iteratorrrzdecoderchunkrvr1r1r2stream_decode_response_unicodes  rccsHd}|dks|dkrt|}|t|krD||||V||7}qdS)z Iterate over slices of a string.rN)r9)stringZ slice_lengthposr1r1r2 iter_slicess  rcCs~tdtg}t|j}|rLzt|j|WStk rJ||YnXzt|j|ddWSt k rx|jYSXdS)zReturns the requested content back in unicode. :param r: Response object to get unicode content from. Tried: 1. charset from content-type 2. fall back and replace all unicode characters :rtype: str zIn requests 3.0, get_unicode_from_response will be removed. For more information, please see the discussion on issue #2266. (This warning should only appear once.)r,rN) rErFrrrr r UnicodeErrorrx TypeError)rZtried_encodingsrr1r1r2get_unicode_from_response s  rzBABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-._~c Cs|d}tdt|D]}||dd}t|dkr|rztt|d}Wn tk rrtd|YnX|tkr|||dd||<qd||||<qd||||<qd |S) zUn-escape any percent-escape sequences in a URI that are unreserved characters. This leaves all reserved, illegal and non-ASCII bytes encoded. :rtype: str %rrr>z%Invalid percent-escape sequence: '%s'N) r+ranger9isalnumchrr)rorUNRESERVED_SETre)uripartsihcr1r1r2unquote_unreserved5s rcCs@d}d}ztt||dWStk r:t||dYSXdS)zRe-quote the given URI. This function passes the given URI through an unquote/quote cycle to ensure that it is fully and consistently quoted. :rtype: str z!#$%&'()*+,/:;=?@[]~z!#$&'()*+,/:;=?@[]~)ZsafeN)rrr)rZsafe_with_percentZsafe_without_percentr1r1r2 requote_uriMs rcCsltdt|d}|d\}}tdttt|d}tdt|d|@}||@||@kS)zThis function allows you to check if an IP belongs to a network subnet Example: returns True if ip = 192.168.1.1 and net = 192.168.1.0/24 returns False if ip = 192.168.1.1 and net = 192.168.100.0/24 :rtype: bool z=Lrrd)structunpacksocket inet_atonr+dotted_netmaskr))ZipZnetZipaddrZnetaddrbitsZnetmaskZnetworkr1r1r2address_in_networkcs rcCs&ddd|>dA}ttd|S)zConverts mask from /xx format to xxx.xxx.xxx.xxx Example: if mask is 24 function returns 255.255.255.0 :rtype: str lr z>I)rZ inet_ntoarpack)maskrr1r1r2rrsrcCs,zt|Wntjk r&YdSXdS)z :rtype: bool FT)rrerror)Z string_ipr1r1r2is_ipv4_address}s rcCs|ddkrzt|dd}Wntk r:YdSX|dksL|dkrPdSzt|ddWqtjk rYdSXndSdS)zV Very simple check of the cidr format in no_proxy variable. :rtype: bool rdrFrrT)countr)r+rorrr)Zstring_networkrr1r1r2 is_valid_cidrs rc csT|dk }|r"tj|}|tj|<z dVW5|rN|dkrDtj|=n |tj|<XdS)zSet the environment variable 'env_name' to 'value' Save previous value, yield, and then restore the previous value stored in the environment variable 'env_name'. If 'value' is None, do nothingN)rAenvironr)Zenv_namerqZ value_changed old_valuer1r1r2 set_environs    rc Cs*dd}|}|dkr|d}t|}|jdkr2dS|rdd|dd d D}t|jr|D]0}t|rt|j|rdSq`|j|kr`dSq`nB|j}|jr|d |j7}|D] }|j |s| |rdSqt d|6zt |j}Wn t t jfk rd }YnXW5QRX|r&dSd S) zL Returns whether we should bypass proxies or not. :rtype: bool cSstj|ptj|SN)rArrupper)kr1r1r2rz'should_bypass_proxies..Nno_proxyTcss|]}|r|VqdSrr1).0r0r1r1r2 sz(should_bypass_proxies.. r,z:{}F)r hostnamer,r+rrrZportrRendswithrrrrZgaierror) rYrZ get_proxyZ no_proxy_argparsedZproxy_ipZhost_with_portr0Zbypassr1r1r2should_bypass_proxiess@      rcCst||driStSdS)zA Return a dict of environment proxies. :rtype: dict )rN)rr)rYrr1r1r2get_environ_proxiess rcCst|pi}t|}|jdkr.||j|dS|jd|j|jd|jdg}d}|D]}||krV||}qpqV|S)zSelect a proxy for the url, if applicable. :param url: The url being for the request :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs Nallz://zall://)r rrscheme)rYZproxiesZurlpartsZ proxy_keysproxyZ proxy_keyr1r1r2 select_proxys rpython-requestscCs d|tfS)zO Return a string representing the default user agent. :rtype: str z%s/%sr)r\r1r1r2default_user_agentsrcCsttdddddS)z9 :rtype: requests.structures.CaseInsensitiveDict z, )ZgzipZdeflatez*/*z keep-alive)z User-AgentzAccept-EncodingZAcceptZ Connection)rrrer1r1r1r2default_headers"s rc Csg}d}||}|s|Std|D]}z|dd\}}Wntk r\|d}}YnXd|di}|dD]F}z|d\}}Wntk rYqYnX|||||<qv||q&|S) zReturn a list of parsed link headers proxies. i.e. Link: ; rel=front; type="image/jpeg",; rel=back;type="image/jpeg" :rtype: list z '"z, * '"r|)rr-r+rorx) rqZlinksZ replace_charsvalrYrlinkrrr1r1r2parse_header_links.s&   rrNr>cCs|dd}|tjtjfkr dS|ddtjkr6dS|ddtjtjfkrRdS|t}|dkrhd S|dkr|dddtkrd S|d ddtkrd S|dkr|ddt krd S|d dt krdSdS)z :rtype: str Nzutf-32rz utf-8-sigr>zutf-16rzutf-8z utf-16-berz utf-16-lez utf-32-bez utf-32-le) r BOM_UTF32_LE BOM_UTF32_BEBOM_UTF8 BOM_UTF16_LE BOM_UTF16_BEr_null_null2_null3)dataZsampleZ nullcountr1r1r2guess_json_utfYs*  rcCs8t||\}}}}}}|s$||}}t||||||fS)zGiven a URL that may or may not have a scheme, prepend the given scheme. Does not replace a present scheme with the one provided as an argument. :rtype: str )r r)rYZ new_schemerrWrPrqueryfragmentr1r1r2prepend_scheme_if_neededys rc CsBt|}zt|jt|jf}Wnttfk r<d}YnX|S)z{Given a url with authentication components, extract them into a tuple of username,password. :rtype: (str,str) )rr)r r ZusernameZpasswordrXr)rYrZauthr1r1r2get_auth_from_urls  rs^\S[^\r\n]*$|^$z^\S[^\r\n]*$|^$c Csf|\}}t|trt}nt}z||s4td|Wn*tk r`td||t|fYnXdS)zVerifies that header value is a string which doesn't contain leading whitespace or return characters. This prevents unintended header injection. :param header: tuple, in the format (name, value). z7Invalid return character or leading space in header: %sz>Value for header {%s: %s} must be of type str or bytes, not %sN)rUr _CLEAN_HEADER_REGEX_BYTE_CLEAN_HEADER_REGEX_STRr.rrtype)rr\rqZpatr1r1r2check_header_validitys   rcCsFt|\}}}}}}|s"||}}|ddd}t|||||dfS)zW Given a url remove the fragment and the authentication part. :rtype: str @rr^r)r rsplitr)rYrrWrPrrrr1r1r2 urldefragauths  rc Cs`t|jdd}|dk rTt|jtrTz||jWq\ttfk rPtdYq\XntddS)zfMove file pointer back to its recorded starting position so it can be read again on redirect. r=Nz;An error occurred when rewinding request body for redirect.z+Unable to rewind request body for redirect.)r`ZbodyrUZ_body_positionrrGr*r)Zprepared_requestZ body_seekr1r1r2 rewind_bodysr)F)F)N)r)a__doc__r contextlibr?rAr-rrsysrgrErfrrrZ_internal_utilsrcompatrrvrr r r r r rrrrrrrrrZcookiesrZ structuresr exceptionsrrrrrOwhereZDEFAULT_CA_BUNDLE_PATHZ DEFAULT_PORTSplatformr3r7rJr[rcrlrrrtr{r}rwrrrrrrrr frozensetrrrrrrrcontextmanagerrrrrrrrencoderrrrrrrrrrrrr1r1r1r2s    D    '  = 3  #   %   =   &    site-packages/pip/_vendor/requests/__pycache__/models.cpython-38.opt-1.pyc000064400000056524151464452720022473 0ustar00U .e@sdZddlZddlZddlZddlmZddlmZddl m Z ddl m Z m Z mZmZddlmZdd lmZdd lmZdd lmZdd lmZmZmZdd lmZmZmZm Z m!Z!m"Z"m#Z#ddl$m%Z%m&Z&ddl'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1ddl2m3Z3m4Z4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;mZ>ddl2m?Z@ddlAmBZBeBjCeBjDeBjEeBjFeBjGfZHdZIdZJdZKGdddeLZMGdddeLZNGdddeNZOGdddeMeNZPGdddeLZQdS) z` requests.models ~~~~~~~~~~~~~~~ This module contains the primary objects that power Requests. N) RequestField)encode_multipart_formdata) parse_url) DecodeErrorReadTimeoutError ProtocolErrorLocationParseError)UnsupportedOperation) default_hooks)CaseInsensitiveDict) HTTPBasicAuth)cookiejar_from_dictget_cookie_header_copy_cookie_jar) HTTPError MissingSchema InvalidURLChunkedEncodingErrorContentDecodingErrorConnectionErrorStreamConsumedError)to_native_stringunicode_is_ascii) guess_filenameget_auth_from_url requote_uristream_decode_response_unicodeto_key_val_listparse_header_links iter_slicesguess_json_utf super_lencheck_header_validity) CallableMapping cookielib urlunparseurlsplit urlencodestrbytesis_py2chardet builtin_str basestring)json)codesi(ic@s0eZdZeddZeddZeddZdS)RequestEncodingMixincCsNg}t|j}|j}|sd}|||j}|rD|d||d|S)zBuild the path URL to use./?)r(urlpathappendqueryjoin)selfr7pr8r:r>?/usr/lib/python3.8/site-packages/pip/_vendor/requests/models.pypath_url=s    zRequestEncodingMixin.path_urlcCst|ttfr|St|dr |St|drg}t|D]j\}}t|tsRt|dsX|g}|D]B}|dk r\|t|tr|dn|t|tr|dn|fq\q6t|ddS|SdS)zEncode parameters in a piece of data. Will successfully encode parameters when passed as a dict or a list of 2-tuples. Order is retained if data is a list of 2-tuples but arbitrary if parameters are supplied as a dict. read__iter__Nutf-8T)Zdoseq) isinstancer*r+hasattrrr/r9encoder))dataresultkZvsvr>r>r?_encode_paramsRs$    z#RequestEncodingMixin._encode_paramscCs|stdnt|tr tdg}t|p,i}t|p8i}|D]|\}}t|ts\t|dsb|g}|D]T}|dk rft|tst|}|t|tr|dn|t|tr| dn|fqfq@|D]\}}d}d} t|t t fr t |dkr|\} } n&t |dkr|\} } }n |\} } }} nt |p,|} |} t| tttfrJ| } n(t| dr`| } n| dkrnqn| } t|| | | d } | j|d || qt|\}}||fS) aBuild the body for a multipart/form-data request. Will successfully encode files when passed as a dict or a list of tuples. Order is retained if data is a list of tuples but arbitrary if parameters are supplied as a dict. The tuples may be 2-tuples (filename, fileobj), 3-tuples (filename, fileobj, contentype) or 4-tuples (filename, fileobj, contentype, custom_headers). zFiles must be provided.zData must not be a string.rBNrCrA)namerGfilenameheaders) content_type) ValueErrorrDr/rrEr+r*r9decoderFtuplelistlenr bytearrayrArZmake_multipartr)filesrGZ new_fieldsZfieldsZfieldvalrJrIZftZfhfnfpZfdataZrfbodyrQr>r>r? _encode_filesmsT                 z"RequestEncodingMixin._encode_filesN)__name__ __module__ __qualname__propertyr@ staticmethodrKr]r>r>r>r?r3<s   r3c@seZdZddZddZdS)RequestHooksMixincCsZ||jkrtd|t|tr2|j||n$t|drV|j|dd|DdS)zProperly register a hook.z1Unsupported event specified, with event name "%s"rBcss|]}t|tr|VqdSN)rDr$).0hr>r>r? s z2RequestHooksMixin.register_hook..N)hooksrRrDr$r9rEextendr<eventhookr>r>r? register_hooks     zRequestHooksMixin.register_hookcCs2z|j||WdStk r,YdSXdS)ziDeregister a previously registered hook. Returns True if the hook existed, False if not. TFN)rhremoverRrjr>r>r?deregister_hooks z!RequestHooksMixin.deregister_hookN)r^r_r`rmror>r>r>r?rcs rcc @s*eZdZdZd ddZddZddZdS) Requesta{A user-created :class:`Request ` object. Used to prepare a :class:`PreparedRequest `, which is sent to the server. :param method: HTTP method to use. :param url: URL to send. :param headers: dictionary of headers to send. :param files: dictionary of {filename: fileobject} files to multipart upload. :param data: the body to attach to the request. If a dictionary or list of tuples ``[(key, value)]`` is provided, form-encoding will take place. :param json: json for the body to attach to the request (if files or data is not specified). :param params: URL parameters to append to the URL. If a dictionary or list of tuples ``[(key, value)]`` is provided, form-encoding will take place. :param auth: Auth handler or (user, pass) tuple. :param cookies: dictionary or CookieJar of cookies to attach to this request. :param hooks: dictionary of callback hooks, for internal usage. Usage:: >>> import requests >>> req = requests.Request('GET', 'https://httpbin.org/get') >>> req.prepare() Nc Cs|dkr gn|}|dkrgn|}|dkr,in|}|dkrr>r?__init__s"zRequest.__init__cCs d|jS)Nzrrr<r>r>r?__repr__szRequest.__repr__c Cs<t}|j|j|j|j|j|j|j|j|j |j |j d |S)zXConstructs a :class:`PreparedRequest ` for transmission and returns it.) rrr7rPrXrGr0rsrtrurh) PreparedRequestpreparerrr7rPrXrGr0rsrtrurhr<r=r>r>r?r{s zRequest.prepare) NNNNNNNNNN)r^r_r`__doc__rvryr{r>r>r>r?rps rpc @seZdZdZddZdddZddZd d Zd d Ze d dZ ddZ ddZ dddZ ddZd ddZddZddZdS)!rzaThe fully mutable :class:`PreparedRequest ` object, containing the exact bytes that will be sent to the server. Generated from either a :class:`Request ` object or manually. Usage:: >>> import requests >>> req = requests.Request('GET', 'https://httpbin.org/get') >>> r = req.prepare() >>> s = requests.Session() >>> s.send(r) cCs0d|_d|_d|_d|_d|_t|_d|_dSrd)rrr7rP_cookiesr\r rh_body_positionrxr>r>r?rv"szPreparedRequest.__init__Nc CsR||||||||||||| ||||| dS)z6Prepares the entire request with the given parameters.N)prepare_method prepare_urlprepare_headersprepare_cookies prepare_body prepare_auth prepare_hooks) r<rrr7rPrXrGrsrtrurhr0r>r>r?r{3s     zPreparedRequest.preparecCs d|jS)Nzrwrxr>r>r?ryEszPreparedRequest.__repr__cCsXt}|j|_|j|_|jdk r*|jnd|_t|j|_|j|_|j|_|j |_ |Srd) rzrrr7rPcopyrr~r\rhrr|r>r>r?rHs zPreparedRequest.copycCs$||_|jdk r t|j|_dS)zPrepares the given HTTP method.N)rrrupper)r<rrr>r>r?rSs zPreparedRequest.prepare_methodcCsDddlm}z|j|ddd}Wn|jk r>tYnX|S)Nr)idnaT)Zuts46rC)Z pip._vendorrrFrSZ IDNAError UnicodeError)hostrr>r>r?_get_idna_encoded_hostYs   z&PreparedRequest._get_idna_encoded_hostc Cs,t|tr|d}ntr"t|nt|}|}d|krR|dsR||_ dSzt |\}}}}}}} Wn,t k r} zt | j W5d} ~ XYnX|sd} | t|d} t| |st d|t|sz||}Wntk rt dYnXn|drt d|pd } | r.| d 7} | |7} |rL| dt|7} |sVd }trt|trr|d }t| tr| d } t|tr|d }t|tr|d }t| tr| d } t|ttfrt|}||} | r |rd || f}n| }tt|| |d|| g}||_ dS)zPrepares the given HTTP URL.utf8:ZhttpNzDInvalid URL {0!r}: No schema supplied. Perhaps you meant http://{0}?z Invalid URL %r: No host suppliedzURL has an invalid label.*r6@r4rCz%s&%s)rDr+rSr,Zunicoder*lstriplower startswithr7rrrargsformatrrrrrrFrKrr')r<r7rsZschemertrZportr8r:ZfragmenteerrorZnetlocZ enc_paramsr>r>r?rcsh                 zPreparedRequest.prepare_urlcCs<t|_|r8|D]"}t||\}}||jt|<qdS)z Prepares the given HTTP headers.N)r rPrqr#r)r<rPheaderrNvaluer>r>r?rs  zPreparedRequest.prepare_headersc Csrd}d}|s6|dk r6d}t|}t|ts6|d}tt|dt|ttt t f g}z t |}Wnt t tfk rd}YnX|r|}t|dddk rz||_Wn ttfk rt|_YnX|rtd|rt||jd<n d|jd <np|r|||\}}n2|rB||}t|ts8t|d r>d}nd }|||rhd |jkrh||jd <||_dS)z"Prepares the given HTTP body data.Nzapplication/jsonrCrBtellz1Streamed bodies and files are mutually exclusive.Content-LengthZchunkedzTransfer-EncodingrAz!application/x-www-form-urlencodedz content-typez Content-Type) complexjsondumpsrDr+rFallrEr/rUrTr%r" TypeErrorAttributeErrorr getattrrrIOErrorOSErrorobjectNotImplementedErrorr.rPr]rKprepare_content_lengthr\)r<rGrXr0r\rQZ is_streamlengthr>r>r?rsL          zPreparedRequest.prepare_bodycCsL|dk r$t|}|rHt||jd<n$|jdkrH|jddkrHd|jd<dS)z>Prepare Content-Length header based on request method and bodyNr)ZGETZHEAD0)r"r.rPrrget)r<r\rr>r>r?r s z&PreparedRequest.prepare_content_lengthr6cCsj|dkr"t|j}t|r|nd}|rft|trDt|dkrDt|}||}|j|j| |j dS)z"Prepares the given HTTP auth data.NrL) rr7anyrDrTrVr __dict__updaterr\)r<rtr7Zurl_authrr>r>r?rs zPreparedRequest.prepare_authcCs@t|tjr||_n t||_t|j|}|dk r<||jd<dS)aPrepares the given HTTP cookie data. This function eventually generates a ``Cookie`` header from the given cookies using cookielib. Due to cookielib's design, the header will not be regenerated if it already exists, meaning this function can only be called once for the life of the :class:`PreparedRequest ` object. Any subsequent calls to ``prepare_cookies`` will have no actual effect, unless the "Cookie" header is removed beforehand. NZCookie)rDr&Z CookieJarr~rrrP)r<ruZ cookie_headerr>r>r?r,s   zPreparedRequest.prepare_cookiescCs&|pg}|D]}||||q dS)zPrepares the given hooks.N)rm)r<rhrkr>r>r?r@szPreparedRequest.prepare_hooks) NNNNNNNNNN)N)r6)r^r_r`r}rvr{ryrrrbrrrrrrrrr>r>r>r?rzs2   V E rzc @seZdZdZdddddddd d d g Zd d ZddZddZddZddZ ddZ ddZ ddZ ddZ eddZed d!Zed"d#Zed$d%Zed&d'Zd;d*d+Zed)d,fd-d.Zed/d0Zed1d2Zd3d4Zed5d6Zd7d8Zd9d:Zd,S)<ResponsezhThe :class:`Response ` object, which contains a server's response to an HTTP request. _content status_coderPr7historyencodingreasonruelapsedrequestcCs^d|_d|_d|_d|_t|_d|_d|_d|_g|_ d|_ t i|_ t d|_d|_dS)NFr)r_content_consumed_nextrr rPrawr7rrrrrudatetimeZ timedeltarrrxr>r>r?rvTs  zResponse.__init__cCs|Srdr>rxr>r>r? __enter__szResponse.__enter__cGs |dSrd)close)r<rr>r>r?__exit__szResponse.__exit__cs js jfddjDS)Ncsi|]}|t|dqSrd)r)reattrrxr>r? sz)Response.__getstate__..)rcontent __attrs__rxr>rxr? __getstate__szResponse.__getstate__cCs:|D]\}}t|||qt|ddt|dddS)NrTr)rqsetattr)r<staterNrr>r>r? __setstate__s zResponse.__setstate__cCs d|jS)Nz)rrxr>r>r?ryszResponse.__repr__cCs|jSakReturns True if :attr:`status_code` is less than 400. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code, is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. okrxr>r>r?__bool__szResponse.__bool__cCs|jSrrrxr>r>r? __nonzero__szResponse.__nonzero__cCs |dS)z,Allows you to use a response as an iterator.) iter_contentrxr>r>r?rBszResponse.__iter__cCs(z |Wntk r"YdSXdS)axReturns True if :attr:`status_code` is less than 400, False if not. This attribute checks if the status code of the response is between 400 and 600 to see if there was a client error or a server error. If the status code is between 200 and 400, this will return True. This is **not** a check to see if the response code is ``200 OK``. FT)raise_for_statusrrxr>r>r?rs  z Response.okcCsd|jko|jtkS)zTrue if this Response is a well-formed HTTP redirect that could have been processed automatically (by :meth:`Session.resolve_redirects`). location)rPrREDIRECT_STATIrxr>r>r? is_redirectszResponse.is_redirectcCsd|jko|jtjtjfkS)z@True if this Response one of the permanent versions of redirect.r)rPrr1Zmoved_permanentlypermanent_redirectrxr>r>r?is_permanent_redirectszResponse.is_permanent_redirectcCs|jS)zTReturns a PreparedRequest for the next request in a redirect chain, if there is one.)rrxr>r>r?nextsz Response.nextcCst|jdS)z7The apparent encoding, provided by the chardet library.r)r-Zdetectrrxr>r>r?apparent_encodingszResponse.apparent_encodingr Fcs|fdd}jr(tjtr(tn"dk rJttsJtdttj}|}jrf|n|}|rxt |}|S)aIterates over the response data. When stream=True is set on the request, this avoids reading the content at once into memory for large responses. The chunk size is the number of bytes it should read into memory. This is not necessarily the length of each item returned as decoding can take place. chunk_size must be of type int or None. A value of None will function differently depending on the value of `stream`. stream=True will read data as it arrives in whatever size the chunks are received. If stream=False, data is returned as a single chunk. If decode_unicode is True, content will be decoded using the best available encoding based on the response. c 3stjdrz jjddD] }|VqWqtk rV}z t|W5d}~XYqtk r~}z t|W5d}~XYqtk r}z t|W5d}~XYqXnj }|sq|Vqd_ dS)NstreamT)Zdecode_content) rErrrrrrrrrAr)chunkr chunk_sizer<r>r?generates   z'Response.iter_content..generateNz.chunk_size must be an int, it is instead a %s.) rrDrboolrintrtyper r)r<rdecode_unicoderZ reused_chunksZ stream_chunksZchunksr>rr?rs  zResponse.iter_contentNccsd}|j||dD]n}|dk r&||}|r6||}n|}|rl|drl|rl|dd|dkrl|}nd}|D] }|Vqtq|dk r|VdS)zIterates over the response data, one line at a time. When stream=True is set on the request, this avoids reading the content at once into memory for large responses. .. note:: This method is not reentrant safe. N)rr)rsplit splitlinespop)r<rrZ delimiterZpendingrlinesliner>r>r? iter_liness $  zResponse.iter_linescCsV|jdkrJ|jrtd|jdks,|jdkr4d|_nd|tpFd|_d|_|jS)z"Content of the response, in bytes.Fz2The content for this response was already consumedrNT)rr RuntimeErrorrrr;rCONTENT_CHUNK_SIZErxr>r>r?r/s zResponse.contentc Cshd}|j}|jstdS|jdkr(|j}zt|j|dd}Wn&ttfk rbt|jdd}YnX|S)aContent of the response, in unicode. If Response.encoding is None, encoding will be guessed using ``chardet``. The encoding of the response content is determined based solely on HTTP headers, following RFC 2616 to the letter. If you can take advantage of non-HTTP knowledge to make a better guess at the encoding, you should set ``r.encoding`` appropriately before accessing this property. Nr6replace)errors)rrr*r LookupErrorr)r<rrr>r>r?textCs z Response.textcKsj|jsZ|jrZt|jdkrZt|j}|dk rZztj|j|f|WStk rXYnXtj|jf|S)zReturns the json-encoded content of a response, if any. :param \*\*kwargs: Optional arguments that ``json.loads`` takes. :raises ValueError: If the response body does not contain valid json. rMN) rrrVr!rloadsrSUnicodeDecodeErrorr)r<kwargsrr>r>r?r0is  z Response.jsoncCsF|jd}i}|rBt|}|D] }|dp6|d}|||<q |S)z8Returns the parsed header links of the response, if any.linkZrelr7)rPrr)r<rllinksrkeyr>r>r?rs  zResponse.linkscCsd}t|jtrDz|jd}WqJtk r@|jd}YqJXn|j}d|jkr`dkrxnnd|j||jf}n,d|jkrdkrnnd|j||jf}|rt||d d S) z2Raises stored :class:`HTTPError`, if one occurred.r6rCz iso-8859-1iiz%s Client Error: %s for url: %siXz%s Server Error: %s for url: %s)ZresponseN)rDrr+rSrrr7r)r<Zhttp_error_msgrr>r>r?rs zResponse.raise_for_statuscCs0|js|jt|jdd}|dk r,|dS)zReleases the connection back to the pool. Once this method has been called the underlying ``raw`` object must not be accessed again. *Note: Should not normally need to be called explicitly.* release_connN)rrrr)r<rr>r>r?rs  zResponse.close)r F)r^r_r`r}rrvrrrrryrrrBrarrrrrrITER_CHUNK_SIZErrrr0rrrr>r>r>r?rJsT/        7  % r)Rr}rsysZencodings.idnaZ encodingsZpip._vendor.urllib3.fieldsrZpip._vendor.urllib3.filepostrZpip._vendor.urllib3.utilrZpip._vendor.urllib3.exceptionsrrrrior rhr Z structuresr rtr rurrr exceptionsrrrrrrrZ_internal_utilsrrZutilsrrrrrrr r!r"r#compatr$r%r&r'r(r)r*r+r,r-r.r/r0rZ status_codesr1ZmovedfoundotherZtemporary_redirectrrZDEFAULT_REDIRECT_LIMITrrrr3rcrprzrr>r>r>r?sD       $08  rJ<site-packages/pip/_vendor/requests/exceptions.py000064400000006175151464452720016141 0ustar00# -*- coding: utf-8 -*- """ requests.exceptions ~~~~~~~~~~~~~~~~~~~ This module contains the set of Requests' exceptions. """ from pip._vendor.urllib3.exceptions import HTTPError as BaseHTTPError class RequestException(IOError): """There was an ambiguous exception that occurred while handling your request. """ def __init__(self, *args, **kwargs): """Initialize RequestException with `request` and `response` objects.""" response = kwargs.pop('response', None) self.response = response self.request = kwargs.pop('request', None) if (response is not None and not self.request and hasattr(response, 'request')): self.request = self.response.request super(RequestException, self).__init__(*args, **kwargs) class HTTPError(RequestException): """An HTTP error occurred.""" class ConnectionError(RequestException): """A Connection error occurred.""" class ProxyError(ConnectionError): """A proxy error occurred.""" class SSLError(ConnectionError): """An SSL error occurred.""" class Timeout(RequestException): """The request timed out. Catching this error will catch both :exc:`~requests.exceptions.ConnectTimeout` and :exc:`~requests.exceptions.ReadTimeout` errors. """ class ConnectTimeout(ConnectionError, Timeout): """The request timed out while trying to connect to the remote server. Requests that produced this error are safe to retry. """ class ReadTimeout(Timeout): """The server did not send any data in the allotted amount of time.""" class URLRequired(RequestException): """A valid URL is required to make a request.""" class TooManyRedirects(RequestException): """Too many redirects.""" class MissingSchema(RequestException, ValueError): """The URL schema (e.g. http or https) is missing.""" class InvalidSchema(RequestException, ValueError): """See defaults.py for valid schemas.""" class InvalidURL(RequestException, ValueError): """The URL provided was somehow invalid.""" class InvalidHeader(RequestException, ValueError): """The header value provided was somehow invalid.""" class InvalidProxyURL(InvalidURL): """The proxy URL provided is invalid.""" class ChunkedEncodingError(RequestException): """The server declared chunked encoding but sent an invalid chunk.""" class ContentDecodingError(RequestException, BaseHTTPError): """Failed to decode response content""" class StreamConsumedError(RequestException, TypeError): """The content for this response was already consumed""" class RetryError(RequestException): """Custom retries logic failed""" class UnrewindableBodyError(RequestException): """Requests encountered an error when trying to rewind a body""" # Warnings class RequestsWarning(Warning): """Base warning for Requests.""" pass class FileModeWarning(RequestsWarning, DeprecationWarning): """A file was opened in text mode, but Requests determined its binary length.""" pass class RequestsDependencyWarning(RequestsWarning): """An imported dependency doesn't match the expected version range.""" pass site-packages/pip/_vendor/requests/status_codes.py000064400000010041151464452720016443 0ustar00# -*- coding: utf-8 -*- r""" The ``codes`` object defines a mapping from common names for HTTP statuses to their numerical codes, accessible either as attributes or as dictionary items. >>> requests.codes['temporary_redirect'] 307 >>> requests.codes.teapot 418 >>> requests.codes['\o/'] 200 Some codes have multiple names, and both upper- and lower-case versions of the names are allowed. For example, ``codes.ok``, ``codes.OK``, and ``codes.okay`` all correspond to the HTTP status code 200. """ from .structures import LookupDict _codes = { # Informational. 100: ('continue',), 101: ('switching_protocols',), 102: ('processing',), 103: ('checkpoint',), 122: ('uri_too_long', 'request_uri_too_long'), 200: ('ok', 'okay', 'all_ok', 'all_okay', 'all_good', '\\o/', '✓'), 201: ('created',), 202: ('accepted',), 203: ('non_authoritative_info', 'non_authoritative_information'), 204: ('no_content',), 205: ('reset_content', 'reset'), 206: ('partial_content', 'partial'), 207: ('multi_status', 'multiple_status', 'multi_stati', 'multiple_stati'), 208: ('already_reported',), 226: ('im_used',), # Redirection. 300: ('multiple_choices',), 301: ('moved_permanently', 'moved', '\\o-'), 302: ('found',), 303: ('see_other', 'other'), 304: ('not_modified',), 305: ('use_proxy',), 306: ('switch_proxy',), 307: ('temporary_redirect', 'temporary_moved', 'temporary'), 308: ('permanent_redirect', 'resume_incomplete', 'resume',), # These 2 to be removed in 3.0 # Client Error. 400: ('bad_request', 'bad'), 401: ('unauthorized',), 402: ('payment_required', 'payment'), 403: ('forbidden',), 404: ('not_found', '-o-'), 405: ('method_not_allowed', 'not_allowed'), 406: ('not_acceptable',), 407: ('proxy_authentication_required', 'proxy_auth', 'proxy_authentication'), 408: ('request_timeout', 'timeout'), 409: ('conflict',), 410: ('gone',), 411: ('length_required',), 412: ('precondition_failed', 'precondition'), 413: ('request_entity_too_large',), 414: ('request_uri_too_large',), 415: ('unsupported_media_type', 'unsupported_media', 'media_type'), 416: ('requested_range_not_satisfiable', 'requested_range', 'range_not_satisfiable'), 417: ('expectation_failed',), 418: ('im_a_teapot', 'teapot', 'i_am_a_teapot'), 421: ('misdirected_request',), 422: ('unprocessable_entity', 'unprocessable'), 423: ('locked',), 424: ('failed_dependency', 'dependency'), 425: ('unordered_collection', 'unordered'), 426: ('upgrade_required', 'upgrade'), 428: ('precondition_required', 'precondition'), 429: ('too_many_requests', 'too_many'), 431: ('header_fields_too_large', 'fields_too_large'), 444: ('no_response', 'none'), 449: ('retry_with', 'retry'), 450: ('blocked_by_windows_parental_controls', 'parental_controls'), 451: ('unavailable_for_legal_reasons', 'legal_reasons'), 499: ('client_closed_request',), # Server Error. 500: ('internal_server_error', 'server_error', '/o\\', '✗'), 501: ('not_implemented',), 502: ('bad_gateway',), 503: ('service_unavailable', 'unavailable'), 504: ('gateway_timeout',), 505: ('http_version_not_supported', 'http_version'), 506: ('variant_also_negotiates',), 507: ('insufficient_storage',), 509: ('bandwidth_limit_exceeded', 'bandwidth'), 510: ('not_extended',), 511: ('network_authentication_required', 'network_auth', 'network_authentication'), } codes = LookupDict(name='status_codes') def _init(): for code, titles in _codes.items(): for title in titles: setattr(codes, title, code) if not title.startswith(('\\', '/')): setattr(codes, title.upper(), code) def doc(code): names = ', '.join('``%s``' % n for n in _codes[code]) return '* %d: %s' % (code, names) global __doc__ __doc__ = (__doc__ + '\n' + '\n'.join(doc(code) for code in sorted(_codes)) if __doc__ is not None else None) _init() site-packages/pip/_vendor/requests/cookies.py000064400000043776151464452720015424 0ustar00# -*- coding: utf-8 -*- """ requests.cookies ~~~~~~~~~~~~~~~~ Compatibility code to be able to use `cookielib.CookieJar` with requests. requests.utils imports from here, so be careful with imports. """ import copy import time import calendar from ._internal_utils import to_native_string from .compat import cookielib, urlparse, urlunparse, Morsel, MutableMapping try: import threading except ImportError: import dummy_threading as threading class MockRequest(object): """Wraps a `requests.Request` to mimic a `urllib2.Request`. The code in `cookielib.CookieJar` expects this interface in order to correctly manage cookie policies, i.e., determine whether a cookie can be set, given the domains of the request and the cookie. The original request object is read-only. The client is responsible for collecting the new headers via `get_new_headers()` and interpreting them appropriately. You probably want `get_cookie_header`, defined below. """ def __init__(self, request): self._r = request self._new_headers = {} self.type = urlparse(self._r.url).scheme def get_type(self): return self.type def get_host(self): return urlparse(self._r.url).netloc def get_origin_req_host(self): return self.get_host() def get_full_url(self): # Only return the response's URL if the user hadn't set the Host # header if not self._r.headers.get('Host'): return self._r.url # If they did set it, retrieve it and reconstruct the expected domain host = to_native_string(self._r.headers['Host'], encoding='utf-8') parsed = urlparse(self._r.url) # Reconstruct the URL as we expect it return urlunparse([ parsed.scheme, host, parsed.path, parsed.params, parsed.query, parsed.fragment ]) def is_unverifiable(self): return True def has_header(self, name): return name in self._r.headers or name in self._new_headers def get_header(self, name, default=None): return self._r.headers.get(name, self._new_headers.get(name, default)) def add_header(self, key, val): """cookielib has no legitimate use for this method; add it back if you find one.""" raise NotImplementedError("Cookie headers should be added with add_unredirected_header()") def add_unredirected_header(self, name, value): self._new_headers[name] = value def get_new_headers(self): return self._new_headers @property def unverifiable(self): return self.is_unverifiable() @property def origin_req_host(self): return self.get_origin_req_host() @property def host(self): return self.get_host() class MockResponse(object): """Wraps a `httplib.HTTPMessage` to mimic a `urllib.addinfourl`. ...what? Basically, expose the parsed HTTP headers from the server response the way `cookielib` expects to see them. """ def __init__(self, headers): """Make a MockResponse for `cookielib` to read. :param headers: a httplib.HTTPMessage or analogous carrying the headers """ self._headers = headers def info(self): return self._headers def getheaders(self, name): self._headers.getheaders(name) def extract_cookies_to_jar(jar, request, response): """Extract the cookies from the response into a CookieJar. :param jar: cookielib.CookieJar (not necessarily a RequestsCookieJar) :param request: our own requests.Request object :param response: urllib3.HTTPResponse object """ if not (hasattr(response, '_original_response') and response._original_response): return # the _original_response field is the wrapped httplib.HTTPResponse object, req = MockRequest(request) # pull out the HTTPMessage with the headers and put it in the mock: res = MockResponse(response._original_response.msg) jar.extract_cookies(res, req) def get_cookie_header(jar, request): """ Produce an appropriate Cookie header string to be sent with `request`, or None. :rtype: str """ r = MockRequest(request) jar.add_cookie_header(r) return r.get_new_headers().get('Cookie') def remove_cookie_by_name(cookiejar, name, domain=None, path=None): """Unsets a cookie by name, by default over all domains and paths. Wraps CookieJar.clear(), is O(n). """ clearables = [] for cookie in cookiejar: if cookie.name != name: continue if domain is not None and domain != cookie.domain: continue if path is not None and path != cookie.path: continue clearables.append((cookie.domain, cookie.path, cookie.name)) for domain, path, name in clearables: cookiejar.clear(domain, path, name) class CookieConflictError(RuntimeError): """There are two cookies that meet the criteria specified in the cookie jar. Use .get and .set and include domain and path args in order to be more specific. """ class RequestsCookieJar(cookielib.CookieJar, MutableMapping): """Compatibility class; is a cookielib.CookieJar, but exposes a dict interface. This is the CookieJar we create by default for requests and sessions that don't specify one, since some clients may expect response.cookies and session.cookies to support dict operations. Requests does not use the dict interface internally; it's just for compatibility with external client code. All requests code should work out of the box with externally provided instances of ``CookieJar``, e.g. ``LWPCookieJar`` and ``FileCookieJar``. Unlike a regular CookieJar, this class is pickleable. .. warning:: dictionary operations that are normally O(1) may be O(n). """ def get(self, name, default=None, domain=None, path=None): """Dict-like get() that also supports optional domain and path args in order to resolve naming collisions from using one cookie jar over multiple domains. .. warning:: operation is O(n), not O(1). """ try: return self._find_no_duplicates(name, domain, path) except KeyError: return default def set(self, name, value, **kwargs): """Dict-like set() that also supports optional domain and path args in order to resolve naming collisions from using one cookie jar over multiple domains. """ # support client code that unsets cookies by assignment of a None value: if value is None: remove_cookie_by_name(self, name, domain=kwargs.get('domain'), path=kwargs.get('path')) return if isinstance(value, Morsel): c = morsel_to_cookie(value) else: c = create_cookie(name, value, **kwargs) self.set_cookie(c) return c def iterkeys(self): """Dict-like iterkeys() that returns an iterator of names of cookies from the jar. .. seealso:: itervalues() and iteritems(). """ for cookie in iter(self): yield cookie.name def keys(self): """Dict-like keys() that returns a list of names of cookies from the jar. .. seealso:: values() and items(). """ return list(self.iterkeys()) def itervalues(self): """Dict-like itervalues() that returns an iterator of values of cookies from the jar. .. seealso:: iterkeys() and iteritems(). """ for cookie in iter(self): yield cookie.value def values(self): """Dict-like values() that returns a list of values of cookies from the jar. .. seealso:: keys() and items(). """ return list(self.itervalues()) def iteritems(self): """Dict-like iteritems() that returns an iterator of name-value tuples from the jar. .. seealso:: iterkeys() and itervalues(). """ for cookie in iter(self): yield cookie.name, cookie.value def items(self): """Dict-like items() that returns a list of name-value tuples from the jar. Allows client-code to call ``dict(RequestsCookieJar)`` and get a vanilla python dict of key value pairs. .. seealso:: keys() and values(). """ return list(self.iteritems()) def list_domains(self): """Utility method to list all the domains in the jar.""" domains = [] for cookie in iter(self): if cookie.domain not in domains: domains.append(cookie.domain) return domains def list_paths(self): """Utility method to list all the paths in the jar.""" paths = [] for cookie in iter(self): if cookie.path not in paths: paths.append(cookie.path) return paths def multiple_domains(self): """Returns True if there are multiple domains in the jar. Returns False otherwise. :rtype: bool """ domains = [] for cookie in iter(self): if cookie.domain is not None and cookie.domain in domains: return True domains.append(cookie.domain) return False # there is only one domain in jar def get_dict(self, domain=None, path=None): """Takes as an argument an optional domain and path and returns a plain old Python dict of name-value pairs of cookies that meet the requirements. :rtype: dict """ dictionary = {} for cookie in iter(self): if ( (domain is None or cookie.domain == domain) and (path is None or cookie.path == path) ): dictionary[cookie.name] = cookie.value return dictionary def __contains__(self, name): try: return super(RequestsCookieJar, self).__contains__(name) except CookieConflictError: return True def __getitem__(self, name): """Dict-like __getitem__() for compatibility with client code. Throws exception if there are more than one cookie with name. In that case, use the more explicit get() method instead. .. warning:: operation is O(n), not O(1). """ return self._find_no_duplicates(name) def __setitem__(self, name, value): """Dict-like __setitem__ for compatibility with client code. Throws exception if there is already a cookie of that name in the jar. In that case, use the more explicit set() method instead. """ self.set(name, value) def __delitem__(self, name): """Deletes a cookie given a name. Wraps ``cookielib.CookieJar``'s ``remove_cookie_by_name()``. """ remove_cookie_by_name(self, name) def set_cookie(self, cookie, *args, **kwargs): if hasattr(cookie.value, 'startswith') and cookie.value.startswith('"') and cookie.value.endswith('"'): cookie.value = cookie.value.replace('\\"', '') return super(RequestsCookieJar, self).set_cookie(cookie, *args, **kwargs) def update(self, other): """Updates this jar with cookies from another CookieJar or dict-like""" if isinstance(other, cookielib.CookieJar): for cookie in other: self.set_cookie(copy.copy(cookie)) else: super(RequestsCookieJar, self).update(other) def _find(self, name, domain=None, path=None): """Requests uses this method internally to get cookie values. If there are conflicting cookies, _find arbitrarily chooses one. See _find_no_duplicates if you want an exception thrown if there are conflicting cookies. :param name: a string containing name of cookie :param domain: (optional) string containing domain of cookie :param path: (optional) string containing path of cookie :return: cookie.value """ for cookie in iter(self): if cookie.name == name: if domain is None or cookie.domain == domain: if path is None or cookie.path == path: return cookie.value raise KeyError('name=%r, domain=%r, path=%r' % (name, domain, path)) def _find_no_duplicates(self, name, domain=None, path=None): """Both ``__get_item__`` and ``get`` call this function: it's never used elsewhere in Requests. :param name: a string containing name of cookie :param domain: (optional) string containing domain of cookie :param path: (optional) string containing path of cookie :raises KeyError: if cookie is not found :raises CookieConflictError: if there are multiple cookies that match name and optionally domain and path :return: cookie.value """ toReturn = None for cookie in iter(self): if cookie.name == name: if domain is None or cookie.domain == domain: if path is None or cookie.path == path: if toReturn is not None: # if there are multiple cookies that meet passed in criteria raise CookieConflictError('There are multiple cookies with name, %r' % (name)) toReturn = cookie.value # we will eventually return this as long as no cookie conflict if toReturn: return toReturn raise KeyError('name=%r, domain=%r, path=%r' % (name, domain, path)) def __getstate__(self): """Unlike a normal CookieJar, this class is pickleable.""" state = self.__dict__.copy() # remove the unpickleable RLock object state.pop('_cookies_lock') return state def __setstate__(self, state): """Unlike a normal CookieJar, this class is pickleable.""" self.__dict__.update(state) if '_cookies_lock' not in self.__dict__: self._cookies_lock = threading.RLock() def copy(self): """Return a copy of this RequestsCookieJar.""" new_cj = RequestsCookieJar() new_cj.set_policy(self.get_policy()) new_cj.update(self) return new_cj def get_policy(self): """Return the CookiePolicy instance used.""" return self._policy def _copy_cookie_jar(jar): if jar is None: return None if hasattr(jar, 'copy'): # We're dealing with an instance of RequestsCookieJar return jar.copy() # We're dealing with a generic CookieJar instance new_jar = copy.copy(jar) new_jar.clear() for cookie in jar: new_jar.set_cookie(copy.copy(cookie)) return new_jar def create_cookie(name, value, **kwargs): """Make a cookie from underspecified parameters. By default, the pair of `name` and `value` will be set for the domain '' and sent on every request (this is sometimes called a "supercookie"). """ result = { 'version': 0, 'name': name, 'value': value, 'port': None, 'domain': '', 'path': '/', 'secure': False, 'expires': None, 'discard': True, 'comment': None, 'comment_url': None, 'rest': {'HttpOnly': None}, 'rfc2109': False, } badargs = set(kwargs) - set(result) if badargs: err = 'create_cookie() got unexpected keyword arguments: %s' raise TypeError(err % list(badargs)) result.update(kwargs) result['port_specified'] = bool(result['port']) result['domain_specified'] = bool(result['domain']) result['domain_initial_dot'] = result['domain'].startswith('.') result['path_specified'] = bool(result['path']) return cookielib.Cookie(**result) def morsel_to_cookie(morsel): """Convert a Morsel object into a Cookie containing the one k/v pair.""" expires = None if morsel['max-age']: try: expires = int(time.time() + int(morsel['max-age'])) except ValueError: raise TypeError('max-age: %s must be integer' % morsel['max-age']) elif morsel['expires']: time_template = '%a, %d-%b-%Y %H:%M:%S GMT' expires = calendar.timegm( time.strptime(morsel['expires'], time_template) ) return create_cookie( comment=morsel['comment'], comment_url=bool(morsel['comment']), discard=False, domain=morsel['domain'], expires=expires, name=morsel.key, path=morsel['path'], port=None, rest={'HttpOnly': morsel['httponly']}, rfc2109=False, secure=bool(morsel['secure']), value=morsel.value, version=morsel['version'] or 0, ) def cookiejar_from_dict(cookie_dict, cookiejar=None, overwrite=True): """Returns a CookieJar from a key/value dictionary. :param cookie_dict: Dict of key/values to insert into CookieJar. :param cookiejar: (optional) A cookiejar to add the cookies to. :param overwrite: (optional) If False, will not replace cookies already in the jar with new ones. :rtype: CookieJar """ if cookiejar is None: cookiejar = RequestsCookieJar() if cookie_dict is not None: names_from_jar = [cookie.name for cookie in cookiejar] for name in cookie_dict: if overwrite or (name not in names_from_jar): cookiejar.set_cookie(create_cookie(name, cookie_dict[name])) return cookiejar def merge_cookies(cookiejar, cookies): """Add cookies to cookiejar and returns a merged CookieJar. :param cookiejar: CookieJar object to add the cookies to. :param cookies: Dictionary or CookieJar object to be added. :rtype: CookieJar """ if not isinstance(cookiejar, cookielib.CookieJar): raise ValueError('You can only merge into CookieJar') if isinstance(cookies, dict): cookiejar = cookiejar_from_dict( cookies, cookiejar=cookiejar, overwrite=False) elif isinstance(cookies, cookielib.CookieJar): try: cookiejar.update(cookies) except AttributeError: for cookie_in_jar in cookies: cookiejar.set_cookie(cookie_in_jar) return cookiejar site-packages/pip/_vendor/requests/packages.py000064400000001267151464452720015533 0ustar00import sys # This code exists for backwards compatibility reasons. # I don't like it either. Just look the other way. :) for package in ('urllib3', 'idna', 'chardet'): vendored_package = "pip._vendor." + package locals()[package] = __import__(vendored_package) # This traversal is apparently necessary such that the identities are # preserved (requests.packages.urllib3.* is urllib3.*) for mod in list(sys.modules): if mod == vendored_package or mod.startswith(vendored_package + '.'): unprefixed_mod = mod[len("pip._vendor."):] sys.modules['pip._vendor.requests.packages.' + unprefixed_mod] = sys.modules[mod] # Kinda cool, though, right? site-packages/pip/_vendor/requests/api.py000064400000014177151464452720014532 0ustar00# -*- coding: utf-8 -*- """ requests.api ~~~~~~~~~~~~ This module implements the Requests API. :copyright: (c) 2012 by Kenneth Reitz. :license: Apache2, see LICENSE for more details. """ from . import sessions def request(method, url, **kwargs): """Constructs and sends a :class:`Request `. :param method: method for the new :class:`Request` object. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary, list of tuples or bytes to send in the query string for the :class:`Request`. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) A JSON serializable Python object to send in the body of the :class:`Request`. :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`. :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`. :param files: (optional) Dictionary of ``'name': file-like-objects`` (or ``{'name': file-tuple}``) for multipart encoding upload. ``file-tuple`` can be a 2-tuple ``('filename', fileobj)``, 3-tuple ``('filename', fileobj, 'content_type')`` or a 4-tuple ``('filename', fileobj, 'content_type', custom_headers)``, where ``'content-type'`` is a string defining the content type of the given file and ``custom_headers`` a dict-like object containing additional headers to add for the file. :param auth: (optional) Auth tuple to enable Basic/Digest/Custom HTTP Auth. :param timeout: (optional) How many seconds to wait for the server to send data before giving up, as a float, or a :ref:`(connect timeout, read timeout) ` tuple. :type timeout: float or tuple :param allow_redirects: (optional) Boolean. Enable/disable GET/OPTIONS/POST/PUT/PATCH/DELETE/HEAD redirection. Defaults to ``True``. :type allow_redirects: bool :param proxies: (optional) Dictionary mapping protocol to the URL of the proxy. :param verify: (optional) Either a boolean, in which case it controls whether we verify the server's TLS certificate, or a string, in which case it must be a path to a CA bundle to use. Defaults to ``True``. :param stream: (optional) if ``False``, the response content will be immediately downloaded. :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair. :return: :class:`Response ` object :rtype: requests.Response Usage:: >>> import requests >>> req = requests.request('GET', 'https://httpbin.org/get') """ # By using the 'with' statement we are sure the session is closed, thus we # avoid leaving sockets open which can trigger a ResourceWarning in some # cases, and look like a memory leak in others. with sessions.Session() as session: return session.request(method=method, url=url, **kwargs) def get(url, params=None, **kwargs): r"""Sends a GET request. :param url: URL for the new :class:`Request` object. :param params: (optional) Dictionary, list of tuples or bytes to send in the query string for the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ kwargs.setdefault('allow_redirects', True) return request('get', url, params=params, **kwargs) def options(url, **kwargs): r"""Sends an OPTIONS request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ kwargs.setdefault('allow_redirects', True) return request('options', url, **kwargs) def head(url, **kwargs): r"""Sends a HEAD request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ kwargs.setdefault('allow_redirects', False) return request('head', url, **kwargs) def post(url, data=None, json=None, **kwargs): r"""Sends a POST request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ return request('post', url, data=data, json=json, **kwargs) def put(url, data=None, **kwargs): r"""Sends a PUT request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ return request('put', url, data=data, **kwargs) def patch(url, data=None, **kwargs): r"""Sends a PATCH request. :param url: URL for the new :class:`Request` object. :param data: (optional) Dictionary, list of tuples, bytes, or file-like object to send in the body of the :class:`Request`. :param json: (optional) json data to send in the body of the :class:`Request`. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ return request('patch', url, data=data, **kwargs) def delete(url, **kwargs): r"""Sends a DELETE request. :param url: URL for the new :class:`Request` object. :param \*\*kwargs: Optional arguments that ``request`` takes. :return: :class:`Response ` object :rtype: requests.Response """ return request('delete', url, **kwargs) site-packages/pip/_vendor/requests/structures.py000064400000005645151464452720016204 0ustar00# -*- coding: utf-8 -*- """ requests.structures ~~~~~~~~~~~~~~~~~~~ Data structures that power Requests. """ from .compat import OrderedDict, Mapping, MutableMapping class CaseInsensitiveDict(MutableMapping): """A case-insensitive ``dict``-like object. Implements all methods and operations of ``MutableMapping`` as well as dict's ``copy``. Also provides ``lower_items``. All keys are expected to be strings. The structure remembers the case of the last key to be set, and ``iter(instance)``, ``keys()``, ``items()``, ``iterkeys()``, and ``iteritems()`` will contain case-sensitive keys. However, querying and contains testing is case insensitive:: cid = CaseInsensitiveDict() cid['Accept'] = 'application/json' cid['aCCEPT'] == 'application/json' # True list(cid) == ['Accept'] # True For example, ``headers['content-encoding']`` will return the value of a ``'Content-Encoding'`` response header, regardless of how the header name was originally stored. If the constructor, ``.update``, or equality comparison operations are given keys that have equal ``.lower()``s, the behavior is undefined. """ def __init__(self, data=None, **kwargs): self._store = OrderedDict() if data is None: data = {} self.update(data, **kwargs) def __setitem__(self, key, value): # Use the lowercased key for lookups, but store the actual # key alongside the value. self._store[key.lower()] = (key, value) def __getitem__(self, key): return self._store[key.lower()][1] def __delitem__(self, key): del self._store[key.lower()] def __iter__(self): return (casedkey for casedkey, mappedvalue in self._store.values()) def __len__(self): return len(self._store) def lower_items(self): """Like iteritems(), but with all lowercase keys.""" return ( (lowerkey, keyval[1]) for (lowerkey, keyval) in self._store.items() ) def __eq__(self, other): if isinstance(other, Mapping): other = CaseInsensitiveDict(other) else: return NotImplemented # Compare insensitively return dict(self.lower_items()) == dict(other.lower_items()) # Copy is required def copy(self): return CaseInsensitiveDict(self._store.values()) def __repr__(self): return str(dict(self.items())) class LookupDict(dict): """Dictionary lookup object.""" def __init__(self, name=None): self.name = name super(LookupDict, self).__init__() def __repr__(self): return '' % (self.name) def __getitem__(self, key): # We allow fall-through here, so values default to None return self.__dict__.get(key, None) def get(self, key, default=None): return self.__dict__.get(key, default) site-packages/pip/_vendor/requests/compat.py000064400000003625151464452720015240 0ustar00# -*- coding: utf-8 -*- """ requests.compat ~~~~~~~~~~~~~~~ This module handles import compatibility issues between Python 2 and Python 3. """ from pip._vendor import chardet import sys # ------- # Pythons # ------- # Syntax sugar. _ver = sys.version_info #: Python 2.x? is_py2 = (_ver[0] == 2) #: Python 3.x? is_py3 = (_ver[0] == 3) # Note: We've patched out simplejson support in pip because it prevents # upgrading simplejson on Windows. # try: # import simplejson as json # except (ImportError, SyntaxError): # # simplejson does not support Python 3.2, it throws a SyntaxError # # because of u'...' Unicode literals. import json # --------- # Specifics # --------- if is_py2: from urllib import ( quote, unquote, quote_plus, unquote_plus, urlencode, getproxies, proxy_bypass, proxy_bypass_environment, getproxies_environment) from urlparse import urlparse, urlunparse, urljoin, urlsplit, urldefrag from urllib2 import parse_http_list import cookielib from Cookie import Morsel from StringIO import StringIO from collections import Callable, Mapping, MutableMapping, OrderedDict builtin_str = str bytes = str str = unicode basestring = basestring numeric_types = (int, long, float) integer_types = (int, long) elif is_py3: from urllib.parse import urlparse, urlunparse, urljoin, urlsplit, urlencode, quote, unquote, quote_plus, unquote_plus, urldefrag from urllib.request import parse_http_list, getproxies, proxy_bypass, proxy_bypass_environment, getproxies_environment from http import cookiejar as cookielib from http.cookies import Morsel from io import StringIO from collections import OrderedDict from collections.abc import Callable, Mapping, MutableMapping builtin_str = str str = str bytes = bytes basestring = (str, bytes) numeric_types = (int, float) integer_types = (int,) site-packages/pip/_vendor/requests/help.py000064400000006772151464452720014713 0ustar00"""Module containing bug report helper(s).""" from __future__ import print_function import json import platform import sys import ssl from pip._vendor import idna from pip._vendor import urllib3 from pip._vendor import chardet from . import __version__ as requests_version try: from pip._vendor.urllib3.contrib import pyopenssl except ImportError: pyopenssl = None OpenSSL = None cryptography = None else: import OpenSSL import cryptography def _implementation(): """Return a dict with the Python implementation and version. Provide both the name and the version of the Python implementation currently running. For example, on CPython 2.7.5 it will return {'name': 'CPython', 'version': '2.7.5'}. This function works best on CPython and PyPy: in particular, it probably doesn't work for Jython or IronPython. Future investigation should be done to work out the correct shape of the code for those platforms. """ implementation = platform.python_implementation() if implementation == 'CPython': implementation_version = platform.python_version() elif implementation == 'PyPy': implementation_version = '%s.%s.%s' % (sys.pypy_version_info.major, sys.pypy_version_info.minor, sys.pypy_version_info.micro) if sys.pypy_version_info.releaselevel != 'final': implementation_version = ''.join([ implementation_version, sys.pypy_version_info.releaselevel ]) elif implementation == 'Jython': implementation_version = platform.python_version() # Complete Guess elif implementation == 'IronPython': implementation_version = platform.python_version() # Complete Guess else: implementation_version = 'Unknown' return {'name': implementation, 'version': implementation_version} def info(): """Generate information for a bug report.""" try: platform_info = { 'system': platform.system(), 'release': platform.release(), } except IOError: platform_info = { 'system': 'Unknown', 'release': 'Unknown', } implementation_info = _implementation() urllib3_info = {'version': urllib3.__version__} chardet_info = {'version': chardet.__version__} pyopenssl_info = { 'version': None, 'openssl_version': '', } if OpenSSL: pyopenssl_info = { 'version': OpenSSL.__version__, 'openssl_version': '%x' % OpenSSL.SSL.OPENSSL_VERSION_NUMBER, } cryptography_info = { 'version': getattr(cryptography, '__version__', ''), } idna_info = { 'version': getattr(idna, '__version__', ''), } system_ssl = ssl.OPENSSL_VERSION_NUMBER system_ssl_info = { 'version': '%x' % system_ssl if system_ssl is not None else '' } return { 'platform': platform_info, 'implementation': implementation_info, 'system_ssl': system_ssl_info, 'using_pyopenssl': pyopenssl is not None, 'pyOpenSSL': pyopenssl_info, 'urllib3': urllib3_info, 'chardet': chardet_info, 'cryptography': cryptography_info, 'idna': idna_info, 'requests': { 'version': requests_version, }, } def main(): """Pretty-print the bug information as JSON.""" print(json.dumps(info(), sort_keys=True, indent=2)) if __name__ == '__main__': main() site-packages/pip/_vendor/requests/hooks.py000064400000001365151464452720015077 0ustar00# -*- coding: utf-8 -*- """ requests.hooks ~~~~~~~~~~~~~~ This module provides the capabilities for the Requests hooks system. Available hooks: ``response``: The response generated from a Request. """ HOOKS = ['response'] def default_hooks(): return {event: [] for event in HOOKS} # TODO: response is the only one def dispatch_hook(key, hooks, hook_data, **kwargs): """Dispatches a hook dictionary on a given piece of data.""" hooks = hooks or {} hooks = hooks.get(key) if hooks: if hasattr(hooks, '__call__'): hooks = [hooks] for hook in hooks: _hook_data = hook(hook_data, **kwargs) if _hook_data is not None: hook_data = _hook_data return hook_data site-packages/pip/_vendor/requests/certs.py000064400000000721151464452720015067 0ustar00#!/usr/bin/env python # -*- coding: utf-8 -*- """ requests.certs ~~~~~~~~~~~~~~ This module returns the preferred default CA certificate bundle. There is only one — the one from the certifi package. If you are packaging Requests, e.g., for a Linux distribution or a managed environment, you can change the definition of where() to return a separately packaged CA bundle. """ from pip._vendor.certifi import where if __name__ == '__main__': print(where()) site-packages/pip/_vendor/requests/__init__.py000064400000007752151464452720015521 0ustar00# -*- coding: utf-8 -*- # __ # /__) _ _ _ _ _/ _ # / ( (- (/ (/ (- _) / _) # / """ Requests HTTP Library ~~~~~~~~~~~~~~~~~~~~~ Requests is an HTTP library, written in Python, for human beings. Basic GET usage: >>> import requests >>> r = requests.get('https://www.python.org') >>> r.status_code 200 >>> 'Python is a programming language' in r.content True ... or POST: >>> payload = dict(key1='value1', key2='value2') >>> r = requests.post('https://httpbin.org/post', data=payload) >>> print(r.text) { ... "form": { "key2": "value2", "key1": "value1" }, ... } The other HTTP methods are supported - see `requests.api`. Full documentation is at . :copyright: (c) 2017 by Kenneth Reitz. :license: Apache 2.0, see LICENSE for more details. """ from pip._vendor import urllib3 from pip._vendor import chardet import warnings from .exceptions import RequestsDependencyWarning def check_compatibility(urllib3_version, chardet_version): urllib3_version = urllib3_version.split('.') assert urllib3_version != ['dev'] # Verify urllib3 isn't installed from git. # Sometimes, urllib3 only reports its version as 16.1. if len(urllib3_version) == 2: urllib3_version.append('0') # Check urllib3 for compatibility. major, minor, patch = urllib3_version # noqa: F811 major, minor, patch = int(major), int(minor), int(patch) # urllib3 >= 1.21.1, <= 1.25 assert major == 1 assert minor >= 21 assert minor <= 25 # Check chardet for compatibility. major, minor, patch = chardet_version.split('.')[:3] major, minor, patch = int(major), int(minor), int(patch) # chardet >= 3.0.2, < 3.1.0 assert major == 3 assert minor < 1 assert patch >= 2 def _check_cryptography(cryptography_version): # cryptography < 1.3.4 try: cryptography_version = list(map(int, cryptography_version.split('.'))) except ValueError: return if cryptography_version < [1, 3, 4]: warning = 'Old version of cryptography ({}) may cause slowdown.'.format(cryptography_version) warnings.warn(warning, RequestsDependencyWarning) # Check imported dependencies for compatibility. try: check_compatibility(urllib3.__version__, chardet.__version__) except (AssertionError, ValueError): warnings.warn("urllib3 ({}) or chardet ({}) doesn't match a supported " "version!".format(urllib3.__version__, chardet.__version__), RequestsDependencyWarning) # Attempt to enable urllib3's SNI support, if possible from pip._internal.utils.compat import WINDOWS if not WINDOWS: try: from pip._vendor.urllib3.contrib import pyopenssl pyopenssl.inject_into_urllib3() # Check cryptography version from cryptography import __version__ as cryptography_version _check_cryptography(cryptography_version) except ImportError: pass # urllib3's DependencyWarnings should be silenced. from pip._vendor.urllib3.exceptions import DependencyWarning warnings.simplefilter('ignore', DependencyWarning) from .__version__ import __title__, __description__, __url__, __version__ from .__version__ import __build__, __author__, __author_email__, __license__ from .__version__ import __copyright__, __cake__ from . import utils from . import packages from .models import Request, Response, PreparedRequest from .api import request, get, head, post, patch, put, delete, options from .sessions import session, Session from .status_codes import codes from .exceptions import ( RequestException, Timeout, URLRequired, TooManyRedirects, HTTPError, ConnectionError, FileModeWarning, ConnectTimeout, ReadTimeout ) # Set default logging handler to avoid "No handler found" warnings. import logging from logging import NullHandler logging.getLogger(__name__).addHandler(NullHandler()) # FileModeWarnings go off per the default. warnings.simplefilter('default', FileModeWarning, append=True) site-packages/pip/_vendor/requests/__version__.py000064400000000664151464452720016236 0ustar00# .-. .-. .-. . . .-. .-. .-. .-. # |( |- |.| | | |- `-. | `-. # ' ' `-' `-`.`-' `-' `-' ' `-' __title__ = 'requests' __description__ = 'Python HTTP for Humans.' __url__ = 'http://python-requests.org' __version__ = '2.22.0' __build__ = 0x022200 __author__ = 'Kenneth Reitz' __author_email__ = 'me@kennethreitz.org' __license__ = 'Apache 2.0' __copyright__ = 'Copyright 2019 Kenneth Reitz' __cake__ = u'\u2728 \U0001f370 \u2728' site-packages/pip/_vendor/distro.py000064400000124363151464452720013411 0ustar00# Copyright 2015,2016,2017 Nir Cohen # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ The ``distro`` package (``distro`` stands for Linux Distribution) provides information about the Linux distribution it runs on, such as a reliable machine-readable distro ID, or version information. It is the recommended replacement for Python's original :py:func:`platform.linux_distribution` function, but it provides much more functionality. An alternative implementation became necessary because Python 3.5 deprecated this function, and Python 3.8 will remove it altogether. Its predecessor function :py:func:`platform.dist` was already deprecated since Python 2.6 and will also be removed in Python 3.8. Still, there are many cases in which access to OS distribution information is needed. See `Python issue 1322 `_ for more information. """ import os import re import sys import json import shlex import logging import argparse import subprocess _UNIXCONFDIR = os.environ.get('UNIXCONFDIR', '/etc') _OS_RELEASE_BASENAME = 'os-release' #: Translation table for normalizing the "ID" attribute defined in os-release #: files, for use by the :func:`distro.id` method. #: #: * Key: Value as defined in the os-release file, translated to lower case, #: with blanks translated to underscores. #: #: * Value: Normalized value. NORMALIZED_OS_ID = { 'ol': 'oracle', # Oracle Enterprise Linux } #: Translation table for normalizing the "Distributor ID" attribute returned by #: the lsb_release command, for use by the :func:`distro.id` method. #: #: * Key: Value as returned by the lsb_release command, translated to lower #: case, with blanks translated to underscores. #: #: * Value: Normalized value. NORMALIZED_LSB_ID = { 'enterpriseenterprise': 'oracle', # Oracle Enterprise Linux 'redhatenterpriseworkstation': 'rhel', # RHEL 6, 7 Workstation 'redhatenterpriseserver': 'rhel', # RHEL 6, 7 Server } #: Translation table for normalizing the distro ID derived from the file name #: of distro release files, for use by the :func:`distro.id` method. #: #: * Key: Value as derived from the file name of a distro release file, #: translated to lower case, with blanks translated to underscores. #: #: * Value: Normalized value. NORMALIZED_DISTRO_ID = { 'redhat': 'rhel', # RHEL 6.x, 7.x } # Pattern for content of distro release file (reversed) _DISTRO_RELEASE_CONTENT_REVERSED_PATTERN = re.compile( r'(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)') # Pattern for base file name of distro release file _DISTRO_RELEASE_BASENAME_PATTERN = re.compile( r'(\w+)[-_](release|version)$') # Base file names to be ignored when searching for distro release file _DISTRO_RELEASE_IGNORE_BASENAMES = ( 'debian_version', 'lsb-release', 'oem-release', _OS_RELEASE_BASENAME, 'system-release' ) def linux_distribution(full_distribution_name=True): """ Return information about the current OS distribution as a tuple ``(id_name, version, codename)`` with items as follows: * ``id_name``: If *full_distribution_name* is false, the result of :func:`distro.id`. Otherwise, the result of :func:`distro.name`. * ``version``: The result of :func:`distro.version`. * ``codename``: The result of :func:`distro.codename`. The interface of this function is compatible with the original :py:func:`platform.linux_distribution` function, supporting a subset of its parameters. The data it returns may not exactly be the same, because it uses more data sources than the original function, and that may lead to different data if the OS distribution is not consistent across multiple data sources it provides (there are indeed such distributions ...). Another reason for differences is the fact that the :func:`distro.id` method normalizes the distro ID string to a reliable machine-readable value for a number of popular OS distributions. """ return _distro.linux_distribution(full_distribution_name) def id(): """ Return the distro ID of the current distribution, as a machine-readable string. For a number of OS distributions, the returned distro ID value is *reliable*, in the sense that it is documented and that it does not change across releases of the distribution. This package maintains the following reliable distro ID values: ============== ========================================= Distro ID Distribution ============== ========================================= "ubuntu" Ubuntu "debian" Debian "rhel" RedHat Enterprise Linux "centos" CentOS "fedora" Fedora "sles" SUSE Linux Enterprise Server "opensuse" openSUSE "amazon" Amazon Linux "arch" Arch Linux "cloudlinux" CloudLinux OS "exherbo" Exherbo Linux "gentoo" GenToo Linux "ibm_powerkvm" IBM PowerKVM "kvmibm" KVM for IBM z Systems "linuxmint" Linux Mint "mageia" Mageia "mandriva" Mandriva Linux "parallels" Parallels "pidora" Pidora "raspbian" Raspbian "oracle" Oracle Linux (and Oracle Enterprise Linux) "scientific" Scientific Linux "slackware" Slackware "xenserver" XenServer "openbsd" OpenBSD "netbsd" NetBSD "freebsd" FreeBSD ============== ========================================= If you have a need to get distros for reliable IDs added into this set, or if you find that the :func:`distro.id` function returns a different distro ID for one of the listed distros, please create an issue in the `distro issue tracker`_. **Lookup hierarchy and transformations:** First, the ID is obtained from the following sources, in the specified order. The first available and non-empty value is used: * the value of the "ID" attribute of the os-release file, * the value of the "Distributor ID" attribute returned by the lsb_release command, * the first part of the file name of the distro release file, The so determined ID value then passes the following transformations, before it is returned by this method: * it is translated to lower case, * blanks (which should not be there anyway) are translated to underscores, * a normalization of the ID is performed, based upon `normalization tables`_. The purpose of this normalization is to ensure that the ID is as reliable as possible, even across incompatible changes in the OS distributions. A common reason for an incompatible change is the addition of an os-release file, or the addition of the lsb_release command, with ID values that differ from what was previously determined from the distro release file name. """ return _distro.id() def name(pretty=False): """ Return the name of the current OS distribution, as a human-readable string. If *pretty* is false, the name is returned without version or codename. (e.g. "CentOS Linux") If *pretty* is true, the version and codename are appended. (e.g. "CentOS Linux 7.1.1503 (Core)") **Lookup hierarchy:** The name is obtained from the following sources, in the specified order. The first available and non-empty value is used: * If *pretty* is false: - the value of the "NAME" attribute of the os-release file, - the value of the "Distributor ID" attribute returned by the lsb_release command, - the value of the "" field of the distro release file. * If *pretty* is true: - the value of the "PRETTY_NAME" attribute of the os-release file, - the value of the "Description" attribute returned by the lsb_release command, - the value of the "" field of the distro release file, appended with the value of the pretty version ("" and "" fields) of the distro release file, if available. """ return _distro.name(pretty) def version(pretty=False, best=False): """ Return the version of the current OS distribution, as a human-readable string. If *pretty* is false, the version is returned without codename (e.g. "7.0"). If *pretty* is true, the codename in parenthesis is appended, if the codename is non-empty (e.g. "7.0 (Maipo)"). Some distributions provide version numbers with different precisions in the different sources of distribution information. Examining the different sources in a fixed priority order does not always yield the most precise version (e.g. for Debian 8.2, or CentOS 7.1). The *best* parameter can be used to control the approach for the returned version: If *best* is false, the first non-empty version number in priority order of the examined sources is returned. If *best* is true, the most precise version number out of all examined sources is returned. **Lookup hierarchy:** In all cases, the version number is obtained from the following sources. If *best* is false, this order represents the priority order: * the value of the "VERSION_ID" attribute of the os-release file, * the value of the "Release" attribute returned by the lsb_release command, * the version number parsed from the "" field of the first line of the distro release file, * the version number parsed from the "PRETTY_NAME" attribute of the os-release file, if it follows the format of the distro release files. * the version number parsed from the "Description" attribute returned by the lsb_release command, if it follows the format of the distro release files. """ return _distro.version(pretty, best) def version_parts(best=False): """ Return the version of the current OS distribution as a tuple ``(major, minor, build_number)`` with items as follows: * ``major``: The result of :func:`distro.major_version`. * ``minor``: The result of :func:`distro.minor_version`. * ``build_number``: The result of :func:`distro.build_number`. For a description of the *best* parameter, see the :func:`distro.version` method. """ return _distro.version_parts(best) def major_version(best=False): """ Return the major version of the current OS distribution, as a string, if provided. Otherwise, the empty string is returned. The major version is the first part of the dot-separated version string. For a description of the *best* parameter, see the :func:`distro.version` method. """ return _distro.major_version(best) def minor_version(best=False): """ Return the minor version of the current OS distribution, as a string, if provided. Otherwise, the empty string is returned. The minor version is the second part of the dot-separated version string. For a description of the *best* parameter, see the :func:`distro.version` method. """ return _distro.minor_version(best) def build_number(best=False): """ Return the build number of the current OS distribution, as a string, if provided. Otherwise, the empty string is returned. The build number is the third part of the dot-separated version string. For a description of the *best* parameter, see the :func:`distro.version` method. """ return _distro.build_number(best) def like(): """ Return a space-separated list of distro IDs of distributions that are closely related to the current OS distribution in regards to packaging and programming interfaces, for example distributions the current distribution is a derivative from. **Lookup hierarchy:** This information item is only provided by the os-release file. For details, see the description of the "ID_LIKE" attribute in the `os-release man page `_. """ return _distro.like() def codename(): """ Return the codename for the release of the current OS distribution, as a string. If the distribution does not have a codename, an empty string is returned. Note that the returned codename is not always really a codename. For example, openSUSE returns "x86_64". This function does not handle such cases in any special way and just returns the string it finds, if any. **Lookup hierarchy:** * the codename within the "VERSION" attribute of the os-release file, if provided, * the value of the "Codename" attribute returned by the lsb_release command, * the value of the "" field of the distro release file. """ return _distro.codename() def info(pretty=False, best=False): """ Return certain machine-readable information items about the current OS distribution in a dictionary, as shown in the following example: .. sourcecode:: python { 'id': 'rhel', 'version': '7.0', 'version_parts': { 'major': '7', 'minor': '0', 'build_number': '' }, 'like': 'fedora', 'codename': 'Maipo' } The dictionary structure and keys are always the same, regardless of which information items are available in the underlying data sources. The values for the various keys are as follows: * ``id``: The result of :func:`distro.id`. * ``version``: The result of :func:`distro.version`. * ``version_parts -> major``: The result of :func:`distro.major_version`. * ``version_parts -> minor``: The result of :func:`distro.minor_version`. * ``version_parts -> build_number``: The result of :func:`distro.build_number`. * ``like``: The result of :func:`distro.like`. * ``codename``: The result of :func:`distro.codename`. For a description of the *pretty* and *best* parameters, see the :func:`distro.version` method. """ return _distro.info(pretty, best) def os_release_info(): """ Return a dictionary containing key-value pairs for the information items from the os-release file data source of the current OS distribution. See `os-release file`_ for details about these information items. """ return _distro.os_release_info() def lsb_release_info(): """ Return a dictionary containing key-value pairs for the information items from the lsb_release command data source of the current OS distribution. See `lsb_release command output`_ for details about these information items. """ return _distro.lsb_release_info() def distro_release_info(): """ Return a dictionary containing key-value pairs for the information items from the distro release file data source of the current OS distribution. See `distro release file`_ for details about these information items. """ return _distro.distro_release_info() def uname_info(): """ Return a dictionary containing key-value pairs for the information items from the distro release file data source of the current OS distribution. """ return _distro.uname_info() def os_release_attr(attribute): """ Return a single named information item from the os-release file data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. See `os-release file`_ for details about these information items. """ return _distro.os_release_attr(attribute) def lsb_release_attr(attribute): """ Return a single named information item from the lsb_release command output data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. See `lsb_release command output`_ for details about these information items. """ return _distro.lsb_release_attr(attribute) def distro_release_attr(attribute): """ Return a single named information item from the distro release file data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. See `distro release file`_ for details about these information items. """ return _distro.distro_release_attr(attribute) def uname_attr(attribute): """ Return a single named information item from the distro release file data source of the current OS distribution. Parameters: * ``attribute`` (string): Key of the information item. Returns: * (string): Value of the information item, if the item exists. The empty string, if the item does not exist. """ return _distro.uname_attr(attribute) class cached_property(object): """A version of @property which caches the value. On access, it calls the underlying function and sets the value in `__dict__` so future accesses will not re-call the property. """ def __init__(self, f): self._fname = f.__name__ self._f = f def __get__(self, obj, owner): assert obj is not None, 'call {} on an instance'.format(self._fname) ret = obj.__dict__[self._fname] = self._f(obj) return ret class LinuxDistribution(object): """ Provides information about a OS distribution. This package creates a private module-global instance of this class with default initialization arguments, that is used by the `consolidated accessor functions`_ and `single source accessor functions`_. By using default initialization arguments, that module-global instance returns data about the current OS distribution (i.e. the distro this package runs on). Normally, it is not necessary to create additional instances of this class. However, in situations where control is needed over the exact data sources that are used, instances of this class can be created with a specific distro release file, or a specific os-release file, or without invoking the lsb_release command. """ def __init__(self, include_lsb=True, os_release_file='', distro_release_file='', include_uname=True): """ The initialization method of this class gathers information from the available data sources, and stores that in private instance attributes. Subsequent access to the information items uses these private instance attributes, so that the data sources are read only once. Parameters: * ``include_lsb`` (bool): Controls whether the `lsb_release command output`_ is included as a data source. If the lsb_release command is not available in the program execution path, the data source for the lsb_release command will be empty. * ``os_release_file`` (string): The path name of the `os-release file`_ that is to be used as a data source. An empty string (the default) will cause the default path name to be used (see `os-release file`_ for details). If the specified or defaulted os-release file does not exist, the data source for the os-release file will be empty. * ``distro_release_file`` (string): The path name of the `distro release file`_ that is to be used as a data source. An empty string (the default) will cause a default search algorithm to be used (see `distro release file`_ for details). If the specified distro release file does not exist, or if no default distro release file can be found, the data source for the distro release file will be empty. * ``include_name`` (bool): Controls whether uname command output is included as a data source. If the uname command is not available in the program execution path the data source for the uname command will be empty. Public instance attributes: * ``os_release_file`` (string): The path name of the `os-release file`_ that is actually used as a data source. The empty string if no distro release file is used as a data source. * ``distro_release_file`` (string): The path name of the `distro release file`_ that is actually used as a data source. The empty string if no distro release file is used as a data source. * ``include_lsb`` (bool): The result of the ``include_lsb`` parameter. This controls whether the lsb information will be loaded. * ``include_uname`` (bool): The result of the ``include_uname`` parameter. This controls whether the uname information will be loaded. Raises: * :py:exc:`IOError`: Some I/O issue with an os-release file or distro release file. * :py:exc:`subprocess.CalledProcessError`: The lsb_release command had some issue (other than not being available in the program execution path). * :py:exc:`UnicodeError`: A data source has unexpected characters or uses an unexpected encoding. """ self.os_release_file = os_release_file or \ os.path.join(_UNIXCONFDIR, _OS_RELEASE_BASENAME) self.distro_release_file = distro_release_file or '' # updated later self.include_lsb = include_lsb self.include_uname = include_uname def __repr__(self): """Return repr of all info """ return \ "LinuxDistribution(" \ "os_release_file={self.os_release_file!r}, " \ "distro_release_file={self.distro_release_file!r}, " \ "include_lsb={self.include_lsb!r}, " \ "include_uname={self.include_uname!r}, " \ "_os_release_info={self._os_release_info!r}, " \ "_lsb_release_info={self._lsb_release_info!r}, " \ "_distro_release_info={self._distro_release_info!r}, " \ "_uname_info={self._uname_info!r})".format( self=self) def linux_distribution(self, full_distribution_name=True): """ Return information about the OS distribution that is compatible with Python's :func:`platform.linux_distribution`, supporting a subset of its parameters. For details, see :func:`distro.linux_distribution`. """ return ( self.name() if full_distribution_name else self.id(), self.version(), self.codename() ) def id(self): """Return the distro ID of the OS distribution, as a string. For details, see :func:`distro.id`. """ def normalize(distro_id, table): distro_id = distro_id.lower().replace(' ', '_') return table.get(distro_id, distro_id) distro_id = self.os_release_attr('id') if distro_id: return normalize(distro_id, NORMALIZED_OS_ID) distro_id = self.lsb_release_attr('distributor_id') if distro_id: return normalize(distro_id, NORMALIZED_LSB_ID) distro_id = self.distro_release_attr('id') if distro_id: return normalize(distro_id, NORMALIZED_DISTRO_ID) distro_id = self.uname_attr('id') if distro_id: return normalize(distro_id, NORMALIZED_DISTRO_ID) return '' def name(self, pretty=False): """ Return the name of the OS distribution, as a string. For details, see :func:`distro.name`. """ name = self.os_release_attr('name') \ or self.lsb_release_attr('distributor_id') \ or self.distro_release_attr('name') \ or self.uname_attr('name') if pretty: name = self.os_release_attr('pretty_name') \ or self.lsb_release_attr('description') if not name: name = self.distro_release_attr('name') \ or self.uname_attr('name') version = self.version(pretty=True) if version: name = name + ' ' + version return name or '' def version(self, pretty=False, best=False): """ Return the version of the OS distribution, as a string. For details, see :func:`distro.version`. """ versions = [ self.os_release_attr('version_id'), self.lsb_release_attr('release'), self.distro_release_attr('version_id'), self._parse_distro_release_content( self.os_release_attr('pretty_name')).get('version_id', ''), self._parse_distro_release_content( self.lsb_release_attr('description')).get('version_id', ''), self.uname_attr('release') ] version = '' if best: # This algorithm uses the last version in priority order that has # the best precision. If the versions are not in conflict, that # does not matter; otherwise, using the last one instead of the # first one might be considered a surprise. for v in versions: if v.count(".") > version.count(".") or version == '': version = v else: for v in versions: if v != '': version = v break if pretty and version and self.codename(): version = u'{0} ({1})'.format(version, self.codename()) return version def version_parts(self, best=False): """ Return the version of the OS distribution, as a tuple of version numbers. For details, see :func:`distro.version_parts`. """ version_str = self.version(best=best) if version_str: version_regex = re.compile(r'(\d+)\.?(\d+)?\.?(\d+)?') matches = version_regex.match(version_str) if matches: major, minor, build_number = matches.groups() return major, minor or '', build_number or '' return '', '', '' def major_version(self, best=False): """ Return the major version number of the current distribution. For details, see :func:`distro.major_version`. """ return self.version_parts(best)[0] def minor_version(self, best=False): """ Return the minor version number of the current distribution. For details, see :func:`distro.minor_version`. """ return self.version_parts(best)[1] def build_number(self, best=False): """ Return the build number of the current distribution. For details, see :func:`distro.build_number`. """ return self.version_parts(best)[2] def like(self): """ Return the IDs of distributions that are like the OS distribution. For details, see :func:`distro.like`. """ return self.os_release_attr('id_like') or '' def codename(self): """ Return the codename of the OS distribution. For details, see :func:`distro.codename`. """ try: # Handle os_release specially since distros might purposefully set # this to empty string to have no codename return self._os_release_info['codename'] except KeyError: return self.lsb_release_attr('codename') \ or self.distro_release_attr('codename') \ or '' def info(self, pretty=False, best=False): """ Return certain machine-readable information about the OS distribution. For details, see :func:`distro.info`. """ return dict( id=self.id(), version=self.version(pretty, best), version_parts=dict( major=self.major_version(best), minor=self.minor_version(best), build_number=self.build_number(best) ), like=self.like(), codename=self.codename(), ) def os_release_info(self): """ Return a dictionary containing key-value pairs for the information items from the os-release file data source of the OS distribution. For details, see :func:`distro.os_release_info`. """ return self._os_release_info def lsb_release_info(self): """ Return a dictionary containing key-value pairs for the information items from the lsb_release command data source of the OS distribution. For details, see :func:`distro.lsb_release_info`. """ return self._lsb_release_info def distro_release_info(self): """ Return a dictionary containing key-value pairs for the information items from the distro release file data source of the OS distribution. For details, see :func:`distro.distro_release_info`. """ return self._distro_release_info def uname_info(self): """ Return a dictionary containing key-value pairs for the information items from the uname command data source of the OS distribution. For details, see :func:`distro.uname_info`. """ return self._uname_info def os_release_attr(self, attribute): """ Return a single named information item from the os-release file data source of the OS distribution. For details, see :func:`distro.os_release_attr`. """ return self._os_release_info.get(attribute, '') def lsb_release_attr(self, attribute): """ Return a single named information item from the lsb_release command output data source of the OS distribution. For details, see :func:`distro.lsb_release_attr`. """ return self._lsb_release_info.get(attribute, '') def distro_release_attr(self, attribute): """ Return a single named information item from the distro release file data source of the OS distribution. For details, see :func:`distro.distro_release_attr`. """ return self._distro_release_info.get(attribute, '') def uname_attr(self, attribute): """ Return a single named information item from the uname command output data source of the OS distribution. For details, see :func:`distro.uname_release_attr`. """ return self._uname_info.get(attribute, '') @cached_property def _os_release_info(self): """ Get the information items from the specified os-release file. Returns: A dictionary containing all information items. """ if os.path.isfile(self.os_release_file): with open(self.os_release_file) as release_file: return self._parse_os_release_content(release_file) return {} @staticmethod def _parse_os_release_content(lines): """ Parse the lines of an os-release file. Parameters: * lines: Iterable through the lines in the os-release file. Each line must be a unicode string or a UTF-8 encoded byte string. Returns: A dictionary containing all information items. """ props = {} lexer = shlex.shlex(lines, posix=True) lexer.whitespace_split = True # The shlex module defines its `wordchars` variable using literals, # making it dependent on the encoding of the Python source file. # In Python 2.6 and 2.7, the shlex source file is encoded in # 'iso-8859-1', and the `wordchars` variable is defined as a byte # string. This causes a UnicodeDecodeError to be raised when the # parsed content is a unicode object. The following fix resolves that # (... but it should be fixed in shlex...): if sys.version_info[0] == 2 and isinstance(lexer.wordchars, bytes): lexer.wordchars = lexer.wordchars.decode('iso-8859-1') tokens = list(lexer) for token in tokens: # At this point, all shell-like parsing has been done (i.e. # comments processed, quotes and backslash escape sequences # processed, multi-line values assembled, trailing newlines # stripped, etc.), so the tokens are now either: # * variable assignments: var=value # * commands or their arguments (not allowed in os-release) if '=' in token: k, v = token.split('=', 1) if isinstance(v, bytes): v = v.decode('utf-8') props[k.lower()] = v else: # Ignore any tokens that are not variable assignments pass if 'version_codename' in props: # os-release added a version_codename field. Use that in # preference to anything else Note that some distros purposefully # do not have code names. They should be setting # version_codename="" props['codename'] = props['version_codename'] elif 'ubuntu_codename' in props: # Same as above but a non-standard field name used on older Ubuntus props['codename'] = props['ubuntu_codename'] elif 'version' in props: # If there is no version_codename, parse it from the version codename = re.search(r'(\(\D+\))|,(\s+)?\D+', props['version']) if codename: codename = codename.group() codename = codename.strip('()') codename = codename.strip(',') codename = codename.strip() # codename appears within paranthese. props['codename'] = codename return props @cached_property def _lsb_release_info(self): """ Get the information items from the lsb_release command output. Returns: A dictionary containing all information items. """ if not self.include_lsb: return {} with open(os.devnull, 'w') as devnull: try: cmd = ('lsb_release', '-a') stdout = subprocess.check_output(cmd, stderr=devnull) except OSError: # Command not found return {} content = stdout.decode(sys.getfilesystemencoding()).splitlines() return self._parse_lsb_release_content(content) @staticmethod def _parse_lsb_release_content(lines): """ Parse the output of the lsb_release command. Parameters: * lines: Iterable through the lines of the lsb_release output. Each line must be a unicode string or a UTF-8 encoded byte string. Returns: A dictionary containing all information items. """ props = {} for line in lines: kv = line.strip('\n').split(':', 1) if len(kv) != 2: # Ignore lines without colon. continue k, v = kv props.update({k.replace(' ', '_').lower(): v.strip()}) return props @cached_property def _uname_info(self): with open(os.devnull, 'w') as devnull: try: cmd = ('uname', '-rs') stdout = subprocess.check_output(cmd, stderr=devnull) except OSError: return {} content = stdout.decode(sys.getfilesystemencoding()).splitlines() return self._parse_uname_content(content) @staticmethod def _parse_uname_content(lines): props = {} match = re.search(r'^([^\s]+)\s+([\d\.]+)', lines[0].strip()) if match: name, version = match.groups() # This is to prevent the Linux kernel version from # appearing as the 'best' version on otherwise # identifiable distributions. if name == 'Linux': return {} props['id'] = name.lower() props['name'] = name props['release'] = version return props @cached_property def _distro_release_info(self): """ Get the information items from the specified distro release file. Returns: A dictionary containing all information items. """ if self.distro_release_file: # If it was specified, we use it and parse what we can, even if # its file name or content does not match the expected pattern. distro_info = self._parse_distro_release_file( self.distro_release_file) basename = os.path.basename(self.distro_release_file) # The file name pattern for user-specified distro release files # is somewhat more tolerant (compared to when searching for the # file), because we want to use what was specified as best as # possible. match = _DISTRO_RELEASE_BASENAME_PATTERN.match(basename) if 'name' in distro_info \ and 'cloudlinux' in distro_info['name'].lower(): distro_info['id'] = 'cloudlinux' elif match: distro_info['id'] = match.group(1) return distro_info else: try: basenames = os.listdir(_UNIXCONFDIR) # We sort for repeatability in cases where there are multiple # distro specific files; e.g. CentOS, Oracle, Enterprise all # containing `redhat-release` on top of their own. basenames.sort() except OSError: # This may occur when /etc is not readable but we can't be # sure about the *-release files. Check common entries of # /etc for information. If they turn out to not be there the # error is handled in `_parse_distro_release_file()`. basenames = ['SuSE-release', 'arch-release', 'base-release', 'centos-release', 'fedora-release', 'gentoo-release', 'mageia-release', 'mandrake-release', 'mandriva-release', 'mandrivalinux-release', 'manjaro-release', 'oracle-release', 'redhat-release', 'sl-release', 'slackware-version'] for basename in basenames: if basename in _DISTRO_RELEASE_IGNORE_BASENAMES: continue match = _DISTRO_RELEASE_BASENAME_PATTERN.match(basename) if match: filepath = os.path.join(_UNIXCONFDIR, basename) distro_info = self._parse_distro_release_file(filepath) if 'name' in distro_info: # The name is always present if the pattern matches self.distro_release_file = filepath distro_info['id'] = match.group(1) if 'cloudlinux' in distro_info['name'].lower(): distro_info['id'] = 'cloudlinux' return distro_info return {} def _parse_distro_release_file(self, filepath): """ Parse a distro release file. Parameters: * filepath: Path name of the distro release file. Returns: A dictionary containing all information items. """ try: with open(filepath) as fp: # Only parse the first line. For instance, on SLES there # are multiple lines. We don't want them... return self._parse_distro_release_content(fp.readline()) except (OSError, IOError): # Ignore not being able to read a specific, seemingly version # related file. # See https://github.com/nir0s/distro/issues/162 return {} @staticmethod def _parse_distro_release_content(line): """ Parse a line from a distro release file. Parameters: * line: Line from the distro release file. Must be a unicode string or a UTF-8 encoded byte string. Returns: A dictionary containing all information items. """ if isinstance(line, bytes): line = line.decode('utf-8') matches = _DISTRO_RELEASE_CONTENT_REVERSED_PATTERN.match( line.strip()[::-1]) distro_info = {} if matches: # regexp ensures non-None distro_info['name'] = matches.group(3)[::-1] if matches.group(2): distro_info['version_id'] = matches.group(2)[::-1] if matches.group(1): distro_info['codename'] = matches.group(1)[::-1] elif line: distro_info['name'] = line.strip() return distro_info _distro = LinuxDistribution() def main(): logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) logger.addHandler(logging.StreamHandler(sys.stdout)) parser = argparse.ArgumentParser(description="OS distro info tool") parser.add_argument( '--json', '-j', help="Output in machine readable format", action="store_true") args = parser.parse_args() if args.json: logger.info(json.dumps(info(), indent=4, sort_keys=True)) else: logger.info('Name: %s', name(pretty=True)) distribution_version = version(pretty=True) logger.info('Version: %s', distribution_version) distribution_codename = codename() logger.info('Codename: %s', distribution_codename) if __name__ == '__main__': main() site-packages/pip/__init__.py000064400000000027151464452720012176 0ustar00__version__ = "19.3.1" site-packages/__pycache__/easy_install.cpython-38.opt-1.pyc000064400000000431151464452720017562 0ustar00U Qab~@s"dZedkrddlmZedS)zRun the EasyInstall command__main__)mainN)__doc____name__Zsetuptools.command.easy_installrrr0/usr/lib/python3.8/site-packages/easy_install.pys site-packages/__pycache__/easy_install.cpython-38.pyc000064400000000431151464452720016623 0ustar00U Qab~@s"dZedkrddlmZedS)zRun the EasyInstall command__main__)mainN)__doc____name__Zsetuptools.command.easy_installrrr0/usr/lib/python3.8/site-packages/easy_install.pys site-packages/pkg_resources/py31compat.py000064400000001056151464452720014525 0ustar00import os import errno import sys from .extern import six def _makedirs_31(path, exist_ok=False): try: os.makedirs(path) except OSError as exc: if not exist_ok or exc.errno != errno.EEXIST: raise # rely on compatibility behavior until mode considerations # and exists_ok considerations are disentangled. # See https://github.com/pypa/setuptools/pull/1083#issuecomment-315168663 needs_makedirs = ( six.PY2 or (3, 4) <= sys.version_info < (3, 4, 1) ) makedirs = _makedirs_31 if needs_makedirs else os.makedirs site-packages/pkg_resources/__pycache__/py31compat.cpython-38.pyc000064400000001130151464452720021004 0ustar00U Qab.@s`ddlZddlZddlZddlmZd ddZejpLdejkoHdknZerVenej Z dS) N)sixFc CsHzt|Wn4tk rB}z|r0|jtjkr2W5d}~XYnXdS)N)osmakedirsOSErrorerrnoZEEXIST)pathexist_okexcr s  site-packages/pkg_resources/__pycache__/__init__.cpython-38.pyc000064400000304004151464452720020551 0ustar00U Qab=G@sdZddlmZddlZddlZddlZddlZddlZddlZddl Z ddl Z ddl Z ddl Z ddl Z ddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlmZz ddlZWnek rddlZYnXze Wne!k r*e"Z YnXddl#m$Z$ddl%m&Z&m'Z'm(Z(ddlm)Z)zddlm*Z*m+Z+m,Z,d Z-Wnek rd Z-YnXdd lm.Z/dd l0m1Z1m2Z2zddl3m4Z5e5j6Wnek rdZ5YnXd dl7m8Z8ddl#m9Z9ddl#m:Z:e;de;de;de;dekr@dkrLnne?de$j@r\dZAdZBdZCdZDdZEdZFdZGdZHdZIdZJdZKdZLdZMdZNdZOdZPdZQdZRdZSGdddeTZUddZViZWddZXddZYd d!ZZd"d#Z[d$d%Z\d&d'Z]d(d)Z^d*d+Z_Z`d,d-Zad.d/d0d1d2d3d4d5d6d7d8d9d:d;dd?d@dAdBdCdDdEdFdGdHdIdJdKdLdMddNddOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrgGZbGdsdIdIecZdGdtdJdJedZeGdudvdveeZfGdwdKdKedZgGdxdLdLedZhiZidyjjej>ZkdzZld{Zmd ZndZod|Zpd}dmZqd~d0ZrgfddZsddZtddZuevdZwevdZxeuZyddRZzdd/Z{e{Z|dd1Z}dd2Z~ddd3Zdd4ZGdd`d`ZGddadaeZGddDdDZGdddeZGddCdCZeZGddMdMe?ZGddEdEZddBZddOZddPZddUZddVZddWZdddXZGddgdgZeqeeGddhdheZGddidieZeGddedeeZeZGdddeZGdddeZGddjdjeZeqe jeGddbdbeZGddcdceZGdddddeZeXdidddkZddd?ZdddZee jedddZeeeddZdddZddZGdddZddZddZddZddZeejeee5dree5jeeXdideXdidddlZddńZddDŽZdd<ZdddnZdd˄Zeejeee jeee5dr,ee5jedd̈́ZeeeddZZddЄZifdd҄ZddԄZddքZdd؄ZddSZevdڡjZevdejejBjZGddHdHZddބZddZGddFdFZGdddeăZGdddeăZeeedZddZGdddeɃZddNZGddGdGe:jj̓ZddZddZddYZddZddTZddZe jdeUd dddZeeփfddZeddZGddpdpeكZdS(aZ Package resource API -------------------- A resource is a logical file contained within a package, or a logical subdirectory thereof. The package resource API expects resource names to have their path parts separated with ``/``, *not* whatever the local path separator is. Do not use os.path operations to manipulate resource names being passed into the API. The package resource API is designed to work with normal filesystem packages, .egg files, and unpacked .egg files. It can also work in a limited way with .zip files and with custom PEP 302 loaders that support the ``get_data()`` method. )absolute_importN) get_importer)six)urllibmapfilter)utime)mkdirrenameunlinkTF)open)isdirsplit) py31compat)appdirs) packagingz&pkg_resources.extern.packaging.versionz)pkg_resources.extern.packaging.specifiersz+pkg_resources.extern.packaging.requirementsz&pkg_resources.extern.packaging.markers)r)rzPython 3.4 or later is requiredc@seZdZdZdS) PEP440Warningza Used when there is an issue with a version or specifier not complying with PEP 440. N__name__ __module__ __qualname____doc__rr:/usr/lib/python3.8/site-packages/pkg_resources/__init__.pyrxsrcCs8ztj|WStjjk r2tj|YSXdSN)rversionVersionInvalidVersion LegacyVersion)vrrr parse_versionsr#cKs"t|tt||dSr)globalsupdate _state_varsdictfromkeys)Zvartypekwrrr_declare_states r*cCs8i}t}tD] \}}|d|||||<q|S)NZ_sget_)r$r&itemsstategkr"rrr __getstate__s r0cCs8t}|D]$\}}|dt|||||q|S)NZ_sset_)r$r+r&r,rrr __setstate__sr1cCs|Sr)copyvalrrr _sget_dictsr5cCs|||dSr)clearr%keyobr-rrr _sset_dictsr:cCs|Sr)r0r3rrr _sget_objectsr;cCs||dSr)r1r7rrr _sset_objectsr<cGsdSrrargsrrrr?cCsbt}t|}|dk r^tjdkr^z&ddtdd|df}Wntk r\YnX|S)aZReturn this platform's maximum compatible version. distutils.util.get_platform() normally reports the minimum version of Mac OS X that would be required to *use* extensions produced by distutils. But what we want when checking compatibility is to know the version of Mac OS X that we are *running*. To allow usage of packages that explicitly require a newer version of Mac OS X, we must also know the current version of the OS. If this condition occurs for any other platform with a version in its platform strings, this function should be extended accordingly. Ndarwinz macosx-%s-%s.r) get_build_platformmacosVersionStringmatchsysplatformjoin _macosx_versgroup ValueError)platmrrrget_supported_platforms  &rOrequire run_script get_providerget_distributionload_entry_point get_entry_mapget_entry_infoiter_entry_pointsresource_stringresource_streamresource_filenameresource_listdirresource_existsresource_isdirdeclare_namespace working_setadd_activation_listenerfind_distributionsset_extraction_pathcleanup_resourcesget_default_cache Environment WorkingSetResourceManager Distribution Requirement EntryPointResolutionErrorVersionConflictDistributionNotFound UnknownExtraExtractionErrorparse_requirements safe_name safe_version get_platformcompatible_platforms yield_linessplit_sections safe_extra to_filenameinvalid_markerevaluate_markerensure_directorynormalize_pathEGG_DIST BINARY_DIST SOURCE_DIST CHECKOUT_DIST DEVELOP_DISTIMetadataProviderIResourceProvider FileMetadata PathMetadata EggMetadata EmptyProviderempty_provider NullProvider EggProviderDefaultProvider ZipProviderregister_finderregister_namespace_handlerregister_loader_typefixup_namespace_packagesrPkgResourcesDeprecationWarningrun_mainAvailableDistributionsc@seZdZdZddZdS)rkz.Abstract base for dependency resolution errorscCs|jjt|jSr) __class__rreprr>selfrrr__repr__szResolutionError.__repr__N)rrrrrrrrrrksc@s<eZdZdZdZeddZeddZddZd d Z d S) rlz An already-installed version conflicts with the requested version. Should be initialized with the installed Distribution and the requested Requirement. z3{self.dist} is installed but {self.req} is requiredcCs |jdSNrr=rrrrdist szVersionConflict.distcCs |jdSNrr=rrrrreqszVersionConflict.reqcCs|jjftSr _templateformatlocalsrrrrreportszVersionConflict.reportcCs|s|S|j|f}t|S)zt If required_by is non-empty, return a version of self that is a ContextualVersionConflict. )r>ContextualVersionConflict)r required_byr>rrr with_contexts zVersionConflict.with_contextN) rrrrrpropertyrrrrrrrrrls  c@s&eZdZdZejdZeddZdS)rz A VersionConflict that accepts a third parameter, the set of the requirements that required the installed Distribution. z by {self.required_by}cCs |jdS)NrCr=rrrrr*sz%ContextualVersionConflict.required_byN)rrrrrlrrrrrrrr"s rc@sHeZdZdZdZeddZeddZeddZd d Z d d Z d S)rmz&A requested distribution was not foundzSThe '{self.req}' distribution was not found and is required by {self.requirers_str}cCs |jdSrr=rrrrr5szDistributionNotFound.reqcCs |jdSrr=rrrr requirers9szDistributionNotFound.requirerscCs|js dSd|jS)Nzthe applicationz, )rrIrrrr requirers_str=sz"DistributionNotFound.requirers_strcCs|jjftSrrrrrrrCszDistributionNotFound.reportcCs|Sr)rrrrr__str__FszDistributionNotFound.__str__N) rrrrrrrrrrrrrrrrm/s   c@seZdZdZdS)rnz>Distribution doesn't have an "extra feature" of the given nameNrrrrrrnJsz{}.{}rrCcCs |t|<dS)aRegister `provider_factory` to make providers for `loader_type` `loader_type` is the type or class of a PEP 302 ``module.__loader__``, and `provider_factory` is a function that, passed a *module* object, returns an ``IResourceProvider`` for that module. N)_provider_factories)Z loader_typeZprovider_factoryrrrrXscCstt|tr$t|p"tt|dSztj|}Wn&tk rXt |tj|}YnXt |dd}t t ||S)z?Return an IResourceProvider for the named module or requirementr __loader__N) isinstancerir_findrPstrrGmodulesKeyError __import__getattr _find_adapterr)Z moduleOrReqmoduleloaderrrrrRbs  cCsd|s\td}|dkrLd}tj|rLttdrLt|}d|krL|d}|| d|dS)Nrz0/System/Library/CoreServices/SystemVersion.plist readPlistZProductVersionrB) rHZmac_verospathexistshasattrplistlibrappendr)_cacherZplistZ plist_contentrrrrJos    rJcCsddd||S)NZppc)ZPowerPCZPower_Macintosh)get)machinerrr _macosx_archsrcCs~ddlm}|}tjdkrz|dszz>t}tddd}dt |dt |d t |fWSt k rxYnX|S) zReturn this platform's string for platform-specific distributions XXX Currently this is the same as ``distutils.util.get_platform()``, but it needs some hacks for Linux and Mac OS X. r)rsrAzmacosx-r _zmacosx-%d.%d-%sr) Z sysconfigrsrGrH startswithrJrunamereplaceintrrL)rsrMrrrrrrDs  rDzmacosx-(\d+)\.(\d+)-(.*)zdarwin-(\d+)\.(\d+)\.(\d+)-(.*)cCs|dks|dks||krdSt|}|rt|}|st|}|rt|d}d|d|df}|dkr||dks|dkr|d krdSd S|d|dks|d |d krd St|dt|dkrd SdSd S) zCan code for the `provided` platform run on the `required` platform? Returns true if either platform is ``None``, or the platforms are equal. XXX Needs compatibility checks for Linux and other unixy OSes. NTrz%s.%srCz10.3z10.4Fr)rErFdarwinVersionStringrrK)ZprovidedZrequiredZreqMacZprovMacZ provDarwinZdversionZ macosversionrrrrts2   cCs<tdj}|d}|||d<t|d||dS)z@Locate distribution `dist_spec` and run its `script_name` scriptrrrNrG _getframe f_globalsr6rPrQ)Z dist_spec script_namensnamerrrrQs  cCs@t|tjrt|}t|tr(t|}t|tsIs the named metadata a directory? (like ``os.path.isdir()``)Nrrrrrmetadata_isdirsz IMetadataProvider.metadata_isdircCsdS)z?List of metadata names in the directory (like ``os.listdir()``)Nrrrrrmetadata_listdirsz"IMetadataProvider.metadata_listdircCsdS)z=Execute the named script in the supplied namespace dictionaryNr)r namespacerrrrQ szIMetadataProvider.run_scriptN) rrrrrrrrrQrrrrrs c@s@eZdZdZddZddZddZdd Zd d Zd d Z dS)rz3An object that provides access to package resourcescCsdS)zdReturn a true filesystem path for `resource_name` `manager` must be an ``IResourceManager``Nrmanager resource_namerrrget_resource_filenamesz'IResourceProvider.get_resource_filenamecCsdS)ziReturn a readable file-like object for `resource_name` `manager` must be an ``IResourceManager``Nrrrrrget_resource_streamsz%IResourceProvider.get_resource_streamcCsdS)zmReturn a string containing the contents of `resource_name` `manager` must be an ``IResourceManager``Nrrrrrget_resource_stringsz%IResourceProvider.get_resource_stringcCsdS)z,Does the package contain the named resource?Nrrrrr has_resourceszIResourceProvider.has_resourcecCsdS)z>Is the named resource a directory? (like ``os.path.isdir()``)Nrrrrrr]"sz IResourceProvider.resource_isdircCsdS)z?List of resource names in the directory (like ``os.listdir()``)Nrrrrrr[%sz"IResourceProvider.resource_listdirN) rrrrrrrrr]r[rrrrr sc@seZdZdZd'ddZeddZeddZd d Zd d Z d dZ d(ddZ ddZ ddZ d)ddZd*ddZd+ddZddZd,dd Zd!d"Zd#d$Zd%d&ZdS)-rfzDA collection of active distributions on sys.path (or a similar list)NcCs>g|_i|_i|_g|_|dkr&tj}|D]}||q*dS)z?Create working set from list of path entries (default=sys.path)N)entries entry_keysby_key callbacksrGr add_entry)rrentryrrr__init__,szWorkingSet.__init__cCsb|}zddlm}Wntk r.|YSXz||Wntk r\||YSX|S)z1 Prepare the master working set. r) __requires__)__main__r ImportErrorrPrl_build_from_requirements)clswsrrrr _build_master9s zWorkingSet._build_mastercCsf|g}t|}||t}|D]}||q"tjD]}||jkr8||q8|jtjdd<|S)zQ Build a working set from a requirement spec. Rewrites sys.path. N)rpresolvereaddrGrrr)rZreq_specrreqsdistsrrrrrrMs    z#WorkingSet._build_from_requirementscCs<|j|g|j|t|dD]}|||dq$dS)aAdd a path item to ``.entries``, finding any distributions on it ``find_distributions(entry, True)`` is used to find distributions corresponding to the path entry, and they are added. `entry` is always appended to ``.entries``, even if it is already present. (This is because ``sys.path`` can contain the same value more than once, and the ``.entries`` of the ``sys.path`` WorkingSet should always equal ``sys.path``.) TFN)r setdefaultrrrar)rrrrrrrcs  zWorkingSet.add_entrycCs|j|j|kS)z9True if `dist` is the active distribution for its project)rrr8rrrrr __contains__rszWorkingSet.__contains__cCs,|j|j}|dk r(||kr(t|||S)aFind a distribution matching requirement `req` If there is an active distribution for the requested project, this returns it as long as it meets the version requirement specified by `req`. But, if there is an active distribution for the project and it does *not* meet the `req` requirement, ``VersionConflict`` is raised. If there is no active distribution for the requested project, ``None`` is returned. N)rrr8rl)rrrrrrrvs  zWorkingSet.findcsfdd|DS)aYield entry point objects from `group` matching `name` If `name` is None, yields all entry points in `group` from all distributions in the working set, otherwise only ones matching both `group` and `name` are yielded (in distribution order). c3s8|]0}|D]}dks*|jkr|VqqdSr)rUvaluesr).0rrrKrrr s  z/WorkingSet.iter_entry_points..rrrKrrrrrWs zWorkingSet.iter_entry_pointscCs>tdj}|d}|||d<||d||dS)z?Locate distribution for `requires` and run `script_name` scriptrrrNr)rrequiresrrrrrrrQs  zWorkingSet.run_scriptccsLi}|jD]<}||jkrq |j|D] }||kr$d||<|j|Vq$q dS)zYield distributions for non-duplicate projects in the working set The yield order is the order in which the items' path entries were added to the working set. rN)rrr)rseenitemr8rrr__iter__s  zWorkingSet.__iter__TFcCs|r|j|j||d|dkr$|j}|j|g}|j|jg}|sV|j|jkrVdS||j|j<|j|krx||j|j|kr||j||dS)aAdd `dist` to working set, associated with `entry` If `entry` is unspecified, it defaults to the ``.location`` of `dist`. On exit from this routine, `entry` is added to the end of the working set's ``.entries`` (if it wasn't already present). `dist` is only added to the working set if it's for a project that doesn't already have a distribution in the set, unless `replace=True`. If it's added, any callbacks registered with the ``subscribe()`` method will be called. rN) insert_onrlocationrrr8rr _added_new)rrrinsertrkeysZkeys2rrrrs      zWorkingSet.addcCsxt|ddd}i}i}g}t} tt} |rt|d} | |krHq.| | |sVq.|| j} | dkr|j | j} | dks| | kr|r|} |dkr| dkrt |j }nt g}t g} |j | | ||d} || j<| dkr| | d}t| ||| | | kr$| | }t| | || | jddd}|||D] }| || j| j| |<qHd|| <q.|S)aList all distributions needed to (recursively) meet `requirements` `requirements` must be a sequence of ``Requirement`` objects. `env`, if supplied, should be an ``Environment`` instance. If not supplied, it defaults to all distributions available within any entry or distribution in the working set. `installer`, if supplied, will be invoked with each requirement that cannot be met by an already-installed distribution; it should return a ``Distribution`` or ``None``. Unless `replace_conflicting=True`, raises a VersionConflict exception if any requirements are found on the path that have the correct name but the wrong version. Otherwise, if an `installer` is supplied it will be invoked to obtain the correct version of the requirement and activate it. `extras` is a list of the extras to be used with these requirements. This is important because extra requirements may look like `my_req; extra = "my_extra"`, which would otherwise be interpreted as a purely optional requirement. Instead, we want to be able to assert that these requirements are truly required. Nrr)replace_conflictingT)list _ReqExtras collections defaultdictsetpop markers_passrr8rrerrf best_matchrmrrlrrextrasextendr project_name)r requirementsenv installerr rZ processedZbestZ to_activateZ req_extrasrrrrrZ dependent_reqZnew_requirementsZnew_requirementrrrrsT            zWorkingSet.resolvec Cs t|}|i}i}|dkr4t|j}||7}n||}|g} tt| j||D]} || D]} | g} z| | ||} WnBt k r}z$||| <|rWYqfn WYqZW5d}~XYqfXtt| j| | t | qZqfqZt|}|||fS)asFind all activatable distributions in `plugin_env` Example usage:: distributions, errors = working_set.find_plugins( Environment(plugin_dirlist) ) # add plugins+libs to sys.path map(working_set.add, distributions) # display errors print('Could not load', errors) The `plugin_env` should be an ``Environment`` instance that contains only distributions that are in the project's "plugin directory" or directories. The `full_env`, if supplied, should be an ``Environment`` contains all currently-available distributions. If `full_env` is not supplied, one is created automatically from the ``WorkingSet`` this method is called on, which will typically mean that every directory on ``sys.path`` will be scanned for distributions. `installer` is a standard installer callback as used by the ``resolve()`` method. The `fallback` flag indicates whether we should attempt to resolve older versions of a plugin if the newest version cannot be resolved. This method returns a 2-tuple: (`distributions`, `error_info`), where `distributions` is a list of the distributions found in `plugin_env` that were loadable, along with any other distributions that are needed to resolve their dependencies. `error_info` is a dictionary mapping unloadable plugin distributions to an exception instance describing the error that occurred. Usually this will be a ``DistributionNotFound`` or ``VersionConflict`` instance. N) r sortrerrrras_requirementrrkr%r'r()rZ plugin_envZfull_envrZfallbackZplugin_projectsZ error_infoZ distributionsrZ shadow_setrrrZ resolveesr"rrr find_plugins's4$      zWorkingSet.find_pluginscGs&|t|}|D]}||q|S)aEnsure that distributions matching `requirements` are activated `requirements` must be a string or a (possibly-nested) sequence thereof, specifying the distributions and versions required. The return value is a sequence of the distributions that needed to be activated to fulfill the requirements; all relevant distributions are included, even if they were already activated in this working set. )rrpr)rrZneededrrrrrP{s  zWorkingSet.requirecCs8||jkrdS|j||s"dS|D] }||q&dS)zInvoke `callback` for all distributions If `existing=True` (default), call on all existing ones, as well. N)rr)rcallbackexistingrrrr subscribes  zWorkingSet.subscribecCs|jD] }||qdSr)r)rrrrrrrs zWorkingSet._added_newcCs,|jdd|j|j|jddfSr)rrr2rrrrrrr0s   zWorkingSet.__getstate__cCs@|\}}}}|dd|_||_||_|dd|_dSr)rr2rrr)rZe_k_b_crr rrrrrr1s    zWorkingSet.__setstate__)N)N)NTF)NNFN)NNT)T)rrrrr classmethodrrrrrrWrQrrrrrPrrr0r1rrrrrf)s4      ] T c@seZdZdZdddZdS)r z> Map each requirement to the extras that demanded it. Ncs2fdd|d|pdD}j p0t|S)z Evaluate markers for req against each extra that demanded it. Return False if the req has a marker and fails evaluation. Otherwise, return True. c3s|]}jd|iVqdS)extraNmarkerevaluate)rr!rrrrsz*_ReqExtras.markers_pass..rr)rr#any)rrrZ extra_evalsrr%rrs z_ReqExtras.markers_pass)N)rrrrrrrrrr sr c@sxeZdZdZdeefddZddZddZdd d Z d d Z d dZ dddZ dddZ ddZddZddZdS)rez5Searchable snapshot of distributions on a search pathNcCs i|_||_||_||dS)a!Snapshot distributions available on a search path Any distributions found on `search_path` are added to the environment. `search_path` should be a sequence of ``sys.path`` items. If not supplied, ``sys.path`` is used. `platform` is an optional string specifying the name of the platform that platform-specific distributions must be compatible with. If unspecified, it defaults to the current platform. `python` is an optional string naming the desired version of Python (e.g. ``'3.6'``); it defaults to the current version. You may explicitly set `platform` (and/or `python`) to ``None`` if you wish to map *all* distributions, not just those compatible with the running platform or Python version. N)_distmaprHpythonscan)r search_pathrHr(rrrrszEnvironment.__init__cCs2|jdkp|jdkp|j|jk}|o0t|j|jS)zIs distribution `dist` acceptable for this environment? The distribution must match the platform and python version requirements specified when this environment was created, or False is returned. N)r( py_versionrtrH)rrZ py_compatrrrcan_adds   zEnvironment.can_addcCs|j|j|dS)z"Remove `dist` from the environmentN)r'r8removerrrrr-szEnvironment.removecCs4|dkrtj}|D]}t|D]}||qqdS)adScan `search_path` for distributions usable in this environment Any distributions found are added to the environment. `search_path` should be a sequence of ``sys.path`` items. If not supplied, ``sys.path`` is used. Only distributions conforming to the platform/python version defined at initialization are added. N)rGrrar)rr*rrrrrr)s  zEnvironment.scancCs|}|j|gS)aReturn a newest-to-oldest list of distributions for `project_name` Uses case-insensitive `project_name` comparison, assuming all the project's distributions use their project's name converted to all lowercase as their key. )lowerr'r)rrZdistribution_keyrrr __getitem__szEnvironment.__getitem__cCsL||rH|rH|j|jg}||krH|||jtddddS)zLAdd `dist` if we ``can_add()`` it and it has not already been added hashcmpTr8reverseN) r, has_versionr'rr8rroperator attrgetter)rrrrrrrs  zEnvironment.addFcCsfz||}Wntk r,|s$d}YnX|dk r:|S||jD]}||krD|SqD|||S)aFind distribution best matching `req` and usable on `working_set` This calls the ``find(req)`` method of the `working_set` to see if a suitable distribution is already active. (This may raise ``VersionConflict`` if an unsuitable version of the project is already active in the specified `working_set`.) If a suitable distribution isn't active, this method returns the newest distribution in the environment that meets the ``Requirement`` in `req`. If no suitable distribution is found, and `installer` is supplied, then the result of calling the environment's ``obtain(req, installer)`` method will be returned. N)rrlr8obtain)rrr_rr rrrrrs  zEnvironment.best_matchcCs|dk r||SdS)aObtain a distribution matching `requirement` (e.g. via download) Obtain a distro that matches requirement (e.g. via download). In the base ``Environment`` class, this routine just returns ``installer(requirement)``, unless `installer` is None, in which case None is returned instead. This method is a hook that allows subclasses to attempt other ways of obtaining a distribution before falling back to the `installer` argument.Nr)rZ requirementrrrrr6+s zEnvironment.obtainccs"|jD]}||r |Vq dS)z=Yield the unique project names of the available distributionsN)r'r rr8rrrr7szEnvironment.__iter__cCsVt|tr||n)rZ archive_namenamesZ extract_pathZ target_pathrrrget_cache_paths    zResourceManager.get_cache_pathcCsVtjdkr|tjdsdSt|j}|tj@s>|tj@rRd|}t |t dS)aN If the default extraction path is overridden and set to an insecure location, such as /tmp, it opens up an opportunity for an attacker to replace an extracted file with an unauthorized payload. Warn the user if a known insecure location is used. See Distribute #375 for more details. ntZwindirNz%s is writable by group/others and vulnerable to attack when used with get_resource_filename. Consider a more secure location (set with .set_extraction_path or the PYTHON_EGG_CACHE environment variable).) rrrenvironstatst_modeS_IWOTHS_IWGRPwarningswarn UserWarning)rmodemsgrrrrLs  z,ResourceManager._warn_unsafe_extraction_pathcCs.tjdkr*t|jdBd@}t||dS)a4Perform any platform-specific postprocessing of `tempname` This is where Mac header rewrites should be done; other platforms don't have anything special they should do. Resource providers should call this method ONLY after successfully extracting a compressed resource. They must NOT call it on resources that are already in the filesystem. `tempname` is the current (temporary) name of the file, and `filename` is the name it will be renamed to by the caller after this routine returns. posiximiN)rrrQrRchmod)rZtempnamefilenamerXrrr postprocesss zResourceManager.postprocesscCs|jrtd||_dS)aSet the base path where resources will be extracted to, if needed. If you do not call this routine before any extractions take place, the path defaults to the return value of ``get_default_cache()``. (Which is based on the ``PYTHON_EGG_CACHE`` environment variable, with various platform-specific fallbacks. See that routine's documentation for more details.) Resources are extracted to subdirectories of this path based upon information given by the ``IResourceProvider``. You may set this to a temporary directory, but then you must call ``cleanup_resources()`` to delete the extracted files when done. There is no guarantee that ``cleanup_resources()`` will be able to remove all extracted files. (Note: you may not change the extraction path for a given resource manager once resources have been extracted, unless you first call ``cleanup_resources()``.) z5Can't change extraction path, files already extractedN)r>rLrArrrrrrbs z#ResourceManager.set_extraction_pathFcCsdS)aB Delete all extracted resource files and directories, returning a list of the file and directory names that could not be successfully removed. This function does not have any concurrency protection, so it should generally only be called when the extraction path is a temporary directory exclusive to a single process. This method is not automatically called; you must call it explicitly or register it as an ``atexit`` function if you wish to ensure cleanup of a temporary directory used for extractions. Nr)rforcerrrrc sz!ResourceManager.cleanup_resources)r)F)rrrrrArr\r]rZrYrXr[rIrN staticmethodrLr]rbrcrrrrrges   cCstjdptjddS)z Return the ``PYTHON_EGG_CACHE`` environment variable or a platform-relevant user cache dir for an app named "Python-Eggs". ZPYTHON_EGG_CACHEz Python-Eggs)Zappname)rrPrrZuser_cache_dirrrrrrds  cCstdd|S)zConvert an arbitrary string to a standard distribution name Any runs of non-alphanumeric/. characters are replaced with a single '-'. [^A-Za-z0-9.]+-)resubrrrrrq%scCsJzttj|WStjjk rD|dd}tdd|YSXdS)zB Convert an arbitrary string to a standard version string rrBrarbN)rrrrr rrcrd)rrrrrr-s  cCstdd|S)zConvert an arbitrary string to a standard 'extra' name Any runs of non-alphanumeric characters are replaced with a single '_', and the result is always lowercased. z[^A-Za-z0-9.-]+r)rcrdr.)r!rrrrw9scCs |ddS)z|Convert a project or version name to its filename-escaped form Any '-' characters are currently replaced with '_'. rbrrrrrrrxBsc CsHz t|Wn6tk rB}zd|_d|_|WYSd}~XYnXdS)zo Validate text as a PEP 508 environment marker; return an exception if invalid or False otherwise. NF)rz SyntaxErrorr\lineno)texterrrryJs c CsJztj|}|WStjjk rD}z t|W5d}~XYnXdS)z Evaluate a PEP 508 environment marker. Return a boolean indicating the marker result in this environment. Raise SyntaxError if marker is invalid. This implementation uses the 'pyparsing' module. N)rZmarkersZMarkerr$Z InvalidMarkerre)rgr!r#rhrrrrzXs   c@seZdZdZdZdZdZddZddZddZ d d Z d d Z d dZ ddZ ddZddZddZddZddZddZddZdd Zd!d"Zd#d$Zd%d&Zed'd(Zd)d*ZdS)+rzETry to implement resources and metadata for arbitrary PEP 302 loadersNcCs(t|dd|_tjt|dd|_dS)Nr__file__r)rrrrdirname module_pathrrrrrrnszNullProvider.__init__cCs||j|Sr)_fnrkrrrrrrrrsz"NullProvider.get_resource_filenamecCst|||Sr)ioBytesIOrrnrrrrusz NullProvider.get_resource_streamcCs|||j|Sr)_getrmrkrnrrrrxsz NullProvider.get_resource_stringcCs|||j|Sr)_hasrmrkrrrrrr{szNullProvider.has_resourcecCs||j|Sr)rmegg_inforrrrr_get_metadata_path~szNullProvider._get_metadata_pathcCs |js |jS||}||Sr)rtrvrrrrrrrrrs zNullProvider.has_metadatac Cst|js dS||}||}tjr(|Sz |dWStk rn}z|jd||7_W5d}~XYnXdS)Nrutf-8z in {} file at path: {}) rtrvrqrPY2decodeUnicodeDecodeErrorreasonr)rrrvalueexcrrrrs   zNullProvider.get_metadatacCst||SrrurrurrrrszNullProvider.get_metadata_linescCs|||j|Sr)_isdirrmrkrsrrrr]szNullProvider.resource_isdircCs|jo|||j|Sr)rtrrmrurrrrszNullProvider.metadata_isdircCs|||j|Sr)_listdirrmrkrsrrrr[szNullProvider.resource_listdircCs|jr|||j|SgSr)rtrrmrurrrrszNullProvider.metadata_listdirc Csd|}||s$tdjft||dd}|dd}||j|}||d<tj |rt | }t ||d}t|||n>dd lm}t|d|d|f||<t ||d} t| ||dS) Nzscripts/z